README (i.e. for the server: exportfs, mountd, statd, nfsd,
sm-notify; for the client: statd / sm-notify before mountall). This
is important to allow locking to work correctly.
svn path=/nixos/branches/boot-order/; revision=22196
When starting multiple VMs, some will have perfectly synchronised
clocks, while others will have their clocks run much slower (say, a
factor of 5).
svn path=/nixos/branches/boot-order/; revision=22195
a daemon (it just starts some kernel threads). In the post-stop
script, stop the kernel threads.
* exportfs: fix the createMountPoints option.
* Mount the nfsd filesystem on /proc/fs/nfsd because mountd prefers
this.
svn path=/nixos/branches/boot-order/; revision=22187
swapfiles cannot be unmounted or even remounted read-only.
* In the remount, pass `-t none' to get a more informative error
message if the filesystem is in use.
svn path=/nixos/branches/boot-order/; revision=22179
it allows the `mountall' task to start mounting filesystems as soon
as udev is running and devices become available (i.e. it doesn't
have to wait for all devices). This means that some jobs should
depend on "stopped udevtrigger" instead of "started udev".
svn path=/nixos/branches/boot-order/; revision=22167
during boot. Mountall ensures that these are done in the right
order. It's informed by udev about devices becoming available. It
emits some Upstart events upon reaching certain states, in
particular ‘local-filesystems’ after all local filesystems have been
mounted successfully, ‘remote-filesystems’ after all network
filesystems have been mounted, and ‘filesystem’ (sic) when all
filesystems have been mounted.
Currently, if a filesystem fails to mount or doesn't exist, then the
mingettys won't start and the boot will appear to hang. This is
because mountall doesn't emit an event for failing filesystems and
waits indefinitely for the filesystems to become available.
* The ‘filesystems’ and ‘swap’ Upstart jobs are gone. (Support for
encrypted swap devices is temporarily gone.)
* Generate a proper /etc/fstab from the ‘fileSystems’ and
‘swapDevices’ options.
svn path=/nixos/branches/boot-order/; revision=22148
`su'.
* The `usermod' from `shadow' allows setting a supplementary group
equal to the user's primary group, so the special hack for the
`nixbld' group is no longer needed.
* Removed /etc/default/passwd since it's not used by the new passwd.
The hash is configured in pam_unix.
* Move some values for `security.setuidPrograms' and
`security.pam.services' to the appropriate modules.
svn path=/nixos/trunk/; revision=22107
This ensures that they're gone by the time the shutdown job runs, so
it doesn't have to stop them itself.
* Don't respawn tasks, as it doesn't seem useful (if they fail they're
likely to fail again if they're restarted).
svn path=/nixos/trunk/; revision=22099
to use the standard (coreutils) tools.
* Use util-linux's `switch_root' to switch over to the target root
FS. It automatically moves over the /dev, /proc and /sys from stage
1, so stage 2 doesn't need to set them up again.
svn path=/nixos/trunk/; revision=22085
or Google Earth) on 64-bit NixOS on NVIDIA hardware. The 32-bit
OpenGL library is symlinked from /var/run/opengl-driver-32, which is
added to the LD_LIBRARY_PATH so that 32-bit binaries can find it.
svn path=/nixos/trunk/; revision=22062
interface name through the derived option networking.ifaces. This
makes it easier to get information about specific interfaces
(e.g. `nodes.router.config.networking.ifaces.eth2.ipAddress').
Really networking.interfaces should be an attribute set.
svn path=/nixos/trunk/; revision=21938
machine can now declare an option `virtualisation.vlans' that causes
it to have network interfaces connected to each listed virtual
network. For instance,
virtualisation.vlans = [ 1 2 ];
causes the machine to have two interfaces (in addition to eth0, used
by the test driver to control the machine): eth1 connected to
network 1 with IP address 192.168.1.<i>, and eth2 connected to
network 2 with address 192.168.2.<i> (where <i> is the index of the
machine in the `nodes' attribute set). On the other hand,
virtualisation.vlans = [ 2 ];
causes the machine to only have an eth1 connected to network 2 with
address 192.168.2.<i>. So each virtual network <n> is assigned the
IP range 192.168.<n>.0/24.
Each virtual network is implemented using a separate multicast
address on the host, so guests really cannot talk to networks to
which they are not connected.
* Added a simple NAT test to demonstrate this.
* Added an option `virtualisation.qemu.options' to specify QEMU
command-line options. Used to factor out some commonality between
the test driver script and the interactive test script.
svn path=/nixos/trunk/; revision=21928