]> @LXC_GENERATE_DATE@ lxc.conf 5 lxc.conf linux container configuration file Description The linux containers (lxc) are always created before being used. This creation defines a set of system resources to be virtualized / isolated when a process is using the container. By default, the pids, sysv ipc and mount points are virtualized and isolated. The other system resources are shared across containers, until they are explicitly defined in the configuration file. For example, if there is no network configuration, the network will be shared between the creator of the container and the container itself, but if the network is specified, a new network stack is created for the container and the container can no longer use the network of its ancestor. The configuration file defines the different system resources to be assigned for the container. At present, the utsname, the network, the mount points, the root file system and the control groups are supported. Each option in the configuration file has the form key = value fitting in one line. The '#' character means the line is a comment. Architecture Allows to set the architecture for the container. For example, set a 32bits architecture for a container running 32bits binaries on a 64bits host. That fix the container scripts which rely on the architecture to do some work like downloading the packages. Specify the architecture for the container. Valid options are , , , Hostname The utsname section defines the hostname to be set for the container. That means the container can set its own hostname without changing the one from the system. That makes the hostname private for the container. specify the hostname for the container Network The network section defines how the network is virtualized in the container. The network virtualization acts at layer two. In order to use the network virtualization, parameters must be specified to define the network interfaces of the container. Several virtual interfaces can be assigned and used in a container even if the system has only one physical network interface. specify what kind of network virtualization to be used for the container. Each time a field is found a new round of network configuration begins. In this way, several network virtualization types can be specified for the same container, as well as assigning several network interfaces for one container. The different virtualization types can be: will create only the loopback interface. a peer network device is created with one side assigned to the container and the other side is attached to a bridge specified by the . If the bridge is not specified, then the veth pair device will be created but not attached to any bridge. Otherwise, the bridge has to be setup before on the system, lxc won't handle any configuration outside of the container. By default lxc choose a name for the network device belonging to the outside of the container, this name is handled by lxc, but if you wish to handle this name yourself, you can tell lxc to set a specific name with the option. a vlan interface is linked with the interface specified by the and assigned to the container. The vlan identifier is specified with the option . a macvlan interface is linked with the interface specified by the and assigned to the container. specifies the mode the macvlan will use to communicate between different macvlan on the same upper device. The accepted modes are , the device never communicates with any other device on the same upper_dev (default), , the new Virtual Ethernet Port Aggregator (VEPA) mode, it assumes that the adjacent bridge returns all frames where both source and destination are local to the macvlan port, i.e. the bridge is set up as a reflective relay. Broadcast frames coming in from the upper_dev get flooded to all macvlan interfaces in VEPA mode, local frames are not delivered locallay, or , it provides the behavior of a simple bridge between different macvlan interfaces on the same port. Frames from one interface to another one get delivered directly and are not sent out externally. Broadcast frames get flooded to all other bridge ports and to the external interface, but when they come back from a reflective relay, we don't deliver them again. Since we know all the MAC addresses, the macvlan bridge mode does not require learning or STP like the bridge module does. an already existing interface specified by the is assigned to the container. specify an action to do for the network. activates the interface. specify the interface to be used for real network traffic. the interface name is dynamically allocated, but if another name is needed because the configuration files being used by the container use a generic name, eg. eth0, this option will rename the interface in the container. the interface mac address is dynamically allocated by default to the virtual interface, but in some cases, this is needed to resolve a mac address conflict or to always have the same link-local ipv6 address specify the ipv4 address to assign to the virtualized interface. Several lines specify several ipv4 addresses. The address is in format x.y.z.t/m, eg. 192.168.1.123/24. The broadcast address should be specified on the same line, right after the ipv4 address. specify the ipv4 address to use as the gateway inside the container. The address is in format x.y.z.t, eg. 192.168.1.123. Can also have the special value , which means to take the primary address from the bridge interface (as specified by the option) and use that as the gateway. is only available when using the and network types. specify the ipv6 address to assign to the virtualized interface. Several lines specify several ipv6 addresses. The address is in format x::y/m, eg. 2003:db8:1:0:214:1234:fe0b:3596/64 specify the ipv6 address to use as the gateway inside the container. The address is in format x::y, eg. 2003:db8:1:0::1 Can also have the special value , which means to take the primary address from the bridge interface (as specified by the option) and use that as the gateway. is only available when using the and network types. add a configuration option to specify a script to be executed after creating and configuring the network used from the host side. The following arguments are passed to the script: container name and config section name (net) Additional arguments depend on the config section employing a script hook; the following are used by the network system: execution context (up), network type (empty/veth/macvlan/phys), Depending on the network type, other arguments may be passed: veth/macvlan/phys. And finally (host-sided) device name. add a configuration option to specify a script to be executed before destroying the network used from the host side. The following arguments are passed to the script: container name and config section name (net) Additional arguments depend on the config section employing a script hook; the following are used by the network system: execution context (down), network type (empty/veth/macvlan/phys), Depending on the network type, other arguments may be passed: veth/macvlan/phys. And finally (host-sided) device name. New pseudo tty instance (devpts) For stricter isolation the container can have its own private instance of the pseudo tty. If set, the container will have a new pseudo tty instance, making this private to it. The value specifies the maximum number of pseudo ttys allowed for a pts instance (this limitation is not implemented yet). Container system console If the container is configured with a root filesystem and the inittab file is setup to use the console, you may want to specify where goes the output of this console. Specify a path to a file where the console output will be written. The keyword 'none' will simply disable the console. This is dangerous once if have a rootfs with a console device file where the application can write, the messages will fall in the host. Console through the ttys If the container is configured with a root filesystem and the inittab file is setup to launch a getty on the ttys. This option will specify the number of ttys to be available for the container. The number of getty in the inittab file of the container should not be greater than the number of ttys specified in this configuration file, otherwise the excess getty sessions will die and respawn indefinitly giving annoying messages on the console. Specify the number of tty to make available to the container. Console devices location LXC consoles are provided through Unix98 PTYs created on the host and bind-mounted over the expected devices in the container. By default, they are bind-mounted over /dev/console and /dev/ttyN. This can prevent package upgrades in the guest. Therefore you can specify a directory location (under /dev under which LXC will create the files and bind-mount over them. These will then be symbolically linked to /dev/console and /dev/ttyN. A package upgrade can then succeed as it is able to remove and replace the symbolic links. Specify a directory under /dev under which to create the container console devices. /dev directory By default, lxc does nothing with the container's /dev. This allows the container's /dev to be set up as needed in the container rootfs. If lxc.autodev is set to 1, then after mounting the container's rootfs LXC will mount a fresh tmpfs under /dev (limited to 100k) and fill in a minimal set of initial devices. This is generally required when starting a container containing a "systemd" based "init" but may be optional at other times. Addional devices in the containers /dev directory may be created through the use of the hook. Set this to 1 to have LXC mount and populate a minimal /dev when starting the container. Mount points The mount points section specifies the different places to be mounted. These mount points will be private to the container and won't be visible by the processes running outside of the container. This is useful to mount /etc, /var or /home for examples. specify a file location in the fstab format, containing the mount informations. If the rootfs is an image file or a device block and the fstab is used to mount a point somewhere in this rootfs, the path of the rootfs mount point should be prefixed with the @LXCROOTFSMOUNT@ default path or the value of if specified. specify a mount point corresponding to a line in the fstab format. Root file system The root file system of the container can be different than that of the host system. specify the root file system for the container. It can be an image file, a directory or a block device. If not specified, the container shares its root file system with the host. where to recursively bind before pivoting. This is to ensure success of the pivot_root 8 syscall. Any directory suffices, the default should generally work. where to pivot the original root file system under , specified relatively to that. The default is mnt. It is created if necessary, and also removed after unmounting everything from it during container setup. Control group The control group section contains the configuration for the different subsystem. lxc does not check the correctness of the subsystem name. This has the disadvantage of not detecting configuration errors until the container is started, but has the advantage of permitting any future subsystem. specify the control group value to be set. The subsystem name is the literal name of the control group subsystem. The permitted names and the syntax of their values is not dictated by LXC, instead it depends on the features of the Linux kernel running at the time the container is started, eg. Capabilities The capabilities can be dropped in the container if this one is run as root. Specify the capability to be dropped in the container. A single line defining several capabilities with a space separation is allowed. The format is the lower case of the capability definition without the "CAP_" prefix, eg. CAP_SYS_MODULE should be specified as sys_module. See capabilities 7 , Startup hooks Startup hooks are programs or scripts which can be executed at various times in a container's lifetime. A hook to be run in the host's namespace before the container ttys, consoles, or mounts are up. A hook to be run in the container's fs namespace but before the rootfs has been set up. This allows for manipulation of the rootfs, i.e. to mount an encrypted filesystem. Mounts done in this hook will not be reflected on the host (apart from mounts propagation), so they will be automatically cleaned up when the container shuts down. A hook to be run in the container's namespace after mounting has been done, but before the pivot_root. A hook to be run in the container's namespace after mounting has been done and after any mount hooks have run, but before the pivot_root, if == 1. The purpose of this hook is to assist in populating the /dev directory of the container when using the autodev option for systemd based containers. The container's /dev directory is relative to the ${} environment variable available when the hook is run. A hook to be run in the container's namespace immediately before executing the container's init. This requires the program to be available in the container. A hook to be run in the host's namespace after the container has been shut down. Startup hooks Environment Variables A number of environment variables are made available to the startup hooks to provide configuration information and assist in the functioning of the hooks. Not all variables are valid in all contexts. In particular, all paths are relative to the host system and, as such, not valid during the hook. The LXC name of the container. Useful for logging messages in commmon log environments. [] Host relative path to the container configuration file. This gives the container to reference the original, top level, configuration file for the container in order to locate any addotional configuration information not otherwise made available. [] The path to the console output of the container if not NULL. [] [] The path to the console log output of the container if not NULL. [] The mount location to which the container is initially bound. This will be the host relative path to the container rootfs for the container instance being started and is where changes should be made for that instance. [] The host relative path to the container root which has been mounted to the rootfs.mount location. [] Examples In addition to the few examples given below, you will find some other examples of configuration file in @DOCDIR@/examples Network This configuration sets up a container to use a veth pair device with one side plugged to a bridge br0 (which has been configured before on the system by the administrator). The virtual network device visible in the container is renamed to eth0. lxc.utsname = myhostname lxc.network.type = veth lxc.network.flags = up lxc.network.link = br0 lxc.network.name = eth0 lxc.network.hwaddr = 4a:49:43:49:79:bf lxc.network.ipv4 = 10.2.3.5/24 10.2.3.255 lxc.network.ipv6 = 2003:db8:1:0:214:1234:fe0b:3597 Control group This configuration will setup several control groups for the application, cpuset.cpus restricts usage of the defined cpu, cpus.share prioritize the control group, devices.allow makes usable the specified devices. lxc.cgroup.cpuset.cpus = 0,1 lxc.cgroup.cpu.shares = 1234 lxc.cgroup.devices.deny = a lxc.cgroup.devices.allow = c 1:3 rw lxc.cgroup.devices.allow = b 8:0 rw Complex configuration This example show a complex configuration making a complex network stack, using the control groups, setting a new hostname, mounting some locations and a changing root file system. lxc.utsname = complex lxc.network.type = veth lxc.network.flags = up lxc.network.link = br0 lxc.network.hwaddr = 4a:49:43:49:79:bf lxc.network.ipv4 = 10.2.3.5/24 10.2.3.255 lxc.network.ipv6 = 2003:db8:1:0:214:1234:fe0b:3597 lxc.network.ipv6 = 2003:db8:1:0:214:5432:feab:3588 lxc.network.type = macvlan lxc.network.flags = up lxc.network.link = eth0 lxc.network.hwaddr = 4a:49:43:49:79:bd lxc.network.ipv4 = 10.2.3.4/24 lxc.network.ipv4 = 192.168.10.125/24 lxc.network.ipv6 = 2003:db8:1:0:214:1234:fe0b:3596 lxc.network.type = phys lxc.network.flags = up lxc.network.link = dummy0 lxc.network.hwaddr = 4a:49:43:49:79:ff lxc.network.ipv4 = 10.2.3.6/24 lxc.network.ipv6 = 2003:db8:1:0:214:1234:fe0b:3297 lxc.cgroup.cpuset.cpus = 0,1 lxc.cgroup.cpu.shares = 1234 lxc.cgroup.devices.deny = a lxc.cgroup.devices.allow = c 1:3 rw lxc.cgroup.devices.allow = b 8:0 rw lxc.mount = /etc/fstab.complex lxc.mount.entry = /lib /root/myrootfs/lib none ro,bind 0 0 lxc.rootfs = /mnt/rootfs.complex lxc.cap.drop = sys_module mknod setuid net_raw lxc.cap.drop = mac_override See Also chroot 1 , pivot_root 8 , fstab 5 &seealso; Author Daniel Lezcano daniel.lezcano@free.fr