Running Flatcar Container Linux on VMware

    These instructions walk through running Flatcar Container Linux on VMware Fusion or ESXi. If you are familiar with another VMware product, you can use these instructions as a starting point.

    Running the VM

    Choosing a channel

    Flatcar Container Linux is designed to be updated automatically with different schedules per channel. You can disable this feature , although we don’t recommend it. Read the release notes for specific features and bug fixes.

    The Stable channel should be used by production clusters. Versions of Flatcar Container Linux are battle-tested within the Beta and Alpha channels before being promoted. The current version is Flatcar Container Linux 3815.2.0.

    curl -LO https://stable.release.flatcar-linux.net/amd64-usr/current/flatcar_production_vmware_ova.ova

    The Alpha channel closely tracks master and is released frequently. The newest versions of system libraries and utilities will be available for testing. The current version is Flatcar Container Linux 3874.0.0.

    curl -LO https://alpha.release.flatcar-linux.net/amd64-usr/current/flatcar_production_vmware_ova.ova

    The Beta channel consists of promoted Alpha releases. The current version is Flatcar Container Linux 3850.1.0.

    curl -LO https://beta.release.flatcar-linux.net/amd64-usr/current/flatcar_production_vmware_ova.ova

    Booting with VMware vSphere/ESXi from the web interface

    Use the vSphere Client/ESXi web interface to deploy the VM as follows:

    1. In the menu, click File > Deploy OVF Template...
    2. In the wizard, specify the location of the OVA file downloaded earlier
    3. Name your VM
    4. Choose “thin provision” for the disk format
    5. Choose your network settings and specify provisioning userdata
    6. Confirm the settings, then click “Finish”

    Uncheck Power on after deployment in order to edit the VM before booting it the first time.

    The last step uploads the files to the ESXi datastore and registers the new VM. You can now tweak VM settings, then power it on.

    Booting with VMware vSphere/ESXi from the command line with ovftool

    Use the ovftool to deploy from the command line as follows:

    ovftool --name=testvm --skipManifestCheck --noSSLVerify --datastore=datastore1 --powerOn=True --net:"VM Network=VM Network" --X:waitForIp --overwrite --powerOffTarget --X:guest:ignition.config.data=$(cat ignition_config.json | base64 --wrap=0) --X:guest:ignition.config.data.encoding=base64 ./flatcar_production_vmware_ova.ova 'vi:///<YOUR_USER>:<ESXI_PASSWORD>@<ESXI_HOST_IP>'
    

    This assumes that you downloaded flatcar_production_vmware_ova.ova to your current folder, and that you want to specify an Ignition config as userdata from ignition_config.json.

    NB: These instructions were tested with an ESXi v5.5 host.

    Booting with VMware Workstation 12 or VMware Fusion

    Run VMware Workstation GUI:

    1. In the menu, click File > Open...
    2. In the wizard, specify the location of the OVA template downloaded earlier
    3. Name your VM, then click Import
    4. (Press Retry if VMware Workstation raises an “OVF specification” warning)
    5. Edit VM settings if necessary and specify provisioning userdata
    6. Start your Flatcar Container Linux VM

    NB: These instructions were tested with a Fusion 8.1 host.

    Installing via PXE or ISO image

    Flatcar Container Linux can also be installed by booting the virtual machine via PXE or the ISO image and then installing Flatcar Container Linux to disk .

    Butane Configs

    Flatcar Container Linux allows you to configure machine parameters, configure networking, launch systemd units on startup, and more via Butane Configs. These configs are then transpiled into Ignition configs and given to booting machines. Head over to the docs to learn about the supported features .

    You can provide a raw Ignition config to Flatcar Container Linux via VMware’s Guestinfo interface .

    As an example, this Butane Config will start an NGINX Docker container and configure private and public static IP addresses:

    variant: flatcar
    version: 1.0.0
    storage:
      files:
        - path: /etc/systemd/network/00-vmware.network
          contents:
            inline: |
              [Match]
              Name=ens192
              [Network]
              DHCP=no
              DNS=1.1.1.1
              DNS=1.0.0.1
              [Address]
              Address=123.45.67.2/29
              [Address]
              Address=10.0.0.2/29
              [Route]
              Destination=0.0.0.0/0
              Gateway=123.45.67.1
              [Route]
              Destination=10.0.0.0/8
              Gateway=10.0.0.1          
    systemd:
      units:
        - name: nginx.service
          enabled: true
          contents: |
            [Unit]
            Description=NGINX example
            After=docker.service
            Requires=docker.service
            [Service]
            TimeoutStartSec=0
            ExecStartPre=-/usr/bin/docker rm --force nginx1
            ExecStart=/usr/bin/docker run --name nginx1 --pull always --log-driver=journald --net host docker.io/nginx:1
            ExecStop=/usr/bin/docker stop nginx1
            Restart=always
            RestartSec=5s
            [Install]
            WantedBy=multi-user.target        
    

    Transpile it to Ignition JSON:

    cat cl.yaml | docker run --rm -i quay.io/coreos/butane:release > ignition.json
    

    For DHCP you don’t need to specify any networkd units.

    After transpilation, the resulting JSON content can be used in guestinfo.ignition.config.data after encoding it to base64 and setting guestinfo.ignition.config.data.encoding to base64. If DHCP is used, the JSON file can also be uploaded to a web server and fetched by Ignition if the HTTP(s) URL is given in guestinfo.ignition.config.url.

    Beginning with Flatcar major version 3248, fetching remote resources in Ignition or with torcx is not only supported with DHCP but also by using guestinfo.afterburn.initrd.network-kargs to define a custom network configuration, here an example for a static IP address .

    IP configuration specified via guestinfo.interface.* and guestinfo.dns.* variables is currently not supported with Ignition and will only work if you provide coreos-cloudinit data (cloud-config or a script) as userdata.

    Templating with Butane Configs and setting up metadata

    On many cloud providers Ignition will run the coreos-metadata.service (which runs afterburn) to set up node metadata . This is not the case with VMware because the network setup is defined by you and nothing generic that afterburn would know about.

    Here’s a Butane configuration example to setup an etcd instance with a custom coreos-metadata.service:

    version: 1.0.0
    variant: flatcar
    systemd:
      units:
        - name: etcd-member.service
          enabled: true
          dropins:
            - name: 20-clct-etcd-member.conf
              contents: |
                [Unit]
                Requires=coreos-metadata.service
                After=coreos-metadata.service
                [Service]
                EnvironmentFile=/run/metadata/coreos
                ExecStart=
                ExecStart=/usr/lib/coreos/etcd-wrapper $ETCD_OPTS --advertise-client-urls="http://${COREOS_CUSTOM_PUBLIC_IPV4}:2379"            
        - name: coreos-metadata.service
          contents: |
            [Unit]
            Description=VMware metadata agent
            After=nss-lookup.target
            After=network-online.target
            Wants=network-online.target
            [Service]
            Type=oneshot
            Restart=on-failure
            RemainAfterExit=yes
            Environment=OUTPUT=/run/metadata/coreos
            ExecStart=/usr/bin/mkdir --parent /run/metadata
            ExecStart=/usr/bin/bash -c 'echo "COREOS_CUSTOM_PRIVATE_IPV4=$(ip addr show ens192 | grep "inet 10." | grep -Po "inet \K[\d.]+")\nCOREOS_CUSTOM_PUBLIC_IPV4=$(ip addr show ens192 | grep -v "inet 10." | grep -Po "inet \K[\d.]+")" > ${OUTPUT}'        
    

    This populates /run/metadata/coreos with variables for a public IP address on interface ens192 (taking the one that is not starting with 10.…) and a private IP address on the same interface (taking the one that is starting with 10.…). You need to adjust this to your network setup. In case you use the guestinfo.interface.* variables you could use /usr/share/oem/bin/vmware-rpctool 'info-get guestinfo.interface.0.ip.0.address' instead of ip addr show … | grep ….

    Using coreos-cloudinit Cloud-Configs

    Ignition is the preferred way of provisioning because it runs in the initramfs and only at first boot. Cloud-Configs are supported, too, but coreos-cloudinit is not actively developed at the moment.

    Both Cloud-Config YAML content and raw bash scripts are supported by coreos-cloudinit. You can provide them to Flatcar Container Linux via VMware’s Guestinfo interface .

    For $public_ipv4 and $private_ipv4 substitutions to work you either need to use static IPs (through guestinfo.interface.* as described below) or you need to write the variables COREOS_PUBLIC_IPV4 and COREOS_PRIVATE_IPV4 to /etc/environment before coreos-cloudinit runs which would require a reboot. Thus, it may be easier to use the coreos-metadata.service approach and write these variables to /run/metadata/coreos. To do so, set EnvironmentFile=/run/metadata/coreos, Requires=coreos-metadata.service, and After=coreos-metadata.service in your systemd unit.

    Besides applying the config itself coreos-cloudinit supports the guestinfo.interface.* variables and will generate a networkd unit from them stored in /run/systemd/network/.

    The guestinfo variables known to coreos-cloudinit are (taken from here ), with <n>, <m>, <l> being numbers starting from 0:

    • guestinfo.hostname used for hostnamectl set-hostname
    • guestinfo.interface.<n>.name used in the [Match] section of the networkd unit (can include wildcards)
    • guestinfo.interface.<n>.mac used in the [Match] section of the networkd unit
    • guestinfo.interface.<n>.dhcp is either yes or no and used in the [Network] section of the networkd unit
    • guestinfo.interface.<n>.role (required to generate a networkd unit for <n>) is either public or private and used for Cloud-Config variable substitions ($public_ipv4 etc) instead of COREOS_PUBLIC_IPV4 from /etc/environment
    • guestinfo.interface.<n>.ip.<m>.address is a static IP address with subnet, e.g., 123.4.5.6/29, used in the [Address] section of the networkd unit
    • guestinfo.interface.<n>.route.<l>.gateway used in the [Route] section of the networkd unit
    • guestinfo.interface.<n>.route.<l>.destination is a IP CIDR, e.g., 0.0.0.0/0, used in the [Route] section of the networkd unit
    • guestinfo.dns.server.<x> used in the [Network] section of any networkd unit
    • guestinfo.dns.domain.<y> used in the [Network] section of any networkd unit
    • guestinfo.(ignition|coreos).config.data, guestinfo.(ignition|coreos).config.data.encoding, and guestinfo.(ignition|coreos).config.url as described in the surrounding sections

    If you rely on $public_ipv4 and $private_ipv4 substitutions through guestinfo.interface.<n>.role but have both IP addresses in one interface you may either use variables in /run/metadata/coreos as written in the previous section or you could provide the second IP address again on a dummy interface with a name that never matches a real interface, just to propagate the IP address to the coreos-cloudinit metadata.

    VMware Guestinfo interface

    Setting Guestinfo options

    The VMware guestinfo interface is a mechanism for VM configuration. Guestinfo properties are stored in the VMX file, or in the VMX representation in host memory. These properties are available to the VM at boot time. Within the VMX, the names of these properties are prefixed with guestinfo.. Guestinfo settings can be injected into VMs in one of four ways:

    • Configure guestinfo in the OVF for deployment. Software like vcloud director manipulates OVF descriptors for guest configuration. For details, check out this VMware blog post about Self-Configuration and the OVF Environment .

    • The ESXi web UI and VMware Workstation Player either directly display the OVF guestinfo variables for editing or allow to add them as parameters in the VM settings before deployment. They can also be changed and added later in the VM settings (but for Ignition configs that requires touch /boot/flatcar/first_boot so that Ignition runs again on the next boot).

    • The ovftool supports guestinfo variables with --X:guest:VARIABLE=value.

    • Set guestinfo keys and values from the Flatcar Container Linux guest itself, by using a VMware Tools command like:

    /usr/share/oem/bin/vmtoolsd --cmd "info-set guestinfo.<variable> <value>"
    
    • Guestinfo keys and values can be set from a VMware Service Console, using the setguestinfo subcommand:
    vmware-cmd /vmfs/volumes/[...]/<VMNAME>/<VMNAME>.vmx setguestinfo guestinfo.<property> <value>
    
    • You can manually modify the VMX and reload it on the VMware Workstation, ESXi host, or in vCenter.

    Guestinfo configuration set via the VMware API or with vmtoolsd from within the Flatcar Container Linux guest itself are stored in VM process memory and are lost on VM shutdown or reboot.

    Defining the Ignition config or coreos-cloudinit Cloud-Config in Guestinfo

    If either the guestinfo.ignition.config.data or the guestinfo.ignition.config.url userdata property contains an Ignition config, Ignition will apply the referenced config on first boot during the initramfs phase. If it contains a Cloud-Config or script, Ignition will enable a service for coreos-cloudinit that will run on every boot and apply the config.

    The userdata is prepared for the guestinfo facility in one of two encoding types, specified in the guestinfo.ignition.config.data.encoding variable:

    Encoding Command
    <elided> sed -e 's/%/%%/g' -e 's/"/%22/g' /path/to/user_data
    base64 base64 -w0 /path/to/user_data
    gz+base64 gzip -c -9 /path/to/user_data | base64 -w0

    base64 (or gz+base64) encoding is mandatory for ESXi, passing unencoded Ignition data will lead to Ignition failures during boot due to lack of escaping in the guestinfo XML data.

    Example

    guestinfo.ignition.config.data = "ewogICJpZ25pdGlvbiI6IHsgInZlcnNpb24iOiAiMi4wLjAiIH0KfQo="
    guestinfo.ignition.config.data.encoding = "base64"
    

    This example will be decoded into the following Ignition config, but a Cloud-Config can be specified the same way in the variable:

    {
      "ignition": { "version": "2.0.0" }
    }
    

    Instead of providing the userdata inline, you can also specify a remote HTTP location in guestinfo.ignition.config.url. Both Ignition and coreos-cloudinit support it but Ignition relies on DHCP in the initramfs which means that it can’t fetch remote resources if you have to use static IPs.

    Logging in

    The VGA console has autologin enabled.

    Networking can take some time to start under VMware. Once it does, you will see the IP when typing ip a or in the VM info that VMware displays.

    You can login to the host at that IP using your SSH key, or the password set in your cloud-config:

    ssh core@YOURIP
    

    Disabling/enabling autologin

    Beginning with Flatcar major version 3185 the kernelArguments directive in Ignition v3 allows to add/remove the flatcar.autologin kernel command line parameter that is set in grub.cfg. The following short Butane YAML config (to be transpiled to Ignition v3 JSON) ensures that the flatcar.autologin kernel parameter gets removed and then as part of the first boot it gets applied to an instant reboot before the instance comes up:

    variant: flatcar
    version: 1.0.0
    kernel_arguments:
      should_not_exist:
        - flatcar.autologin
    

    With should_exist instead of should_not_exist the argument would be added if it isn’t set in grub.cfg already.

    Read more about setting kernel command line parameters this way here .

    In case you want to disable the autologin on the console with Ignition v2 where no kernelArguments directive exists, you can use the following directive in your Container Linux Config YAML. To take effect it requires an additional reboot.

    storage:
      filesystems:
        - name: oem
          mount:
            device: /dev/disk/by-label/OEM
            format: btrfs
            label: OEM
      files:
        - path: /grub.cfg
          filesystem: oem
          mode: 0644
          contents:
            inline: |
              set oem_id="vmware"
              set linux_append=""          
    

    To take effect directly on first boot, the alternative is to create a [email protected] drop-in, here a CLC snippet:

    systemd:
      units:
        - name: [email protected]
          dropins:
            - name: 10-autologin.conf
              contents: |
                [Service]
                ExecStart=
                ExecStart=-/sbin/agetty --noclear %I $TERM
    

    Using Flatcar Container Linux

    Now that you have a machine booted, it’s time to explore. Check out the Flatcar Container Linux Quickstart guide, or dig into more specific topics .