Closed Kramerican closed 6 years ago
stgraber@castiana:~$ lxc launch ubuntu:bionic bionic
Creating bionic
Starting bionic
stgraber@castiana:~$ lxc list bionic
+--------+---------+----------------------+----------------------------------------------+------------+-----------+
| NAME | STATE | IPV4 | IPV6 | TYPE | SNAPSHOTS |
+--------+---------+----------------------+----------------------------------------------+------------+-----------+
| bionic | RUNNING | 10.166.11.129 (eth0) | 2001:470:b368:4242:216:3eff:fe40:9c2c (eth0) | PERSISTENT | 0 |
+--------+---------+----------------------+----------------------------------------------+------------+-----------+
Anything special about your bionic container?
Can you post lxc config show --expanded NAME
?
Ah beaverballs ... I was really hoping this would not be something specific to my system(s). Anyway, as requested lxc config show --expanded NAME
architecture: x86_64
config:
image.architecture: amd64
image.description: ubuntu 18.04 LTS amd64 (release) (20180426.2)
image.label: release
image.os: ubuntu
image.release: bionic
image.serial: "20180426.2"
image.version: "18.04"
limits.cpu.allowance: 50%
limits.memory: 4GB
limits.memory.enforce: soft
linux.kernel_modules: ip_tables,ip6_tables
volatile.base_image: b36ec647e374da4816104a98807633a2cc387488083d3776557081c4d0333618
volatile.eth0.hwaddr: 00:16:3e:bf:58:5a
volatile.idmap.base: "0"
volatile.idmap.next: '[{"Isuid":true,"Isgid":false,"Hostid":100000,"Nsid":0,"Maprange":65536},{"Isuid":false,"Isgid":true,"Hostid":100000,"Nsid":0,"Maprange":65536}]'
volatile.last_state.idmap: '[{"Isuid":true,"Isgid":false,"Hostid":100000,"Nsid":0,"Maprange":65536},{"Isuid":false,"Isgid":true,"Hostid":100000,"Nsid":0,"Maprange":65536}]'
volatile.last_state.power: RUNNING
devices:
eth0:
name: eth0
nictype: bridged
parent: br0
type: nic
root:
path: /
pool: lxd
size: 60GB
type: disk
ephemeral: false
profiles:
- webdock
stateful: false
description: ""
Thank you for your assistance as always
Ok, that looks reasonable.
What does ps fauxww
look like inside the container?
ps fauxww
root@bionic-template:~# ps fauxwww
USER PID %CPU %MEM VSZ RSS TTY STAT START TIME COMMAND
root 263 0.0 0.0 21900 2584 ? Ss 08:50 0:00 bash
root 273 0.0 0.0 37792 2156 ? R+ 08:50 0:00 \_ ps fauxwww
root 1 0.0 0.0 77484 6004 ? Ss 08:07 0:00 /sbin/init
root 58 0.0 0.0 42104 2016 ? Ss 08:07 0:00 /lib/systemd/systemd-udevd
root 62 0.0 0.0 78436 7604 ? Ss 08:07 0:00 /lib/systemd/systemd-journald
systemd+ 112 0.0 0.0 71816 3844 ? Ss 08:07 0:00 /lib/systemd/systemd-networkd
systemd+ 113 0.0 0.0 70608 4024 ? Ss 08:07 0:00 /lib/systemd/systemd-resolved
root 150 0.0 0.0 61996 3780 ? Ss 08:09 0:00 /lib/systemd/systemd-logind
daemon 154 0.0 0.0 28332 1568 ? Ss 08:09 0:00 /usr/sbin/atd -f
root 156 0.0 0.0 31748 1880 ? Ss 08:09 0:00 /usr/sbin/cron -f
syslog 159 0.0 0.0 197636 2824 ? Ssl 08:09 0:00 /usr/sbin/rsyslogd -n
message+ 160 0.0 0.0 49928 2836 ? Ss 08:09 0:00 /usr/bin/dbus-daemon --system --address=systemd: --nofork --nopidfile --systemd-activation --syslog-only
root 163 0.0 0.0 170860 14264 ? Ssl 08:09 0:00 /usr/bin/python3 /usr/bin/networkd-dispatcher
root 164 0.0 0.0 287980 4496 ? Ssl 08:09 0:00 /usr/lib/accountsservice/accounts-daemon
root 165 0.0 0.0 1210256 14000 ? Ssl 08:09 0:00 /usr/lib/snapd/snapd
root 166 0.0 0.0 72296 3652 ? Ss 08:09 0:00 /usr/sbin/sshd -D
root 173 0.0 0.0 16412 1484 console Ss+ 08:09 0:00 /sbin/agetty -o -p -- \u --noclear --keep-baud console 115200,38400,9600 linux
root 179 0.0 0.0 288868 4480 ? Ssl 08:09 0:00 /usr/lib/policykit-1/polkitd --no-debug
Also, what kernel are you running? You mentioned 18.04, so that's LXD 3.0.0 from the archive running on Ubuntu 18.04 with the 4.15 kernel correct?
@stgraber Yes that is correct, although the command outputs here are from my test container on my 16.04 system. I just provisioned a container with lxc init ubuntu:bionic
on my 18.04 system and ran the same commands:
lxc config show --expanded NAME
architecture: x86_64
config:
boot.autostart: "true"
image.architecture: amd64
image.description: ubuntu 18.04 LTS amd64 (release) (20180426.2)
image.label: release
image.os: ubuntu
image.release: bionic
image.serial: "20180426.2"
image.version: "18.04"
limits.cpu.allowance: 50%
limits.memory: 8GB
limits.memory.enforce: soft
linux.kernel_modules: ip_tables,ip6_tables
volatile.base_image: b36ec647e374da4816104a98807633a2cc387488083d3776557081c4d0333618
volatile.eth0.hwaddr: 00:16:3e:d3:ff:75
volatile.idmap.base: "0"
volatile.idmap.next: '[{"Isuid":true,"Isgid":false,"Hostid":100000,"Nsid":0,"Maprange":65536},{"Isuid":false,"Isgid":true,"Hostid":100000,"Nsid":0,"Maprange":65536}]'
volatile.last_state.idmap: '[{"Isuid":true,"Isgid":false,"Hostid":100000,"Nsid":0,"Maprange":65536},{"Isuid":false,"Isgid":true,"Hostid":100000,"Nsid":0,"Maprange":65536}]'
volatile.last_state.power: RUNNING
devices:
eth0:
name: eth0
nictype: bridged
parent: br0
type: nic
root:
limits.read: 40000iops
limits.write: 10000iops
path: /
pool: lxd
size: 50GB
type: disk
ephemeral: false
profiles:
- webdockssd50
stateful: false
description: ""
ps fauxwww
root@bionic-template:~# ps fauxwww
USER PID %CPU %MEM VSZ RSS TTY STAT START TIME COMMAND
root 169 0.0 0.0 21900 2572 ? Ss 09:54 0:00 bash
root 294 0.0 0.0 37792 2104 ? R+ 09:55 0:00 \_ ps fauxwww
root 1 0.0 0.0 77544 5980 ? Ss 09:53 0:00 /sbin/init
root 54 0.0 0.0 78436 7580 ? Ss 09:53 0:00 /lib/systemd/systemd-journald
root 63 0.0 0.0 42104 2104 ? Ss 09:53 0:00 /lib/systemd/systemd-udevd
systemd+ 166 0.0 0.0 71816 3932 ? Ss 09:53 0:00 /lib/systemd/systemd-networkd
systemd+ 167 0.0 0.0 70608 4148 ? Ss 09:53 0:00 /lib/systemd/systemd-resolved
root 212 0.0 0.0 767224 14912 ? Ssl 09:55 0:00 /usr/lib/snapd/snapd
root 213 0.0 0.0 170864 14232 ? Ssl 09:55 0:00 /usr/bin/python3 /usr/bin/networkd-dispatcher
root 217 0.0 0.0 61996 3792 ? Ss 09:55 0:00 /lib/systemd/systemd-logind
syslog 218 0.0 0.0 197636 2812 ? Ssl 09:55 0:00 /usr/sbin/rsyslogd -n
root 219 0.0 0.0 287980 4432 ? Ssl 09:55 0:00 /usr/lib/accountsservice/accounts-daemon
message+ 221 0.0 0.0 49928 2880 ? Ss 09:55 0:00 /usr/bin/dbus-daemon --system --address=systemd: --nofork --nopidfile --systemd-activation --syslog-only
daemon 235 0.0 0.0 28332 1612 ? Ss 09:55 0:00 /usr/sbin/atd -f
root 236 0.0 0.0 31748 1996 ? Ss 09:55 0:00 /usr/sbin/cron -f
root 237 0.0 0.0 72296 3712 ? Ss 09:55 0:00 /usr/sbin/sshd -D
root 239 0.0 0.0 288868 4432 ? Ssl 09:55 0:00 /usr/lib/policykit-1/polkitd --no-debug
root 244 0.0 0.0 16412 1496 console Ss+ 09:55 0:00 /sbin/agetty -o -p -- \u --noclear --keep-baud console 115200,38400,9600 linux
for completeness, on my 16.04 system inside the container uname -r
shows 4.13.0-32-generic
And on 18.04 4.15.0-20-generic
What storage backend are you using?
@stgraber ZFS :)
Interesting questions you are asking! I would have thought we'd be looking more specifically at the network interface. I'd love to know what theories / causes you are thinking might be at work here ...
No luck reproducing it here on a bunch of systems. Our automated testing also shows that this image has come up with working network for weeks now...
In your container, can you try running systemctl start systemd-networkd
see if that causes the same error to show up in dmesg?
Oww man - this is basically a worst case scenario then. There's something in my setup messing with networking in a Bionic container specifically, but not Zesty or Xenial. On the surface this is rather bizarre. I really hate wasting your time, so I really hope this isn't down to some retarded config I have done on my end.
In any case, I am not sure if I follow your instructions correctly here, but I execute systemctl start systemd-networkd
and it completes with no message (success I assume). Running dmesg
in the container gives me a dmesg: read kernel buffer failed: Operation not permitted
which is expected AFAIK, and running dmesg on the host reveals nothing new (no new messages) - and cat /var/log/syslog on host doesn't reveal anything new (no new messages).
If we assume I have something "unrelated" to LXD in my setup here which is messing things up, any suggestions on steps I could take to narrow things down?
That line:
Apr 29 20:25:16 krellide kernel: [6056887.323323] audit: type=1400 audit(1525026316.029:23532): apparmor="DENIED" operation="mount" info="failed flags match" error=-13 profile="lxd-bionic-template-xlemp72_</var/lib/lxd>" name="/run/systemd/unit-root/var/lib/lxcfs/" pid=19482 comm="(networkd)" flags="ro, nosuid, nodev, remount, bind"
Is clearly what's causing the network problem, most likely it's preventing systemd-networkd from starting dhclient or something, causing the network problem you're seeing...
Can you try:
lxc launch ubuntu:18.04 test lxc exec test bash
journalctl | cat
cat /proc/1/mountinfo
That may help figure out what's going on.
Sure thing :)
root@test:~# journalctl | cat
-- Logs begin at Mon 2018-04-30 14:48:10 UTC, end at Mon 2018-04-30 14:48:12 UTC. --
Apr 30 14:48:10 test systemd-journald[55]: Journal started
Apr 30 14:48:10 test systemd-journald[55]: Runtime journal (/run/log/journal/897ae676489745b4b5f6303bdde70ccf) is 8.0M, max 3.1G, 3.1G free.
Apr 30 14:48:10 test systemd-sysctl[58]: Couldn't write '4 4 1 7' to 'kernel/printk', ignoring: Permission denied
Apr 30 14:48:10 test systemd-sysctl[58]: Couldn't write '1' to 'kernel/kptr_restrict', ignoring: Permission denied
Apr 30 14:48:10 test systemd-sysctl[58]: Couldn't write '1' to 'fs/protected_hardlinks', ignoring: Permission denied
Apr 30 14:48:10 test systemd-sysctl[58]: Couldn't write '1' to 'fs/protected_symlinks', ignoring: Permission denied
Apr 30 14:48:10 test keyboard-setup.sh[52]: Couldn't get a file descriptor referring to the console
Apr 30 14:48:10 test systemd-sysctl[58]: Couldn't write '1024' to 'fs/inotify/max_user_instances', ignoring: Permission denied
Apr 30 14:48:10 test keyboard-setup.sh[52]: Couldn't get a file descriptor referring to the console
Apr 30 14:48:10 test systemd-sysctl[58]: Couldn't write '176' to 'kernel/sysrq', ignoring: Permission denied
Apr 30 14:48:10 test systemd-sysctl[58]: Couldn't write '1' to 'kernel/yama/ptrace_scope', ignoring: Permission denied
Apr 30 14:48:10 test systemd-sysctl[58]: Couldn't write '65536' to 'vm/mmap_min_addr', ignoring: Permission denied
Apr 30 14:48:10 test systemd-sysctl[58]: Couldn't write 'fq_codel' to 'net/core/default_qdisc', ignoring: No such file or directory
Apr 30 14:48:10 test systemd[1]: Started Apply Kernel Variables.
Apr 30 14:48:10 test systemd-journald[55]: Forwarding to syslog missed 1 messages.
Apr 30 14:48:10 test systemd[1]: systemd-journal-flush.service: Failed to reset devices.list: Operation not permitted
Apr 30 14:48:10 test systemd[1]: Starting Flush Journal to Persistent Storage...
Apr 30 14:48:10 test systemd[1]: Started udev Kernel Device Manager.
Apr 30 14:48:10 test systemd-journald[55]: Time spent on flushing to /var is 1.671ms for 18 entries.
Apr 30 14:48:10 test systemd-journald[55]: System journal (/var/log/journal/897ae676489745b4b5f6303bdde70ccf) is 512B, max 4.0G, 3.9G free.
Apr 30 14:48:10 test systemd[1]: Started Set the console keyboard layout.
Apr 30 14:48:10 test keyboard-setup.sh[52]: setupcon: We are not on the console, the console is left unconfigured.
Apr 30 14:48:10 test systemd[1]: Reached target Local File Systems (Pre).
Apr 30 14:48:10 test systemd[1]: Reached target Local File Systems.
Apr 30 14:48:10 test systemd[1]: apparmor.service: Failed to reset devices.list: Operation not permitted
Apr 30 14:48:10 test systemd[1]: Starting AppArmor initialization...
Apr 30 14:48:10 test systemd[1]: console-setup.service: Failed to reset devices.list: Operation not permitted
Apr 30 14:48:10 test systemd[1]: Starting Set console font and keymap...
Apr 30 14:48:10 test systemd[1]: ebtables.service: Failed to reset devices.list: Operation not permitted
Apr 30 14:48:10 test systemd[1]: Starting ebtables ruleset management...
Apr 30 14:48:10 test systemd[1]: plymouth-read-write.service: Failed to reset devices.list: Operation not permitted
Apr 30 14:48:10 test systemd[1]: Starting Tell Plymouth To Write Out Runtime Data...
Apr 30 14:48:10 test systemd[1]: Started Set console font and keymap.
Apr 30 14:48:10 test systemd[1]: Started Tell Plymouth To Write Out Runtime Data.
Apr 30 14:48:10 test systemd[1]: Started Flush Journal to Persistent Storage.
Apr 30 14:48:10 test systemd[1]: Started ebtables ruleset management.
Apr 30 14:48:10 test apparmor[85]: * Not starting AppArmor in container
Apr 30 14:48:10 test apparmor[85]: ...done.
Apr 30 14:48:10 test systemd[1]: systemd-tmpfiles-setup.service: Failed to reset devices.list: Operation not permitted
Apr 30 14:48:10 test systemd[1]: Starting Create Volatile Files and Directories...
Apr 30 14:48:10 test systemd[1]: Started AppArmor initialization.
Apr 30 14:48:10 test systemd[1]: Started Create Volatile Files and Directories.
Apr 30 14:48:10 test systemd[1]: systemd-update-utmp.service: Failed to reset devices.list: Operation not permitted
Apr 30 14:48:10 test systemd[1]: Starting Update UTMP about System Boot/Shutdown...
Apr 30 14:48:10 test systemd[1]: Reached target System Time Synchronized.
Apr 30 14:48:10 test systemd[1]: Started udev Coldplug all Devices.
Apr 30 14:48:10 test systemd[1]: Started Dispatch Password Requests to Console Directory Watch.
Apr 30 14:48:10 test systemd[1]: Reached target Local Encrypted Volumes.
Apr 30 14:48:10 test systemd[1]: Started Update UTMP about System Boot/Shutdown.
Apr 30 14:48:11 test cloud-init[53]: Cloud-init v. 18.2 running 'init-local' at Mon, 30 Apr 2018 14:48:10 +0000. Up 0.00 seconds.
Apr 30 14:48:11 test systemd[1]: Started Initial cloud-init job (pre-networking).
Apr 30 14:48:11 test systemd[1]: Reached target Network (Pre).
Apr 30 14:48:11 test systemd[1]: systemd-networkd.service: Failed to reset devices.list: Operation not permitted
Apr 30 14:48:11 test systemd[1]: Starting Network Service...
Apr 30 14:48:11 test systemd-networkd[123]: Enumeration completed
Apr 30 14:48:11 test systemd[1]: Started Network Service.
Apr 30 14:48:11 test systemd[1]: systemd-networkd-wait-online.service: Failed to reset devices.list: Operation not permitted
Apr 30 14:48:11 test systemd[1]: Starting Wait for Network to be Configured...
Apr 30 14:48:11 test systemd[1]: systemd-resolved.service: Failed to reset devices.list: Operation not permitted
Apr 30 14:48:11 test systemd[1]: Failed to set devices.allow on /system.slice/systemd-resolved.service: Operation not permitted
Apr 30 14:48:11 test systemd[1]: Failed to set devices.allow on /system.slice/systemd-resolved.service: Operation not permitted
Apr 30 14:48:11 test systemd[1]: Failed to set devices.allow on /system.slice/systemd-resolved.service: Operation not permitted
Apr 30 14:48:11 test systemd[1]: Failed to set devices.allow on /system.slice/systemd-resolved.service: Operation not permitted
Apr 30 14:48:11 test systemd[1]: Failed to set devices.allow on /system.slice/systemd-resolved.service: Operation not permitted
Apr 30 14:48:11 test systemd[1]: Failed to set devices.allow on /system.slice/systemd-resolved.service: Operation not permitted
Apr 30 14:48:11 test systemd[1]: Failed to set devices.allow on /system.slice/systemd-resolved.service: Operation not permitted
Apr 30 14:48:11 test systemd[1]: Failed to set devices.allow on /system.slice/systemd-resolved.service: Operation not permitted
Apr 30 14:48:11 test systemd[1]: Starting Network Name Resolution...
Apr 30 14:48:11 test systemd-resolved[125]: Positive Trust Anchors:
Apr 30 14:48:11 test systemd-resolved[125]: . IN DS 19036 8 2 49aac11d7b6f6446702e54a1607371607a1a41855200fd2ce1cdde32f24e8fb5
Apr 30 14:48:11 test systemd-resolved[125]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d
Apr 30 14:48:11 test systemd-resolved[125]: Negative trust anchors: 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa corp home internal intranet lan local private test
Apr 30 14:48:11 test systemd-resolved[125]: Using system hostname 'test'.
Apr 30 14:48:11 test systemd[1]: Started Network Name Resolution.
Apr 30 14:48:11 test systemd[1]: Reached target Host and Network Name Lookups.
Apr 30 14:48:11 test systemd[1]: Reached target Network.
Apr 30 14:48:12 test systemd-networkd[123]: eth0: Gained IPv6LL
root@test:~# cat /proc/1/mountinfo
430 431 0:80 /rootfs / rw,relatime master:98 - zfs lxd/containers/test rw,xattr,noacl
475 430 0:81 / /dev rw,relatime - tmpfs none rw,size=492k,mode=755,uid=100000,gid=100000
476 430 0:83 / /proc rw,nosuid,nodev,noexec,relatime - proc proc rw
477 430 0:84 / /sys rw,relatime - sysfs sysfs rw
478 475 0:6 /fuse /dev/fuse rw,nosuid,relatime master:2 - devtmpfs udev rw,size=32854604k,nr_inodes=8213651,mode=755
479 475 0:6 /net/tun /dev/net/tun rw,nosuid,relatime master:2 - devtmpfs udev rw,size=32854604k,nr_inodes=8213651,mode=755
480 476 0:44 / /proc/sys/fs/binfmt_misc rw,relatime master:82 - binfmt_misc binfmt_misc rw
481 477 0:42 / /sys/fs/fuse/connections rw,relatime master:29 - fusectl fusectl rw
482 477 0:28 / /sys/fs/pstore rw,nosuid,nodev,noexec,relatime master:11 - pstore pstore rw
483 477 0:8 / /sys/kernel/debug rw,relatime master:24 - debugfs debugfs rw
484 483 0:11 / /sys/kernel/debug/tracing rw,relatime master:26 - tracefs tracefs rw
485 477 0:7 / /sys/kernel/security rw,nosuid,nodev,noexec,relatime master:8 - securityfs securityfs rw
486 475 0:19 / /dev/mqueue rw,relatime master:25 - mqueue mqueue rw
487 475 0:47 / /dev/lxd rw,relatime master:88 - tmpfs tmpfs rw,size=100k,mode=755
488 475 0:46 /test /dev/.lxd-mounts rw,relatime master:86 - tmpfs tmpfs rw,size=100k,mode=711
489 476 0:45 /proc/cpuinfo /proc/cpuinfo rw,nosuid,nodev,relatime master:84 - fuse.lxcfs lxcfs rw,user_id=0,group_id=0,allow_other
490 476 0:45 /proc/diskstats /proc/diskstats rw,nosuid,nodev,relatime master:84 - fuse.lxcfs lxcfs rw,user_id=0,group_id=0,allow_other
491 476 0:45 /proc/meminfo /proc/meminfo rw,nosuid,nodev,relatime master:84 - fuse.lxcfs lxcfs rw,user_id=0,group_id=0,allow_other
492 476 0:45 /proc/stat /proc/stat rw,nosuid,nodev,relatime master:84 - fuse.lxcfs lxcfs rw,user_id=0,group_id=0,allow_other
493 476 0:45 /proc/swaps /proc/swaps rw,nosuid,nodev,relatime master:84 - fuse.lxcfs lxcfs rw,user_id=0,group_id=0,allow_other
494 476 0:45 /proc/uptime /proc/uptime rw,nosuid,nodev,relatime master:84 - fuse.lxcfs lxcfs rw,user_id=0,group_id=0,allow_other
495 430 0:45 / /var/lib/lxcfs rw,nosuid,nodev,relatime master:84 - fuse.lxcfs lxcfs rw,user_id=0,group_id=0,allow_other
496 475 0:6 /null /dev/null rw,nosuid,relatime master:2 - devtmpfs udev rw,size=32854604k,nr_inodes=8213651,mode=755
497 475 0:6 /zero /dev/zero rw,nosuid,relatime master:2 - devtmpfs udev rw,size=32854604k,nr_inodes=8213651,mode=755
498 475 0:6 /full /dev/full rw,nosuid,relatime master:2 - devtmpfs udev rw,size=32854604k,nr_inodes=8213651,mode=755
499 475 0:6 /urandom /dev/urandom rw,nosuid,relatime master:2 - devtmpfs udev rw,size=32854604k,nr_inodes=8213651,mode=755
500 475 0:6 /random /dev/random rw,nosuid,relatime master:2 - devtmpfs udev rw,size=32854604k,nr_inodes=8213651,mode=755
501 475 0:6 /tty /dev/tty rw,nosuid,relatime master:2 - devtmpfs udev rw,size=32854604k,nr_inodes=8213651,mode=755
502 475 0:22 /4 /dev/console rw,nosuid,noexec,relatime master:3 - devpts devpts rw,gid=5,mode=620,ptmxmode=000
432 475 0:85 / /dev/pts rw,relatime - devpts devpts rw,gid=100005,mode=620,ptmxmode=666
433 475 0:85 /ptmx /dev/ptmx rw,relatime - devpts devpts rw,gid=100005,mode=620,ptmxmode=666
434 475 0:86 / /dev/shm rw,nosuid,nodev - tmpfs tmpfs rw,uid=100000,gid=100000
435 430 0:87 / /run rw,nosuid,nodev - tmpfs tmpfs rw,mode=755,uid=100000,gid=100000
436 435 0:88 / /run/lock rw,nosuid,nodev,noexec,relatime - tmpfs tmpfs rw,size=5120k,uid=100000,gid=100000
437 477 0:89 / /sys/fs/cgroup ro,nosuid,nodev,noexec - tmpfs tmpfs ro,mode=755,uid=100000,gid=100000
438 437 0:27 / /sys/fs/cgroup/systemd rw,nosuid,nodev,noexec,relatime - cgroup cgroup rw,xattr,release_agent=/lib/systemd/systemd-cgroups-agent,name=systemd
439 437 0:36 / /sys/fs/cgroup/blkio rw,nosuid,nodev,noexec,relatime - cgroup cgroup rw,blkio
440 437 0:32 / /sys/fs/cgroup/cpu,cpuacct rw,nosuid,nodev,noexec,relatime - cgroup cgroup rw,cpu,cpuacct
441 437 0:35 / /sys/fs/cgroup/net_cls,net_prio rw,nosuid,nodev,noexec,relatime - cgroup cgroup rw,net_cls,net_prio
442 437 0:29 / /sys/fs/cgroup/perf_event rw,nosuid,nodev,noexec,relatime - cgroup cgroup rw,perf_event
443 437 0:31 / /sys/fs/cgroup/rdma rw,nosuid,nodev,noexec,relatime - cgroup cgroup rw,rdma
444 437 0:39 / /sys/fs/cgroup/pids rw,nosuid,nodev,noexec,relatime - cgroup cgroup rw,pids
445 437 0:37 / /sys/fs/cgroup/freezer rw,nosuid,nodev,noexec,relatime - cgroup cgroup rw,freezer
446 437 0:38 / /sys/fs/cgroup/memory rw,nosuid,nodev,noexec,relatime - cgroup cgroup rw,memory
447 437 0:30 / /sys/fs/cgroup/cpuset rw,nosuid,nodev,noexec,relatime - cgroup cgroup rw,cpuset
448 437 0:34 / /sys/fs/cgroup/hugetlb rw,nosuid,nodev,noexec,relatime - cgroup cgroup rw,hugetlb
449 437 0:33 / /sys/fs/cgroup/devices rw,nosuid,nodev,noexec,relatime - cgroup cgroup rw,devices
What do you see in /run/systemd/network/10-netplan-eth0.network
?
And lets do networkctl status eth0
too
cat /run/systemd/network/10-netplan-eth0.network
[Match]
Name=eth0
[Network]
DHCP=ipv4
[DHCP]
UseMTU=true
RouteMetric=100
root@test:~# networkctl status eth0
● 246: eth0
Link File: n/a
Network File: /run/systemd/network/10-netplan-eth0.network
Type: ether
State: degraded (configuring)
HW Address: 00:16:3e:8d:f6:22 (Xensource, Inc.)
Address: fe80::216:3eff:fe8d:f622
Seems to me, comparing journalctl output from a working Xenial container that your hunch is right in that DHCP isn't coming up
Tried...
sudo ln -s /etc/apparmor.d/usr.sbin.tcpdump /etc/apparmor.d/disable/
# Then container restart. No effect.
Any suggestions of a fix appreciated.
Complex/Manual steps are fine.
Thanks.
The problem appears to be in cloud-init.
The /etc/netplan/50-cloud-init.yaml runs + silently fails.
Cloud Init log output shows all public address assigned + the DHCP address fails to be assigned + there's no error.
Refer to Upstream Ubuntu Bug: /etc/netplan/50-cloud-init.yaml fails silently with no dhcp address assigned I opened this morning.
The info you've posted above is non-netplan type networking.
I'm using pure Netplan at machine level + inside containers, so my setup will be very different than yours.
@davidfavor Fair enough - I was grasping at straws :) I have delete my comment as not to crud up this thread too much. Seems like you may have found the cause - good work David!
@Kramerican oh, so in your case you had some custom network config for your containers? It wasn't a straight lxc launch ubuntu:bionic blah
without any additional modification?
@davidfavor do you see something in your dmesg
, like the DENIAL listed above here?
And to be clear, the problem is also specifically with Ubuntu 18.04 containers running on either a 18.04 or 16.04 host, non-18.04 containers work fine for you right?
@stgraber Yes and no. Just doing straing lxc launch ubuntu:bionic blah
without any additional modification also resulted in the apparmor error message and dhcp not coming up. So I don't think it's my network config per se at fault here
Ok, I don't suppose you've got that happening on a system that you could give me access to? That'd be way too easy to debug :)
@stgraber Of course! I'll send you a mail with details post-haste :)
@stgraber, Apologies. I should have been more clear about machine/container OS.
The container OS versions are yakkety, zesty, artful.
These containers run on machines running yakkety, zesty, artful, bionic.
As soon as any machine upgrades to bionic, all containers fail.
When Netplan processes /etc/netplan/50-cloud-init.yaml there are no errors or warnings. The DHCP address simply isn't defined. The cloud-init log files provide no hint of possible problem (see data in Ubuntu cloud-init ticket referenced above).
Bionic (18.04) container work find running on machines where yakkety, zesty, artful are installed.
It's only the machine update to bionic which causes all containers to fail.
dmesg | grep -i denial
Both machine level + container level produce no output.
This relates to all old containers + also the new test container I just launched which has zero modifications.
@stgraber, I can provide you full access to the machine having problem.
Skype me at ID davidfavor to talk about access.
Geez. Bionic packaging is all screwed up.
1) dnsmasq-base isn't installed at machine level.
FIX: apt-get install dnsmasq-base
2) named is running ahead of LXD setup which hijacks all interfaces, so named somehow listens on 10.X.X.X before dnsmasq, producing /var/log/syslog messages of...
dnsmasq[1846]: failed to create listening socket for 10.245.137.1: Address already in use
FIX: apt-get purge bind9
3) After reboot some progress is made.
LXD never starts + at least an error message is produced which might be useful.
Get http://unix.socket/1.0: dial unix /var/snap/lxd/common/lxd/unix.socket: connect: no such file or directory" t=2018-05-01T13:11:38+0000
So lxdbr0 never becomes available + dnsmasq can't run to route over 10.X.X.X address.
Let me know what to debug next.
net10 # snap refresh
All snaps up to date.
net10 # snap list
Name Version Rev Tracking Developer Notes
core 16-2.32.5 4486 stable canonical core
lxd 3.0.0 6862 stable canonical -
Machine history. This machine had an Artful install running. Fresh install. No update from any previous OS.
LXD was installed via SNAP, so no migration of containers was done.
Containers used "lxc copy" from other machines, to arrive at this machine.
Hum...
net10 # service lxd restart
Failed to restart lxd.service: Unit lxd.service not found.
net10 # apt-file search lxd.service
lxd: /lib/systemd/system/lxd.service
So... next question is... if there's no /lib/systemd/system/lxd.service installed via snap install of LXD, how can lxd be restarted from command line.
snap start lxd
Produces loop of Failed connecting to LXD daemon (attempt 690): Get http://unix.socket/1.0: dial unix /var/snap/lxd/common/lxd/unix.socket: connect: no such file or directory till 660 timeout of waitready.
@davidfavor sounds like you're mixing the LXD deb and snap here. When you're using the snap, you don't need dnsmasq-base installed on the host as it's included inside the snap.
The systemd unit for the snap is snap.lxd.daemon, you shouldn't have a lxd.service or lxd.socket unit around.
I suspect the problem in your case was very much bind9 conflicting with the LXD dnsmasq. To fix that, removing bind9 and then bouncing LXD with systemctl reload snap.lxd.daemon
should take care of it.
You should have had a dnsmasq error in the LXD log at /var/snap/lxd/common/lxd/logs/lxd.log indicating some kind of port/address conflict. After getting rid of the conflicting and bouncing LXD, things should work again.
No APT + SNAP mixing on this machine.
net10 # dpkg -l | grep lxd
net10 # dpkg -l | grep lxc
There is no bind9/named conflict anymore as I've removed the packages + rebooted.
systemctl reload snap.lxd.daemon
Job for snap.lxd.daemon.service failed because the control process exited with error code.
See "systemctl status snap.lxd.daemon.service" and "journalctl -xe" for details.
Because lxd.daemon trapped in "waitready 600" after boot.
So let me know if I should just kill the "lxd waitready --timeout=600" process + try the restart.
Had to kill the "lxd waitready --timeout=600" process, because the code seems broken.
I'm guessing the code's broken as /var/log/syslog shows...
Connecting to LXD daemon (attempt 1030)
Which far exceeds --timeout=600 passed to lxd waitready.
Yeah, kill that lxd waitready
and then try the reload again, if that still fails, show the output of:
Issuing - systemctl reload snap.lxd.daemon - after killing off - lxd waitready --timeout=600 - just starts another - lxd waitready --timeout=600 - process, so this fails also.
net10 # systemctl status snap.lxd.daemon
● snap.lxd.daemon.service - Service for snap application lxd.daemon
Loaded: loaded (/etc/systemd/system/snap.lxd.daemon.service; enabled; vendor preset: enabled)
Active: active (running) (Result: exit-code) since Tue 2018-05-01 08:50:33 CDT; 3min 41s ago
Process: 34806 ExecReload=/usr/bin/snap run --command=reload lxd.daemon (code=exited, status=1/FAILURE)
Main PID: 24257 (daemon.start)
Tasks: 0 (limit: 4915)
CGroup: /system.slice/snap.lxd.daemon.service
‣ 24257 /bin/sh /snap/lxd/6862/commands/daemon.start
May 01 08:50:34 net10.bizcooker.com lxd.daemon[24257]: ==> Escaping the systemd cgroups
May 01 08:50:34 net10.bizcooker.com lxd.daemon[24257]: ==> Escaping the systemd process resource limits
May 01 08:50:34 net10.bizcooker.com lxd.daemon[24257]: => Re-using existing LXCFS
May 01 08:50:34 net10.bizcooker.com lxd.daemon[24257]: => Starting LXD
May 01 08:50:34 net10.bizcooker.com lxd.daemon[24257]: lvl=warn msg="CGroup memory swap accounting is disabled, swap limits will be ignored." t=2018-05-01T13:50:34+0000
May 01 08:52:29 net10.bizcooker.com systemd[1]: Reloading Service for snap application lxd.daemon.
May 01 08:52:29 net10.bizcooker.com systemd[1]: Reloaded Service for snap application lxd.daemon.
May 01 08:53:05 net10.bizcooker.com systemd[1]: Reloading Service for snap application lxd.daemon.
May 01 08:53:05 net10.bizcooker.com systemd[1]: snap.lxd.daemon.service: Control process exited, code=exited status=1
May 01 08:53:05 net10.bizcooker.com systemd[1]: Reload failed for Service for snap application lxd.daemon.
-- Reboot --
May 01 08:10:27 net10.bizcooker.com systemd[1]: Started Service for snap application lxd.daemon.
May 01 08:10:31 net10.bizcooker.com lxd.daemon[1047]: => Preparing the system
May 01 08:10:31 net10.bizcooker.com lxd.daemon[1047]: ==> Loading snap configuration
May 01 08:10:31 net10.bizcooker.com lxd.daemon[1047]: ==> Setting up mntns symlink
May 01 08:10:31 net10.bizcooker.com lxd.daemon[1047]: ==> Setting up kmod wrapper
May 01 08:10:31 net10.bizcooker.com lxd.daemon[1047]: ==> Preparing /boot
May 01 08:10:31 net10.bizcooker.com lxd.daemon[1047]: ==> Preparing a clean copy of /run
May 01 08:10:31 net10.bizcooker.com lxd.daemon[1047]: ==> Preparing a clean copy of /etc
May 01 08:10:32 net10.bizcooker.com lxd.daemon[1047]: ==> Setting up ceph configuration
May 01 08:10:32 net10.bizcooker.com lxd.daemon[1047]: ==> Setting up LVM configuration
May 01 08:10:32 net10.bizcooker.com lxd.daemon[1047]: ==> Rotating logs
May 01 08:10:32 net10.bizcooker.com lxd.daemon[1047]: ==> Setting up ZFS (0.7)
May 01 08:10:32 net10.bizcooker.com lxd.daemon[1047]: ==> Escaping the systemd cgroups
May 01 08:10:32 net10.bizcooker.com lxd.daemon[1047]: ==> Escaping the systemd process resource limits
May 01 08:10:32 net10.bizcooker.com lxd.daemon[1047]: => Starting LXCFS
May 01 08:10:32 net10.bizcooker.com lxd.daemon[1047]: => Starting LXD
May 01 08:10:32 net10.bizcooker.com lxd.daemon[1047]: mount namespace: 5
May 01 08:10:32 net10.bizcooker.com lxd.daemon[1047]: hierarchies:
May 01 08:10:32 net10.bizcooker.com lxd.daemon[1047]: 0: fd: 6: devices
May 01 08:10:32 net10.bizcooker.com lxd.daemon[1047]: 1: fd: 7: freezer
May 01 08:10:32 net10.bizcooker.com lxd.daemon[1047]: 2: fd: 8: cpuset
May 01 08:10:32 net10.bizcooker.com lxd.daemon[1047]: 3: fd: 9: perf_event
May 01 08:10:32 net10.bizcooker.com lxd.daemon[1047]: 4: fd: 10: rdma
May 01 08:10:32 net10.bizcooker.com lxd.daemon[1047]: 5: fd: 11: blkio
May 01 08:10:32 net10.bizcooker.com lxd.daemon[1047]: 6: fd: 12: cpu,cpuacct
May 01 08:10:32 net10.bizcooker.com lxd.daemon[1047]: 7: fd: 13: hugetlb
May 01 08:10:32 net10.bizcooker.com lxd.daemon[1047]: 8: fd: 14: net_cls,net_prio
May 01 08:10:32 net10.bizcooker.com lxd.daemon[1047]: 9: fd: 15: memory
May 01 08:10:32 net10.bizcooker.com lxd.daemon[1047]: 10: fd: 16: pids
May 01 08:10:32 net10.bizcooker.com lxd.daemon[1047]: 11: fd: 17: name=systemd
May 01 08:10:32 net10.bizcooker.com lxd.daemon[1047]: 12: fd: 18: unified
May 01 08:20:33 net10.bizcooker.com systemd[1]: snap.lxd.daemon.service: Main process exited, code=exited, status=1/FAILURE
May 01 08:20:33 net10.bizcooker.com systemd[1]: snap.lxd.daemon.service: Failed with result 'exit-code'.
May 01 08:20:33 net10.bizcooker.com systemd[1]: snap.lxd.daemon.service: Service hold-off time over, scheduling restart.
May 01 08:20:33 net10.bizcooker.com systemd[1]: snap.lxd.daemon.service: Scheduled restart job, restart counter is at 1.
May 01 08:20:33 net10.bizcooker.com systemd[1]: Stopped Service for snap application lxd.daemon.
May 01 08:20:33 net10.bizcooker.com systemd[1]: Started Service for snap application lxd.daemon.
May 01 08:20:33 net10.bizcooker.com lxd.daemon[6776]: => Preparing the system
May 01 08:20:33 net10.bizcooker.com lxd.daemon[6776]: ==> Loading snap configuration
May 01 08:20:33 net10.bizcooker.com lxd.daemon[6776]: ==> Setting up mntns symlink
May 01 08:20:33 net10.bizcooker.com lxd.daemon[6776]: ==> Setting up kmod wrapper
May 01 08:20:33 net10.bizcooker.com lxd.daemon[6776]: ==> Preparing /boot
May 01 08:20:33 net10.bizcooker.com lxd.daemon[6776]: ==> Preparing a clean copy of /run
May 01 08:20:33 net10.bizcooker.com lxd.daemon[6776]: ==> Preparing a clean copy of /etc
May 01 08:20:33 net10.bizcooker.com lxd.daemon[6776]: ==> Setting up ceph configuration
May 01 08:20:33 net10.bizcooker.com lxd.daemon[6776]: ==> Setting up LVM configuration
May 01 08:20:33 net10.bizcooker.com lxd.daemon[6776]: ==> Rotating logs
May 01 08:20:34 net10.bizcooker.com lxd.daemon[6776]: ==> Setting up ZFS (0.7)
May 01 08:20:34 net10.bizcooker.com lxd.daemon[6776]: ==> Escaping the systemd cgroups
May 01 08:20:34 net10.bizcooker.com lxd.daemon[6776]: ==> Escaping the systemd process resource limits
May 01 08:20:34 net10.bizcooker.com lxd.daemon[6776]: => Re-using existing LXCFS
May 01 08:20:34 net10.bizcooker.com lxd.daemon[6776]: => Starting LXD
May 01 08:20:34 net10.bizcooker.com lxd.daemon[6776]: lvl=warn msg="CGroup memory swap accounting is disabled, swap limits will be ignored." t=2018-05-01T13:20:34+0000
May 01 08:30:34 net10.bizcooker.com systemd[1]: snap.lxd.daemon.service: Main process exited, code=exited, status=1/FAILURE
May 01 08:30:34 net10.bizcooker.com systemd[1]: snap.lxd.daemon.service: Failed with result 'exit-code'.
May 01 08:30:34 net10.bizcooker.com systemd[1]: snap.lxd.daemon.service: Service hold-off time over, scheduling restart.
May 01 08:30:34 net10.bizcooker.com systemd[1]: snap.lxd.daemon.service: Scheduled restart job, restart counter is at 2.
May 01 08:30:34 net10.bizcooker.com systemd[1]: Stopped Service for snap application lxd.daemon.
May 01 08:30:34 net10.bizcooker.com systemd[1]: Started Service for snap application lxd.daemon.
May 01 08:30:34 net10.bizcooker.com lxd.daemon[12938]: => Preparing the system
May 01 08:30:34 net10.bizcooker.com lxd.daemon[12938]: ==> Loading snap configuration
May 01 08:30:34 net10.bizcooker.com lxd.daemon[12938]: ==> Setting up mntns symlink
May 01 08:30:34 net10.bizcooker.com lxd.daemon[12938]: ==> Setting up kmod wrapper
May 01 08:30:34 net10.bizcooker.com lxd.daemon[12938]: ==> Preparing /boot
May 01 08:30:34 net10.bizcooker.com lxd.daemon[12938]: ==> Preparing a clean copy of /run
May 01 08:30:34 net10.bizcooker.com lxd.daemon[12938]: ==> Preparing a clean copy of /etc
May 01 08:30:34 net10.bizcooker.com lxd.daemon[12938]: ==> Setting up ceph configuration
May 01 08:30:34 net10.bizcooker.com lxd.daemon[12938]: ==> Setting up LVM configuration
May 01 08:30:34 net10.bizcooker.com lxd.daemon[12938]: ==> Rotating logs
May 01 08:30:34 net10.bizcooker.com lxd.daemon[12938]: ==> Setting up ZFS (0.7)
May 01 08:30:34 net10.bizcooker.com lxd.daemon[12938]: ==> Escaping the systemd cgroups
May 01 08:30:34 net10.bizcooker.com lxd.daemon[12938]: ==> Escaping the systemd process resource limits
May 01 08:30:34 net10.bizcooker.com lxd.daemon[12938]: => Re-using existing LXCFS
May 01 08:30:34 net10.bizcooker.com lxd.daemon[12938]: => Starting LXD
May 01 08:30:34 net10.bizcooker.com lxd.daemon[12938]: lvl=warn msg="CGroup memory swap accounting is disabled, swap limits will be ignored." t=2018-05-01T13:30:34+0000
May 01 08:34:20 net10.bizcooker.com systemd[1]: Reloading Service for snap application lxd.daemon.
May 01 08:34:20 net10.bizcooker.com systemd[1]: snap.lxd.daemon.service: Control process exited, code=exited status=1
May 01 08:34:20 net10.bizcooker.com systemd[1]: Reload failed for Service for snap application lxd.daemon.
May 01 08:34:34 net10.bizcooker.com systemd[1]: Reloading Service for snap application lxd.daemon.
May 01 08:34:34 net10.bizcooker.com systemd[1]: snap.lxd.daemon.service: Control process exited, code=exited status=1
May 01 08:34:34 net10.bizcooker.com systemd[1]: Reload failed for Service for snap application lxd.daemon.
May 01 08:40:34 net10.bizcooker.com systemd[1]: snap.lxd.daemon.service: Main process exited, code=exited, status=1/FAILURE
May 01 08:40:34 net10.bizcooker.com systemd[1]: snap.lxd.daemon.service: Failed with result 'exit-code'.
May 01 08:40:34 net10.bizcooker.com systemd[1]: snap.lxd.daemon.service: Service hold-off time over, scheduling restart.
May 01 08:40:34 net10.bizcooker.com systemd[1]: snap.lxd.daemon.service: Scheduled restart job, restart counter is at 3.
May 01 08:40:34 net10.bizcooker.com systemd[1]: Stopped Service for snap application lxd.daemon.
May 01 08:40:34 net10.bizcooker.com systemd[1]: Started Service for snap application lxd.daemon.
May 01 08:40:34 net10.bizcooker.com lxd.daemon[18639]: => Preparing the system
May 01 08:40:34 net10.bizcooker.com lxd.daemon[18639]: ==> Loading snap configuration
May 01 08:40:34 net10.bizcooker.com lxd.daemon[18639]: ==> Setting up mntns symlink
May 01 08:40:34 net10.bizcooker.com lxd.daemon[18639]: ==> Setting up kmod wrapper
May 01 08:40:34 net10.bizcooker.com lxd.daemon[18639]: ==> Preparing /boot
May 01 08:40:35 net10.bizcooker.com lxd.daemon[18639]: ==> Preparing a clean copy of /run
May 01 08:40:35 net10.bizcooker.com lxd.daemon[18639]: ==> Preparing a clean copy of /etc
May 01 08:40:35 net10.bizcooker.com lxd.daemon[18639]: ==> Setting up ceph configuration
May 01 08:40:35 net10.bizcooker.com lxd.daemon[18639]: ==> Setting up LVM configuration
May 01 08:40:35 net10.bizcooker.com lxd.daemon[18639]: ==> Rotating logs
May 01 08:40:35 net10.bizcooker.com lxd.daemon[18639]: ==> Setting up ZFS (0.7)
May 01 08:40:35 net10.bizcooker.com lxd.daemon[18639]: ==> Escaping the systemd cgroups
May 01 08:40:35 net10.bizcooker.com lxd.daemon[18639]: ==> Escaping the systemd process resource limits
May 01 08:40:35 net10.bizcooker.com lxd.daemon[18639]: => Re-using existing LXCFS
May 01 08:40:35 net10.bizcooker.com lxd.daemon[18639]: => Starting LXD
May 01 08:40:35 net10.bizcooker.com lxd.daemon[18639]: lvl=warn msg="CGroup memory swap accounting is disabled, swap limits will be ignored." t=2018-05-01T13:40:35+0000
May 01 08:50:33 net10.bizcooker.com systemd[1]: snap.lxd.daemon.service: Main process exited, code=exited, status=143/n/a
May 01 08:50:33 net10.bizcooker.com systemd[1]: snap.lxd.daemon.service: Failed with result 'exit-code'.
May 01 08:50:33 net10.bizcooker.com systemd[1]: snap.lxd.daemon.service: Service hold-off time over, scheduling restart.
May 01 08:50:33 net10.bizcooker.com systemd[1]: snap.lxd.daemon.service: Scheduled restart job, restart counter is at 4.
May 01 08:50:33 net10.bizcooker.com systemd[1]: Stopped Service for snap application lxd.daemon.
May 01 08:50:33 net10.bizcooker.com systemd[1]: Started Service for snap application lxd.daemon.
May 01 08:50:33 net10.bizcooker.com lxd.daemon[24257]: => Preparing the system
May 01 08:50:33 net10.bizcooker.com lxd.daemon[24257]: ==> Loading snap configuration
May 01 08:50:33 net10.bizcooker.com lxd.daemon[24257]: ==> Setting up mntns symlink
May 01 08:50:33 net10.bizcooker.com lxd.daemon[24257]: ==> Setting up kmod wrapper
May 01 08:50:33 net10.bizcooker.com lxd.daemon[24257]: ==> Preparing /boot
May 01 08:50:33 net10.bizcooker.com lxd.daemon[24257]: ==> Preparing a clean copy of /run
May 01 08:50:33 net10.bizcooker.com lxd.daemon[24257]: ==> Preparing a clean copy of /etc
May 01 08:50:33 net10.bizcooker.com lxd.daemon[24257]: ==> Setting up ceph configuration
May 01 08:50:33 net10.bizcooker.com lxd.daemon[24257]: ==> Setting up LVM configuration
May 01 08:50:33 net10.bizcooker.com lxd.daemon[24257]: ==> Rotating logs
May 01 08:50:34 net10.bizcooker.com lxd.daemon[24257]: ==> Setting up ZFS (0.7)
May 01 08:50:34 net10.bizcooker.com lxd.daemon[24257]: ==> Escaping the systemd cgroups
May 01 08:50:34 net10.bizcooker.com lxd.daemon[24257]: ==> Escaping the systemd process resource limits
May 01 08:50:34 net10.bizcooker.com lxd.daemon[24257]: => Re-using existing LXCFS
May 01 08:50:34 net10.bizcooker.com lxd.daemon[24257]: => Starting LXD
May 01 08:50:34 net10.bizcooker.com lxd.daemon[24257]: lvl=warn msg="CGroup memory swap accounting is disabled, swap limits will be ignored." t=2018-05-01T13:50:34+0000
May 01 08:52:29 net10.bizcooker.com systemd[1]: Reloading Service for snap application lxd.daemon.
May 01 08:52:29 net10.bizcooker.com systemd[1]: Reloaded Service for snap application lxd.daemon.
May 01 08:53:05 net10.bizcooker.com systemd[1]: Reloading Service for snap application lxd.daemon.
May 01 08:53:05 net10.bizcooker.com systemd[1]: snap.lxd.daemon.service: Control process exited, code=exited status=1
May 01 08:53:05 net10.bizcooker.com systemd[1]: Reload failed for Service for snap application lxd.daemon.
Ok, so LXD is pretty unhappy about something.
What do you get if you run:
lxd --debug --group lxd
Dude! You've got me laughing now.
LXD == unhappy. Yep.
net10 # lxd --debug --group lxd
DBUG[05-01|14:08:03] Connecting to a local LXD over a Unix socket
DBUG[05-01|14:08:03] Sending request to LXD etag= method=GET url=http://unix.socket/1.0
DBUG[05-01|14:08:03] Got response struct from LXD
DBUG[05-01|14:08:03]
{
"config": {
"core.https_address": ":8443",
"core.trust_password": true
},
"api_extensions": [
"storage_zfs_remove_snapshots",
"container_host_shutdown_timeout",
"container_stop_priority",
"container_syscall_filtering",
"auth_pki",
"container_last_used_at",
"etag",
"patch",
"usb_devices",
"https_allowed_credentials",
"image_compression_algorithm",
"directory_manipulation",
"container_cpu_time",
"storage_zfs_use_refquota",
"storage_lvm_mount_options",
"network",
"profile_usedby",
"container_push",
"container_exec_recording",
"certificate_update",
"container_exec_signal_handling",
"gpu_devices",
"container_image_properties",
"migration_progress",
"id_map",
"network_firewall_filtering",
"network_routes",
"storage",
"file_delete",
"file_append",
"network_dhcp_expiry",
"storage_lvm_vg_rename",
"storage_lvm_thinpool_rename",
"network_vlan",
"image_create_aliases",
"container_stateless_copy",
"container_only_migration",
"storage_zfs_clone_copy",
"unix_device_rename",
"storage_lvm_use_thinpool",
"storage_rsync_bwlimit",
"network_vxlan_interface",
"storage_btrfs_mount_options",
"entity_description",
"image_force_refresh",
"storage_lvm_lv_resizing",
"id_map_base",
"file_symlinks",
"container_push_target",
"network_vlan_physical",
"storage_images_delete",
"container_edit_metadata",
"container_snapshot_stateful_migration",
"storage_driver_ceph",
"storage_ceph_user_name",
"resource_limits",
"storage_volatile_initial_source",
"storage_ceph_force_osd_reuse",
"storage_block_filesystem_btrfs",
"resources",
"kernel_limits",
"storage_api_volume_rename",
"macaroon_authentication",
"network_sriov",
"console",
"restrict_devlxd",
"migration_pre_copy",
"infiniband",
"maas_network",
"devlxd_events",
"proxy",
"network_dhcp_gateway",
"file_get_symlink",
"network_leases",
"unix_device_hotplug",
"storage_api_local_volume_handling",
"operation_description",
"clustering",
"event_lifecycle",
"storage_api_remote_volume_handling",
"nvidia_runtime"
],
"api_status": "stable",
"api_version": "1.0",
"auth": "trusted",
"public": false,
"auth_methods": [
"tls"
],
"environment": {
"addresses": [
":8443"
],
"architectures": [
"x86_64",
"i686"
],
"certificate": "-----BEGIN CERTIFICATE-----\nMIIFTzCCAzegAwIBAgIQGsJzQkIE1hqVEWDhRnSHmTANBgkqhkiG9w0BAQsFADAz\nMRwwGgYDVQQKExNsaW51eGNvbnRhaW5lcnMub3JnMRMwEQYDVQQDDApyb290QG5l\ndDEwMB4XDTE4MDMwNTE4MjIyMVoXDTI4MDMwMjE4MjIyMVowMzEcMBoGA1UEChMT\nbGludXhjb250YWluZXJzLm9yZzETMBEGA1UEAwwKcm9vdEBuZXQxMDCCAiIwDQYJ\nKoZIhvcNAQEBBQADggIPADCCAgoCggIBANQGHoWaz4caeFjFOok26mENtjhrhm/G\nJ3+9JJEdJreC6Jm3hqo6pypZVgxPZ5sdYRWWr1cSWJ4QgvkuKGGfprHsSxyJ2loL\n2yROd4uvWcMUIe3f+1B5lrSR5aC+RRFCaC8g/G+EfOWJvZDNJmAqTBDkZqtrgUR/\ng5UA4WTtwcDo8P2a+LnlcjXrdNCnyJn2Uqy6M1i5SPaPffgLL709nlWyvb1tBUg3\nshs139ZeMaxI9ZbO3Ucilv8qgQXPLUuzJofDAa9So2wDhtaL2Vd4zZWWSZXoBU9f\nqa8R+cEMpm4yWpNQ1l3RymSsshidkG9YGbGp2bfYq5nFAU4vDQX4HFNWXRQh9HTJ\n2JZHai4gEgtsiz9XUW32HcFYPp2+I0lERwYAfuMvYo5Exyfufm3VsvTQhc+7X8Vh\nCylSqqPieKYCzd1P11gOU6kDkBGTUCddnR5JgbCQaW/uwkATcyvHiuWmb8KKjMDU\nyYmoGlG8UM1zUi2pZYEi+y1VnMaMAty3CzM78IgU2kmHbViiVXjMu3VDgxX5h+IM\nP+2CZagMcsZ+aEMTh4lcgacC0xubNJkDISWlDReaGqZlD9SIHpSFq785rRxlMahR\nifHA2TPS78PUF/IoNOb56NOjFcQovgWq+tgp/EM3kE6SdoXqLbSyVruS8swdp2hX\n6cwW2w6WNwiTAgMBAAGjXzBdMA4GA1UdDwEB/wQEAwIFoDATBgNVHSUEDDAKBggr\nBgEFBQcDATAMBgNVHRMBAf8EAjAAMCgGA1UdEQQhMB+CBW5ldDEwhwSeRX6vhxAm\nB1MAAGCTrwAAAAAAAAAAMA0GCSqGSIb3DQEBCwUAA4ICAQCabzlzG4+ISjk6e9Ca\nt0CZ6HpnYCZfd7uPc/Wqc2emfse1NOfpdTY/x98o3mSi74x4wi7JHZ/CAQM6XnSo\nGJeHmKO1UPGS8xzgRew7wOrxcQY1b/txTicpGbzJA5n1MXHzxzBLs1DM7gttUCwn\n9qDLWGeZAgECJnBjOk2Y/0Wt0PPWnzMn2T7POf9R4h2gIyZHRfdJOqXDI/RMf+Zh\nh5cw3AvS9zRpy92QOzhGy9QxdbFE+ZViO0NzA2RC8lMzZprqbrPWglGOo/vnK7Vx\n8wPECiCzdDxZ90yuK5Hk9dVx+haXlitQJLEQ0/S2D1ZA9SES6Wm+kzucANdRB0JX\ne9GfTPWw6uyFBYB5aBppjy06lSJXDTdi0lqjbKWFER/3ZI6Dy1xYOdv6Z2i3ZRhM\nAN615gIb21jCnZltQvj+ICp2UGh6nQxo/2y80nJCMSPBO/R4+qnIBqcEUh3lDKHt\n0OYlAirfYqKE6lqrbOHjpwIdsyMfWTY9/rppZotV7m2oacBEq5aMU+gZqbN57vnb\nV6kJbSs0e7ayjgeB/p5OIgARphSiuTCHNTBhG0pRu4JN4PzFLXVTzqkt1drQOKAp\n+SdJnr0hvPZSxPXRPOTer73PvyyCASwiUE3ED2G2ESF6kPjmHES+PGd6tiIpkFbs\n1sJpykR/jOxj6Z4gce/gJQbcHQ==\n-----END CERTIFICATE-----\n",
"certificate_fingerprint": "e61592926f6a55523d6edaf736b38aa71d75d53d409a8684b74c1a72057e4876",
"driver": "lxc",
"driver_version": "3.0.0",
"kernel": "Linux",
"kernel_architecture": "x86_64",
"kernel_version": "4.15.0-20-generic",
"server": "lxd",
"server_pid": 38205,
"server_version": "3.0.0",
"storage": "dir",
"storage_version": "1",
"server_clustered": false,
"server_name": "net10.bizcooker.com"
}
}
EROR[05-01|14:08:03] Failed to start the daemon: LXD is already running
INFO[05-01|14:08:03] Starting shutdown sequence
DBUG[05-01|14:08:03] Not unmounting temporary filesystems (containers are still running)
INFO[05-01|14:08:03] Saving simplestreams cache
INFO[05-01|14:08:03] Saved simplestreams cache
Error: LXD is already running
already running hmm, can you paste a full ps fauxww
? I wonder if there are some leftovers somewhere in there.
Rebooting now. To provide clean info, as killing of lxd waitready created many partially running containers + made the ps 1000s of lines long.
After reboot, with lxd waitready hung...
net10 # lxd --debug --group lxd
INFO[05-01|14:34:38] LXD 3.0.0 is starting in normal mode path=/var/snap/lxd/common/lxd
INFO[05-01|14:34:38] Kernel uid/gid map:
INFO[05-01|14:34:38] - u 0 0 4294967295
INFO[05-01|14:34:38] - g 0 0 4294967295
INFO[05-01|14:34:38] Configured LXD uid/gid map:
INFO[05-01|14:34:38] - u 0 1000000 1000000000
INFO[05-01|14:34:38] - g 0 1000000 1000000000
WARN[05-01|14:34:38] CGroup memory swap accounting is disabled, swap limits will be ignored.
INFO[05-01|14:34:38] Initializing local database
INFO[05-01|14:34:38] Initializing database gateway
INFO[05-01|14:34:38] Start database node address= id=1
INFO[05-01|14:34:38] Raft: Initial configuration (index=1): [{Suffrage:Voter ID:1 Address:0}]
INFO[05-01|14:34:38] Raft: Node at 0 [Leader] entering Leader state
INFO[05-01|14:34:38] LXD isn't socket activated
INFO[05-01|14:34:38] Starting /dev/lxd handler:
INFO[05-01|14:34:38] - binding devlxd socket socket=/var/snap/lxd/common/lxd/devlxd/sock
INFO[05-01|14:34:38] REST API daemon:
INFO[05-01|14:34:38] - binding Unix socket socket=/var/snap/lxd/common/lxd/unix.socket
INFO[05-01|14:34:38] - binding TCP socket socket=[::]:8443
INFO[05-01|14:34:38] Initializing global database
# hangs here, likely waiting for lxd waitready to end
lxc command hangs same place...
net10 # lxd --debug --group lxd
DBUG[05-01|14:35:33] Connecting to a local LXD over a Unix socket
DBUG[05-01|14:35:33] Sending request to LXD etag= method=GET url=http://unix.socket/1.0
INFO[05-01|14:35:33] LXD 3.0.0 is starting in normal mode path=/var/snap/lxd/common/lxd
INFO[05-01|14:35:33] Kernel uid/gid map:
INFO[05-01|14:35:33] - u 0 0 4294967295
INFO[05-01|14:35:33] - g 0 0 4294967295
INFO[05-01|14:35:33] Configured LXD uid/gid map:
INFO[05-01|14:35:33] - u 0 1000000 1000000000
INFO[05-01|14:35:33] - g 0 1000000 1000000000
WARN[05-01|14:35:33] CGroup memory swap accounting is disabled, swap limits will be ignored.
INFO[05-01|14:35:33] Initializing local database
INFO[05-01|14:35:33] Initializing database gateway
INFO[05-01|14:35:33] Start database node address= id=1
INFO[05-01|14:35:33] Raft: Initial configuration (index=1): [{Suffrage:Voter ID:1 Address:0}]
INFO[05-01|14:35:33] Raft: Node at 0 [Leader] entering Leader state
INFO[05-01|14:35:33] LXD isn't socket activated
DBUG[05-01|14:35:33] Connecting to a local LXD over a Unix socket
DBUG[05-01|14:35:33] Sending request to LXD etag= method=GET url=http://unix.socket/1.0
DBUG[05-01|14:35:33] Detected stale unix socket, deleting
DBUG[05-01|14:35:33] Detected stale unix socket, deleting
INFO[05-01|14:35:33] Starting /dev/lxd handler:
INFO[05-01|14:35:33] - binding devlxd socket socket=/var/snap/lxd/common/lxd/devlxd/sock
INFO[05-01|14:35:33] REST API daemon:
INFO[05-01|14:35:33] - binding Unix socket socket=/var/snap/lxd/common/lxd/unix.socket
INFO[05-01|14:35:33] - binding TCP socket socket=[::]:8443
INFO[05-01|14:35:33] Initializing global database
Bottom line of problem seems to be /var/snap/lxd/common/lxd/unix.socket never gets created.
Let me know how to debug the process which creates the socket + maybe that will net some better information.
lxd --debug finally came back. Took 2+ minutes to create socket.
net10 # r
lxd --debug --group lxd
INFO[05-01|14:38:55] LXD 3.0.0 is starting in normal mode path=/var/snap/lxd/common/lxd
INFO[05-01|14:38:55] Kernel uid/gid map:
INFO[05-01|14:38:55] - u 0 0 4294967295
INFO[05-01|14:38:55] - g 0 0 4294967295
INFO[05-01|14:38:55] Configured LXD uid/gid map:
INFO[05-01|14:38:55] - u 0 1000000 1000000000
INFO[05-01|14:38:55] - g 0 1000000 1000000000
WARN[05-01|14:38:55] CGroup memory swap accounting is disabled, swap limits will be ignored.
INFO[05-01|14:38:55] Initializing local database
INFO[05-01|14:38:55] Initializing database gateway
INFO[05-01|14:38:55] Start database node address= id=1
INFO[05-01|14:38:55] Raft: Initial configuration (index=1): [{Suffrage:Voter ID:1 Address:0}]
INFO[05-01|14:38:55] Raft: Node at 0 [Leader] entering Leader state
INFO[05-01|14:38:55] LXD isn't socket activated
INFO[05-01|14:38:55] Starting /dev/lxd handler:
INFO[05-01|14:38:55] - binding devlxd socket socket=/var/snap/lxd/common/lxd/devlxd/sock
INFO[05-01|14:38:55] REST API daemon:
INFO[05-01|14:38:55] - binding Unix socket socket=/var/snap/lxd/common/lxd/unix.socket
INFO[05-01|14:38:55] - binding TCP socket socket=[::]:8443
INFO[05-01|14:38:55] Initializing global database
INFO[05-01|14:40:11] Initializing storage pools
DBUG[05-01|14:40:11] Initializing and checking storage pool "default".
DBUG[05-01|14:40:11] Initializing a DIR driver.
DBUG[05-01|14:40:11] Checking DIR storage pool "default".
DBUG[05-01|14:40:11] Initializing a DIR driver.
INFO[05-01|14:40:11] Initializing networks
INFO[05-01|14:40:12] Loading configuration
DBUG[05-01|14:40:12] Initialized inotify with file descriptor 16
INFO[05-01|14:40:12] Pruning expired images
DBUG[05-01|14:40:12] handling ip=@ method=GET url=/1.0
INFO[05-01|14:40:12] Done pruning expired images
INFO[05-01|14:40:12] Expiring log files
INFO[05-01|14:40:12] Updating instance types
DBUG[05-01|14:40:12] handling ip=@ method=GET url=/internal/ready
INFO[05-01|14:40:12] Updating images
DBUG[05-01|14:40:12] Processing image alias=18.04 fp=b36ec647e374da4816104a98807633a2cc387488083d3776557081c4d0333618 protocol=simplestreams server=https://cloud-images.ubuntu.com/releases
DBUG[05-01|14:40:12] Connecting to a remote simplestreams server
INFO[05-01|14:40:13] Done expiring log files
DBUG[05-01|14:40:13] handling ip=@ method=GET url=/1.0
DBUG[05-01|14:40:13] handling ip=@ method=GET url=/internal/ready
DBUG[05-01|14:40:14] handling ip=@ method=GET url=/1.0
DBUG[05-01|14:40:14] handling ip=@ method=GET url=/1.0
DBUG[05-01|14:40:14] handling ip=@ method=GET url=/internal/ready
DBUG[05-01|14:40:14] handling ip=@ method=GET url=/1.0
DBUG[05-01|14:40:14] handling ip=@ method=GET url=/1.0
DBUG[05-01|14:40:14] handling ip=@ method=GET url=/internal/ready
INFO[05-01|14:40:15] Done updating instance types
DBUG[05-01|14:40:15] handling ip=@ method=GET url=/1.0
DBUG[05-01|14:40:15] handling ip=@ method=GET url=/1.0
DBUG[05-01|14:40:15] handling ip=@ method=GET url=/internal/ready
DBUG[05-01|14:40:16] handling ip=@ method=GET url=/1.0
DBUG[05-01|14:40:16] handling ip=@ method=GET url=/1.0
DBUG[05-01|14:40:16] handling ip=@ method=GET url=/internal/ready
DBUG[05-01|14:40:16] Image already exists in the db image=b36ec647e374da4816104a98807633a2cc387488083d3776557081c4d0333618
DBUG[05-01|14:40:16] Already up to date fp=b36ec647e374da4816104a98807633a2cc387488083d3776557081c4d0333618
DBUG[05-01|14:40:16] Processing image alias=17.10 fp=f7febb8cbebc6aa8a993eb1ce534963a6b288fde23b9594bb3ba4560704dd65c protocol=simplestreams server=https://cloud-images.ubuntu.com/releases
DBUG[05-01|14:40:16] Using SimpleStreams cache entry expiry=2018-05-01T15:40:16+0000 server=https://cloud-images.ubuntu.com/releases
DBUG[05-01|14:40:16] Image already exists in the db image=f7febb8cbebc6aa8a993eb1ce534963a6b288fde23b9594bb3ba4560704dd65c
DBUG[05-01|14:40:16] Already up to date fp=f7febb8cbebc6aa8a993eb1ce534963a6b288fde23b9594bb3ba4560704dd65c
INFO[05-01|14:40:16] Done updating images
DBUG[05-01|14:40:17] handling ip=@ method=GET url=/1.0
DBUG[05-01|14:40:17] handling ip=@ method=GET url=/1.0
DBUG[05-01|14:40:17] handling ip=@ method=GET url=/internal/ready
Sigh... And socket is unusable...
net10 # ll /var/snap/lxd/common/lxd/unix.socket
srw-rw---- 1 root lxd 0 May 1 09:42 /var/snap/lxd/common/lxd/unix.socket=
net10 # lxc list
Error: Get http://unix.socket/1.0: dial unix /var/snap/lxd/common/lxd/unix.socket: connect: connection refused
And you still have that first shell running right, you didn't ctrl+c that shell you're running "lxd --debug --group lxd" from?
Looks like your database took two minutes to come online which is a bit unusual, @freeekanayaka is already looking into a similar issue.
Yes I left the original - lxd --debug --group lxd - command running, which is when I saw the 2min lag.
If I run the above command again, then I get the same hang at Initializing global database again.
Running this command again, I get...
net10# lxd --debug --group lxd
INFO[05-01|14:51:04] LXD 3.0.0 is starting in normal mode path=/var/snap/lxd/common/lxd
INFO[05-01|14:51:04] Kernel uid/gid map:
INFO[05-01|14:51:04] - u 0 0 4294967295
INFO[05-01|14:51:04] - g 0 0 4294967295
INFO[05-01|14:51:04] Configured LXD uid/gid map:
INFO[05-01|14:51:04] - u 0 1000000 1000000000
INFO[05-01|14:51:04] - g 0 1000000 1000000000
WARN[05-01|14:51:04] CGroup memory swap accounting is disabled, swap limits will be ignored.
INFO[05-01|14:51:04] Initializing local database
INFO[05-01|14:51:04] Initializing database gateway
INFO[05-01|14:51:04] Start database node address= id=1
INFO[05-01|14:51:04] Raft: Initial configuration (index=1): [{Suffrage:Voter ID:1 Address:0}]
INFO[05-01|14:51:04] Raft: Node at 0 [Leader] entering Leader state
INFO[05-01|14:51:04] LXD isn't socket activated
INFO[05-01|14:51:04] Starting /dev/lxd handler:
INFO[05-01|14:51:04] - binding devlxd socket socket=/var/snap/lxd/common/lxd/devlxd/sock
INFO[05-01|14:51:04] REST API daemon:
INFO[05-01|14:51:04] - binding Unix socket socket=/var/snap/lxd/common/lxd/unix.socket
INFO[05-01|14:51:04] - binding TCP socket socket=[::]:8443
INFO[05-01|14:51:04] Initializing global database
2018/05/01 14:51:50 http: TLS handshake error from 196.52.43.109:6666: tls: client offered an unsupported, maximum protocol version of 302
DBUG[05-01|14:52:04] Database error: failed to begin transaction: gRPC BEGIN response error: rpc error: code = Unknown desc = failed to handle BEGIN request: FSM out of sync: timed out enqueuing operation
EROR[05-01|14:52:04] Failed to start the daemon: failed to open cluster database: failed to ensure schema: failed to begin transaction: gRPC BEGIN response error: rpc error: code = Unknown desc = failed to handle BEGIN request: FSM out of sync: timed out enqueuing operation
INFO[05-01|14:52:04] Starting shutdown sequence
INFO[05-01|14:52:04] Stopping REST API handler:
INFO[05-01|14:52:04] - closing socket socket=[::]:8443
INFO[05-01|14:52:04] - closing socket socket=/var/snap/lxd/common/lxd/unix.socket
INFO[05-01|14:52:04] Stopping /dev/lxd handler
INFO[05-01|14:52:04] - closing socket socket=/var/snap/lxd/common/lxd/devlxd/sock
INFO[05-01|14:52:04] Stop database gateway
INFO[05-01|14:52:04] Stop raft instance
INFO[05-01|14:52:05] Stopping REST API handler:
INFO[05-01|14:52:05] Stopping /dev/lxd handler
INFO[05-01|14:52:05] Stopping REST API handler:
INFO[05-01|14:52:05] Stopping /dev/lxd handler
DBUG[05-01|14:52:05] Not unmounting temporary filesystems (containers are still running)
INFO[05-01|14:52:05] Saving simplestreams cache
INFO[05-01|14:52:05] Saved simplestreams cache
Error: failed to open cluster database: failed to ensure schema: failed to begin transaction: gRPC BEGIN response error: rpc error: code = Unknown desc = failed to handle BEGIN request: FSM out of sync: timed out enqueuing operation
Ah... I also see lxd.db is missing.
net10 # ll /var/snap/lxd/common/lxd
total 136K
drwx------ 2 root root 4.0K Mar 5 12:58 cache/
drwx--x--x 2 root root 4.0K Apr 30 13:57 containers/
drwx------ 3 root root 4.0K Apr 30 06:14 database/
drwx--x--x 18 root root 4.0K Apr 30 13:57 devices/
drwxr-xr-x 2 root root 4.0K Mar 5 12:22 devlxd/
drwx------ 2 root root 4.0K Mar 5 12:22 disks/
drwx------ 2 root root 4.0K Apr 30 13:54 images/
drwx------ 18 root root 4.0K May 1 09:40 logs/
drwx--x--x 3 root root 4.0K Mar 5 12:24 networks/
drwx------ 4 root root 4.0K Mar 8 07:59 security/
drwx--x--x 2 root root 4.0K Mar 5 12:22 shmounts/
drwx------ 2 root root 4.0K Mar 5 12:22 snapshots/
drwx--x--x 3 root root 4.0K Mar 5 12:24 storage-pools/
-rw-r--r-- 1 root root 74K Apr 3 03:52 lxd.db.bak
-rw-r--r-- 1 root root 1.9K Mar 5 12:22 server.crt
-rw------- 1 root root 3.2K Mar 5 12:22 server.key
Tried with LXD v2.21 on Ubuntu 16.04 and LXD v3.0.0 on 18.04 (system upgraded from 16.04)
Networking does not come up and container does not get an Ip assigned on my network bridge.
On both my 16.04 and 18.04 host system, a xenial image comes up just fine.
I have tried provisioning from
ubuntu:bionic
as well asimages:ubuntu/bionic/amd64
with identical results./var/log/syslog
on the host shows in all cases lines similar toThese lines are not present in syslog when provisioning other versions of Ubuntu (Xenial/Zesty). Interestingly upgrading an existing Xenial container to Bionic does not cause any networking issues.
Without knowing much about apparmor, I am assuming that the
DENIED ... networkd
line is an indicator of the culprit here. Any assistance would be much appreciated :)