Closed alborotogarcia closed 3 years ago
I'm not sure if this is related, but avahi appears to be attempting to manage all your container interfaces by joining them to multicast groups. I've seen this cause problems in the past (example: https://github.com/k3s-io/k3s/issues/2599#issuecomment-828659931) - can you try disabling or uninstalling avahi from your host?
Still getting the same error disabling both avahi-daemon service and socket..
pi@kpi4:~ $ systemctl status avahi-daemon.service
● avahi-daemon.service - Avahi mDNS/DNS-SD Stack
Loaded: loaded (/lib/systemd/system/avahi-daemon.service; disabled; vendor preset: enabled)
Active: inactive (dead) since Tue 2021-07-06 22:56:55 CEST; 48min ago
Main PID: 475 (code=exited, status=0/SUCCESS)
Status: "avahi-daemon 0.7 starting up."
Jul 05 21:25:21 kpi4 avahi-daemon[475]: New relevant interface eth0.IPv4 for mDNS.
Jul 05 21:25:21 kpi4 avahi-daemon[475]: Registering new address record for 192.168.1.223 on eth0.IPv4.
Jul 06 22:56:55 kpi4 avahi-daemon[475]: Got SIGTERM, quitting.
Jul 06 22:56:55 kpi4 avahi-daemon[475]: Leaving mDNS multicast group on interface docker0.IPv4 with address 172.17.0.1.
Jul 06 22:56:55 kpi4 systemd[1]: Stopping Avahi mDNS/DNS-SD Stack...
Jul 06 22:56:55 kpi4 avahi-daemon[475]: Leaving mDNS multicast group on interface eth0.IPv6 with address fe80::eae7:f8c5:2fc:d0df.
Jul 06 22:56:55 kpi4 avahi-daemon[475]: Leaving mDNS multicast group on interface eth0.IPv4 with address 192.168.1.223.
Jul 06 22:56:55 kpi4 avahi-daemon[475]: avahi-daemon 0.7 exiting.
Jul 06 22:56:55 kpi4 systemd[1]: avahi-daemon.service: Succeeded.
Jul 06 22:56:55 kpi4 systemd[1]: Stopped Avahi mDNS/DNS-SD Stack.
pi@kpi4:~ $ systemctl status avahi-daemon.socket
● avahi-daemon.socket - Avahi mDNS/DNS-SD Stack Activation Socket
Loaded: loaded (/lib/systemd/system/avahi-daemon.socket; disabled; vendor preset: enabled)
Active: inactive (dead) since Tue 2021-07-06 22:57:23 CEST; 47min ago
Listen: /run/avahi-daemon/socket (Stream)
Jul 05 21:25:09 kpi4 systemd[1]: Listening on Avahi mDNS/DNS-SD Stack Activation Socket.
Jul 06 22:57:23 kpi4 systemd[1]: avahi-daemon.socket: Succeeded.
Jul 06 22:57:23 kpi4 systemd[1]: Closed Avahi mDNS/DNS-SD Stack Activation Socket.
pi@kpi4:~ $ journalctl -xb0 -f -u k3s
(*v1.MountPropagationMode)(nil), SubPathExpr:""}, v1.VolumeMount{Name:"host", ReadOnly:false, MountPath:"/rootfs", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(0x400d34fbf0), SubPathExpr:""}, v1.VolumeMount{Name:"lib-modules", ReadOnly:true, MountPath:"/lib/modules", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil), SubPathExpr:""}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), StartupProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(0x400d34fc00), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"IfNotPresent", SecurityContext:(*v1.SecurityContext)(0x4014f49ce0), Stdin:false, StdinOnce:false, TTY:false}}, EphemeralContainers:[]v1.EphemeralContainer(nil), RestartPolicy:"Always", TerminationGracePeriodSeconds:(*int64)(0x4011fb2bd8), ActiveDeadlineSeconds:(*int64)(nil), DNSPolicy:"ClusterFirst", NodeSelector:map[string]string(nil), ServiceAccountName:"longhorn-service-account", DeprecatedServiceAccount:"longhorn-service-account", AutomountServiceAccountToken:(*bool)(nil), NodeName:"", HostNetwork:false, HostPID:true, HostIPC:false, ShareProcessNamespace:(*bool)(nil), SecurityContext:(*v1.PodSecurityContext)(0x4008c58700), ImagePullSecrets:[]v1.LocalObjectReference{v1.LocalObjectReference{Name:"regcred"}}, Hostname:"", Subdomain:"", Affinity:(*v1.Affinity)(nil), SchedulerName:"default-scheduler", Tolerations:[]v1.Toleration(nil), HostAliases:[]v1.HostAlias(nil), PriorityClassName:"", Priority:(*int32)(nil), DNSConfig:(*v1.PodDNSConfig)(nil), ReadinessGates:[]v1.PodReadinessGate(nil), RuntimeClassName:(*string)(nil), EnableServiceLinks:(*bool)(nil), PreemptionPolicy:(*v1.PreemptionPolicy)(nil), Overhead:v1.ResourceList(nil), TopologySpreadConstraints:[]v1.TopologySpreadConstraint(nil), SetHostnameAsFQDN:(*bool)(nil)}}, UpdateStrategy:v1.DaemonSetUpdateStrategy{Type:"RollingUpdate", RollingUpdate:(*v1.RollingUpdateDaemonSet)(0x400d34fc40)}, MinReadySeconds:0, RevisionHistoryLimit:(*int32)(0x4011fb2bf8)}, Status:v1.DaemonSetStatus{CurrentNumberScheduled:0, NumberMisscheduled:0, DesiredNumberScheduled:0, NumberReady:0, ObservedGeneration:0, UpdatedNumberScheduled:0, NumberAvailable:0, NumberUnavailable:0, CollisionCount:(*int32)(nil), Conditions:[]v1.DaemonSetCondition(nil)}}: Operation cannot be fulfilled on daemonsets.apps "longhorn-csi-plugin": the object has been modified; please apply your changes to the latest version and try again
Jul 06 23:05:15 kpi4 k3s[19003]: I0706 23:05:15.550283 19003 trace.go:205] Trace[670189661]: "GuaranteedUpdate etcd3" type:*apps.ReplicaSet (06-Jul-2021 23:05:14.597) (total time: 953ms):
Jul 06 23:05:15 kpi4 k3s[19003]: Trace[670189661]: ---"Transaction committed" 934ms (23:05:00.550)
Jul 06 23:05:15 kpi4 k3s[19003]: Trace[670189661]: [953.075615ms] [953.075615ms] END
Jul 06 23:05:15 kpi4 k3s[19003]: I0706 23:05:15.550727 19003 trace.go:205] Trace[1448889774]: "Update" url:/apis/apps/v1/namespaces/longhorn-system/replicasets/csi-resizer-6859cb947f/status,user-agent:k3s/v1.21.2+k3s1 (linux/arm64) kubernetes/5a67e8d/system:serviceaccount:kube-system:replicaset-controller,client:127.0.0.1,accept:application/vnd.kubernetes.protobuf, */*,protocol:HTTP/2.0 (06-Jul-2021 23:05:14.590) (total time: 959ms):
Jul 06 23:05:15 kpi4 k3s[19003]: Trace[1448889774]: ---"Object stored in database" 956ms (23:05:00.550)
Jul 06 23:05:15 kpi4 k3s[19003]: Trace[1448889774]: [959.885742ms] [959.885742ms] END
Jul 06 23:05:15 kpi4 k3s[19003]: I0706 23:05:15.883988 19003 kubelet_getters.go:300] "Path does not exist" path="/var/lib/kubelet/pods/4bc6e79a-e236-40f1-b3da-8e0e3a2652a4/volumes"
Jul 06 23:05:16 kpi4 k3s[19003]: E0706 23:05:16.476420 19003 cadvisor_stats_provider.go:415] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods/besteffort/pod58ecc470-54e1-4c2e-9541-9d73991a63b6\": RecentStats: unable to find data in memory cache], [\"/kubepods/besteffort/pod5d164fe7-52c0-4a82-95b7-b98cd0e98b45\": RecentStats: unable to find data in memory cache], [\"/kubepods/besteffort/pod1e49ea54-5e1d-45f4-bb66-a0edc9fd5d33\": RecentStats: unable to find data in memory cache]"
Jul 06 23:05:18 kpi4 k3s[19003]: I0706 23:05:18.151717 19003 trace.go:205] Trace[1400083480]: "Create" url:/api/v1/namespaces/longhorn-system/events,user-agent:k3s/v1.21.2+k3s1 (linux/arm64) kubernetes/5a67e8d,client:127.0.0.1,accept:application/vnd.kubernetes.protobuf,application/json,protocol:HTTP/1.1 (06-Jul-2021 23:05:17.296) (total time: 854ms):
Jul 06 23:05:18 kpi4 k3s[19003]: Trace[1400083480]: ---"Object stored in database" 854ms (23:05:00.151)
Jul 06 23:05:18 kpi4 k3s[19003]: Trace[1400083480]: [854.805635ms] [854.805635ms] END
Jul 06 23:05:22 kpi4 k3s[19003]: E0706 23:05:22.190072 19003 cadvisor_stats_provider.go:415] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods/besteffort/pod1e49ea54-5e1d-45f4-bb66-a0edc9fd5d33\": RecentStats: unable to find data in memory cache], [\"/kubepods/besteffort/pod5d164fe7-52c0-4a82-95b7-b98cd0e98b45\": RecentStats: unable to find data in memory cache], [\"/kubepods/besteffort/pod58ecc470-54e1-4c2e-9541-9d73991a63b6\": RecentStats: unable to find data in memory cache]"
Jul 06 23:05:28 kpi4 k3s[19003]: I0706 23:05:28.747872 19003 trace.go:205] Trace[442557945]: "GuaranteedUpdate etcd3" type:*coordination.Lease (06-Jul-2021 23:05:24.855) (total time: 3891ms):
Jul 06 23:05:28 kpi4 k3s[19003]: Trace[442557945]: ---"Transaction committed" 3890ms (23:05:00.747)
Jul 06 23:05:28 kpi4 k3s[19003]: Trace[442557945]: [3.891711395s] [3.891711395s] END
Jul 06 23:05:28 kpi4 k3s[19003]: I0706 23:05:28.748673 19003 trace.go:205] Trace[1405767360]: "Update" url:/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/cert-manager-cainjector-leader-election,user-agent:cainjector/v0.0.0 (linux/arm64) kubernetes/$Format/leader-election,client:192.168.1.224,accept:application/json, */*,protocol:HTTP/1.1 (06-Jul-2021 23:05:24.855) (total time: 3893ms):
Jul 06 23:05:28 kpi4 k3s[19003]: Trace[1405767360]: ---"Object stored in database" 3892ms (23:05:00.747)
Jul 06 23:05:28 kpi4 k3s[19003]: Trace[1405767360]: [3.893193845s] [3.893193845s] END
Jul 06 23:05:31 kpi4 k3s[19003]: I0706 23:05:31.728917 19003 trace.go:205] Trace[624710701]: "GuaranteedUpdate etcd3" type:*core.Node (06-Jul-2021 23:05:30.449) (total time: 1278ms):
Jul 06 23:05:31 kpi4 k3s[19003]: Trace[624710701]: ---"Transaction committed" 1274ms (23:05:00.728)
Jul 06 23:05:31 kpi4 k3s[19003]: Trace[624710701]: [1.278964291s] [1.278964291s] END
Jul 06 23:05:31 kpi4 k3s[19003]: I0706 23:05:31.729974 19003 trace.go:205] Trace[2062471273]: "Patch" url:/api/v1/nodes/knode2/status,user-agent:k3s/v1.21.2+k3s1 (linux/arm64) kubernetes/5a67e8d,client:192.168.1.225,accept:application/vnd.kubernetes.protobuf,application/json,protocol:HTTP/1.1 (06-Jul-2021 23:05:30.449) (total time: 1280ms):
Jul 06 23:05:31 kpi4 k3s[19003]: Trace[2062471273]: ---"Object stored in database" 1275ms (23:05:00.729)
Jul 06 23:05:31 kpi4 k3s[19003]: Trace[2062471273]: [1.280333798s] [1.280333798s] END
Jul 06 23:05:33 kpi4 k3s[19003]: I0706 23:05:33.028928 19003 trace.go:205] Trace[1160769760]: "GuaranteedUpdate etcd3" type:*coordination.Lease (06-Jul-2021 23:05:31.846) (total time: 1181ms):
Jul 06 23:05:33 kpi4 k3s[19003]: Trace[1160769760]: ---"Transaction committed" 1180ms (23:05:00.028)
Jul 06 23:05:33 kpi4 k3s[19003]: Trace[1160769760]: [1.181863717s] [1.181863717s] END
Jul 06 23:05:33 kpi4 k3s[19003]: I0706 23:05:33.029246 19003 trace.go:205] Trace[1689466342]: "Update" url:/apis/coordination.k8s.io/v1/namespaces/longhorn-system/leases/external-attacher-leader-driver-longhorn-io,user-agent:csi-attacher/v0.0.0 (linux/arm64) kubernetes/$Format,client:192.168.1.224,accept:application/json, */*,protocol:HTTP/1.1 (06-Jul-2021 23:05:31.846) (total time: 1182ms):
Jul 06 23:05:33 kpi4 k3s[19003]: Trace[1689466342]: ---"Object stored in database" 1182ms (23:05:00.028)
Jul 06 23:05:33 kpi4 k3s[19003]: Trace[1689466342]: [1.182969263s] [1.182969263s] END
Jul 06 23:05:54 kpi4 k3s[19003]: I0706 23:05:54.093440 19003 scope.go:111] "RemoveContainer" containerID="883f8773c1d547c3f05d9609bc5ca27b596800960e78be9bf97828f0b3e70039"
Jul 06 23:06:14 kpi4 k3s[19003]: I0706 23:06:14.062223 19003 trace.go:205] Trace[1030437697]: "GuaranteedUpdate etcd3" type:*core.ConfigMap (06-Jul-2021 23:06:12.316) (total time: 1745ms):
Jul 06 23:06:14 kpi4 k3s[19003]: Trace[1030437697]: ---"Transaction committed" 1744ms (23:06:00.062)
Jul 06 23:06:14 kpi4 k3s[19003]: Trace[1030437697]: [1.745158718s] [1.745158718s] END
Jul 06 23:06:14 kpi4 k3s[19003]: I0706 23:06:14.062559 19003 trace.go:205] Trace[1057281326]: "Update" url:/api/v1/namespaces/kube-system/configmaps/cert-manager-cainjector-leader-election,user-agent:cainjector/v0.0.0 (linux/arm64) kubernetes/$Format/leader-election,client:192.168.1.224,accept:application/json, */*,protocol:HTTP/1.1 (06-Jul-2021 23:06:12.316) (total time: 1745ms):
Jul 06 23:06:14 kpi4 k3s[19003]: Trace[1057281326]: ---"Object stored in database" 1745ms (23:06:00.062)
Jul 06 23:06:14 kpi4 k3s[19003]: Trace[1057281326]: [1.745982452s] [1.745982452s] END
Jul 06 23:06:15 kpi4 k3s[19003]: I0706 23:06:15.308475 19003 trace.go:205] Trace[1033211365]: "GuaranteedUpdate etcd3" type:*coordination.Lease (06-Jul-2021 23:06:14.086) (total time: 1221ms):
Jul 06 23:06:15 kpi4 k3s[19003]: Trace[1033211365]: ---"Transaction committed" 1220ms (23:06:00.308)
Jul 06 23:06:15 kpi4 k3s[19003]: Trace[1033211365]: [1.221603821s] [1.221603821s] END
Jul 06 23:06:15 kpi4 k3s[19003]: I0706 23:06:15.308802 19003 trace.go:205] Trace[1515918817]: "Update" url:/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/cert-manager-cainjector-leader-election,user-agent:cainjector/v0.0.0 (linux/arm64) kubernetes/$Format/leader-election,client:192.168.1.224,accept:application/json, */*,protocol:HTTP/1.1 (06-Jul-2021 23:06:14.086) (total time: 1222ms):
Jul 06 23:06:15 kpi4 k3s[19003]: Trace[1515918817]: ---"Object stored in database" 1221ms (23:06:00.308)
Jul 06 23:06:15 kpi4 k3s[19003]: Trace[1515918817]: [1.222424648s] [1.222424648s] END
Jul 06 23:06:24 kpi4 k3s[19003]: I0706 23:06:24.800090 19003 csi_plugin.go:99] kubernetes.io/csi: Trying to validate a new CSI Driver with name: driver.longhorn.io endpoint: /var/lib/kubelet/plugins/driver.longhorn.io/csi.sock versions: 1.0.0
Jul 06 23:06:24 kpi4 k3s[19003]: I0706 23:06:24.800180 19003 csi_plugin.go:112] kubernetes.io/csi: Register new plugin with name: driver.longhorn.io at endpoint: /var/lib/kubelet/plugins/driver.longhorn.io/csi.sock
Jul 06 23:06:24 kpi4 k3s[19003]: E0706 23:06:24.888921 19003 nodeinfomanager.go:566] Invalid attach limit value 0 cannot be added to CSINode object for "driver.longhorn.io"
Jul 06 23:06:39 kpi4 k3s[19003]: I0706 23:06:39.501865 19003 trace.go:205] Trace[1052799017]: "GuaranteedUpdate etcd3" type:*core.ConfigMap (06-Jul-2021 23:06:36.197) (total time: 3304ms):
Jul 06 23:06:39 kpi4 k3s[19003]: Trace[1052799017]: ---"Transaction committed" 3303ms (23:06:00.501)
Jul 06 23:06:39 kpi4 k3s[19003]: Trace[1052799017]: [3.304448084s] [3.304448084s] END
Jul 06 23:06:39 kpi4 k3s[19003]: I0706 23:06:39.502403 19003 trace.go:205] Trace[393239039]: "Update" url:/api/v1/namespaces/kube-system/configmaps/cert-manager-cainjector-leader-election,user-agent:cainjector/v0.0.0 (linux/arm64) kubernetes/$Format/leader-election,client:192.168.1.224,accept:application/json, */*,protocol:HTTP/1.1 (06-Jul-2021 23:06:36.196) (total time: 3305ms):
Jul 06 23:06:39 kpi4 k3s[19003]: Trace[393239039]: ---"Object stored in database" 3304ms (23:06:00.501)
Jul 06 23:06:39 kpi4 k3s[19003]: Trace[393239039]: [3.305488445s] [3.305488445s] END
Jul 06 23:06:40 kpi4 k3s[19003]: I0706 23:06:40.828365 19003 trace.go:205] Trace[1240402593]: "GuaranteedUpdate etcd3" type:*coordination.Lease (06-Jul-2021 23:06:39.534) (total time: 1293ms):
Jul 06 23:06:40 kpi4 k3s[19003]: Trace[1240402593]: ---"Transaction committed" 1292ms (23:06:00.828)
Jul 06 23:06:40 kpi4 k3s[19003]: Trace[1240402593]: [1.293659525s] [1.293659525s] END
Jul 06 23:06:40 kpi4 k3s[19003]: I0706 23:06:40.828723 19003 trace.go:205] Trace[1506576956]: "Update" url:/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/cert-manager-cainjector-leader-election,user-agent:cainjector/v0.0.0 (linux/arm64) kubernetes/$Format/leader-election,client:192.168.1.224,accept:application/json, */*,protocol:HTTP/1.1 (06-Jul-2021 23:06:39.533) (total time: 1294ms):
Jul 06 23:06:40 kpi4 k3s[19003]: Trace[1506576956]: ---"Object stored in database" 1293ms (23:06:00.828)
Jul 06 23:06:40 kpi4 k3s[19003]: Trace[1506576956]: [1.294764905s] [1.294764905s] END
Jul 06 23:06:43 kpi4 k3s[19003]: I0706 23:06:43.297530 19003 trace.go:205] Trace[118167507]: "GuaranteedUpdate etcd3" type:*core.Endpoints (06-Jul-2021 23:06:42.405) (total time: 892ms):
Jul 06 23:06:43 kpi4 k3s[19003]: Trace[118167507]: ---"Transaction committed" 891ms (23:06:00.297)
Jul 06 23:06:43 kpi4 k3s[19003]: Trace[118167507]: [892.32055ms] [892.32055ms] END
Jul 06 23:06:43 kpi4 k3s[19003]: I0706 23:06:43.297845 19003 trace.go:205] Trace[1338257210]: "Update" url:/api/v1/namespaces/kube-system/endpoints/rancher.io-local-path,user-agent:local-path-provisioner/v0.0.0 (linux/arm64) kubernetes/$Format,client:10.42.0.4,accept:application/json, */*,protocol:HTTP/1.1 (06-Jul-2021 23:06:42.404) (total time: 893ms):
Jul 06 23:06:43 kpi4 k3s[19003]: Trace[1338257210]: ---"Object stored in database" 892ms (23:06:00.297)
Jul 06 23:06:43 kpi4 k3s[19003]: Trace[1338257210]: [893.115784ms] [893.115784ms] END
Jul 06 23:08:06 kpi4 k3s[19003]: I0706 23:08:06.860661 19003 trace.go:205] Trace[894065588]: "GuaranteedUpdate etcd3" type:*core.Endpoints (06-Jul-2021 23:08:04.949) (total time: 1911ms):
Jul 06 23:08:06 kpi4 k3s[19003]: Trace[894065588]: ---"Transaction committed" 1910ms (23:08:00.860)
Jul 06 23:08:06 kpi4 k3s[19003]: Trace[894065588]: [1.911532738s] [1.911532738s] END
Jul 06 23:08:06 kpi4 k3s[19003]: I0706 23:08:06.860969 19003 trace.go:205] Trace[259400425]: "Update" url:/api/v1/namespaces/kube-system/endpoints/rancher.io-local-path,user-agent:local-path-provisioner/v0.0.0 (linux/arm64) kubernetes/$Format,client:10.42.0.4,accept:application/json, */*,protocol:HTTP/1.1 (06-Jul-2021 23:08:04.948) (total time: 1912ms):
Jul 06 23:08:06 kpi4 k3s[19003]: Trace[259400425]: ---"Object stored in database" 1911ms (23:08:00.860)
Jul 06 23:08:06 kpi4 k3s[19003]: Trace[259400425]: [1.912385195s] [1.912385195s] END
Jul 06 23:08:08 kpi4 k3s[19003]: I0706 23:08:08.846552 19003 trace.go:205] Trace[733674686]: "GuaranteedUpdate etcd3" type:*core.ConfigMap (06-Jul-2021 23:08:07.241) (total time: 1605ms):
Jul 06 23:08:08 kpi4 k3s[19003]: Trace[733674686]: ---"Transaction committed" 1604ms (23:08:00.846)
Jul 06 23:08:08 kpi4 k3s[19003]: Trace[733674686]: [1.605464429s] [1.605464429s] END
Jul 06 23:08:08 kpi4 k3s[19003]: I0706 23:08:08.846850 19003 trace.go:205] Trace[67028535]: "Update" url:/api/v1/namespaces/kube-system/configmaps/cert-manager-cainjector-leader-election,user-agent:cainjector/v0.0.0 (linux/arm64) kubernetes/$Format/leader-election,client:192.168.1.224,accept:application/json, */*,protocol:HTTP/1.1 (06-Jul-2021 23:08:07.240) (total time: 1606ms):
Jul 06 23:08:08 kpi4 k3s[19003]: Trace[67028535]: ---"Object stored in database" 1605ms (23:08:00.846)
Jul 06 23:08:08 kpi4 k3s[19003]: Trace[67028535]: [1.606263109s] [1.606263109s] END
Jul 06 23:08:09 kpi4 k3s[19003]: I0706 23:08:09.553583 19003 trace.go:205] Trace[239102217]: "GuaranteedUpdate etcd3" type:*coordination.Lease (06-Jul-2021 23:08:08.869) (total time: 683ms):
Jul 06 23:08:09 kpi4 k3s[19003]: Trace[239102217]: ---"Transaction committed" 682ms (23:08:00.553)
Jul 06 23:08:09 kpi4 k3s[19003]: Trace[239102217]: [683.879442ms] [683.879442ms] END
Jul 06 23:08:09 kpi4 k3s[19003]: I0706 23:08:09.553890 19003 trace.go:205] Trace[827278646]: "Update" url:/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/cert-manager-cainjector-leader-election,user-agent:cainjector/v0.0.0 (linux/arm64) kubernetes/$Format/leader-election,client:192.168.1.224,accept:application/json, */*,protocol:HTTP/1.1 (06-Jul-2021 23:08:08.869) (total time: 684ms):
Jul 06 23:08:09 kpi4 k3s[19003]: Trace[827278646]: ---"Object stored in database" 684ms (23:08:00.553)
Jul 06 23:08:09 kpi4 k3s[19003]: Trace[827278646]: [684.683973ms] [684.683973ms] END
Jul 06 23:08:10 kpi4 k3s[19003]: I0706 23:08:10.812927 19003 trace.go:205] Trace[80803981]: "GuaranteedUpdate etcd3" type:*coordination.Lease (06-Jul-2021 23:08:09.592) (total time: 1219ms):
Jul 06 23:08:10 kpi4 k3s[19003]: Trace[80803981]: ---"Transaction committed" 1218ms (23:08:00.812)
Jul 06 23:08:10 kpi4 k3s[19003]: Trace[80803981]: [1.219993009s] [1.219993009s] END
Jul 06 23:08:10 kpi4 k3s[19003]: I0706 23:08:10.813351 19003 trace.go:205] Trace[412982872]: "Update" url:/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/kpi4,user-agent:k3s/v1.21.2+k3s1 (linux/arm64) kubernetes/5a67e8d,client:127.0.0.1,accept:application/vnd.kubernetes.protobuf,application/json,protocol:HTTP/1.1 (06-Jul-2021 23:08:09.592) (total time: 1220ms):
Jul 06 23:08:10 kpi4 k3s[19003]: Trace[412982872]: ---"Object stored in database" 1220ms (23:08:00.813)
Jul 06 23:08:10 kpi4 k3s[19003]: Trace[412982872]: [1.220683466s] [1.220683466s] END
Jul 06 23:08:11 kpi4 k3s[19003]: I0706 23:08:11.595104 19003 trace.go:205] Trace[67238373]: "GuaranteedUpdate etcd3" type:*core.Endpoints (06-Jul-2021 23:08:10.929) (total time: 665ms):
Jul 06 23:08:11 kpi4 k3s[19003]: Trace[67238373]: ---"Transaction committed" 663ms (23:08:00.594)
Jul 06 23:08:11 kpi4 k3s[19003]: Trace[67238373]: [665.105041ms] [665.105041ms] END
Jul 06 23:08:11 kpi4 k3s[19003]: I0706 23:08:11.595400 19003 trace.go:205] Trace[1995907290]: "Update" url:/api/v1/namespaces/kube-system/endpoints/rancher.io-local-path,user-agent:local-path-provisioner/v0.0.0 (linux/arm64) kubernetes/$Format,client:10.42.0.4,accept:application/json, */*,protocol:HTTP/1.1 (06-Jul-2021 23:08:10.929) (total time: 666ms):
Jul 06 23:08:11 kpi4 k3s[19003]: Trace[1995907290]: ---"Object stored in database" 665ms (23:08:00.595)
Jul 06 23:08:11 kpi4 k3s[19003]: Trace[1995907290]: [666.037441ms] [666.037441ms] END
Jul 06 23:08:53 kpi4 k3s[19003]: W0706 23:08:53.797420 19003 sysinfo.go:203] Nodes topology is not available, providing CPU topology
Jul 06 23:13:53 kpi4 k3s[19003]: W0706 23:13:53.797631 19003 sysinfo.go:203] Nodes topology is not available, providing CPU topology
Jul 06 23:18:53 kpi4 k3s[19003]: W0706 23:18:53.795917 19003 sysinfo.go:203] Nodes topology is not available, providing CPU topology
Jul 06 23:23:42 kpi4 k3s[19003]: time="2021-07-06T23:23:42.498422650+02:00" level=error msg="Compact failed: failed to compact to revision 4327: database is locked"
Jul 06 23:23:53 kpi4 k3s[19003]: W0706 23:23:53.797186 19003 sysinfo.go:203] Nodes topology is not available, providing CPU topology
Jul 06 23:28:53 kpi4 k3s[19003]: W0706 23:28:53.799131 19003 sysinfo.go:203] Nodes topology is not available, providing CPU topology
pi@kpi4:~ $ journalctl -xb0 -f
-- Logs begin at Mon 2021-07-05 21:25:05 CEST. --
Jul 06 23:45:13 kpi4 systemd[1]: run-containerd-runc-k8s.io-27a5e7c74fab26bc92d6cf65d06247ea9f03f177d72197406749c7612f5fd3a0-runc.NFFaIl.mount: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://www.debian.org/support
--
-- The unit run-containerd-runc-k8s.io-27a5e7c74fab26bc92d6cf65d06247ea9f03f177d72197406749c7612f5fd3a0-runc.NFFaIl.mount has successfully entered the 'dead' state.
Jul 06 23:45:18 kpi4 systemd[666]: run-containerd-runc-k8s.io-27a5e7c74fab26bc92d6cf65d06247ea9f03f177d72197406749c7612f5fd3a0-runc.CPgcgF.mount: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://www.debian.org/support
--
-- The unit UNIT has successfully entered the 'dead' state.
Jul 06 23:45:18 kpi4 systemd[1]: run-containerd-runc-k8s.io-27a5e7c74fab26bc92d6cf65d06247ea9f03f177d72197406749c7612f5fd3a0-runc.CPgcgF.mount: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://www.debian.org/support
I checked that I can create empty volumes and write in them but that unit keeps on failing.. Is there anything I am missing that I can check @brandond ? I have created a gist with the full k3s journal
https://gist.github.com/alborotogarcia/546cf2fbc14932dd250bea1de887fcab
It fails only on raspberry 4b not on other arm64 devices I have.. But this error trace makes the raspberry node unresposive after a while..
I attach some logs from local-path
storage csi-attacher-5dbfd9bf46-hsf47 1/1 Running 3 58m 10.42.0.15 kpi4 <none> <none>
kube-system local-path-provisioner-5ff76fc89d-m2t6q 1/1 Running 7 64m 10.42.0.4 kpi4 <none> <none>
cert-manager cert-manager-cainjector-69d885bf55-65bhk 1/1 Running 9 63m 10.42.0.8 kpi4 <none> <none>
k logs -n kube-system pod/local-path-provisioner-5ff76fc89d-m2t6q
I0707 02:44:33.468680 1 leaderelection.go:242] attempting to acquire leader lease kube-system/rancher.io-local-path...
I0707 02:44:51.171328 1 leaderelection.go:252] successfully acquired lease kube-system/rancher.io-local-path
I0707 02:44:51.172856 1 controller.go:773] Starting provisioner controller rancher.io/local-path_local-path-provisioner-5ff76fc89d-m2t6q_42229385-ddd5-4822-8054-5ea804c390ad!
I0707 02:44:51.174361 1 event.go:281] Event(v1.ObjectReference{Kind:"Endpoints", Namespace:"kube-system", Name:"rancher.io-local-path", UID:"eac70bb8-86cc-43bb-aab6-da1d1ee38b9e", APIVersion:"v1", ResourceVersion:"10857", FieldPath:""}): type: 'Normal' reason: 'LeaderElection' local-path-provisioner-5ff76fc89d-m2t6q_42229385-ddd5-4822-8054-5ea804c390ad became leader
I0707 02:44:51.273181 1 controller.go:822] Started provisioner controller rancher.io/local-path_local-path-provisioner-5ff76fc89d-m2t6q_42229385-ddd5-4822-8054-5ea804c390ad!
I0707 02:56:44.138114 1 leaderelection.go:288] failed to renew lease kube-system/rancher.io-local-path: failed to tryAcquireOrRenew context deadline exceeded
F0707 02:56:44.138184 1 controller.go:851] leaderelection lost
Are you running Longhorn on a Pi 4? Are you using MMC/SD, or external storage? I suspect that you don't have sufficient disk IO throughput to support the control-plane, longhorn, and workloads.
Apparently the raspberry pi4 has overcome unresponsiveness (it used to fail within 24h range) by disabling avahi, but the error still persists @brandond
kube-system sealed-secrets-controller-6ff6764757-9kvxt 1/1 Running 3 38h 10.42.8.4 knode2 <none> <none>
kube-system traefik-97b44b794-r8pj8 1/1 Running 0 38h 10.42.8.3 knode2 <none> <none>
ml yatai-service-6d598d979c-wxlsd 1/1 Running 10 20h 10.42.8.31 knode2 <none> <none>
argo-rollouts argo-rollouts-58dc89bcfb-zfn9c 1/1 Running 0 14h 10.42.4.30 knode1 <none> <none>
storage csi-provisioner-578dc5df8d-wbvn6 1/1 Running 81 38h 10.42.4.12 knode1 <none> <none>
storage csi-resizer-85fbd8459d-k2m6t 1/1 Running 119 38h 10.42.4.14 knode1 <none> <none>
storage csi-attacher-5dbfd9bf46-zdcwn 1/1 Running 78 38h 10.42.4.10 knode1 <none> <none>
storage csi-snapshotter-94bbf64bd-kbpr6 1/1 Running 105 38h 10.42.8.15 knode2 <none> <none>
storage csi-resizer-85fbd8459d-2kqh7 1/1 Running 84 38h 10.42.8.13 knode2 <none> <none>
storage csi-provisioner-578dc5df8d-hxnj9 1/1 Running 89 38h 10.42.4.11 knode1 <none> <none>
storage csi-provisioner-578dc5df8d-9mmd2 1/1 Running 99 38h 10.42.8.14 knode2 <none> <none>
storage csi-snapshotter-94bbf64bd-kcx59 1/1 Running 86 38h 10.42.0.20 kpi4 <none> <none>
storage csi-attacher-5dbfd9bf46-k4sh5 1/1 Running 111 38h 10.42.8.12 knode2 <none> <none>
storage csi-snapshotter-94bbf64bd-wtzn6 1/1 Running 90 38h 10.42.4.13 knode1 <none> <none>
storage csi-attacher-5dbfd9bf46-96g6m 1/1 Running 80 38h 10.42.8.11 knode2 <none> <none>
storage csi-resizer-85fbd8459d-q54dm 1/1 Running 75 38h 10.42.0.19 kpi4 <none> <none>
argo workflow-controller-cb97dc8d5-fxfjv 1/1 Running 233 38h 10.42.4.17 knode1 <none> <none>
kube-system local-path-provisioner-5ff76fc89d-ktq9t 1/1 Running 337 38h 10.42.0.4 kpi4 <none> <none>
cert-manager cert-manager-cainjector-69d885bf55-k48dq 0/1 CrashLoopBackOff 375 38h 10.42.8.2 knode2 <none> <none>
Wrt storage I am pursuing to use longhorn on my k3s cluster you're right but the raspi 4b uses a ssd as main disk on a usb3 port..
root@kpi4:/home/pi# lshw -class disk -class storage
*-usb
description: Mass storage device
product: AS2115
vendor: ASMedia
physical id: 1
bus info: usb@2:1
logical name: scsi0
version: 0.01
serial: 00000000000000000000
capabilities: usb-3.00 scsi emulated
configuration: driver=usb-storage speed=5000Mbit/s
*-disk
description: SCSI Disk
product: 2115
vendor: ASMT
physical id: 0.0.0
bus info: scsi@0:0.0.0
logical name: /dev/sda
version: 0
serial: 00000000000000000000
size: 223GiB (240GB)
capabilities: partitioned partitioned:dos
configuration: ansiversion=6 logicalsectorsize=512 sectorsize=512 signature=6393f405
pi@kpi4:~ $ cat /etc/fstab
proc /proc proc defaults 0 0
PARTUUID=6393f405-01 /boot vfat defaults 0 2
PARTUUID=6393f405-02 / ext4 defaults,noatime 0 1
Disk /dev/sda: 223.6 GiB, 240057409536 bytes, 468862128 sectors
Disk model: 2115
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes
Disklabel type: dos
Disk identifier: 0x6393f405
Device Boot Start End Sectors Size Id Type
/dev/sda1 8192 532479 524288 256M c W95 FAT32 (LBA)
/dev/sda2 532480 468862127 468329648 223.3G 83 Linux
root@kpi4:/home/pi# hdparm -I /dev/sda
/dev/sda:
ATA device, with non-removable media
Model Number: KingDian P10 240GB
Serial Number: 2018071900719
Firmware Revision: R0522A0
Media Serial Num:
Media Manufacturer:
Transport: Serial, ATA8-AST, SATA 1.0a, SATA II Extensions, SATA Rev 2.5, SATA Rev 2.6, SATA Rev 3.0
Standards:
Used: unknown (minor revision code 0x0110)
Supported: 9 8 7 6 5
Likely used: 9
Configuration:
Logical max current
cylinders 16383 16383
heads 16 16
sectors/track 63 63
--
CHS current addressable sectors: 16514064
LBA user addressable sectors: 268435455
LBA48 user addressable sectors: 468862128
Logical Sector size: 512 bytes
Physical Sector size: 512 bytes
Logical Sector-0 offset: 0 bytes
device size with M = 1024*1024: 228936 MBytes
device size with M = 1000*1000: 240057 MBytes (240 GB)
cache/buffer size = unknown
Form Factor: 2.5 inch
Nominal Media Rotation Rate: Solid State Device
Capabilities:
LBA, IORDY(can be disabled)
Queue depth: 32
Standby timer values: spec'd by Standard, no device specific minimum
R/W multiple sector transfer: Max = 1 Current = 1
Advanced power management level: 128
DMA: mdma0 mdma1 mdma2 udma0 udma1 udma2 udma3 udma4 udma5 *udma6
Cycle time: min=120ns recommended=120ns
PIO: pio0 pio1 pio2 pio3 pio4
Cycle time: no flow control=120ns IORDY flow control=120ns
Commands/features:
Enabled Supported:
* SMART feature set
Security Mode feature set
* Power Management feature set
* Write cache
* Look-ahead
* Host Protected Area feature set
* WRITE_BUFFER command
* READ_BUFFER command
* DOWNLOAD_MICROCODE
* Advanced Power Management feature set
SET_MAX security extension
* 48-bit Address feature set
* Mandatory FLUSH_CACHE
* FLUSH_CACHE_EXT
* SMART error logging
* SMART self-test
* General Purpose Logging feature set
* WRITE_{DMA|MULTIPLE}_FUA_EXT
* WRITE_UNCORRECTABLE_EXT command
* {READ,WRITE}_DMA_EXT_GPL commands
* Segmented DOWNLOAD_MICROCODE
* Gen1 signaling speed (1.5Gb/s)
* Gen2 signaling speed (3.0Gb/s)
* Gen3 signaling speed (6.0Gb/s)
* Native Command Queueing (NCQ)
* Phy event counters
* READ_LOG_DMA_EXT equivalent to READ_LOG_EXT
DMA Setup Auto-Activate optimization
* Software settings preservation
* SANITIZE feature set
* BLOCK_ERASE_EXT command
* DOWNLOAD MICROCODE DMA command
* WRITE BUFFER DMA command
* READ BUFFER DMA command
* Data Set Management TRIM supported (limit 8 blocks)
* Deterministic read ZEROs after TRIM
Security:
Master password revision code = 65534
supported
not enabled
not locked
not frozen
not expired: security count
supported: enhanced erase
2min for SECURITY ERASE UNIT. 2min for ENHANCED SECURITY ERASE UNIT.
Do you think it may work out swapping it as an agent node?
I've never seen that message before, and it appears to be coming from systemd... so I'm not sure what the deal is. I have a couple Raspberry Pi 4s that I run as k3s servers as well (on Ubuntu, with external SSD), but I haven't personally tried Longhorn on them. It says it's successfully entering the dead state, which I'm not sure is a failure - is it possible that the log level just got turned up somewhere in your systemd config?
Same error appeared to me on prebuilt kernels, at least bootstrapping a k3s cluster, when I add longhorn it becomes more frequently probably due to IO throughtput as you say @brandond I enabled debug mode rebuilding the kernel.. as far as I remember CONFIG_RT_GROUP_SCHED, CONFIG_CGROUP_HUGETLB were missing on the built-in kernel.. that's why I rebuilt it.. Could you check if longhorn does good for you ? And if so tell me the ssd you use please :) I've spent last week on this and I am running out of ideas.. I was thinking on using an SD card and the ssd for longhorn data-dir.. but micro SD cards are slow as hell..
I've just realized that one of my nodes (knode2) has reached inotify watchers limits by opening sublime text.. @brandond
In addition, there are some diagnostic messages that I didn't check
jetson32@knode2:~$ journalctl -k |tail -n35
Jul 08 22:07:32 knode2 kernel: IPv6: eth0: IPv6 duplicate address fe80::d06b:afff:fec6:db30 detected!
Jul 08 22:07:40 knode2 kernel: cni0: port 23(veth3ba1b017) entered disabled state
Jul 08 22:07:40 knode2 kernel: device veth3ba1b017 left promiscuous mode
Jul 08 22:07:40 knode2 kernel: cni0: port 23(veth3ba1b017) entered disabled state
Jul 08 22:07:49 knode2 kernel: scsi host2: iSCSI Initiator over TCP/IP
Jul 08 22:07:49 knode2 kernel: scsi 2:0:0:0: RAID IET Controller 0001 PQ: 0 ANSI: 5
Jul 08 22:07:49 knode2 kernel: scsi 2:0:0:1: Direct-Access IET VIRTUAL-DISK 0001 PQ: 0 ANSI: 5
Jul 08 22:07:49 knode2 kernel: sd 2:0:0:1: [sda] 4194304 512-byte logical blocks: (2.15 GB/2.00 GiB)
Jul 08 22:07:49 knode2 kernel: sd 2:0:0:1: [sda] Write Protect is off
Jul 08 22:07:49 knode2 kernel: sd 2:0:0:1: [sda] Mode Sense: 69 00 10 08
Jul 08 22:07:49 knode2 kernel: sd 2:0:0:1: [sda] Write cache: enabled, read cache: enabled, supports DPO and FUA
Jul 08 22:07:57 knode2 kernel: sd 2:0:0:1: [sda] Attached SCSI disk
Jul 08 22:09:28 knode2 kernel: sd 2:0:0:1: [sda] Synchronizing SCSI cache
Jul 08 22:10:19 knode2 kernel: scsi host2: iSCSI Initiator over TCP/IP
Jul 08 22:10:19 knode2 kernel: scsi 2:0:0:0: RAID IET Controller 0001 PQ: 0 ANSI: 5
Jul 08 22:10:19 knode2 kernel: scsi 2:0:0:1: Direct-Access IET VIRTUAL-DISK 0001 PQ: 0 ANSI: 5
Jul 08 22:10:19 knode2 kernel: sd 2:0:0:1: [sda] 20971520 512-byte logical blocks: (10.7 GB/10.0 GiB)
Jul 08 22:10:19 knode2 kernel: sd 2:0:0:1: [sda] Write Protect is off
Jul 08 22:10:19 knode2 kernel: sd 2:0:0:1: [sda] Mode Sense: 69 00 10 08
Jul 08 22:10:19 knode2 kernel: sd 2:0:0:1: [sda] Write cache: enabled, read cache: enabled, supports DPO and FUA
Jul 08 22:10:19 knode2 kernel: sd 2:0:0:1: [sda] Attached SCSI disk
Jul 08 22:12:19 knode2 kernel: EXT4-fs (sda): mounted filesystem with ordered data mode. Opts: (null)
Jul 08 22:12:20 knode2 kernel: IPVS: Creating netns size=1928 id=26
Jul 08 22:12:20 knode2 kernel: IPv6: ADDRCONF(NETDEV_UP): eth0: link is not ready
Jul 08 22:12:20 knode2 kernel: IPv6: ADDRCONF(NETDEV_UP): vethe2d3af3a: link is not ready
Jul 08 22:12:20 knode2 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): vethe2d3af3a: link becomes ready
Jul 08 22:12:20 knode2 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): eth0: link becomes ready
Jul 08 22:12:20 knode2 kernel: cni0: port 23(vethe2d3af3a) entered blocking state
Jul 08 22:12:20 knode2 kernel: cni0: port 23(vethe2d3af3a) entered disabled state
Jul 08 22:12:20 knode2 kernel: device vethe2d3af3a entered promiscuous mode
Jul 08 22:12:20 knode2 kernel: cni0: port 23(vethe2d3af3a) entered blocking state
Jul 08 22:12:20 knode2 kernel: cni0: port 23(vethe2d3af3a) entered forwarding state
Jul 08 22:12:20 knode2 kernel: IPv6: eth0: IPv6 duplicate address fe80::d845:e7ff:fe7d:5790 detected!
Jul 08 22:15:08 knode2 kernel: t19x-arm-smmu 12000000.iommu: Unhandled context fault: smmu0, iova=0x4fc0ff000, fsynr=0x20003, cb=0, sid=86(0x56 - PCIE0), pgd=855adf003, pud=855adf003, pmd=7e4268003, pte=0
Jul 08 22:20:14 knode2 kernel: t19x-arm-smmu 12000000.iommu: Unhandled context fault: smmu1, iova=0x4fc86f000, fsynr=0x380003, cb=0, sid=86(0x56 - PCIE0), pgd=855adf003, pud=855adf003, pmd=839d8e003, pte=0
pi@kpi4:~ $ journalctl -k |tail -n 30
Jul 08 22:04:38 kpi4 kernel: cni0: port 17(veth18d8e7dc) entered disabled state
Jul 08 22:04:38 kpi4 kernel: device veth18d8e7dc entered promiscuous mode
Jul 08 22:04:38 kpi4 kernel: cni0: port 17(veth18d8e7dc) entered blocking state
Jul 08 22:04:38 kpi4 kernel: cni0: port 17(veth18d8e7dc) entered forwarding state
Jul 08 22:04:44 kpi4 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): vethc50adaea: link becomes ready
Jul 08 22:04:44 kpi4 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): eth0: link becomes ready
Jul 08 22:04:44 kpi4 kernel: cni0: port 18(vethc50adaea) entered blocking state
Jul 08 22:04:44 kpi4 kernel: cni0: port 18(vethc50adaea) entered disabled state
Jul 08 22:04:44 kpi4 kernel: device vethc50adaea entered promiscuous mode
Jul 08 22:04:44 kpi4 kernel: cni0: port 18(vethc50adaea) entered blocking state
Jul 08 22:04:44 kpi4 kernel: cni0: port 18(vethc50adaea) entered forwarding state
Jul 08 22:04:45 kpi4 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): eth0: link becomes ready
Jul 08 22:04:45 kpi4 kernel: cni0: port 19(vethcc1f3d40) entered blocking state
Jul 08 22:04:45 kpi4 kernel: cni0: port 19(vethcc1f3d40) entered disabled state
Jul 08 22:04:45 kpi4 kernel: device vethcc1f3d40 entered promiscuous mode
Jul 08 22:04:45 kpi4 kernel: cni0: port 19(vethcc1f3d40) entered blocking state
Jul 08 22:04:45 kpi4 kernel: cni0: port 19(vethcc1f3d40) entered forwarding state
Jul 08 22:05:32 kpi4 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): veth6891ad26: link becomes ready
Jul 08 22:05:32 kpi4 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): eth0: link becomes ready
Jul 08 22:05:32 kpi4 kernel: cni0: port 20(veth6891ad26) entered blocking state
Jul 08 22:05:32 kpi4 kernel: cni0: port 20(veth6891ad26) entered disabled state
Jul 08 22:05:32 kpi4 kernel: device veth6891ad26 entered promiscuous mode
Jul 08 22:05:32 kpi4 kernel: cni0: port 20(veth6891ad26) entered blocking state
Jul 08 22:05:32 kpi4 kernel: cni0: port 20(veth6891ad26) entered forwarding state
Jul 08 22:05:32 kpi4 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): eth0: link becomes ready
Jul 08 22:05:32 kpi4 kernel: cni0: port 21(veth969f545b) entered blocking state
Jul 08 22:05:32 kpi4 kernel: cni0: port 21(veth969f545b) entered disabled state
Jul 08 22:05:32 kpi4 kernel: device veth969f545b entered promiscuous mode
Jul 08 22:05:32 kpi4 kernel: cni0: port 21(veth969f545b) entered blocking state
Jul 08 22:05:32 kpi4 kernel: cni0: port 21(veth969f545b) entered forwarding state
jetson16@knode1:~$ journalctl -k |tail -n 30
Jul 08 22:10:07 knode1 kernel: device veth47620c56 entered promiscuous mode
Jul 08 22:10:07 knode1 kernel: cni0: port 29(veth47620c56) entered blocking state
Jul 08 22:10:07 knode1 kernel: cni0: port 29(veth47620c56) entered forwarding state
Jul 08 22:10:07 knode1 kernel: IPv6: eth0: IPv6 duplicate address fe80::c2d:a4ff:fe62:f965 detected!
Jul 08 22:10:09 knode1 kernel: scsi host2: iSCSI Initiator over TCP/IP
Jul 08 22:10:09 knode1 kernel: scsi 2:0:0:0: RAID IET Controller 0001 PQ: 0 ANSI: 5
Jul 08 22:10:09 knode1 kernel: scsi 2:0:0:1: Direct-Access IET VIRTUAL-DISK 0001 PQ: 0 ANSI: 5
Jul 08 22:10:09 knode1 kernel: sd 2:0:0:1: [sda] 4194304 512-byte logical blocks: (2.15 GB/2.00 GiB)
Jul 08 22:10:09 knode1 kernel: sd 2:0:0:1: [sda] Write Protect is off
Jul 08 22:10:09 knode1 kernel: sd 2:0:0:1: [sda] Mode Sense: 69 00 10 08
Jul 08 22:10:09 knode1 kernel: sd 2:0:0:1: [sda] Write cache: enabled, read cache: enabled, supports DPO and FUA
Jul 08 22:10:09 knode1 kernel: sd 2:0:0:1: [sda] Attached SCSI disk
Jul 08 22:12:06 knode1 kernel: EXT4-fs (sda): mounted filesystem with ordered data mode. Opts: (null)
Jul 08 22:12:06 knode1 kernel: IPVS: Creating netns size=1928 id=35
Jul 08 22:12:06 knode1 kernel: IPv6: ADDRCONF(NETDEV_UP): eth0: link is not ready
Jul 08 22:12:06 knode1 kernel: IPv6: ADDRCONF(NETDEV_UP): vethe8d73c70: link is not ready
Jul 08 22:12:06 knode1 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): vethe8d73c70: link becomes ready
Jul 08 22:12:06 knode1 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): eth0: link becomes ready
Jul 08 22:12:06 knode1 kernel: cni0: port 30(vethe8d73c70) entered blocking state
Jul 08 22:12:06 knode1 kernel: cni0: port 30(vethe8d73c70) entered disabled state
Jul 08 22:12:06 knode1 kernel: device vethe8d73c70 entered promiscuous mode
Jul 08 22:12:06 knode1 kernel: cni0: port 30(vethe8d73c70) entered blocking state
Jul 08 22:12:06 knode1 kernel: cni0: port 30(vethe8d73c70) entered forwarding state
Jul 08 22:12:06 knode1 kernel: IPv6: eth0: IPv6 duplicate address fe80::cc79:a7ff:fe05:29d9 detected!
Jul 08 22:12:42 knode1 kernel: cni0: port 22(veth1cefc6c4) entered disabled state
Jul 08 22:12:42 knode1 kernel: device veth1cefc6c4 left promiscuous mode
Jul 08 22:12:42 knode1 kernel: cni0: port 22(veth1cefc6c4) entered disabled state
Jul 08 22:13:35 knode1 kernel: cni0: port 23(vethd53af24c) entered disabled state
Jul 08 22:13:35 knode1 kernel: device vethd53af24c left promiscuous mode
Jul 08 22:13:35 knode1 kernel: cni0: port 23(vethd53af24c) entered disabled state
@alborotogarcia Were you able to solve the problem?
@heiderich in the end I think it got solved by making a clean debian install (try debian 11 bullseye better) and set pi4 as a worker node.. you shouldn't have many problems, nowadays it gets even easier to set an external ssd as main filesystem on a raspberry pi 4, though I still notice that under high workloads it's the node that would get the most unresponsive, so mind balancing and get metrics..
Environmental Info: K3s Version: v1.21.2+k3s1
Node(s) CPU architecture, OS, and Version:
Master node (Debian 10):
Worker nodes (Ubuntu Bionic 18.04):
Verifying binaries in /var/lib/rancher/k3s/data/67f8817e7c5b5a7e397a9c75dafbfe2db3696b18c79ba3c51f746e1f960c8f14/bin:
System:
Limits:
info: reading kernel config from /proc/config.gz ...
Generally Necessary:
Optional Features:
STATUS: pass
cgroup v2
pi@kpi4:~ $ ls /sys/fs/cgroup/memory cgroup.clone_children kubepods memory.kmem.max_usage_in_bytes memory.kmem.tcp.usage_in_bytes memory.memsw.limit_in_bytes memory.pressure_level memory.use_hierarchy user.slice cgroup.event_control memory.failcnt memory.kmem.slabinfo memory.kmem.usage_in_bytes memory.memsw.max_usage_in_bytes memory.soft_limit_in_bytes notify_on_release cgroup.procs memory.force_empty memory.kmem.tcp.failcnt memory.limit_in_bytes memory.memsw.usage_in_bytes memory.stat release_agent cgroup.sane_behavior memory.kmem.failcnt memory.kmem.tcp.limit_in_bytes memory.max_usage_in_bytes memory.move_charge_at_immigrate memory.swappiness system.slice init.scope memory.kmem.limit_in_bytes memory.kmem.tcp.max_usage_in_bytes memory.memsw.failcnt memory.oom_control memory.usage_in_bytes tasks
/etc/systemd/system/k3s.service
[Unit] Description=Lightweight Kubernetes Documentation=https://k3s.io After=network-online.target
[Service] Type=notify ExecStartPre=-/sbin/modprobe br_netfilter ExecStartPre=-/sbin/modprobe overlay ExecStart=/usr/local/bin/k3s server --disable servicelb --tls-san 192.168.1.226,example.com,192.168.1.227,cluster.local --kubelet-arg=max-pods=1022 --kube-controller-manager-arg=node-cidr-mask-size=22 --flannel-backend=host-gw KillMode=process Delegate=yes
Having non-zero Limit*s causes performance problems due to accounting overhead
in the kernel. We recommend using cgroups to do container-local accounting.
LimitNOFILE=1048576 LimitNPROC=infinity LimitCORE=infinity TasksMax=infinity TimeoutStartSec=0 Restart=always RestartSec=5s
[Install] WantedBy=multi-user.target
/etc/systemd/system/k3s-node.service
[Unit] Description=Lightweight Kubernetes Documentation=https://k3s.io After=network-online.target
[Service] Type=notify ExecStartPre=-/sbin/modprobe br_netfilter ExecStartPre=-/sbin/modprobe overlay ExecStart=/usr/local/bin/k3s agent --server https://kpi4:6443 --token K10d77ab80ea82087dfcdbe21d8a6b781c9a5c176db43e89d31596dc78763168a91::server:6892c7aa82636ad69e592a03f0d1d4c5 KillMode=process Delegate=yes
Having non-zero Limit*s causes performance problems due to accounting overhead
in the kernel. We recommend using cgroups to do container-local accounting.
LimitNOFILE=1048576 LimitNPROC=infinity LimitCORE=infinity TasksMax=infinity TimeoutStartSec=0 Restart=always RestartSec=5s
[Install] WantedBy=multi-user.target
LAST SEEN TYPE REASON OBJECT MESSAGE 29m Normal Starting node/kpi4 Starting kubelet. 29m Warning InvalidDiskCapacity node/kpi4 invalid capacity 0 on image filesystem 29m Normal NodeHasSufficientMemory node/kpi4 Node kpi4 status is now: NodeHasSufficientMemory 29m Normal NodeHasNoDiskPressure node/kpi4 Node kpi4 status is now: NodeHasNoDiskPressure 29m Normal NodeHasSufficientPID node/kpi4 Node kpi4 status is now: NodeHasSufficientPID 29m Normal NodeAllocatableEnforced node/kpi4 Updated Node Allocatable limit across pods 29m Normal Starting node/kpi4 Starting kube-proxy. 28m Normal NodeReady node/kpi4 Node kpi4 status is now: NodeReady 28m Normal Synced node/kpi4 Node synced successfully 28m Normal RegisteredNode node/kpi4 Node kpi4 event: Registered Node kpi4 in Controller 28m Normal Starting node/knode1 Starting kubelet. 28m Warning InvalidDiskCapacity node/knode1 invalid capacity 0 on image filesystem 28m Normal Starting node/knode2 Starting kubelet. 28m Warning InvalidDiskCapacity node/knode2 invalid capacity 0 on image filesystem 28m Normal NodeHasSufficientMemory node/knode1 Node knode1 status is now: NodeHasSufficientMemory 28m Normal NodeHasNoDiskPressure node/knode1 Node knode1 status is now: NodeHasNoDiskPressure 28m Normal NodeHasSufficientPID node/knode1 Node knode1 status is now: NodeHasSufficientPID 28m Normal NodeHasSufficientMemory node/knode2 Node knode2 status is now: NodeHasSufficientMemory 28m Normal NodeHasNoDiskPressure node/knode2 Node knode2 status is now: NodeHasNoDiskPressure 28m Normal Synced node/knode1 Node synced successfully 28m Normal NodeHasSufficientPID node/knode2 Node knode2 status is now: NodeHasSufficientPID 28m Normal NodeAllocatableEnforced node/knode1 Updated Node Allocatable limit across pods 28m Normal Starting node/knode1 Starting kube-proxy. 28m Normal Synced node/knode2 Node synced successfully 28m Normal NodeAllocatableEnforced node/knode2 Updated Node Allocatable limit across pods 28m Normal Starting node/knode2 Starting kube-proxy. 28m Normal RegisteredNode node/knode1 Node knode1 event: Registered Node knode1 in Controller 28m Normal RegisteredNode node/knode2 Node knode2 event: Registered Node knode2 in Controller 28m Normal NodeReady node/knode1 Node knode1 status is now: NodeReady 28m Normal NodeReady node/knode2 Node knode2 status is now: NodeReady 19m Warning FailedToCreateEndpoint endpoints/longhorn-backend Failed to create endpoint for service longhorn-system/longhorn-backend: endpoints "longhorn-backend" already exists
● run-k3s-containerd-io.containerd.runtime.v2.task-k8s.io-6ebbd28459368680767dd0a519affa8f9dc21aa1b73cc88c6f3d80140f950771-rootfs.mount - /run/k3s/containerd/io.containerd.runtime.v2.task/k8s.io/6ebbd28459368680767dd0a519affa8f9dc21aa1b73cc88c6f3d80140f950771/rootfs Loaded: loaded (/proc/self/mountinfo) Active: active (mounted) since Mon 2021-07-05 20:40:34 CEST; 4min 44s ago Where: /run/k3s/containerd/io.containerd.runtime.v2.task/k8s.io/6ebbd28459368680767dd0a519affa8f9dc21aa1b73cc88c6f3d80140f950771/rootfs What: overlay
Where=/run/k3s/containerd/io.containerd.runtime.v2.task/k8s.io/6ebbd28459368680767dd0a519affa8f9dc21aa1b73cc88c6f3d80140f950771/rootfs What=overlay Options=rw,relatime,lowerdir=/var/lib/rancher/k3s/agent/containerd/io.containerd.snapshotter.v1.overlayfs/snapshots/64/fs:/var/lib/rancher/k3s/agent/containerd/io.containerd.snapshotter.v1.overlayfs/snapshots/63/fs:/var/lib/rancher/k3s/agent/containerd/io.containerd.snapshotter.v1.overlayfs/snapshots/62/fs:/var/lib/rancher/k3s/agent/containerd/io.containerd.snapshotter.v1.overlayfs/snapshots/61/fs:/var/lib/rancher/k3s/agent/containerd/io.containerd.snapshotter.v1.overlayfs/snapshots/60/fs:/var/lib/rancher/k3s/agent/containerd/io.containerd.snapshotter.v1.overlayfs/snapshots/59/fs:/var/lib/rancher/k3s/agent/containerd/io.containerd.snapshotter.v1.overlayfs/snapshots/58/fs:/var/lib/rancher/k3s/agent/containerd/io.containerd.snapshotter.v1.overlayfs/snapshots/52/fs:/var/lib/rancher/k3s/agent/containerd/io.containerd.snapshotter.v1.overlayfs/snapshots/37/fs,upperdir=/var/lib/rancher/k3s/agent/containerd/io.containerd.snapshotter.v1.overlayfs/snapshots/65/fs,workdir=/var/lib/rancher/k3s/agent/containerd/io.containerd.snapshotter.v1.overlayfs/snapshots/65/work,index=off Type=overlay TimeoutUSec=1min 30s ControlPID=0 DirectoryMode=0755 SloppyOptions=no LazyUnmount=no ForceUnmount=no Result=success UID=[not set] GID=[not set] Slice=system.slice MemoryCurrent=[not set] CPUUsageNSec=[not set] TasksCurrent=[not set] IPIngressBytes=18446744073709551615 IPIngressPackets=18446744073709551615 IPEgressBytes=18446744073709551615 IPEgressPackets=18446744073709551615 Delegate=no CPUAccounting=no CPUWeight=[not set] StartupCPUWeight=[not set] CPUShares=[not set] StartupCPUShares=[not set] CPUQuotaPerSecUSec=infinity IOAccounting=no IOWeight=[not set] StartupIOWeight=[not set] BlockIOAccounting=no BlockIOWeight=[not set] StartupBlockIOWeight=[not set] MemoryAccounting=yes MemoryMin=0 MemoryLow=0 MemoryHigh=infinity MemoryMax=infinity MemorySwapMax=infinity MemoryLimit=infinity DevicePolicy=auto TasksAccounting=yes TasksMax=4915 IPAccounting=no UMask=0022 LimitCPU=infinity LimitCPUSoft=infinity LimitFSIZE=infinity LimitFSIZESoft=infinity LimitDATA=infinity LimitDATASoft=infinity LimitSTACK=infinity LimitSTACKSoft=8388608 LimitCORE=infinity LimitCORESoft=0 LimitRSS=infinity LimitRSSSoft=infinity LimitNOFILE=524288 LimitNOFILESoft=1024 LimitAS=infinity LimitASSoft=infinity LimitNPROC=30204 LimitNPROCSoft=30204 LimitMEMLOCK=65536 LimitMEMLOCKSoft=65536 LimitLOCKS=infinity LimitLOCKSSoft=infinity LimitSIGPENDING=30204 LimitSIGPENDINGSoft=30204 LimitMSGQUEUE=819200 LimitMSGQUEUESoft=819200 LimitNICE=0 LimitNICESoft=0 LimitRTPRIO=0 LimitRTPRIOSoft=0 LimitRTTIME=infinity LimitRTTIMESoft=infinity OOMScoreAdjust=0 Nice=0 IOSchedulingClass=0 IOSchedulingPriority=0 CPUSchedulingPolicy=0 CPUSchedulingPriority=0 TimerSlackNSec=50000 CPUSchedulingResetOnFork=no NonBlocking=no StandardInput=null StandardInputData= StandardOutput=journal StandardError=inherit TTYReset=no TTYVHangup=no TTYVTDisallocate=no SyslogPriority=30 SyslogLevelPrefix=yes SyslogLevel=6 SyslogFacility=3 LogLevelMax=-1 LogRateLimitIntervalUSec=0 LogRateLimitBurst=0 SecureBits=0 CapabilityBoundingSet=cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf AmbientCapabilities= DynamicUser=no RemoveIPC=no MountFlags= PrivateTmp=no PrivateDevices=no ProtectKernelTunables=no ProtectKernelModules=no ProtectControlGroups=no PrivateNetwork=no PrivateUsers=no PrivateMounts=no ProtectHome=no ProtectSystem=no SameProcessGroup=yes UtmpMode=init IgnoreSIGPIPE=yes NoNewPrivileges=no SystemCallErrorNumber=0 LockPersonality=no RuntimeDirectoryPreserve=no RuntimeDirectoryMode=0755 StateDirectoryMode=0755 CacheDirectoryMode=0755 LogsDirectoryMode=0755 ConfigurationDirectoryMode=0755 MemoryDenyWriteExecute=no RestrictRealtime=no RestrictNamespaces=no MountAPIVFS=no KeyringMode=shared KillMode=control-group KillSignal=15 FinalKillSignal=9 SendSIGKILL=yes SendSIGHUP=no WatchdogSignal=6 Id=run-k3s-containerd-io.containerd.runtime.v2.task-k8s.io-6ebbd28459368680767dd0a519affa8f9dc21aa1b73cc88c6f3d80140f950771-rootfs.mount Names=run-k3s-containerd-io.containerd.runtime.v2.task-k8s.io-6ebbd28459368680767dd0a519affa8f9dc21aa1b73cc88c6f3d80140f950771-rootfs.mount Requires=system.slice -.mount Conflicts=umount.target Before=umount.target local-fs.target After=system.slice -.mount systemd-journald.socket local-fs-pre.target RequiresMountsFor=/run/k3s/containerd/io.containerd.runtime.v2.task/k8s.io/6ebbd28459368680767dd0a519affa8f9dc21aa1b73cc88c6f3d80140f950771 Description=/run/k3s/containerd/io.containerd.runtime.v2.task/k8s.io/6ebbd28459368680767dd0a519affa8f9dc21aa1b73cc88c6f3d80140f950771/rootfs LoadState=loaded ActiveState=active SubState=mounted SourcePath=/proc/self/mountinfo StateChangeTimestamp=Mon 2021-07-05 20:40:34 CEST StateChangeTimestampMonotonic=4149692709 InactiveExitTimestamp=Mon 2021-07-05 20:40:34 CEST InactiveExitTimestampMonotonic=4149692709 ActiveEnterTimestamp=Mon 2021-07-05 20:40:34 CEST ActiveEnterTimestampMonotonic=4149692709 ActiveExitTimestampMonotonic=0 InactiveEnterTimestampMonotonic=0 CanStart=yes CanStop=yes CanReload=yes CanIsolate=no StopWhenUnneeded=no RefuseManualStart=no RefuseManualStop=no AllowIsolate=no DefaultDependencies=yes OnFailureJobMode=replace IgnoreOnIsolate=yes NeedDaemonReload=no JobTimeoutUSec=infinity JobRunningTimeoutUSec=infinity JobTimeoutAction=none ConditionResult=no AssertResult=no ConditionTimestampMonotonic=0 AssertTimestampMonotonic=0 Transient=no Perpetual=no StartLimitIntervalUSec=10s StartLimitBurst=5 StartLimitAction=none FailureAction=none FailureActionExitStatus=-1 SuccessAction=none SuccessActionExitStatus=-1 InvocationID=33c6a692294f4930a8a262de75da7ae0 CollectMode=inactive
root@kpi4:/home/pi# k3s ctr plugins list | grep snapshotter io.containerd.snapshotter.v1 overlayfs linux/arm64/v8 ok
io.containerd.snapshotter.v1 native linux/arm64/v8 ok
io.containerd.snapshotter.v1 fuse-overlayfs linux/arm64/v8 ok
Jul 05 20:28:55 kpi4 kernel: IPVS: [wrr] scheduler registered. Jul 05 20:28:55 kpi4 kernel: IPVS: [sh] scheduler registered. Jul 05 20:28:55 kpi4 k3s[12002]: E0705 20:28:55.529790 12002 node.go:161] Failed to retrieve node info: nodes "kpi4" is forbidden: User "system:kube-proxy" cannot get resource "nodes" in API group "" at the cluster scope Jul 05 20:28:55 kpi4 k3s[12002]: I0705 20:28:55.596000 12002 dynamic_cafile_content.go:167] Starting client-ca-bundle::/var/lib/rancher/k3s/agent/client-ca.crt Jul 05 20:28:55 kpi4 k3s[12002]: W0705 20:28:55.708937 12002 sysinfo.go:203] Nodes topology is not available, providing CPU topology Jul 05 20:28:55 kpi4 k3s[12002]: I0705 20:28:55.722155 12002 server.go:660] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Jul 05 20:28:55 kpi4 k3s[12002]: I0705 20:28:55.724476 12002 container_manager_linux.go:291] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Jul 05 20:28:55 kpi4 k3s[12002]: I0705 20:28:55.725754 12002 container_manager_linux.go:296] "Creating Container Manager object based on Node Config" nodeConfig={RuntimeCgroupsName: SystemCgroupsName: KubeletCgroupsName: ContainerRuntime:remote CgroupsPerQOS:true CgroupRo Jul 05 20:28:55 kpi4 k3s[12002]: I0705 20:28:55.727209 12002 topology_manager.go:120] "Creating topology manager with policy per scope" topologyPolicyName="none" topologyScopeName="container" Jul 05 20:28:55 kpi4 k3s[12002]: I0705 20:28:55.728431 12002 container_manager_linux.go:327] "Initializing Topology Manager" policy="none" scope="container" Jul 05 20:28:55 kpi4 k3s[12002]: I0705 20:28:55.729568 12002 container_manager_linux.go:332] "Creating device plugin manager" devicePluginEnabled=true Jul 05 20:28:55 kpi4 k3s[12002]: I0705 20:28:55.732465 12002 kubelet.go:404] "Attempting to sync node with API server" Jul 05 20:28:55 kpi4 k3s[12002]: I0705 20:28:55.734277 12002 kubelet.go:272] "Adding static pod path" path="/var/lib/rancher/k3s/agent/pod-manifests" Jul 05 20:28:55 kpi4 k3s[12002]: I0705 20:28:55.736942 12002 kubelet.go:283] "Adding apiserver pod source" Jul 05 20:28:55 kpi4 k3s[12002]: I0705 20:28:55.738428 12002 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Jul 05 20:28:55 kpi4 k3s[12002]: I0705 20:28:55.742594 12002 kuberuntime_manager.go:222] "Container runtime initialized" containerRuntime="containerd" version="v1.4.4-k3s2" apiVersion="v1alpha2" ...skipping... Jul 05 20:40:34 kpi4 systemd[1]: run-containerd-runc-k8s.io-6ebbd28459368680767dd0a519affa8f9dc21aa1b73cc88c6f3d80140f950771-runc.cDmIjD.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://www.debian.org/support -- -- The unit run-containerd-runc-k8s.io-6ebbd28459368680767dd0a519affa8f9dc21aa1b73cc88c6f3d80140f950771-runc.cDmIjD.mount has successfully entered the 'dead' state. Jul 05 20:40:34 kpi4 systemd[652]: run-containerd-runc-k8s.io-6ebbd28459368680767dd0a519affa8f9dc21aa1b73cc88c6f3d80140f950771-runc.cDmIjD.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://www.debian.org/support -- -- The unit UNIT has successfully entered the 'dead' state. Jul 05 20:40:35 kpi4 systemd[652]: tmp-ctd\x2dvolume667618187.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://www.debian.org/support -- -- The unit UNIT has successfully entered the 'dead' state. Jul 05 20:40:35 kpi4 systemd[1]: tmp-ctd\x2dvolume667618187.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://www.debian.org/support -- -- The unit tmp-ctd\x2dvolume667618187.mount has successfully entered the 'dead' state. Jul 05 20:40:35 kpi4 dhcpcd[489]: vethe5a04940: probing for an IPv4LL address Jul 05 20:40:40 kpi4 dhcpcd[489]: vethe5a04940: using IPv4LL address 169.254.9.128 Jul 05 20:40:40 kpi4 dhcpcd[489]: vethe5a04940: adding route to 169.254.0.0/16 Jul 05 20:40:40 kpi4 avahi-daemon[482]: Joining mDNS multicast group on interface vethe5a04940.IPv4 with address 169.254.9.128. Jul 05 20:40:40 kpi4 avahi-daemon[482]: New relevant interface vethe5a04940.IPv4 for mDNS.