labring / sealos

Sealos is a production-ready Kubernetes distribution. You can run any Docker image on sealos, start high availability databases like mysql/pgsql/redis/mongo, develop applications using any Programming language.
https://cloud.sealos.io
Apache License 2.0
13.84k stars 2.07k forks source link

BUG: 用一键脚本安装时候在安装 Ingress-nginx-controller 和 Kubeblocks.这步显示couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request #4670

Closed binaryYuki closed 5 months ago

binaryYuki commented 5 months ago

Sealos Version

v5.0.0

How to reproduce the bug?

The same as #4528.

多个 pod 报错:

/home/ubuntu# kubectl get pod -A 

NAMESPACE       NAME                                             READY   STATUS             RESTARTS        AGE
cert-manager    cert-manager-655bf9748f-svr76                    1/1     Running            0               4m44s
cert-manager    cert-manager-cainjector-7985fb445b-zzm68         0/1     CrashLoopBackOff   4 (89s ago)     4m44s
cert-manager    cert-manager-webhook-6dc9656f89-2jqcq            0/1     CrashLoopBackOff   6 (11s ago)     4m44s
ingress-nginx   ingress-nginx-admission-create-lwhfh             0/1     Error              4 (56s ago)     99s
kube-system     cilium-4r4gx                                     1/1     Running            0               5m29s
kube-system     cilium-operator-59b49dc487-zphvr                 1/1     Running            1 (3m16s ago)   5m29s
kube-system     coredns-565d847f94-2cl2z                         0/1     Running            0               5m29s
kube-system     coredns-565d847f94-664pl                         0/1     Running            0               5m29s
kube-system     etcd-instance-20240409-1035                      1/1     Running            0               5m55s
kube-system     kube-apiserver-instance-20240409-1035            1/1     Running            0               5m54s
kube-system     kube-controller-manager-instance-20240409-1035   1/1     Running            1 (3m7s ago)    5m54s
kube-system     kube-proxy-k6wvl                                 1/1     Running            0               5m29s
kube-system     kube-scheduler-instance-20240409-1035            1/1     Running            1 (3m16s ago)   5m53s
kube-system     metrics-server-755f77d4d5-jm4lr                  0/1     CrashLoopBackOff   4 (65s ago)     4m37s
openebs         openebs-localpv-provisioner-79f4c678cd-9qc69     0/1     CrashLoopBackOff   4 (86s ago)     4m41s

Name: ingress-nginx-admission-create-lwhfh Namespace: ingress-nginx Priority: 0 Service Account: ingress-nginx-admission Node: instance-20240409-1035/10.0.0.118 Start Time: Tue, 09 Apr 2024 10:01:24 +0000 Labels: app.kubernetes.io/component=admission-webhook app.kubernetes.io/instance=ingress-nginx app.kubernetes.io/managed-by=Helm app.kubernetes.io/name=ingress-nginx app.kubernetes.io/part-of=ingress-nginx app.kubernetes.io/version=1.5.1 controller-uid=70f07f8d-096f-4a60-9197-940c7e3f33c1 helm.sh/chart=ingress-nginx-4.4.2 job-name=ingress-nginx-admission-create Annotations: Status: Running IP: 10.0.0.245 IPs: IP: 10.0.0.245 Controlled By: Job/ingress-nginx-admission-create Containers: create: Container ID: containerd://a2286ddb49d53a6e90f81ebb0ebb99d4120ae01f6bdda90f93d0fc35beb1cedc Image: registry.k8s.io/ingress-nginx/kube-webhook-certgen:v20220916-gd32f8c343 Image ID: sealos.hub:5000/ingress-nginx/kube-webhook-certgen@sha256:58534b91f48893167034470c8faaa4000bf2133f788d43351ebf54972e138bc3 Port: Host Port: Args: create --host=ingress-nginx-controller-admission,ingress-nginx-controller-admission.$(POD_NAMESPACE).svc --namespace=$(POD_NAMESPACE) --secret-name=ingress-nginx-admission State: Waiting Reason: CrashLoopBackOff Last State: Terminated Reason: Error Exit Code: 1 Started: Tue, 09 Apr 2024 10:04:13 +0000 Finished: Tue, 09 Apr 2024 10:04:13 +0000 Ready: False Restart Count: 5 Environment: POD_NAMESPACE: ingress-nginx (v1:metadata.namespace) Mounts: /var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-rvkm2 (ro) Conditions: Type Status Initialized True Ready False ContainersReady False PodScheduled True Volumes: kube-api-access-rvkm2: Type: Projected (a volume that contains injected data from multiple sources) TokenExpirationSeconds: 3607 ConfigMapName: kube-root-ca.crt ConfigMapOptional: DownwardAPI: true QoS Class: BestEffort Node-Selectors: kubernetes.io/os=linux Tolerations: node.kubernetes.io/not-ready:NoExecute op=Exists for 300s node.kubernetes.io/unreachable:NoExecute op=Exists for 300s Events: Type Reason Age From Message


Normal Scheduled 4m46s default-scheduler Successfully assigned ingress-nginx/ingress-nginx-admission-create-lwhfh to instance-20240409-1035 Normal Pulling 4m46s kubelet Pulling image "registry.k8s.io/ingress-nginx/kube-webhook-certgen:v20220916-gd32f8c343" Normal Pulled 4m45s kubelet Successfully pulled image "registry.k8s.io/ingress-nginx/kube-webhook-certgen:v20220916-gd32f8c343" in 889.527886ms (889.535366ms including waiting) Normal Created 3m19s (x5 over 4m45s) kubelet Created container create Normal Started 3m19s (x5 over 4m44s) kubelet Started container create Normal Pulled 3m19s (x4 over 4m43s) kubelet Container image "registry.k8s.io/ingress-nginx/kube-webhook-certgen:v20220916-gd32f8c343" already present on machine Warning BackOff 3m7s (x9 over 4m42s) kubelet Back-off restarting failed container

What is the expected behavior?

/

What do you see instead?

No response

Operating environment

- Sealos version:
- Docker version:
- Kubernetes version:
- Operating system:
- Runtime environment:
- Cluster size:
- Additional information:

Additional information

No response

binaryYuki commented 5 months ago

附日志:

NAMESPACE      NAME                                             READY   STATUS             RESTARTS        AGEcert-manager   cert-manager-655bf9748f-svr76                    1/1     Running            0               24m
cert-manager   cert-manager-cainjector-7985fb445b-zzm68         0/1     CrashLoopBackOff   9 (86s ago)     24m
cert-manager   cert-manager-webhook-6dc9656f89-2jqcq            0/1     CrashLoopBackOff   10 (102s ago)   24m
kube-system    cilium-4r4gx                                     1/1     Running            0               25m
kube-system    cilium-operator-59b49dc487-zphvr                 1/1     Running            1 (22m ago)     25m
kube-system    coredns-565d847f94-2cl2z                         0/1     Running            0               25m
kube-system    coredns-565d847f94-664pl                         0/1     Running            0               25m
kube-system    etcd-instance-20240409-1035                      1/1     Running            0               25m
kube-system    kube-apiserver-instance-20240409-1035            1/1     Running            0               25m
kube-system    kube-controller-manager-instance-20240409-1035   1/1     Running            1 (22m ago)     25m
kube-system    kube-proxy-k6wvl                                 1/1     Running            0               25m
kube-system    kube-scheduler-instance-20240409-1035            1/1     Running            1 (22m ago)     25m
kube-system    metrics-server-755f77d4d5-jm4lr                  0/1     CrashLoopBackOff   9 (59s ago)     24m
openebs        openebs-localpv-provisioner-79f4c678cd-9qc69     0/1     CrashLoopBackOff   9 (102s ago)    24m
root@instance-20240409-1035:/home/ubuntu# kubectl describe pod metrics-server-755f77d4d5-jm4lr 
Error from server (NotFound): pods "metrics-server-755f77d4d5-jm4lr" not found
root@instance-20240409-1035:/home/ubuntu# kubectl describe pod 
No resources found in default namespace.
root@instance-20240409-1035:/home/ubuntu# kubectl describe pod -A
Name:             cert-manager-655bf9748f-svr76
Namespace:        cert-manager
Priority:         0
Service Account:  cert-manager
Node:             instance-20240409-1035/10.0.0.118
Start Time:       Tue, 09 Apr 2024 09:58:19 +0000
Labels:           app=cert-manager
                  app.kubernetes.io/component=controller
                  app.kubernetes.io/instance=cert-manager
                  app.kubernetes.io/name=cert-manager
                  app.kubernetes.io/version=v1.8.0
                  pod-template-hash=655bf9748f
Annotations:      prometheus.io/path: /metrics
                  prometheus.io/port: 9402
                  prometheus.io/scrape: true
Status:           Running
IP:               10.0.0.153
IPs:
  IP:           10.0.0.153
Controlled By:  ReplicaSet/cert-manager-655bf9748f
Containers:
  cert-manager:
    Container ID:  containerd://4d5e1661db748d837a912289d83e1907267e206bd087d7ac6f9a7c9a45cc0d06
    Image:         quay.io/jetstack/cert-manager-controller:v1.8.0
    Image ID:      sealos.hub:5000/jetstack/cert-manager-controller@sha256:2927259596823c8ccb4ebec10aef48955880c9be13dbc5f70ccb05a11a91eee9
    Port:          9402/TCP
    Host Port:     0/TCP
    Args:
      --v=2
      --cluster-resource-namespace=$(POD_NAMESPACE)
      --leader-election-namespace=kube-system
    State:          Running
      Started:      Tue, 09 Apr 2024 09:58:26 +0000
    Ready:          True
    Restart Count:  0
    Environment:
      POD_NAMESPACE:  cert-manager (v1:metadata.namespace)
    Mounts:
      /var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-tng7d (ro)
Conditions:
  Type              Status
  Initialized       True 
  Ready             True 
  ContainersReady   True 
  PodScheduled      True 
Volumes:
  kube-api-access-tng7d:
    Type:                    Projected (a volume that contains injected data from multiple sources)
    TokenExpirationSeconds:  3607
    ConfigMapName:           kube-root-ca.crt
    ConfigMapOptional:       <nil>
    DownwardAPI:             true
QoS Class:                   BestEffort
Node-Selectors:              kubernetes.io/os=linux
Tolerations:                 node.kubernetes.io/not-ready:NoExecute op=Exists for 300s
                             node.kubernetes.io/unreachable:NoExecute op=Exists for 300s
Events:
  Type    Reason     Age   From               Message
  ----    ------     ----  ----               -------
  Normal  Scheduled  25m   default-scheduler  Successfully assigned cert-manager/cert-manager-655bf9748f-svr76 to instance-20240409-1035
  Normal  Pulling    25m   kubelet            Pulling image "quay.io/jetstack/cert-manager-controller:v1.8.0"
  Normal  Pulled     25m   kubelet            Successfully pulled image "quay.io/jetstack/cert-manager-controller:v1.8.0" in 2.549649125s (3.921623501s including waiting)
  Normal  Created    25m   kubelet            Created container cert-manager
  Normal  Started    25m   kubelet            Started container cert-manager

Name:             cert-manager-cainjector-7985fb445b-zzm68
Namespace:        cert-manager
Priority:         0
Service Account:  cert-manager-cainjector
Node:             instance-20240409-1035/10.0.0.118
Start Time:       Tue, 09 Apr 2024 09:58:19 +0000
Labels:           app=cainjector
                  app.kubernetes.io/component=cainjector
                  app.kubernetes.io/instance=cert-manager
                  app.kubernetes.io/name=cainjector
                  app.kubernetes.io/version=v1.8.0
                  pod-template-hash=7985fb445b
Annotations:      <none>
Status:           Running
IP:               10.0.0.192
IPs:
  IP:           10.0.0.192
Controlled By:  ReplicaSet/cert-manager-cainjector-7985fb445b
Containers:
  cert-manager:
    Container ID:  containerd://f4206b3d2f2fc46b40a3888ddba13b23a0e2e7ab8cff5a4cef8af810ebf97d9a
    Image:         quay.io/jetstack/cert-manager-cainjector:v1.8.0
    Image ID:      sealos.hub:5000/jetstack/cert-manager-cainjector@sha256:c747297ef5de75cb4b00d92edaee451e8068bc3d34b2e5e5502810935a1590c4
    Port:          <none>
    Host Port:     <none>
    Args:
      --v=2
      --leader-election-namespace=kube-system
    State:          Waiting
      Reason:       CrashLoopBackOff
    Last State:     Terminated
      Reason:       Error
      Exit Code:    1
      Started:      Tue, 09 Apr 2024 10:21:11 +0000
      Finished:     Tue, 09 Apr 2024 10:21:11 +0000
    Ready:          False
    Restart Count:  9
    Environment:
      POD_NAMESPACE:  cert-manager (v1:metadata.namespace)
    Mounts:
      /var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-bfklh (ro)
Conditions:
  Type              Status
  Initialized       True 
  Ready             False 
  ContainersReady   False 
  PodScheduled      True 
Volumes:
  kube-api-access-bfklh:
    Type:                    Projected (a volume that contains injected data from multiple sources)
    TokenExpirationSeconds:  3607
    ConfigMapName:           kube-root-ca.crt
    ConfigMapOptional:       <nil>
    DownwardAPI:             true
QoS Class:                   BestEffort
Node-Selectors:              kubernetes.io/os=linux
Tolerations:                 node.kubernetes.io/not-ready:NoExecute op=Exists for 300s
                             node.kubernetes.io/unreachable:NoExecute op=Exists for 300s
Events:
  Type     Reason     Age                 From               Message
  ----     ------     ----                ----               -------
  Normal   Scheduled  25m                 default-scheduler  Successfully assigned cert-manager/cert-manager-cainjector-7985fb445b-zzm68 to instance-20240409-1035
  Normal   Pulling    25m                 kubelet            Pulling image "quay.io/jetstack/cert-manager-cainjector:v1.8.0"
  Normal   Pulled     25m                 kubelet            Successfully pulled image "quay.io/jetstack/cert-manager-cainjector:v1.8.0" in 1.401366613s (1.401510614s including waiting)
  Normal   Created    21m (x5 over 25m)   kubelet            Created container cert-manager
  Normal   Started    21m (x5 over 25m)   kubelet            Started container cert-manager
  Normal   Pulled     21m (x4 over 25m)   kubelet            Container image "quay.io/jetstack/cert-manager-cainjector:v1.8.0" already present on machine
  Warning  BackOff    1s (x106 over 23m)  kubelet            Back-off restarting failed container

Name:             cert-manager-webhook-6dc9656f89-2jqcq
Namespace:        cert-manager
Priority:         0
Service Account:  cert-manager-webhook
Node:             instance-20240409-1035/10.0.0.118
Start Time:       Tue, 09 Apr 2024 09:58:19 +0000
Labels:           app=webhook
                  app.kubernetes.io/component=webhook
                  app.kubernetes.io/instance=cert-manager
                  app.kubernetes.io/name=webhook
                  app.kubernetes.io/version=v1.8.0
                  pod-template-hash=6dc9656f89
Annotations:      <none>
Status:           Running
IP:               10.0.0.110
IPs:
  IP:           10.0.0.110
Controlled By:  ReplicaSet/cert-manager-webhook-6dc9656f89
Containers:
  cert-manager:
    Container ID:  containerd://d2d8355c8f032f400fffefa09d01cf6d46b700997a957b74fbae374703e517d4
    Image:         quay.io/jetstack/cert-manager-webhook:v1.8.0
    Image ID:      sealos.hub:5000/jetstack/cert-manager-webhook@sha256:b0a6937263daf19958855ad4c687317ff96f5ab36f2a557231969a6ca352afac
    Port:          10250/TCP
    Host Port:     0/TCP
    Args:
      --v=2
      --secure-port=10250
      --dynamic-serving-ca-secret-namespace=$(POD_NAMESPACE)
      --dynamic-serving-ca-secret-name=cert-manager-webhook-ca
      --dynamic-serving-dns-names=cert-manager-webhook,cert-manager-webhook.cert-manager,cert-manager-webhook.cert-manager.svc
    State:          Waiting
      Reason:       CrashLoopBackOff
    Last State:     Terminated
      Reason:       Error
      Exit Code:    1
      Started:      Tue, 09 Apr 2024 10:20:55 +0000
      Finished:     Tue, 09 Apr 2024 10:20:55 +0000
    Ready:          False
    Restart Count:  10
    Liveness:       http-get http://:6080/livez delay=60s timeout=1s period=10s #success=1 #failure=3
    Readiness:      http-get http://:6080/healthz delay=5s timeout=1s period=5s #success=1 #failure=3
    Environment:
      POD_NAMESPACE:  cert-manager (v1:metadata.namespace)
    Mounts:
      /var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-96d7w (ro)
Conditions:
  Type              Status
  Initialized       True 
  Ready             False 
  ContainersReady   False 
  PodScheduled      True 
Volumes:
  kube-api-access-96d7w:
    Type:                    Projected (a volume that contains injected data from multiple sources)
    TokenExpirationSeconds:  3607
    ConfigMapName:           kube-root-ca.crt
    ConfigMapOptional:       <nil>
    DownwardAPI:             true
QoS Class:                   BestEffort
Node-Selectors:              kubernetes.io/os=linux
Tolerations:                 node.kubernetes.io/not-ready:NoExecute op=Exists for 300s
                             node.kubernetes.io/unreachable:NoExecute op=Exists for 300s
Events:
  Type     Reason     Age                  From               Message
  ----     ------     ----                 ----               -------
  Normal   Scheduled  25m                  default-scheduler  Successfully assigned cert-manager/cert-manager-webhook-6dc9656f89-2jqcq to instance-20240409-1035
  Normal   Pulling    25m                  kubelet            Pulling image "quay.io/jetstack/cert-manager-webhook:v1.8.0"
  Normal   Pulled     24m                  kubelet            Successfully pulled image "quay.io/jetstack/cert-manager-webhook:v1.8.0" in 1.041179064s (4.892909431s including waiting)
  Normal   Created    24m                  kubelet            Created container cert-manager
  Normal   Started    24m                  kubelet            Started container cert-manager
  Warning  Unhealthy  23m (x17 over 24m)   kubelet            Readiness probe failed: Get "http://10.0.0.110:6080/healthz": dial tcp 10.0.0.110:6080: connect: connection refused
  Warning  Unhealthy  23m (x3 over 23m)    kubelet            Liveness probe failed: Get "http://10.0.0.110:6080/livez": dial tcp 10.0.0.110:6080: connect: connection refused
  Normal   Killing    23m                  kubelet            Container cert-manager failed liveness probe, will be restarted
  Warning  BackOff    5m1s (x95 over 23m)  kubelet            Back-off restarting failed container

Name:                 cilium-4r4gx
Namespace:            kube-system
Priority:             2000001000
Priority Class Name:  system-node-critical
Service Account:      cilium
Node:                 instance-20240409-1035/10.0.0.118
Start Time:           Tue, 09 Apr 2024 09:57:34 +0000
Labels:               controller-revision-hash=5b77d79447
                      k8s-app=cilium
                      pod-template-generation=1
Annotations:          container.apparmor.security.beta.kubernetes.io/apply-sysctl-overwrites: unconfined
                      container.apparmor.security.beta.kubernetes.io/cilium-agent: unconfined
                      container.apparmor.security.beta.kubernetes.io/clean-cilium-state: unconfined
                      container.apparmor.security.beta.kubernetes.io/mount-cgroup: unconfined
Status:               Running
IP:                   10.0.0.118
IPs:
  IP:           10.0.0.118
Controlled By:  DaemonSet/cilium
Init Containers:
  mount-cgroup:
    Container ID:  containerd://2884d23e2fbb3b243b90b029f29a27c81aa2edd30a3f5fd3d9553b57f3731055
    Image:         quay.io/cilium/cilium:v1.12.14
    Image ID:      sealos.hub:5000/cilium/cilium@sha256:7ce0308f599539515252a06a34549d9861861fb6392beb98b7458683e47dca8e
    Port:          <none>
    Host Port:     <none>
    Command:
      sh
      -ec
      cp /usr/bin/cilium-mount /hostbin/cilium-mount;
      nsenter --cgroup=/hostproc/1/ns/cgroup --mount=/hostproc/1/ns/mnt "${BIN_PATH}/cilium-mount" $CGROUP_ROOT;
      rm /hostbin/cilium-mount

    State:          Terminated
      Reason:       Completed
      Exit Code:    0
      Started:      Tue, 09 Apr 2024 09:57:53 +0000
      Finished:     Tue, 09 Apr 2024 09:57:53 +0000
    Ready:          True
    Restart Count:  0
    Environment:
      CGROUP_ROOT:  /run/cilium/cgroupv2
      BIN_PATH:     /opt/cni/bin
    Mounts:
      /hostbin from cni-path (rw)
      /hostproc from hostproc (rw)
      /var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-ml8xf (ro)
  apply-sysctl-overwrites:
    Container ID:  containerd://f75e23877dd9e8477c25fe4dfaa70dc4df5fd8f79e5aadff3094492a1937b53c
    Image:         quay.io/cilium/cilium:v1.12.14
    Image ID:      sealos.hub:5000/cilium/cilium@sha256:7ce0308f599539515252a06a34549d9861861fb6392beb98b7458683e47dca8e
    Port:          <none>
    Host Port:     <none>
    Command:
      sh
      -ec
      cp /usr/bin/cilium-sysctlfix /hostbin/cilium-sysctlfix;
      nsenter --mount=/hostproc/1/ns/mnt "${BIN_PATH}/cilium-sysctlfix";
      rm /hostbin/cilium-sysctlfix

    State:          Terminated
      Reason:       Completed
      Exit Code:    0
      Started:      Tue, 09 Apr 2024 09:58:01 +0000
      Finished:     Tue, 09 Apr 2024 09:58:01 +0000
    Ready:          True
    Restart Count:  0
    Environment:
      BIN_PATH:  /opt/cni/bin
    Mounts:
      /hostbin from cni-path (rw)
      /hostproc from hostproc (rw)
      /var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-ml8xf (ro)
  mount-bpf-fs:
    Container ID:  containerd://98d684ff89657cd35be6cc6665abdc1e81c2e4c577e293a42181c274d3319804
    Image:         quay.io/cilium/cilium:v1.12.14
    Image ID:      sealos.hub:5000/cilium/cilium@sha256:7ce0308f599539515252a06a34549d9861861fb6392beb98b7458683e47dca8e
    Port:          <none>
    Host Port:     <none>
    Command:
      /bin/bash
      -c
      --
    Args:
      mount | grep "/sys/fs/bpf type bpf" || mount -t bpf bpf /sys/fs/bpf
    State:          Terminated
      Reason:       Completed
      Exit Code:    0
      Started:      Tue, 09 Apr 2024 09:58:02 +0000
      Finished:     Tue, 09 Apr 2024 09:58:02 +0000
    Ready:          True
    Restart Count:  0
    Environment:    <none>
    Mounts:
      /sys/fs/bpf from bpf-maps (rw)
      /var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-ml8xf (ro)
  clean-cilium-state:
    Container ID:  containerd://4046039585c8cdec2cb6b98f2d79e2effe7228d9021c0c78cff4144ff058c50d
    Image:         quay.io/cilium/cilium:v1.12.14
    Image ID:      sealos.hub:5000/cilium/cilium@sha256:7ce0308f599539515252a06a34549d9861861fb6392beb98b7458683e47dca8e
    Port:          <none>
    Host Port:     <none>
    Command:
      /init-container.sh
    State:          Terminated
      Reason:       Completed
      Exit Code:    0
      Started:      Tue, 09 Apr 2024 09:58:04 +0000
      Finished:     Tue, 09 Apr 2024 09:58:04 +0000
    Ready:          True
    Restart Count:  0
    Requests:
      cpu:     100m
      memory:  100Mi
    Environment:
      CILIUM_ALL_STATE:         <set to the key 'clean-cilium-state' of config map 'cilium-config'>      Optional: true
      CILIUM_BPF_STATE:         <set to the key 'clean-cilium-bpf-state' of config map 'cilium-config'>  Optional: true
      KUBERNETES_SERVICE_HOST:  apiserver.cluster.local
      KUBERNETES_SERVICE_PORT:  6443
    Mounts:
      /run/cilium/cgroupv2 from cilium-cgroup (rw)
      /sys/fs/bpf from bpf-maps (rw)
      /var/run/cilium from cilium-run (rw)
      /var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-ml8xf (ro)
  install-cni-binaries:
    Container ID:  containerd://4f1832231a8f6c167f63aeaf3342512d1bfa130062ee8983a51a5229128355d3
    Image:         quay.io/cilium/cilium:v1.12.14
    Image ID:      sealos.hub:5000/cilium/cilium@sha256:7ce0308f599539515252a06a34549d9861861fb6392beb98b7458683e47dca8e
    Port:          <none>
    Host Port:     <none>
    Command:
      /install-plugin.sh
    State:          Terminated
      Reason:       Completed
      Exit Code:    0
      Started:      Tue, 09 Apr 2024 09:58:04 +0000
      Finished:     Tue, 09 Apr 2024 09:58:05 +0000
    Ready:          True
    Restart Count:  0
    Requests:
      cpu:        100m
      memory:     10Mi
    Environment:  <none>
    Mounts:
      /host/opt/cni/bin from cni-path (rw)
      /var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-ml8xf (ro)
Containers:
  cilium-agent:
    Container ID:  containerd://77985d178df9343391c6b470c049030d024f8c72b41e9ffb19392ece8098bfd2
    Image:         quay.io/cilium/cilium:v1.12.14
    Image ID:      sealos.hub:5000/cilium/cilium@sha256:7ce0308f599539515252a06a34549d9861861fb6392beb98b7458683e47dca8e
    Port:          <none>
    Host Port:     <none>
    Command:
      cilium-agent
    Args:
      --config-dir=/tmp/cilium/config-map
    State:          Running
      Started:      Tue, 09 Apr 2024 09:58:06 +0000
    Ready:          True
    Restart Count:  0
    Liveness:       http-get http://127.0.0.1:9879/healthz delay=0s timeout=5s period=30s #success=1 #failure=10
    Readiness:      http-get http://127.0.0.1:9879/healthz delay=0s timeout=5s period=30s #success=1 #failure=3
    Startup:        http-get http://127.0.0.1:9879/healthz delay=0s timeout=1s period=2s #success=1 #failure=105
    Environment:
      K8S_NODE_NAME:               (v1:spec.nodeName)
      CILIUM_K8S_NAMESPACE:       kube-system (v1:metadata.namespace)
      CILIUM_CLUSTERMESH_CONFIG:  /var/lib/cilium/clustermesh/
      CILIUM_CNI_CHAINING_MODE:   <set to the key 'cni-chaining-mode' of config map 'cilium-config'>  Optional: true
      CILIUM_CUSTOM_CNI_CONF:     <set to the key 'custom-cni-conf' of config map 'cilium-config'>    Optional: true
      KUBERNETES_SERVICE_HOST:    apiserver.cluster.local
      KUBERNETES_SERVICE_PORT:    6443
    Mounts:
      /host/etc/cni/net.d from etc-cni-netd (rw)
      /host/proc/sys/kernel from host-proc-sys-kernel (rw)
      /host/proc/sys/net from host-proc-sys-net (rw)
      /lib/modules from lib-modules (ro)
      /run/xtables.lock from xtables-lock (rw)
      /sys/fs/bpf from bpf-maps (rw)
      /tmp/cilium/config-map from cilium-config-path (ro)
      /var/lib/cilium/clustermesh from clustermesh-secrets (ro)
      /var/lib/cilium/tls/hubble from hubble-tls (ro)
      /var/run/cilium from cilium-run (rw)
      /var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-ml8xf (ro)
Conditions:
  Type              Status
  Initialized       True 
  Ready             True 
  ContainersReady   True 
  PodScheduled      True 
Volumes:
  cilium-run:
    Type:          HostPath (bare host directory volume)
    Path:          /var/run/cilium
    HostPathType:  DirectoryOrCreate
  bpf-maps:
    Type:          HostPath (bare host directory volume)
    Path:          /sys/fs/bpf
    HostPathType:  DirectoryOrCreate
  hostproc:
    Type:          HostPath (bare host directory volume)
    Path:          /proc
    HostPathType:  Directory
  cilium-cgroup:
    Type:          HostPath (bare host directory volume)
    Path:          /run/cilium/cgroupv2
    HostPathType:  DirectoryOrCreate
  cni-path:
    Type:          HostPath (bare host directory volume)
    Path:          /opt/cni/bin
    HostPathType:  DirectoryOrCreate
  etc-cni-netd:
    Type:          HostPath (bare host directory volume)
    Path:          /etc/cni/net.d
    HostPathType:  DirectoryOrCreate
  lib-modules:
    Type:          HostPath (bare host directory volume)
    Path:          /lib/modules
    HostPathType:  
  xtables-lock:
    Type:          HostPath (bare host directory volume)
    Path:          /run/xtables.lock
    HostPathType:  FileOrCreate
  clustermesh-secrets:
    Type:        Secret (a volume populated by a Secret)
    SecretName:  cilium-clustermesh
    Optional:    true
  cilium-config-path:
    Type:      ConfigMap (a volume populated by a ConfigMap)
    Name:      cilium-config
    Optional:  false
  host-proc-sys-net:
    Type:          HostPath (bare host directory volume)
    Path:          /proc/sys/net
    HostPathType:  Directory
  host-proc-sys-kernel:
    Type:          HostPath (bare host directory volume)
    Path:          /proc/sys/kernel
    HostPathType:  Directory
  hubble-tls:
    Type:                Projected (a volume that contains injected data from multiple sources)
    SecretName:          hubble-server-certs
    SecretOptionalName:  0x40008f9c9a
  kube-api-access-ml8xf:
    Type:                    Projected (a volume that contains injected data from multiple sources)
    TokenExpirationSeconds:  3607
    ConfigMapName:           kube-root-ca.crt
    ConfigMapOptional:       <nil>
    DownwardAPI:             true
QoS Class:                   Burstable
Node-Selectors:              kubernetes.io/os=linux
Tolerations:                 op=Exists
                             node.kubernetes.io/disk-pressure:NoSchedule op=Exists
                             node.kubernetes.io/memory-pressure:NoSchedule op=Exists
                             node.kubernetes.io/network-unavailable:NoSchedule op=Exists
                             node.kubernetes.io/not-ready:NoExecute op=Exists
                             node.kubernetes.io/pid-pressure:NoSchedule op=Exists
                             node.kubernetes.io/unreachable:NoExecute op=Exists
                             node.kubernetes.io/unschedulable:NoSchedule op=Exists
Events:
  Type     Reason     Age                From               Message
  ----     ------     ----               ----               -------
  Normal   Scheduled  25m                default-scheduler  Successfully assigned kube-system/cilium-4r4gx to instance-20240409-1035
  Normal   Pulling    25m                kubelet            Pulling image "quay.io/cilium/cilium:v1.12.14"
  Normal   Pulled     25m                kubelet            Successfully pulled image "quay.io/cilium/cilium:v1.12.14" in 18.483941886s (18.483995566s including waiting)
  Normal   Created    25m                kubelet            Created container mount-cgroup
  Normal   Started    25m                kubelet            Started container mount-cgroup
  Normal   Pulled     25m                kubelet            Container image "quay.io/cilium/cilium:v1.12.14" already present on machine
  Normal   Created    25m                kubelet            Created container apply-sysctl-overwrites
  Normal   Started    25m                kubelet            Started container apply-sysctl-overwrites
  Normal   Pulled     25m                kubelet            Container image "quay.io/cilium/cilium:v1.12.14" already present on machine
  Normal   Created    25m                kubelet            Created container mount-bpf-fs
  Normal   Started    25m                kubelet            Started container mount-bpf-fs
  Normal   Pulled     25m                kubelet            Container image "quay.io/cilium/cilium:v1.12.14" already present on machine
  Normal   Created    25m                kubelet            Created container clean-cilium-state
  Normal   Started    25m                kubelet            Started container clean-cilium-state
  Normal   Pulled     25m                kubelet            Container image "quay.io/cilium/cilium:v1.12.14" already present on machine
  Normal   Created    25m                kubelet            Created container install-cni-binaries
  Normal   Started    25m                kubelet            Started container install-cni-binaries
  Normal   Pulled     25m                kubelet            Container image "quay.io/cilium/cilium:v1.12.14" already present on machine
  Normal   Created    25m                kubelet            Created container cilium-agent
  Normal   Started    25m                kubelet            Started container cilium-agent
  Warning  Unhealthy  25m (x6 over 25m)  kubelet            Startup probe failed: Get "http://127.0.0.1:9879/healthz": dial tcp 127.0.0.1:9879: connect: connection refused

Name:                 cilium-operator-59b49dc487-zphvr
Namespace:            kube-system
Priority:             2000000000
Priority Class Name:  system-cluster-critical
Service Account:      cilium-operator
Node:                 instance-20240409-1035/10.0.0.118
Start Time:           Tue, 09 Apr 2024 09:57:34 +0000
Labels:               io.cilium/app=operator
                      name=cilium-operator
                      pod-template-hash=59b49dc487
Annotations:          <none>
Status:               Running
IP:                   10.0.0.118
IPs:
  IP:           10.0.0.118
Controlled By:  ReplicaSet/cilium-operator-59b49dc487
Containers:
  cilium-operator:
    Container ID:  containerd://7de041179905c067c96595a64f2ae5c9f6cbfafb79d9173be63fa02ad51f27ce
    Image:         quay.io/cilium/operator:v1.12.14
    Image ID:      sealos.hub:5000/cilium/operator@sha256:e29749f441e3450a4bf4a15033dedc55e8928c3075b229d4c65d2a9768a4db8c
    Port:          <none>
    Host Port:     <none>
    Command:
      cilium-operator
    Args:
      --config-dir=/tmp/cilium/config-map
      --debug=$(CILIUM_DEBUG)
    State:       Running
      Started:   Tue, 09 Apr 2024 10:00:05 +0000
    Last State:  Terminated
      Reason:    Error
      Message:   Rs="[10.0.0.0/8]" ipv6CIDRs="[]" subsys=ipam-allocator-clusterpool
level=info msg="Removing Cilium Node Taints or Setting Cilium Is Up Condition for Kubernetes Nodes" k8sNamespace=kube-system label-selector="k8s-app=cilium" remove-cilium-node-taints=true set-cilium-is-up-condition=true subsys=cilium-operator
level=info msg="Starting to synchronize CiliumNode custom resources" subsys=cilium-operator
level=info msg="Starting to garbage collect stale CiliumNode custom resources" subsys=watchers
level=info msg="Garbage collected status/nodes in Cilium Clusterwide Network Policies found=0, gc=0" subsys=cilium-operator
level=info msg="Garbage collected status/nodes in Cilium Network Policies found=0, gc=0" subsys=cilium-operator
level=info msg="CiliumNodes caches synced with Kubernetes" subsys=cilium-operator
level=info msg="Starting to garbage collect stale CiliumEndpoint custom resources" subsys=cilium-operator
level=info msg="Starting CRD identity garbage collector" interval=15m0s subsys=cilium-operator
level=info msg="Starting CNP derivative handler" subsys=cilium-operator
level=info msg="Starting CCNP derivative handler" subsys=cilium-operator
level=info msg="Initialization complete" subsys=cilium-operator
Failed to update lock: etcdserver: request timed out
level=error msg="Failed to update lock: etcdserver: request timed out" subsys=klog
level=info msg="failed to renew lease kube-system/cilium-operator-resource-lock: timed out waiting for the condition" subsys=klog
Failed to release lock: Operation cannot be fulfilled on leases.coordination.k8s.io "cilium-operator-resource-lock": the object has been modified; please apply your changes to the latest version and try again
level=error msg="Failed to release lock: Operation cannot be fulfilled on leases.coordination.k8s.io \"cilium-operator-resource-lock\": the object has been modified; please apply your changes to the latest version and try again" subsys=klog
level=info msg="Leader election lost" operator-id=instance-20240409-1035-mVfidDJfRI subsys=cilium-operator

      Exit Code:    1
      Started:      Tue, 09 Apr 2024 09:58:03 +0000
      Finished:     Tue, 09 Apr 2024 09:59:47 +0000
    Ready:          True
    Restart Count:  1
    Liveness:       http-get http://127.0.0.1:9234/healthz delay=60s timeout=3s period=10s #success=1 #failure=3
    Environment:
      K8S_NODE_NAME:             (v1:spec.nodeName)
      CILIUM_K8S_NAMESPACE:     kube-system (v1:metadata.namespace)
      CILIUM_DEBUG:             <set to the key 'debug' of config map 'cilium-config'>  Optional: true
      KUBERNETES_SERVICE_HOST:  apiserver.cluster.local
      KUBERNETES_SERVICE_PORT:  6443
    Mounts:
      /tmp/cilium/config-map from cilium-config-path (ro)
      /var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-sp4jp (ro)
Conditions:
  Type              Status
  Initialized       True 
  Ready             True 
  ContainersReady   True 
  PodScheduled      True 
Volumes:
  cilium-config-path:
    Type:      ConfigMap (a volume populated by a ConfigMap)
    Name:      cilium-config
    Optional:  false
  kube-api-access-sp4jp:
    Type:                    Projected (a volume that contains injected data from multiple sources)
    TokenExpirationSeconds:  3607
    ConfigMapName:           kube-root-ca.crt
    ConfigMapOptional:       <nil>
    DownwardAPI:             true
QoS Class:                   BestEffort
Node-Selectors:              kubernetes.io/os=linux
Tolerations:                 op=Exists
Events:
  Type     Reason     Age                From               Message
  ----     ------     ----               ----               -------
  Normal   Scheduled  25m                default-scheduler  Successfully assigned kube-system/cilium-operator-59b49dc487-zphvr to instance-20240409-1035
  Normal   Pulling    25m                kubelet            Pulling image "quay.io/cilium/operator:v1.12.14"
  Normal   Pulled     25m                kubelet            Successfully pulled image "quay.io/cilium/operator:v1.12.14" in 8.590662366s (26.380157037s including waiting)
  Warning  Unhealthy  23m (x3 over 24m)  kubelet            Liveness probe failed: Get "http://127.0.0.1:9234/healthz": context deadline exceeded (Client.Timeout exceeded while awaiting headers)
  Normal   Killing    23m                kubelet            Container cilium-operator failed liveness probe, will be restarted
  Normal   Pulled     23m                kubelet            Container image "quay.io/cilium/operator:v1.12.14" already present on machine
  Normal   Created    23m (x2 over 25m)  kubelet            Created container cilium-operator
  Normal   Started    23m (x2 over 25m)  kubelet            Started container cilium-operator

Name:                 coredns-565d847f94-2cl2z
Namespace:            kube-system
Priority:             2000000000
Priority Class Name:  system-cluster-critical
Service Account:      coredns
Node:                 instance-20240409-1035/10.0.0.118
Start Time:           Tue, 09 Apr 2024 09:58:16 +0000
Labels:               k8s-app=kube-dns
                      pod-template-hash=565d847f94
Annotations:          <none>
Status:               Running
IP:                   10.0.0.60
IPs:
  IP:           10.0.0.60
Controlled By:  ReplicaSet/coredns-565d847f94
Containers:
  coredns:
    Container ID:  containerd://cae4df91ea8354ab3b146bcf7b274a61297e567dffa124adbb5f0f2192a8e588
    Image:         registry.k8s.io/coredns/coredns:v1.9.3
    Image ID:      sealos.hub:5000/coredns/coredns@sha256:85c4677668f548da72cc0194663ae9d412cac8f239551ca90c0941c9c1ad8685
    Ports:         53/UDP, 53/TCP, 9153/TCP
    Host Ports:    0/UDP, 0/TCP, 0/TCP
    Args:
      -conf
      /etc/coredns/Corefile
    State:          Running
      Started:      Tue, 09 Apr 2024 09:58:22 +0000
    Ready:          False
    Restart Count:  0
    Limits:
      memory:  170Mi
    Requests:
      cpu:        100m
      memory:     70Mi
    Liveness:     http-get http://:8080/health delay=60s timeout=5s period=10s #success=1 #failure=5
    Readiness:    http-get http://:8181/ready delay=0s timeout=1s period=10s #success=1 #failure=3
    Environment:  <none>
    Mounts:
      /etc/coredns from config-volume (ro)
      /var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-vndjm (ro)
Conditions:
  Type              Status
  Initialized       True 
  Ready             False 
  ContainersReady   False 
  PodScheduled      True 
Volumes:
  config-volume:
    Type:      ConfigMap (a volume populated by a ConfigMap)
    Name:      coredns
    Optional:  false
  kube-api-access-vndjm:
    Type:                    Projected (a volume that contains injected data from multiple sources)
    TokenExpirationSeconds:  3607
    ConfigMapName:           kube-root-ca.crt
    ConfigMapOptional:       <nil>
    DownwardAPI:             true
QoS Class:                   Burstable
Node-Selectors:              kubernetes.io/os=linux
Tolerations:                 CriticalAddonsOnly op=Exists
                             node-role.kubernetes.io/control-plane:NoSchedule
                             node-role.kubernetes.io/master:NoSchedule
                             node.kubernetes.io/not-ready:NoExecute op=Exists for 300s
                             node.kubernetes.io/unreachable:NoExecute op=Exists for 300s
Events:
  Type     Reason            Age                 From               Message
  ----     ------            ----                ----               -------
  Warning  FailedScheduling  25m                 default-scheduler  0/1 nodes are available: 1 node(s) had untolerated taint {node.kubernetes.io/not-ready: }. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.
  Normal   Scheduled         25m                 default-scheduler  Successfully assigned kube-system/coredns-565d847f94-2cl2z to instance-20240409-1035
  Normal   Pulled            25m                 kubelet            Container image "registry.k8s.io/coredns/coredns:v1.9.3" already present on machine
  Normal   Created           25m                 kubelet            Created container coredns
  Normal   Started           25m                 kubelet            Started container coredns
  Warning  Unhealthy         25m                 kubelet            Readiness probe failed: Get "http://10.0.0.60:8181/ready": context deadline exceeded (Client.Timeout exceeded while awaiting headers)
  Warning  Unhealthy         0s (x171 over 25m)  kubelet            Readiness probe failed: HTTP probe failed with statuscode: 503

Name:                 coredns-565d847f94-664pl
Namespace:            kube-system
Priority:             2000000000
Priority Class Name:  system-cluster-critical
Service Account:      coredns
Node:                 instance-20240409-1035/10.0.0.118
Start Time:           Tue, 09 Apr 2024 09:58:16 +0000
Labels:               k8s-app=kube-dns
                      pod-template-hash=565d847f94
Annotations:          <none>
Status:               Running
IP:                   10.0.0.83
IPs:
  IP:           10.0.0.83
Controlled By:  ReplicaSet/coredns-565d847f94
Containers:
  coredns:
    Container ID:  containerd://62679cd5ea971640c21216ecfd1c069e5451a2ba6cf831577218b318ea543078
    Image:         registry.k8s.io/coredns/coredns:v1.9.3
    Image ID:      sealos.hub:5000/coredns/coredns@sha256:85c4677668f548da72cc0194663ae9d412cac8f239551ca90c0941c9c1ad8685
    Ports:         53/UDP, 53/TCP, 9153/TCP
    Host Ports:    0/UDP, 0/TCP, 0/TCP
    Args:
      -conf
      /etc/coredns/Corefile
    State:          Running
      Started:      Tue, 09 Apr 2024 09:58:21 +0000
    Ready:          False
    Restart Count:  0
    Limits:
      memory:  170Mi
    Requests:
      cpu:        100m
      memory:     70Mi
    Liveness:     http-get http://:8080/health delay=60s timeout=5s period=10s #success=1 #failure=5
    Readiness:    http-get http://:8181/ready delay=0s timeout=1s period=10s #success=1 #failure=3
    Environment:  <none>
    Mounts:
      /etc/coredns from config-volume (ro)
      /var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-9hzx6 (ro)
Conditions:
  Type              Status
  Initialized       True 
  Ready             False 
  ContainersReady   False 
  PodScheduled      True 
Volumes:
  config-volume:
    Type:      ConfigMap (a volume populated by a ConfigMap)
    Name:      coredns
    Optional:  false
  kube-api-access-9hzx6:
    Type:                    Projected (a volume that contains injected data from multiple sources)
    TokenExpirationSeconds:  3607
    ConfigMapName:           kube-root-ca.crt
    ConfigMapOptional:       <nil>
    DownwardAPI:             true
QoS Class:                   Burstable
Node-Selectors:              kubernetes.io/os=linux
Tolerations:                 CriticalAddonsOnly op=Exists
                             node-role.kubernetes.io/control-plane:NoSchedule
                             node-role.kubernetes.io/master:NoSchedule
                             node.kubernetes.io/not-ready:NoExecute op=Exists for 300s
                             node.kubernetes.io/unreachable:NoExecute op=Exists for 300s
Events:
  Type     Reason            Age                 From               Message
  ----     ------            ----                ----               -------
  Warning  FailedScheduling  25m (x2 over 25m)   default-scheduler  0/1 nodes are available: 1 node(s) had untolerated taint {node.kubernetes.io/not-ready: }. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.
  Normal   Scheduled         25m                 default-scheduler  Successfully assigned kube-system/coredns-565d847f94-664pl to instance-20240409-1035
  Normal   Pulled            25m                 kubelet            Container image "registry.k8s.io/coredns/coredns:v1.9.3" already present on machine
  Normal   Created           25m                 kubelet            Created container coredns
  Normal   Started           25m                 kubelet            Started container coredns
  Warning  Unhealthy         25m (x2 over 25m)   kubelet            Readiness probe failed: Get "http://10.0.0.83:8181/ready": context deadline exceeded (Client.Timeout exceeded while awaiting headers)
  Warning  Unhealthy         0s (x170 over 25m)  kubelet            Readiness probe failed: HTTP probe failed with statuscode: 503

Name:                 etcd-instance-20240409-1035
Namespace:            kube-system
Priority:             2000001000
Priority Class Name:  system-node-critical
Node:                 instance-20240409-1035/10.0.0.118
Start Time:           Tue, 09 Apr 2024 09:57:10 +0000
Labels:               component=etcd
                      tier=control-plane
Annotations:          kubeadm.kubernetes.io/etcd.advertise-client-urls: https://10.0.0.118:2379
                      kubernetes.io/config.hash: 255ac202a4b6285ecf9f95463fa3c5f3
                      kubernetes.io/config.mirror: 255ac202a4b6285ecf9f95463fa3c5f3
                      kubernetes.io/config.seen: 2024-04-09T09:57:00.446193904Z
                      kubernetes.io/config.source: file
Status:               Running
IP:                   10.0.0.118
IPs:
  IP:           10.0.0.118
Controlled By:  Node/instance-20240409-1035
Containers:
  etcd:
    Container ID:  containerd://fb8b18f601c4f3460adf82c17a42220e9d3d7bf2b924f9e0599683427ed03d25
    Image:         registry.k8s.io/etcd:3.5.6-0
    Image ID:      sealos.hub:5000/etcd@sha256:d8aa60417894c5563344589602f2ba5c63613d112c34a694cb84ff1e529c1975
    Port:          <none>
    Host Port:     <none>
    Command:
      etcd
      --advertise-client-urls=https://10.0.0.118:2379
      --cert-file=/etc/kubernetes/pki/etcd/server.crt
      --client-cert-auth=true
      --data-dir=/var/lib/etcd
      --experimental-initial-corrupt-check=true
      --experimental-watch-progress-notify-interval=5s
      --initial-advertise-peer-urls=https://10.0.0.118:2380
      --initial-cluster=instance-20240409-1035=https://10.0.0.118:2380
      --key-file=/etc/kubernetes/pki/etcd/server.key
      --listen-client-urls=https://127.0.0.1:2379,https://10.0.0.118:2379
      --listen-metrics-urls=http://0.0.0.0:2381
      --listen-peer-urls=https://10.0.0.118:2380
      --name=instance-20240409-1035
      --peer-cert-file=/etc/kubernetes/pki/etcd/peer.crt
      --peer-client-cert-auth=true
      --peer-key-file=/etc/kubernetes/pki/etcd/peer.key
      --peer-trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt
      --snapshot-count=10000
      --trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt
    State:          Running
      Started:      Tue, 09 Apr 2024 09:57:01 +0000
    Ready:          True
    Restart Count:  0
    Requests:
      cpu:        100m
      memory:     100Mi
    Liveness:     http-get http://0.0.0.0:2381/health%3Fexclude=NOSPACE&serializable=true delay=10s timeout=15s period=10s #success=1 #failure=8
    Startup:      http-get http://0.0.0.0:2381/health%3Fserializable=false delay=10s timeout=15s period=10s #success=1 #failure=24
    Environment:  <none>
    Mounts:
      /etc/kubernetes/pki/etcd from etcd-certs (rw)
      /var/lib/etcd from etcd-data (rw)
Conditions:
  Type              Status
  Initialized       True 
  Ready             True 
  ContainersReady   True 
  PodScheduled      True 
Volumes:
  etcd-certs:
    Type:          HostPath (bare host directory volume)
    Path:          /etc/kubernetes/pki/etcd
    HostPathType:  DirectoryOrCreate
  etcd-data:
    Type:          HostPath (bare host directory volume)
    Path:          /var/lib/etcd
    HostPathType:  DirectoryOrCreate
QoS Class:         Burstable
Node-Selectors:    <none>
Tolerations:       :NoExecute op=Exists
Events:            <none>

Name:                 kube-apiserver-instance-20240409-1035
Namespace:            kube-system
Priority:             2000001000
Priority Class Name:  system-node-critical
Node:                 instance-20240409-1035/10.0.0.118
Start Time:           Tue, 09 Apr 2024 09:57:10 +0000
Labels:               component=kube-apiserver
                      tier=control-plane
Annotations:          kubeadm.kubernetes.io/kube-apiserver.advertise-address.endpoint: 10.0.0.118:6443
                      kubernetes.io/config.hash: 5d0e812d4276731d589ddecbc2f799fe
                      kubernetes.io/config.mirror: 5d0e812d4276731d589ddecbc2f799fe
                      kubernetes.io/config.seen: 2024-04-09T09:57:09.539430631Z
                      kubernetes.io/config.source: file
Status:               Running
IP:                   10.0.0.118
IPs:
  IP:           10.0.0.118
Controlled By:  Node/instance-20240409-1035
Containers:
  kube-apiserver:
    Container ID:  containerd://8c0455b07745403222ac4461d4ca98a74c03c61fdbf7717e28ada76b8b680106
    Image:         registry.k8s.io/kube-apiserver:v1.25.6
    Image ID:      sealos.hub:5000/kube-apiserver@sha256:563e1fc5acd5e5f1b3e5914fe146a236f0c1feb92c3cc8966b5e1fd776fe2d4f
    Port:          <none>
    Host Port:     <none>
    Command:
      kube-apiserver
      --advertise-address=10.0.0.118
      --allow-privileged=true
      --audit-log-format=json
      --audit-log-maxage=7
      --audit-log-maxbackup=10
      --audit-log-maxsize=100
      --audit-log-path=/var/log/kubernetes/audit.log
      --audit-policy-file=/etc/kubernetes/audit-policy.yml
      --authorization-mode=Node,RBAC
      --client-ca-file=/etc/kubernetes/pki/ca.crt
      --enable-admission-plugins=NodeRestriction
      --enable-aggregator-routing=true
      --enable-bootstrap-token-auth=true
      --etcd-cafile=/etc/kubernetes/pki/etcd/ca.crt
      --etcd-certfile=/etc/kubernetes/pki/apiserver-etcd-client.crt
      --etcd-keyfile=/etc/kubernetes/pki/apiserver-etcd-client.key
      --etcd-servers=https://127.0.0.1:2379
      --feature-gates=EphemeralContainers=true
      --kubelet-client-certificate=/etc/kubernetes/pki/apiserver-kubelet-client.crt
      --kubelet-client-key=/etc/kubernetes/pki/apiserver-kubelet-client.key
      --kubelet-preferred-address-types=InternalIP,ExternalIP,Hostname
      --proxy-client-cert-file=/etc/kubernetes/pki/front-proxy-client.crt
      --proxy-client-key-file=/etc/kubernetes/pki/front-proxy-client.key
      --requestheader-allowed-names=front-proxy-client
      --requestheader-client-ca-file=/etc/kubernetes/pki/front-proxy-ca.crt
      --requestheader-extra-headers-prefix=X-Remote-Extra-
      --requestheader-group-headers=X-Remote-Group
      --requestheader-username-headers=X-Remote-User
      --secure-port=6443
      --service-account-issuer=https://kubernetes.default.svc.cluster.local
      --service-account-key-file=/etc/kubernetes/pki/sa.pub
      --service-account-signing-key-file=/etc/kubernetes/pki/sa.key
      --service-cluster-ip-range=10.96.0.0/22
      --tls-cert-file=/etc/kubernetes/pki/apiserver.crt
      --tls-private-key-file=/etc/kubernetes/pki/apiserver.key
    State:          Running
      Started:      Tue, 09 Apr 2024 09:57:01 +0000
    Ready:          True
    Restart Count:  0
    Requests:
      cpu:        250m
    Liveness:     http-get https://10.0.0.118:6443/livez delay=10s timeout=15s period=10s #success=1 #failure=8
    Readiness:    http-get https://10.0.0.118:6443/readyz delay=0s timeout=15s period=1s #success=1 #failure=3
    Startup:      http-get https://10.0.0.118:6443/livez delay=10s timeout=15s period=10s #success=1 #failure=24
    Environment:  <none>
    Mounts:
      /etc/ca-certificates from etc-ca-certificates (ro)
      /etc/kubernetes from audit (rw)
      /etc/kubernetes/pki from k8s-certs (ro)
      /etc/localtime from localtime (ro)
      /etc/pki from etc-pki (ro)
      /etc/ssl/certs from ca-certs (ro)
      /usr/local/share/ca-certificates from usr-local-share-ca-certificates (ro)
      /usr/share/ca-certificates from usr-share-ca-certificates (ro)
      /var/log/kubernetes from audit-log (rw)
Conditions:
  Type              Status
  Initialized       True 
  Ready             True 
  ContainersReady   True 
  PodScheduled      True 
Volumes:
  audit:
    Type:          HostPath (bare host directory volume)
    Path:          /etc/kubernetes
    HostPathType:  DirectoryOrCreate
  audit-log:
    Type:          HostPath (bare host directory volume)
    Path:          /var/log/kubernetes
    HostPathType:  DirectoryOrCreate
  ca-certs:
    Type:          HostPath (bare host directory volume)
    Path:          /etc/ssl/certs
    HostPathType:  DirectoryOrCreate
  etc-ca-certificates:
    Type:          HostPath (bare host directory volume)
    Path:          /etc/ca-certificates
    HostPathType:  DirectoryOrCreate
  etc-pki:
    Type:          HostPath (bare host directory volume)
    Path:          /etc/pki
    HostPathType:  DirectoryOrCreate
  k8s-certs:
    Type:          HostPath (bare host directory volume)
    Path:          /etc/kubernetes/pki
    HostPathType:  DirectoryOrCreate
  localtime:
    Type:          HostPath (bare host directory volume)
    Path:          /etc/localtime
    HostPathType:  File
  usr-local-share-ca-certificates:
    Type:          HostPath (bare host directory volume)
    Path:          /usr/local/share/ca-certificates
    HostPathType:  DirectoryOrCreate
  usr-share-ca-certificates:
    Type:          HostPath (bare host directory volume)
    Path:          /usr/share/ca-certificates
    HostPathType:  DirectoryOrCreate
QoS Class:         Burstable
Node-Selectors:    <none>
Tolerations:       :NoExecute op=Exists
Events:
  Type     Reason     Age                 From     Message
  ----     ------     ----                ----     -------
  Warning  Unhealthy  24m (x4 over 24m)   kubelet  Liveness probe failed: HTTP probe failed with statuscode: 500
  Warning  Unhealthy  10m (x35 over 25m)  kubelet  Readiness probe failed: HTTP probe failed with statuscode: 500

Name:                 kube-controller-manager-instance-20240409-1035
Namespace:            kube-system
Priority:             2000001000
Priority Class Name:  system-node-critical
Node:                 instance-20240409-1035/10.0.0.118
Start Time:           Tue, 09 Apr 2024 09:57:10 +0000
Labels:               component=kube-controller-manager
                      tier=control-plane
Annotations:          kubernetes.io/config.hash: 1de705f356ebbf18bae10f071a771ac6
                      kubernetes.io/config.mirror: 1de705f356ebbf18bae10f071a771ac6
                      kubernetes.io/config.seen: 2024-04-09T09:57:09.539431911Z
                      kubernetes.io/config.source: file
Status:               Running
IP:                   10.0.0.118
IPs:
  IP:           10.0.0.118
Controlled By:  Node/instance-20240409-1035
Containers:
  kube-controller-manager:
    Container ID:  containerd://0393aa990fe41b2a5360ae4721abbb66198d0a0709069456032a9ac9c3acf9ae
    Image:         registry.k8s.io/kube-controller-manager:v1.25.6
    Image ID:      sealos.hub:5000/kube-controller-manager@sha256:6b8a8b7135b88f8d1ef427e8ff63385c0b05bc938d7aed208cac5d3b49adee2b
    Port:          <none>
    Host Port:     <none>
    Command:
      kube-controller-manager
      --allocate-node-cidrs=true
      --authentication-kubeconfig=/etc/kubernetes/controller-manager.conf
      --authorization-kubeconfig=/etc/kubernetes/controller-manager.conf
      --bind-address=0.0.0.0
      --client-ca-file=/etc/kubernetes/pki/ca.crt
      --cluster-cidr=100.64.0.0/10
      --cluster-name=kubernetes
      --cluster-signing-cert-file=/etc/kubernetes/pki/ca.crt
      --cluster-signing-duration=876000h
      --cluster-signing-key-file=/etc/kubernetes/pki/ca.key
      --controllers=*,bootstrapsigner,tokencleaner
      --feature-gates=EphemeralContainers=true
      --kubeconfig=/etc/kubernetes/controller-manager.conf
      --leader-elect=true
      --requestheader-client-ca-file=/etc/kubernetes/pki/front-proxy-ca.crt
      --root-ca-file=/etc/kubernetes/pki/ca.crt
      --service-account-private-key-file=/etc/kubernetes/pki/sa.key
      --service-cluster-ip-range=10.96.0.0/22
      --use-service-account-credentials=true
    State:          Running
      Started:      Tue, 09 Apr 2024 09:59:59 +0000
    Last State:     Terminated
      Reason:       Error
      Exit Code:    1
      Started:      Tue, 09 Apr 2024 09:57:01 +0000
      Finished:     Tue, 09 Apr 2024 09:59:56 +0000
    Ready:          True
    Restart Count:  1
    Requests:
      cpu:        200m
    Liveness:     http-get https://:10257/healthz delay=10s timeout=15s period=10s #success=1 #failure=8
    Startup:      http-get https://:10257/healthz delay=10s timeout=15s period=10s #success=1 #failure=24
    Environment:  <none>
    Mounts:
      /etc/ca-certificates from etc-ca-certificates (ro)
      /etc/kubernetes/controller-manager.conf from kubeconfig (ro)
      /etc/kubernetes/pki from k8s-certs (ro)
      /etc/localtime from localtime (ro)
      /etc/pki from etc-pki (ro)
      /etc/ssl/certs from ca-certs (ro)
      /usr/libexec/kubernetes/kubelet-plugins/volume/exec from flexvolume-dir (rw)
      /usr/local/share/ca-certificates from usr-local-share-ca-certificates (ro)
      /usr/share/ca-certificates from usr-share-ca-certificates (ro)
Conditions:
  Type              Status
  Initialized       True 
  Ready             True 
  ContainersReady   True 
  PodScheduled      True 
Volumes:
  ca-certs:
    Type:          HostPath (bare host directory volume)
    Path:          /etc/ssl/certs
    HostPathType:  DirectoryOrCreate
  etc-ca-certificates:
    Type:          HostPath (bare host directory volume)
    Path:          /etc/ca-certificates
    HostPathType:  DirectoryOrCreate
  etc-pki:
    Type:          HostPath (bare host directory volume)
    Path:          /etc/pki
    HostPathType:  DirectoryOrCreate
  flexvolume-dir:
    Type:          HostPath (bare host directory volume)
    Path:          /usr/libexec/kubernetes/kubelet-plugins/volume/exec
    HostPathType:  DirectoryOrCreate
  k8s-certs:
    Type:          HostPath (bare host directory volume)
    Path:          /etc/kubernetes/pki
    HostPathType:  DirectoryOrCreate
  kubeconfig:
    Type:          HostPath (bare host directory volume)
    Path:          /etc/kubernetes/controller-manager.conf
    HostPathType:  FileOrCreate
  localtime:
    Type:          HostPath (bare host directory volume)
    Path:          /etc/localtime
    HostPathType:  File
  usr-local-share-ca-certificates:
    Type:          HostPath (bare host directory volume)
    Path:          /usr/local/share/ca-certificates
    HostPathType:  DirectoryOrCreate
  usr-share-ca-certificates:
    Type:          HostPath (bare host directory volume)
    Path:          /usr/share/ca-certificates
    HostPathType:  DirectoryOrCreate
QoS Class:         Burstable
Node-Selectors:    <none>
Tolerations:       :NoExecute op=Exists
Events:
  Type     Reason     Age                From     Message
  ----     ------     ----               ----     -------
  Warning  Unhealthy  23m (x4 over 24m)  kubelet  Liveness probe failed: Get "https://10.0.0.118:10257/healthz": net/http: TLS handshake timeout
  Warning  Unhealthy  23m                kubelet  Liveness probe failed: Get "https://10.0.0.118:10257/healthz": read tcp 10.0.0.118:2768->10.0.0.118:10257: read: connection reset by peer
  Warning  Unhealthy  23m                kubelet  Liveness probe failed: Get "https://10.0.0.118:10257/healthz": dial tcp 10.0.0.118:10257: connect: connection refused
  Normal   Pulled     23m                kubelet  Container image "registry.k8s.io/kube-controller-manager:v1.25.6" already present on machine
  Normal   Created    23m                kubelet  Created container kube-controller-manager
  Normal   Started    23m                kubelet  Started container kube-controller-manager

Name:                 kube-proxy-k6wvl
Namespace:            kube-system
Priority:             2000001000
Priority Class Name:  system-node-critical
Service Account:      kube-proxy
Node:                 instance-20240409-1035/10.0.0.118
Start Time:           Tue, 09 Apr 2024 09:57:34 +0000
Labels:               controller-revision-hash=5f6bcf49c
                      k8s-app=kube-proxy
                      pod-template-generation=1
Annotations:          <none>
Status:               Running
IP:                   10.0.0.118
IPs:
  IP:           10.0.0.118
Controlled By:  DaemonSet/kube-proxy
Containers:
  kube-proxy:
    Container ID:  containerd://cae85050bd6fe87e5b64a610e09dea726b7b381bcda28e2d1edd7bb9c0114f66
    Image:         registry.k8s.io/kube-proxy:v1.25.6
    Image ID:      sealos.hub:5000/kube-proxy@sha256:21a119d3450366e8b3405a4bb0c6585889086eed37f1c04a076f9d8d3db20597
    Port:          <none>
    Host Port:     <none>
    Command:
      /usr/local/bin/kube-proxy
      --config=/var/lib/kube-proxy/config.conf
      --hostname-override=$(NODE_NAME)
    State:          Running
      Started:      Tue, 09 Apr 2024 09:57:35 +0000
    Ready:          True
    Restart Count:  0
    Environment:
      NODE_NAME:   (v1:spec.nodeName)
    Mounts:
      /lib/modules from lib-modules (ro)
      /run/xtables.lock from xtables-lock (rw)
      /var/lib/kube-proxy from kube-proxy (rw)
      /var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-6n877 (ro)
Conditions:
  Type              Status
  Initialized       True 
  Ready             True 
  ContainersReady   True 
  PodScheduled      True 
Volumes:
  kube-proxy:
    Type:      ConfigMap (a volume populated by a ConfigMap)
    Name:      kube-proxy
    Optional:  false
  xtables-lock:
    Type:          HostPath (bare host directory volume)
    Path:          /run/xtables.lock
    HostPathType:  FileOrCreate
  lib-modules:
    Type:          HostPath (bare host directory volume)
    Path:          /lib/modules
    HostPathType:  
  kube-api-access-6n877:
    Type:                    Projected (a volume that contains injected data from multiple sources)
    TokenExpirationSeconds:  3607
    ConfigMapName:           kube-root-ca.crt
    ConfigMapOptional:       <nil>
    DownwardAPI:             true
QoS Class:                   BestEffort
Node-Selectors:              kubernetes.io/os=linux
Tolerations:                 op=Exists
                             node.kubernetes.io/disk-pressure:NoSchedule op=Exists
                             node.kubernetes.io/memory-pressure:NoSchedule op=Exists
                             node.kubernetes.io/network-unavailable:NoSchedule op=Exists
                             node.kubernetes.io/not-ready:NoExecute op=Exists
                             node.kubernetes.io/pid-pressure:NoSchedule op=Exists
                             node.kubernetes.io/unreachable:NoExecute op=Exists
                             node.kubernetes.io/unschedulable:NoSchedule op=Exists
Events:
  Type    Reason     Age   From               Message
  ----    ------     ----  ----               -------
  Normal  Scheduled  25m   default-scheduler  Successfully assigned kube-system/kube-proxy-k6wvl to instance-20240409-1035
  Normal  Pulled     25m   kubelet            Container image "registry.k8s.io/kube-proxy:v1.25.6" already present on machine
  Normal  Created    25m   kubelet            Created container kube-proxy
  Normal  Started    25m   kubelet            Started container kube-proxy

Name:                 kube-scheduler-instance-20240409-1035
Namespace:            kube-system
Priority:             2000001000
Priority Class Name:  system-node-critical
Node:                 instance-20240409-1035/10.0.0.118
Start Time:           Tue, 09 Apr 2024 09:57:09 +0000
Labels:               component=kube-scheduler
                      tier=control-plane
Annotations:          kubernetes.io/config.hash: f6f8901ecf2a1baa407655f583adee6a
                      kubernetes.io/config.mirror: f6f8901ecf2a1baa407655f583adee6a
                      kubernetes.io/config.seen: 2024-04-09T09:57:09.539433271Z
                      kubernetes.io/config.source: file
Status:               Running
IP:                   10.0.0.118
IPs:
  IP:           10.0.0.118
Controlled By:  Node/instance-20240409-1035
Containers:
  kube-scheduler:
    Container ID:  containerd://4ebba71736ed3ebdfc6ffd40c714e33b6c106bd81e0c16948bba001e754b388a
    Image:         registry.k8s.io/kube-scheduler:v1.25.6
    Image ID:      sealos.hub:5000/kube-scheduler@sha256:2290d4128c2ffc0d5e8b3f21312703f02e4f94e46666e8c8143bbacb21c35269
    Port:          <none>
    Host Port:     <none>
    Command:
      kube-scheduler
      --authentication-kubeconfig=/etc/kubernetes/scheduler.conf
      --authorization-kubeconfig=/etc/kubernetes/scheduler.conf
      --bind-address=0.0.0.0
      --feature-gates=EphemeralContainers=true
      --kubeconfig=/etc/kubernetes/scheduler.conf
      --leader-elect=true
    State:          Running
      Started:      Tue, 09 Apr 2024 10:00:05 +0000
    Last State:     Terminated
      Reason:       Error
      Exit Code:    1
      Started:      Tue, 09 Apr 2024 09:57:01 +0000
      Finished:     Tue, 09 Apr 2024 09:59:47 +0000
    Ready:          True
    Restart Count:  1
    Requests:
      cpu:        100m
    Liveness:     http-get https://:10259/healthz delay=10s timeout=15s period=10s #success=1 #failure=8
    Startup:      http-get https://:10259/healthz delay=10s timeout=15s period=10s #success=1 #failure=24
    Environment:  <none>
    Mounts:
      /etc/kubernetes/scheduler.conf from kubeconfig (ro)
      /etc/localtime from localtime (ro)
Conditions:
  Type              Status
  Initialized       True 
  Ready             True 
  ContainersReady   True 
  PodScheduled      True 
Volumes:
  kubeconfig:
    Type:          HostPath (bare host directory volume)
    Path:          /etc/kubernetes/scheduler.conf
    HostPathType:  FileOrCreate
  localtime:
    Type:          HostPath (bare host directory volume)
    Path:          /etc/localtime
    HostPathType:  File
QoS Class:         Burstable
Node-Selectors:    <none>
Tolerations:       :NoExecute op=Exists
Events:
  Type     Reason     Age                From     Message
  ----     ------     ----               ----     -------
  Warning  Unhealthy  23m (x5 over 24m)  kubelet  Liveness probe failed: Get "https://10.0.0.118:10259/healthz": net/http: TLS handshake timeout
  Warning  Unhealthy  23m                kubelet  Liveness probe failed: Get "https://10.0.0.118:10259/healthz": read tcp 10.0.0.118:9190->10.0.0.118:10259: read: connection reset by peer
  Warning  Unhealthy  23m                kubelet  Liveness probe failed: Get "https://10.0.0.118:10259/healthz": dial tcp 10.0.0.118:10259: connect: connection refused
  Normal   Pulled     23m                kubelet  Container image "registry.k8s.io/kube-scheduler:v1.25.6" already present on machine
  Normal   Created    23m                kubelet  Created container kube-scheduler
  Normal   Started    23m                kubelet  Started container kube-scheduler

Name:                 metrics-server-755f77d4d5-jm4lr
Namespace:            kube-system
Priority:             2000000000
Priority Class Name:  system-cluster-critical
Service Account:      metrics-server
Node:                 instance-20240409-1035/10.0.0.118
Start Time:           Tue, 09 Apr 2024 09:58:26 +0000
Labels:               app.kubernetes.io/instance=metrics-server
                      app.kubernetes.io/name=metrics-server
                      pod-template-hash=755f77d4d5
Annotations:          <none>
Status:               Running
IP:                   10.0.0.46
IPs:
  IP:           10.0.0.46
Controlled By:  ReplicaSet/metrics-server-755f77d4d5
Containers:
  metrics-server:
    Container ID:  containerd://7a11751ddbeeaabe845d1de3c47619762743d7f959e0172d23747122464132fb
    Image:         registry.k8s.io/metrics-server/metrics-server:v0.6.4
    Image ID:      sealos.hub:5000/metrics-server/metrics-server@sha256:401a4b9796f3f80c1f03d22cd7b1a26839f515a36032ef49b682c237e5848ab3
    Port:          10250/TCP
    Host Port:     0/TCP
    Args:
      --secure-port=10250
      --cert-dir=/tmp
      --kubelet-preferred-address-types=InternalIP,ExternalIP,Hostname
      --kubelet-use-node-status-port
      --metric-resolution=15s
      --kubelet-insecure-tls=true
      --kubelet-preferred-address-types=InternalIP
    State:          Waiting
      Reason:       CrashLoopBackOff
    Last State:     Terminated
      Reason:       Error
      Exit Code:    2
      Started:      Tue, 09 Apr 2024 10:21:37 +0000
      Finished:     Tue, 09 Apr 2024 10:21:38 +0000
    Ready:          False
    Restart Count:  9
    Requests:
      cpu:        100m
      memory:     200Mi
    Liveness:     http-get https://:https/livez delay=0s timeout=1s period=10s #success=1 #failure=3
    Readiness:    http-get https://:https/readyz delay=20s timeout=1s period=10s #success=1 #failure=3
    Environment:  <none>
    Mounts:
      /tmp from tmp (rw)
      /var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-nvjpr (ro)
Conditions:
  Type              Status
  Initialized       True 
  Ready             False 
  ContainersReady   False 
  PodScheduled      True 
Volumes:
  tmp:
    Type:       EmptyDir (a temporary directory that shares a pod's lifetime)
    Medium:     
    SizeLimit:  <unset>
  kube-api-access-nvjpr:
    Type:                    Projected (a volume that contains injected data from multiple sources)
    TokenExpirationSeconds:  3607
    ConfigMapName:           kube-root-ca.crt
    ConfigMapOptional:       <nil>
    DownwardAPI:             true
QoS Class:                   Burstable
Node-Selectors:              <none>
Tolerations:                 node.kubernetes.io/not-ready:NoExecute op=Exists for 300s
                             node.kubernetes.io/unreachable:NoExecute op=Exists for 300s
Events:
  Type     Reason     Age                   From               Message
  ----     ------     ----                  ----               -------
  Normal   Scheduled  25m                   default-scheduler  Successfully assigned kube-system/metrics-server-755f77d4d5-jm4lr to instance-20240409-1035
  Normal   Pulling    25m                   kubelet            Pulling image "registry.k8s.io/metrics-server/metrics-server:v0.6.4"
  Normal   Pulled     23m                   kubelet            Successfully pulled image "registry.k8s.io/metrics-server/metrics-server:v0.6.4" in 9.687327685s (1m48.296207273s including waiting)
  Warning  Unhealthy  23m                   kubelet            Liveness probe failed: Get "https://10.0.0.46:10250/livez": dial tcp 10.0.0.46:10250: connect: connection refused
  Normal   Started    22m (x4 over 23m)     kubelet            Started container metrics-server
  Normal   Created    21m (x5 over 23m)     kubelet            Created container metrics-server
  Normal   Pulled     21m (x4 over 23m)     kubelet            Container image "registry.k8s.io/metrics-server/metrics-server:v0.6.4" already present on machine
  Warning  BackOff    4m58s (x92 over 23m)  kubelet            Back-off restarting failed container

Name:             openebs-localpv-provisioner-79f4c678cd-9qc69
Namespace:        openebs
Priority:         0
Service Account:  openebs
Node:             instance-20240409-1035/10.0.0.118
Start Time:       Tue, 09 Apr 2024 09:58:22 +0000
Labels:           app=openebs
                  component=localpv-provisioner
                  name=openebs-localpv-provisioner
                  openebs.io/component-name=openebs-localpv-provisioner
                  openebs.io/version=3.4.0
                  pod-template-hash=79f4c678cd
                  release=openebs
Annotations:      <none>
Status:           Running
IP:               10.0.0.172
IPs:
  IP:           10.0.0.172
Controlled By:  ReplicaSet/openebs-localpv-provisioner-79f4c678cd
Containers:
  openebs-localpv-provisioner:
    Container ID:  containerd://9a5f7260419da747578ee2a96dcb0987828e04f7a89be8ad6b36662923b1c187
    Image:         openebs/provisioner-localpv:3.4.0
    Image ID:      sealos.hub:5000/openebs/provisioner-localpv@sha256:ac6da48d6f93933f735e86567d9d857349e4fb17e809afba39ebf9d5d9eb66d0
    Port:          <none>
    Host Port:     <none>
    Args:
      --bd-time-out=$(BDC_BD_BIND_RETRIES)
    State:          Waiting
      Reason:       CrashLoopBackOff
    Last State:     Terminated
      Reason:       Error
      Exit Code:    1
      Started:      Tue, 09 Apr 2024 10:20:55 +0000
      Finished:     Tue, 09 Apr 2024 10:20:55 +0000
    Ready:          False
    Restart Count:  9
    Liveness:       exec [sh -c test `pgrep -c "^provisioner-loc.*"` = 1] delay=30s timeout=1s period=60s #success=1 #failure=3
    Environment:
      BDC_BD_BIND_RETRIES:          12
      OPENEBS_NAMESPACE:            openebs
      NODE_NAME:                     (v1:spec.nodeName)
      OPENEBS_SERVICE_ACCOUNT:       (v1:spec.serviceAccountName)
      OPENEBS_IO_ENABLE_ANALYTICS:  true
      OPENEBS_IO_BASE_PATH:         /var/openebs/local
      OPENEBS_IO_HELPER_IMAGE:      openebs/linux-utils:3.4.0
      OPENEBS_IO_INSTALLER_TYPE:    charts-helm
      LEADER_ELECTION_ENABLED:      true
    Mounts:
      /var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-hs2l9 (ro)
Conditions:
  Type              Status
  Initialized       True 
  Ready             False 
  ContainersReady   False 
  PodScheduled      True 
Volumes:
  kube-api-access-hs2l9:
    Type:                    Projected (a volume that contains injected data from multiple sources)
    TokenExpirationSeconds:  3607
    ConfigMapName:           kube-root-ca.crt
    ConfigMapOptional:       <nil>
    DownwardAPI:             true
QoS Class:                   BestEffort
Node-Selectors:              <none>
Tolerations:                 node.kubernetes.io/not-ready:NoExecute op=Exists for 300s
                             node.kubernetes.io/unreachable:NoExecute op=Exists for 300s
Events:
  Type     Reason     Age                  From               Message
  ----     ------     ----                 ----               -------
  Normal   Scheduled  25m                  default-scheduler  Successfully assigned openebs/openebs-localpv-provisioner-79f4c678cd-9qc69 to instance-20240409-1035
  Normal   Pulling    25m                  kubelet            Pulling image "openebs/provisioner-localpv:3.4.0"
  Normal   Pulled     23m                  kubelet            Successfully pulled image "openebs/provisioner-localpv:3.4.0" in 1m38.32376667s (1m41.541746387s including waiting)
  Normal   Created    21m (x5 over 23m)    kubelet            Created container openebs-localpv-provisioner
  Normal   Started    21m (x5 over 23m)    kubelet            Started container openebs-localpv-provisioner
  Normal   Pulled     21m (x4 over 23m)    kubelet            Container image "openebs/provisioner-localpv:3.4.0" already present on machine
  Warning  BackOff    5m2s (x88 over 23m)  kubelet            Back-off restarting failed container
willzhang commented 5 months ago
binaryYuki commented 5 months ago

Hi,

感谢您的解答!

我目前只发现这个问题会出现在尝试将 sealos 安装在 aarch 架构的 ubuntu 上 我尝试解决一下

谢谢!

sealos-ci-robot commented 5 months ago

Bot detected the issue body's language is not English, translate it automatically. 👯👭🏻🧑‍🤝‍🧑👫🧑🏿‍🤝‍🧑🏻👩🏾‍🤝‍👨🏿👬🏿


Hi,

Thank you for your answer!

I only found that this problem will occur when trying to install sealos on aarch architecture ubuntu. I will try to solve it.

Thanks!