unchama / kube-cluster-on-proxmox

Proxmox環境でサクッと作ってサクっと壊せる高可用性なkubernetesクラスタを作ってみる
MIT License
51 stars 4 forks source link

kubeadm init直後だとkube-prometheus-stackは入らない #5

Closed unchama closed 2 years ago

unchama commented 2 years ago

workerが生えてないとダメそう

https://github.com/unchama/kube-cluster-on-proxmox/blob/a0fbd0a2005f8d3cd6791fee13c52355d79e1d21/scripts/k8s-node-setup.sh#L317-L321

cloudinit@unc-k8s-cp-1:~$ helm install prometheus prometheus-community/kube-prometheus-stack     --create-namespace     --namespace monitoring --debug
install.go:178: [debug] Original chart version: ""
install.go:195: [debug] CHART PATH: /home/cloudinit/.cache/helm/repository/kube-prometheus-stack-35.3.0.tgz

client.go:128: [debug] creating 1 resource(s)
install.go:151: [debug] CRD alertmanagerconfigs.monitoring.coreos.com is already present. Skipping.
client.go:128: [debug] creating 1 resource(s)
install.go:151: [debug] CRD alertmanagers.monitoring.coreos.com is already present. Skipping.
client.go:128: [debug] creating 1 resource(s)
install.go:151: [debug] CRD podmonitors.monitoring.coreos.com is already present. Skipping.
client.go:128: [debug] creating 1 resource(s)
install.go:151: [debug] CRD probes.monitoring.coreos.com is already present. Skipping.
client.go:128: [debug] creating 1 resource(s)
install.go:151: [debug] CRD prometheuses.monitoring.coreos.com is already present. Skipping.
client.go:128: [debug] creating 1 resource(s)
install.go:151: [debug] CRD prometheusrules.monitoring.coreos.com is already present. Skipping.
client.go:128: [debug] creating 1 resource(s)
install.go:151: [debug] CRD servicemonitors.monitoring.coreos.com is already present. Skipping.
client.go:128: [debug] creating 1 resource(s)
install.go:151: [debug] CRD thanosrulers.monitoring.coreos.com is already present. Skipping.
client.go:128: [debug] creating 1 resource(s)
client.go:310: [debug] Starting delete for "prometheus-kube-prometheus-admission" ServiceAccount
client.go:128: [debug] creating 1 resource(s)
client.go:310: [debug] Starting delete for "prometheus-kube-prometheus-admission" ClusterRole
client.go:128: [debug] creating 1 resource(s)
client.go:310: [debug] Starting delete for "prometheus-kube-prometheus-admission" ClusterRoleBinding
client.go:128: [debug] creating 1 resource(s)
client.go:310: [debug] Starting delete for "prometheus-kube-prometheus-admission" Role
client.go:128: [debug] creating 1 resource(s)
client.go:310: [debug] Starting delete for "prometheus-kube-prometheus-admission" RoleBinding
client.go:128: [debug] creating 1 resource(s)
client.go:310: [debug] Starting delete for "prometheus-kube-prometheus-admission-create" Job
client.go:128: [debug] creating 1 resource(s)
client.go:540: [debug] Watching for changes to Job prometheus-kube-prometheus-admission-create with timeout of 5m0s
client.go:568: [debug] Add/Modify event for prometheus-kube-prometheus-admission-create: ADDED
client.go:607: [debug] prometheus-kube-prometheus-admission-create: Jobs active: 0, jobs failed: 0, jobs succeeded: 0
client.go:568: [debug] Add/Modify event for prometheus-kube-prometheus-admission-create: MODIFIED
client.go:607: [debug] prometheus-kube-prometheus-admission-create: Jobs active: 1, jobs failed: 0, jobs succeeded: 0
Error: INSTALLATION FAILED: failed pre-install: timed out waiting for the condition
helm.go:84: [debug] failed pre-install: timed out waiting for the condition
INSTALLATION FAILED
main.newInstallCmd.func2
        helm.sh/helm/v3/cmd/helm/install.go:127
github.com/spf13/cobra.(*Command).execute
        github.com/spf13/cobra@v1.4.0/command.go:856
github.com/spf13/cobra.(*Command).ExecuteC
        github.com/spf13/cobra@v1.4.0/command.go:974
github.com/spf13/cobra.(*Command).Execute
        github.com/spf13/cobra@v1.4.0/command.go:902
main.main
        helm.sh/helm/v3/cmd/helm/helm.go:83
runtime.main
        runtime/proc.go:255
runtime.goexit
        runtime/asm_amd64.s:1581
cloudinit@unc-k8s-cp-1:~$
cloudinit@unc-k8s-cp-1:~$ kubectl get po -n monitoring
NAME                                                READY   STATUS    RESTARTS   AGE
prometheus-kube-prometheus-admission-create-9w72t   0/1     Pending   0          2m31s
cloudinit@unc-k8s-cp-1:~$ kubectl describe po prometheus-kube-prometheus-admission-create-9w72t -n monitoring
Name:           prometheus-kube-prometheus-admission-create-9w72t
Namespace:      monitoring
Priority:       0
Node:           <none>
Labels:         app=kube-prometheus-stack-admission-create
                app.kubernetes.io/instance=prometheus
                app.kubernetes.io/managed-by=Helm
                app.kubernetes.io/part-of=kube-prometheus-stack
                app.kubernetes.io/version=35.3.0
                chart=kube-prometheus-stack-35.3.0
                controller-uid=ce63181c-cf39-4924-b9f6-558f207f65ae
                heritage=Helm
                job-name=prometheus-kube-prometheus-admission-create
                release=prometheus
Annotations:    <none>
Status:         Pending
IP:
IPs:            <none>
Controlled By:  Job/prometheus-kube-prometheus-admission-create
Containers:
  create:
    Image:      k8s.gcr.io/ingress-nginx/kube-webhook-certgen:v1.1.1
    Port:       <none>
    Host Port:  <none>
    Args:
      create
      --host=prometheus-kube-prometheus-operator,prometheus-kube-prometheus-operator.monitoring.svc
      --namespace=monitoring
      --secret-name=prometheus-kube-prometheus-admission
    Environment:  <none>
    Mounts:
      /var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-ptjxl (ro)
Conditions:
  Type           Status
  PodScheduled   False
Volumes:
  kube-api-access-ptjxl:
    Type:                    Projected (a volume that contains injected data from multiple sources)
    TokenExpirationSeconds:  3607
    ConfigMapName:           kube-root-ca.crt
    ConfigMapOptional:       <nil>
    DownwardAPI:             true
QoS Class:                   BestEffort
Node-Selectors:              <none>
Tolerations:                 node.kubernetes.io/not-ready:NoExecute op=Exists for 300s
                             node.kubernetes.io/unreachable:NoExecute op=Exists for 300s
Events:
  Type     Reason            Age    From               Message
  ----     ------            ----   ----               -------
  Warning  FailedScheduling  2m40s  default-scheduler  0/1 nodes are available: 1 node(s) had untolerated taint {node-role.kubernetes.io/master: }. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.
cloudinit@unc-k8s-cp-1:~$
unchama commented 2 years ago

c5d2916