cloudinit@unc-k8s-cp-1:~$ helm install prometheus prometheus-community/kube-prometheus-stack --create-namespace --namespace monitoring --debug
install.go:178: [debug] Original chart version: ""
install.go:195: [debug] CHART PATH: /home/cloudinit/.cache/helm/repository/kube-prometheus-stack-35.3.0.tgz
client.go:128: [debug] creating 1 resource(s)
install.go:151: [debug] CRD alertmanagerconfigs.monitoring.coreos.com is already present. Skipping.
client.go:128: [debug] creating 1 resource(s)
install.go:151: [debug] CRD alertmanagers.monitoring.coreos.com is already present. Skipping.
client.go:128: [debug] creating 1 resource(s)
install.go:151: [debug] CRD podmonitors.monitoring.coreos.com is already present. Skipping.
client.go:128: [debug] creating 1 resource(s)
install.go:151: [debug] CRD probes.monitoring.coreos.com is already present. Skipping.
client.go:128: [debug] creating 1 resource(s)
install.go:151: [debug] CRD prometheuses.monitoring.coreos.com is already present. Skipping.
client.go:128: [debug] creating 1 resource(s)
install.go:151: [debug] CRD prometheusrules.monitoring.coreos.com is already present. Skipping.
client.go:128: [debug] creating 1 resource(s)
install.go:151: [debug] CRD servicemonitors.monitoring.coreos.com is already present. Skipping.
client.go:128: [debug] creating 1 resource(s)
install.go:151: [debug] CRD thanosrulers.monitoring.coreos.com is already present. Skipping.
client.go:128: [debug] creating 1 resource(s)
client.go:310: [debug] Starting delete for "prometheus-kube-prometheus-admission" ServiceAccount
client.go:128: [debug] creating 1 resource(s)
client.go:310: [debug] Starting delete for "prometheus-kube-prometheus-admission" ClusterRole
client.go:128: [debug] creating 1 resource(s)
client.go:310: [debug] Starting delete for "prometheus-kube-prometheus-admission" ClusterRoleBinding
client.go:128: [debug] creating 1 resource(s)
client.go:310: [debug] Starting delete for "prometheus-kube-prometheus-admission" Role
client.go:128: [debug] creating 1 resource(s)
client.go:310: [debug] Starting delete for "prometheus-kube-prometheus-admission" RoleBinding
client.go:128: [debug] creating 1 resource(s)
client.go:310: [debug] Starting delete for "prometheus-kube-prometheus-admission-create" Job
client.go:128: [debug] creating 1 resource(s)
client.go:540: [debug] Watching for changes to Job prometheus-kube-prometheus-admission-create with timeout of 5m0s
client.go:568: [debug] Add/Modify event for prometheus-kube-prometheus-admission-create: ADDED
client.go:607: [debug] prometheus-kube-prometheus-admission-create: Jobs active: 0, jobs failed: 0, jobs succeeded: 0
client.go:568: [debug] Add/Modify event for prometheus-kube-prometheus-admission-create: MODIFIED
client.go:607: [debug] prometheus-kube-prometheus-admission-create: Jobs active: 1, jobs failed: 0, jobs succeeded: 0
Error: INSTALLATION FAILED: failed pre-install: timed out waiting for the condition
helm.go:84: [debug] failed pre-install: timed out waiting for the condition
INSTALLATION FAILED
main.newInstallCmd.func2
helm.sh/helm/v3/cmd/helm/install.go:127
github.com/spf13/cobra.(*Command).execute
github.com/spf13/cobra@v1.4.0/command.go:856
github.com/spf13/cobra.(*Command).ExecuteC
github.com/spf13/cobra@v1.4.0/command.go:974
github.com/spf13/cobra.(*Command).Execute
github.com/spf13/cobra@v1.4.0/command.go:902
main.main
helm.sh/helm/v3/cmd/helm/helm.go:83
runtime.main
runtime/proc.go:255
runtime.goexit
runtime/asm_amd64.s:1581
cloudinit@unc-k8s-cp-1:~$
cloudinit@unc-k8s-cp-1:~$ kubectl get po -n monitoring
NAME READY STATUS RESTARTS AGE
prometheus-kube-prometheus-admission-create-9w72t 0/1 Pending 0 2m31s
cloudinit@unc-k8s-cp-1:~$ kubectl describe po prometheus-kube-prometheus-admission-create-9w72t -n monitoring
Name: prometheus-kube-prometheus-admission-create-9w72t
Namespace: monitoring
Priority: 0
Node: <none>
Labels: app=kube-prometheus-stack-admission-create
app.kubernetes.io/instance=prometheus
app.kubernetes.io/managed-by=Helm
app.kubernetes.io/part-of=kube-prometheus-stack
app.kubernetes.io/version=35.3.0
chart=kube-prometheus-stack-35.3.0
controller-uid=ce63181c-cf39-4924-b9f6-558f207f65ae
heritage=Helm
job-name=prometheus-kube-prometheus-admission-create
release=prometheus
Annotations: <none>
Status: Pending
IP:
IPs: <none>
Controlled By: Job/prometheus-kube-prometheus-admission-create
Containers:
create:
Image: k8s.gcr.io/ingress-nginx/kube-webhook-certgen:v1.1.1
Port: <none>
Host Port: <none>
Args:
create
--host=prometheus-kube-prometheus-operator,prometheus-kube-prometheus-operator.monitoring.svc
--namespace=monitoring
--secret-name=prometheus-kube-prometheus-admission
Environment: <none>
Mounts:
/var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-ptjxl (ro)
Conditions:
Type Status
PodScheduled False
Volumes:
kube-api-access-ptjxl:
Type: Projected (a volume that contains injected data from multiple sources)
TokenExpirationSeconds: 3607
ConfigMapName: kube-root-ca.crt
ConfigMapOptional: <nil>
DownwardAPI: true
QoS Class: BestEffort
Node-Selectors: <none>
Tolerations: node.kubernetes.io/not-ready:NoExecute op=Exists for 300s
node.kubernetes.io/unreachable:NoExecute op=Exists for 300s
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Warning FailedScheduling 2m40s default-scheduler 0/1 nodes are available: 1 node(s) had untolerated taint {node-role.kubernetes.io/master: }. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.
cloudinit@unc-k8s-cp-1:~$
workerが生えてないとダメそう
https://github.com/unchama/kube-cluster-on-proxmox/blob/a0fbd0a2005f8d3cd6791fee13c52355d79e1d21/scripts/k8s-node-setup.sh#L317-L321