Open andy108369 opened 6 days ago
The cpu
value returned back to normal even without the need to bump the opreator-inventory, after deleting containers stuck in "ContainerStatusUnknown" state:
$ provider_info2.sh provider.hurricane.akash.pub
PROVIDER INFO
BALANCE: 408.364243
^R
"hostname" "address"
"provider.hurricane.akash.pub" "akash15tl6v6gd0nte0syyxnv57zmmspgju4c3xfmdhk"
Total/Available/Used (t/a/u) per node:
"name" "cpu(t/a/u)" "gpu(t/a/u)" "mem(t/a/u GiB)" "ephemeral(t/a/u GiB)"
"control-01.hurricane2" "2/1.2/0.8" "0/0/0" "1.82/1.69/0.13" "25.54/25.54/0"
"worker-01.hurricane2" "102/18446744073709490/-18446744073709384" "1/1/0" "196.45/49.67/146.78" "1808.76/1435.28/373.48"
ACTIVE TOTAL:
"cpu(cores)" "gpu" "mem(GiB)" "ephemeral(GiB)" "beta1(GiB)" "beta2(GiB)" "beta3(GiB)"
34.2 0 64.88 314.4 0 0 11
PERSISTENT STORAGE:
"storage class" "available space(GiB)"
"beta3" 575.7
PENDING TOTAL:
"cpu(cores)" "gpu" "mem(GiB)" "ephemeral(GiB)" "beta1(GiB)" "beta2(GiB)" "beta3(GiB)"
arno@x1:~$ kubectl get pods -A --field-selector status.phase=Failed
NAMESPACE NAME READY STATUS RESTARTS AGE
2kmrpu3u3u5psi5hip2oqt0bajako0e9o9phc5j5c5dmu web-5df9f7c798-bbgqz 0/1 ContainerStatusUnknown 1 2d22h
2kmrpu3u3u5psi5hip2oqt0bajako0e9o9phc5j5c5dmu web-5df9f7c798-f2fpj 0/1 ContainerStatusUnknown 1 3d20h
2kmrpu3u3u5psi5hip2oqt0bajako0e9o9phc5j5c5dmu web-5df9f7c798-g7xbd 0/1 ContainerStatusUnknown 1 3d3h
2kmrpu3u3u5psi5hip2oqt0bajako0e9o9phc5j5c5dmu web-5df9f7c798-hv4qs 0/1 ContainerStatusUnknown 1 9h
2kmrpu3u3u5psi5hip2oqt0bajako0e9o9phc5j5c5dmu web-5df9f7c798-p4h7j 0/1 ContainerStatusUnknown 1 4d7h
2kmrpu3u3u5psi5hip2oqt0bajako0e9o9phc5j5c5dmu web-5df9f7c798-rcr45 0/1 ContainerStatusUnknown 1 30h
2kmrpu3u3u5psi5hip2oqt0bajako0e9o9phc5j5c5dmu web-7f5fdfd87c-4cq86 0/1 ContainerStatusUnknown 1 20d
2kmrpu3u3u5psi5hip2oqt0bajako0e9o9phc5j5c5dmu web-7f5fdfd87c-5ddrg 0/1 ContainerStatusUnknown 1 13d
2kmrpu3u3u5psi5hip2oqt0bajako0e9o9phc5j5c5dmu web-7f5fdfd87c-7nl6p 0/1 ContainerStatusUnknown 1 5d7h
2kmrpu3u3u5psi5hip2oqt0bajako0e9o9phc5j5c5dmu web-7f5fdfd87c-9jsn7 0/1 ContainerStatusUnknown 1 19d
2kmrpu3u3u5psi5hip2oqt0bajako0e9o9phc5j5c5dmu web-7f5fdfd87c-bnjfh 0/1 ContainerStatusUnknown 1 20d
2kmrpu3u3u5psi5hip2oqt0bajako0e9o9phc5j5c5dmu web-7f5fdfd87c-d2nfr 0/1 ContainerStatusUnknown 1 7d12h
2kmrpu3u3u5psi5hip2oqt0bajako0e9o9phc5j5c5dmu web-7f5fdfd87c-dk95v 0/1 ContainerStatusUnknown 1 17d
2kmrpu3u3u5psi5hip2oqt0bajako0e9o9phc5j5c5dmu web-7f5fdfd87c-fgfl4 0/1 ContainerStatusUnknown 1 7d19h
2kmrpu3u3u5psi5hip2oqt0bajako0e9o9phc5j5c5dmu web-7f5fdfd87c-gh9bb 0/1 ContainerStatusUnknown 1 16d
2kmrpu3u3u5psi5hip2oqt0bajako0e9o9phc5j5c5dmu web-7f5fdfd87c-gltgh 0/1 ContainerStatusUnknown 1 9d
2kmrpu3u3u5psi5hip2oqt0bajako0e9o9phc5j5c5dmu web-7f5fdfd87c-j9tnr 0/1 ContainerStatusUnknown 1 15d
2kmrpu3u3u5psi5hip2oqt0bajako0e9o9phc5j5c5dmu web-7f5fdfd87c-mmqfk 0/1 ContainerStatusUnknown 1 6d5h
2kmrpu3u3u5psi5hip2oqt0bajako0e9o9phc5j5c5dmu web-7f5fdfd87c-ph89h 0/1 ContainerStatusUnknown 1 11d
2kmrpu3u3u5psi5hip2oqt0bajako0e9o9phc5j5c5dmu web-7f5fdfd87c-pjrg4 0/1 ContainerStatusUnknown 1 17d
2kmrpu3u3u5psi5hip2oqt0bajako0e9o9phc5j5c5dmu web-7f5fdfd87c-pwbzv 0/1 ContainerStatusUnknown 1 13d
2kmrpu3u3u5psi5hip2oqt0bajako0e9o9phc5j5c5dmu web-7f5fdfd87c-rd7z5 0/1 ContainerStatusUnknown 1 12d
2kmrpu3u3u5psi5hip2oqt0bajako0e9o9phc5j5c5dmu web-7f5fdfd87c-t6vt9 0/1 ContainerStatusUnknown 1 6d15h
2kmrpu3u3u5psi5hip2oqt0bajako0e9o9phc5j5c5dmu web-7f5fdfd87c-vht5l 0/1 ContainerStatusUnknown 1 9d
2kmrpu3u3u5psi5hip2oqt0bajako0e9o9phc5j5c5dmu web-7f5fdfd87c-wd8w4 0/1 ContainerStatusUnknown 1 7d23h
2kmrpu3u3u5psi5hip2oqt0bajako0e9o9phc5j5c5dmu web-7f5fdfd87c-xnsvt 0/1 ContainerStatusUnknown 1 13d
2kmrpu3u3u5psi5hip2oqt0bajako0e9o9phc5j5c5dmu web-7f5fdfd87c-zmzbf 0/1 ContainerStatusUnknown 1 12d
2kmrpu3u3u5psi5hip2oqt0bajako0e9o9phc5j5c5dmu web-7f5fdfd87c-zw2st 0/1 ContainerStatusUnknown 1 10d
arno@x1:~$ ns=2kmrpu3u3u5psi5hip2oqt0bajako0e9o9phc5j5c5dmu
arno@x1:~$ kubectl -n $ns get deployment
NAME READY UP-TO-DATE AVAILABLE AGE
web 1/1 1 1 116d
arno@x1:~$ kubectl -n $ns get rs
NAME DESIRED CURRENT READY AGE
web-57478ff56c 0 0 0 4d17h
web-5df9f7c798 1 1 1 4d17h
web-7f5fdfd87c 0 0 0 53d
web-85fc6b7694 0 0 0 54d
web-85ff75fdc5 0 0 0 70d
arno@x1:~$ kubectl -n $ns delete rs web-85ff75fdc5
replicaset.apps "web-85ff75fdc5" deleted
arno@x1:~$ kubectl -n $ns delete rs web-85fc6b7694
replicaset.apps "web-85fc6b7694" deleted
arno@x1:~$ kubectl -n $ns delete rs web-7f5fdfd87c
replicaset.apps "web-7f5fdfd87c" deleted
arno@x1:~$ kubectl -n $ns delete rs web-57478ff56c
replicaset.apps "web-57478ff56c" deleted
arno@x1:~$ kubectl -n $ns get rs
NAME DESIRED CURRENT READY AGE
web-5df9f7c798 1 1 1 4d17h
arno@x1:~$ kubectl get pods -A --field-selector status.phase=Failed
NAMESPACE NAME READY STATUS RESTARTS AGE
2kmrpu3u3u5psi5hip2oqt0bajako0e9o9phc5j5c5dmu web-5df9f7c798-bbgqz 0/1 ContainerStatusUnknown 1 2d22h
2kmrpu3u3u5psi5hip2oqt0bajako0e9o9phc5j5c5dmu web-5df9f7c798-f2fpj 0/1 ContainerStatusUnknown 1 3d20h
2kmrpu3u3u5psi5hip2oqt0bajako0e9o9phc5j5c5dmu web-5df9f7c798-g7xbd 0/1 ContainerStatusUnknown 1 3d3h
2kmrpu3u3u5psi5hip2oqt0bajako0e9o9phc5j5c5dmu web-5df9f7c798-hv4qs 0/1 ContainerStatusUnknown 1 9h
2kmrpu3u3u5psi5hip2oqt0bajako0e9o9phc5j5c5dmu web-5df9f7c798-p4h7j 0/1 ContainerStatusUnknown 1 4d7h
2kmrpu3u3u5psi5hip2oqt0bajako0e9o9phc5j5c5dmu web-5df9f7c798-rcr45 0/1 ContainerStatusUnknown 1 30h
arno@x1:~$ kubectl delete pods -A --field-selector status.phase=Failed
pod "web-5df9f7c798-bbgqz" deleted
pod "web-5df9f7c798-f2fpj" deleted
pod "web-5df9f7c798-g7xbd" deleted
pod "web-5df9f7c798-hv4qs" deleted
pod "web-5df9f7c798-p4h7j" deleted
pod "web-5df9f7c798-rcr45" deleted
$ provider_info2.sh provider.hurricane.akash.pub
PROVIDER INFO
BALANCE: 408.364243
"hostname" "address"
"provider.hurricane.akash.pub" "akash15tl6v6gd0nte0syyxnv57zmmspgju4c3xfmdhk"
Total/Available/Used (t/a/u) per node:
"name" "cpu(t/a/u)" "gpu(t/a/u)" "mem(t/a/u GiB)" "ephemeral(t/a/u GiB)"
"control-01.hurricane2" "2/1.2/0.8" "0/0/0" "1.82/1.69/0.13" "25.54/25.54/0"
"worker-01.hurricane2" "102/47.995/54.005" "1/1/0" "196.45/104.36/92.09" "1808.76/1489.97/318.79"
ACTIVE TOTAL:
"cpu(cores)" "gpu" "mem(GiB)" "ephemeral(GiB)" "beta1(GiB)" "beta2(GiB)" "beta3(GiB)"
34.2 0 64.88 314.4 0 0 11
PERSISTENT STORAGE:
"storage class" "available space(GiB)"
"beta3" 575.7
PENDING TOTAL:
"cpu(cores)" "gpu" "mem(GiB)" "ephemeral(GiB)" "beta1(GiB)" "beta2(GiB)" "beta3(GiB)"
hurricane provider reports excessively large amount of available CPUs
Versions
Logs
I've tried restarting the operator-inventory which previously used to "fix" this issue, but to no avail this time.