Closed Pothulapati closed 2 years ago
So, Digging more on the specific logs. It seems more to be a issue around fedora
and filesystem mounts inside kuberentes. Similar issues have been reported with other minimal kubernetes tools as per https://github.com/kubernetes-sigs/kind/issues/2411
The work around here seems to be to manually mounting that into the container. I'll update the specific user on the same, and see from there!
Closing this as its not a specific GItpod issue and only happens with fedora
and a filesystem configuration that already seems to have a fix.
Feel free to re-open if more users fall into this specific issue so that we can document or do something else here.
Bug description
For users, Who are on cgroupsv2 there seems to be an issue during the startup of
k3s
preventing thelocal-preview
to workLogs are as below
``` + REQUIRED_CORES=4 + nproc + total_cores=24 + '[' 24 -lt 4 ] + echo 'Gitpod Domain: preview.gitpod-self-hosted.com' Gitpod Domain: preview.gitpod-self-hosted.com + '[' -f /sys/fs/cgroup/cgroup.controllers ] + date -Iseconds + echo '[2022-08-12T08:16:03+00:00] [CgroupV2 Fix] Evacuating Root Cgroup ...' [2022-08-12T08:16:03+00:00] [CgroupV2 Fix] Evacuating Root Cgroup ... + mkdir -p /sys/fs/cgroup/init + busybox xargs -rn1 + sed -e 's/ / +/g' -e s/^/+/ + date -Iseconds + echo '[2022-08-12T08:16:03+00:00] [CgroupV2 Fix] Done' [2022-08-12T08:16:03+00:00] [CgroupV2 Fix] Done + mount --make-shared /sys/fs/cgroup + mount --make-shared /proc + mount --make-shared /var/gitpod + mkcert -install Created a new local CA đĨ Installing to the system store is not yet supported on this Linux đŖ but Firefox and/or Chrome/Chromium will still work. You can also manually install the root certificate at "/.local/share/mkcert/rootCA.pem". + cat //.local/share/mkcert/rootCA.pem + cat //.local/share/mkcert/rootCA.pem + FN_CACERT=./ca.pem + FN_SSLCERT=./ssl.crt + FN_SSLKEY=./ssl.key + cat //.local/share/mkcert/rootCA.pem + mkcert -cert-file ./ssl.crt -key-file ./ssl.key '*.ws.preview.gitpod-self-hosted.com' '*.preview.gitpod-self-hosted.com' preview.gitpod-self-hosted.com reg.preview.gitpod-self-hosted.com registry.default.svc.cluster.local gitpod.default ws-manager.default.svc ws-manager ws-manager-dev registry-facade server ws-manager-bridge ws-proxy ws-manager ws-daemon.default.svc ws-daemon wsdaemon Created a new certificate valid for the following names đ - "*.ws.preview.gitpod-self-hosted.com" - "*.preview.gitpod-self-hosted.com" - "preview.gitpod-self-hosted.com" - "reg.preview.gitpod-self-hosted.com" - "registry.default.svc.cluster.local" - "gitpod.default" - "ws-manager.default.svc" - "ws-manager" - "ws-manager-dev" - "registry-facade" - "server" - "ws-manager-bridge" - "ws-proxy" - "ws-manager" - "ws-daemon.default.svc" - "ws-daemon" - "wsdaemon" Reminder: X.509 wildcards only go one level deep, so this won't match a.b.ws.preview.gitpod-self-hosted.com âšī¸ The certificate is at "./ssl.crt" and the key at "./ssl.key" â It will expire on 12 November 2024 đ + base64 -w0 + CACERT='LS0tLS1CRUdJTiBDRVJUSUZJQ0FURS0tLS0tCk1JSUVlVENDQXVHZ0F3SUJBZ0lRQmptcFJzTFllZXVFRXJod3FiNDFJVEFOQmdrcWhraUc5dzBCQVFzRkFEQlYKTVI0d0hBWURWUVFLRXhWdGEyTmxjblFnWkdWMlpXeHZjRzFsYm5RZ1EwRXhGVEFUQmdOVkJBc1REREJsWldNNQpNbU15Wm1ZME1ERWNNQm9HQTFVRUF4TVRiV3RqWlhKMElEQmxaV001TW1NeVptWTBNREFlRncweU1qQTRNVEl3Ck9ERTJNRE5hRncwek1qQTRNVEl3T0RFMk1ETmFNRlV4SGpBY0JnTlZCQW9URlcxclkyVnlkQ0JrWlhabGJHOXcKYldWdWRDQkRRVEVWTUJNR0ExVUVDeE1NTUdWbFl6a3lZekptWmpRd01Sd3dHZ1lEVlFRREV4TnRhMk5sY25RZwpNR1ZsWXpreVl6Sm1aalF3TUlJQm9qQU5CZ2txaGtpRzl3MEJBUUVGQUFPQ0FZOEFNSUlCaWdLQ0FZRUE1bUJKClJFeVV2cm1XbTJRNmlKclQ5KzQ2YXh4ZUxuRFBReWtPVktoTVZLZDBmdGR0emNzVVNtOFVRYjJWN2JmelovQzYKajJUVXJsbVVuc0ptTFpKSjZtOHN2YUw1c1NDODRXOS9ybCtrOG9Kb01Ta3R5dnoyekNxeWtEZGFiSWppQ2wrQQpnQk1GUFlEMyt3K3d4TzcrVUNRVDA3VFBiaFBIK09yU1JmQkt0MVVyU2NLTE5iM1ZnRW9ScCsxT28vYkpWcThMClZ5dG5rMnhoV2JWQ3FKMzlkSUxVOXYzV2QzMUxLdzFNckNTc01mNGhBeUxCT3JnZTVDQ1FGTGRpMmZXZ0ZDVisKQUxwUi9VZ1BRbU51aWtVb21ZVTl4ekttcEVPbG9vYUlVb2N0VUl4Tzg1Uk1CSFk4eGFhQkNXK3R2STJlYjdyagpCOEpDaERLZVNneCtPNzhCSHBSZ3ZuT2QvTkFoZ3ZURmlkUHgzZTVkY3h6S3VNRGtrM2RJcitLNFdWaWh1TkJSCmNpY09ER21lSUlURGlHMEJ3UTdkRFRaM2xCZXR4eGFMNkJMMHVndkthWllLMlU2TFg0TnpLdmJ5N0dVN0hFR3AKOHR1eCtzaWYrSGVjUVNHK2VKZGFpdFY0Vy9xN09vcmlFWkpxSUJpZ0ZiZnFkY2tTOWhnVDQra0U5ZTl6QWdNQgpBQUdqUlRCRE1BNEdBMVVkRHdFQi93UUVBd0lDQkRBU0JnTlZIUk1CQWY4RUNEQUdBUUgvQWdFQU1CMEdBMVVkCkRnUVdCQlRPQk8rUW9KUHcyckViRG1lTHNiYUpsR1R5bGpBTkJna3Foa2lHOXcwQkFRc0ZBQU9DQVlFQVFzT1kKdUtEcXdscWdhamczSGFjTlVWYkVmU2N1T3NoZU5PZjNWRmxvTVB5SVVONVNNbXRyYzcrU3RvNGh4dk9ja2Ezdgp6WWxqeVZocE5CWEdmUWRyK1hHWFBteEVpVWVlMHREbGYvaXU3TFI3akpGY3k1dzZhbDdRMEJiQ0hWaEI1b0Z2CmFqSkNzNllOOCtGVm5RRkxndXVZWkVXVllSem1kUXdINmliV3lPTDk3Q3BTVUZ0dzlJcGZTcjl4c2t6RmthaW4KVTUva1dTcjNWRzRkblZ0SmVZa3ZKSUZ3VlNtTVkvVTM4dXJ2VTMzTFVkOXBhTVpRQ0duR1hEY0UvVkZqcDBndApLQnkwRUw3K25tdHE3cWlzckVyYzlzVnQvZ3N3Mzhua1VoUkVoakJPQTVVK2JraHZuRDZzNGMxM2g1dUQrMVMxClcreFNFR0FMUnl1U0s3OTZyNG8zWlFURkptWlpXUEphdEhBcWgwbFZBajBhVVRQNjZrUHd2ZHVKTWE5R1lwcVIKTCtjaGdydks5dFlLbUVFVjROZnBQSlJQbzhLLzFSMXJOaUFTeWp6YWduT3hSb0xQdk1qZXNraTR0em9XNnNZMwpncng4L0pYWU1hYmYxaHR0WWZpbjJ3bTNZM3FvNWNYdjllWkF3cjhRbk8zak5nRzJ2TkxhUXpBL29hUTAKLS0tLS1FTkQgQ0VSVElGSUNBVEUtLS0tLQo=' + base64 -w0 + SSLCERT=LS0tLS1CRUdJTiBDRVJUSUZJQ0FURS0tLS0tCk1JSUZYakNDQThhZ0F3SUJBZ0lSQU9BSU9sTlVTR2JOd2FuQVdhZnY1ekV3RFFZSktvWklodmNOQVFFTEJRQXcKVlRFZU1Cd0dBMVVFQ2hNVmJXdGpaWEowSUdSbGRtVnNiM0J0Wlc1MElFTkJNUlV3RXdZRFZRUUxFd3d3WldWagpPVEpqTW1abU5EQXhIREFhQmdOVkJBTVRFMjFyWTJWeWRDQXdaV1ZqT1RKak1tWm1OREF3SGhjTk1qSXdPREV5Ck1EZ3hOakF6V2hjTk1qUXhNVEV5TURneE5qQXpXakJBTVNjd0pRWURWUVFLRXg1dGEyTmxjblFnWkdWMlpXeHYKY0cxbGJuUWdZMlZ5ZEdsbWFXTmhkR1V4RlRBVEJnTlZCQXNURERCbFpXTTVNbU15Wm1ZME1EQ0NBU0l3RFFZSgpLb1pJaHZjTkFRRUJCUUFEZ2dFUEFEQ0NBUW9DZ2dFQkFMbCt0NDgzdEZadVhUZWZiU2JSbGlBVWFuSk1zUEpPCjBCMXc2bnJiSm84NVRSQXhWWUo1UTlGZ3gwb3N4aFhPdWoxNzZSLzN0WGhzRkZLcXRPQ24rQ3E5T21nOFVLeUgKYXRETEZLcG9DczZEcks5amtUWUhHSGxuZzY3dFQydzdTNlljcHJER0w2ZXZLWkI5NmVibHVYbW84Sk9oQ1VqYwpPdUVFdUxjcFR4eUpmN1ZPeVUvY2Vtb3Nvd1I3MFRwN3ZXOG9QMmRLK20vbzkyODFZWnBJcFFTVmNKMXNrN0xRClAxakR2M3pmb3JKUFNydXdBTjN3N2VueEdoYjBmVmk5Tisxdmd5NDlBdUhkWDhETUgzemtldzE0Y09IcjFTU3QKRzRNRmVVd2xQWmxGRUU2b09SNzhRQ0RRbHBnOGFScGtnUHgzTlFscXl2NmtvVFhUbEpLeUhza0NBd0VBQWFPQwpBYnd3Z2dHNE1BNEdBMVVkRHdFQi93UUVBd0lGb0RBVEJnTlZIU1VFRERBS0JnZ3JCZ0VGQlFjREFUQWZCZ05WCkhTTUVHREFXZ0JUT0JPK1FvSlB3MnJFYkRtZUxzYmFKbEdUeWxqQ0NBVzRHQTFVZEVRU0NBV1V3Z2dGaGdpTXEKTG5kekxuQnlaWFpwWlhjdVoybDBjRzlrTFhObGJHWXRhRzl6ZEdWa0xtTnZiWUlnS2k1d2NtVjJhV1YzTG1kcApkSEJ2WkMxelpXeG1MV2h2YzNSbFpDNWpiMjJDSG5CeVpYWnBaWGN1WjJsMGNHOWtMWE5sYkdZdGFHOXpkR1ZrCkxtTnZiWUlpY21WbkxuQnlaWFpwWlhjdVoybDBjRzlrTFhObGJHWXRhRzl6ZEdWa0xtTnZiWUlpY21WbmFYTjAKY25rdVpHVm1ZWFZzZEM1emRtTXVZMngxYzNSbGNpNXNiMk5oYklJT1oybDBjRzlrTG1SbFptRjFiSFNDRm5kegpMVzFoYm1GblpYSXVaR1ZtWVhWc2RDNXpkbU9DQ25kekxXMWhibUZuWlhLQ0RuZHpMVzFoYm1GblpYSXRaR1YyCmdnOXlaV2RwYzNSeWVTMW1ZV05oWkdXQ0JuTmxjblpsY29JUmQzTXRiV0Z1WVdkbGNpMWljbWxrWjJXQ0NIZHoKTFhCeWIzaDVnZ3AzY3kxdFlXNWhaMlZ5Z2hWM2N5MWtZV1Z0YjI0dVpHVm1ZWFZzZEM1emRtT0NDWGR6TFdSaApaVzF2Ym9JSWQzTmtZV1Z0YjI0d0RRWUpLb1pJaHZjTkFRRUxCUUFEZ2dHQkFGM282L05oSmtvNXNJR3BTYVQ4ClVCS1ZtR3JPcVVWeExIaDVTd3F3MUlCeDYydE1mTENTOE1OM3BCbENaeWM4UHc4aTBPaTJCVGQweFlFYjVOZ1kKNGQyZnRzbStzZ2pUcVYwbWRHU3cvTXNSNG1UQzFPZTJyemtrTS9NOTVHODZiWDRsVFBBNnFSSVpQa1MwZVo4ZApER2lrckJpaEhoRFJwOHBTeWhVUzRxZUhqdVVnQ3ZWZ3hBMkx3ZW10SDUyNUZTeFBPc0xuRE9uQXNaMG1sN1VzCmp0bDNSV05ZWXNGeFJnYmdENVNWcEs5OHM0MTdzZEVyQXJnZTRqV1Njb3NubjhWbUhVSXNGdG5wZmhUZ2w1YXoKNFo2SzlOSWVZM1VydUF3bjdsWWdxTVBQOXpuTCthamNwSHNITkNwZWJkV1VhTC8zZXdCVFNaQ3ZNTURmeTFQZAo4Nm83UUJ4OHhpbWRTOHZCTURZV0RnSHBHVU5xS2JaWDZ2UDRNQzMrUzZ6NkJZTkczVU1QUUFwd1MrMDJSMURlCnFJNGJjcUpvSGozR0hJUmVsODdsYUN5clVUOVgwaUl0SmcyRFlpcTdrTmtJQWlGU3h6d2gvRE12WWZpL1dOUzAKZlpYMnEyMUpod2xNeElzbU1SL2xaR2w2QmUrNHpacGRnd296UmpEa2oxS2ZvUT09Ci0tLS0tRU5EIENFUlRJRklDQVRFLS0tLS0K + base64 -w0 + SSLKEY=LS0tLS1CRUdJTiBQUklWQVRFIEtFWS0tLS0tCk1JSUV2Z0lCQURBTkJna3Foa2lHOXcwQkFRRUZBQVNDQktnd2dnU2tBZ0VBQW9JQkFRQzVmcmVQTjdSV2JsMDMKbjIwbTBaWWdGR3B5VExEeVR0QWRjT3A2MnlhUE9VMFFNVldDZVVQUllNZEtMTVlWenJvOWUra2Y5N1Y0YkJSUwpxclRncC9ncXZUcG9QRkNzaDJyUXl4U3FhQXJPZzZ5dlk1RTJCeGg1WjRPdTdVOXNPMHVtSEthd3hpK25yeW1RCmZlbm01Ymw1cVBDVG9RbEkzRHJoQkxpM0tVOGNpWCsxVHNsUDNIcHFMS01FZTlFNmU3MXZLRDluU3ZwdjZQZHYKTldHYVNLVUVsWENkYkpPeTBEOVl3Nzk4MzZLeVQwcTdzQURkOE8zcDhSb1c5SDFZdlRmdGI0TXVQUUxoM1YvQQp6Qjk4NUhzTmVIRGg2OVVrclJ1REJYbE1KVDJaUlJCT3FEa2UvRUFnMEphWVBHa2FaSUQ4ZHpVSmFzcitwS0UxCjA1U1NzaDdKQWdNQkFBRUNnZ0VBSy9CY1FzeUxKejRWVHF1eEMxVHlIcjgzUjhQcTFqcmRDVnhKN3JnaXRpSjQKb3JGTTlBOE5oWGRMUGNMRldUMFMyS1dWWDBFcDkxQ0NyK0pIM2o5cmhaUTFWYU9UNklwYlB3SWI3eEdlSGJVTApIckNUSVIwbEt2emVNSDErSnNFVTlsQXJIQXlXRlQ1a3RobGRZcGhnQ3ZWOXB6cXFIRnd1aGthOENvYjZlbU9nClYwOGl5WC9SQk9aY1dzc0ttTmtlbUdQSCtVZnh3dlJMWHkzdzFNdHlEU05nTFVHVlhjL3hZTjZWekVIVjg0YUUKOVBEcXhLTjRFT29BNHhXYW5ENzZJNlcwV0tHVXE4eUc2V1NndnZqVmRacjZSdFpZYTVDYVZsemoxaFhEMHNJKwoyNFFaMjgrKzFLcWZHcEQyWnZCZ0hxTjZISVc2aXhsQ1dzN1hqdUxxUVFLQmdRRGdzNlBRSjBOMUpqeS9STDViCkV4RVF2YXJvNVJ0QzNXZUhoOXVxMmxuTmp6bXczaHpubkFSUEZxMG4xdGZZL2JIMGlGQjdMUXNiemRHc0lFeHoKL2h0a0hlYVBuWHNEQnRSVGovUU16TThEWlVjNURQMVpMaC9RMHRvclVzenR0VVIxQ3NHWklhV3E1UlM2MXJXRQpJRXgwWFQ5UTEyb3V1NG5UQmtCN3BzdXhUUUtCZ1FEVFZRMk5ibm1VdU1SRWtBOHFmemVnL2pzS0xDdzhvdlpaCm5mcUhyR2RPMXprWTJpRCtXcVo4TjRqdFdrL0lURUVkQlNzbkpDbkFHSWJVTCsvR2tkRUVZYlhPVXZ3NUJiRjQKaVFzWHRTTU1iWXRqL3lNaWE2OHFRSEVPZUJrbmxEaEFpYlRNK2pZbTN4bEQwelFwWWF6UFcvRi9SZHdWS0ZQRgozNXZEN3ZXbGJRS0JnRFQ0NUpON3ppRmVCRkFyQ3AwNTMzb00zSy9PNHlCZVJidmp3VnVENGt2ZGlnSXlPcW8zClU2UzVlZFM4aDJJMlhLK0RPMFh1bG9IVmdhcU1hcm1sbkJ0OEdSQ2VWWk9mRm9za2txbzUxa3U4b28vR2lpdHQKL2o0aWx5QkRndUEvTFlaU0pOWE80dGxvNi93b0JkN0NKb1FBUDU3MVNhait1VDB3YWg4OGNTUzVBb0dCQUk1SAp0ZzhoY05PekxkaW5VTDZnMWZnYkVlN0FYS3dhWDFkb3FCS04vU08wZlNtQk9qTmxIcStFeURoYzFGZ2JGcitPCkNrYVk3MDc0ZEZZSlRCcFpjK3JLU2hmMkFQLzNHRXY1b0RFKzc3RGZVN2hvUHVSZXNaajF0K2d3N1dhYlFPQWEKbGxKbXB1eTJ5WkREY2x2bCtlM0ZqaXJOQXVadnR5OENaQ0dmRVYxbEFvR0JBTFpGalZhMTNyMFEwV0tydDhPeQpGZVVLOTBvWExydEZYMDRNWGNqWE4rTUJBbHJzQnVVaVRCQndUSE9zMUNCWXo4M2tienFydDRZZ3p4MnV6SmtGClU3a1F0MTdOM0YwMUNTYVAyWWo2MHZwNGpyanZjVHhUM0N1eWpFTHNrUjlaUWFyV2wycDRwdEt6TDJpSHpVNGwKUEVLV0F3YVdUQklDdmtBNk5DeHJmL0VzCi0tLS0tRU5EIFBSSVZBVEUgS0VZLS0tLS0K + mkdir -p /var/lib/rancher/k3s/server/manifests/gitpod + cat + cat + cat + cat + cat + cat + cat + /gitpod-installer init + yq e -i '.domain = "preview.gitpod-self-hosted.com"' config.yaml + yq e -i '.certificate.name = "https-certificates"' config.yaml + yq e -i '.certificate.kind = "secret"' config.yaml + yq e -i '.customCACert.name = "ca-key-pair"' config.yaml + yq e -i '.customCACert.kind = "secret"' config.yaml + yq e -i '.observability.logLevel = "debug"' config.yaml + yq e -i '.workspace.runtime.containerdSocket = "/run/k3s/containerd/containerd.sock"' config.yaml + yq e -i '.workspace.runtime.containerdRuntimeDir = "/var/lib/rancher/k3s/agent/containerd/io.containerd.runtime.v2.task/k8s.io/"' config.yaml + yq e -i '.experimental.telemetry.data.platform = "local-preview"' config.yaml + echo 'extracting images to download ahead...' extracting images to download ahead... + /gitpod-installer render --use-experimental-config --config config.yaml + grep image: + sed 's/ *//g' + sed s/image://g + sed 's/\"//g' + sed s/^-//g + sort + uniq + rm -rf /var/lib/rancher/k3s/server/manifests/gitpod + /bin/k3s server --disable traefik --node-label 'gitpod.io/workload_meta=true' --node-label 'gitpod.io/workload_ide=true' --node-label 'gitpod.io/workload_workspace_services=true' --node-label 'gitpod.io/workload_workspace_regular=true' --node-label 'gitpod.io/workload_workspace_headless=true' + run_telemetry + sleep 100 time="2022-08-12T08:16:05.816831079Z" level=info msg="Starting k3s v1.21.12+k3s1 (1db3ab57)" time="2022-08-12T08:16:05.820069160Z" level=info msg="Configuring sqlite3 database connection pooling: maxIdleConns=2, maxOpenConns=0, connMaxLifetime=0s" time="2022-08-12T08:16:05.820095179Z" level=info msg="Configuring database table schema and indexes, this may take a moment..." time="2022-08-12T08:16:05.823604719Z" level=info msg="Database tables and indexes are up to date" time="2022-08-12T08:16:05.825212439Z" level=info msg="Kine listening on unix://kine.sock" time="2022-08-12T08:16:05.832049589Z" level=info msg="certificate CN=system:admin,O=system:masters signed by CN=k3s-client-ca@1660292165: notBefore=2022-08-12 08:16:05 +0000 UTC notAfter=2023-08-12 08:16:05 +0000 UTC" time="2022-08-12T08:16:05.832422958Z" level=info msg="certificate CN=system:kube-controller-manager signed by CN=k3s-client-ca@1660292165: notBefore=2022-08-12 08:16:05 +0000 UTC notAfter=2023-08-12 08:16:05 +0000 UTC" time="2022-08-12T08:16:05.832786488Z" level=info msg="certificate CN=system:kube-scheduler signed by CN=k3s-client-ca@1660292165: notBefore=2022-08-12 08:16:05 +0000 UTC notAfter=2023-08-12 08:16:05 +0000 UTC" time="2022-08-12T08:16:05.833154687Z" level=info msg="certificate CN=system:apiserver,O=system:masters signed by CN=k3s-client-ca@1660292165: notBefore=2022-08-12 08:16:05 +0000 UTC notAfter=2023-08-12 08:16:05 +0000 UTC" time="2022-08-12T08:16:05.833515132Z" level=info msg="certificate CN=system:kube-proxy signed by CN=k3s-client-ca@1660292165: notBefore=2022-08-12 08:16:05 +0000 UTC notAfter=2023-08-12 08:16:05 +0000 UTC" time="2022-08-12T08:16:05.833824882Z" level=info msg="certificate CN=system:k3s-controller signed by CN=k3s-client-ca@1660292165: notBefore=2022-08-12 08:16:05 +0000 UTC notAfter=2023-08-12 08:16:05 +0000 UTC" time="2022-08-12T08:16:05.834199373Z" level=info msg="certificate CN=k3s-cloud-controller-manager signed by CN=k3s-client-ca@1660292165: notBefore=2022-08-12 08:16:05 +0000 UTC notAfter=2023-08-12 08:16:05 +0000 UTC" time="2022-08-12T08:16:05.834834682Z" level=info msg="certificate CN=kube-apiserver signed by CN=k3s-server-ca@1660292165: notBefore=2022-08-12 08:16:05 +0000 UTC notAfter=2023-08-12 08:16:05 +0000 UTC" time="2022-08-12T08:16:05.835439835Z" level=info msg="certificate CN=system:auth-proxy signed by CN=k3s-request-header-ca@1660292165: notBefore=2022-08-12 08:16:05 +0000 UTC notAfter=2023-08-12 08:16:05 +0000 UTC" time="2022-08-12T08:16:05.836021905Z" level=info msg="certificate CN=etcd-server signed by CN=etcd-server-ca@1660292165: notBefore=2022-08-12 08:16:05 +0000 UTC notAfter=2023-08-12 08:16:05 +0000 UTC" time="2022-08-12T08:16:05.836345942Z" level=info msg="certificate CN=etcd-client signed by CN=etcd-server-ca@1660292165: notBefore=2022-08-12 08:16:05 +0000 UTC notAfter=2023-08-12 08:16:05 +0000 UTC" time="2022-08-12T08:16:05.836947858Z" level=info msg="certificate CN=etcd-peer signed by CN=etcd-peer-ca@1660292165: notBefore=2022-08-12 08:16:05 +0000 UTC notAfter=2023-08-12 08:16:05 +0000 UTC" time="2022-08-12T08:16:06.038885186Z" level=info msg="certificate CN=k3s,O=k3s signed by CN=k3s-server-ca@1660292165: notBefore=2022-08-12 08:16:05 +0000 UTC notAfter=2023-08-12 08:16:06 +0000 UTC" time="2022-08-12T08:16:06.039235231Z" level=info msg="Active TLS secret (ver=) (count 9): map[listener.cattle.io/cn-0eec92c2ff40:0eec92c2ff40 listener.cattle.io/cn-10.43.0.1:10.43.0.1 listener.cattle.io/cn-127.0.0.1:127.0.0.1 listener.cattle.io/cn-172.17.0.2:172.17.0.2 listener.cattle.io/cn-kubernetes:kubernetes listener.cattle.io/cn-kubernetes.default:kubernetes.default listener.cattle.io/cn-kubernetes.default.svc:kubernetes.default.svc listener.cattle.io/cn-kubernetes.default.svc.cluster.local:kubernetes.default.svc.cluster.local listener.cattle.io/cn-localhost:localhost listener.cattle.io/fingerprint:SHA1=3A8D99E27B0DD569E0E54209560EBAD6A15370C2]" time="2022-08-12T08:16:06.043934218Z" level=info msg="Running kube-apiserver --advertise-port=6443 --allow-privileged=true --anonymous-auth=false --api-audiences=https://kubernetes.default.svc.cluster.local,k3s --authorization-mode=Node,RBAC --bind-address=127.0.0.1 --cert-dir=/var/lib/rancher/k3s/server/tls/temporary-certs --client-ca-file=/var/lib/rancher/k3s/server/tls/client-ca.crt --enable-admission-plugins=NodeRestriction --etcd-servers=unix://kine.sock --insecure-port=0 --kubelet-certificate-authority=/var/lib/rancher/k3s/server/tls/server-ca.crt --kubelet-client-certificate=/var/lib/rancher/k3s/server/tls/client-kube-apiserver.crt --kubelet-client-key=/var/lib/rancher/k3s/server/tls/client-kube-apiserver.key --profiling=false --proxy-client-cert-file=/var/lib/rancher/k3s/server/tls/client-auth-proxy.crt --proxy-client-key-file=/var/lib/rancher/k3s/server/tls/client-auth-proxy.key --requestheader-allowed-names=system:auth-proxy --requestheader-client-ca-file=/var/lib/rancher/k3s/server/tls/request-header-ca.crt --requestheader-extra-headers-prefix=X-Remote-Extra- --requestheader-group-headers=X-Remote-Group --requestheader-username-headers=X-Remote-User --secure-port=6444 --service-account-issuer=https://kubernetes.default.svc.cluster.local --service-account-key-file=/var/lib/rancher/k3s/server/tls/service.key --service-account-signing-key-file=/var/lib/rancher/k3s/server/tls/service.key --service-cluster-ip-range=10.43.0.0/16 --service-node-port-range=30000-32767 --storage-backend=etcd3 --tls-cert-file=/var/lib/rancher/k3s/server/tls/serving-kube-apiserver.crt --tls-private-key-file=/var/lib/rancher/k3s/server/tls/serving-kube-apiserver.key" Flag --insecure-port has been deprecated, This flag has no effect now and will be removed in v1.24. I0812 08:16:06.045241 982 server.go:656] external host was not specified, using 172.17.0.2 I0812 08:16:06.045489 982 server.go:195] Version: v1.21.12+k3s1 time="2022-08-12T08:16:06.047613716Z" level=info msg="Running kube-scheduler --address=127.0.0.1 --bind-address=127.0.0.1 --kubeconfig=/var/lib/rancher/k3s/server/cred/scheduler.kubeconfig --leader-elect=false --port=10251 --profiling=false --secure-port=0" time="2022-08-12T08:16:06.047696511Z" level=info msg="Waiting for API server to become available" time="2022-08-12T08:16:06.048079188Z" level=info msg="Running kube-controller-manager --address=127.0.0.1 --allocate-node-cidrs=true --bind-address=127.0.0.1 --cluster-cidr=10.42.0.0/16 --cluster-signing-kube-apiserver-client-cert-file=/var/lib/rancher/k3s/server/tls/client-ca.crt --cluster-signing-kube-apiserver-client-key-file=/var/lib/rancher/k3s/server/tls/client-ca.key --cluster-signing-kubelet-client-cert-file=/var/lib/rancher/k3s/server/tls/client-ca.crt --cluster-signing-kubelet-client-key-file=/var/lib/rancher/k3s/server/tls/client-ca.key --cluster-signing-kubelet-serving-cert-file=/var/lib/rancher/k3s/server/tls/server-ca.crt --cluster-signing-kubelet-serving-key-file=/var/lib/rancher/k3s/server/tls/server-ca.key --cluster-signing-legacy-unknown-cert-file=/var/lib/rancher/k3s/server/tls/client-ca.crt --cluster-signing-legacy-unknown-key-file=/var/lib/rancher/k3s/server/tls/client-ca.key --configure-cloud-routes=false --controllers=*,-service,-route,-cloud-node-lifecycle --kubeconfig=/var/lib/rancher/k3s/server/cred/controller.kubeconfig --leader-elect=false --port=10252 --profiling=false --root-ca-file=/var/lib/rancher/k3s/server/tls/server-ca.crt --secure-port=0 --service-account-private-key-file=/var/lib/rancher/k3s/server/tls/service.key --use-service-account-credentials=true" time="2022-08-12T08:16:06.048564706Z" level=info msg="Running cloud-controller-manager --allocate-node-cidrs=true --bind-address=127.0.0.1 --cloud-provider=k3s --cluster-cidr=10.42.0.0/16 --configure-cloud-routes=false --kubeconfig=/var/lib/rancher/k3s/server/cred/cloud-controller.kubeconfig --leader-elect=false --node-status-update-frequency=1m0s --port=0 --profiling=false" time="2022-08-12T08:16:06.049409197Z" level=info msg="Node token is available at /var/lib/rancher/k3s/server/token" time="2022-08-12T08:16:06.049471724Z" level=info msg="To join node to cluster: k3s agent -s https://172.17.0.2:6443 -t ${NODE_TOKEN}" time="2022-08-12T08:16:06.050472979Z" level=info msg="Wrote kubeconfig /etc/rancher/k3s/k3s.yaml" time="2022-08-12T08:16:06.050495761Z" level=info msg="Run: k3s kubectl" time="2022-08-12T08:16:06.081109692Z" level=info msg="certificate CN=0eec92c2ff40 signed by CN=k3s-server-ca@1660292165: notBefore=2022-08-12 08:16:05 +0000 UTC notAfter=2023-08-12 08:16:06 +0000 UTC" time="2022-08-12T08:16:06.083549359Z" level=info msg="certificate CN=system:node:0eec92c2ff40,O=system:nodes signed by CN=k3s-client-ca@1660292165: notBefore=2022-08-12 08:16:05 +0000 UTC notAfter=2023-08-12 08:16:06 +0000 UTC" time="2022-08-12T08:16:06.115071300Z" level=info msg="Module overlay was already loaded" time="2022-08-12T08:16:06.115118759Z" level=info msg="Module nf_conntrack was already loaded" time="2022-08-12T08:16:06.115133657Z" level=info msg="Module br_netfilter was already loaded" time="2022-08-12T08:16:06.115772162Z" level=warning msg="Failed to load kernel module iptable_nat with modprobe" time="2022-08-12T08:16:06.120262158Z" level=info msg="Set sysctl 'net/netfilter/nf_conntrack_max' to 786432" time="2022-08-12T08:16:06.120282025Z" level=error msg="Failed to set sysctl: open /proc/sys/net/netfilter/nf_conntrack_max: permission denied" time="2022-08-12T08:16:06.120304958Z" level=info msg="Set sysctl 'net/netfilter/nf_conntrack_tcp_timeout_established' to 86400" time="2022-08-12T08:16:06.120342278Z" level=info msg="Set sysctl 'net/netfilter/nf_conntrack_tcp_timeout_close_wait' to 3600" time="2022-08-12T08:16:06.121456674Z" level=info msg="Logging containerd to /var/lib/rancher/k3s/agent/containerd/containerd.log" time="2022-08-12T08:16:06.121543336Z" level=info msg="Running containerd -c /var/lib/rancher/k3s/agent/etc/containerd/config.toml -a /run/k3s/containerd/containerd.sock --state /run/k3s/containerd --root /var/lib/rancher/k3s/agent/containerd" I0812 08:16:06.297130 982 shared_informer.go:240] Waiting for caches to sync for node_authorizer I0812 08:16:06.297713 982 plugins.go:158] Loaded 12 mutating admission controller(s) successfully in the following order: NamespaceLifecycle,LimitRanger,ServiceAccount,NodeRestriction,TaintNodesByCondition,Priority,DefaultTolerationSeconds,DefaultStorageClass,StorageObjectInUseProtection,RuntimeClass,DefaultIngressClass,MutatingAdmissionWebhook. I0812 08:16:06.297725 982 plugins.go:161] Loaded 10 validating admission controller(s) successfully in the following order: LimitRanger,ServiceAccount,Priority,PersistentVolumeClaimResize,RuntimeClass,CertificateApproval,CertificateSigning,CertificateSubjectRestriction,ValidatingAdmissionWebhook,ResourceQuota. I0812 08:16:06.298286 982 plugins.go:158] Loaded 12 mutating admission controller(s) successfully in the following order: NamespaceLifecycle,LimitRanger,ServiceAccount,NodeRestriction,TaintNodesByCondition,Priority,DefaultTolerationSeconds,DefaultStorageClass,StorageObjectInUseProtection,RuntimeClass,DefaultIngressClass,MutatingAdmissionWebhook. I0812 08:16:06.298293 982 plugins.go:161] Loaded 10 validating admission controller(s) successfully in the following order: LimitRanger,ServiceAccount,Priority,PersistentVolumeClaimResize,RuntimeClass,CertificateApproval,CertificateSigning,CertificateSubjectRestriction,ValidatingAdmissionWebhook,ResourceQuota. I0812 08:16:06.316827 982 instance.go:283] Using reconciler: lease I0812 08:16:06.339782 982 rest.go:130] the default service ipfamily for this cluster is: IPv4 W0812 08:16:06.560698 982 genericapiserver.go:425] Skipping API node.k8s.io/v1alpha1 because it has no resources. W0812 08:16:06.567690 982 genericapiserver.go:425] Skipping API rbac.authorization.k8s.io/v1alpha1 because it has no resources. W0812 08:16:06.570225 982 genericapiserver.go:425] Skipping API scheduling.k8s.io/v1alpha1 because it has no resources. W0812 08:16:06.574576 982 genericapiserver.go:425] Skipping API storage.k8s.io/v1alpha1 because it has no resources. W0812 08:16:06.576312 982 genericapiserver.go:425] Skipping API flowcontrol.apiserver.k8s.io/v1alpha1 because it has no resources. W0812 08:16:06.580634 982 genericapiserver.go:425] Skipping API apps/v1beta2 because it has no resources. W0812 08:16:06.580647 982 genericapiserver.go:425] Skipping API apps/v1beta1 because it has no resources. I0812 08:16:06.587943 982 plugins.go:158] Loaded 12 mutating admission controller(s) successfully in the following order: NamespaceLifecycle,LimitRanger,ServiceAccount,NodeRestriction,TaintNodesByCondition,Priority,DefaultTolerationSeconds,DefaultStorageClass,StorageObjectInUseProtection,RuntimeClass,DefaultIngressClass,MutatingAdmissionWebhook. I0812 08:16:06.587963 982 plugins.go:161] Loaded 10 validating admission controller(s) successfully in the following order: LimitRanger,ServiceAccount,Priority,PersistentVolumeClaimResize,RuntimeClass,CertificateApproval,CertificateSigning,CertificateSubjectRestriction,ValidatingAdmissionWebhook,ResourceQuota. time="2022-08-12T08:16:07.123372414Z" level=info msg="Containerd is now running" time="2022-08-12T08:16:07.128440351Z" level=info msg="Connecting to proxy" url="wss://127.0.0.1:6443/v1-k3s/connect" time="2022-08-12T08:16:07.134845191Z" level=info msg="Handling backend connection request [0eec92c2ff40]" time="2022-08-12T08:16:07.135543047Z" level=info msg="Running kubelet --address=0.0.0.0 --anonymous-auth=false --authentication-token-webhook=true --authorization-mode=Webhook --cgroup-driver=cgroupfs --client-ca-file=/var/lib/rancher/k3s/agent/client-ca.crt --cloud-provider=external --cluster-dns=10.43.0.10 --cluster-domain=cluster.local --cni-bin-dir=/bin --cni-conf-dir=/var/lib/rancher/k3s/agent/etc/cni/net.d --container-runtime-endpoint=unix:///run/k3s/containerd/containerd.sock --container-runtime=remote --containerd=/run/k3s/containerd/containerd.sock --eviction-hard=imagefs.available<5%,nodefs.available<5% --eviction-minimum-reclaim=imagefs.available=10%,nodefs.available=10% --fail-swap-on=false --healthz-bind-address=127.0.0.1 --hostname-override=0eec92c2ff40 --kubeconfig=/var/lib/rancher/k3s/agent/kubelet.kubeconfig --kubelet-cgroups=/k3s --node-labels=gitpod.io/workload_meta=true,gitpod.io/workload_ide=true,gitpod.io/workload_workspace_services=true,gitpod.io/workload_workspace_regular=true,gitpod.io/workload_workspace_headless=true --pod-manifest-path=/var/lib/rancher/k3s/agent/pod-manifests --read-only-port=0 --resolv-conf=/etc/resolv.conf --runtime-cgroups=/k3s --serialize-image-pulls=false --tls-cert-file=/var/lib/rancher/k3s/agent/serving-kubelet.crt --tls-private-key-file=/var/lib/rancher/k3s/agent/serving-kubelet.key" Flag --cloud-provider has been deprecated, will be removed in 1.23, in favor of removing cloud provider code from Kubelet. Flag --cni-bin-dir has been deprecated, will be removed along with dockershim. Flag --cni-conf-dir has been deprecated, will be removed along with dockershim. Flag --containerd has been deprecated, This is a cadvisor flag that was mistakenly registered with the Kubelet. Due to legacy concerns, it will follow the standard CLI deprecation timeline before being removed. I0812 08:16:07.136994 982 server.go:436] "Kubelet version" kubeletVersion="v1.21.12+k3s1" time="2022-08-12T08:16:07.141740169Z" level=info msg="Waiting to retrieve kube-proxy configuration; server is not ready: https://127.0.0.1:6443/v1-k3s/readyz: 500 Internal Server Error" W0812 08:16:07.159096 982 manager.go:159] Cannot detect current cgroup on cgroup v2 I0812 08:16:07.159114 982 dynamic_cafile_content.go:167] Starting client-ca-bundle::/var/lib/rancher/k3s/agent/client-ca.crt I0812 08:16:07.426947 982 dynamic_cafile_content.go:167] Starting request-header::/var/lib/rancher/k3s/server/tls/request-header-ca.crt I0812 08:16:07.426977 982 dynamic_cafile_content.go:167] Starting client-ca-bundle::/var/lib/rancher/k3s/server/tls/client-ca.crt I0812 08:16:07.427071 982 dynamic_serving_content.go:130] Starting serving-cert::/var/lib/rancher/k3s/server/tls/serving-kube-apiserver.crt::/var/lib/rancher/k3s/server/tls/serving-kube-apiserver.key I0812 08:16:07.427243 982 secure_serving.go:202] Serving securely on 127.0.0.1:6444 I0812 08:16:07.427275 982 controller.go:83] Starting OpenAPI AggregationController I0812 08:16:07.427292 982 tlsconfig.go:240] Starting DynamicServingCertificateController I0812 08:16:07.427511 982 available_controller.go:475] Starting AvailableConditionController I0812 08:16:07.427530 982 cache.go:32] Waiting for caches to sync for AvailableConditionController controller I0812 08:16:07.427598 982 apiservice_controller.go:97] Starting APIServiceRegistrationController I0812 08:16:07.427606 982 cache.go:32] Waiting for caches to sync for APIServiceRegistrationController controller I0812 08:16:07.427598 982 apf_controller.go:307] Starting API Priority and Fairness config controller I0812 08:16:07.427696 982 autoregister_controller.go:141] Starting autoregister controller I0812 08:16:07.427713 982 cache.go:32] Waiting for caches to sync for autoregister controller I0812 08:16:07.427752 982 crdregistration_controller.go:111] Starting crd-autoregister controller I0812 08:16:07.427761 982 shared_informer.go:240] Waiting for caches to sync for crd-autoregister I0812 08:16:07.427813 982 customresource_discovery_controller.go:209] Starting DiscoveryController I0812 08:16:07.427862 982 dynamic_serving_content.go:130] Starting aggregator-proxy-cert::/var/lib/rancher/k3s/server/tls/client-auth-proxy.crt::/var/lib/rancher/k3s/server/tls/client-auth-proxy.key I0812 08:16:07.427904 982 dynamic_cafile_content.go:167] Starting client-ca-bundle::/var/lib/rancher/k3s/server/tls/client-ca.crt I0812 08:16:07.427933 982 dynamic_cafile_content.go:167] Starting request-header::/var/lib/rancher/k3s/server/tls/request-header-ca.crt I0812 08:16:07.427982 982 controller.go:86] Starting OpenAPI controller I0812 08:16:07.428004 982 naming_controller.go:291] Starting NamingConditionController I0812 08:16:07.428024 982 establishing_controller.go:76] Starting EstablishingController I0812 08:16:07.428047 982 nonstructuralschema_controller.go:192] Starting NonStructuralSchemaConditionController I0812 08:16:07.428066 982 apiapproval_controller.go:186] Starting KubernetesAPIApprovalPolicyConformantConditionController I0812 08:16:07.428092 982 crd_finalizer.go:266] Starting CRDFinalizer I0812 08:16:07.427879 982 cluster_authentication_trust_controller.go:440] Starting cluster_authentication_trust_controller controller I0812 08:16:07.428113 982 shared_informer.go:240] Waiting for caches to sync for cluster_authentication_trust_controller I0812 08:16:07.436811 982 controller.go:611] quota admission added evaluator for: namespaces E0812 08:16:07.442856 982 controller.go:151] Unable to perform initial Kubernetes service initialization: Service "kubernetes" is invalid: spec.clusterIPs: Invalid value: []string{"10.43.0.1"}: failed to allocated ip:10.43.0.1 with error:cannot allocate resources of type serviceipallocations at this time E0812 08:16:07.443914 982 controller.go:156] Unable to remove old endpoints from kubernetes service: StorageError: key not found, Code: 1, Key: /registry/masterleases/172.17.0.2, ResourceVersion: 0, AdditionalErrorMsg: I0812 08:16:07.497582 982 shared_informer.go:247] Caches are synced for node_authorizer I0812 08:16:07.528585 982 cache.go:39] Caches are synced for APIServiceRegistrationController controller I0812 08:16:07.528593 982 cache.go:39] Caches are synced for AvailableConditionController controller I0812 08:16:07.528617 982 shared_informer.go:247] Caches are synced for crd-autoregister I0812 08:16:07.528627 982 cache.go:39] Caches are synced for autoregister controller I0812 08:16:07.528631 982 shared_informer.go:247] Caches are synced for cluster_authentication_trust_controller I0812 08:16:07.528681 982 apf_controller.go:312] Running API Priority and Fairness config worker I0812 08:16:08.427145 982 controller.go:132] OpenAPI AggregationController: action for item : Nothing (removed from the queue). I0812 08:16:08.433206 982 storage_scheduling.go:132] created PriorityClass system-node-critical with value 2000001000 I0812 08:16:08.436475 982 storage_scheduling.go:132] created PriorityClass system-cluster-critical with value 2000000000 I0812 08:16:08.436494 982 storage_scheduling.go:148] all system priority classes are created successfully or already exist. I0812 08:16:08.440889 982 controller.go:132] OpenAPI AggregationController: action for item k8s_internal_local_delegation_chain_0000000000: Nothing (removed from the queue). I0812 08:16:08.735792 982 controller.go:611] quota admission added evaluator for: roles.rbac.authorization.k8s.io I0812 08:16:08.762067 982 controller.go:611] quota admission added evaluator for: rolebindings.rbac.authorization.k8s.io W0812 08:16:08.868399 982 lease.go:233] Resetting endpoints for master service "kubernetes" to [172.17.0.2] I0812 08:16:08.869072 982 controller.go:611] quota admission added evaluator for: endpoints I0812 08:16:08.871999 982 controller.go:611] quota admission added evaluator for: endpointslices.discovery.k8s.io time="2022-08-12T08:16:09.438574406Z" level=info msg="Kube API server is now running" time="2022-08-12T08:16:09.438588242Z" level=info msg="Waiting for cloud-controller-manager privileges to become available" time="2022-08-12T08:16:09.438607428Z" level=info msg="k3s is up and running" Flag --address has been deprecated, see --bind-address instead. I0812 08:16:09.440778 982 controllermanager.go:175] Version: v1.21.12+k3s1 I0812 08:16:09.441056 982 deprecated_insecure_serving.go:56] Serving insecurely on 127.0.0.1:10252 time="2022-08-12T08:16:09.445728980Z" level=info msg="Creating CRD addons.k3s.cattle.io" time="2022-08-12T08:16:09.448354755Z" level=info msg="Creating CRD helmcharts.helm.cattle.io" time="2022-08-12T08:16:09.450539475Z" level=info msg="Creating CRD helmchartconfigs.helm.cattle.io" time="2022-08-12T08:16:09.456079627Z" level=info msg="Waiting for CRD addons.k3s.cattle.io to become available" time="2022-08-12T08:16:09.958336729Z" level=info msg="Done waiting for CRD addons.k3s.cattle.io to become available" time="2022-08-12T08:16:09.958359381Z" level=info msg="Waiting for CRD helmcharts.helm.cattle.io to become available" time="2022-08-12T08:16:10.461081554Z" level=info msg="Done waiting for CRD helmcharts.helm.cattle.io to become available" time="2022-08-12T08:16:10.461110187Z" level=info msg="Waiting for CRD helmchartconfigs.helm.cattle.io to become available" time="2022-08-12T08:16:10.963896560Z" level=info msg="Done waiting for CRD helmchartconfigs.helm.cattle.io to become available" time="2022-08-12T08:16:10.970242681Z" level=info msg="Writing static file: /var/lib/rancher/k3s/server/static/charts/traefik-crd-10.14.100.tgz" time="2022-08-12T08:16:10.970464105Z" level=info msg="Writing static file: /var/lib/rancher/k3s/server/static/charts/traefik-10.14.100.tgz" time="2022-08-12T08:16:10.970571647Z" level=info msg="Writing manifest: /var/lib/rancher/k3s/server/manifests/ccm.yaml" time="2022-08-12T08:16:10.970708513Z" level=info msg="Writing manifest: /var/lib/rancher/k3s/server/manifests/local-storage.yaml" time="2022-08-12T08:16:10.970821985Z" level=info msg="Writing manifest: /var/lib/rancher/k3s/server/manifests/metrics-server/aggregated-metrics-reader.yaml" time="2022-08-12T08:16:10.970899490Z" level=info msg="Writing manifest: /var/lib/rancher/k3s/server/manifests/metrics-server/auth-delegator.yaml" time="2022-08-12T08:16:10.970984710Z" level=info msg="Writing manifest: /var/lib/rancher/k3s/server/manifests/metrics-server/auth-reader.yaml" time="2022-08-12T08:16:10.971059410Z" level=info msg="Writing manifest: /var/lib/rancher/k3s/server/manifests/metrics-server/metrics-apiservice.yaml" time="2022-08-12T08:16:10.971173022Z" level=info msg="Writing manifest: /var/lib/rancher/k3s/server/manifests/metrics-server/metrics-server-deployment.yaml" time="2022-08-12T08:16:10.971264383Z" level=info msg="Writing manifest: /var/lib/rancher/k3s/server/manifests/metrics-server/metrics-server-service.yaml" time="2022-08-12T08:16:10.971354913Z" level=info msg="Writing manifest: /var/lib/rancher/k3s/server/manifests/metrics-server/resource-reader.yaml" time="2022-08-12T08:16:10.971507439Z" level=info msg="Writing manifest: /var/lib/rancher/k3s/server/manifests/coredns.yaml" time="2022-08-12T08:16:10.971604410Z" level=info msg="Writing manifest: /var/lib/rancher/k3s/server/manifests/rolebindings.yaml" time="2022-08-12T08:16:11.072515742Z" level=info msg="Starting k3s.cattle.io/v1, Kind=Addon controller" time="2022-08-12T08:16:11.072581535Z" level=info msg="Creating deploy event broadcaster" time="2022-08-12T08:16:11.072588378Z" level=info msg="Starting /v1, Kind=Secret controller" I0812 08:16:11.073943 982 controller.go:611] quota admission added evaluator for: addons.k3s.cattle.io time="2022-08-12T08:16:11.074234499Z" level=info msg="Waiting for control-plane node 0eec92c2ff40 startup: nodes \"0eec92c2ff40\" not found" time="2022-08-12T08:16:11.075007015Z" level=info msg="Active TLS secret k3s-serving (ver=217) (count 9): map[listener.cattle.io/cn-0eec92c2ff40:0eec92c2ff40 listener.cattle.io/cn-10.43.0.1:10.43.0.1 listener.cattle.io/cn-127.0.0.1:127.0.0.1 listener.cattle.io/cn-172.17.0.2:172.17.0.2 listener.cattle.io/cn-kubernetes:kubernetes listener.cattle.io/cn-kubernetes.default:kubernetes.default listener.cattle.io/cn-kubernetes.default.svc:kubernetes.default.svc listener.cattle.io/cn-kubernetes.default.svc.cluster.local:kubernetes.default.svc.cluster.local listener.cattle.io/cn-localhost:localhost listener.cattle.io/fingerprint:SHA1=3A8D99E27B0DD569E0E54209560EBAD6A15370C2]" time="2022-08-12T08:16:11.077015435Z" level=info msg="Event(v1.ObjectReference{Kind:\"Addon\", Namespace:\"kube-system\", Name:\"ccm\", UID:\"c9c03f71-5a16-45fb-9a6e-004d4c9b06f5\", APIVersion:\"k3s.cattle.io/v1\", ResourceVersion:\"218\", FieldPath:\"\"}): type: 'Normal' reason: 'ApplyingManifest' Applying manifest at \"/var/lib/rancher/k3s/server/manifests/ccm.yaml\"" time="2022-08-12T08:16:11.089420458Z" level=info msg="Event(v1.ObjectReference{Kind:\"Addon\", Namespace:\"kube-system\", Name:\"ccm\", UID:\"c9c03f71-5a16-45fb-9a6e-004d4c9b06f5\", APIVersion:\"k3s.cattle.io/v1\", ResourceVersion:\"218\", FieldPath:\"\"}): type: 'Normal' reason: 'AppliedManifest' Applied manifest at \"/var/lib/rancher/k3s/server/manifests/ccm.yaml\"" time="2022-08-12T08:16:11.095992882Z" level=info msg="Event(v1.ObjectReference{Kind:\"Addon\", Namespace:\"kube-system\", Name:\"custom-coredns\", UID:\"c3aebc38-3dec-4eaf-9e08-b2bc6401a00e\", APIVersion:\"k3s.cattle.io/v1\", ResourceVersion:\"224\", FieldPath:\"\"}): type: 'Normal' reason: 'ApplyingManifest' Applying manifest at \"/var/lib/rancher/k3s/server/manifests/custom-coredns.yaml\"" I0812 08:16:11.100249 982 controller.go:611] quota admission added evaluator for: serviceaccounts I0812 08:16:11.131772 982 controller.go:611] quota admission added evaluator for: deployments.apps time="2022-08-12T08:16:11.141691402Z" level=info msg="Event(v1.ObjectReference{Kind:\"Addon\", Namespace:\"kube-system\", Name:\"custom-coredns\", UID:\"c3aebc38-3dec-4eaf-9e08-b2bc6401a00e\", APIVersion:\"k3s.cattle.io/v1\", ResourceVersion:\"224\", FieldPath:\"\"}): type: 'Normal' reason: 'AppliedManifest' Applied manifest at \"/var/lib/rancher/k3s/server/manifests/custom-coredns.yaml\"" I0812 08:16:11.197515 982 request.go:668] Waited for 1.048042351s due to client-side throttling, not priority and fairness, request: GET:https://127.0.0.1:6444/apis/networking.k8s.io/v1beta1?timeout=32s time="2022-08-12T08:16:11.203364037Z" level=info msg="Event(v1.ObjectReference{Kind:\"Addon\", Namespace:\"kube-system\", Name:\"gitpod\", UID:\"ebfe0f6c-8d5d-4718-ba84-f027e31a0bf8\", APIVersion:\"k3s.cattle.io/v1\", ResourceVersion:\"235\", FieldPath:\"\"}): type: 'Normal' reason: 'ApplyingManifest' Applying manifest at \"/var/lib/rancher/k3s/server/manifests/gitpod.yaml\"" I0812 08:16:11.247607 982 controller.go:611] quota admission added evaluator for: poddisruptionbudgets.policy time="2022-08-12T08:16:11.450678162Z" level=info msg="Starting /v1, Kind=Pod controller" time="2022-08-12T08:16:11.450673663Z" level=info msg="Starting /v1, Kind=Node controller" time="2022-08-12T08:16:11.450686517Z" level=info msg="Starting /v1, Kind=Service controller" time="2022-08-12T08:16:11.450692619Z" level=info msg="Starting /v1, Kind=Endpoints controller" time="2022-08-12T08:16:11.474233579Z" level=info msg="Starting helm.cattle.io/v1, Kind=HelmChartConfig controller" time="2022-08-12T08:16:11.474270348Z" level=info msg="Starting helm.cattle.io/v1, Kind=HelmChart controller" time="2022-08-12T08:16:11.474279916Z" level=info msg="Starting batch/v1, Kind=Job controller" time="2022-08-12T08:16:11.646083340Z" level=info msg="Cluster dns configmap has been set successfully" I0812 08:16:11.766730 982 serving.go:354] Generated self-signed cert in-memory W0812 08:16:12.097454 982 authentication.go:308] No authentication-kubeconfig provided in order to lookup client-ca-file in configmap/extension-apiserver-authentication in kube-system, so client certificate authentication won't work. W0812 08:16:12.097464 982 authentication.go:332] No authentication-kubeconfig provided in order to lookup requestheader-client-ca-file in configmap/extension-apiserver-authentication in kube-system, so request-header client certificate authentication won't work. W0812 08:16:12.097471 982 authorization.go:184] No authorization-kubeconfig provided, so SubjectAccessReview of authorization tokens won't work. I0812 08:16:12.099334 982 controllermanager.go:142] Version: v1.21.12+k3s1 I0812 08:16:12.100032 982 secure_serving.go:202] Serving securely on 127.0.0.1:10258 I0812 08:16:12.100091 982 tlsconfig.go:240] Starting DynamicServingCertificateController time="2022-08-12T08:16:12.142875071Z" level=info msg="Running kube-proxy --cluster-cidr=10.42.0.0/16 --conntrack-max-per-core=0 --conntrack-tcp-timeout-close-wait=0s --conntrack-tcp-timeout-established=0s --healthz-bind-address=127.0.0.1 --hostname-override=0eec92c2ff40 --kubeconfig=/var/lib/rancher/k3s/agent/kubeproxy.kubeconfig --proxy-mode=iptables" W0812 08:16:12.143151 982 server.go:224] WARNING: all flags other than --config, --write-config-to, and --cleanup are deprecated. Please begin using a config file ASAP. W0812 08:16:12.143654 982 proxier.go:659] Failed to read file /lib/modules/5.18.13-200.fc36.x86_64/modules.builtin with error open /lib/modules/5.18.13-200.fc36.x86_64/modules.builtin: no such file or directory. You can ignore this message when kube-proxy is running inside container without mounting /lib/modules W0812 08:16:12.144174 982 proxier.go:669] Failed to load kernel module ip_vs with modprobe. You can ignore this message when kube-proxy is running inside container without mounting /lib/modules W0812 08:16:12.144642 982 proxier.go:669] Failed to load kernel module ip_vs_rr with modprobe. You can ignore this message when kube-proxy is running inside container without mounting /lib/modules W0812 08:16:12.145163 982 proxier.go:669] Failed to load kernel module ip_vs_wrr with modprobe. You can ignore this message when kube-proxy is running inside container without mounting /lib/modules W0812 08:16:12.145407 982 proxier.go:669] Failed to load kernel module ip_vs_sh with modprobe. You can ignore this message when kube-proxy is running inside container without mounting /lib/modules W0812 08:16:12.145669 982 proxier.go:669] Failed to load kernel module nf_conntrack with modprobe. You can ignore this message when kube-proxy is running inside container without mounting /lib/modules time="2022-08-12T08:16:12.147410011Z" level=warning msg="Running modprobe ip_vs failed with message: `modprobe: can't change directory to '5.18.13-200.fc36.x86_64': No such file or directory`, error: exit status 1" E0812 08:16:12.154777 982 node.go:161] Failed to retrieve node info: nodes "0eec92c2ff40" not found W0812 08:16:12.163245 982 fs.go:214] stat failed on /dev/mapper/luks-b627714b-8701-4a2a-bcbb-2bf0fb1d0957 with error: no such file or directory W0812 08:16:12.186700 982 info.go:53] Couldn't collect info from any of the files in "/etc/machine-id,/var/lib/dbus/machine-id" I0812 08:16:12.187152 982 server.go:660] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" I0812 08:16:12.187303 982 container_manager_linux.go:291] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] I0812 08:16:12.187389 982 container_manager_linux.go:296] "Creating Container Manager object based on Node Config" nodeConfig={RuntimeCgroupsName:/k3s SystemCgroupsName: KubeletCgroupsName:/k3s ContainerRuntime:remote CgroupsPerQOS:true CgroupRoot:/ CgroupDriver:cgroupfs KubeletRootDir:/var/lib/kubelet ProtectKernelDefaults:false NodeAllocatableConfig:{KubeReservedCgroupName: SystemReservedCgroupName: ReservedSystemCPUs: EnforceNodeAllocatable:map[pods:{}] KubeReserved:map[] SystemReserved:map[] HardEvictionThresholds:[{Signal:imagefs.available Operator:LessThan Value:{Quantity:Steps to reproduce
Run
local-preview
on a linux distro with cgroups v2Workspace affected
No response
Expected behavior
No response
Example repository
No response
Anything else?
No response