Altinity / clickhouse-operator

Altinity Kubernetes Operator for ClickHouse creates, configures and manages ClickHouse® clusters running on Kubernetes
https://altinity.com
Apache License 2.0
1.93k stars 462 forks source link

clickhouse installation: statefulset pod stuck in `CrashLoopBackOff` w/ log `filesystem error: in weakly_canonical:` #1509

Closed fidiego closed 2 weeks ago

fidiego commented 2 months ago

Description

I have a clickhouse-installation version 24.3.7 running on an EKS cluster. One of my Statefulsets is stuck in CrashLoopBackOff. Re-starting the pod has no effect. I am hesitant to delete it manually in case I lose the underlying volume causing data loss.

What is the recommended course of action?

Error Logs

clickhouse Processing configuration file '/etc/clickhouse-server/config.xml'.
clickhouse Merging configuration file '/etc/clickhouse-server/conf.d/chop-generated-hostname-ports.xml'.
clickhouse Merging configuration file '/etc/clickhouse-server/conf.d/chop-generated-macros.xml'.
clickhouse Merging configuration file '/etc/clickhouse-server/config.d/01-clickhouse-01-listen.xml'.
clickhouse Merging configuration file '/etc/clickhouse-server/config.d/01-clickhouse-02-logger.xml'.
clickhouse Merging configuration file '/etc/clickhouse-server/config.d/01-clickhouse-03-query_log.xml'.
clickhouse Merging configuration file '/etc/clickhouse-server/config.d/01-clickhouse-04-part_log.xml'.
clickhouse Merging configuration file '/etc/clickhouse-server/config.d/01-clickhouse-05-trace_log.xml'.
clickhouse Merging configuration file '/etc/clickhouse-server/config.d/chop-generated-remote_servers.xml'.
clickhouse Merging configuration file '/etc/clickhouse-server/config.d/disks.xml'.
clickhouse Merging configuration file '/etc/clickhouse-server/config.d/s3.xml'.
clickhouse Logging debug to /var/log/clickhouse-server/clickhouse-server.log
clickhouse std::exception. Code: 1001, type: std::__1::__fs::filesystem::filesystem_error, e.what() = filesystem error: in weakly_canonical: Input/output error ["/var/log/clickhouse-server/clickhouse-server.log"], Stack trace (when copying this message, always include the lines below):
clickhouse 
clickhouse 0. std::system_error::system_error(std::error_code, String const&) @ 0x0000000016dc7975
clickhouse 1. std::__fs::filesystem::filesystem_error::filesystem_error[abi:v15000](String const&, std::__fs::filesystem::path const&, std::error_code) @ 0x000000000c8e9301
clickhouse 2. void std::__fs::filesystem::__throw_filesystem_error[abi:v15000]<String&, std::__fs::filesystem::path const&, std::error_code const&>(String&, std::__fs::filesystem::path const&, std::error_code const&) @ 0x0000000016d7e65a
clickhouse 3. std::__fs::filesystem::detail::(anonymous namespace)::ErrorHandler<std::__fs::filesystem::path>::report(std::error_code const&) const (.llvm.1044553589876245905) @ 0x0000000016d7fc63
clickhouse 4. std::__fs::filesystem::__weakly_canonical(std::__fs::filesystem::path const&, std::error_code*) @ 0x0000000016d85b68
clickhouse 5. Loggers::buildLoggers(Poco::Util::AbstractConfiguration&, Poco::Logger&, String const&) @ 0x000000000cb57efe
clickhouse 6. BaseDaemon::initialize(Poco::Util::Application&) @ 0x000000000cb3900f
clickhouse 7. DB::Server::initialize(Poco::Util::Application&) @ 0x000000000ca16698
clickhouse 8. Poco::Util::Application::run() @ 0x000000001498923a
clickhouse 9. DB::Server::run() @ 0x000000000ca16451
clickhouse 10. Poco::Util::ServerApplication::run(int, char**) @ 0x00000000149920f9
clickhouse 11. mainEntryClickHouseServer(int, char**) @ 0x000000000ca1224a
clickhouse 12. main @ 0x000000000764f098
clickhouse 13. ? @ 0x00007f7e757c5083
clickhouse 14. _start @ 0x0000000005de8f2e
clickhouse 
clickhouse Cannot print extra info for Poco::Exception (version 24.3.7.30 (official build))
Stream closed EOF for clickhouse/chi-clickhouse-installation-simple-1-1-0 (clickhouse)
State

```txt Name: chi-clickhouse-installation-simple-1-1-0 Namespace: clickhouse Priority: 0 Service Account: default Node: ip-10-132-130-000.us-west-2.compute.internal/10.132.130.000 Start Time: Thu, 12 Sep 2024 16:38:00 -0500 Labels: apps.kubernetes.io/pod-index=0 clickhouse.altinity.com/app=chop clickhouse.altinity.com/chi=clickhouse-installation clickhouse.altinity.com/cluster=simple clickhouse.altinity.com/namespace=clickhouse clickhouse.altinity.com/ready=yes clickhouse.altinity.com/replica=1 clickhouse.altinity.com/shard=1 controller-revision-hash=chi-clickhouse-installation-simple-1-1-56cd87d759 statefulset.kubernetes.io/pod-name=chi-clickhouse-installation-simple-1-1-0 Annotations: Status: Running IP: 10.132.130.00 IPs: IP: 10.132.130.00 Controlled By: StatefulSet/chi-clickhouse-installation-simple-1-1 Containers: clickhouse: Container ID: containerd://5a3ab0c1331f8e1dc25c610edcccbcdc059822a49c80b47f90e882af9e046471 Image: clickhouse/clickhouse-server:24.3.7 Image ID: docker.io/clickhouse/clickhouse-server@sha256:e55fe12eb1964663d5f4cb8b633444b6dfd1233124bd3cbbf1daacfa815d3c59 Ports: 9000/TCP, 8123/TCP, 9009/TCP Host Ports: 0/TCP, 0/TCP, 0/TCP State: Running Started: Thu, 12 Sep 2024 20:39:14 -0500 Last State: Terminated Reason: Error Exit Code: 137 Started: Thu, 12 Sep 2024 20:31:59 -0500 Finished: Thu, 12 Sep 2024 20:33:58 -0500 Ready: False Restart Count: 56 Liveness: http-get http://:http/ping delay=60s timeout=1s period=3s #success=1 #failure=10 Readiness: http-get http://:http/ping delay=10s timeout=1s period=3s #success=1 #failure=3 Environment: CLICKHOUSE_ALWAYS_RUN_INITDB_SCRIPTS: true AWS_STS_REGIONAL_ENDPOINTS: regional AWS_DEFAULT_REGION: us-west-2 AWS_REGION: us-west-2 AWS_ROLE_ARN: arn:aws:iam::0000000000000:role/company-clickhouse-role-prod AWS_WEB_IDENTITY_TOKEN_FILE: /var/run/secrets/eks.amazonaws.com/serviceaccount/token Mounts: /docker-entrypoint-initdb.d from bootstrap-configmap-volume (rw) /etc/clickhouse-server/conf.d/ from chi-clickhouse-installation-deploy-confd-simple-1-1 (rw) /etc/clickhouse-server/config.d/ from chi-clickhouse-installation-common-configd (rw) /etc/clickhouse-server/users.d/ from chi-clickhouse-installation-common-usersd (rw) /var/lib/clickhouse from data-volume-template (rw) /var/log/clickhouse-server from log-volume-template (rw) /var/run/secrets/eks.amazonaws.com/serviceaccount from aws-iam-token (ro) /var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-8sp2v (ro) clickhouse-log: Container ID: containerd://dceeebea25731fe1a0bbba3b0ca2789a626bdaeba8e6ebf358fa6cb429cb9f51 Image: registry.access.redhat.com/ubi8/ubi-minimal:latest Image ID: registry.access.redhat.com/ubi8/ubi-minimal@sha256:a47c89f02b39a98290f88204ed3d162845db0a0c464b319c2596cfd1e94b444e Port: Host Port: Command: /bin/sh -c -- Args: while true; do sleep 30; done; State: Running Started: Thu, 12 Sep 2024 16:38:35 -0500 Ready: True Restart Count: 0 Environment: AWS_STS_REGIONAL_ENDPOINTS: regional AWS_DEFAULT_REGION: us-west-2 AWS_REGION: us-west-2 AWS_ROLE_ARN: arn:aws:iam::000000000000:role/company-clickhouse-role-prod AWS_WEB_IDENTITY_TOKEN_FILE: /var/run/secrets/eks.amazonaws.com/serviceaccount/token Mounts: /var/lib/clickhouse from data-volume-template (rw) /var/log/clickhouse-server from log-volume-template (rw) /var/run/secrets/eks.amazonaws.com/serviceaccount from aws-iam-token (ro) /var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-8sp2v (ro) Conditions: Type Status PodReadyToStartContainers True Initialized True Ready False ContainersReady False PodScheduled True Volumes: aws-iam-token: Type: Projected (a volume that contains injected data from multiple sources) TokenExpirationSeconds: 86400 data-volume-template: Type: PersistentVolumeClaim (a reference to a PersistentVolumeClaim in the same namespace) ClaimName: data-volume-template-chi-clickhouse-installation-simple-1-1-0 ReadOnly: false log-volume-template: Type: PersistentVolumeClaim (a reference to a PersistentVolumeClaim in the same namespace) ClaimName: log-volume-template-chi-clickhouse-installation-simple-1-1-0 ReadOnly: false bootstrap-configmap-volume: Type: ConfigMap (a volume populated by a ConfigMap) Name: bootstrap-configmap Optional: false chi-clickhouse-installation-common-configd: Type: ConfigMap (a volume populated by a ConfigMap) Name: chi-clickhouse-installation-common-configd Optional: false chi-clickhouse-installation-common-usersd: Type: ConfigMap (a volume populated by a ConfigMap) Name: chi-clickhouse-installation-common-usersd Optional: false chi-clickhouse-installation-deploy-confd-simple-1-1: Type: ConfigMap (a volume populated by a ConfigMap) Name: chi-clickhouse-installation-deploy-confd-simple-1-1 Optional: false kube-api-access-8sp2v: Type: Projected (a volume that contains injected data from multiple sources) TokenExpirationSeconds: 3607 ConfigMapName: kube-root-ca.crt ConfigMapOptional: DownwardAPI: true QoS Class: BestEffort Node-Selectors: clickhouse-installation=true Tolerations: installation=clickhouse-installation:NoSchedule node.kubernetes.io/not-ready:NoExecute op=Exists for 300s node.kubernetes.io/unreachable:NoExecute op=Exists for 300s Events: Type Reason Age From Message ---- ------ ---- ---- ------- Warning Unhealthy 7m26s (x2104 over 4h1m) kubelet Readiness probe failed: Get "http://10.132.130.00:8123/ping": dial tcp 10.132.130.00:8123: connect: connection refused Warning BackOff 2m20s (x574 over 3h48m) kubelet Back-off restarting failed container clickhouse in pod chi-clickhouse-installation-simple-1-1-0_clickhouse(510d3812-9037-4452-adc9-417531893542) ```

Slach commented 1 month ago

could you share kubectl get chi -n clickhouse clickhouse-installation -o yaml?

fidiego commented 1 month ago

@Slach sure thing.

kubectl get chi -n clickhouse clickhouse-installation -o yaml

```yaml apiVersion: clickhouse.altinity.com/v1 kind: ClickHouseInstallation metadata: annotations: kubectl.kubernetes.io/last-applied-configuration: | {"apiVersion":"clickhouse.altinity.com/v1","kind":"ClickHouseInstallation","metadata":{"annotations":{},"name":"clickhouse-installation","namespace":"clickhouse"},"spec":{"configuration":{"clusters":[{"layout":{"replicasCount":3,"shardsCount":3},"name":"simple","templates":{"podTemplate":"clickhouse:24.3.7","serviceTemplate":"clickhouse:24.3.7"}}],"files":{"config.d/disks.xml":"\u003cclickhouse\u003e\n \u003cstorage_configuration\u003e\n \u003cdisks\u003e\n \u003cs3_disk\u003e\n \u003ctype\u003es3\u003c/type\u003e\n \u003cendpoint\u003ehttps://company-clickhouse-env.s3.amazonaws.com/tables/\u003c/endpoint\u003e\n \u003cuse_environment_credentials\u003etrue\u003c/use_environment_credentials\u003e\n \u003cmetadata_path\u003e/var/lib/clickhouse/disks/s3_disk/\u003c/metadata_path\u003e\n \u003c/s3_disk\u003e\n \u003cs3_cache\u003e\n \u003ctype\u003ecache\u003c/type\u003e\n \u003cdisk\u003es3_disk\u003c/disk\u003e\n \u003cpath\u003e/var/lib/clickhouse/disks/s3_cache/\u003c/path\u003e\n \u003cmax_size\u003e10Gi\u003c/max_size\u003e\n \u003c/s3_cache\u003e\n \u003c/disks\u003e\n \u003cpolicies\u003e\n \u003cs3_main\u003e\n \u003cvolumes\u003e\n \u003cmain\u003e\n \u003cdisk\u003es3_disk\u003c/disk\u003e\n \u003c/main\u003e\n \u003c/volumes\u003e\n \u003c/s3_main\u003e\n \u003c/policies\u003e\n \u003c/storage_configuration\u003e\n\u003c/clickhouse\u003e\n","config.d/s3.xml":"\u003cclickhouse\u003e\n \u003cs3\u003e\n \u003cuse_environment_credentials\u003etrue\u003c/use_environment_credentials\u003e\n \u003c/s3\u003e\n\u003c/clickhouse\u003e\n"},"users":{"username/networks/ip":["0.0.0.0/0"],"username/password_sha256_hex":"xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx"}},"defaults":{"templates":{"dataVolumeClaimTemplate":"data-volume-template","logVolumeClaimTemplate":"log-volume-template","serviceTemplate":"clickhouse:24.3.7"}},"templates":{"podTemplates":[{"name":"clickhouse:24.3.7","spec":{"containers":[{"env":[{"name":"CLICKHOUSE_ALWAYS_RUN_INITDB_SCRIPTS","value":"true"}],"image":"clickhouse/clickhouse-server:24.3.7","name":"clickhouse","volumeMounts":[{"mountPath":"/var/lib/clickhouse","name":"data-volume-template"},{"mountPath":"/var/log/clickhouse-server","name":"log-volume-template"},{"mountPath":"/docker-entrypoint-initdb.d","name":"bootstrap-configmap-volume"}]}],"nodeSelector":{"clickhouse-installation":"true"},"tolerations":[{"effect":"NoSchedule","key":"installation","operator":"Equal","value":"clickhouse-installation"}],"volumes":[{"configMap":{"name":"bootstrap-configmap"},"name":"bootstrap-configmap-volume"}]}}],"serviceTemplates":[{"metadata":{"annotations":{"external-dns.alpha.kubernetes.io/internal-hostname":"clickhouse.company.us-west-2.env.company.cloud","external-dns.alpha.kubernetes.io/ttl":"60"}},"name":"clickhouse:24.3.7","spec":{"ports":[{"name":"http","port":8123},{"name":"client","port":9000}]}}],"volumeClaimTemplates":[{"name":"data-volume-template","spec":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"100Gi"}}}},{"name":"log-volume-template","spec":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"4Gi"}}}}]}}} creationTimestamp: "2024-08-21T13:49:55Z" finalizers: - finalizer.clickhouseinstallation.altinity.com generation: 1 name: clickhouse-installation namespace: clickhouse resourceVersion: "415622675" uid: 75211649-036d-4cb3-a535-d9a4bab082c0 spec: configuration: clusters: - layout: replicasCount: 3 shardsCount: 3 name: simple templates: podTemplate: clickhouse:24.3.7 serviceTemplate: clickhouse:24.3.7 files: config.d/disks.xml: | s3 https://company-clickhouse-env.s3.amazonaws.com/tables/ true /var/lib/clickhouse/disks/s3_disk/ cache s3_disk /var/lib/clickhouse/disks/s3_cache/ 10Gi

s3_disk
config.d/s3.xml: | true users: username/networks/ip: - 0.0.0.0/0 username/password_sha256_hex: xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx defaults: templates: dataVolumeClaimTemplate: data-volume-template logVolumeClaimTemplate: log-volume-template serviceTemplate: clickhouse:24.3.7 templates: podTemplates: - name: clickhouse:24.3.7 spec: containers: - env: - name: CLICKHOUSE_ALWAYS_RUN_INITDB_SCRIPTS value: "true" image: clickhouse/clickhouse-server:24.3.7 name: clickhouse volumeMounts: - mountPath: /var/lib/clickhouse name: data-volume-template - mountPath: /var/log/clickhouse-server name: log-volume-template - mountPath: /docker-entrypoint-initdb.d name: bootstrap-configmap-volume nodeSelector: clickhouse-installation: "true" tolerations: - effect: NoSchedule key: installation operator: Equal value: clickhouse-installation volumes: - configMap: name: bootstrap-configmap name: bootstrap-configmap-volume serviceTemplates: - metadata: annotations: external-dns.alpha.kubernetes.io/internal-hostname: clickhouse.company.us-west-2.env.company.cloud external-dns.alpha.kubernetes.io/ttl: "60" name: clickhouse:24.3.7 spec: ports: - name: http port: 8123 - name: client port: 9000 volumeClaimTemplates: - name: data-volume-template spec: accessModes: - ReadWriteOnce resources: requests: storage: 100Gi - name: log-volume-template spec: accessModes: - ReadWriteOnce resources: requests: storage: 4Gi status: action: Update ConfigMap clickhouse/chi-clickhouse-installation-common-usersd actions: - '2024-09-13T12:35:44.391764656Z reconcile completed successfully, task id: 520089d3-fa7e-4d8d-932c-2f5dafc1b21b' - '2024-09-13T12:07:33.838481812Z reconcile started, task id: 520089d3-fa7e-4d8d-932c-2f5dafc1b21b' - '2024-09-13T03:45:34.579600835Z reconcile completed successfully, task id: f5a11dd9-268b-4c9a-ab70-9c9a92de40b8' - '2024-09-13T03:31:05.141469604Z reconcile started, task id: f5a11dd9-268b-4c9a-ab70-9c9a92de40b8' - '2024-09-12T12:37:54.81024906Z reconcile completed successfully, task id: 49d76b1f-d2ff-43f3-b016-355f021bc573' - '2024-09-12T12:21:41.717022529Z reconcile started, task id: 49d76b1f-d2ff-43f3-b016-355f021bc573' - '2024-09-12T12:07:27.264047468Z reconcile started, task id: 2acada9a-e3ba-4234-ac65-94d16d402ba3' - '2024-09-12T10:47:40.524694896Z reconcile completed successfully, task id: 379fea7c-8cb8-4757-ab0c-ba130b30ea04' - '2024-09-12T09:44:30.147849804Z reconcile started, task id: 379fea7c-8cb8-4757-ab0c-ba130b30ea04' - '2024-09-12T02:09:56.90045236Z reconcile completed successfully, task id: 28e6faf5-a626-4822-878f-ebda6950bf31' chop-commit: 2dd9eca chop-date: 2024-08-12T14:15:16 chop-ip: 10.132.130.39 chop-version: 0.23.7 clusters: 1 endpoint: clickhouse-clickhouse-installation.clickhouse.svc.cluster.local errors: - '2024-09-13T13:04:45.448738646Z FAILED to reconcile ConfigMap: chi-clickhouse-installation-common-usersd CHI: clickhouse-installation ' - '2024-09-13T13:04:43.452622218Z Update ConfigMap clickhouse/chi-clickhouse-installation-common-usersd failed with error client rate limiter Wait returned an error: context canceled' - '2024-09-13T12:44:43.6014188Z FAILED to reconcile ConfigMap: chi-clickhouse-installation-common-usersd CHI: clickhouse-installation ' - '2024-09-13T12:44:41.60058812Z Update ConfigMap clickhouse/chi-clickhouse-installation-common-usersd failed with error client rate limiter Wait returned an error: context canceled' - '2024-09-13T12:18:01.035018186Z Update Service: clickhouse/clickhouse-clickhouse-installation failed with error: just recreate the service in case of service type change ''ClusterIP''=>''''' - '2024-09-13T09:34:58.265686457Z FAILED to reconcile ConfigMap: chi-clickhouse-installation-common-usersd CHI: clickhouse-installation ' - '2024-09-13T09:34:56.26912994Z Update ConfigMap clickhouse/chi-clickhouse-installation-common-usersd failed with error client rate limiter Wait returned an error: context canceled' - '2024-09-13T09:34:52.468149832Z FAILED to reconcile ConfigMap: chi-clickhouse-installation-common-usersd CHI: clickhouse-installation ' - '2024-09-13T09:34:50.470358556Z Update ConfigMap clickhouse/chi-clickhouse-installation-common-usersd failed with error client rate limiter Wait returned an error: context canceled' - '2024-09-13T04:36:35.263627446Z FAILED to reconcile ConfigMap: chi-clickhouse-installation-common-usersd CHI: clickhouse-installation ' - '2024-09-13T04:36:33.263382059Z Update ConfigMap clickhouse/chi-clickhouse-installation-common-usersd failed with error client rate limiter Wait returned an error: context canceled' - '2024-09-13T03:32:27.328779347Z Update Service: clickhouse/clickhouse-clickhouse-installation failed with error: just recreate the service in case of service type change ''ClusterIP''=>''''' - '2024-09-12T14:03:21.848868176Z FAILED to reconcile ConfigMap: chi-clickhouse-installation-common-usersd CHI: clickhouse-installation ' - '2024-09-12T14:03:19.848689143Z Update ConfigMap clickhouse/chi-clickhouse-installation-common-usersd failed with error client rate limiter Wait returned an error: context canceled' fqdns: - chi-clickhouse-installation-simple-0-0.clickhouse.svc.cluster.local - chi-clickhouse-installation-simple-0-1.clickhouse.svc.cluster.local - chi-clickhouse-installation-simple-0-2.clickhouse.svc.cluster.local - chi-clickhouse-installation-simple-1-0.clickhouse.svc.cluster.local - chi-clickhouse-installation-simple-1-1.clickhouse.svc.cluster.local - chi-clickhouse-installation-simple-1-2.clickhouse.svc.cluster.local - chi-clickhouse-installation-simple-2-0.clickhouse.svc.cluster.local - chi-clickhouse-installation-simple-2-1.clickhouse.svc.cluster.local - chi-clickhouse-installation-simple-2-2.clickhouse.svc.cluster.local hosts: 9 hostsWithTablesCreated: - chi-clickhouse-installation-simple-0-0.clickhouse.svc.cluster.local - chi-clickhouse-installation-simple-0-1.clickhouse.svc.cluster.local - chi-clickhouse-installation-simple-0-2.clickhouse.svc.cluster.local - chi-clickhouse-installation-simple-2-0.clickhouse.svc.cluster.local - chi-clickhouse-installation-simple-1-0.clickhouse.svc.cluster.local - chi-clickhouse-installation-simple-2-1.clickhouse.svc.cluster.local - chi-clickhouse-installation-simple-1-1.clickhouse.svc.cluster.local - chi-clickhouse-installation-simple-2-2.clickhouse.svc.cluster.local - chi-clickhouse-installation-simple-1-2.clickhouse.svc.cluster.local normalized: apiVersion: clickhouse.altinity.com/v1 kind: ClickHouseInstallation metadata: creationTimestamp: "2024-08-21T13:49:55Z" finalizers: - finalizer.clickhouseinstallation.altinity.com generation: 1 name: clickhouse-installation namespace: clickhouse resourceVersion: "415622632" uid: 75211649-036d-4cb3-a535-d9a4bab082c0 spec: configuration: clusters: - layout: replicas: - name: "0" shards: - httpPort: 8123 interserverHTTPPort: 9009 name: 0-0 tcpPort: 9000 templates: dataVolumeClaimTemplate: data-volume-template logVolumeClaimTemplate: log-volume-template podTemplate: clickhouse:24.3.7 serviceTemplate: clickhouse:24.3.7 - httpPort: 8123 interserverHTTPPort: 9009 name: 1-0 tcpPort: 9000 templates: dataVolumeClaimTemplate: data-volume-template logVolumeClaimTemplate: log-volume-template podTemplate: clickhouse:24.3.7 serviceTemplate: clickhouse:24.3.7 - httpPort: 8123 interserverHTTPPort: 9009 name: 2-0 tcpPort: 9000 templates: dataVolumeClaimTemplate: data-volume-template logVolumeClaimTemplate: log-volume-template podTemplate: clickhouse:24.3.7 serviceTemplate: clickhouse:24.3.7 shardsCount: 3 templates: dataVolumeClaimTemplate: data-volume-template logVolumeClaimTemplate: log-volume-template podTemplate: clickhouse:24.3.7 serviceTemplate: clickhouse:24.3.7 - name: "1" shards: - httpPort: 8123 interserverHTTPPort: 9009 name: 0-1 tcpPort: 9000 templates: dataVolumeClaimTemplate: data-volume-template logVolumeClaimTemplate: log-volume-template podTemplate: clickhouse:24.3.7 serviceTemplate: clickhouse:24.3.7 - httpPort: 8123 interserverHTTPPort: 9009 name: 1-1 tcpPort: 9000 templates: dataVolumeClaimTemplate: data-volume-template logVolumeClaimTemplate: log-volume-template podTemplate: clickhouse:24.3.7 serviceTemplate: clickhouse:24.3.7 - httpPort: 8123 interserverHTTPPort: 9009 name: 2-1 tcpPort: 9000 templates: dataVolumeClaimTemplate: data-volume-template logVolumeClaimTemplate: log-volume-template podTemplate: clickhouse:24.3.7 serviceTemplate: clickhouse:24.3.7 shardsCount: 3 templates: dataVolumeClaimTemplate: data-volume-template logVolumeClaimTemplate: log-volume-template podTemplate: clickhouse:24.3.7 serviceTemplate: clickhouse:24.3.7 - name: "2" shards: - httpPort: 8123 interserverHTTPPort: 9009 name: 0-2 tcpPort: 9000 templates: dataVolumeClaimTemplate: data-volume-template logVolumeClaimTemplate: log-volume-template podTemplate: clickhouse:24.3.7 serviceTemplate: clickhouse:24.3.7 - httpPort: 8123 interserverHTTPPort: 9009 name: 1-2 tcpPort: 9000 templates: dataVolumeClaimTemplate: data-volume-template logVolumeClaimTemplate: log-volume-template podTemplate: clickhouse:24.3.7 serviceTemplate: clickhouse:24.3.7 - httpPort: 8123 interserverHTTPPort: 9009 name: 2-2 tcpPort: 9000 templates: dataVolumeClaimTemplate: data-volume-template logVolumeClaimTemplate: log-volume-template podTemplate: clickhouse:24.3.7 serviceTemplate: clickhouse:24.3.7 shardsCount: 3 templates: dataVolumeClaimTemplate: data-volume-template logVolumeClaimTemplate: log-volume-template podTemplate: clickhouse:24.3.7 serviceTemplate: clickhouse:24.3.7 replicasCount: 3 shards: - internalReplication: "True" name: "0" replicas: - httpPort: 8123 interserverHTTPPort: 9009 name: 0-0 tcpPort: 9000 templates: dataVolumeClaimTemplate: data-volume-template logVolumeClaimTemplate: log-volume-template podTemplate: clickhouse:24.3.7 serviceTemplate: clickhouse:24.3.7 - httpPort: 8123 interserverHTTPPort: 9009 name: 0-1 tcpPort: 9000 templates: dataVolumeClaimTemplate: data-volume-template logVolumeClaimTemplate: log-volume-template podTemplate: clickhouse:24.3.7 serviceTemplate: clickhouse:24.3.7 - httpPort: 8123 interserverHTTPPort: 9009 name: 0-2 tcpPort: 9000 templates: dataVolumeClaimTemplate: data-volume-template logVolumeClaimTemplate: log-volume-template podTemplate: clickhouse:24.3.7 serviceTemplate: clickhouse:24.3.7 replicasCount: 3 templates: dataVolumeClaimTemplate: data-volume-template logVolumeClaimTemplate: log-volume-template podTemplate: clickhouse:24.3.7 serviceTemplate: clickhouse:24.3.7 - internalReplication: "True" name: "1" replicas: - httpPort: 8123 interserverHTTPPort: 9009 name: 1-0 tcpPort: 9000 templates: dataVolumeClaimTemplate: data-volume-template logVolumeClaimTemplate: log-volume-template podTemplate: clickhouse:24.3.7 serviceTemplate: clickhouse:24.3.7 - httpPort: 8123 interserverHTTPPort: 9009 name: 1-1 tcpPort: 9000 templates: dataVolumeClaimTemplate: data-volume-template logVolumeClaimTemplate: log-volume-template podTemplate: clickhouse:24.3.7 serviceTemplate: clickhouse:24.3.7 - httpPort: 8123 interserverHTTPPort: 9009 name: 1-2 tcpPort: 9000 templates: dataVolumeClaimTemplate: data-volume-template logVolumeClaimTemplate: log-volume-template podTemplate: clickhouse:24.3.7 serviceTemplate: clickhouse:24.3.7 replicasCount: 3 templates: dataVolumeClaimTemplate: data-volume-template logVolumeClaimTemplate: log-volume-template podTemplate: clickhouse:24.3.7 serviceTemplate: clickhouse:24.3.7 - internalReplication: "True" name: "2" replicas: - httpPort: 8123 interserverHTTPPort: 9009 name: 2-0 tcpPort: 9000 templates: dataVolumeClaimTemplate: data-volume-template logVolumeClaimTemplate: log-volume-template podTemplate: clickhouse:24.3.7 serviceTemplate: clickhouse:24.3.7 - httpPort: 8123 interserverHTTPPort: 9009 name: 2-1 tcpPort: 9000 templates: dataVolumeClaimTemplate: data-volume-template logVolumeClaimTemplate: log-volume-template podTemplate: clickhouse:24.3.7 serviceTemplate: clickhouse:24.3.7 - httpPort: 8123 interserverHTTPPort: 9009 name: 2-2 tcpPort: 9000 templates: dataVolumeClaimTemplate: data-volume-template logVolumeClaimTemplate: log-volume-template podTemplate: clickhouse:24.3.7 serviceTemplate: clickhouse:24.3.7 replicasCount: 3 templates: dataVolumeClaimTemplate: data-volume-template logVolumeClaimTemplate: log-volume-template podTemplate: clickhouse:24.3.7 serviceTemplate: clickhouse:24.3.7 shardsCount: 3 name: simple schemaPolicy: replica: All shard: All templates: dataVolumeClaimTemplate: data-volume-template logVolumeClaimTemplate: log-volume-template podTemplate: clickhouse:24.3.7 serviceTemplate: clickhouse:24.3.7 files: config.d/disks.xml: | s3 https://company-clickhouse-env.s3.amazonaws.com/tables/ true /var/lib/clickhouse/disks/s3_disk/ cache s3_disk /var/lib/clickhouse/disks/s3_cache/ 10Gi
s3_disk
config.d/s3.xml: | true users: clickhouse_operator/networks/ip: - 10.132.130.39 clickhouse_operator/password_sha256_hex: 716b36073a90c6fe1d445ac1af85f4777c5b7a155cea359961826a030513e448 clickhouse_operator/profile: clickhouse_operator default/networks/host_regexp: (chi-clickhouse-installation-[^.]+\d+-\d+|clickhouse\-clickhouse-installation)\.clickhouse\.svc\.cluster\.local$ default/networks/ip: - ::1 - 127.0.0.1 - 10.132.130.206 - 10.132.130.78 - 10.132.130.47 - 10.132.130.108 - 10.132.130.241 - 10.132.130.26 - 10.132.130.249 - 10.132.130.231 - 10.132.130.150 default/profile: default default/quota: default username/networks/host_regexp: (chi-clickhouse-installation-[^.]+\d+-\d+|clickhouse\-clickhouse-installation)\.clickhouse\.svc\.cluster\.local$ username/networks/ip: - ::1 - 127.0.0.1 - 0.0.0.0/0 username/password_sha256_hex: xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx username/profile: default username/quota: default defaults: replicasUseFQDN: "False" storageManagement: {} templates: dataVolumeClaimTemplate: data-volume-template logVolumeClaimTemplate: log-volume-template serviceTemplate: clickhouse:24.3.7 reconciling: cleanup: reconcileFailedObjects: configMap: Retain pvc: Retain service: Retain statefulSet: Retain unknownObjects: configMap: Delete pvc: Delete service: Delete statefulSet: Delete configMapPropagationTimeout: 10 policy: unspecified stop: "False" taskID: de831f6d-5d63-4370-be33-77dd5fac449e templates: PodTemplatesIndex: {} ServiceTemplatesIndex: {} VolumeClaimTemplatesIndex: {} podTemplates: - metadata: creationTimestamp: null name: clickhouse:24.3.7 spec: containers: - env: - name: CLICKHOUSE_ALWAYS_RUN_INITDB_SCRIPTS value: "true" image: clickhouse/clickhouse-server:24.3.7 name: clickhouse resources: {} volumeMounts: - mountPath: /var/lib/clickhouse name: data-volume-template - mountPath: /var/log/clickhouse-server name: log-volume-template - mountPath: /docker-entrypoint-initdb.d name: bootstrap-configmap-volume nodeSelector: clickhouse-installation: "true" tolerations: - effect: NoSchedule key: installation operator: Equal value: clickhouse-installation volumes: - configMap: name: bootstrap-configmap name: bootstrap-configmap-volume zone: {} serviceTemplates: - metadata: annotations: external-dns.alpha.kubernetes.io/internal-hostname: clickhouse.company.us-west-2.env.company.cloud external-dns.alpha.kubernetes.io/ttl: "60" creationTimestamp: null name: clickhouse:24.3.7 spec: ports: - name: http port: 8123 targetPort: 0 - name: client port: 9000 targetPort: 0 volumeClaimTemplates: - metadata: creationTimestamp: null name: data-volume-template spec: accessModes: - ReadWriteOnce resources: requests: storage: 100Gi - metadata: creationTimestamp: null name: log-volume-template spec: accessModes: - ReadWriteOnce resources: requests: storage: 4Gi templating: policy: manual troubleshoot: "False" normalizedCompleted: apiVersion: clickhouse.altinity.com/v1 kind: ClickHouseInstallation metadata: creationTimestamp: "2024-08-21T13:49:55Z" finalizers: - finalizer.clickhouseinstallation.altinity.com generation: 1 name: clickhouse-installation namespace: clickhouse resourceVersion: "415591205" uid: 75211649-036d-4cb3-a535-d9a4bab082c0 spec: configuration: clusters: - layout: replicas: - name: "0" shards: - httpPort: 8123 interserverHTTPPort: 9009 name: 0-0 tcpPort: 9000 templates: dataVolumeClaimTemplate: data-volume-template logVolumeClaimTemplate: log-volume-template podTemplate: clickhouse:24.3.7 serviceTemplate: clickhouse:24.3.7 - httpPort: 8123 interserverHTTPPort: 9009 name: 1-0 tcpPort: 9000 templates: dataVolumeClaimTemplate: data-volume-template logVolumeClaimTemplate: log-volume-template podTemplate: clickhouse:24.3.7 serviceTemplate: clickhouse:24.3.7 - httpPort: 8123 interserverHTTPPort: 9009 name: 2-0 tcpPort: 9000 templates: dataVolumeClaimTemplate: data-volume-template logVolumeClaimTemplate: log-volume-template podTemplate: clickhouse:24.3.7 serviceTemplate: clickhouse:24.3.7 shardsCount: 3 templates: dataVolumeClaimTemplate: data-volume-template logVolumeClaimTemplate: log-volume-template podTemplate: clickhouse:24.3.7 serviceTemplate: clickhouse:24.3.7 - name: "1" shards: - httpPort: 8123 interserverHTTPPort: 9009 name: 0-1 tcpPort: 9000 templates: dataVolumeClaimTemplate: data-volume-template logVolumeClaimTemplate: log-volume-template podTemplate: clickhouse:24.3.7 serviceTemplate: clickhouse:24.3.7 - httpPort: 8123 interserverHTTPPort: 9009 name: 1-1 tcpPort: 9000 templates: dataVolumeClaimTemplate: data-volume-template logVolumeClaimTemplate: log-volume-template podTemplate: clickhouse:24.3.7 serviceTemplate: clickhouse:24.3.7 - httpPort: 8123 interserverHTTPPort: 9009 name: 2-1 tcpPort: 9000 templates: dataVolumeClaimTemplate: data-volume-template logVolumeClaimTemplate: log-volume-template podTemplate: clickhouse:24.3.7 serviceTemplate: clickhouse:24.3.7 shardsCount: 3 templates: dataVolumeClaimTemplate: data-volume-template logVolumeClaimTemplate: log-volume-template podTemplate: clickhouse:24.3.7 serviceTemplate: clickhouse:24.3.7 - name: "2" shards: - httpPort: 8123 interserverHTTPPort: 9009 name: 0-2 tcpPort: 9000 templates: dataVolumeClaimTemplate: data-volume-template logVolumeClaimTemplate: log-volume-template podTemplate: clickhouse:24.3.7 serviceTemplate: clickhouse:24.3.7 - httpPort: 8123 interserverHTTPPort: 9009 name: 1-2 tcpPort: 9000 templates: dataVolumeClaimTemplate: data-volume-template logVolumeClaimTemplate: log-volume-template podTemplate: clickhouse:24.3.7 serviceTemplate: clickhouse:24.3.7 - httpPort: 8123 interserverHTTPPort: 9009 name: 2-2 tcpPort: 9000 templates: dataVolumeClaimTemplate: data-volume-template logVolumeClaimTemplate: log-volume-template podTemplate: clickhouse:24.3.7 serviceTemplate: clickhouse:24.3.7 shardsCount: 3 templates: dataVolumeClaimTemplate: data-volume-template logVolumeClaimTemplate: log-volume-template podTemplate: clickhouse:24.3.7 serviceTemplate: clickhouse:24.3.7 replicasCount: 3 shards: - internalReplication: "True" name: "0" replicas: - httpPort: 8123 interserverHTTPPort: 9009 name: 0-0 tcpPort: 9000 templates: dataVolumeClaimTemplate: data-volume-template logVolumeClaimTemplate: log-volume-template podTemplate: clickhouse:24.3.7 serviceTemplate: clickhouse:24.3.7 - httpPort: 8123 interserverHTTPPort: 9009 name: 0-1 tcpPort: 9000 templates: dataVolumeClaimTemplate: data-volume-template logVolumeClaimTemplate: log-volume-template podTemplate: clickhouse:24.3.7 serviceTemplate: clickhouse:24.3.7 - httpPort: 8123 interserverHTTPPort: 9009 name: 0-2 tcpPort: 9000 templates: dataVolumeClaimTemplate: data-volume-template logVolumeClaimTemplate: log-volume-template podTemplate: clickhouse:24.3.7 serviceTemplate: clickhouse:24.3.7 replicasCount: 3 templates: dataVolumeClaimTemplate: data-volume-template logVolumeClaimTemplate: log-volume-template podTemplate: clickhouse:24.3.7 serviceTemplate: clickhouse:24.3.7 - internalReplication: "True" name: "1" replicas: - httpPort: 8123 interserverHTTPPort: 9009 name: 1-0 tcpPort: 9000 templates: dataVolumeClaimTemplate: data-volume-template logVolumeClaimTemplate: log-volume-template podTemplate: clickhouse:24.3.7 serviceTemplate: clickhouse:24.3.7 - httpPort: 8123 interserverHTTPPort: 9009 name: 1-1 tcpPort: 9000 templates: dataVolumeClaimTemplate: data-volume-template logVolumeClaimTemplate: log-volume-template podTemplate: clickhouse:24.3.7 serviceTemplate: clickhouse:24.3.7 - httpPort: 8123 interserverHTTPPort: 9009 name: 1-2 tcpPort: 9000 templates: dataVolumeClaimTemplate: data-volume-template logVolumeClaimTemplate: log-volume-template podTemplate: clickhouse:24.3.7 serviceTemplate: clickhouse:24.3.7 replicasCount: 3 templates: dataVolumeClaimTemplate: data-volume-template logVolumeClaimTemplate: log-volume-template podTemplate: clickhouse:24.3.7 serviceTemplate: clickhouse:24.3.7 - internalReplication: "True" name: "2" replicas: - httpPort: 8123 interserverHTTPPort: 9009 name: 2-0 tcpPort: 9000 templates: dataVolumeClaimTemplate: data-volume-template logVolumeClaimTemplate: log-volume-template podTemplate: clickhouse:24.3.7 serviceTemplate: clickhouse:24.3.7 - httpPort: 8123 interserverHTTPPort: 9009 name: 2-1 tcpPort: 9000 templates: dataVolumeClaimTemplate: data-volume-template logVolumeClaimTemplate: log-volume-template podTemplate: clickhouse:24.3.7 serviceTemplate: clickhouse:24.3.7 - httpPort: 8123 interserverHTTPPort: 9009 name: 2-2 tcpPort: 9000 templates: dataVolumeClaimTemplate: data-volume-template logVolumeClaimTemplate: log-volume-template podTemplate: clickhouse:24.3.7 serviceTemplate: clickhouse:24.3.7 replicasCount: 3 templates: dataVolumeClaimTemplate: data-volume-template logVolumeClaimTemplate: log-volume-template podTemplate: clickhouse:24.3.7 serviceTemplate: clickhouse:24.3.7 shardsCount: 3 name: simple schemaPolicy: replica: All shard: All templates: dataVolumeClaimTemplate: data-volume-template logVolumeClaimTemplate: log-volume-template podTemplate: clickhouse:24.3.7 serviceTemplate: clickhouse:24.3.7 files: config.d/disks.xml: | s3 https://company-clickhouse-env.s3.amazonaws.com/tables/ true /var/lib/clickhouse/disks/s3_disk/ cache s3_disk /var/lib/clickhouse/disks/s3_cache/ 10Gi
s3_disk
config.d/s3.xml: | true users: clickhouse_operator/networks/ip: - 10.132.130.39 clickhouse_operator/password_sha256_hex: 716b36073a90c6fe1d445ac1af85f4777c5b7a155cea359961826a030513e448 clickhouse_operator/profile: clickhouse_operator default/networks/host_regexp: (chi-clickhouse-installation-[^.]+\d+-\d+|clickhouse\-clickhouse-installation)\.clickhouse\.svc\.cluster\.local$ default/networks/ip: - ::1 - 127.0.0.1 - 10.132.130.22 - 10.132.130.78 - 10.132.130.230 - 10.132.130.128 - 10.132.130.139 - 10.132.130.231 - 10.132.130.150 - 10.132.130.108 - 10.132.130.47 default/profile: default default/quota: default username/networks/host_regexp: (chi-clickhouse-installation-[^.]+\d+-\d+|clickhouse\-clickhouse-installation)\.clickhouse\.svc\.cluster\.local$ username/networks/ip: - ::1 - 127.0.0.1 - 0.0.0.0/0 username/password_sha256_hex: xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx username/profile: default username/quota: default defaults: replicasUseFQDN: "False" storageManagement: {} templates: dataVolumeClaimTemplate: data-volume-template logVolumeClaimTemplate: log-volume-template serviceTemplate: clickhouse:24.3.7 reconciling: cleanup: reconcileFailedObjects: configMap: Retain pvc: Retain service: Retain statefulSet: Retain unknownObjects: configMap: Delete pvc: Delete service: Delete statefulSet: Delete configMapPropagationTimeout: 10 policy: unspecified stop: "False" taskID: 5e9a0378-fa89-4029-8dfe-fcffc4c37bae templates: PodTemplatesIndex: {} ServiceTemplatesIndex: {} VolumeClaimTemplatesIndex: {} podTemplates: - metadata: creationTimestamp: null name: clickhouse:24.3.7 spec: containers: - env: - name: CLICKHOUSE_ALWAYS_RUN_INITDB_SCRIPTS value: "true" image: clickhouse/clickhouse-server:24.3.7 name: clickhouse resources: {} volumeMounts: - mountPath: /var/lib/clickhouse name: data-volume-template - mountPath: /var/log/clickhouse-server name: log-volume-template - mountPath: /docker-entrypoint-initdb.d name: bootstrap-configmap-volume nodeSelector: clickhouse-installation: "true" tolerations: - effect: NoSchedule key: installation operator: Equal value: clickhouse-installation volumes: - configMap: name: bootstrap-configmap name: bootstrap-configmap-volume zone: {} serviceTemplates: - metadata: annotations: external-dns.alpha.kubernetes.io/internal-hostname: clickhouse.company.us-west-2.env.company.cloud external-dns.alpha.kubernetes.io/ttl: "60" creationTimestamp: null name: clickhouse:24.3.7 spec: ports: - name: http port: 8123 targetPort: 0 - name: client port: 9000 targetPort: 0 volumeClaimTemplates: - metadata: creationTimestamp: null name: data-volume-template spec: accessModes: - ReadWriteOnce resources: requests: storage: 100Gi - metadata: creationTimestamp: null name: log-volume-template spec: accessModes: - ReadWriteOnce resources: requests: storage: 4Gi templating: policy: manual troubleshoot: "False" pod-ips: - 10.132.130.206 - 10.132.130.78 - 10.132.130.47 - 10.132.130.108 - 10.132.130.241 - 10.132.130.26 - 10.132.130.249 - 10.132.130.231 - 10.132.130.150 pods: - chi-clickhouse-installation-simple-0-0-0 - chi-clickhouse-installation-simple-0-1-0 - chi-clickhouse-installation-simple-0-2-0 - chi-clickhouse-installation-simple-1-0-0 - chi-clickhouse-installation-simple-1-1-0 - chi-clickhouse-installation-simple-1-2-0 - chi-clickhouse-installation-simple-2-0-0 - chi-clickhouse-installation-simple-2-1-0 - chi-clickhouse-installation-simple-2-2-0 shards: 3 status: Completed taskID: 5e9a0378-fa89-4029-8dfe-fcffc4c37bae taskIDsCompleted: - 5e9a0378-fa89-4029-8dfe-fcffc4c37bae - c4535835-bbd7-4705-843e-445d04cd2ee4 - e6521aa3-c6f3-4bd8-9eb3-42ace89e0c04 - f3cb30fc-35fd-4723-840c-5d280efa0576 - ac33babd-5975-4059-b872-01cbd4c7c184 - 8697fb2d-1dba-44f8-b2c3-e8a7dcda26c7 - 7351d1a4-f5c2-4568-b3fd-0a49d6e7898d - 66d1a4b8-c473-4a95-87e7-44e7747c82ce - 8388aa98-62d8-401d-840b-1b604d7ae6cd - 18f504a6-187f-4784-aeaa-5f967d40439a taskIDsStarted: - 520089d3-fa7e-4d8d-932c-2f5dafc1b21b - f5a11dd9-268b-4c9a-ab70-9c9a92de40b8 - 49d76b1f-d2ff-43f3-b016-355f021bc573 - 2acada9a-e3ba-4234-ac65-94d16d402ba3 - 379fea7c-8cb8-4757-ab0c-ba130b30ea04 - 28e6faf5-a626-4822-878f-ebda6950bf31 - 5851304b-ae19-4b5c-89ca-bf50675b73ee - 83c46bbc-c173-412d-8fcd-8537683f916d - 0c4282bd-cf56-429f-b980-fdc5db20a3ee - 38cef402-0dea-4d74-a056-cb617225f57a ```

Slach commented 1 month ago
        resources:
          requests:
            storage: 4Gi

maybe this is out of space?

could you share kubectl get storageclass -o wide?

fidiego commented 1 month ago

I see. Easy fix, in that case.

NAME            PROVISIONER             RECLAIMPOLICY   VOLUMEBINDINGMODE      ALLOWVOLUMEEXPANSION   AGE
gp2 (default)   kubernetes.io/aws-ebs   Delete          WaitForFirstConsumer   false                  2y135d
Slach commented 1 month ago

could you share? kubectl get storageclass gp2 -o yaml

fidiego commented 3 weeks ago

of course @Slach

apiVersion: storage.k8s.io/v1
kind: StorageClass
metadata:
  annotations:
    kubectl.kubernetes.io/last-applied-configuration: |
      {"apiVersion":"storage.k8s.io/v1","kind":"StorageClass","metadata":{"annotations":{"storageclass.kubernetes.io/is-default-class":"true"},"name":"gp2"},"parameters":{"fsType":
"ext4","type":"gp2"},"provisioner":"kubernetes.io/aws-ebs","volumeBindingMode":"WaitForFirstConsumer"}
    storageclass.kubernetes.io/is-default-class: "true"
  creationTimestamp: "2022-05-05T16:36:20Z"
  name: gp2
  resourceVersion: "325"
  uid: 7249518f-xxxx-xxxx-xxxx-c150dc4feedb
parameters:
  fsType: ext4
  type: gp2
provisioner: kubernetes.io/aws-ebs
reclaimPolicy: Delete
volumeBindingMode: WaitForFirstConsumer
Slach commented 2 weeks ago

try to increase size for logs in log-volume-template in volumeClaimTemplates

resources:
 requests:
   storage: 8Gi

or make log rotation more aggressive something like that

spec:
  configuration:
    settings:
      logger/size: 1000M
      logger/count: 3