Open liad5h opened 1 year ago
Hi same issue with helm chart and consul 1.14.2
Any update ?
I faced to the same case, is there any solution yet?
Been having the same issue for a while now and opened #2509 before seeing this one. On another read its a different issue but similiar behavior... the dataplane just self destructs. It seems like the consul-dataplane doesn't reach a healthy state so it exits.
Community Note
Overview of the Issue
I am using consul on kubernetes with connect enabled. I am trying to use connect to allow secure communication between my two apps - static-server & static-client. Connect works great for the first container I start, doesn't matter which one it is. The second pod always has the consul-dataplane container restarting due to readiness probe.
I did not experience this issue with consul 1.12.x and chart version < 1.x.
Consul version 1.14.2 Chart version 1.0.2
Reproduction Steps
Steps to reproduce this issue, eg:
Helm Values
``` global: name: consul datacenter: eu-central-1-qa enabled: false gossipEncryption: secretName: consul-gossip secretKey: key acls: manageSystemACLs: true bootstrapToken: secretName: consul-bootstrap-acl secretKey: token metrics: enabled: true enableAgentMetrics: true agentMetricsRetentionTime: "1m" defaultPrometheusScrapePath: "/metrics" enableGatewayMetrics: false client: enabled: false server: enabled: true replicas: 1 exposeGossipAndRPCPorts: true connect: true extraConfig: | { "performance": { "raft_multiplier": 1 }, "telemetry": { "disable_hostname": true } } # resources: requests: memory: "1Gi" cpu: "250m" limits: memory: "2Gi" cpu: "1000m" terminatingGateways: enabled: true prometheus: enabled: false connectInject: enabled: true replicas: 1 default: false cni: enabled: true logLevel: info cniBinDir: "/opt/cni/bin" cniNetDir: "/etc/cni/net.d" transparentProxy: defaultEnabled: false metrics: defaultEnabled: true defaultEnableMerging: true defaultPrometheusScrapePort: 20200 defaultPrometheusScrapePath: "/metrics" resources: requests: memory: "50Mi" cpu: "50m" limits: memory: "250Mi" cpu: "300m" ui: enabled: true metrics: enabled: true provider: "prometheus" service: type: ClusterIP ```Run:
Server info
``` agent: check_monitors = 0 check_ttls = 0 checks = 0 services = 0 build: prerelease = revision = 0ba7a401 version = 1.14.2 version_metadata = consul: acl = enabled bootstrap = true known_datacenters = 1 leader = true leader_addr = 10.209.55.131:8300 server = true raft: applied_index = 8756 commit_index = 8756 fsm_pending = 0 last_contact = 0 last_log_index = 8756 last_log_term = 22 last_snapshot_index = 0 last_snapshot_term = 0 latest_configuration = [{Suffrage:Voter ID:66d802f6-563e-1223-da38-5a907b19f317 Address:10.209.55.131:8300} {Suffrage:Voter ID:1e7eaab7-6d3a-273c-38e0-136b669a3555 Address:10.209.16.195:8300}] latest_configuration_index = 0 num_peers = 1 protocol_version = 3 protocol_version_max = 3 protocol_version_min = 0 snapshot_version_max = 1 snapshot_version_min = 0 state = Leader term = 22 runtime: arch = amd64 cpu_count = 4 goroutines = 202 max_procs = 4 os = linux version = go1.19.2 serf_lan: coordinate_resets = 0 encrypted = true event_queue = 0 event_time = 4 failed = 0 health_score = 0 intent_queue = 0 left = 0 member_time = 2 members = 2 query_queue = 0 query_time = 1 serf_wan: coordinate_resets = 0 encrypted = true event_queue = 0 event_time = 1 failed = 0 health_score = 0 intent_queue = 0 left = 0 member_time = 3 members = 2 query_queue = 0 query_time = 1 ```Operating system and Environment details
Chart version 1.0.2 (Consul 1.14.x) AWS EKS 1.21
Log Fragments
kubernetes events from the pod: