Open rwlove opened 1 year ago
Hey thank you for your report. Could you share how did you run/deploy the project? So I can replicate it myself. Suggestions:
inmemory
and with replication_factor: 1
. If that works it means you have a network issuecmd/loki
. Hey thank you for your report. Could you share how did you run/deploy the project? So I can replicate it myself. Suggestions:
I deployed it via Flux2 (GitOps infrastructure), so I'm not sure how easily you could reproduce. If there's something else I can share, I'd be glad to.
Hey thank you for your report. Could you share how did you run/deploy the project? So I can replicate it myself. Suggestions:
- Try running with replicas=1, ring store=
inmemory
and withreplication_factor: 1
. If that works it means you have a network issue
Read and Write pods start just fine with the above configuration.
As far as I can tell my network is fine. Any suggestions one what to look for?
KVStore: Ok Disabled
Kubernetes: Ok 1.25 (v1.25.5) [linux/amd64]
Kubernetes APIs: ["cilium/v2::CiliumClusterwideNetworkPolicy", "cilium/v2::CiliumEndpoint", "cilium/v2::CiliumLocalRedirectPolicy", "cilium/v2::CiliumNetworkPolicy", "cilium/v2::CiliumNode", "core/v1::Namespace", "core/v1::Node", "core/v1::Pods", "core/v1::Service", "discovery/v1::EndpointSlice", "networking.k8s.io/v1::NetworkPolicy"]
KubeProxyReplacement: Strict [enp1s0 192.168.4.8 (Direct Routing)]
Host firewall: Disabled
CNI Chaining: none
CNI Config file: CNI configuration file management disabled
Cilium: Ok 1.13.0 (v1.13.0-c9723a8d)
NodeMonitor: Listening for events on 6 CPUs with 64x4096 of shared memory
Cilium health daemon: Ok
IPAM: IPv4: 31/254 allocated from 11.0.5.0/24,
IPv6 BIG TCP: Disabled
BandwidthManager: Disabled
Host Routing: Legacy
Masquerading: IPTables [IPv4: Enabled, IPv6: Disabled]
Controller Status: 149/149 healthy
Proxy Status: OK, ip 11.0.5.151, 0 redirects active on ports 10000-20000
Global Identity Range: min 256, max 65535
Hubble: Ok Current/Max Flows: 4095/4095 (100.00%), Flows/s: 90.27 Metrics: Ok
Encryption: Disabled
Cluster health: 12/12 reachable (2023-03-06T22:22:09Z)```
Describe the bug Loki deployment fails because Read and Write pods crash with unexplained "failed services" message.
I'd be happy to learn that this is a configuration error, but I'm not sure what next steps to take to debug this.
To Reproduce Steps to reproduce the behavior: Deploy Loki via Helm Chat with Rook Ceph bucket for storage and Read and Write pods crash with cryptic message.
Expected behavior Read and Write pods don't crash and deployment succeeds.
Environment:
Screenshots, Promtail config, or terminal output