Open kladiv opened 1 year ago
in my case, When the querier starts, a connection is made to the query frontend.
However, since the querier accesses through the service of the query frontend, it seems that 1:1 mapping may not be possible if there are multiple query frontends.
So, if you increase the querier or decrease the query frontend, it seems that the querier and the query frontend are connected.
eg) query-frontend : "1" , querier : "3"
I guess it's not related to replicas ratio. I guess it should be related to this model: https://grafana.com/docs/loki/latest/configuration/query-frontend/#grpc-mode-pull-model
I'm also getting trouble with this.
It seems that publishNotReadyAddresses
is missing from the querier headless Service
. May this matter?
Same here, with 2 querier and 2 query-frontend pods. As @kladiv mentioned, I changed https://github.com/grafana/helm-charts/blob/main/charts/loki-distributed/values.yaml#L181 to the headless-service and then it worked. Not sure if that's the solution though, or if side-effects can be expected.
In my case, it seems that the change to the headless service does not work normally. Querier has four replicas and query frontend has two replicas, each with an autoscaling option enabled, resulting in a crashloopbackoff (distributor / ingester / querier / queryFrontend).
I'm also facing this error.
If I disable the queryScheduler it works fine
We're seeing this as well..
Please check latest release. Frontend address was adjusted in loki-distributed-0.69.13
: https://github.com/grafana/helm-charts/commit/3829417e0d113d24ea82ff9f0c6c631d20f95822
I no longer see this issue in helm-loki-5.2.0
chart.
We also deployed the 5.2.0
Helm chart to some of our environments today and the issue appears to be resolved. :+1:
I encountered the same issue in mimir-distributed
helm chart and I resolved it by configuring frontend_worker.scheduler_address
parameter. More info here: https://grafana.com/docs/mimir/latest/references/configuration-parameters/#frontend_worker
Using s3 all the way solved this for me. Using filesystem with loki-distributed did make some weird problems
I have the same error with grafana/loki
(simple scalable) deployment Helm Chart version 5.2.0. It deploys 3 loki-read
pods and only one gives that error, the other two is happy.
Edit: After restarting that failing pod, it becomes healthy.
I am having this same issue. My environment is running on istio with mutual tls enabled. If I disable mutual tls everything works fine.
❯ helm ls -n loki
NAME NAMESPACE REVISION UPDATED STATUS CHART APP VERSION
loki loki 2 2024-03-21 15:57:58.90046894 -0400 EDT deployed loki-distributed-0.78.3 2.9.4
Hello, concerning issue https://github.com/grafana/helm-charts/issues/2028 i still got error below when
queryFrontend.replicas
is 2:I checked and the headless service seems present:
Helm chart version is 0.69.9
Why i'm still got this?
Could it be 'caused by the spec below in values.yaml that should point the headless endpoint? https://github.com/grafana/helm-charts/blob/main/charts/loki-distributed/values.yaml#L181
Thank you