Open Jeinhaus opened 1 year ago
Hey there 👋
Yes, whenever loki.file_match returns a new set of targets, the loki.source.file component will restart all of its readers. We've have had a TODO to look into this for a while now.
Could you share a CPU profile to see if there's any impact on performance on this, or is it just the extra logs that are being annoying?
Hi, thanks for the quick response. Good to know, that you already have a TODO for this.
Sorry, I don't have a CPU profile ready. The performance impact is very minimal (if at all) as far as I can see from our Prometheus metrics.
For now, I think we'll just live with the extra logs or might increase the log level if the log volume is too much.
Thank you for your help.
Given that this is currently expected behavior, I'm changing this from a bug
to an enhancement
to remove the annoying log volume.
This issue has not had any activity in the past 30 days, so the needs-attention
label has been added to it.
If the opened issue is a bug, check to see if a newer release fixed your issue. If it is no longer relevant, please feel free to close this issue.
The needs-attention
label signals to maintainers that something has fallen through the cracks. No action is needed by you; your issue will be kept open and you do not have to respond to this comment. The label will be removed the next time this job runs if there is new activity.
Thank you for your contributions!
Any movement on this enhancement? Setting the logging level to warn
isn't great, else I think we're into dropping these logs on Alloy.
What's wrong?
We use the
loki.source.file
component to read pod logs in our EKS cluster. The config (see below) is similar to the setup described in the documentation. So far, it works great. But we noticed, that whenever a pod is created or destroyed in the cluster, the Grafana agent will stop all existing tail processes and start them again (see logs below). This produces (depending on the number of pods in the cluster, I guess) around 1k log lines per pod change.Is this intended behavior, that all tail processes are restarted, or did we misconfigure something? If it's intended behaviour, we might just not log
info
log messages to avoid the high log volume.Steps to reproduce
Start or stop a new pod in the cluster and observe the log-volume of grafana-agent.
System information
5.10.186-179.751.amzn2.x86_64
Software version
grafana/agent:v0.36.1
Configuration
Logs