Open mitesh-devops opened 3 years ago
This issue is stale because it has been open 30 days with no activity. Remove stale label or comment or this will be closed in 5 days.
This issue was closed because it has been stalled for 5 days with no activity.
Hi @mitesh-devops, were you able to solve this issue? I'm facing the same, see #4155
Same here.
same here
Hi, Actually, it was more issue of the output plugin. In my case, output pluging was Elasticsearch so sizing the shards correctly and including rollover ILM policies helped me to get this fixed. Also, try adding more RAM to diagnose.
any solution regrading this ?
@kc-dot-io Did you find solution for this ?
I had same issue. But I think, may be I found which is main factor. I think Refresh_Interval, Buffer_Max_Size and Buffer_Chunk_Size is key-factor. I fixed this configuration after then there are not exist same error.
Has this been resolved? This seems to be related to not removal of inode during file removal process i.e. (the inode entry not being removed from files_static, files_event[https://github.com/fluent/fluent-bit/blob/master/plugins/in_tail/tail_config.h#L129, https://github.com/fluent/fluent-bit/blob/master/plugins/in_tail/tail_file.c#1108])
Hi all, it finally worked for our project after replacing Docker with Containerd and by applying the following config: https://github.com/vmware-samples/vcenter-event-broker-appliance/tree/development/files/configs/fluentbit/templates
https://github.com/fluent/fluent-bit/issues/4155#issuecomment-1046239927
There is a bug in v1.8.13, we have just published the new container image for v1.8.14, please upgrade:
fluent/fluent-bit:1.8.14
Same here (v1.9.1)
Still having the issue with v1.9.2 too. any fix soon?
Same in 1.9.3
Same issue guys
Same in v1.9.3
Same in 1.9.4
Any resolution for this, processes a few logs then falls over ? Docker image: fluent/fluent-bit:1.9.3-debug
Has anyone tried turning Skip_Long_Lines to off ? Some of my logs were dismissed, because they were too long I guess, this did the trick
@Jonathan-w6d yes. This will help for a while, depending on how much data you are processing. The root issue is a memory leak. Increasing buffer / skipping long lines helps get it working but the issue will still reoccur every so often. I have to restart our fluentbit pods every 90-100 days or so.
The same issue is in v2.0.10.
[2023/04/04 21:06:11] [debug] [input:tail:tail.3] scan_blog add(): dismissed: /logs/server.out, inode 277548
[2023/04/04 21:06:11] [debug] [input:tail:tail.4] scan_blog add(): dismissed: /logs/server.FATAL, inode 262809
The same issue is in v2.1.1. [2023/05/22 23:34:52] [debug] [input:tail:tail.0] scan_blog add(): dismissed: /tomcat-logs/platform-2023-05-22.debug.log, inode 57744220 [2023/05/22 23:34:52] [debug] [input:tail:tail.0] 0 new files found on path '/tomcat-logs/platform*.log'
Same issue, has this been resolved?
Same issue using latest version
I faced the same question. Since tail input only reads the last offset of the file - make sure you are actually writing to the correct files that you are reading from, that your filters are correct and that the lines are not too long (because you have Skip_Long_Lines turned on). You can check yourself by using Read_from_Head set to true under [INPUT]
Any updates? Because the same thing happened on v2.1.10
Still same issue here
I Have the same issue https://github.com/fluent/helm-charts/issues/415:
kubectl logs -n fluent-bit -f fluent-bit-6341f45a-j8h2x | grep -i volume-test
[2023/10/27 17:04:05] [debug] [input:tail:tail.0] inode=1062251 with offset=2026 appended as /var/log/containers/volume-test_exp_volume-test-2c25a3b1342924cffab9fffb48b2f0a971fcf9f10009e4c382a37bc09075134b.log
[2023/10/27 17:04:05] [debug] [input:tail:tail.0] scan_glob add(): /var/log/containers/volume-test_exp_volume-test-2c25a3b1342924cffab9fffb48b2f0a971fcf9f10009e4c382a37bc09075134b.log, inode 1062251
[2023/10/27 17:04:05] [debug] [input:tail:tail.0] inode=1062251 file=/var/log/containers/volume-test_exp_volume-test-2c25a3b1342924cffab9fffb48b2f0a971fcf9f10009e4c382a37bc09075134b.log promote to TAIL_EVENT
[2023/10/27 17:04:05] [ info] [input:tail:tail.0] inotify_fs_add(): inode=1062251 watch_fd=23 name=/var/log/containers/volume-test_exp_volume-test-2c25a3b1342924cffab9fffb48b2f0a971fcf9f10009e4c382a37bc09075134b.log
[2023/10/27 17:05:04] [debug] [input:tail:tail.0] scan_blog add(): dismissed: /var/log/containers/volume-test_exp_volume-test-2c25a3b1342924cffab9fffb48b2f0a971fcf9f10009e4c382a37bc09075134b.log, inode 1062251
[2023/10/27 17:06:04] [debug] [input:tail:tail.0] scan_blog add(): dismissed: /var/log/containers/volume-test_exp_volume-test-2c25a3b1342924cffab9fffb48b2f0a971fcf9f10009e4c382a37bc09075134b.log, inode 1062251
[2023/10/27 17:07:04] [debug] [input:tail:tail.0] scan_blog add(): dismissed: /var/log/containers/volume-test_exp_volume-test-2c25a3b1342924cffab9fffb48b2f0a971fcf9f10009e4c382a37bc09075134b.log, inode 1062251
[2023/10/27 17:08:04] [debug] [input:tail:tail.0] scan_blog add(): dismissed: /var/log/containers/volume-test_exp_volume-test-2c25a3b1342924cffab9fffb48b2f0a971fcf9f10009e4c382a37bc09075134b.log, inode 1062251
same here with version 2.2.1, have any solution?
Is anyone actually going to fix this? This issue needs to be re-opened.
I still have this with version 3.0.3.
Please re-open the issue, seems to be persistent in all recent versions
Can we please get an update on this issue? Thanks!
Switched to filebeat as fluent-bit unusable in production due to this issue. Sorry guys, but 3 years of nothing is a little strange... @edsiper maybe you need to look on this?
Hi folks, I have create this branch where I've added some trace level debug messages for some code paths that could slip through the cracks and I'm currently creating a set of container images that you'll be able to access in ghcr.io/fluent/fluent-bit/test/leonardo-master-issue_4014_debug_addition
once the packaging process is complete.
I'd really appreciate it if those of you who are currently able to reproduce the issue took the time to either build from source or run those containers (setting log_level
to trace
) and share those logs with me (You can find me in the public slack server as Leonardo Alminana
if you'd rather share that information privately) to expedite the troubleshooting process.
Quick update: the container images are already available so please, it's really important for at least someone to run them and report the results.
I'll try to find some time this week to deploy the image @leonardo-albertovich . Where are the logs output to? Stdout/err? Or can I define a sink?
The logs are produced in the standard way so you don't need to do anything special (other than setting log_level
to trace
)
I am seeing the same issue
I'm still getting the below error. However, the issue was raised when multiline processing was enabled. what could be the issue? image -> public.ecr.aws/aws-observability/aws-for-fluent-bit:2.32.2.20241008 (Fluent Bit v1.9.10) [2024/10/16 04:03:58] [debug] [input:tail:tail.0] scan_blog add(): dismissed: /var/log/containers/master-ms-deployment-5db5876558-w5gjj_master-ms-e2927e4cfeb05a89f229d7622edaefc750e48568d65415e983980535031ddd41.log, inode 75503248
I would request some help for following fluent-bit logs which is causing logs skip for our many apps.It works correctly for few hours of the restart of fluent-bit.
logs 1:
frequency: around 15-20 entries every 5 minutes when container starts and after few hours, it's the only logs which fluentbit prints and all logs get stopped.
logs 2:
frequency: very few times
log 3:
frequency: Sometimes only.
Rest of the logs looks normal and correctly gathering and writing to elasticsearch.
Cluster Information:
Expected behaviour: It shouldn't skip logs and should work without restart after sometime.
fluent-bit.conf
fluent-bit.yaml