Closed pieterjanpintens closed 1 year ago
We also tried the gzip command extractor but that fails too, fixing that one might be easier
This issue has been automatically marked as stale because it has been open 90 days with no activity. Remove stale label or comment or this issue will be closed in 30 days
This issue was automatically closed because of stale in 30 days
Is your feature request related to a problem? Please describe.
We try to parse ALB logs from AWS. They are pushed to s3. This is set up by AWS, we don't have any control on this. AWS perform gzip before putting them on s3.
We notice that fluentd sometimes has errors related to these files, complaining that it is not a valid gzipped file. These events go to sqs dlq eventually. When investigating further we saw that these events point to files with size 0. We are not sure why aws puts those files but this is outside our control.
Since the size is in the event (see bellow). Would it be possible add a check to skip these events instead of trying to send them to gzip extractor or can the gzip extractor be improved to work properly in this scenario?
Describe the solution you'd like
Either gzip should cope with files of size 0 or we should be able to prevent that these files reach the extractor phase by using some config option like: skip_zero_bytes_files: bool or something like that.
Describe alternatives you've considered
Live with the errors in fluentd
Additional context
No response