Open andrejshapal opened 7 months ago
Our environments seem to be similar, and we also face the same issue. Does your thanos_blocks_meta_synced report a lot of duplicates as well?
This is an image from bitnami's bucketweb that shows "duplicate" blocks with different resolution
@PantelisK Different resolution is expected behaviour when downsampling is turned on. In our case it is:
@andrejshapal thanks for the clarification. I though the duplicate was a problem :)
Thanos, Prometheus and Golang version used: We bump to latest Thanos on every Monday
Object Storage Provider: gc bucket
What happened: I was searching for resolution of my issue and found a lot of similar issues on git and here in slack. It is very scary none of the issues had a proper answer or solution. Basically, we run compactor with args:
and we have noticed spam in logs like:
The compactor goes through the same blocks in frequent loop trying to mark them and fire warning.
What you expected to happen: My theory is that the default deletion delay is 48h. And wait period is 5m. Therefore compactor runs every 5 minutes and tries to mark the blocks which were marked before but not deleted due to 48h delay. If this is correct, then the warning should not be fired if the deletion mark was created < deletion delay. And maybe the default wait-period should be increased (to 48h?). I don't know if it is really needed to run compaction each 5m and run over the same blocks...
How to reproduce it (as minimally and precisely as possible): It should be reproducible if there are any marked for deletion blocks.
Full logs to relevant components:
Anything else we need to know: I also think this leads to increased expenses, no? Not sure how it works, but compactor probably downloads some data from bucket and this is paid operation?
Environment: