Closed fredsig closed 1 week ago
Just realised that I was missing running the janitor component (so, no splits were being deleted). Once I've enabled it, janitor now is catching up with the high number of splits to be deleted, so far errors are gone.
Thanks for letting us know. Let's close this then. Looking rapidly at your chart, I am a tiny bit scared by the pending merge operation chart. It seems to be ok, but you do not have much margin. If you see the pyramid pattern not being able to reach 0 periodically, you might want to tweak the merge_concurrency parameter.
Hi there!
Started to look into testing Quickwit, running latest helm chart in EKS with v0.8.2, 1 indexer pod with local attached EBS volume for /qwdata (gp3 120G), doing 3.20 MB/s, 3k docs/s indexing on a test index (config copied from default otel logs index with added retention settings). Storage backend in S3, metastore in PostgreSQL. No problems with searching/indexing (from what I can see), but observing the following errors (every 10m). It seems to have started after the index reached a certain size:
Is this expected (should be a WARNING) or is there something I've missed on the configuration.
This was gathered by != "INFO" in my query so here's the complete sequence of logs:
qw-logs.txt
Here's some metrics (from a 1 hour view)
Metastore p95 latency looks very fast, I see no noticeable spikes.
Some k8s metrics for the quickwit namespace (indexer pod has been configured with 4G RAM, no OOM issues):
EBS volume stats (indexer /qwdata):
Index config: test-otel_logs.json
My helm chart config section: