Closed wrahmann closed 2 years ago
We are also seeing high CPU usage in our environment but never crosses beyond 1. I read its fluentd which can not go beyond 1 CPU. Our consumption always around 1 CPU only during peak with below settings, whereas there is no limits configured and can consume as much as required.
resources: requests: cpu: 1 memory: 1
We do see issues when 1 CPU is consumed where events are delayed for hours. We wanted to check if we can somehow provide more resources where fluentd can consume and expedite processing. Is there any way we can configure multi-process? (I read tail do not support it but do we have any sample for SCK where it can consume more CPU?)
I'm wondering the same thing. We're firing up SCK on our test environment for the first time, and immediately it jumped up higher than any other project we have on the cluster. We're using a combination of the default values.yaml and the OCP4 example here: https://github.com/splunk/splunk-connect-for-kubernetes/blob/develop/helm-chart/splunk-connect-for-kubernetes/examples/openshift4-logging-only.yaml
Here is our values.yaml in its messy glory.
Ah, we found a possible fix. It appears that using JSON with OpenShift is a bad idea(tm), and causes all sorts of noise. When we switched the logging to CRI, it started working as we expected.
This issue is stale because it has been open for 30 days with no activity.
This issue was closed because it has been inactive for 14 days since being marked as stale.
Hi,
For Splunkconnect , the defaults values for CPU and memory allocation is as follows:
resources:
limits:
However, it does not provide any detail when should we change these values? What is the logging throughput that can be supported by these values?
Also please let me know how buffer settings are impacted if we change the cpu and memory?
buffer: "@type": memory total_limit_size: 600m chunk_limit_size: 20m chunk_limit_records: 100000 flush_interval: 5s flush_thread_count: 1 overflow_action: block retry_max_times: 5 retry_type: periodic
Wajih