Closed anu251989 closed 1 year ago
I have enabled the celery inspect command for checking celery health status. whenever the airflow redis pod restarted the worker pod disconnecting with redis and not processing any messages. This command checking celery status and taking restart worker pod. livenessProbe: exec: command:
I disabled the celery health checks and worker utilization came down. I still need to find out the reason for scheduler utilization.
How to fix the celery issue between redis and workers? if celery health checks disabled, memory leak issue resolved. but if worker disconnected with redis, the workers stay idle without processing any messages.
We upgraded to 2.4.3 airflow version. The airflow redis pod killed and worker pods missed the heartbeat with worker then not processing any tasks. staying as idle. The redis and workers should resume the connection once it is up.
This issue has been automatically marked as stale because it has not had activity in 60 days. It will be closed in 7 days if no further activity occurs.
Thank you for your contributions.
Issues never become stale if any of the following is true:
lifecycle/frozen
label@anu251989 are you still having this issue?
This issue has been automatically marked as stale because it has not had activity in 60 days. It will be closed in 7 days if no further activity occurs.
Thank you for your contributions.
Issues never become stale if any of the following is true:
lifecycle/frozen
labelHello @thesuperzapper, I'm noticing a similar issue to the one described in this issue.
I'll be doing some troubleshooting in the next couple of days but figured I should post in here too for visibility.
I'll start by adding a few extra tools to my custom airflow image to see what's consuming memory.
@albertjmr sounds good, please share any information you have.
Checks
User-Community Airflow Helm Chart
.Chart Version
8.6.0
Kubernetes Version
Helm Version
Description
The scheduler and Worker pods memory keep on increasing day by day. The worker and scheduler pods scaled up max pods.
Relevant Logs
No response
Custom Helm Values