Closed didlawowo closed 2 weeks ago
Hey there, @didlawowo! I'm here to help with any bugs, questions, or contributions you have. Let's tackle this issue together!
The pod restarts and errors you're experiencing with the latest version of the Langfuse Helm chart seem to be related to the liveness and readiness probes failing. The logs indicate that the probes are unable to connect to the specified endpoints, resulting in the pod being marked as unhealthy and subsequently restarted.
Here are a few steps you can take to troubleshoot and resolve the issue:
Check Service Availability: Ensure that the Langfuse service is actually running and accessible on the expected port (3000). You can try accessing the service directly from within the cluster using a tool like curl
to verify connectivity.
Increase Probe Timeouts: The current timeout for both liveness and readiness probes is set to 1 second. This might be too short if the service takes longer to respond. Consider increasing the timeout to give the service more time to respond.
Review Recent Changes: If the previous version was working fine, review any changes made in the latest version of the Helm chart that might affect the service's startup time or resource usage. There might be changes that require additional configuration or resources.
Check Resource Limits: Ensure that the pod has sufficient CPU and memory resources allocated. Insufficient resources can lead to slow startup times and probe failures.
Database Connectivity: Since Langfuse relies on a Postgres database, ensure that the database is accessible and properly configured. Any issues with the database connectivity can affect the readiness of the application.
Logs and Metrics: Check the application logs and any available metrics for more detailed error messages or performance issues that might be causing the probes to fail.
By addressing these areas, you should be able to identify and resolve the cause of the pod restarts.
To continue talking to Dosu, mention @dosu.
can you share logs from the container instance?
@didlawowo Could you also share the resources that you've allocated to the pod? How much traffic does it usually receive over a minute?
got this error:
Error: Required database environment variables are not set. Provide a postgres url for DATABASE_URL.
in fact this env var is not available in helm chart.
@didlawowo Could you share the values.yaml
file that you use?
In our default setting we deploy a postgres instance and set this environment. If this is toggled off, you'll have to provide this value using the additionalEnv
fields.
I did an upgrade today to helm version 0.6.0 . i have set postgresql.deploy = true
in values.yaml
and i have the same issue. I upgraded from version 0.2.1
same here, the default helm chart of langfuse-k8s no longer works
(command helm install release-langfuse langfuse/langfuse -n langfuse
)
Probably the new commit modified some configurations and the user must manually provide the variables.
kubectl logs release-langfuse-78559588bd-s6st8
Error: Required database environment variables are not set. Provide a postgres url for DATABASE_URL.
Thank you for the report. I've noticed a missing .auth
in one of the recent changes. This should be fixed with https://github.com/langfuse/langfuse-k8s/commit/2c7800c1b23fd0b1c64ecfb391c974635fd5522e in the new 0.7.0 version.
i have deployed with helm previous version was working like a charm and got some restart / error