-
Notifications
You must be signed in to change notification settings - Fork 35
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Helm Deployment - Web Pod Crashing Due to Failed Migration #47
Comments
Hey @WendesonSilva-neuro! I'm here to help you with the issue you're facing. Feel free to ask me anything about bugs, questions, or becoming a contributor. To resolve the migration issue causing the
By following these steps, you should be able to address the migration failure and get the To continue talking to Dosu, mention @dosu. Help Dosu learn! Give it feedback: Great Response | Irrelevant Answer | Incorrect Sources | Too Verbose | Hallucination | Other |
Hey @WendesonSilva-neuro , This issue may happen, if Kubernetes kills the langfuse-web container before it completes its migrations. With the startupProbe and livenessProbe defaults it should have about 45s to perform the migrations before being killed. Did you change those settings in any way? If it's a fresh stack, you may try to reinstall the chart with a longer startupProbe duration or, in case you want to make the current installation work, you can connect to the Postgres database and remove the last entry in the Let me know if that helps or if you encounter any other problems. |
Hi @Steffen911, Thank you for your response! I updated the Helm repository and applied the following values.yaml configuration, but unfortunately, the issue persists:
I also extended the livenessProbe settings for both langfuse-web and langfuse-worker pods to allow more time for startup. Despite these adjustments, the langfuse-web pod continues to enter a CrashLoopBackOff state with the same migration error:
I am using an external PostgreSQL database hosted on RDS. Could you confirm if any additional steps are required to resolve this issue? Should I follow the manual migration rollback process you suggested, or is there another recommended approach given this setup? Thanks again for your assistance! |
Hey @WendesonSilva-neuro , Can you share the contents of the _prisma_migrations table with me? The timestamps in there should indicate whether we're talking about a timing issue where more time might help. |
Hi @Steffen911, _**Prisma schema loaded from packages/shared/prisma/schema.prisma 273 migrations found in prisma/migrations No pending migrations to apply. I Tried to create a database on Azure Cloud also but error is same. I'm doing fresh installation. Attaching Screen shot. I have tried with increasing probes timeout also. |
@arhamhamood1306 Can you delete all tables in the clickhouse cluster and restart the web container? |
@Steffen911 Can you please share command to delete tables in clinkhouse cluster. FYI I have tried to delete the clickhouse pods but issue is there. |
@arhamhamood1306 One option is to delete the PVCs associated with ClickHouse, or you connect with the clickhouse cli and run |
@Steffen911 It seems to work now after deleting the PVCs. Thanks. |
While deploying Langfuse using the official Helm guide: Self-Hosting with Kubernetes Helm, the langfuse-web pod enters a CrashLoopBackOff state. Logs from the pod indicate a migration failure during startup:
Steps to Reproduce:
Expected Behavior:
The deployment should succeed, and all pods should be running without errors.
Actual Behavior:
The langfuse-web pod fails during startup due to a failed Prisma migration (20230918180320_add_indices).
Additional Context:
What is the recommended way to resolve this migration issue?
Let me know if additional logs or debugging information would help.
Thank you!
The text was updated successfully, but these errors were encountered: