Prompt Management: Caching on SDK-level #1098
clemra
announced in
Announcements
Replies: 1 comment
-
We have made an update to this behavior to further optimize performance. Whenever a cached version is available it is returned immediately while checking whether an updated prompt version is available is done in the background. https://langfuse.com/changelog/2024-09-04-prompt-management-zero-latency |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
The latest release of the Python and JS/TS Langfuse SDK includes a new prompt caching feature that improves the reliability and performance of your applications.
After launching prompt management in Langfuse, we've enhanced its performance and reliability in the Python and JS/TS SDKs through prompt caching.
Example (Python SDK)
This feature stores prompts in the client SDKs' memory, cutting down on server requests and boosting your applications' reliability and speed. With a default TTL of 60 seconds, caching settings can be adjusted for individual prompts within the SDK. Should a refetch of a prompt fail, the SDK will automatically revert to the cached version, ensuring your application's availability is not compromised.
Looking forward to hearing your thoughts!
cc @hassiebp
Beta Was this translation helpful? Give feedback.
All reactions