Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

LLM API calls cost management #48

Open
renehonig opened this issue Jan 19, 2024 · 1 comment
Open

LLM API calls cost management #48

renehonig opened this issue Jan 19, 2024 · 1 comment
Assignees

Comments

@renehonig
Copy link
Member

Is your feature request related to a problem? Please describe.
As an operations person I want to be sure that the use of LLM calls to external API's does not become cost prohibitive.

Describe the solution you'd like
The guidance engine uses an external service provider for LLM inference for which costs are charged to Alkemio. Whilst individual calls are relatively cheap, the costs could still become prohibitive if the usage patterns are different from what is expected, e.g. due to a bug or some type of misuse by the user. The majority of controls should be managed through the server, but it is probably prudent to build in some simple reporting/rate limiting on the guidance engine side.

Describe alternatives you've considered
Full control of the API calls to the engine on the server side.

Additional context
Over time, more granular controls may shift to the server side. Also, the use of LLM is expected to grow, which will create additional/different requiremenrts for cost management.

@techsmyth
Copy link
Member

@renehonig moving this + other guidance engine issues for new functionality to icebox pending the wider Roadmap prioritization.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

2 participants