-
Notifications
You must be signed in to change notification settings - Fork 1.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[REQUEST] Make Ollama URL customizable #505
Comments
@ferpizza User can always edit any parameters to the models, including base_url like this |
In my case, the issue with Ollama was happening at the initial Setup screen, where we have no chance to edit these settings ... and it was failing to download models and move forward (and I didn't wanted to go into advance mode to get to the screen you are sharing). I workaround it by adding a sed command to the official docker entrypoint, and was able to download models and get a working connection with Ollama running on an independent container. But I think that having this on an EnvVar is better and cleaner. (also, my workaround is not bulletproof)
|
Also, on the screen you share, I still see |
@ferpizza yes we did not consider alternate Ollama endpoint in first setup screen. User are expected to use Skip setup and Set the model in the UI manually later. An env var specifically for Ollama might probably help, we will consider this feature. |
Reference Issues
No response
Summary
Currently, Ollama URL is hardcoded to
http://localhost:11434/api
This becomes an issue when deploying Kotaemon in more advanced scenarios, like running different containers for Kotaemon and Ollama.
Basic Example
Kotaemon docker image doesn't include Ollama, so one approach to tackle this is to run a 2nd docker image with Ollama in it.
The challenge is that on this scenario, Ollama is no longer available on
http://localhost:11434/api
, but on some other URL likehttp://ollama:11434/api
.As Ollama URL is not a setting we can manage through EnvVars, the above deployment is not possible without having to tweak the code.
Drawbacks
We are unable to run Kotaemon and Ollama on different VMs/Containers, and this is a common pattern when working with contaiers (eg. kubernetes, docker-compose)
Additional information
No response
The text was updated successfully, but these errors were encountered: