-
-
Notifications
You must be signed in to change notification settings - Fork 2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
llama-cpp of LocaAI 2.23.0 doesn't load custom .gguf files #4199
Labels
Comments
Maybe related to #4170 |
Debug log:
What does |
Ah, most probably related to ggerganov/llama.cpp#10244 |
... fixed it by using a newer gguf from https://huggingface.co/mradermacher/dolphin-2.7-mixtral-8x7b-GGUF |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
LocalAI version:
2.23.0, podman installation, tried with
latest-gpu-nvidia-cuda-12
andlatest-aio-gpu-nvidia-cuda-12
`
Environment, CPU architecture, OS, and Version:
Standard Debian 12 (96 GB memory) with two Nvidia Tesla P40 gpus (each 24 GB memory)
Describe the bug
LocalAI fails to load custom .gguf files. In this case its
dolphin-2.5-mixtral-8x7b.Q5_K_M.gguf
. Following error message appears with all backendsERR [llama-cpp] Failed loading model, trying with fallback 'llama-cpp-fallback', error: failed to load model with internal loader: could not load model: rpc error: code = Canceled desc =
It loads and works without issue on version 2.22.1.
To Reproduce
Update LocalAI images to version 2.23.0 and load the "custom" model
dolphin-2.7-mixtral-8x7b.Q5_K_M.gguf
.Expected behavior
Loading should work.
Logs
Multiple lines of following messages for all backends.
ERR [llama-cpp] Failed loading model, trying with fallback 'llama-cpp-fallback', error: failed to load model with internal loader: could not load model: rpc error: code = Canceled desc =
The text was updated successfully, but these errors were encountered: