You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
it's because llama.cpp needs to be updated to support those versions. I will be updating llama.cpp soon. It's tricky because it can break code (3 times have had to start over so far, and had to make some major changes even in Lumina, the last time I updated llama.cpp).
If you want to use those models NOW, just use AIToolkit, its designed in a way that I can frequently update llama.cpp more often.
Hi Jarrod.
I tried loading some DeepSeek Distil Qwen based models with Lumina; below you can find the list of models I tried to load:
But all of them, during loading, give the following error:
"print_info: file size = 11.22 GiB (6.78 BPW)
llama_model_load: error loading model: error loading model vocabulary: unknown pre-tokenizer type: 'deepseek-r1-qwen'"
Lumina Log.log
Not that maybe Lumina should be compiled with the latest version of Llama.cpp?
Thanks!
The text was updated successfully, but these errors were encountered: