You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I have checked the documentation of the related framework and cannot find useful information.
I have searched the issues and there is not a similar one.
Information about environment
OS: Windows 10 & 11
Python: 3.13
CPU
Description
Steps to reproduce
I tried converting Qwen2.5-3B-Instruct into Q2_K quantization on two different machines. The output of the compressed model is always nonsense: ,“||9"363的76...5 31367244一246“),).请-264“3))-64))5761595431636843467435565846"):4843)"),\n5353"34“ ), 3\"6)) the"24\n\n964
But it seems that this only happens to Q2_K.
Platforms:
Windows 10 with llama.cpp build b4846.
Windows 11 with llama.cpp build b4520.
It is optional for Q2_K quantization. I don't think the gibberish output is caused by the missing of importance matrix. The importance matrix will decrease perplexity. But the perplexity of gibberish output seems to be infinity.
Model Series
Qwen2.5
What are the models used?
Qwen2.5-3B-Instruct
What is the scenario where the problem happened?
Qwen2.5-3B-Instruct with Q2_K quantization, Llama.cpp
Is this badcase known and can it be solved using avaiable techniques?
Information about environment
Description
Steps to reproduce
I tried converting
Qwen2.5-3B-Instruct
intoQ2_K
quantization on two different machines. The output of the compressed model is always nonsense:,“||9"363的76...5 31367244一246“),).请-264“3))-64))5761595431636843467435565846"):4843)"),\n5353"34“ ), 3\"6)) the"24\n\n964
But it seems that this only happens to
Q2_K
.Platforms:
Windows 10 with llama.cpp build b4846.
Windows 11 with llama.cpp build b4520.
Original model:
https://huggingface.co/Qwen/Qwen2.5-3B-Instruct
Conversion script:
python convert_hf_to_gguf.py ***\Qwen2.5-7B-Instruct --outfile ***\Qwen2.5-7B-Instruct-FP16.gguf
Quantization script:
llama-quantize.exe ***\Qwen2.5-3B-Instruct-FP16.gguf ***\Qwen2.5-3B-Instruct-Q2_K.gguf Q2_K
Model testing script:
llama-cli.exe -m ***\Qwen2.5-3B-Instruct-Q2_K.gguf
Expected results
The results are expected to be ordinary output of words.
Attempts to fix
I have tried several ways to fix this but no help, including:
Anything else helpful for investigation
None.
The text was updated successfully, but these errors were encountered: