Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Error when evaluating the qlora-merged model. #427

Closed
vhientran opened this issue Nov 12, 2024 · 1 comment
Closed

Error when evaluating the qlora-merged model. #427

vhientran opened this issue Nov 12, 2024 · 1 comment

Comments

@vhientran
Copy link

Hi Authors,

Thanks so much for releasing the great source code. I finetuned LlaMA-2-7B using QLoRA successfully on a node of 8GPUs A100 NVIDIA, and saved the merged model well. However, when loading this created model with vLLM, I got the error below:

model = vllm.LLM(
[rank0]:             ^^^^^^^^^
[rank0]:   File "/miniconda3/envs/open-instruct/lib/python3.11/site-packages/vllm/entrypoints/llm.py", line 177, in __init__
[rank0]:     self.llm_engine = LLMEngine.from_engine_args(
[rank0]:                       ^^^^^^^^^^^^^^^^^^^^^^^^^^^
[rank0]:   File "/miniconda3/envs/open-instruct/lib/python3.11/site-packages/vllm/engine/llm_engine.py", line 573, in from_engine_args
[rank0]:     engine = cls(
[rank0]:              ^^^^
[rank0]: **KeyError: 'layers.11.mlp.down_proj.weight'**
Loading safetensors checkpoint shards:   0% Completed | 0/3 [00:00<?, ?it/s]

It seems the difference about the structure when loading the model. Could you please give me some suggestions to fix it? Many thanks for your help!

@hamishivi
Copy link
Collaborator

hamishivi commented Jan 8, 2025

Hi, are your vLLM and transformers vesrion up to date? Sometimes mismatches there can cause issues. Otherwise, its hard to say without more details - do you have example commands you used for training and merging?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants