Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Garbage output of Llama-2-13B-chat model after qlora finetuning #274

Open
cywsg opened this issue Oct 25, 2023 · 0 comments
Open

Garbage output of Llama-2-13B-chat model after qlora finetuning #274

cywsg opened this issue Oct 25, 2023 · 0 comments

Comments

@cywsg
Copy link

cywsg commented Oct 25, 2023

I have finetuned the Llama-2-13B-chat model using lora for a document summarization task. The original text is much longer than the model's context length of 4k. I segmented the text into multiple segments with each less than 3K tokens. I performed model inference after model finetuning (adapter was merged to the base model). There are garbage outputs in some segments such as duplicated (similar) sentences or paragraphs. There are some strange patterns too such as 2 or 3 words repeated sequentially before a full stop. Any idea or thought?

@cywsg cywsg changed the title Garbage output of Llama-2-13B-chat model after lora finetuning Garbage output of Llama-2-13B-chat model after qlora finetuning Oct 25, 2023
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant