-
Notifications
You must be signed in to change notification settings - Fork 59
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Llama-3-8B #75
Comments
@hsb1995 |
Professor, thank you for your full work. I really don't know how GQA is handled as you mentioned Can I understand what you said that I kept the original "generateAct_scale.shift" file unchanged to obtain the "act_scales" and "act_shifts" files. |
Hey, professor. I gave it a try. [2024-04-24 17:14:17 root](omniquant.py 50): INFO Starting ... |
@ChenMnZ hello,I also meet some problems like this. |
I have obtained the weight offset factor for llama3-8b, but there was a unique mismatch issue during my compression process.
My scaling factor code has not been changed, but there was a dimension issue when I started compressing.
The parameter settings are as follows:
--model
${}$Llama-3-8b/
--epochs
20
--output_dir
${}$llama-3-8b-w6a6/
--eval_ppl
--wbits
6
--abits
6
--lwc
--let
--net
Llama-3-8b
--tasks
arc_easy,arc_challenge,boolq,hellaswag,winogrande,piqa
The text was updated successfully, but these errors were encountered: