-
Notifications
You must be signed in to change notification settings - Fork 17
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
the inference of OFA-Sys/gsm8k-rft-llama13b2-u13b has shape error: 13Bllama2的u13b版本推理时出现矩阵形状错误 #14
Comments
I'm not using accelerate and your script, I'm just using it as a object of LlamaForCausalLM and using bnb quantize for inference. But i don't think that would cause problem. |
Here is the minimal reproduction. |
my environment |
What is your transformers version? |
It's 4.33.2 |
try transformers==4.29.2 env see issue 9. |
If I want to do some work with new transformer, can I just do some modify to the config to make it work. Do you know what lead to this problem? |
I have no idea how it work on new version; you may train a new model based on our code. |
There seems no people tried your 13b2-u13b version and I may be the first one. But I got 'RuntimeError: mat1 and mat2 shapes cannot be multiplied (111x5120 and 1x2560)' on my inference. While the 7b version works well.
The text was updated successfully, but these errors were encountered: