-
Notifications
You must be signed in to change notification settings - Fork 1.7k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
add gptqmodel support #2247
base: main
Are you sure you want to change the base?
add gptqmodel support #2247
Conversation
Thanks for this PR to add support for gptqmodel. Let's wait for the transformers PR to be merged before proceeding with this one. |
Signed-off-by: jiqing-feng <[email protected]>
Hi @BenjaminBossan , there is a weird pytest error I never met,
The pytest works only when I removed this code |
Testing changes contain:
For gptq testing:
|
This comes from a pytest plugin we use to monitor code coverage. Running (btw. code coverage is great to check if the new code you added is covered by unit tests) |
Signed-off-by: jiqing-feng <[email protected]>
* add get_gptq_quant_linear * cleanup * rename to quant_linear * rename to get_gptqmodel_quant_linear * rename to QuantLinear * fix get device_map * import hf_select_quant_linear * pass checkpoint_format * fix lora * if is_gptqmodel_available, pass beckend auto_trainable * pass backend auto_trainable * cleanup * Update gptq.py --------- Co-authored-by: Qubitium-ModelCloud <[email protected]>
@BenjaminBossan This PR is ready and tested under cpu, intel/xpu, and nvidia/cuda. However, please note this PR is 1 of 3 in a cumulative set of 3 prs where the primary PR is in Optimum PR huggingface/optimum#2064 (awaiting review/approval) while this and Transformer PR huggingface/transformers#35012 are dependent on getting the Optimum PR merged first. |
Thanks for the update. Let's wait for the optimum and the transformers PRs to be merged first, since changes there could affect this PR. Feel free to ping me as soon as those PRs were merged. |
Support gptqmodel, we plan to replace autogptq by gptqmodel in the future.