You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I am trying to use GPTQ quantization support on my Jetson Orin Nano Super. The docs say that I can use either auto-gptq or gptqmodel for this. However, when only installing gptqmodel - since auto-gptq is supposed to be the deprecated version - I get the error:
ImportError: Loading a GPTQ quantized model requires the auto-gptq library (`pip install auto-gptq`)
Am I doing something wrong or are the docs not up to date?
Thanks!
Lukas
System Info:
Machine: aarch64
System: Linux
Distribution: Ubuntu 22.04 Jammy Jellyfish
Release: 5.15.148-tegra
Python: 3.10.12
CUDA: 12.6.68
Information
The official example scripts
My own modified scripts
Tasks
An officially supported task in the examples folder (such as GLUE/SQuAD, ...)
ENV: Auto setting PYTORCH_CUDA_ALLOC_CONF='expandable_segments:True' for memory saving.
ENV: Auto setting CUDA_DEVICE_ORDER=PCI_BUS_ID for compatibililty.
Traceback (most recent call last):
File "/home/lukas/Development/db-agent/test.py", line 39, in <module>
quantized_model = AutoModelForCausalLM.from_pretrained(model_id, device_map="cuda", quantization_config=gptq_config)
File "/home/lukas/Development/db-agent/venv/lib/python3.10/site-packages/transformers/models/auto/auto_factory.py", line 564, in from_pretrained
return model_class.from_pretrained(
File "/home/lukas/Development/db-agent/venv/lib/python3.10/site-packages/transformers/modeling_utils.py", line 3620, in from_pretrained
hf_quantizer.validate_environment(
File "/home/lukas/Development/db-agent/venv/lib/python3.10/site-packages/transformers/quantizers/quantizer_gptq.py", line 59, in validate_environment
raise ImportError(
ImportError: Loading a GPTQ quantized model requires the auto-gptq library (`pip install auto-gptq`)
Expected behavior
No error
The text was updated successfully, but these errors were encountered:
Yes you are right @lukazso, sorry it's not part of the release yet, It will be in the next release 4.49 in the coming days, for now you can just install the main branch, it's stable
I am trying to use GPTQ quantization support on my Jetson Orin Nano Super. The docs say that I can use either
auto-gptq
orgptqmodel
for this. However, when only installing gptqmodel - since auto-gptq is supposed to be the deprecated version - I get the error:Am I doing something wrong or are the docs not up to date?
Thanks!
Lukas
System Info:
Machine: aarch64
System: Linux
Distribution: Ubuntu 22.04 Jammy Jellyfish
Release: 5.15.148-tegra
Python: 3.10.12
CUDA: 12.6.68
Information
Tasks
examples
folder (such as GLUE/SQuAD, ...)Reproduction
Minimum example to reproduce:
Here the complete traceback:
Expected behavior
No error
The text was updated successfully, but these errors were encountered: