You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Describe the bug
Finetuning in 2:4 sparsity w4a16 example fails with multiple GPUs
Expected behavior
The finetuning step expected to train successfully with multi GPUs
Environment
Using four NVIDIA A10 GPUs on aws notebook instance
To Reproduce
cd examples/quantization_2of4_sparse_w4a16
python llama7b_sparse_w4a16.py
Errors
RuntimeError: Expected all tensors to be on the same device, but found at least two devices, cuda:3 and cuda:0!
Additional context
I have seen example of examples/finetuning/example_fsdp_config.yaml
But not sure how to use it for finetuning in the above example.
Also would like to know if there is any paper that discusses this approach?
Is the finetuning being done on the whole model and in what precision?
Will finetuning of quantized models further with lora adapters be supported?
Thanks
Arun
The text was updated successfully, but these errors were encountered:
Thanks for this nice repo.
Describe the bug
Finetuning in 2:4 sparsity w4a16 example fails with multiple GPUs
Expected behavior
The finetuning step expected to train successfully with multi GPUs
Environment
Using four NVIDIA A10 GPUs on aws notebook instance
To Reproduce
cd examples/quantization_2of4_sparse_w4a16
python llama7b_sparse_w4a16.py
Errors
RuntimeError: Expected all tensors to be on the same device, but found at least two devices, cuda:3 and cuda:0!
Additional context
I have seen example of examples/finetuning/example_fsdp_config.yaml
But not sure how to use it for finetuning in the above example.
Also would like to know if there is any paper that discusses this approach?
Is the finetuning being done on the whole model and in what precision?
Will finetuning of quantized models further with lora adapters be supported?
Thanks
Arun
The text was updated successfully, but these errors were encountered: