You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
There seems to be a bug when training LoRA currently where the amount of memory consumed by LoRA exceeds that of standard full fine-tuning. This issue is exacerbated when training an 8B model with FSDP, but can be observed with DeepSpeed as well.
The text was updated successfully, but these errors were encountered:
There seems to be a bug when training LoRA currently where the amount of memory consumed by LoRA exceeds that of standard full fine-tuning. This issue is exacerbated when training an 8B model with FSDP, but can be observed with DeepSpeed as well.
The text was updated successfully, but these errors were encountered: