-
Notifications
You must be signed in to change notification settings - Fork 553
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Feature] If i want to start Fine-tuning, must the flashattention be installed? #776
Comments
It is not required; however, installing it will significantly improve the training speed. |
does this code must use Ampere GPU? |
you can try to set |
my file is the same as you |
{ |
+1,how can I finetune the model without the flash-attn? |
to install Flash_attn, please pull Nvidia devel image, refer in: https://catalog.ngc.nvidia.com/orgs/nvidia/containers/cuda/tags Example: nvcr.io/nvidia/cuda:12.6.3-cudnn-devel-ubuntu20.04 |
Thank you so much! Help a lot! |
I have another question regarding the use of LoRA. Do I only need to set --freeze_backbone False? However, after training, I found that in the config.json, "use_backbone_lora": 0. What else should I do if I want to finetune the visual encoder as well? Thank you in advance for your help! |
Motivation
GPUS=2 PER_DEVICE_BATCH_SIZE=1 sh ./****2_5_2b_lora.sh
mistake
FlashAttention is not installed.
FlashAttention is not installed.
flash-attention
package not found, consider installing for better performance: No module named 'flash_attn'.Current
flash-attenton
does not supportwindow_size
. Either upgrade or useattn_implementation='eager'
.flash-attention
package not found, consider installing for better performance: No module named 'flash_attn'.Current
flash-attenton
does not supportwindow_size
. Either upgrade or useattn_implementation='eager'
.The text was updated successfully, but these errors were encountered: