-
Notifications
You must be signed in to change notification settings - Fork 39
About the GPU memory usage #62
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
I can train the model on 1 3090 with bs=1. But when I use 2 or more GPUs, the memory is not enough. |
The same question. |
+1 |
Multi-GPU parallel training consumes additional memory compared to single-GPU training, which is likely the cause. |
请问bs参数是在哪里设置啊? |
Hi, wonderful work!
I wonder what's the peak memory usage when training the model with HR size of 512 and batchsize of 1 for one GPU. It seams that a 24GB GPU not enough to run the training? Or I did something wrong?
The text was updated successfully, but these errors were encountered: