Skip to content

About the GPU memory usage #62

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Open
prokaryote-hgy opened this issue Jul 14, 2024 · 5 comments
Open

About the GPU memory usage #62

prokaryote-hgy opened this issue Jul 14, 2024 · 5 comments

Comments

@prokaryote-hgy
Copy link

Hi, wonderful work!

I wonder what's the peak memory usage when training the model with HR size of 512 and batchsize of 1 for one GPU. It seams that a 24GB GPU not enough to run the training? Or I did something wrong?

@Renzhihan
Copy link

I can train the model on 1 3090 with bs=1. But when I use 2 or more GPUs, the memory is not enough.

@prokaryote-hgy
Copy link
Author

I can train the model on 1 3090 with bs=1. But when I use 2 or more GPUs, the memory is not enough.

The same question.

@liyuantsao
Copy link

I can train the model on 1 3090 with bs=1. But when I use 2 or more GPUs, the memory is not enough.

+1

@cswry
Copy link
Owner

cswry commented Jul 22, 2024

I can train the model on 1 3090 with bs=1. But when I use 2 or more GPUs, the memory is not enough.

Multi-GPU parallel training consumes additional memory compared to single-GPU training, which is likely the cause.

@Moonsofang
Copy link

请问bs参数是在哪里设置啊?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

5 participants