Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

RTX 4090 finetune with CUDA_OUT_OF_memory #18

Closed
clpoz opened this issue Oct 30, 2024 · 2 comments
Closed

RTX 4090 finetune with CUDA_OUT_OF_memory #18

clpoz opened this issue Oct 30, 2024 · 2 comments

Comments

@clpoz
Copy link

clpoz commented Oct 30, 2024

As the title, I think it should be fine with 24GB cuda_memory, when I start finetuing, it shows only take 18GB, while 3 hours later, I found it show errors with cuda_out_of_memory. Anyone have ideas? I don't know why?

@clpoz
Copy link
Author

clpoz commented Oct 30, 2024

Ok, I found it, after did a validation and saved a checkpoint, the VRAM usage added about 3GB and never drop back, how could it happen? how could I change setting to solve it?

@little51
Copy link
Contributor

try reduce block_size or per_device_train_batch_size per_device_eval_batch_size

@clpoz clpoz closed this as completed Jan 7, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants