Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Evaluate on a single 24GB/32GB GPU #29

Open
lemyx opened this issue Jan 16, 2024 · 1 comment
Open

Evaluate on a single 24GB/32GB GPU #29

lemyx opened this issue Jan 16, 2024 · 1 comment

Comments

@lemyx
Copy link

lemyx commented Jan 16, 2024

Hi, on a single 4090 GPU with 24GB memory, the following command will cause out-of-memory.

python main.py mmlu --model_name llama --model_path huggyllama/llama-7b

After that, I try executing the command on A100-40GB, the nvidia-smi result is
image

It seems that neither 4090/3090 with 24GB memory or V100 with 32GB memory cannot test Llama-7B on mmlu under above command.

So how to evaluate llama-7b on mmlu on 24GB or 32GB GPU? any more options to enable?

Thanks

@lemyx
Copy link
Author

lemyx commented Jan 17, 2024

It seems that the CUDA memory will increase during execution of the script

image

Maybe related to maximum sequence length
image

Finally, the inference can be finished on a single A100-40GB card
image

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant