Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

OOM Issues in MMLU Evaluation with lm_eval Using vllm as Backend #2490

Open
wchen61 opened this issue Nov 14, 2024 · 0 comments
Open

OOM Issues in MMLU Evaluation with lm_eval Using vllm as Backend #2490

wchen61 opened this issue Nov 14, 2024 · 0 comments

Comments

@wchen61
Copy link

wchen61 commented Nov 14, 2024

I have recently encountered frequent OOM issues while running MMLU evaluation tasks using lm_eval with vllm as the backend. After some investigation, I found that this issue arises because the sampling parameters used by lm_eval are inconsistent with the default parameters used by vllm to estimate peak memory usage. In fact, the memory used by lm_eval during actual execution is much higher than the peak memory estimated by vllm (for the Meta-Llama-8B-Instruct model, using batch size auto, 10GB vs 50GB).

I have already reported this issue to vllm to see if they can provide a way for lm_eval to configure the default sampling parameters to resolve this issue. In the meantime, modifying the default sampling parameters in vllm could temporarily bypass this OOM issue, and I believe this workaround could be helpful for others encountering the same problem.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant