Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

40 GB model #6

Open
DHOFM opened this issue Mar 23, 2023 · 2 comments
Open

40 GB model #6

DHOFM opened this issue Mar 23, 2023 · 2 comments

Comments

@DHOFM
Copy link

DHOFM commented Mar 23, 2023

Hi,
thanks for your nice repo. You mention 2 3090:
`

The following hardware is needed to run different models in MiniLLM:

Model GPU Memory Requirements Compatible GPUs
llama-7b-4bit 6GB RTX 2060, 3050, 3060
llama-13b-4bit 10GB GTX 1080, RTX 2060, 3060, 3080
llama-30b-4bit 20GB RTX 3080, A5000, 3090, 4090, V100
llama-65b-4bit 40GB A100, 2x3090, 2x4090, A40, A6000

`

So when I try the 60B Version with 2 RTX 3090 I get an OOM - how can I use both GPUs?

Kind regards,

Dirk

@kuleshov
Copy link
Owner

Ah, sorry, I haven't implemented that feature yet :) happy to merge a PR if someone does that first

@kuleshov
Copy link
Owner

for what it's worth, i was running the 65B model on one a6000

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants