Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

12GB card #109

Open
arthurwolf opened this issue Jul 30, 2023 · 2 comments
Open

12GB card #109

arthurwolf opened this issue Jul 30, 2023 · 2 comments

Comments

@arthurwolf
Copy link

My card has 12GB of RAM, that's not a case covered anywhere i could see. Would this allow me to do more (run the larger models, etc)? Any chances to get instructions for larger cards?

Thanks!

@miko8422
Copy link

miko8422 commented Jan 1, 2024

You can run Quantized 7B model on your pc, but with the full version which is the not quantized version of 7B, you won't be able to run it, because it will literally eat 12G of RAM.
Just use a cloud server or just use the quantized version if you want to explore prompt engineering.

@miko8422
Copy link

miko8422 commented Jan 1, 2024

You can run Quantized 7B model on your pc, but with the full version which is the not quantized version of 7B, you won't be able to run it, because it will literally eat 12G of RAM. Just use a cloud server or just use the quantized version if you want to explore prompt engineering.

I also have 12G of RAM on my own pc, but I wasn't been able to run the offical 7B model. By the way, I'm using 4070... I start to hate why I didn't bought the 4090 to run this model because I wan't to do some tuning on it.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants