Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Solution - CUDA capability sm_86 #8

Open
int-rnd opened this issue Jul 25, 2023 · 0 comments
Open

Solution - CUDA capability sm_86 #8

int-rnd opened this issue Jul 25, 2023 · 0 comments

Comments

@int-rnd
Copy link

int-rnd commented Jul 25, 2023

So I had some limited access to some Tesla A10s - they are of course sm_86 and not compatible...

But if you want to make them work just add compatibly and build your own docker image locally.

Dockerfile

FROM devforth/gpt-j-6b-gpu
RUN pip3 install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cu118 -U
CMD uvicorn web:app --port 8080 --host 0.0.0.0

Then run
docker build -t CustomImageName .
and run
docker run -p8080:8080 --gpus all --rm -it CustomImageName

Credit to this project and https://pytorch.org/get-started/locally/

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant