Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[BUG] Failed Beginner Installation: Unable to gather CUDA version on (Manjaro) Arch Linux #87

Closed
Endebert opened this issue Mar 20, 2024 · 1 comment
Labels
bug Something isn't working

Comments

@Endebert
Copy link

Just want to report that the "beginner" installation with a clean clone of the repo does not work properly on all linux systems, at least not on mine. In my case, it failed to infer the CUDA version and as a result did not install all required dependencies.

The issue is that my CUDA installation is in /opt/cuda, therefore the installation script cannot infer the CUDA version and proceeds with using requirements-nowheel.txt, which does not include exllamav2.

System Info:

OS: Manjaro Linux x86_64 
Kernel: 6.7.7-1-MANJARO
CPU: AMD Ryzen 9 3900X (12) @ 3.800GHz
GPU: NVIDIA GeForce RTX 4090
Shell: fish 3.7.0

Python 3.11.8
repo @ commit b74603db59ee99712439a38da2550ab692eec52d

The issue primarily seems to be this:

$ echo $CUDA_PATH
/opt/cuda

Therefore, cuda_version = pathlib.Path(CUDA_PATH).name from the python code does not work for my system.

I'm using the regular CUDA package installed via pacman: cuda-12.3.2-1, so nothing special.

I assume there must be another way to infer the installed cuda version, though it doesn't seem as trivial as expected. I found this related StackOverflow question: https://stackoverflow.com/q/9727688

Parsing nvcc output might be viable, but apparently that's not always available:

$ nvcc -V
nvcc: NVIDIA (R) Cuda compiler driver
Copyright (c) 2005-2023 NVIDIA Corporation
Built on Wed_Nov_22_10:17:15_PST_2023
Cuda compilation tools, release 12.3, V12.3.107
Build cuda_12.3.r12.3/compiler.33567101_0

Or parsing cublas_version.txt from the installation, though I don't know if that is platform agnostic.

$ cat /opt/cuda/cublas_version.txt 
CUBLAS Version 12.3.4.1

Maybe there's some python code floating around somewhere that does it properly for all systems, but I haven't found it. It seems other projects either require a manual platform-specific installation of torch, or have an interactive installation script which asks for the GPU vendor.

@Endebert Endebert added the bug Something isn't working label Mar 20, 2024
@kingbri1
Copy link
Member

kingbri1 commented Mar 21, 2024

Fixed in #88. Closing. Please see #49 for the full reasoning.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working
Projects
None yet
Development

No branches or pull requests

2 participants