Skip to content

Multi-GPU model loading issue after v0.0.11 #418

Answered by dagbdagb
LPCTSTR asked this question in Q&A
Discussion options

You must be logged in to vote

To expand on this: My GPUs are numbered:
0: Quadro RTX 3000 (6GB)
1: NVIDIA GeForce RTX 3090
2: NVIDIA GeForce RTX 3090

If I do:
export CUDA_VISIBLE_DEVICES=2,1,0
... before starting exui, and set up my split as '4,21,23', I get to load Llama-3-70B-Instruct-exl2-5.0bpw successfully.

GPU VRAM is then loaded in the order 0,2,1, for some reason.

dagb@p53 ~ $ nvidia-smi 
Wed Apr 24 02:38:54 2024       
+-----------------------------------------------------------------------------------------+
| NVIDIA-SMI 550.67                 Driver Version: 550.67         CUDA Version: 12.4     |
|-----------------------------------------+------------------------+----------------------+
| GPU  Name        …

Replies: 2 comments 3 replies

Comment options

You must be logged in to vote
0 replies
Comment options

You must be logged in to vote
3 replies
@dagbdagb
Comment options

Answer selected by LPCTSTR
@LPCTSTR
Comment options

@dagbdagb
Comment options

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
3 participants