-
-
Notifications
You must be signed in to change notification settings - Fork 221
Issues: turboderp/exllama
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
When will the bfloat16 type of GPTQ algorithm be supported?
#310
opened Dec 20, 2023 by
Kelang-Tian
updated Dec 20, 2023
Does it support safetytensor formate?>
#309
opened Nov 28, 2023 by
lucasjinreal
updated Nov 28, 2023
Using Exllama backend requires all the modules to be on GPU - how?
#306
opened Nov 6, 2023 by
tigerinus
updated Nov 6, 2023
CodeLLaMA + LoRA: RuntimeError: CUDA error: an illegal memory access was encountered
#290
opened Sep 15, 2023 by
juanps90
updated Oct 12, 2023
OSError: CUDA_HOME environment variable is not set.
#291
opened Sep 17, 2023 by
jamesbraza
updated Sep 29, 2023
Changing hyper-parameters after initilization without reloading weights from disk.
#299
opened Sep 28, 2023 by
kmccleary3301
updated Sep 28, 2023
Tried to build setup exllama but encountering ninja related errors, can someone please help me?
#258
opened Aug 22, 2023 by
BwandoWando
updated Sep 25, 2023
GPU Usage Keeps High Even Without Inference Load
#253
opened Aug 19, 2023 by
leonxia1018
updated Sep 13, 2023
Progress on the rewrite for older cards (Like the P40)
#279
opened Sep 8, 2023 by
TimyIsCool
updated Sep 10, 2023
RoPE Frequency Base and Frequency Scale Support
#262
opened Aug 28, 2023 by
ChrisCates
updated Sep 9, 2023
Previous Next
ProTip!
Exclude everything labeled
bug
with -label:bug.