-
-
Notifications
You must be signed in to change notification settings - Fork 1.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Issue] Triton Compilation Error in Unsloth Fine-Tuning Script on Kernel 5.4.0 #1336
Comments
same issue on V100. but it works fine on T4 |
I also have a V100 and I'm getting this error too. |
I also encountered the same error on v100. |
Temporary solution:
wait for the final solution. |
Which torch version did you use? It seems that torch 2.5.1 isn't compatible
|
@hykilpikonna pytorh '2.4.0+cu121' |
Apologies everyone! @LiaoPan @hykilpikonna @ergosumdre @gityeop @chengju-zhou I added a flag to disable some other kernels - I'm unsure if it worked though. Torch 2.5 and torch 2.4 should be now supported - sadly Colab got rid of V100s so I can't test them - so I'm assuming a specific kernel from Apple's Cut Cross Entropy package is the one causing the issues. Please try updating Unsloth without dependencies if that works!
|
By the way to get Torch 2.4 - simply run |
Confirmed working. Thanks!
|
Hmm... unsloth 2024.11.11 + unsloth-zoo 2024.11.8 is showing the same error for me.
Install log:
pip freeze: accelerate==1.0.1
aiofiles==23.2.1
aiohappyeyeballs==2.4.3
aiohttp==3.11.7
aiosignal==1.3.1
annotated-types==0.7.0
anyio==4.6.2.post1
attrs==24.2.0
av==13.1.0
bitsandbytes==0.44.1
certifi==2024.8.30
charset-normalizer==3.4.0
click==8.1.7
contourpy==1.3.1
cut-cross-entropy==24.11.4
cycler==0.12.1
datasets==3.1.0
dill==0.3.8
docstring_parser==0.16
einops==0.8.0
fastapi==0.115.5
ffmpy==0.4.0
filelock @ file:///home/conda/feedstock_root/build_artifacts/filelock_1726613473834/work
fire==0.7.0
fonttools==4.55.0
frozenlist==1.5.0
fsspec==2024.9.0
gmpy2 @ file:///home/conda/feedstock_root/build_artifacts/gmpy2_1725379831219/work
gradio==4.44.1
gradio_client==1.3.0
h11==0.14.0
hf_transfer==0.1.8
httpcore==1.0.7
httpx==0.27.2
huggingface-hub==0.26.2
idna==3.10
importlib_resources==6.4.5
jieba==0.42.1
Jinja2 @ file:///home/conda/feedstock_root/build_artifacts/jinja2_1715127149914/work
joblib==1.4.2
kiwisolver==1.4.7
-e git+https://github.com/hiyouga/LLaMA-Factory.git@00031b1a66ade1c2665ce7a069a756cccbcb07f1#egg=llamafactory
markdown-it-py==3.0.0
MarkupSafe==2.1.5
matplotlib==3.9.2
mdurl==0.1.2
modelscope==1.20.1
mpmath @ file:///home/conda/feedstock_root/build_artifacts/mpmath_1678228039184/work
multidict==6.1.0
multiprocess==0.70.16
networkx @ file:///home/conda/feedstock_root/build_artifacts/bld/rattler-build_networkx_1731521053/work
nltk==3.9.1
numpy==1.26.4
orjson==3.10.12
packaging==24.2
pandas==2.2.3
peft==0.12.0
pillow==10.4.0
propcache==0.2.0
protobuf==3.20.3
psutil==6.1.0
pyarrow==18.1.0
pydantic==2.10.1
pydantic_core==2.27.1
pydub==0.25.1
Pygments==2.18.0
pyparsing==3.2.0
python-dateutil==2.9.0.post0
python-multipart==0.0.17
pytz==2024.2
PyYAML @ file:///home/conda/feedstock_root/build_artifacts/pyyaml_1725456139051/work
regex==2024.11.6
requests==2.32.3
rich==13.9.4
rouge-chinese==1.0.3
ruff==0.8.0
safetensors==0.4.5
scipy==1.14.1
semantic-version==2.10.0
sentencepiece==0.2.0
shellingham==1.5.4
shtab==1.7.1
six==1.16.0
sniffio==1.3.1
sse-starlette==2.1.3
starlette==0.41.3
sympy==1.13.1
termcolor==2.5.0
tiktoken==0.8.0
tokenizers==0.20.3
tomlkit==0.12.0
torch==2.5.1
tqdm==4.67.1
transformers==4.46.1
triton==3.1.0
trl==0.8.6
typeguard==4.4.1
typer==0.13.1
typing_extensions @ file:///home/conda/feedstock_root/build_artifacts/typing_extensions_1717802530399/work
tyro==0.8.14
tzdata==2024.2
unsloth==2024.11.11
unsloth_zoo==2024.11.8
urllib3==2.2.3
uvicorn==0.32.1
websockets==12.0
xformers==0.0.28.post3
xxhash==3.5.0
yarl==1.18.0 |
Using Torch 2.4.0 did not solve the issue. Complete installation: conda create --name unsloth_240 python=3.10 pytorch=2.4.0 pytorch-cuda=12.1 \
cudatoolkit xformers -c pytorch -c nvidia -c xformers -y
conda activate unsloth_240
pip install "unsloth[colab-new] @ git+https://github.com/unslothai/unsloth.git"
pip install --no-deps "trl<0.9.0" peft accelerate bitsandbytes
# Then installed LLaMA-Factory
cd LLaMA-Factory
pip install -e ".[torch,metrics]" Using Torch 2.3.0 with Triton 2.3.0 led to a different error:
|
@hykilpikonna Sorry about that - do you know GPU, CUDA version etc - could you take a screenshot of the stats section Unsloth prints out thanks |
This is my setup: |
My GPU and CUDA version:
Torch 2.4.0 Stats section:
Torch 2.5.1:
Since you said you don't have access to a V100 to test things, I would be happy to give you access to my system if you want to debug anything. If you can send me a ssh public key I'll give you ssh access. |
Description
When trying to run Unsloth fine-tuning script, encountering a Triton compilation error related to ReduceOpToLLVM.cpp.
Error Message
System Information
Code
Additional Context
Steps to Reproduce
Questions
The text was updated successfully, but these errors were encountered: