-
Notifications
You must be signed in to change notification settings - Fork 29
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
error when using fp8 #412
Comments
This is primarily for reproduction purposes, allowing the use of a deterministic algorithm whenever possible. You may set this to False if needed. Could you let me know which device you are using? We have tested this on A100 and Gaudi, and both work fine. |
H100. As far as I know, FP8 is not supported on a100. How do you conduct Fp8 quantization on A100? or is it supported on A100 already? |
BTW, I have a question about int_sym. why doesn't the
|
Although the quantized model cannot run on an A100, the tuning process can still be performed on an A100. |
there is a cast op in |
yes |
This variant, known as Full Range Sym, is detailed in our blog |
@WeiweiZhang1 please help add an arg to disable the use_deterministic_algorithms |
workaround #417 |
Hi, I tried to do quantization with FP8, and I met the following error:
Why set
torch.use_deterministic_algorithms(True, warn_only=True)
?The text was updated successfully, but these errors were encountered: