-
Notifications
You must be signed in to change notification settings - Fork 70
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Quark quantization inference errors #137
Comments
|
Hi cyndwith, I've tried to convert onnx model with opset=11 again. This issue is gone. (Squeeze-11/Squeeze-13 are with different structures) KernelParamGenPass.cpp:2130] xir::Op{name = (Squeeze_output_0_DequantizeLinear_Output), type = transpose}. This order: (0,3,2, And then it just finished without anything. Thanks |
Please try exporting/update the ONNX model to use opset=17, which is the recommended version for Quark quantization. |
Hi cyndwith, |
It should function with opset_version=17. Could you provide more information about the network so I can replicate this error? |
Dear great authors,
I've tried to use quark quantization to generate int8 model under NPU.
However, while inferring the model, it shows the following errors:
The snippet model is like: (Use onnx.load to check the node info and it seems there's attribute generated after quark quantization )
Error Unrecognized attribute: axes for operator Squeeze (Actually the original model is split but is converted into slice)
Could you please help to comment it?
Thanks
The text was updated successfully, but these errors were encountered: