We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
机器配置:32G GPU
实验步骤:
问题: 在转onnx之前,模型可以被加载到显存并成功执行inference; 转onnx之后,无论是否继续转为精度fp16的模型,都因显存不足而报错。 请问,对于当前的32G显存gpu, 有什么方法,可以使用转化后的onnx模型来做预测吗?
The text was updated successfully, but these errors were encountered:
No branches or pull requests
机器配置:32G GPU
实验步骤:
使用GPU:模型加载完毕之前,程序会因显存不足报错;
使用CPU,能加载成功
问题:
在转onnx之前,模型可以被加载到显存并成功执行inference;
转onnx之后,无论是否继续转为精度fp16的模型,都因显存不足而报错。
请问,对于当前的32G显存gpu, 有什么方法,可以使用转化后的onnx模型来做预测吗?
The text was updated successfully, but these errors were encountered: