Questions about using XNNPACK Execution Provider #18112
Unanswered
postech-sms
asked this question in
Other Q&A
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hi, all
I have questions about using XNNPACK Execution Provider. (linux-arm64 ubuntu:20.04)
I built onnxruntime including xnnpack from the v1.16.0 tag in the onnxruntime repository. (use --use_xnnpack)
Using my built files,
I checked that XNNExecutionProvider works well For the Full Precision Model.
(I mean that convolution layers are mapped into XNNExecutionProvider)
However, when I converted Full Precision Model to INT8 Quantization Model,
Convolution layers weren't mapped into XNNExecutionProvider
(Just mapped into CPUExecutionProvider)
I confirmed in the official documentation that XNNExecutionProvider also supports QLinearConv. (ai.onnx:QLinearConv)
(https://github.com/microsoft/onnxruntime/blob/gh-pages/docs/execution-providers/Xnnpack-ExecutionProvider.md#supported-ops)
But I don't understand why XNNPackExecutionProvider doesn't work properly in my INT8 model.
This may be an issue that occurred because I am a newbie, so please understand.
Thanks.
Beta Was this translation helpful? Give feedback.
All reactions