Model quantization #17906
Unanswered
AndreiNanescu
asked this question in
Other Q&A
Model quantization
#17906
Replies: 2 comments
-
Are you using ORT format model for a mobile scenario? |
Beta Was this translation helpful? Give feedback.
0 replies
-
Converting to an ORT format model is the last step for mobile scenarios where binary size is critical. The tools work on ONNX models to be more general purpose, so please quantize the ONNX model and convert to ORT format after that as needed. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hello, I have a question regarding model quantization. I am currently using an ORT format model and would like to know if the quantization tools work with it, or if I need to quantize the ONNX model and then convert it to the ORT format. Thanks!
Beta Was this translation helpful? Give feedback.
All reactions