You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I have been using the Yunet model and tried the quantized version to speed up inference but I got slower results, both in my own code and trying your demo. I checked the benchmarks and the quantized model is announced to be slower too : is this an expected behavior ?
For context I use default backend and I have a Intel(R) Core(TM) i5-10300H CPU. To be clear, I loaded the int8 onnx file from the github and did not use the quantization script.
The text was updated successfully, but these errors were encountered:
Hello ! Thanks for the great work.
I have been using the Yunet model and tried the quantized version to speed up inference but I got slower results, both in my own code and trying your demo. I checked the benchmarks and the quantized model is announced to be slower too : is this an expected behavior ?
For context I use default backend and I have a Intel(R) Core(TM) i5-10300H CPU. To be clear, I loaded the int8 onnx file from the github and did not use the quantization script.
The text was updated successfully, but these errors were encountered: