You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
however for every single convolution, the weights are quantized correctly to int8; bias is not quantized to be int8 but int32 as seen by the screenshots. What would be causing this issue? I suspect this is the issue that causes my conversion to trt engine to fail. It fails when reaching the conv bias node
To reproduce
I run quantize_static on a yolor model.
Urgency
No response
Platform
Linux
OS Version
Ubuntu 22.04
ONNX Runtime Installation
Released Package
ONNX Runtime Version or Commit ID
1.16
ONNX Runtime API
Python
Architecture
X86
Execution Provider
Default CPU
Execution Provider Library Version
No response
The text was updated successfully, but these errors were encountered:
Describe the issue
I have quantized my model with these settings
however for every single convolution, the weights are quantized correctly to int8; bias is not quantized to be int8 but int32 as seen by the screenshots. What would be causing this issue? I suspect this is the issue that causes my conversion to trt engine to fail. It fails when reaching the conv bias node
To reproduce
I run quantize_static on a yolor model.
Urgency
No response
Platform
Linux
OS Version
Ubuntu 22.04
ONNX Runtime Installation
Released Package
ONNX Runtime Version or Commit ID
1.16
ONNX Runtime API
Python
Architecture
X86
Execution Provider
Default CPU
Execution Provider Library Version
No response
The text was updated successfully, but these errors were encountered: