[Performance] OpenVINO EP Produces incorrect inference results #20357
Labels
ep:OpenVINO
issues related to OpenVINO execution provider
performance
issues related to performance regressions
quantization
issues related to quantization
stale
issues that have not been addressed in a while; categorized by a bot
Describe the issue
When using
onnxruntime-openvino==1.16.0
on python3.8, I'm seeing inference results that do not match what the model produces in PyTorch, for certain models. I'm running with theOpenVINOExecutionProvider
and using theGPU
device. This seems to be a regression, as when I run with version 1.14.0, I do not see the difference in scores.This is possibly related to #19975, where I previously left a comment.
To reproduce
See below for a script that reproduces the issue. The results of the
BrokenModel
will disagree, but the results of the other two models will be very close, if not the same.You'll need to install
torch
,onnxruntime-openvino
, andnumpy
to run this script.Urgency
No response
Platform
Linux
OS Version
Ubuntu 20.04
ONNX Runtime Installation
Released Package
ONNX Runtime Version or Commit ID
1.16.0
ONNX Runtime API
Python
Architecture
X64
Execution Provider
OpenVINO
Execution Provider Library Version
bundled with
onnxruntime-openvino
Model File
No response
Is this a quantized model?
Yes
The text was updated successfully, but these errors were encountered: