[Performance] The 16-bit quantization QDQ model cannot be accelerated by CUDA #21478
Labels
ep:CUDA
issues related to the CUDA execution provider
performance
issues related to performance regressions
quantization
issues related to quantization
stale
issues that have not been addressed in a while; categorized by a bot
Describe the issue
My GPU is V100 CUDA Version: 12.0 or 11.8
CPU is Intel(R) Xeon(R) Gold 6271C CPU @ 2.60GHz
I tested the performance of
A8W8
andA16W16
quantization models onCPU
andCUDA
respectively. The performance ofA16W16
quantization model on CUDA is even worse than that of CPU.Summary:
Moreover, The
A16W8
orA8W16
quantization models also have the similar issues.To reproduce
This issue can be reproduced by using the relevant files in performance.zip. The reproduction commands and results are as follows,
then you will receive the following results:
Urgency
Urgent
Platform
Linux
OS Version
Ubuntu 22.04
ONNX Runtime Installation
Released Package
ONNX Runtime Version or Commit ID
1.18.1
ONNX Runtime API
Python
Architecture
X64
Execution Provider
CUDA
Execution Provider Library Version
CUDA12/CUDA11.8
Model File
No response
Is this a quantized model?
Yes
The text was updated successfully, but these errors were encountered: