Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Failed to quantize ResNet50 model to int8 #3644

Closed
gyulaz-htec opened this issue Nov 20, 2024 · 1 comment
Closed

Failed to quantize ResNet50 model to int8 #3644

gyulaz-htec opened this issue Nov 20, 2024 · 1 comment

Comments

@gyulaz-htec
Copy link
Collaborator

gyulaz-htec commented Nov 20, 2024

I'm trying to load a resnet50 model with quantize_int8 using calibration data, but getting the following error: LLVM ERROR: Expected to find GEMM, convolution, or attention op, and didn't
The error comes from MLIR:
https://github.com/ROCm/rocMLIR/blob/45d830dbbc15fe84c41d95585f526a50719020eb/mlir/lib/Dialect/Rock/Tuning/RockTuningImpl.cpp#L493

Description:

cc @TedThemistokleous @pfultz2

I can compile the list of reproduce steps or a pre-built docker image if needed

@gyulaz-htec
Copy link
Collaborator Author

This turned out to be an user error from my side. The input data was in the wrong format for the int8 quantization (I wanted to load int8 data but float was the expected). I'm closing this issue.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant