You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
This turned out to be an user error from my side. The input data was in the wrong format for the int8 quantization (I wanted to load int8 data but float was the expected). I'm closing this issue.
I'm trying to load a resnet50 model with quantize_int8 using calibration data, but getting the following error:
LLVM ERROR: Expected to find GEMM, convolution, or attention op, and didn't
The error comes from MLIR:
https://github.com/ROCm/rocMLIR/blob/45d830dbbc15fe84c41d95585f526a50719020eb/mlir/lib/Dialect/Rock/Tuning/RockTuningImpl.cpp#L493
Description:
There is a description in the script about what it changes in the model: https://github.com/ROCm/migraphx-mlperf/blob/mlperf_resurrection/prototypes/inference_v2/resnet50/scripts/rn50_graphsurgeon.py#L322-L366
cc @TedThemistokleous @pfultz2
I can compile the list of reproduce steps or a pre-built docker image if needed
The text was updated successfully, but these errors were encountered: