Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

MatMulIntegerToFloat reference update for tests #19333

Merged
merged 7 commits into from
Feb 15, 2024
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Original file line number Diff line number Diff line change
Expand Up @@ -882,7 +882,7 @@ struct OperatorDescTraits<DML_QUANTIZED_LINEAR_MATRIX_MULTIPLY_OPERATOR_DESC>
template <>
struct OperatorDescTraits<DML_MATRIX_MULTIPLY_INTEGER_TO_FLOAT_OPERATOR_DESC>
{
static constexpr DML_OPERATOR_TYPE Type = (DML_OPERATOR_TYPE) DML_OPERATOR_MATRIX_MULTIPLY_INTEGER_TO_FLOAT;
static constexpr DML_OPERATOR_TYPE Type = DML_OPERATOR_MATRIX_MULTIPLY_INTEGER_TO_FLOAT;
};

template <>
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -826,14 +826,6 @@ class QLinearMatMulHelper : public MatMulHelperBase
QLinearMatMulHelper(const Info_t& info, const Shape_t& shape) : MatMulHelperBase(info, shape, 0, 3) {}
};

class MatMulIntegerToFloatHelper : public MatMulHelperBase
{
public:
template<typename Info_t, typename Shape_t>
MatMulIntegerToFloatHelper(const Info_t& info, const Shape_t& shape) : MatMulHelperBase(info, shape, 0, 1) {}
};


class TopKHelper
{
void Initialize(
Expand Down Expand Up @@ -1752,7 +1744,7 @@ using ShapeInferenceHelper_Identity16 = GetOutputShapeAsInputShapeHelper;
using ShapeInferenceHelper_MatMul = MatMulHelper;
using ShapeInferenceHelper_MatMulInteger = MatMulHelper;
using ShapeInferenceHelper_DynamicQuantizeMatMul = MatMulHelper;
using ShapeInferenceHelper_MatMulIntegerToFloat = MatMulIntegerToFloatHelper;
using ShapeInferenceHelper_MatMulIntegerToFloat = MatMulHelper;
using ShapeInferenceHelper_QLinearMatMul = QLinearMatMulHelper;
using ShapeInferenceHelper_QLinearAdd = GetBroadcastedOutputShapeHelper;
using ShapeInferenceHelper_DynamicQuantizeLinear = GetOutputShapeAsInputShapeHelper;
Expand Down
371 changes: 305 additions & 66 deletions onnxruntime/test/contrib_ops/matmul_integer_to_float_test.cc

Large diffs are not rendered by default.

28 changes: 0 additions & 28 deletions onnxruntime/test/testdata/matmul_integer_to_float.py
Original file line number Diff line number Diff line change
Expand Up @@ -79,34 +79,6 @@ def GenerateModel(model_name, sign_i, sign_w, output_type_fp16, has_zp=True, bia

if __name__ == "__main__":
GenerateModel("matmul_integer_to_float16_int8.onnx", sign_i=False, sign_w=True, output_type_fp16=True)
GenerateModel("matmul_integer_to_float16_uint8.onnx", sign_i=False, sign_w=False, output_type_fp16=True)
GenerateModel(
"matmul_integer_to_float16_int8_bias.onnx",
sign_i=False,
sign_w=True,
output_type_fp16=True,
has_zp=False,
bias=True,
)
GenerateModel(
"matmul_integer_to_float16_uint8_bias.onnx",
sign_i=False,
sign_w=False,
output_type_fp16=True,
has_zp=False,
bias=True,
)

GenerateModel("matmul_integer_to_float16_int8_int8.onnx", sign_i=True, sign_w=True, output_type_fp16=True)
GenerateModel(
"matmul_integer_to_float16_int8_int8_bias.onnx",
sign_i=True,
sign_w=True,
output_type_fp16=True,
has_zp=False,
bias=True,
)

GenerateModel("matmul_integer_to_float_int8.onnx", sign_i=False, sign_w=True, output_type_fp16=False)
GenerateModel("matmul_integer_to_float_uint8.onnx", sign_i=False, sign_w=False, output_type_fp16=False)
GenerateModel(
Expand Down
51 changes: 0 additions & 51 deletions onnxruntime/test/testdata/matmul_integer_to_float16_int8.onnx

This file was deleted.

49 changes: 0 additions & 49 deletions onnxruntime/test/testdata/matmul_integer_to_float16_int8_bias.onnx

This file was deleted.

51 changes: 0 additions & 51 deletions onnxruntime/test/testdata/matmul_integer_to_float16_int8_int8.onnx

This file was deleted.

This file was deleted.

Loading