Skip to content

Commit

Permalink
lint
Browse files Browse the repository at this point in the history
  • Loading branch information
chenfucn committed Jul 2, 2024
1 parent 3fc508c commit b845171
Show file tree
Hide file tree
Showing 2 changed files with 2 additions and 2 deletions.
2 changes: 1 addition & 1 deletion onnxruntime/contrib_ops/cuda/quantization/matmul_nbits.cc
Original file line number Diff line number Diff line change
Expand Up @@ -40,7 +40,7 @@ Status MatMulNBits<MLFloat16>::PrepackedGemm(
zero_points_ptr, zero_points_size,
Y->MutableData<MLFloat16>(), Y->Shape().Size());
}
#endif // !USE_ROCM
#endif // !USE_ROCM

template <typename T>
Status MatMulNBits<T>::ComputeInternal(OpKernelContext* ctx) const {
Expand Down
2 changes: 1 addition & 1 deletion onnxruntime/contrib_ops/cuda/quantization/matmul_nbits.h
Original file line number Diff line number Diff line change
Expand Up @@ -42,7 +42,7 @@ class MatMulNBits final : public CudaKernel {
[[maybe_unused]] Tensor* Y) const {
return ORT_MAKE_STATUS(ONNXRUNTIME, NOT_IMPLEMENTED, "Prepacked gemm is not supported for MatMulNBits op.");
}
#endif // !USE_ROCM
#endif // !USE_ROCM

Status ComputeInternal(OpKernelContext* context) const override;

Expand Down

0 comments on commit b845171

Please sign in to comment.