From c113d122861ad8360e863575e5835f60eeae76b3 Mon Sep 17 00:00:00 2001 From: Nitika Shanker Date: Tue, 26 Mar 2024 23:33:02 +0000 Subject: [PATCH] #6680: Reverting move op change --- tt_eager/tt_dnn/op_library/move/move_op.hpp | 3 +-- 1 file changed, 1 insertion(+), 2 deletions(-) diff --git a/tt_eager/tt_dnn/op_library/move/move_op.hpp b/tt_eager/tt_dnn/op_library/move/move_op.hpp index adbb75f8184..ca60eae31d8 100644 --- a/tt_eager/tt_dnn/op_library/move/move_op.hpp +++ b/tt_eager/tt_dnn/op_library/move/move_op.hpp @@ -138,8 +138,7 @@ inline Tensor move_sharded(Tensor& input_tensor, std::optional& me shard_mem_config.shard_spec = shard_spec; auto output_tensor = create_sharded_device_tensor(input_shape, input_dtype, input_layout, input_tensor.device(), shard_mem_config); if (input_tensor.buffer()->address() == output_tensor.buffer()->address()) { - cout << "move_op.hpp: WARNING!! No space to move the tensor. Move op's input address == output address. Not moving the tensor." << endl; - return output_tensor; + TT_FATAL(false, "No space to move the tensor. Move op's input address == output address. No-op move unsupported."); } MoveOpParallelizationStrategy move_op_parallelization_strategy = MoveOpParallelizationStrategy::MULTI_CORE_SHARDED; auto output = operation::run(Move{output_mem_config, move_op_parallelization_strategy}, {input_tensor, output_tensor}).at(0);