From cb5bc55cff5badcd5a4858f67b686d665e66b127 Mon Sep 17 00:00:00 2001 From: Bill Teng Date: Fri, 1 Dec 2023 04:57:00 +0000 Subject: [PATCH] #0: disable layernorm pytest due to WH error --- .../unit_testing/test_single_core_fused_ops.py | 1 + 1 file changed, 1 insertion(+) diff --git a/tests/tt_eager/python_api_testing/unit_testing/test_single_core_fused_ops.py b/tests/tt_eager/python_api_testing/unit_testing/test_single_core_fused_ops.py index 78aa982ba43..af32de829df 100644 --- a/tests/tt_eager/python_api_testing/unit_testing/test_single_core_fused_ops.py +++ b/tests/tt_eager/python_api_testing/unit_testing/test_single_core_fused_ops.py @@ -36,6 +36,7 @@ def test_softmax(shape, device): @pytest.mark.parametrize("shape", [[1, 1, 32, 32], [1, 1, 32, 128]]) +@skip_for_wormhole_b0("Watcher error, see issue#4107") def test_layernorm(shape, device): torch.manual_seed(1234) x = torch.randn(shape).bfloat16().float()