From 1c25fe55800b05ffd2d89d2c6f69b39820bc5c4f Mon Sep 17 00:00:00 2001 From: Vincent Wang Date: Tue, 31 Oct 2023 13:53:11 +0800 Subject: [PATCH] Fix PoliCheck (#18180) Fix PoliCheck by changing some words, which was from Triton flash attention's original code. --- .../python/training/ort_triton/kernel/_flash_attn.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/orttraining/orttraining/python/training/ort_triton/kernel/_flash_attn.py b/orttraining/orttraining/python/training/ort_triton/kernel/_flash_attn.py index 40398b33d8f04..03bb0f4373d8d 100644 --- a/orttraining/orttraining/python/training/ort_triton/kernel/_flash_attn.py +++ b/orttraining/orttraining/python/training/ort_triton/kernel/_flash_attn.py @@ -393,7 +393,7 @@ def _bwd_kernel_one_col_block( dk = tl.zeros([BLOCK_N, BLOCK_HEADDIM], dtype=tl.float32) # There seems to be some problem with Triton pipelining that makes results wrong for # headdim=64, seqlen=(113, 255), bias_type='matrix'. In this case the for loop - # may have zero step, and pipelining with the bias matrix could screw it up. + # may have zero step, and pipelining with the bias matrix could cause the problem. # So we just exit early. if begin_m >= seqlen_q: dv_ptrs = DV + (offs_n[:, None] * stride_dvn + offs_d[None, :])