From 34a1490e069dc3ce2ce30c3a37615cbfec39f5a6 Mon Sep 17 00:00:00 2001 From: Gustavo de Rosa Date: Tue, 16 Jan 2024 16:05:38 +0000 Subject: [PATCH] Update modeling_phi.py --- modeling_phi.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/modeling_phi.py b/modeling_phi.py index 5fad744..62dd3b5 100644 --- a/modeling_phi.py +++ b/modeling_phi.py @@ -509,7 +509,7 @@ class PhiFlashAttention2(PhiAttention): value_states = value_states.to(target_dtype) attn_output = self._flash_attention_forward( - query_states, key_states, value_states, attention_mask, q_len, dropout=attn_dropout, softmax_scale=1.0 + query_states, key_states, value_states, attention_mask, q_len, dropout=attn_dropout, softmax_scale=None ) attn_output = attn_output.reshape(bsz, q_len, self.hidden_size).contiguous()