We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 570daa1 commit feff99fCopy full SHA for feff99f
python/paddle/nn/functional/flash_attention.py
@@ -81,7 +81,7 @@ def _math_attention(
81
82
83
def _select_sdp_cuda(head_dim):
84
- if head_dim < 128:
+ if head_dim <= 128:
85
return "flash_attn"
86
else:
87
return "mem_efficient"
0 commit comments