You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
it inserts is_causal argument twice, from kwargs and explicitely. causes TypeError: transformers.modeling_flash_attention_utils._flash_attention_forward() got multiple values for keyword argument 'is_causal'
the kwargs get is_causal from GPT2Attention.forward()
Expected behavior
use is_causal just once
please add GPT2 to your release tests suite
The text was updated successfully, but these errors were encountered:
System Info
4.48.dev ubuntu18, py3.11
Who can help?
@ArthurZucker based on #35235
@Cyrilvallez based on #35342
Information
Tasks
examples
folder (such as GLUE/SQuAD, ...)Reproduction
use GPT2 with flash attn.
in
transformers/src/transformers/integrations/flash_attention.py
Line 47 in 608e163
it inserts
is_causal
argument twice, from kwargs and explicitely. causesTypeError: transformers.modeling_flash_attention_utils._flash_attention_forward() got multiple values for keyword argument 'is_causal'
the kwargs get
is_causal
from GPT2Attention.forward()Expected behavior
use
is_causal
just onceplease add GPT2 to your release tests suite
The text was updated successfully, but these errors were encountered: