You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
@lucidrains Thanks for answering me. Really appreciated!
Causal self-attention requires a triangle-like attention mask to mask out future tokens. In this code, I did find the interface you left for input_mask.
However, I didn't find the spot you initiate the attention mask. That's something confused me.
Found out that Efficient attention doesn't work on Causal attention scenario as mentioned here.
cmsflash/efficient-attention#4
So I doubt if the causal in this code really works?
The text was updated successfully, but these errors were encountered: