Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Improvements to attention mask handling #111

Merged
merged 1 commit into from
Oct 20, 2023
Merged

Improvements to attention mask handling #111

merged 1 commit into from
Oct 20, 2023

Conversation

cbalioglu
Copy link
Contributor

This PR (1) improves the implementation of both CausalAttentionMask and ALiBiMask (2) accepts self_attn_mask_factory in MultiheadAttention and StandardTransformerEncoder for Longformer-based architectures.

@facebook-github-bot facebook-github-bot added the CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. label Oct 20, 2023
@cbalioglu cbalioglu merged commit 62c4430 into main Oct 20, 2023
18 checks passed
@cbalioglu cbalioglu deleted the mask branch October 20, 2023 16:36
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed.
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants