Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[WIP] Add flex attention for gpt2 #34861

Draft
wants to merge 1 commit into
base: main
Choose a base branch
from

Conversation

mayankagarwals
Copy link
Contributor

What does this PR do?

Adding flex_attention for Gpt2 model following #34809

Who can review?

@Rocketknight1
Copy link
Member

cc @ArthurZucker for review, but in the meantime @mayankagarwals you can make the tests pass by doing pip install transformers[quality] and then make fixup in the repo directory

@mayankagarwals
Copy link
Contributor Author

Hi @Rocketknight1

Thanks a ton, got it.

Waiting for discussions to close here #34896. Thinking that PR should follow similar guidelines to keep one standard design of implementation of generic attention block.

@vasqu Let me know once major design bits are closed! I'm following your PR for gpt neox and aligned with everything there.

@vasqu
Copy link
Contributor

vasqu commented Nov 24, 2024

@mayankagarwals Sure I'll let you know when stuff clears up!

@ArthurZucker
Copy link
Collaborator

Having a look at your PR @vasqu, I think quite a few things were fixed since so will check!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

4 participants