Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

add Flash Attention Support for Helsinki-NLP/opus models #36169

Open
AghaDurrani opened this issue Feb 13, 2025 · 2 comments
Open

add Flash Attention Support for Helsinki-NLP/opus models #36169

AghaDurrani opened this issue Feb 13, 2025 · 2 comments
Labels
Feature request Request for a new feature Good Second Issue Issues that are more difficult to do than "Good First" issues - give it a try if you want!

Comments

@AghaDurrani
Copy link

AghaDurrani commented Feb 13, 2025

Feature request

I would like to propose adding support for flash attention to the Helsinki-NLP/opus models in the Hugging Face library. Judging by what I am seeing here, it seems not implemented.

Motivation

performance gains: adding flash attention would allow for significant speedups

Your contribution

I am happy to help test the implementation if needed. Thank you for considering this enhancement, and I look forward to the discussion!

@AghaDurrani AghaDurrani added the Feature request Request for a new feature label Feb 13, 2025
@IliasAarab
Copy link

Second this!

@Rocketknight1
Copy link
Member

Sounds like a cool idea! If anyone opens a PR for this, please ping me or @ArthurZucker for review.

@ArthurZucker ArthurZucker added the Good Second Issue Issues that are more difficult to do than "Good First" issues - give it a try if you want! label Feb 19, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Feature request Request for a new feature Good Second Issue Issues that are more difficult to do than "Good First" issues - give it a try if you want!
Projects
None yet
Development

No branches or pull requests

4 participants