Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

generality to FLUX model #8

Open
youngwanLEE opened this issue Sep 19, 2024 · 3 comments
Open

generality to FLUX model #8

youngwanLEE opened this issue Sep 19, 2024 · 3 comments

Comments

@youngwanLEE
Copy link

Hi, I'm impressed by your work.

I wonder whether the proposed method can be applied to FLUX, the more recent SoTA T2I model, which has more complex transformer blocks.

@hahnyuan
Copy link
Collaborator

Thanks.
Regarding your question about applying the proposed method to FLUX, it sounds like a challenging yet exciting opportunity. The differences in structure with MMDiT could certainly introduce complexities, but it might also offer new avenues for optimization.

Recently, we are exploring how to improve the method of DiTFastAttn. And I believe it will achieve a significant acceleration on FLUX and SD3. It will release with our next paper.

@feifeibear
Copy link

feifeibear commented Oct 11, 2024

Hello @hahnyuan @youngwanLEE

I am Jiarui Fang from the xDiT project (https://github.com/xdit-project/xDiT), and I took notice of the DiTFastAttn work for the first time its released. I found that its motivation, redundancy in diffusion models, is similar to that of PipeFusion. While PipeFusion leverages this redundancy to address parallel performance issues, DiTFastAttn use it to reduce computation on a single GPU.

We have also implemented DiTFastAttn in xDiT and cited your work. We are also exploring the use cases of DiTFastAttn in Flux and SD3. I am not sure if there is an opportunity for collaboration.

@hahnyuan
Copy link
Collaborator

Hello @feifeibear
I've recently been looking into what FLUX and SD3 can do, and I see some potential for enhancements in the CrossAttention between text and images in MMDiT. I'm looking forward to the opportunity to collaborate with you on this!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants