-
Notifications
You must be signed in to change notification settings - Fork 8
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
generality to FLUX model #8
Comments
Thanks. Recently, we are exploring how to improve the method of DiTFastAttn. And I believe it will achieve a significant acceleration on FLUX and SD3. It will release with our next paper. |
Hello @hahnyuan @youngwanLEE I am Jiarui Fang from the xDiT project (https://github.com/xdit-project/xDiT), and I took notice of the DiTFastAttn work for the first time its released. I found that its motivation, redundancy in diffusion models, is similar to that of PipeFusion. While PipeFusion leverages this redundancy to address parallel performance issues, DiTFastAttn use it to reduce computation on a single GPU. We have also implemented DiTFastAttn in xDiT and cited your work. We are also exploring the use cases of DiTFastAttn in Flux and SD3. I am not sure if there is an opportunity for collaboration. |
Hello @feifeibear |
Hi, I'm impressed by your work.
I wonder whether the proposed method can be applied to FLUX, the more recent SoTA T2I model, which has more complex transformer blocks.
The text was updated successfully, but these errors were encountered: