Skip to content

[FlashAttention] Sync from upstream tensor desc implementation (part 3) #4520

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Draft
wants to merge 1 commit into
base: main
Choose a base branch
from

Conversation

whitneywhtsang
Copy link
Contributor

No description provided.

@whitneywhtsang whitneywhtsang self-assigned this Jun 17, 2025
david-hls pushed a commit to david-hls/intel-xpu-backend-for-triton that referenced this pull request Jun 18, 2025
@whitneywhtsang whitneywhtsang force-pushed the whitneywhtsang/flash_attention_3 branch from 9235ab5 to 565774d Compare June 18, 2025 04:05
@whitneywhtsang whitneywhtsang force-pushed the whitneywhtsang/flash_attention_3 branch from 565774d to 8f21d09 Compare June 18, 2025 16:22
@whitneywhtsang
Copy link
Contributor Author

Changes in this PR cause performance degradation, will remeasure when we have the new scheduling.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

1 participant