Skip to content

Conversation

@ClarkChin08
Copy link

@ClarkChin08 ClarkChin08 commented Nov 6, 2025

No description provided.

@ClarkChin08 ClarkChin08 changed the title [Draft] Add VarLen support to new flash attention api Add VarLen support to new flash attention api Nov 7, 2025
Copy link

@petercad petercad left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

varlen-related changes look OK (reviewed causal mask changes in a separate PR). Just to check, we have some accuracy issues with the old FlashAttention + varlen + causal code when the driver stopped zero-initializing memory. Should double-check that it's not happening here.

@tdeng5 tdeng5 requested a review from taozha2 November 9, 2025 23:42
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

5 participants