Skip to content

Conversation

@keshavvinayak01
Copy link
Contributor

@keshavvinayak01 keshavvinayak01 commented Nov 26, 2025

Following the discussion from #22441.

I ran the entire flex_attention_hop implementation with randomised input tensors, (Also see llvm/torch-mlir#4366) through aot.export and compared against eager mode, and I noticed no accuracy losses (On CPU)

…ionOp -> LinalgExt::AttentionOp

Signed-off-by: Keshav Vinayak Jha <[email protected]>
Signed-off-by: Keshav Vinayak Jha <[email protected]>
Signed-off-by: Keshav Vinayak Jha <[email protected]>
Signed-off-by: Keshav Vinayak Jha <[email protected]>
Signed-off-by: Keshav Vinayak Jha <[email protected]>
@keshavvinayak01 keshavvinayak01 marked this pull request as ready for review November 27, 2025 11:32
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant