Skip to content

[Codegen] Remove attention transpose intrinsic hacks#23633

Merged
Groverkss merged 2 commits intoiree-org:mainfrom
Groverkss:remove_attention_hacks
Mar 4, 2026
Merged

[Codegen] Remove attention transpose intrinsic hacks#23633
Groverkss merged 2 commits intoiree-org:mainfrom
Groverkss:remove_attention_hacks

Conversation

@Groverkss
Copy link
Contributor

After #23631 we will be using the analysis to determine conflicts, which allows us to undo a bunch of hacks we did earlier to not have conflicts between the chained mma in attention.

This patch simply sets the intrinsic to be col major in KernelConfig and rest of the pipeline automatically picks up if there needs to be a intermediate conflict using shared memory or not.

@Groverkss
Copy link
Contributor Author

Depends on #23631

@Groverkss Groverkss force-pushed the remove_attention_hacks branch from 1e1df58 to 42dfadf Compare March 4, 2026 14:32
Copy link
Contributor

@sommerlukas sommerlukas left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Just small stuff.

Do we still have a e2e or pipeline test that doesn't use col_major = true to ensure that path still works?

@Groverkss Groverkss enabled auto-merge (squash) March 4, 2026 15:52
@Groverkss Groverkss merged commit 1ea0b66 into iree-org:main Mar 4, 2026
80 of 83 checks passed
bangtianliu added a commit that referenced this pull request Mar 13, 2026
…thon binding (#23757)

PR #23633 removes the `attention_qk_matmul` and `attention_pv_matmul`
marker attributes from attention decomposition configs, replacing them
with a new `col_major = true` parameter on MMA intrinsics.

This PR updates the Python bindings to support the `col_major
`parameter, enabling the tuner to generate attention configs compatible
with the new approach.

After this PR, I will also add the required changes to the tuner side. 

Assisted-by:  [Claude Code](https://claude.ai/code)

Signed-off-by: Bangtian Liu <[email protected]>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants