Skip to content

Attention refactor bug with MHA #4149

@shivadbhavsar

Description

@shivadbhavsar

Some models using multi head attention from onnx are no longer fusing attention properly after the attention refactor was merged in.

Metadata

Metadata

Assignees

Labels

bugSomething isn't working

Type

No type

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions