Remove graph breaks for torch.compile() in flash_attention_forward when Lllama Model is padding free tuned #7878
self-pr-slow-ci.yml
on: pull_request
Find models to run slow tests
0s
Matrix: Run all tests for the model