Skip to content

Remove graph breaks for torch.compile() in flash_attention_forward when Lllama Model is padding free tuned #7878

Remove graph breaks for torch.compile() in flash_attention_forward when Lllama Model is padding free tuned

Remove graph breaks for torch.compile() in flash_attention_forward when Lllama Model is padding free tuned #7878

Re-run triggered October 18, 2024 11:54
Status Skipped
Total duration 5s
Artifacts

self-pr-slow-ci.yml

on: pull_request
Find models to run slow tests
0s
Find models to run slow tests
Matrix: Run all tests for the model
Fit to window
Zoom out
Zoom in