Skip to content

Implements dual-chunk-flash-attn backend for dual chunk attention with sparse attention support #3092

Implements dual-chunk-flash-attn backend for dual chunk attention with sparse attention support

Implements dual-chunk-flash-attn backend for dual chunk attention with sparse attention support #3092

Triggered via pull request January 27, 2025 05:59
Status Success
Total duration 13s
Artifacts

cleanup_pr_body.yml

on: pull_request_target
update-description
5s
update-description
Fit to window
Zoom out
Zoom in

Annotations

1 warning
update-description
ubuntu-latest pipelines will use ubuntu-24.04 soon. For more details, see https://github.com/actions/runner-images/issues/10636