Skip to content

Implements dual-chunk-flash-attn backend for dual chunk attention with sparse attention support #3092

Implements dual-chunk-flash-attn backend for dual chunk attention with sparse attention support

Implements dual-chunk-flash-attn backend for dual chunk attention with sparse attention support #3092

Annotations

1 warning

update-description

succeeded Jan 27, 2025 in 5s