Skip to content

Pull requests: Dao-AILab/flash-attention

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Reviews
Assignee
Filter by who’s assigned
Assigned to nobody Loading
Sort

Pull requests list

feat: add to support float8 kvcache in fa4
#1914 opened Sep 28, 2025 by yicwang Loading…
fix forward and backward kernel
#1907 opened Sep 24, 2025 by rz2778 Loading…
Feature/varlen rotray
#1899 opened Sep 19, 2025 by mhoangvslev Loading…
Improve setup.py
#1859 opened Sep 3, 2025 by cyyever Loading…
Refactors to enable FlexAttention
#1840 opened Aug 26, 2025 by drisspg Loading…
feat: Implement Sink Attention
#1819 opened Aug 18, 2025 by aoxy Loading…
feat: blocksparse support
#1784 opened Jul 30, 2025 by guangyunh-nv Draft
[CI] build upon manylinux, improve compatibility
#1780 opened Jul 29, 2025 by zipzou Loading…
Change the update method of the sub-module
#1774 opened Jul 25, 2025 by RealTapeL Loading…
add var_len case for benchmark_mla_decode
#1770 opened Jul 22, 2025 by XiaobingSuper Loading…
[AMD] Torch Compile Issues
#1756 opened Jul 15, 2025 by micmelesse Loading…
Suppress warnings in windows compilation
#1748 opened Jul 10, 2025 by XXXXRT666 Loading…
Theoretically make compiling from pip quicker
#1703 opened Jun 8, 2025 by whrit Loading…
fix: fa3 backward check qkv with qkv_scale and dqkv
#1686 opened May 29, 2025 by yuyu5333 Loading…
Fix/deterministic dk dv
#1678 opened May 26, 2025 by yuWeiCute Loading…
Fix a bug in flash_attn_triton.py
#1668 opened May 15, 2025 by AminDarabi Loading…
ProTip! Find all pull requests that aren't related to any open issues with -linked:issue.