Skip to content

Pull requests: Dao-AILab/flash-attention

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Reviews
Assignee
Filter by who’s assigned
Assigned to nobody Loading
Sort

Pull requests list

support cpu run fa triton kernel
#1938 opened Oct 15, 2025 by hellozmz Draft
[CUTE] Enable Pack GQA for score mods
#1937 opened Oct 14, 2025 by drisspg Loading…
Fix backward preprocess kernel in FlashAttention3
#1936 opened Oct 14, 2025 by rfbr Loading…
windows error error C2039:
#1932 opened Oct 12, 2025 by Granddyser Loading…
[NFC] Trivial fix to silence linter
#1928 opened Oct 8, 2025 by jduprat Loading…
feat: add to support float8 kvcache in fa4
#1914 opened Sep 28, 2025 by yicwang Loading…
fix forward and backward kernel
#1907 opened Sep 24, 2025 by rz2778 Loading…
Feature/varlen rotray
#1899 opened Sep 19, 2025 by mhoangvslev Loading…
Improve setup.py
#1859 opened Sep 3, 2025 by cyyever Loading…
feat: Implement Sink Attention
#1819 opened Aug 18, 2025 by aoxy Loading…
feat: blocksparse support
#1784 opened Jul 30, 2025 by guangyunh-nv Draft
[CI] build upon manylinux, improve compatibility
#1780 opened Jul 29, 2025 by zipzou Loading…
Change the update method of the sub-module
#1774 opened Jul 25, 2025 by RealTapeL Loading…
add var_len case for benchmark_mla_decode
#1770 opened Jul 22, 2025 by XiaobingSuper Loading…
Suppress warnings in windows compilation
#1748 opened Jul 10, 2025 by XXXXRT666 Loading…
Theoretically make compiling from pip quicker
#1703 opened Jun 8, 2025 by whrit Loading…
fix: fa3 backward check qkv with qkv_scale and dqkv
#1686 opened May 29, 2025 by yuyu5333 Loading…
Fix/deterministic dk dv
#1678 opened May 26, 2025 by yuWeiCute Loading…
ProTip! no:milestone will show everything without a milestone.