Skip to content

Pull requests: Dao-AILab/flash-attention

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Reviews
Assignee
Filter by who’s assigned
Assigned to nobody Loading
Sort

Pull requests list

[Cute] Fix: arg pass in cute flash-attn inferface
#2101 opened Dec 27, 2025 by SeanLi-OI Loading…
Fix softmax incorrect row_max issue
#2083 opened Dec 17, 2025 by imbr92 Loading…
Fix TypeError when ColumnParallelLinear is None
#2080 opened Dec 17, 2025 by ailuntz Loading…
Reduce Chance of Build OOM
#2079 opened Dec 17, 2025 by Qubitium Loading…
Add missing code highlighting to the README
#2061 opened Dec 10, 2025 by bryant1410 Loading…
Update README.md
#2058 opened Dec 10, 2025 by eduardoruiz1999 Loading…
[AMD ROCm] Enable CK backend for ROCm gfx12
#2054 opened Dec 8, 2025 by hyoon1 Loading…
Ko3n1g/ci/torch2.9 for cuda129
#2044 opened Dec 3, 2025 by ko3n1g Draft
Disable abi3 for free-threaded python
#2034 opened Nov 25, 2025 by kevmo314 Loading…
[Cute,Fwd,Sm100] Support q_stage=1 for inference
#1993 opened Nov 7, 2025 by timmy-feng Loading…
[Cute,Fwd,Sm90] Support KV cache
#1992 opened Nov 6, 2025 by imbr92 Loading…
Make wheel name and version consistent
#1956 opened Oct 22, 2025 by bobingm Loading…
fix: nan when m_i_new=-inf in online softmax
#1948 opened Oct 20, 2025 by tongyx361 Loading…
support cpu run fa triton kernel
#1938 opened Oct 15, 2025 by hellozmz Draft
windows error error C2039:
#1932 opened Oct 12, 2025 by Granddyser Loading…
feat: add to support float8 kvcache in fa4
#1914 opened Sep 28, 2025 by yicwang Loading…
fix forward and backward kernel
#1907 opened Sep 24, 2025 by rz2778 Loading…
Feature/varlen rotray
#1899 opened Sep 19, 2025 by mhoangvslev Loading…
ProTip! Exclude everything labeled bug with -label:bug.