Skip to content

Pull requests: Dao-AILab/flash-attention

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Reviews
Assignee
Filter by who’s assigned
Assigned to nobody Loading
Sort

Pull requests list

[WIP] support the flash api for Ascend
#2246 opened Feb 10, 2026 by AnyFree813 Loading…
Add gfx1150/gfx1151 (RDNA 3.5) to RDNA_ARCHS
#2243 opened Feb 9, 2026 by rwfsmith Loading…
[AMD] Migrate to Triton Backend to Aiter
#2230 opened Feb 4, 2026 by micmelesse Loading…
Nicer headdim error message
#2227 opened Feb 4, 2026 by drisspg Loading…
[WIP] varlen blocksparsity
#2224 opened Feb 2, 2026 by reubenconducts Draft
[Ai-assisted] CLC work stealing
#2218 opened Jan 31, 2026 by drisspg Loading…
Add loc info & Fix api changes for CuTeDSL 4.4
#2204 opened Jan 23, 2026 by keithzzzzz Loading…
BWD sm100 2cta
#2202 opened Jan 23, 2026 by tzadouri Loading…
[Cute, SM100] Fix comment in tmem_p_offset
#2201 opened Jan 22, 2026 by Edenzzzz Loading…
Warn when ninja is missing
#2191 opened Jan 17, 2026 by blueberrycongee Loading…
[Cute][Testing] Protyping a fast test mode for Cute
#2188 opened Jan 16, 2026 by drisspg Loading…
[Cute] Add torch.compile support for FA4
#2164 opened Jan 9, 2026 by gilfordting Loading…
[Cute,Fwd,Sm100] fp8 e4m3 and e5m2 support
#2109 opened Dec 29, 2025 by dcw02 Loading…
refactor llama test
#2107 opened Dec 29, 2025 by m3ngyang Loading…
ProTip! Exclude everything labeled bug with -label:bug.