Skip to content

Pull requests: Dao-AILab/flash-attention

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Reviews
Assignee
Filter by who’s assigned
Assigned to nobody Loading
Sort

Pull requests list

Windows compile fix for MSVC
#1716 opened Jun 14, 2025 by loscrossos Loading…
Theoretically make compiling from pip quicker
#1703 opened Jun 8, 2025 by whrit Loading…
fix: fa3 backward check qkv with qkv_scale and dqkv
#1686 opened May 29, 2025 by yuyu5333 Loading…
Fix/deterministic dk dv
#1678 opened May 26, 2025 by yuWeiCute Loading…
Fix a bug in flash_attn_triton.py
#1668 opened May 15, 2025 by AminDarabi Loading…
Fix typos in multiple files
#1655 opened May 8, 2025 by co63oc Loading…
Add compile support for flash_attn_with_kvcache
#1554 opened Mar 25, 2025 by ani300 Loading…
Support cuda 12.8.1 and SBSA wheels
#1507 opened Feb 25, 2025 by johnnynunez Loading…
Remove ninja runtime dependency
#1484 opened Feb 9, 2025 by kevmo314 Loading…
Add missing tests/__init__.py
#1405 opened Dec 20, 2024 by BioGeek Loading…
Fix deprecation warnings
#1382 opened Dec 12, 2024 by rongou Loading…
flashattnvarlen support tree attention
#1188 opened Aug 30, 2024 by efsotr Loading…
ProTip! Find all pull requests that aren't related to any open issues with -linked:issue.