-
Notifications
You must be signed in to change notification settings - Fork 1.8k
Pull requests: Dao-AILab/flash-attention
Author
Label
Projects
Milestones
Reviews
Assignee
Sort
Pull requests list
Useuful command to install flash faster on behamoth clusters
#1660
opened May 10, 2025 by
sleepingcat4
Loading…
Patch
RPATH
of compiled Linux library to locate PyTorch and CUDA libraries in virtual env
#1634
opened Apr 30, 2025 by
sisp
Loading…
feat: support to tile K and V separately in FA3 backward
#1626
opened Apr 28, 2025 by
beginlner
Loading…
add checks for zero elements input of triton LayerNorm impl
#1621
opened Apr 27, 2025 by
Luciennnnnnn
Loading…
Add PT compileable support for flash_attn_with_kvcache
#1592
opened Apr 14, 2025 by
jataylo
Loading…
feat: fa3 custom ops for compatibility with PT Compile
#1590
opened Apr 13, 2025 by
zhangheng408
Loading…
[BugFix] Fix a wrong reference to seqlen_k variable in the fwd_splitkv kernel
#1455
opened Jan 21, 2025 by
muoshuosha
Loading…
wrap func into torch ops to avoid torch.compile graphbreaks
#1333
opened Nov 13, 2024 by
kumarkrishna
Loading…
Promote wheels as alternative to pip install flash-attn
#1297
opened Oct 25, 2024 by
simonw
Loading…
fix: in newer versions of triton, tl.dot should take as input only q …
#1288
opened Oct 21, 2024 by
EdouardYvinec
Loading…
the test_flash_attn.py it's actually in parent directory
#1167
opened Aug 21, 2024 by
ArtificialZeng
Loading…
Previous Next
ProTip!
Find all pull requests that aren't related to any open issues with -linked:issue.