-
Notifications
You must be signed in to change notification settings - Fork 1.4k
Issues: Dao-AILab/flash-attention
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
IncompatibleTypeErrorImpl('invalid operands of type pointer<int64> and triton.language.int32')
#1439
opened Jan 11, 2025 by
wuyouliaoxi
ERROR: No matching distribution found for flash-attn==2.6.3+cu123torch2.4cxx11abifalse
#1423
opened Jan 6, 2025 by
carolynsoo
Unable to install
flash-attn
even if I first install torch
alone
#1421
opened Jan 3, 2025 by
ytxmobile98
Is there a plan to support flash_attn_varlen_backward with fp8
#1420
opened Jan 3, 2025 by
gaodaheng
flash_attn_with_kvcache
discrepancy slicing kv_cache / cache_seqlens
#1417
opened Jan 1, 2025 by
jeromeku
looking for a test to verify cache correctness in
flash_attn_with_kvcache
#1414
opened Dec 26, 2024 by
chakpongchung
Performance Impact of Using Three Warps per Group (WG) in FA3 Compared to Two WGs
#1413
opened Dec 24, 2024 by
ziyuhuang123
UnboundLocalError: local variable 'out' referenced before assignment
#1412
opened Dec 24, 2024 by
chuangzhidan
Why Does FA3 Use Registers Instead of Directly Accessing SMEM with WGMMA on SM90?
#1407
opened Dec 23, 2024 by
ziyuhuang123
Understanding the Role of arrive in NamedBarrier Synchronization
#1400
opened Dec 19, 2024 by
ziyuhuang123
Previous Next
ProTip!
Mix and match filters to narrow down what you’re looking for.