Skip to content

Commit

Permalink
[Bugfix][Kernel] FA3 Fix - RuntimeError: This flash attention build o…
Browse files Browse the repository at this point in the history
…nly supports pack_gqa (for build size reasons). (#12405)

Signed-off-by: Lucas Wilkinson <[email protected]>
  • Loading branch information
LucasWilkinson authored Jan 24, 2025
1 parent df5dafa commit 3132a93
Showing 1 changed file with 1 addition and 1 deletion.
2 changes: 1 addition & 1 deletion CMakeLists.txt
Original file line number Diff line number Diff line change
Expand Up @@ -576,7 +576,7 @@ else()
FetchContent_Declare(
vllm-flash-attn
GIT_REPOSITORY https://github.com/vllm-project/flash-attention.git
GIT_TAG 0aff05f577e8a10086066a00618609199b25231d
GIT_TAG 9732b0ce005d1e6216864788502d5570004678f5
GIT_PROGRESS TRUE
# Don't share the vllm-flash-attn build between build types
BINARY_DIR ${CMAKE_BINARY_DIR}/vllm-flash-attn
Expand Down

0 comments on commit 3132a93

Please sign in to comment.