cuda: add flash attention + test
This commit is contained in:
parent
4f4bf35f46
commit
f7bcfb0566
3 changed files with 526 additions and 1 deletions
|
@ -52,6 +52,8 @@ llama_build_and_test_executable(test-backend-ops.cpp)
|
|||
|
||||
llama_build_and_test_executable(test-rope.cpp)
|
||||
|
||||
llama_build_executable(test-flash-attention.cpp)
|
||||
|
||||
# dummy executable - not installed
|
||||
get_filename_component(TEST_TARGET test-c.c NAME_WE)
|
||||
add_executable(${TEST_TARGET} test-c.c)
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue