cuda: add flash attention + test

This commit is contained in:
FSSRepo 2024-01-17 16:38:28 -05:00
parent 4f4bf35f46
commit f7bcfb0566
3 changed files with 526 additions and 1 deletions

View file

@ -52,6 +52,8 @@ llama_build_and_test_executable(test-backend-ops.cpp)
llama_build_and_test_executable(test-rope.cpp)
llama_build_executable(test-flash-attention.cpp)
# dummy executable - not installed
get_filename_component(TEST_TARGET test-c.c NAME_WE)
add_executable(${TEST_TARGET} test-c.c)