llama.cpp/ggml/src/ggml-vulkan
Jeff Bolz 44e18ef939
vulkan: fix coopmat2 flash attention for non-contiguous inputs (#11281)
Add code similar to mul_mm_cm2 to force alignment of strides, to avoid
a performance regression.

Add noncontiguous FA tests in test-backend-ops.

Fixes #11268.
2025-01-18 09:26:50 +01:00
..
cmake fix: ggml: fix vulkan-shaders-gen build (#10448) 2025-01-15 14:17:42 +01:00
vulkan-shaders vulkan: fix coopmat2 flash attention for non-contiguous inputs (#11281) 2025-01-18 09:26:50 +01:00
CMakeLists.txt fix: ggml: fix vulkan-shaders-gen build (#10448) 2025-01-15 14:17:42 +01:00
ggml-vulkan.cpp vulkan: fix coopmat2 flash attention for non-contiguous inputs (#11281) 2025-01-18 09:26:50 +01:00