llama.cpp/ggml/src
Andreas (Andi) Kunar bf5a81df37
ggml : fix build on Windows with Snapdragon X (#8531)
* Improvements for Windows with Snapdragon X

* Revert "Improvements for Windows with Snapdragon X"

This reverts commit bf21397ae5.

* Improvements for Windows with Snapdragon X

* WOA build clarifications

* WIndows on ARM build clarifications

* cmake build for Windows clarifications

* Update docs/build.md

Co-authored-by: Georgi Gerganov <ggerganov@gmail.com>

---------

Co-authored-by: AndreasKunar <andreaskmsn.com>
Co-authored-by: Georgi Gerganov <ggerganov@gmail.com>
2024-07-25 19:01:00 +03:00
..
ggml-cann [CANN] Add Ascend NPU backend (#6035) 2024-07-17 14:23:50 +03:00
ggml-cuda Allow all RDNA2 archs to use sdot4 intrinsic (#8629) 2024-07-23 10:50:40 +02:00
ggml-sycl [SYCL] fix multi-gpu issue on sycl (#8554) 2024-07-25 19:45:18 +08:00
kompute@4565194ed7 llama : reorganize source code + improve CMake (#8006) 2024-06-26 18:33:02 +03:00
kompute-shaders llama : reorganize source code + improve CMake (#8006) 2024-06-26 18:33:02 +03:00
llamafile ggml : move sgemm sources to llamafile subfolder (#8394) 2024-07-10 15:23:29 +03:00
vulkan-shaders Vulkan IQ4_NL Support (#8613) 2024-07-23 10:56:49 +02:00
CMakeLists.txt Re-add erroneously removed -fsycl from GGML_EXTRA_LIBS (#8667) 2024-07-24 11:55:26 +01:00
ggml-aarch64.c ggml : fix build on Windows with Snapdragon X (#8531) 2024-07-25 19:01:00 +03:00
ggml-aarch64.h ggml : minor naming changes (#8433) 2024-07-12 10:46:02 +03:00
ggml-alloc.c CUDA: fix partial offloading for ne0 % 256 != 0 (#8572) 2024-07-18 23:48:47 +02:00
ggml-backend-impl.h llama : reorganize source code + improve CMake (#8006) 2024-06-26 18:33:02 +03:00
ggml-backend.c CUDA: fix partial offloading for ne0 % 256 != 0 (#8572) 2024-07-18 23:48:47 +02:00
ggml-blas.cpp ggml : add NVPL BLAS support (#8329) (#8425) 2024-07-11 18:49:15 +02:00
ggml-cann.cpp [CANN] Add Ascend NPU backend (#6035) 2024-07-17 14:23:50 +03:00
ggml-common.h ggml : add AArch64 optimized GEMV and GEMM Q4 kernels (#5780) 2024-07-10 15:14:51 +03:00
ggml-cuda.cu CUDA: fix partial offloading for ne0 % 256 != 0 (#8572) 2024-07-18 23:48:47 +02:00
ggml-impl.h ggml : add AArch64 optimized GEMV and GEMM Q4 kernels (#5780) 2024-07-10 15:14:51 +03:00
ggml-kompute.cpp llama : reorganize source code + improve CMake (#8006) 2024-06-26 18:33:02 +03:00
ggml-metal.m ggml : fix quant dot product with odd number of blocks (#8549) 2024-07-19 17:17:27 +02:00
ggml-metal.metal ggml : fix quant dot product with odd number of blocks (#8549) 2024-07-19 17:17:27 +02:00
ggml-quants.c ggml: fix compile error for RISC-V (#8623) 2024-07-22 10:56:45 +03:00
ggml-quants.h ggml : minor naming changes (#8433) 2024-07-12 10:46:02 +03:00
ggml-rpc.cpp llama : reorganize source code + improve CMake (#8006) 2024-06-26 18:33:02 +03:00
ggml-sycl.cpp [SYCL] add concat through dim 1/2 (#8483) 2024-07-15 19:32:15 +08:00
ggml-vulkan.cpp Vulkan IQ4_NL Support (#8613) 2024-07-23 10:56:49 +02:00
ggml.c ggml : add and use ggml_cpu_has_llamafile() (#8664) 2024-07-25 12:37:42 +03:00