From 37bcad7d6dbfe2a8f94591a818f693c005a0db95 Mon Sep 17 00:00:00 2001 From: bandoti <141645996+bandoti@users.noreply.github.com> Date: Tue, 25 Jun 2024 16:27:24 -0300 Subject: [PATCH] Update README.md --- README.md | 15 +++++++++++++++ 1 file changed, 15 insertions(+) diff --git a/README.md b/README.md index a54ee3951..ea8e19863 100644 --- a/README.md +++ b/README.md @@ -578,6 +578,21 @@ Building the program with BLAS support may lead to some performance improvements | LLAMA_CUDA_KQUANTS_ITER | 1 or 2 | 2 | Number of values processed per iteration and per HIP thread for Q2_K and Q6_K quantization formats. Setting this value to 1 can improve performance for slow GPUs. | - #### Vulkan + ##### Windows (MSYS2) + Install [MSYS2](https://www.msys2.org/) and then run the following commands in a UCRT terminal to install dependencies. + ```sh + pacman -S git \ + mingw-w64-ucrt-x86_64-gcc \ + mingw-w64-ucrt-x86_64-cmake \ + mingw-w64-ucrt-x86_64-vulkan-devel \ + mingw-w64-ucrt-x86_64-shaderc \ + mingw-w64-ucrt-x86_64-python3 + ``` + Switch into `llama.cpp` directory and build using CMake. + ```sh + cmake -B build -DLLAMA_VULKAN=ON + cmake --build build --config Release + ``` **With docker**: