fix: missing CLBLAS documentation

This commit is contained in:
Carsten Seeger 2023-05-09 10:14:22 +02:00
parent 41654efea8
commit 28086b52de

View file

@ -299,6 +299,22 @@ Building the program with BLAS support may lead to some performance improvements
cmake --build . --config Release
```
- clBLAS
This provides BLAS acceleration using the CUDA cores of your GPU. Make sure to have the cblas installed.
- Using `make`:
```bash
make LLAMA_CLBLAS=1
```
- Using `CMake`:
```bash
mkdir build
cd build
cmake .. -DLLAMA_CLBLAS=ON
cmake --build . --config Release
```
Note: Because llama.cpp uses multiple CUDA streams for matrix multiplication results [are not guaranteed to be reproducible](https://docs.nvidia.com/cuda/cublas/index.html#results-reproducibility). If you need reproducibility, set `GGML_CUDA_MAX_STREAMS` in the file `ggml-cuda.cu` to 1.
### Prepare Data & Run