fix: missing CLBLAS documentation
This commit is contained in:
parent
41654efea8
commit
28086b52de
1 changed files with 16 additions and 0 deletions
16
README.md
16
README.md
|
@ -299,6 +299,22 @@ Building the program with BLAS support may lead to some performance improvements
|
|||
cmake --build . --config Release
|
||||
```
|
||||
|
||||
- clBLAS
|
||||
|
||||
This provides BLAS acceleration using the CUDA cores of your GPU. Make sure to have the cblas installed.
|
||||
- Using `make`:
|
||||
```bash
|
||||
make LLAMA_CLBLAS=1
|
||||
```
|
||||
- Using `CMake`:
|
||||
|
||||
```bash
|
||||
mkdir build
|
||||
cd build
|
||||
cmake .. -DLLAMA_CLBLAS=ON
|
||||
cmake --build . --config Release
|
||||
```
|
||||
|
||||
Note: Because llama.cpp uses multiple CUDA streams for matrix multiplication results [are not guaranteed to be reproducible](https://docs.nvidia.com/cuda/cublas/index.html#results-reproducibility). If you need reproducibility, set `GGML_CUDA_MAX_STREAMS` in the file `ggml-cuda.cu` to 1.
|
||||
|
||||
### Prepare Data & Run
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue