Add GGML_USE_BLAS flag to llama.cpp and update BLAS documentation
This commit is contained in:
parent
e11bd856d5
commit
4a4a3420de
2 changed files with 9 additions and 1 deletions
|
@ -125,6 +125,14 @@ This provides BLAS acceleration using only the CPU. Make sure to have OpenBLAS i
|
||||||
make GGML_OPENBLAS=1
|
make GGML_OPENBLAS=1
|
||||||
```
|
```
|
||||||
|
|
||||||
|
- On Mac (with Intel GPU):
|
||||||
|
```bash
|
||||||
|
brew install openblas
|
||||||
|
EXPORT PKG_CONFIG_PATH=<openblas_path>
|
||||||
|
make GGML_OPENBLAS=1 GGML_NO_METAL=1
|
||||||
|
```
|
||||||
|
|
||||||
|
|
||||||
- Using `CMake` on Linux:
|
- Using `CMake` on Linux:
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
|
|
|
@ -17424,7 +17424,7 @@ bool llama_supports_mlock(void) {
|
||||||
|
|
||||||
bool llama_supports_gpu_offload(void) {
|
bool llama_supports_gpu_offload(void) {
|
||||||
#if defined(GGML_USE_CUDA) || defined(GGML_USE_METAL) || defined(GGML_USE_VULKAN) || \
|
#if defined(GGML_USE_CUDA) || defined(GGML_USE_METAL) || defined(GGML_USE_VULKAN) || \
|
||||||
defined(GGML_USE_SYCL) || defined(GGML_USE_KOMPUTE) || defined(GGML_USE_RPC)
|
defined(GGML_USE_SYCL) || defined(GGML_USE_KOMPUTE) || defined(GGML_USE_RPC) || defined(GGML_USE_BLAS)
|
||||||
// Defined when llama.cpp is compiled with support for offloading model layers to GPU.
|
// Defined when llama.cpp is compiled with support for offloading model layers to GPU.
|
||||||
return true;
|
return true;
|
||||||
#else
|
#else
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue