From 055e01b1390b2728527cc28e1f1f2089062df155 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?Johannes=20G=C3=A4=C3=9Fler?= Date: Mon, 27 Jan 2025 18:33:34 +0100 Subject: [PATCH] try CI fix --- include/llama.h | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/include/llama.h b/include/llama.h index 69006594d..18ced30d7 100644 --- a/include/llama.h +++ b/include/llama.h @@ -1328,7 +1328,7 @@ extern "C" { typedef bool (*llama_opt_param_filter)(const struct ggml_tensor * tensor, void * userdata); // always returns true - bool llama_opt_param_filter_all(const struct ggml_tensor * tensor, void * userdata); + LLAMA_API bool llama_opt_param_filter_all(const struct ggml_tensor * tensor, void * userdata); struct llama_opt_params { uint32_t n_ctx_train; // assumed context size post training, use context size specified in llama_context if 0