From 458df7413c4f060c3b04027aab7d0aaeb971be10 Mon Sep 17 00:00:00 2001 From: Evan Jones Date: Tue, 2 May 2023 17:43:50 -0400 Subject: [PATCH] avoid designated initializers --- llama.cpp | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/llama.cpp b/llama.cpp index 52ebc6757..85af4dc49 100644 --- a/llama.cpp +++ b/llama.cpp @@ -2498,7 +2498,7 @@ size_t llama_copy_state_data(struct llama_context * ctx, uint8_t * dest) { if (kv_size) { const size_t elt_size = ggml_element_size(kv_self.k); char buffer[4096]; - ggml_context * cpy_ctx = ggml_init({ sizeof(buffer), buffer, .no_alloc = true }); + ggml_context * cpy_ctx = ggml_init({ sizeof(buffer), buffer, /* no_alloc */ true }); ggml_cgraph gf{}; gf.n_threads = 1; @@ -2602,7 +2602,7 @@ size_t llama_set_state_data(struct llama_context * ctx, const uint8_t * src) { const size_t elt_size = ggml_element_size(kv_self.k); char buffer[4096]; - ggml_context * cpy_ctx = ggml_init({ sizeof(buffer), buffer, .no_alloc = true }); + ggml_context * cpy_ctx = ggml_init({ sizeof(buffer), buffer, /* no_alloc */ true }); ggml_cgraph gf{}; gf.n_threads = 1;