From 2b9a9bff2b0d70bc8c88cd3c756305129da4cbd5 Mon Sep 17 00:00:00 2001 From: Georgi Gerganov Date: Sat, 24 Feb 2024 10:41:21 +0200 Subject: [PATCH] minor : fix MPI builds --- llama.cpp | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/llama.cpp b/llama.cpp index a69c86e6a..4b257c472 100644 --- a/llama.cpp +++ b/llama.cpp @@ -7733,7 +7733,7 @@ static void llama_graph_compute( ggml_cgraph * gf, int n_threads) { #ifdef GGML_USE_MPI - const int64_t n_layer = lctx.hparams.n_layer; + const int64_t n_layer = lctx.model.hparams.n_layer; ggml_mpi_graph_compute_pre(lctx.ctx_mpi, gf, n_layer); #endif