From 50cfa1fc36e0b1f43255cf744dd89be447d63dec Mon Sep 17 00:00:00 2001 From: Georgi Gerganov Date: Wed, 10 Jan 2024 16:09:20 +0200 Subject: [PATCH] minor : cleanup trailing whitespaces --- llama.cpp | 8 ++++---- 1 file changed, 4 insertions(+), 4 deletions(-) diff --git a/llama.cpp b/llama.cpp index e211d5d45..f2027f552 100644 --- a/llama.cpp +++ b/llama.cpp @@ -3145,13 +3145,13 @@ static void llm_load_print_meta(llama_model_loader & ml, llama_model & model) { LLAMA_LOG_INFO("%s: rope_finetuned = %s\n", __func__, hparams.rope_finetuned ? "yes" : "unknown"); LLAMA_LOG_INFO("%s: model type = %s\n", __func__, llama_model_type_name(model.type)); LLAMA_LOG_INFO("%s: model ftype = %s\n", __func__, llama_model_ftype_name(model.ftype).c_str()); - if (ml.n_elements >= 1e12) { + if (ml.n_elements >= 1e12) { LLAMA_LOG_INFO("%s: model params = %.2f T\n", __func__, ml.n_elements*1e-12); - } else if (ml.n_elements >= 1e9) { + } else if (ml.n_elements >= 1e9) { LLAMA_LOG_INFO("%s: model params = %.2f B\n", __func__, ml.n_elements*1e-9); - } else if (ml.n_elements >= 1e6) { + } else if (ml.n_elements >= 1e6) { LLAMA_LOG_INFO("%s: model params = %.2f M\n", __func__, ml.n_elements*1e-6); - } else { + } else { LLAMA_LOG_INFO("%s: model params = %.2f K\n", __func__, ml.n_elements*1e-3); } if (ml.n_bytes < GiB) {