From 39a2c89a305453bd3623db70e8693da21ee83f85 Mon Sep 17 00:00:00 2001 From: akawrykow Date: Tue, 22 Aug 2023 17:25:04 -0700 Subject: [PATCH] [gguf] Print the date --- llama.cpp | 3 +++ 1 file changed, 3 insertions(+) diff --git a/llama.cpp b/llama.cpp index 6c5da1309..d2a4676ec 100644 --- a/llama.cpp +++ b/llama.cpp @@ -1279,6 +1279,7 @@ static void llama_model_load_internal( std::string general_name = "n/a"; std::string general_arch = "n/a"; + std::string general_date = "n/a"; // read hparams { @@ -1336,6 +1337,7 @@ static void llama_model_load_internal( // get general kv GGUF_GET(general_name, gguf_get_val_str, GGUF_TYPE_STRING, false, "general.name"); GGUF_GET(general_arch, gguf_get_val_str, GGUF_TYPE_STRING, false, "general.architecture"); + GGUF_GET(general_date, gguf_get_val_str, GGUF_TYPE_STRING, false, "general.date"); // special tokens GGUF_GET(vocab.special_bos_id, gguf_get_val_u32, GGUF_TYPE_UINT32, false, "tokenizer.ggml.bos_token_id"); @@ -1445,6 +1447,7 @@ static void llama_model_load_internal( // general kv LLAMA_LOG_INFO("%s: general.name = %s\n", __func__, general_name.c_str()); + LLAMA_LOG_INFO("%s: general.date = %s\n", __func__, general_date.c_str()); // special tokens if (vocab.special_bos_id != -1) { LLAMA_LOG_INFO( "%s: BOS token = %d '%s'\n", __func__, vocab.special_bos_id, vocab.id_to_token[vocab.special_bos_id].text.c_str() ); }