diff --git a/examples/baby-llama/baby-llama-text.cpp b/examples/baby-llama/baby-llama-text.cpp index c5abf66a4..9f2ff9034 100644 --- a/examples/baby-llama/baby-llama-text.cpp +++ b/examples/baby-llama/baby-llama-text.cpp @@ -1199,10 +1199,12 @@ int main(int argc, char ** argv) { struct llama_context * lctx = llama_init_from_file(fn_model, llama_params); + printf("%s: tokenize training data\n", __func__); std::vector train_tokens; if (tokenize_file(lctx, fn_train, train_tokens) < 0) { fprintf(stderr, "%s: failed to tokenize file '%s'\n", __func__, fn_train); } + printf("%s: number of training tokens: %d\n", __func__, train_tokens.size()); struct my_llama_model model; model.hparams.n_vocab = llama_n_vocab(lctx); @@ -1225,7 +1227,7 @@ int main(int argc, char ** argv) { model.ctx = ggml_init(lcparams); kv_self.ctx = model.ctx; - printf("init model\n"); + printf("%s: init model\n", __func__); init_model(&model); set_param_model(&model); randomize_model(&model, 1337, 0.0f, 1.0f, -1.0f, +1.0f); @@ -1238,6 +1240,8 @@ int main(int argc, char ** argv) { int n_tokens = model.hparams.n_ctx; int n_vocab = model.hparams.n_vocab; + printf("%s: begin training\n", __func__); + for (int ex=0; ex