From c2ad2b02f3d4124246df5b5400cb971a15c7b55c Mon Sep 17 00:00:00 2001 From: supermy Date: Fri, 24 Nov 2023 14:18:01 +0800 Subject: [PATCH] =?UTF-8?q?=E5=BE=AE=E8=B0=83=E8=84=9A=E6=9C=AC=E4=BC=98?= =?UTF-8?q?=E5=8C=96?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- examples/finetune/README.md | 10 +++++----- examples/finetune/finetune.sh | 14 +++++++------- 2 files changed, 12 insertions(+), 12 deletions(-) diff --git a/examples/finetune/README.md b/examples/finetune/README.md index 688061c8d..99111d150 100644 --- a/examples/finetune/README.md +++ b/examples/finetune/README.md @@ -28,38 +28,38 @@ wget https://raw.githubusercontent.com/brunoklein99/deep-learning-notes/master/s # ./bin/main -m open-llama-3b-v2-q8_0.gguf --lora lora-open-llama-3b-v2-q8_0-shakespeare-LATEST.bin # mac os mistral +dataset=slqm \ data_dir="../models/" \ model_dir="../models/ggmls/" \ model_name="openbuddy-mistral-7b-v13.1-q2_k" \ -dataset=slqm \ sh examples/finetune/finetune.sh >${model_name}.log #云电脑 mistral +dataset=slqm \ data_dir="../" \ model_dir="../" \ -dataset=slqm \ model_name="openbuddy-mistral-7b-v13.1-q2_k" \ sh examples/finetune/finetune.sh >${model_name}.log # mac os llama2 +dataset=slqm \ data_dir="../models/" \ model_dir="../models/ggmls/" \ model_name="chinese-llama-2-7b-16k.Q2_K" \ -dataset=slqm \ sh examples/finetune/finetune.sh >${model_name}.log #云电脑 llama2 +dataset="slqm" \ data_dir="../" \ model_dir="../" \ -dataset=slqm \ model_name="chinese-llama-2-7b-16k.Q2_K" \ sh examples/finetune/finetune.sh >${model_name}.log #云电脑 finetune-bc +dataset="slqm" \ data_dir="../" \ model_dir="../" \ -dataset=slqm \ model_name="bc2-7b-chat-q2_k" \ cmd="finetune-bc" \ sh examples/finetune/finetune.sh >${model_name}.log diff --git a/examples/finetune/finetune.sh b/examples/finetune/finetune.sh index 4abe4818c..db7eba67a 100644 --- a/examples/finetune/finetune.sh +++ b/examples/finetune/finetune.sh @@ -12,22 +12,22 @@ fi ./${finetune} \ --train-data ${data_dir}/${dataset}.txt \ --model-base ${model_dir}/${model_name}.gguf \ - --checkpoint-in ${model_dir}/chk/chk-${model_name}-LATEST.gguf \ - --checkpoint-out ${model_dir}/chk/chk-${model_name}-ITERATION.gguf \ - --lora-out ${model_dir}/lora/lora-${model_name}-ITERATION.bin \ + --checkpoint-in ${model_dir}/chk/chk-${dataset}-${model_name}-LATEST.gguf \ + --checkpoint-out ${model_dir}/chk/chk-${dataset}-${model_name}-ITERATION.gguf \ + --lora-out ${model_dir}/lora/lora-${dataset}-${model_name}-ITERATION.bin \ --threads 4 --ctx 64 --batch 4 --adam-iter 1 --save-every 5 \ --lora-r 8 --lora-alpha 16 \ + --grad-acc 1 \ + --escape \ + --epochs 3 \ --use-checkpointing -# # --grad-acc 1 \ -# # --use-flash \ -# # --escape \ # # --seed 1 ./export-lora \ --model-base ${model_dir}/${model_name}.gguf \ --model-out ${model_dir}/${dataset}-${model_name}.gguf \ - --lora-scaled ${model_dir}/lora/lora-${model_name}-LATEST.bin 1.0 + --lora-scaled ${model_dir}/lora/lora-${dataset}-${model_name}-LATEST.bin 1.0 ./main \ -m ${model_dir}/${dataset}-${model_name}.gguf \