diff --git a/README.md b/README.md index 468a5e606..fe9fbb5bf 100644 --- a/README.md +++ b/README.md @@ -116,15 +116,15 @@ cd data && ./download.sh alpaca && cd - # Saves lora only ./scripts/run_finetune_with_lora.sh \ - --model_name_or_path gpt2 \ + --model_name_or_path facebook/galactica-1.3b \ --dataset_path data/alpaca/train \ - --output_lora_path output_models/finetuned_gpt2_lora + --output_lora_path output_models/finetuned_galactica_lora # Saves lora and merges into original model ./scripts/run_finetune_with_lora_save_aggregated_weights.sh \ - --model_name_or_path gpt2 \ + --model_name_or_path facebook/galactica-1.3b \ --dataset_path data/alpaca/train \ - --output_model_path output_models/finetuned_gpt2 + --output_model_path output_models/finetuned_galactica ``` ### Inference diff --git a/scripts/run_finetune_with_lora.sh b/scripts/run_finetune_with_lora.sh index 54bab49eb..e99989c8e 100755 --- a/scripts/run_finetune_with_lora.sh +++ b/scripts/run_finetune_with_lora.sh @@ -41,7 +41,7 @@ mkdir -p ${output_dir} ${log_dir} deepspeed ${deepspeed_args} \ examples/finetune.py \ - --model_name_or_path facebook/galactica-1.3b \ + --model_name_or_path ${model_name_or_path} \ --dataset_path ${dataset_path} \ --output_dir ${output_dir} --overwrite_output_dir \ --num_train_epochs 0.01 \ @@ -53,7 +53,7 @@ deepspeed ${deepspeed_args} \ --save_aggregated_lora 0\ --deepspeed configs/ds_config_zero2.json \ --fp16 \ - --run_name finetune_with_lora \ + --run_name ${exp_id} \ --validation_split_percentage 0 \ --logging_steps 20 \ --do_train \ diff --git a/scripts/run_finetune_with_lora_save_aggregated_weights.sh b/scripts/run_finetune_with_lora_save_aggregated_weights.sh index c27c1f2bc..2f27c9619 100755 --- a/scripts/run_finetune_with_lora_save_aggregated_weights.sh +++ b/scripts/run_finetune_with_lora_save_aggregated_weights.sh @@ -41,7 +41,7 @@ mkdir -p ${output_dir} ${log_dir} deepspeed ${deepspeed_args} \ examples/finetune.py \ - --model_name_or_path facebook/galactica-1.3b \ + --model_name_or_path ${model_name_or_path} \ --dataset_path ${dataset_path} \ --output_dir ${output_dir} --overwrite_output_dir \ --num_train_epochs 0.01 \ @@ -53,14 +53,10 @@ deepspeed ${deepspeed_args} \ --save_aggregated_lora 1\ --deepspeed configs/ds_config_zero2.json \ --fp16 \ - --run_name finetune_with_lora \ + --run_name ${exp_id} \ --validation_split_percentage 0 \ --logging_steps 20 \ --do_train \ - --do_eval \ - --evaluation_strategy "steps" \ - --eval_steps 1000 \ - --eval_dataset_path ${eval_dataset_path} \ --ddp_timeout 72000 \ --save_steps 5000 \ --dataloader_num_workers 1 \