From 14e6e69f6d70e6d9cd99c723a0dbeb0edf4f78ae Mon Sep 17 00:00:00 2001 From: Junyang Lin Date: Sun, 8 Oct 2023 14:50:28 +0800 Subject: [PATCH] Update finetune_lora_ds.sh --- finetune/finetune_lora_ds.sh | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/finetune/finetune_lora_ds.sh b/finetune/finetune_lora_ds.sh index 96aa7b7..30f7883 100644 --- a/finetune/finetune_lora_ds.sh +++ b/finetune/finetune_lora_ds.sh @@ -34,7 +34,7 @@ torchrun $DISTRIBUTED_ARGS finetune.py \ --save_strategy "steps" \ --save_steps 1000 \ --save_total_limit 10 \ - --learning_rate 1e-5 \ + --learning_rate 3e-4 \ --weight_decay 0.1 \ --adam_beta2 0.95 \ --warmup_ratio 0.01 \ @@ -45,4 +45,4 @@ torchrun $DISTRIBUTED_ARGS finetune.py \ --lazy_preprocess True \ --use_lora \ --gradient_checkpointing \ - --deepspeed finetune/ds_config_zero2.json \ No newline at end of file + --deepspeed finetune/ds_config_zero2.json