diff --git a/finetune/finetune_lora_ds.sh b/finetune/finetune_lora_ds.sh index 96aa7b7..30f7883 100644 --- a/finetune/finetune_lora_ds.sh +++ b/finetune/finetune_lora_ds.sh @@ -34,7 +34,7 @@ torchrun $DISTRIBUTED_ARGS finetune.py \ --save_strategy "steps" \ --save_steps 1000 \ --save_total_limit 10 \ - --learning_rate 1e-5 \ + --learning_rate 3e-4 \ --weight_decay 0.1 \ --adam_beta2 0.95 \ --warmup_ratio 0.01 \ @@ -45,4 +45,4 @@ torchrun $DISTRIBUTED_ARGS finetune.py \ --lazy_preprocess True \ --use_lora \ --gradient_checkpointing \ - --deepspeed finetune/ds_config_zero2.json \ No newline at end of file + --deepspeed finetune/ds_config_zero2.json