chore: change lr

This commit is contained in:
wql 2024-09-26 11:10:31 +08:00
parent 0b496d2211
commit cc8d2e7ea0
2 changed files with 9 additions and 9 deletions

View File

@ -25,14 +25,14 @@ for ((i=0; i<repeat_cnt; i++)); do
# bash run_once.sh inference Baichuan2-7B 1 500
# bash run_once.sh inference Baichuan2-7B 1 1000
bash run_once.sh lora_sft ChatGLM2-6B 8 500
bash run_once.sh lora_sft ChatGLM2-6B 8 1000
bash run_once.sh lora_sft ChatGLM2-6B 1 500
bash run_once.sh lora_sft ChatGLM2-6B 1 1000
bash run_once.sh inference ChatGLM2-6B 8 500
bash run_once.sh inference ChatGLM2-6B 8 1000
bash run_once.sh inference ChatGLM2-6B 1 500
bash run_once.sh inference ChatGLM2-6B 1 1000
# bash run_once.sh lora_sft ChatGLM2-6B 8 500
# bash run_once.sh lora_sft ChatGLM2-6B 8 1000
# bash run_once.sh lora_sft ChatGLM2-6B 1 500
# bash run_once.sh lora_sft ChatGLM2-6B 1 1000
# bash run_once.sh inference ChatGLM2-6B 8 500
# bash run_once.sh inference ChatGLM2-6B 8 1000
# bash run_once.sh inference ChatGLM2-6B 1 500
# bash run_once.sh inference ChatGLM2-6B 1 1000
bash run_once.sh lora_sft Llama2-7B 8 500
bash run_once.sh lora_sft Llama2-7B 8 1000

View File

@ -25,7 +25,7 @@ overwrite_output_dir: true
### train
per_device_train_batch_size: 2
gradient_accumulation_steps: 8
learning_rate: 1.0e-4
learning_rate: 1.0e-5
num_train_epochs: 10.0
lr_scheduler_type: cosine
warmup_ratio: 0.1