change: change yaml

This commit is contained in:
wql 2024-08-20 15:48:16 +08:00
parent 07b328ee23
commit d8a730dcfe
25 changed files with 48 additions and 24 deletions

View File

@ -5,12 +5,12 @@ CUDA_VISIBLE_DEVICES=0 llamafactory-cli train results/lora_sft/Llama2-7B/llama2_
CUDA_VISIBLE_DEVICES=0 llamafactory-cli train results/lora_sft/Llama2-7B/llama2_lora_sft_2_single.yaml | tee results/lora_sft/Llama2-7B/llama2_lora_sft_2_single.txt
CUDA_VISIBLE_DEVICES=0 llamafactory-cli train results/lora_sft/Llama2-7B/llama2_lora_sft_3_single.yaml | tee results/lora_sft/Llama2-7B/llama2_lora_sft_3_single.txt
FORCE_TORCHRUN=1 llamafactory-cli train results/inference/Llama2-7B/llama2_predict_1.yaml | tee results/inference/Llama2-7B/llama2_predict_1.txt
FORCE_TORCHRUN=1 llamafactory-cli train results/inference/Llama2-7B/llama2_predict_2.yaml | tee results/inference/Llama2-7B/llama2_predict_2.txt
FORCE_TORCHRUN=1 llamafactory-cli train results/inference/Llama2-7B/llama2_predict_3.yaml | tee results/inference/Llama2-7B/llama2_predict_3.txt
CUDA_VISIBLE_DEVICES=0 llamafactory-cli train results/inference/Llama2-7B/llama2_predict_1_single.yaml | tee results/inference/Llama2-7B/llama2_predict_1_single.txt
CUDA_VISIBLE_DEVICES=0 llamafactory-cli train results/inference/Llama2-7B/llama2_predict_2_single.yaml | tee results/inference/Llama2-7B/llama2_predict_2_single.txt
CUDA_VISIBLE_DEVICES=0 llamafactory-cli train results/inference/Llama2-7B/llama2_predict_3_single.yaml | tee results/inference/Llama2-7B/llama2_predict_3_single.txt
# FORCE_TORCHRUN=1 llamafactory-cli train results/inference/Llama2-7B/llama2_predict_1.yaml | tee results/inference/Llama2-7B/llama2_predict_1.txt
# FORCE_TORCHRUN=1 llamafactory-cli train results/inference/Llama2-7B/llama2_predict_2.yaml | tee results/inference/Llama2-7B/llama2_predict_2.txt
# FORCE_TORCHRUN=1 llamafactory-cli train results/inference/Llama2-7B/llama2_predict_3.yaml | tee results/inference/Llama2-7B/llama2_predict_3.txt
# CUDA_VISIBLE_DEVICES=0 llamafactory-cli train results/inference/Llama2-7B/llama2_predict_1_single.yaml | tee results/inference/Llama2-7B/llama2_predict_1_single.txt
# CUDA_VISIBLE_DEVICES=0 llamafactory-cli train results/inference/Llama2-7B/llama2_predict_2_single.yaml | tee results/inference/Llama2-7B/llama2_predict_2_single.txt
# CUDA_VISIBLE_DEVICES=0 llamafactory-cli train results/inference/Llama2-7B/llama2_predict_3_single.yaml | tee results/inference/Llama2-7B/llama2_predict_3_single.txt
# FORCE_TORCHRUN=1 llamafactory-cli train results/lora_sft/Baichuan2-7B/Baichuan2_lora_sft_1.yaml | tee results/lora_sft/Baichuan2-7B/Baichuan2_lora_sft_1.txt
# FORCE_TORCHRUN=1 llamafactory-cli train results/lora_sft/Baichuan2-7B/Baichuan2_lora_sft_2.yaml | tee results/lora_sft/Baichuan2-7B/Baichuan2_lora_sft_2.txt
@ -19,12 +19,12 @@ CUDA_VISIBLE_DEVICES=0 llamafactory-cli train results/inference/Llama2-7B/llama2
# CUDA_VISIBLE_DEVICES=0 llamafactory-cli train results/lora_sft/Baichuan2-7B/Baichuan2_lora_sft_2_single.yaml | tee results/lora_sft/Baichuan2-7B/Baichuan2_lora_sft_2_single.txt
# CUDA_VISIBLE_DEVICES=0 llamafactory-cli train results/lora_sft/Baichuan2-7B/Baichuan2_lora_sft_3_single.yaml | tee results/lora_sft/Baichuan2-7B/Baichuan2_lora_sft_3_single.txt
# FORCE_TORCHRUN=1 llamafactory-cli train results/lora_sft/ChatGLM2-6B/ChatGLM2_lora_sft_1.yaml | tee results/lora_sft/ChatGLM2-6B/ChatGLM2_lora_sft_1.txt
# FORCE_TORCHRUN=1 llamafactory-cli train results/lora_sft/ChatGLM2-6B/ChatGLM2_lora_sft_2.yaml | tee results/lora_sft/ChatGLM2-6B/ChatGLM2_lora_sft_2.txt
# FORCE_TORCHRUN=1 llamafactory-cli train results/lora_sft/ChatGLM2-6B/ChatGLM2_lora_sft_3.yaml | tee results/lora_sft/ChatGLM2-6B/ChatGLM2_lora_sft_3.txt
# CUDA_VISIBLE_DEVICES=0 llamafactory-cli train results/lora_sft/ChatGLM2-6B/ChatGLM2_lora_sft_1_single.yaml | tee results/lora_sft/ChatGLM2-6B/ChatGLM2_lora_sft_1_single.txt
# CUDA_VISIBLE_DEVICES=0 llamafactory-cli train results/lora_sft/ChatGLM2-6B/ChatGLM2_lora_sft_2_single.yaml | tee results/lora_sft/ChatGLM2-6B/ChatGLM2_lora_sft_2_single.txt
# CUDA_VISIBLE_DEVICES=0 llamafactory-cli train results/lora_sft/ChatGLM2-6B/ChatGLM2_lora_sft_3_single.yaml | tee results/lora_sft/ChatGLM2-6B/ChatGLM2_lora_sft_3_single.txt
FORCE_TORCHRUN=1 llamafactory-cli train results/lora_sft/ChatGLM2-6B/ChatGLM2_lora_sft_1.yaml | tee results/lora_sft/ChatGLM2-6B/ChatGLM2_lora_sft_1.txt
FORCE_TORCHRUN=1 llamafactory-cli train results/lora_sft/ChatGLM2-6B/ChatGLM2_lora_sft_2.yaml | tee results/lora_sft/ChatGLM2-6B/ChatGLM2_lora_sft_2.txt
FORCE_TORCHRUN=1 llamafactory-cli train results/lora_sft/ChatGLM2-6B/ChatGLM2_lora_sft_3.yaml | tee results/lora_sft/ChatGLM2-6B/ChatGLM2_lora_sft_3.txt
CUDA_VISIBLE_DEVICES=0 llamafactory-cli train results/lora_sft/ChatGLM2-6B/ChatGLM2_lora_sft_1_single.yaml | tee results/lora_sft/ChatGLM2-6B/ChatGLM2_lora_sft_1_single.txt
CUDA_VISIBLE_DEVICES=0 llamafactory-cli train results/lora_sft/ChatGLM2-6B/ChatGLM2_lora_sft_2_single.yaml | tee results/lora_sft/ChatGLM2-6B/ChatGLM2_lora_sft_2_single.txt
CUDA_VISIBLE_DEVICES=0 llamafactory-cli train results/lora_sft/ChatGLM2-6B/ChatGLM2_lora_sft_3_single.yaml | tee results/lora_sft/ChatGLM2-6B/ChatGLM2_lora_sft_3_single.txt
# FORCE_TORCHRUN=1 llamafactory-cli train results/lora_sft/Qwen-7B/Qwen_lora_sft_1.yaml | tee results/lora_sft/Qwen-7B/Qwen_lora_sft_1.txt
# FORCE_TORCHRUN=1 llamafactory-cli train results/lora_sft/Qwen-7B/Qwen_lora_sft_2.yaml | tee results/lora_sft/Qwen-7B/Qwen_lora_sft_2.txt

View File

@ -32,6 +32,7 @@ warmup_ratio: 0.1
bf16: true
ddp_timeout: 180000000
max_steps: 1000
include_num_input_tokens_seen: true
### eval
val_size: 0.1

View File

@ -32,6 +32,7 @@ warmup_ratio: 0.1
bf16: true
ddp_timeout: 180000000
max_steps: 1000
include_num_input_tokens_seen: true
### eval
val_size: 0.1

View File

@ -32,6 +32,7 @@ warmup_ratio: 0.1
bf16: true
ddp_timeout: 180000000
max_steps: 1000
include_num_input_tokens_seen: true
### eval
val_size: 0.1

View File

@ -32,6 +32,7 @@ warmup_ratio: 0.1
bf16: true
ddp_timeout: 180000000
max_steps: 1000
include_num_input_tokens_seen: true
### eval
val_size: 0.1

View File

@ -32,6 +32,7 @@ warmup_ratio: 0.1
bf16: true
ddp_timeout: 180000000
max_steps: 1000
include_num_input_tokens_seen: true
### eval
val_size: 0.1

View File

@ -32,6 +32,7 @@ warmup_ratio: 0.1
bf16: true
ddp_timeout: 180000000
max_steps: 1000
include_num_input_tokens_seen: true
### eval
val_size: 0.1

View File

@ -1,5 +1,5 @@
### model
model_name_or_path: ZhipuAI/chatglm2-6b
model_name_or_path: ~/data
### method
stage: sft
@ -32,6 +32,7 @@ warmup_ratio: 0.1
bf16: true
ddp_timeout: 180000000
max_steps: 1000
include_num_input_tokens_seen: true
### eval
val_size: 0.1

View File

@ -1,5 +1,5 @@
### model
model_name_or_path: ZhipuAI/chatglm2-6b
model_name_or_path: ~/data
### method
stage: sft
@ -32,6 +32,7 @@ warmup_ratio: 0.1
bf16: true
ddp_timeout: 180000000
max_steps: 1000
include_num_input_tokens_seen: true
### eval
val_size: 0.1

View File

@ -1,5 +1,5 @@
### model
model_name_or_path: ZhipuAI/chatglm2-6b
model_name_or_path: ~/data
### method
stage: sft
@ -32,6 +32,7 @@ warmup_ratio: 0.1
bf16: true
ddp_timeout: 180000000
max_steps: 1000
include_num_input_tokens_seen: true
### eval
val_size: 0.1

View File

@ -1,5 +1,5 @@
### model
model_name_or_path: ZhipuAI/chatglm2-6b
model_name_or_path: ~/data
### method
stage: sft
@ -32,6 +32,7 @@ warmup_ratio: 0.1
bf16: true
ddp_timeout: 180000000
max_steps: 1000
include_num_input_tokens_seen: true
### eval
val_size: 0.1

View File

@ -1,5 +1,5 @@
### model
model_name_or_path: ZhipuAI/chatglm2-6b
model_name_or_path: ~/data
### method
stage: sft
@ -32,6 +32,7 @@ warmup_ratio: 0.1
bf16: true
ddp_timeout: 180000000
max_steps: 1000
include_num_input_tokens_seen: true
### eval
val_size: 0.1

View File

@ -1,5 +1,5 @@
### model
model_name_or_path: ZhipuAI/chatglm2-6b
model_name_or_path: ~/data
### method
stage: sft
@ -32,6 +32,7 @@ warmup_ratio: 0.1
bf16: true
ddp_timeout: 180000000
max_steps: 1000
include_num_input_tokens_seen: true
### eval
val_size: 0.1

View File

@ -31,7 +31,8 @@ lr_scheduler_type: cosine
warmup_ratio: 0.1
fp16: true
ddp_timeout: 180000000
max_steps: 500
max_steps: 1000
include_num_input_tokens_seen: true
### eval
val_size: 0.1

View File

@ -31,7 +31,8 @@ lr_scheduler_type: cosine
warmup_ratio: 0.1
fp16: true
ddp_timeout: 180000000
max_steps: 500
max_steps: 1000
include_num_input_tokens_seen: true
### eval
val_size: 0.1

View File

@ -31,7 +31,8 @@ lr_scheduler_type: cosine
warmup_ratio: 0.1
fp16: true
ddp_timeout: 180000000
max_steps: 500
max_steps: 1000
include_num_input_tokens_seen: true
### eval
val_size: 0.1

View File

@ -31,7 +31,8 @@ lr_scheduler_type: cosine
warmup_ratio: 0.1
fp16: true
ddp_timeout: 180000000
max_steps: 500
max_steps: 1000
include_num_input_tokens_seen: true
### eval
val_size: 0.1

View File

@ -31,7 +31,8 @@ lr_scheduler_type: cosine
warmup_ratio: 0.1
fp16: true
ddp_timeout: 180000000
max_steps: 500
max_steps: 1000
include_num_input_tokens_seen: true
### eval
val_size: 0.1

View File

@ -31,7 +31,8 @@ lr_scheduler_type: cosine
warmup_ratio: 0.1
fp16: true
ddp_timeout: 180000000
max_steps: 500
max_steps: 1000
include_num_input_tokens_seen: true
### eval
val_size: 0.1

View File

@ -32,6 +32,7 @@ warmup_ratio: 0.1
bf16: true
ddp_timeout: 180000000
max_steps: 1000
include_num_input_tokens_seen: true
### eval
val_size: 0.1

View File

@ -32,6 +32,7 @@ warmup_ratio: 0.1
bf16: true
ddp_timeout: 180000000
max_steps: 1000
include_num_input_tokens_seen: true
### eval
val_size: 0.1

View File

@ -32,6 +32,7 @@ warmup_ratio: 0.1
bf16: true
ddp_timeout: 180000000
max_steps: 1000
include_num_input_tokens_seen: true
### eval
val_size: 0.1

View File

@ -32,6 +32,7 @@ warmup_ratio: 0.1
bf16: true
ddp_timeout: 180000000
max_steps: 1000
include_num_input_tokens_seen: true
### eval
val_size: 0.1

View File

@ -32,6 +32,7 @@ warmup_ratio: 0.1
bf16: true
ddp_timeout: 180000000
max_steps: 1000
include_num_input_tokens_seen: true
### eval
val_size: 0.1

View File

@ -32,6 +32,7 @@ warmup_ratio: 0.1
bf16: true
ddp_timeout: 180000000
max_steps: 1000
include_num_input_tokens_seen: true
### eval
val_size: 0.1