From c6a4d43c068cc6e1f680e7c2f08603800b598039 Mon Sep 17 00:00:00 2001 From: wql Date: Thu, 5 Sep 2024 07:05:47 +0000 Subject: [PATCH] fix: remove no need test file --- ...ichuan-7B_4_gpu_50_step_20240905050501.yml | 31 ------------------- ...chuan-7B_4_gpu_50_step_20240905050958.yaml | 31 ------------------- ...chuan-7B_4_gpu_50_step_20240905051039.yaml | 31 ------------------- ..._Qwen-7B_4_gpu_50_step_20240905052241.yaml | 31 ------------------- ..._Qwen-7B_4_gpu_50_step_20240905053758.yaml | 31 ------------------- ...g.txt &\n train_pid=1720\n echo Start" | 0 ..._Qwen-7B_4_gpu_50_step_20240905064243.yaml | 31 ------------------- .../log.txt | 0 ..._Qwen-7B_4_gpu_50_step_20240905064736.yaml | 31 ------------------- 9 files changed, 217 deletions(-) delete mode 100644 results/lora_sft_Baichuan-7B_4_gpu_50_step_20240905050501/lora_sft_Baichuan-7B_4_gpu_50_step_20240905050501.yml delete mode 100644 results/lora_sft_Baichuan-7B_4_gpu_50_step_20240905050958/lora_sft_Baichuan-7B_4_gpu_50_step_20240905050958.yaml delete mode 100644 results/lora_sft_Baichuan-7B_4_gpu_50_step_20240905051039/lora_sft_Baichuan-7B_4_gpu_50_step_20240905051039.yaml delete mode 100644 results/lora_sft_Qwen-7B_4_gpu_50_step_20240905052241/lora_sft_Qwen-7B_4_gpu_50_step_20240905052241.yaml delete mode 100644 results/lora_sft_Qwen-7B_4_gpu_50_step_20240905053758/lora_sft_Qwen-7B_4_gpu_50_step_20240905053758.yaml delete mode 100644 "results/lora_sft_Qwen-7B_4_gpu_50_step_20240905064243/log.txt &\n train_pid=1720\n echo Start" delete mode 100644 results/lora_sft_Qwen-7B_4_gpu_50_step_20240905064243/lora_sft_Qwen-7B_4_gpu_50_step_20240905064243.yaml delete mode 100644 results/lora_sft_Qwen-7B_4_gpu_50_step_20240905064736/log.txt delete mode 100644 results/lora_sft_Qwen-7B_4_gpu_50_step_20240905064736/lora_sft_Qwen-7B_4_gpu_50_step_20240905064736.yaml diff --git a/results/lora_sft_Baichuan-7B_4_gpu_50_step_20240905050501/lora_sft_Baichuan-7B_4_gpu_50_step_20240905050501.yml b/results/lora_sft_Baichuan-7B_4_gpu_50_step_20240905050501/lora_sft_Baichuan-7B_4_gpu_50_step_20240905050501.yml deleted file mode 100644 index 26507813..00000000 --- a/results/lora_sft_Baichuan-7B_4_gpu_50_step_20240905050501/lora_sft_Baichuan-7B_4_gpu_50_step_20240905050501.yml +++ /dev/null @@ -1,31 +0,0 @@ -bf16: true -cutoff_len: 1024 -dataset: belle_1m -ddp_timeout: 180000000 -do_train: true -eval_steps: 500 -eval_strategy: steps -finetuning_type: lora -gradient_accumulation_steps: 8 -include_num_input_tokens_seen: true -include_tokens_per_second: true -learning_rate: 0.0001 -logging_steps: 3 -lora_target: all -lr_scheduler_type: cosine -max_samples: 10000 -max_steps: '50' -model_name_or_path: '' -num_train_epochs: 10.0 -output_dir: ./results/lora_sft_Baichuan-7B_4_gpu_50_step_20240905050501 -overwrite_cache: true -overwrite_output_dir: true -per_device_eval_batch_size: 2 -per_device_train_batch_size: 2 -plot_loss: true -preprocessing_num_workers: 16 -save_steps: 500 -stage: sft -template: '' -val_size: 0.1 -warmup_ratio: 0.1 diff --git a/results/lora_sft_Baichuan-7B_4_gpu_50_step_20240905050958/lora_sft_Baichuan-7B_4_gpu_50_step_20240905050958.yaml b/results/lora_sft_Baichuan-7B_4_gpu_50_step_20240905050958/lora_sft_Baichuan-7B_4_gpu_50_step_20240905050958.yaml deleted file mode 100644 index e041b60b..00000000 --- a/results/lora_sft_Baichuan-7B_4_gpu_50_step_20240905050958/lora_sft_Baichuan-7B_4_gpu_50_step_20240905050958.yaml +++ /dev/null @@ -1,31 +0,0 @@ -bf16: true -cutoff_len: 1024 -dataset: belle_1m -ddp_timeout: 180000000 -do_train: true -eval_steps: 500 -eval_strategy: steps -finetuning_type: lora -gradient_accumulation_steps: 8 -include_num_input_tokens_seen: true -include_tokens_per_second: true -learning_rate: 0.0001 -logging_steps: 3 -lora_target: all -lr_scheduler_type: cosine -max_samples: 10000 -max_steps: '50' -model_name_or_path: '' -num_train_epochs: 10.0 -output_dir: ./results/lora_sft_Baichuan-7B_4_gpu_50_step_20240905050958 -overwrite_cache: true -overwrite_output_dir: true -per_device_eval_batch_size: 2 -per_device_train_batch_size: 2 -plot_loss: true -preprocessing_num_workers: 16 -save_steps: 500 -stage: sft -template: '' -val_size: 0.1 -warmup_ratio: 0.1 diff --git a/results/lora_sft_Baichuan-7B_4_gpu_50_step_20240905051039/lora_sft_Baichuan-7B_4_gpu_50_step_20240905051039.yaml b/results/lora_sft_Baichuan-7B_4_gpu_50_step_20240905051039/lora_sft_Baichuan-7B_4_gpu_50_step_20240905051039.yaml deleted file mode 100644 index 84e13b18..00000000 --- a/results/lora_sft_Baichuan-7B_4_gpu_50_step_20240905051039/lora_sft_Baichuan-7B_4_gpu_50_step_20240905051039.yaml +++ /dev/null @@ -1,31 +0,0 @@ -bf16: true -cutoff_len: 1024 -dataset: belle_1m -ddp_timeout: 180000000 -do_train: true -eval_steps: 500 -eval_strategy: steps -finetuning_type: lora -gradient_accumulation_steps: 8 -include_num_input_tokens_seen: true -include_tokens_per_second: true -learning_rate: 0.0001 -logging_steps: 3 -lora_target: all -lr_scheduler_type: cosine -max_samples: 10000 -max_steps: '50' -model_name_or_path: '' -num_train_epochs: 10.0 -output_dir: ./results/lora_sft_Baichuan-7B_4_gpu_50_step_20240905051039 -overwrite_cache: true -overwrite_output_dir: true -per_device_eval_batch_size: 2 -per_device_train_batch_size: 2 -plot_loss: true -preprocessing_num_workers: 16 -save_steps: 500 -stage: sft -template: '' -val_size: 0.1 -warmup_ratio: 0.1 diff --git a/results/lora_sft_Qwen-7B_4_gpu_50_step_20240905052241/lora_sft_Qwen-7B_4_gpu_50_step_20240905052241.yaml b/results/lora_sft_Qwen-7B_4_gpu_50_step_20240905052241/lora_sft_Qwen-7B_4_gpu_50_step_20240905052241.yaml deleted file mode 100644 index 2a1de0fe..00000000 --- a/results/lora_sft_Qwen-7B_4_gpu_50_step_20240905052241/lora_sft_Qwen-7B_4_gpu_50_step_20240905052241.yaml +++ /dev/null @@ -1,31 +0,0 @@ -bf16: true -cutoff_len: 1024 -dataset: belle_1m -ddp_timeout: 180000000 -do_train: true -eval_steps: 500 -eval_strategy: steps -finetuning_type: lora -gradient_accumulation_steps: 8 -include_num_input_tokens_seen: true -include_tokens_per_second: true -learning_rate: 0.0001 -logging_steps: 3 -lora_target: all -lr_scheduler_type: cosine -max_samples: 10000 -max_steps: '50' -model_name_or_path: ../../models/Qwen-7B -num_train_epochs: 10.0 -output_dir: ./results/lora_sft_Qwen-7B_4_gpu_50_step_20240905052241 -overwrite_cache: true -overwrite_output_dir: true -per_device_eval_batch_size: 2 -per_device_train_batch_size: 2 -plot_loss: true -preprocessing_num_workers: 16 -save_steps: 500 -stage: sft -template: qwen -val_size: 0.1 -warmup_ratio: 0.1 diff --git a/results/lora_sft_Qwen-7B_4_gpu_50_step_20240905053758/lora_sft_Qwen-7B_4_gpu_50_step_20240905053758.yaml b/results/lora_sft_Qwen-7B_4_gpu_50_step_20240905053758/lora_sft_Qwen-7B_4_gpu_50_step_20240905053758.yaml deleted file mode 100644 index caa1505f..00000000 --- a/results/lora_sft_Qwen-7B_4_gpu_50_step_20240905053758/lora_sft_Qwen-7B_4_gpu_50_step_20240905053758.yaml +++ /dev/null @@ -1,31 +0,0 @@ -bf16: true -cutoff_len: 1024 -dataset: belle_1m -ddp_timeout: 180000000 -do_train: true -eval_steps: 500 -eval_strategy: steps -finetuning_type: lora -gradient_accumulation_steps: 8 -include_num_input_tokens_seen: true -include_tokens_per_second: true -learning_rate: 0.0001 -logging_steps: 3 -lora_target: all -lr_scheduler_type: cosine -max_samples: 10000 -max_steps: '50' -model_name_or_path: ../../models/Qwen-7B -num_train_epochs: 10.0 -output_dir: ./results/lora_sft_Qwen-7B_4_gpu_50_step_20240905053758 -overwrite_cache: true -overwrite_output_dir: true -per_device_eval_batch_size: 2 -per_device_train_batch_size: 2 -plot_loss: true -preprocessing_num_workers: 16 -save_steps: 500 -stage: sft -template: qwen -val_size: 0.1 -warmup_ratio: 0.1 diff --git "a/results/lora_sft_Qwen-7B_4_gpu_50_step_20240905064243/log.txt &\n train_pid=1720\n echo Start" "b/results/lora_sft_Qwen-7B_4_gpu_50_step_20240905064243/log.txt &\n train_pid=1720\n echo Start" deleted file mode 100644 index e69de29b..00000000 diff --git a/results/lora_sft_Qwen-7B_4_gpu_50_step_20240905064243/lora_sft_Qwen-7B_4_gpu_50_step_20240905064243.yaml b/results/lora_sft_Qwen-7B_4_gpu_50_step_20240905064243/lora_sft_Qwen-7B_4_gpu_50_step_20240905064243.yaml deleted file mode 100644 index 4631b614..00000000 --- a/results/lora_sft_Qwen-7B_4_gpu_50_step_20240905064243/lora_sft_Qwen-7B_4_gpu_50_step_20240905064243.yaml +++ /dev/null @@ -1,31 +0,0 @@ -bf16: true -cutoff_len: 1024 -dataset: belle_1m -ddp_timeout: 180000000 -do_train: true -eval_steps: 500 -eval_strategy: steps -finetuning_type: lora -gradient_accumulation_steps: 8 -include_num_input_tokens_seen: true -include_tokens_per_second: true -learning_rate: 0.0001 -logging_steps: 3 -lora_target: all -lr_scheduler_type: cosine -max_samples: 10000 -max_steps: '50' -model_name_or_path: ../../models/Qwen-7B -num_train_epochs: 10.0 -output_dir: ./results/lora_sft_Qwen-7B_4_gpu_50_step_20240905064243 -overwrite_cache: true -overwrite_output_dir: true -per_device_eval_batch_size: 2 -per_device_train_batch_size: 2 -plot_loss: true -preprocessing_num_workers: 16 -save_steps: 500 -stage: sft -template: qwen -val_size: 0.1 -warmup_ratio: 0.1 diff --git a/results/lora_sft_Qwen-7B_4_gpu_50_step_20240905064736/log.txt b/results/lora_sft_Qwen-7B_4_gpu_50_step_20240905064736/log.txt deleted file mode 100644 index e69de29b..00000000 diff --git a/results/lora_sft_Qwen-7B_4_gpu_50_step_20240905064736/lora_sft_Qwen-7B_4_gpu_50_step_20240905064736.yaml b/results/lora_sft_Qwen-7B_4_gpu_50_step_20240905064736/lora_sft_Qwen-7B_4_gpu_50_step_20240905064736.yaml deleted file mode 100644 index fe61b1e4..00000000 --- a/results/lora_sft_Qwen-7B_4_gpu_50_step_20240905064736/lora_sft_Qwen-7B_4_gpu_50_step_20240905064736.yaml +++ /dev/null @@ -1,31 +0,0 @@ -bf16: true -cutoff_len: 1024 -dataset: belle_1m -ddp_timeout: 180000000 -do_train: true -eval_steps: 500 -eval_strategy: steps -finetuning_type: lora -gradient_accumulation_steps: 8 -include_num_input_tokens_seen: true -include_tokens_per_second: true -learning_rate: 0.0001 -logging_steps: 3 -lora_target: all -lr_scheduler_type: cosine -max_samples: 10000 -max_steps: '50' -model_name_or_path: ../../models/Qwen-7B -num_train_epochs: 10.0 -output_dir: ./results/lora_sft_Qwen-7B_4_gpu_50_step_20240905064736 -overwrite_cache: true -overwrite_output_dir: true -per_device_eval_batch_size: 2 -per_device_train_batch_size: 2 -plot_loss: true -preprocessing_num_workers: 16 -save_steps: 500 -stage: sft -template: qwen -val_size: 0.1 -warmup_ratio: 0.1