diff --git a/examples/README.md b/examples/README.md index f396ee40..34d5f198 100644 --- a/examples/README.md +++ b/examples/README.md @@ -192,7 +192,7 @@ llamafactory-cli train examples/extras/badam/llama3_full_sft.yaml #### Full-Parameter Fine-Tuning using Adam-mini ```bash -llamafactory-cli train examples/extras/adam_mini/llama3_full_sft.yaml +llamafactory-cli train examples/extras/adam_mini/qwen2_full_sft.yaml ``` #### LoRA+ Fine-Tuning diff --git a/examples/README_zh.md b/examples/README_zh.md index e179d77d..037a7fe6 100644 --- a/examples/README_zh.md +++ b/examples/README_zh.md @@ -192,7 +192,7 @@ llamafactory-cli train examples/extras/badam/llama3_full_sft.yaml #### 使用 Adam-mini 进行全参数训练 ```bash -llamafactory-cli train examples/extras/adam_mini/llama3_full_sft.yaml +llamafactory-cli train examples/extras/adam_mini/qwen2_full_sft.yaml ``` #### LoRA+ 微调 diff --git a/examples/extras/adam_mini/llama3_full_sft.yaml b/examples/extras/adam_mini/qwen2_full_sft.yaml similarity index 84% rename from examples/extras/adam_mini/llama3_full_sft.yaml rename to examples/extras/adam_mini/qwen2_full_sft.yaml index c0c4740b..3637a8e8 100644 --- a/examples/extras/adam_mini/llama3_full_sft.yaml +++ b/examples/extras/adam_mini/qwen2_full_sft.yaml @@ -1,5 +1,5 @@ ### model -model_name_or_path: meta-llama/Meta-Llama-3-8B-Instruct +model_name_or_path: Qwen/Qwen2-1.5B-Instruct ### method stage: sft @@ -9,14 +9,14 @@ use_adam_mini: true ### dataset dataset: identity,alpaca_en_demo -template: llama3 +template: qwen cutoff_len: 1024 max_samples: 1000 overwrite_cache: true preprocessing_num_workers: 16 ### output -output_dir: saves/llama3-8b/full/sft +output_dir: saves/qwen2-1_5b/full/sft logging_steps: 10 save_steps: 500 plot_loss: true