From 5d45adf47dc9942ec0aff1e35913d027cca607a0 Mon Sep 17 00:00:00 2001 From: hiyouga <467089858@qq.com> Date: Tue, 28 May 2024 13:44:22 +0800 Subject: [PATCH] fix #3931 --- .dockerignore | 2 ++ README.md | 2 +- README_zh.md | 2 +- 3 files changed, 4 insertions(+), 2 deletions(-) diff --git a/.dockerignore b/.dockerignore index ce67d58a..2ac0e11d 100644 --- a/.dockerignore +++ b/.dockerignore @@ -4,6 +4,8 @@ .venv cache data +hf_cache +output examples .dockerignore .gitattributes diff --git a/README.md b/README.md index 78312e07..974b30d0 100644 --- a/README.md +++ b/README.md @@ -45,7 +45,7 @@ Choose your path: ## Features - **Various models**: LLaMA, LLaVA, Mistral, Mixtral-MoE, Qwen, Yi, Gemma, Baichuan, ChatGLM, Phi, etc. -- **Integrated methods**: (Continuous) pre-training, (multimodal) supervised fine-tuning, reward modeling, PPO, DPO, KTO and ORPO. +- **Integrated methods**: (Continuous) pre-training, (multimodal) supervised fine-tuning, reward modeling, PPO, DPO, KTO, ORPO, etc. - **Scalable resources**: 32-bit full-tuning, 16-bit freeze-tuning, 16-bit LoRA and 2/4/8-bit QLoRA via AQLM/AWQ/GPTQ/LLM.int8. - **Advanced algorithms**: GaLore, BAdam, DoRA, LongLoRA, LLaMA Pro, Mixture-of-Depths, LoRA+, LoftQ and Agent tuning. - **Practical tricks**: FlashAttention-2, Unsloth, RoPE scaling, NEFTune and rsLoRA. diff --git a/README_zh.md b/README_zh.md index 5acf3dd1..7106bbab 100644 --- a/README_zh.md +++ b/README_zh.md @@ -45,7 +45,7 @@ https://github.com/hiyouga/LLaMA-Factory/assets/16256802/ec36a9dd-37f4-4f72-81bd ## 项目特色 - **多种模型**:LLaMA、LLaVA、Mistral、Mixtral-MoE、Qwen、Yi、Gemma、Baichuan、ChatGLM、Phi 等等。 -- **集成方法**:(增量)预训练、(多模态)指令监督微调、奖励模型训练、PPO 训练、DPO 训练、KTO 训练和 ORPO 训练。 +- **集成方法**:(增量)预训练、(多模态)指令监督微调、奖励模型训练、PPO 训练、DPO 训练、KTO 训练、ORPO 训练等等。 - **多种精度**:32 比特全参数微调、16 比特冻结微调、16 比特 LoRA 微调和基于 AQLM/AWQ/GPTQ/LLM.int8 的 2/4/8 比特 QLoRA 微调。 - **先进算法**:GaLore、BAdam、DoRA、LongLoRA、LLaMA Pro、Mixture-of-Depths、LoRA+、LoftQ 和 Agent 微调。 - **实用技巧**:FlashAttention-2、Unsloth、RoPE scaling、NEFTune 和 rsLoRA。