LLaMA-Factory-310P3/prepare_yaml_file.py

48 lines
1.4 KiB
Python
Raw Normal View History

2024-09-05 11:28:19 +08:00
import sys
import yaml
def main():
run_type = sys.argv[1]
model = sys.argv[2]
max_steps = sys.argv[3]
run_name = sys.argv[4]
output_dir = sys.argv[5]
if run_type == "lora_sft":
2024-09-05 13:03:22 +08:00
yaml_file = './results/lora_sft_template.yaml'
2024-09-05 11:28:19 +08:00
elif run_type == "inference":
2024-09-05 13:03:22 +08:00
yaml_file = './results/predict_template.yaml'
2024-09-05 13:04:54 +08:00
model_name_or_path = ""
template = ""
2024-09-05 11:28:19 +08:00
if model == "9g-8B":
model_name_or_path = "../../models/sft_8b_v2"
template = ""
elif model == "Baichuan2-7B":
model_name_or_path = "../../models/Baichuan-7B"
template = "baichuan"
elif model == "ChatGLM2-6B":
model_name_or_path = "../../models/chatglm2-6b"
template = "chatglm2"
elif model == "Llama2-7B":
model_name_or_path = "../../models/llama-2-7b-ms"
template = "llama2"
elif model == "Qwen-7B":
model_name_or_path = "../../models/Qwen-7B"
template = "qwen"
config = None
with open(yaml_file, 'r', encoding='utf-8') as f:
config = yaml.load(f.read(), Loader=yaml.FullLoader)
config['model_name_or_path'] = model_name_or_path
config['template'] = template
config['output_dir'] = output_dir
if run_type == "lora_sft":
config['max_steps'] = max_steps
with open(f'{output_dir}/{run_name}.yml', 'w', encoding='utf-8') as f:
yaml.dump(data=config, stream=f, allow_unicode=True)
if __name__ == "__main__":
main()