From ad6e7c76c7b5590ea6ca55c4f76db2d3206b5987 Mon Sep 17 00:00:00 2001 From: hiyouga Date: Thu, 10 Aug 2023 15:20:44 +0800 Subject: [PATCH] fix webui val size --- src/llmtuner/webui/components/sft.py | 6 +++--- src/llmtuner/webui/locales.py | 4 ++-- src/llmtuner/webui/runner.py | 6 +++--- 3 files changed, 8 insertions(+), 8 deletions(-) diff --git a/src/llmtuner/webui/components/sft.py b/src/llmtuner/webui/components/sft.py index 678693b9..9db57ac9 100644 --- a/src/llmtuner/webui/components/sft.py +++ b/src/llmtuner/webui/components/sft.py @@ -38,7 +38,7 @@ def create_sft_tab(top_elems: Dict[str, "Component"], runner: "Runner") -> Dict[ value="cosine", choices=[scheduler.value for scheduler in SchedulerType] ) max_grad_norm = gr.Textbox(value="1.0") - dev_ratio = gr.Slider(value=0, minimum=0, maximum=1, step=0.001) + val_size = gr.Slider(value=0, minimum=0, maximum=1, step=0.001) with gr.Accordion(label="Advanced config", open=False) as advanced_tab: with gr.Row(): @@ -88,7 +88,7 @@ def create_sft_tab(top_elems: Dict[str, "Component"], runner: "Runner") -> Dict[ gradient_accumulation_steps, lr_scheduler_type, max_grad_norm, - dev_ratio, + val_size, logging_steps, save_steps, warmup_steps, @@ -122,7 +122,7 @@ def create_sft_tab(top_elems: Dict[str, "Component"], runner: "Runner") -> Dict[ gradient_accumulation_steps=gradient_accumulation_steps, lr_scheduler_type=lr_scheduler_type, max_grad_norm=max_grad_norm, - dev_ratio=dev_ratio, + val_size=val_size, advanced_tab=advanced_tab, logging_steps=logging_steps, save_steps=save_steps, diff --git a/src/llmtuner/webui/locales.py b/src/llmtuner/webui/locales.py index ba9a73bd..9bf3d7a9 100644 --- a/src/llmtuner/webui/locales.py +++ b/src/llmtuner/webui/locales.py @@ -227,9 +227,9 @@ LOCALES = { "info": "用于梯度裁剪的范数。" } }, - "dev_ratio": { + "val_size": { "en": { - "label": "Dev ratio", + "label": "Val size", "info": "Proportion of data in the dev set." }, "zh": { diff --git a/src/llmtuner/webui/runner.py b/src/llmtuner/webui/runner.py index 763ff614..fddc4070 100644 --- a/src/llmtuner/webui/runner.py +++ b/src/llmtuner/webui/runner.py @@ -83,7 +83,7 @@ class Runner: gradient_accumulation_steps: int, lr_scheduler_type: str, max_grad_norm: str, - dev_ratio: float, + val_size: float, logging_steps: int, save_steps: int, warmup_steps: int, @@ -139,8 +139,8 @@ class Runner: output_dir=output_dir ) - if dev_ratio > 1e-6: - args["dev_ratio"] = dev_ratio + if val_size > 1e-6: + args["val_size"] = val_size args["evaluation_strategy"] = "steps" args["eval_steps"] = save_steps args["load_best_model_at_end"] = True