From f91a9a250afa0c2e1554b2c81c1860d7968dc816 Mon Sep 17 00:00:00 2001 From: Liuww <49828783+liuwwang@users.noreply.github.com> Date: Wed, 24 Jul 2024 14:31:29 +0800 Subject: [PATCH] fix: Repair the issue where quantization failed after merging the adapter. --- src/llamafactory/webui/components/export.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/src/llamafactory/webui/components/export.py b/src/llamafactory/webui/components/export.py index 86fad2aa..86a84720 100644 --- a/src/llamafactory/webui/components/export.py +++ b/src/llamafactory/webui/components/export.py @@ -66,7 +66,7 @@ def save_model( error = ALERTS["err_no_dataset"][lang] elif export_quantization_bit not in GPTQ_BITS and not checkpoint_path: error = ALERTS["err_no_adapter"][lang] - elif export_quantization_bit in GPTQ_BITS and isinstance(checkpoint_path, list): + elif export_quantization_bit in GPTQ_BITS and checkpoint_path and isinstance(checkpoint_path, list): error = ALERTS["err_gptq_lora"][lang] if error: