fix: Repair the issue where quantization failed after merging the adapter.
This commit is contained in:
parent
bb0a37dc06
commit
f91a9a250a
|
@ -66,7 +66,7 @@ def save_model(
|
||||||
error = ALERTS["err_no_dataset"][lang]
|
error = ALERTS["err_no_dataset"][lang]
|
||||||
elif export_quantization_bit not in GPTQ_BITS and not checkpoint_path:
|
elif export_quantization_bit not in GPTQ_BITS and not checkpoint_path:
|
||||||
error = ALERTS["err_no_adapter"][lang]
|
error = ALERTS["err_no_adapter"][lang]
|
||||||
elif export_quantization_bit in GPTQ_BITS and isinstance(checkpoint_path, list):
|
elif export_quantization_bit in GPTQ_BITS and checkpoint_path and isinstance(checkpoint_path, list):
|
||||||
error = ALERTS["err_gptq_lora"][lang]
|
error = ALERTS["err_gptq_lora"][lang]
|
||||||
|
|
||||||
if error:
|
if error:
|
||||||
|
|
Loading…
Reference in New Issue