This commit is contained in:
hiyouga 2023-12-28 18:47:19 +08:00
parent e165354fac
commit 65c5b0477c
4 changed files with 8 additions and 4 deletions

View File

@ -479,7 +479,9 @@ python src/export_model.py \
--adapter_name_or_path path_to_checkpoint \
--template default \
--finetuning_type lora \
--export_dir path_to_export
--export_dir path_to_export \
--export_size 2 \
--export_legacy_format False
```
> [!WARNING]

View File

@ -479,7 +479,9 @@ python src/export_model.py \
--adapter_name_or_path path_to_checkpoint \
--template default \
--finetuning_type lora \
--export_dir path_to_export
--export_dir path_to_export \
--export_size 2 \
--export_legacy_format False
```
> [!WARNING]

View File

@ -102,7 +102,7 @@ class ModelArguments:
default=1024,
metadata={"help": "The maximum length of the model inputs used for quantization."}
)
export_lecacy_format: Optional[bool] = field(
export_legacy_format: Optional[bool] = field(
default=False,
metadata={"help": "Whether or not to save the `.bin` files instead of `.safetensors`."}
)

View File

@ -50,7 +50,7 @@ def export_model(args: Optional[Dict[str, Any]] = None):
model.save_pretrained(
save_directory=model_args.export_dir,
max_shard_size="{}GB".format(model_args.export_size),
safe_serialization=(not model_args.export_lecacy_format)
safe_serialization=(not model_args.export_legacy_format)
)
try: