diff --git a/data/dataset_info.json b/data/dataset_info.json index 4a135794..86015600 100644 --- a/data/dataset_info.json +++ b/data/dataset_info.json @@ -190,7 +190,7 @@ "tags": { "role_tag": "role", "content_tag": "content", - "user_tag": "human", + "user_tag": "user", "assistant_tag": "assistant" }, "formatting": "sharegpt" diff --git a/src/llmtuner/model/adapter.py b/src/llmtuner/model/adapter.py index b0dc6489..14863a10 100644 --- a/src/llmtuner/model/adapter.py +++ b/src/llmtuner/model/adapter.py @@ -63,8 +63,8 @@ def init_adapter( freeze_modules = {"all"} for name, _ in model.named_modules(): - if "0." in name: - freeze_modules.add(name.split("0.")[-1].split(".")[0]) + if ".0." in name: + freeze_modules.add(name.split(".0.")[-1].split(".")[0]) trainable_layers = [] for module_name in finetuning_args.name_module_trainable: