From 7fbe8add8f449358c9815c5ba8a2052a2d874dab Mon Sep 17 00:00:00 2001 From: hiyouga Date: Wed, 24 Apr 2024 23:42:59 +0800 Subject: [PATCH] fix log level --- src/llmtuner/model/utils/rope.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/src/llmtuner/model/utils/rope.py b/src/llmtuner/model/utils/rope.py index 9163253b..93ab8929 100644 --- a/src/llmtuner/model/utils/rope.py +++ b/src/llmtuner/model/utils/rope.py @@ -30,7 +30,7 @@ def configure_rope(config: "PretrainedConfig", model_args: "ModelArguments", is_ current_max_length = getattr(config, "max_position_embeddings", None) if current_max_length and model_args.model_max_length > current_max_length: - logger.warning( + logger.info( "Enlarge max model length from {} to {}.".format(current_max_length, model_args.model_max_length) ) setattr(config, "max_position_embeddings", model_args.model_max_length)