From 751dd77bc076af3a13884a609ec71c089c516f1f Mon Sep 17 00:00:00 2001 From: hoshi-hiyouga Date: Thu, 6 Jun 2024 23:38:09 +0800 Subject: [PATCH] Update cli.py --- src/llamafactory/cli.py | 4 ---- 1 file changed, 4 deletions(-) diff --git a/src/llamafactory/cli.py b/src/llamafactory/cli.py index 8a229a38..19a104bd 100644 --- a/src/llamafactory/cli.py +++ b/src/llamafactory/cli.py @@ -71,10 +71,6 @@ def main(): export_model() elif command == Command.TRAIN: if get_device_count() > 0: - # NOTE (MengqingCao): why use torchrun when only one accelerator is available? - # DeepSpeed only warp model with DeepSpeedEngine when launching by distributed launcher, - # e.g., torchrun, causing some feature missing - # sa: https://github.com/huggingface/transformers/issues/24309 master_addr = os.environ.get("MASTER_ADDR", "127.0.0.1") master_port = os.environ.get("MASTER_PORT", str(random.randint(20001, 29999))) logger.info("Initializing distributed tasks at: {}:{}".format(master_addr, master_port))