add llamafactory-cli env

This commit is contained in:
hiyouga 2024-06-06 01:28:14 +08:00
parent 67fe822324
commit d4908d5708
4 changed files with 65 additions and 5 deletions

View File

@ -20,10 +20,10 @@ body:
attributes:
label: System Info
description: |
Please share your system info with us. You can run the command **transformers-cli env** and copy-paste its output below.
请提供您的系统信息。您可以在命令行运行 **transformers-cli env** 并将其输出复制到该文本框中。
Please share your system info with us. You can run the command **llamafactory-cli env** and copy-paste its output below.
请提供您的系统信息。您可以在命令行运行 **llamafactory-cli env** 并将其输出复制到该文本框中。
placeholder: transformers version, platform, python version, ...
placeholder: llamafactory version, platform, python version, ...
- type: textarea
id: reproduction

View File

@ -8,6 +8,7 @@ from . import launcher
from .api.app import run_api
from .chat.chat_model import run_chat
from .eval.evaluator import run_eval
from .extras.env import VERSION, print_env
from .extras.logging import get_logger
from .extras.misc import get_device_count
from .train.tuner import export_model, run_exp
@ -29,8 +30,6 @@ USAGE = (
+ "-" * 70
)
VERSION = "0.7.2.dev0"
WELCOME = (
"-" * 58
+ "\n"
@ -50,6 +49,7 @@ logger = get_logger(__name__)
class Command(str, Enum):
API = "api"
CHAT = "chat"
ENV = "env"
EVAL = "eval"
EXPORT = "export"
TRAIN = "train"
@ -65,6 +65,8 @@ def main():
run_api()
elif command == Command.CHAT:
run_chat()
elif command == Command.ENV:
print_env()
elif command == Command.EVAL:
run_eval()
elif command == Command.EXPORT:

View File

@ -0,0 +1,54 @@
import platform
import accelerate
import datasets
import peft
import torch
import transformers
import trl
from transformers.utils import is_bitsandbytes_available, is_torch_cuda_available, is_torch_npu_available
from .packages import is_deepspeed_available, is_vllm_available
VERSION = "0.7.2.dev0"
def print_env() -> None:
info = {
"`llamafactory` version": VERSION,
"Platform": platform.platform(),
"Python version": platform.python_version(),
"PyTorch version": torch.__version__,
"Transformers version": transformers.__version__,
"Datasets version": datasets.__version__,
"Accelerate version": accelerate.__version__,
"PEFT version": peft.__version__,
"TRL version": trl.__version__,
}
if is_torch_cuda_available():
info["PyTorch version"] += " (GPU)"
info["GPU type"] = torch.cuda.get_device_name()
if is_torch_npu_available():
info["PyTorch version"] += " (NPU)"
info["NPU type"] = torch.npu.get_device_name()
info["CANN version"] = torch.version.cann
if is_deepspeed_available():
import deepspeed # type: ignore
info["DeepSpeed version"] = deepspeed.__version__
if is_bitsandbytes_available():
import bitsandbytes
info["Bitsandbytes version"] = bitsandbytes.__version__
if is_vllm_available():
import vllm
info["vLLM version"] = vllm.__version__
print("\n".join(["- {}: {}".format(key, value) for key, value in info.items()]) + "\n")

View File

@ -20,6 +20,10 @@ def _get_package_version(name: str) -> "Version":
return version.parse("0.0.0")
def is_deepspeed_available():
return _is_package_available("deepspeed")
def is_fastapi_available():
return _is_package_available("fastapi")