From e1f4e53915fc4dcc309e2b1bea27f6d11f63083a Mon Sep 17 00:00:00 2001 From: hiyouga Date: Wed, 15 May 2024 19:20:11 +0800 Subject: [PATCH] add NPU docker images --- README.md | 7 ++++++- README_zh.md | 5 +++++ 2 files changed, 11 insertions(+), 1 deletion(-) diff --git a/README.md b/README.md index 826512c6..0643e8ef 100644 --- a/README.md +++ b/README.md @@ -342,7 +342,7 @@ To enable FlashAttention-2 on the Windows platform, you need to install the prec
For Ascend NPU users -To utilize Ascend NPU devices for (distributed) training and inference, you need to install the **[torch-npu](https://gitee.com/ascend/pytorch)** package and the **[Ascend CANN Kernels](https://www.hiascend.com/developer/download/community/result?module=cann)**. +To utilize Ascend NPU devices for (distributed) training and inference, you need to install the **[torch-npu](https://gitee.com/ascend/pytorch)** library and the **[Ascend CANN Kernels](https://www.hiascend.com/developer/download/community/result?module=cann)**. | Requirement | Minimum | Recommend | | ------------ | ------- | --------- | @@ -351,6 +351,11 @@ To utilize Ascend NPU devices for (distributed) training and inference, you need | torch-npu | 2.2.0 | 2.2.0 | | deepspeed | 0.13.2 | 0.13.2 | +Docker image: + +- 32GB: [Download page](http://mirrors.cn-central-221.ovaijisuan.com/detail/130.html) +- 64GB: Coming soon + Remember to use `ASCEND_RT_VISIBLE_DEVICES` instead of `CUDA_VISIBLE_DEVICES` to specify the device to use. If you cannot infer model on NPU devices, try setting `do_sample: false` in the configurations. diff --git a/README_zh.md b/README_zh.md index d41ff13a..47f7e111 100644 --- a/README_zh.md +++ b/README_zh.md @@ -351,6 +351,11 @@ pip install https://github.com/jllllll/bitsandbytes-windows-webui/releases/downl | torch-npu | 2.2.0 | 2.2.0 | | deepspeed | 0.13.2 | 0.13.2 | +Docker 镜像: + +- 32GB:[下载地址](http://mirrors.cn-central-221.ovaijisuan.com/detail/130.html) +- 64GB:敬请期待 + 请记得使用 `ASCEND_RT_VISIBLE_DEVICES` 而非 `CUDA_VISIBLE_DEVICES` 来指定您使用的设备。 如果遇到无法正常推理的情况,请尝试设置 `do_sample: false`。