This commit is contained in:
hiyouga 2024-06-25 01:51:29 +08:00
parent d0f953bf5b
commit efb81b25ec
6 changed files with 14 additions and 8 deletions

View File

@ -448,7 +448,8 @@ docker build -f ./docker/docker-cuda/Dockerfile \
-t llamafactory:latest .
docker run -dit --gpus=all \
-v ./hf_cache:/root/.cache/huggingface/ \
-v ./hf_cache:/root/.cache/huggingface \
-v ./ms_cache:/root/.cache/modelscope \
-v ./data:/app/data \
-v ./output:/app/output \
-p 7860:7860 \
@ -471,7 +472,8 @@ docker build -f ./docker/docker-npu/Dockerfile \
# Change `device` upon your resources
docker run -dit \
-v ./hf_cache:/root/.cache/huggingface/ \
-v ./hf_cache:/root/.cache/huggingface \
-v ./ms_cache:/root/.cache/modelscope \
-v ./data:/app/data \
-v ./output:/app/output \
-v /usr/local/dcmi:/usr/local/dcmi \

View File

@ -448,7 +448,8 @@ docker build -f ./docker/docker-cuda/Dockerfile \
-t llamafactory:latest .
docker run -dit --gpus=all \
-v ./hf_cache:/root/.cache/huggingface/ \
-v ./hf_cache:/root/.cache/huggingface \
-v ./ms_cache:/root/.cache/modelscope \
-v ./data:/app/data \
-v ./output:/app/output \
-p 7860:7860 \
@ -471,7 +472,8 @@ docker build -f ./docker/docker-npu/Dockerfile \
# 根据您的资源更改 `device`
docker run -dit \
-v ./hf_cache:/root/.cache/huggingface/ \
-v ./hf_cache:/root/.cache/huggingface \
-v ./ms_cache:/root/.cache/modelscope \
-v ./data:/app/data \
-v ./output:/app/output \
-v /usr/local/dcmi:/usr/local/dcmi \

View File

@ -36,7 +36,7 @@ RUN EXTRA_PACKAGES="metrics"; \
pip uninstall -y transformer-engine flash-attn
# Set up volumes
VOLUME [ "/root/.cache/huggingface/", "/app/data", "/app/output" ]
VOLUME [ "/root/.cache/huggingface", "/root/.cache/modelscope", "/app/data", "/app/output" ]
# Expose port 7860 for the LLaMA Board
ENV GRADIO_SERVER_PORT 7860

View File

@ -10,7 +10,8 @@ services:
PIP_INDEX: https://pypi.org/simple
container_name: llamafactory
volumes:
- ./hf_cache:/root/.cache/huggingface/
- ./hf_cache:/root/.cache/huggingface
- ./ms_cache:/root/.cache/modelscope
- ./data:/app/data
- ./output:/app/output
ports:

View File

@ -30,7 +30,7 @@ RUN EXTRA_PACKAGES="torch-npu,metrics"; \
pip uninstall -y transformer-engine flash-attn
# Set up volumes
VOLUME [ "/root/.cache/huggingface/", "/app/data", "/app/output" ]
VOLUME [ "/root/.cache/huggingface", "/root/.cache/modelscope", "/app/data", "/app/output" ]
# Expose port 7860 for the LLaMA Board
ENV GRADIO_SERVER_PORT 7860

View File

@ -8,7 +8,8 @@ services:
PIP_INDEX: https://pypi.org/simple
container_name: llamafactory
volumes:
- ./hf_cache:/root/.cache/huggingface/
- ./hf_cache:/root/.cache/huggingface
- ./ms_cache:/root/.cache/modelscope
- ./data:/app/data
- ./output:/app/output
- /usr/local/dcmi:/usr/local/dcmi