Skip to content

Commit

Permalink
Browse files Browse the repository at this point in the history
  • Loading branch information
hiyouga authored and xtchen96 committed Jul 17, 2024
1 parent 313c9a1 commit 35d8ac0
Show file tree
Hide file tree
Showing 6 changed files with 14 additions and 8 deletions.
6 changes: 4 additions & 2 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -448,7 +448,8 @@ docker build -f ./docker/docker-cuda/Dockerfile \
-t llamafactory:latest .

docker run -dit --gpus=all \
-v ./hf_cache:/root/.cache/huggingface/ \
-v ./hf_cache:/root/.cache/huggingface \
-v ./ms_cache:/root/.cache/modelscope \
-v ./data:/app/data \
-v ./output:/app/output \
-p 7860:7860 \
Expand All @@ -471,7 +472,8 @@ docker build -f ./docker/docker-npu/Dockerfile \

# Change `device` upon your resources
docker run -dit \
-v ./hf_cache:/root/.cache/huggingface/ \
-v ./hf_cache:/root/.cache/huggingface \
-v ./ms_cache:/root/.cache/modelscope \
-v ./data:/app/data \
-v ./output:/app/output \
-v /usr/local/dcmi:/usr/local/dcmi \
Expand Down
6 changes: 4 additions & 2 deletions README_zh.md
Original file line number Diff line number Diff line change
Expand Up @@ -448,7 +448,8 @@ docker build -f ./docker/docker-cuda/Dockerfile \
-t llamafactory:latest .

docker run -dit --gpus=all \
-v ./hf_cache:/root/.cache/huggingface/ \
-v ./hf_cache:/root/.cache/huggingface \
-v ./ms_cache:/root/.cache/modelscope \
-v ./data:/app/data \
-v ./output:/app/output \
-p 7860:7860 \
Expand All @@ -471,7 +472,8 @@ docker build -f ./docker/docker-npu/Dockerfile \

# 根据您的资源更改 `device`
docker run -dit \
-v ./hf_cache:/root/.cache/huggingface/ \
-v ./hf_cache:/root/.cache/huggingface \
-v ./ms_cache:/root/.cache/modelscope \
-v ./data:/app/data \
-v ./output:/app/output \
-v /usr/local/dcmi:/usr/local/dcmi \
Expand Down
2 changes: 1 addition & 1 deletion docker/docker-cuda/Dockerfile
Original file line number Diff line number Diff line change
Expand Up @@ -36,7 +36,7 @@ RUN EXTRA_PACKAGES="metrics"; \
pip uninstall -y transformer-engine flash-attn

# Set up volumes
VOLUME [ "/root/.cache/huggingface/", "/app/data", "/app/output" ]
VOLUME [ "/root/.cache/huggingface", "/root/.cache/modelscope", "/app/data", "/app/output" ]

# Expose port 7860 for the LLaMA Board
ENV GRADIO_SERVER_PORT 7860
Expand Down
3 changes: 2 additions & 1 deletion docker/docker-cuda/docker-compose.yml
Original file line number Diff line number Diff line change
Expand Up @@ -10,7 +10,8 @@ services:
PIP_INDEX: https://pypi.org/simple
container_name: llamafactory
volumes:
- ./hf_cache:/root/.cache/huggingface/
- ./hf_cache:/root/.cache/huggingface
- ./ms_cache:/root/.cache/modelscope
- ./data:/app/data
- ./output:/app/output
ports:
Expand Down
2 changes: 1 addition & 1 deletion docker/docker-npu/Dockerfile
Original file line number Diff line number Diff line change
Expand Up @@ -30,7 +30,7 @@ RUN EXTRA_PACKAGES="torch-npu,metrics"; \
pip uninstall -y transformer-engine flash-attn

# Set up volumes
VOLUME [ "/root/.cache/huggingface/", "/app/data", "/app/output" ]
VOLUME [ "/root/.cache/huggingface", "/root/.cache/modelscope", "/app/data", "/app/output" ]

# Expose port 7860 for the LLaMA Board
ENV GRADIO_SERVER_PORT 7860
Expand Down
3 changes: 2 additions & 1 deletion docker/docker-npu/docker-compose.yml
Original file line number Diff line number Diff line change
Expand Up @@ -8,7 +8,8 @@ services:
PIP_INDEX: https://pypi.org/simple
container_name: llamafactory
volumes:
- ./hf_cache:/root/.cache/huggingface/
- ./hf_cache:/root/.cache/huggingface
- ./ms_cache:/root/.cache/modelscope
- ./data:/app/data
- ./output:/app/output
- /usr/local/dcmi:/usr/local/dcmi
Expand Down

0 comments on commit 35d8ac0

Please sign in to comment.