@@ -36,7 +36,7 @@ RUN EXTRA_PACKAGES="metrics"; \
|
||||
pip uninstall -y transformer-engine flash-attn
|
||||
|
||||
# Set up volumes
|
||||
VOLUME [ "/root/.cache/huggingface/", "/app/data", "/app/output" ]
|
||||
VOLUME [ "/root/.cache/huggingface", "/root/.cache/modelscope", "/app/data", "/app/output" ]
|
||||
|
||||
# Expose port 7860 for the LLaMA Board
|
||||
ENV GRADIO_SERVER_PORT 7860
|
||||
|
||||
@@ -10,7 +10,8 @@ services:
|
||||
PIP_INDEX: https://pypi.org/simple
|
||||
container_name: llamafactory
|
||||
volumes:
|
||||
- ./hf_cache:/root/.cache/huggingface/
|
||||
- ./hf_cache:/root/.cache/huggingface
|
||||
- ./ms_cache:/root/.cache/modelscope
|
||||
- ./data:/app/data
|
||||
- ./output:/app/output
|
||||
ports:
|
||||
|
||||
@@ -30,7 +30,7 @@ RUN EXTRA_PACKAGES="torch-npu,metrics"; \
|
||||
pip uninstall -y transformer-engine flash-attn
|
||||
|
||||
# Set up volumes
|
||||
VOLUME [ "/root/.cache/huggingface/", "/app/data", "/app/output" ]
|
||||
VOLUME [ "/root/.cache/huggingface", "/root/.cache/modelscope", "/app/data", "/app/output" ]
|
||||
|
||||
# Expose port 7860 for the LLaMA Board
|
||||
ENV GRADIO_SERVER_PORT 7860
|
||||
|
||||
@@ -8,7 +8,8 @@ services:
|
||||
PIP_INDEX: https://pypi.org/simple
|
||||
container_name: llamafactory
|
||||
volumes:
|
||||
- ./hf_cache:/root/.cache/huggingface/
|
||||
- ./hf_cache:/root/.cache/huggingface
|
||||
- ./ms_cache:/root/.cache/modelscope
|
||||
- ./data:/app/data
|
||||
- ./output:/app/output
|
||||
- /usr/local/dcmi:/usr/local/dcmi
|
||||
|
||||
Reference in New Issue
Block a user