Update Docker configurations for deployment improvements
Revised `IPEXLLM_RELEASE_REPO` value and adjusted file and path references for consistency. Updated `docker-compose.yml` with refined environment variables, device mapping, restart policies, and added necessary port bindings for better functionality and maintainability.
This commit is contained in:
+1
-1
@@ -24,7 +24,7 @@ RUN mkdir -p /tmp/gpu && \
|
|||||||
rm *.deb
|
rm *.deb
|
||||||
|
|
||||||
# Install Ollama Portable Zip
|
# Install Ollama Portable Zip
|
||||||
ARG IPEXLLM_RELEASE_REPO=intel/ipex-llm
|
ARG IPEXLLM_RELEASE_REPO=ipex-llm/ipex-llm
|
||||||
ARG IPEXLLM_RELEASE_VERSON=v2.2.0
|
ARG IPEXLLM_RELEASE_VERSON=v2.2.0
|
||||||
ARG IPEXLLM_PORTABLE_ZIP_FILENAME=ollama-ipex-llm-2.2.0-ubuntu.tgz
|
ARG IPEXLLM_PORTABLE_ZIP_FILENAME=ollama-ipex-llm-2.2.0-ubuntu.tgz
|
||||||
RUN cd / && \
|
RUN cd / && \
|
||||||
|
|||||||
+34
-5
@@ -4,20 +4,37 @@ services:
|
|||||||
context: .
|
context: .
|
||||||
dockerfile: Dockerfile
|
dockerfile: Dockerfile
|
||||||
args:
|
args:
|
||||||
IPEXLLM_RELEASE_REPO: intel/ipex-llm
|
IPEXLLM_RELEASE_REPO: ipex-llm/ipex-llm
|
||||||
IPEXLLM_RELEASE_VERSON: v2.2.0
|
IPEXLLM_RELEASE_VERSON: v2.2.0
|
||||||
IPEXLLM_PORTABLE_ZIP_FILENAME: ollama-ipex-llm-2.2.0-ubuntu.tar.gz
|
IPEXLLM_PORTABLE_ZIP_FILENAME: ollama-ipex-llm-2.2.0-ubuntu.tgz
|
||||||
container_name: ollama-intel-gpu
|
container_name: ollama-intel-gpu
|
||||||
restart: always
|
restart: unless-stopped
|
||||||
devices:
|
devices:
|
||||||
- /dev/dri:/dev/dri
|
# - /dev/dri:/dev/dri
|
||||||
|
- /dev/dri/renderD128:/dev/dri/renderD128
|
||||||
volumes:
|
volumes:
|
||||||
|
- /tmp/.X11-unix:/tmp/.X11-unix
|
||||||
- ollama-intel-gpu:/root/.ollama
|
- ollama-intel-gpu:/root/.ollama
|
||||||
environment:
|
environment:
|
||||||
- ONEAPI_DEVICE_SELECTOR=level_zero:0
|
- ONEAPI_DEVICE_SELECTOR=level_zero:0
|
||||||
- IPEX_LLM_NUM_CTX=16384
|
- IPEX_LLM_NUM_CTX=16384
|
||||||
|
- LD_LIBRARY_PATH:/opt/intel/oneapi/compiler/2024.2/lib
|
||||||
|
- DISPLAY=${DISPLAY}
|
||||||
|
- OLLAMA_DEFAULT_KEEPALIVE="6h"
|
||||||
|
- OLLAMA_HOST=0.0.0.0
|
||||||
|
- OLLAMA_KEEP_ALIVE=24h
|
||||||
|
- OLLAMA_MAX_LOADED_MODELS=1
|
||||||
|
- OLLAMA_MAX_QUEUE=512
|
||||||
|
- OLLAMA_MAX_VRAM=0
|
||||||
|
#- OLLAMA_NOHISTORY=false
|
||||||
|
#- OLLAMA_NOPRUNE=false
|
||||||
|
- OLLAMA_NUM_PARALLEL=1
|
||||||
|
|
||||||
|
ports:
|
||||||
|
- 11435:11434
|
||||||
|
|
||||||
ollama-webui:
|
ollama-webui:
|
||||||
image: ghcr.io/open-webui/open-webui
|
image: ghcr.io/open-webui/open-webui:latest
|
||||||
container_name: ollama-webui
|
container_name: ollama-webui
|
||||||
volumes:
|
volumes:
|
||||||
- ollama-webui:/app/backend/data
|
- ollama-webui:/app/backend/data
|
||||||
@@ -27,6 +44,18 @@ services:
|
|||||||
- ${OLLAMA_WEBUI_PORT-3000}:8080
|
- ${OLLAMA_WEBUI_PORT-3000}:8080
|
||||||
environment:
|
environment:
|
||||||
- OLLAMA_BASE_URL=http://ollama-intel-gpu:11434
|
- OLLAMA_BASE_URL=http://ollama-intel-gpu:11434
|
||||||
|
- OLLAMA_DEFAULT_KEEPALIVE="6h"
|
||||||
|
#- OPENAI_API_BASE_URL=
|
||||||
|
#- OPENAI_API_KEY=
|
||||||
|
#
|
||||||
|
# AUTOMATIC1111_BASE_URL="http://localhost:7860"
|
||||||
|
|
||||||
|
- ENABLE_RAG_WEB_SEARCH=True
|
||||||
|
|
||||||
|
# DO NOT TRACK
|
||||||
|
- SCARF_NO_ANALYTICS=true
|
||||||
|
- DO_NOT_TRACK=true
|
||||||
|
- ANONYMIZED_TELEMETRY=false
|
||||||
extra_hosts:
|
extra_hosts:
|
||||||
- host.docker.internal:host-gateway
|
- host.docker.internal:host-gateway
|
||||||
restart: unless-stopped
|
restart: unless-stopped
|
||||||
|
|||||||
Reference in New Issue
Block a user