Update Docker configurations for deployment improvements

Revised `IPEXLLM_RELEASE_REPO` value and adjusted file and path references for consistency. Updated `docker-compose.yml` with refined environment variables, device mapping, restart policies, and added necessary port bindings for better functionality and maintainability.
This commit is contained in:
2025-04-22 17:56:04 +01:00
parent 8172339ca5
commit 17592946fa
2 changed files with 35 additions and 6 deletions
+1 -1
View File
@@ -24,7 +24,7 @@ RUN mkdir -p /tmp/gpu && \
rm *.deb rm *.deb
# Install Ollama Portable Zip # Install Ollama Portable Zip
ARG IPEXLLM_RELEASE_REPO=intel/ipex-llm ARG IPEXLLM_RELEASE_REPO=ipex-llm/ipex-llm
ARG IPEXLLM_RELEASE_VERSON=v2.2.0 ARG IPEXLLM_RELEASE_VERSON=v2.2.0
ARG IPEXLLM_PORTABLE_ZIP_FILENAME=ollama-ipex-llm-2.2.0-ubuntu.tgz ARG IPEXLLM_PORTABLE_ZIP_FILENAME=ollama-ipex-llm-2.2.0-ubuntu.tgz
RUN cd / && \ RUN cd / && \
+34 -5
View File
@@ -4,20 +4,37 @@ services:
context: . context: .
dockerfile: Dockerfile dockerfile: Dockerfile
args: args:
IPEXLLM_RELEASE_REPO: intel/ipex-llm IPEXLLM_RELEASE_REPO: ipex-llm/ipex-llm
IPEXLLM_RELEASE_VERSON: v2.2.0 IPEXLLM_RELEASE_VERSON: v2.2.0
IPEXLLM_PORTABLE_ZIP_FILENAME: ollama-ipex-llm-2.2.0-ubuntu.tar.gz IPEXLLM_PORTABLE_ZIP_FILENAME: ollama-ipex-llm-2.2.0-ubuntu.tgz
container_name: ollama-intel-gpu container_name: ollama-intel-gpu
restart: always restart: unless-stopped
devices: devices:
- /dev/dri:/dev/dri # - /dev/dri:/dev/dri
- /dev/dri/renderD128:/dev/dri/renderD128
volumes: volumes:
- /tmp/.X11-unix:/tmp/.X11-unix
- ollama-intel-gpu:/root/.ollama - ollama-intel-gpu:/root/.ollama
environment: environment:
- ONEAPI_DEVICE_SELECTOR=level_zero:0 - ONEAPI_DEVICE_SELECTOR=level_zero:0
- IPEX_LLM_NUM_CTX=16384 - IPEX_LLM_NUM_CTX=16384
- LD_LIBRARY_PATH:/opt/intel/oneapi/compiler/2024.2/lib
- DISPLAY=${DISPLAY}
- OLLAMA_DEFAULT_KEEPALIVE="6h"
- OLLAMA_HOST=0.0.0.0
- OLLAMA_KEEP_ALIVE=24h
- OLLAMA_MAX_LOADED_MODELS=1
- OLLAMA_MAX_QUEUE=512
- OLLAMA_MAX_VRAM=0
#- OLLAMA_NOHISTORY=false
#- OLLAMA_NOPRUNE=false
- OLLAMA_NUM_PARALLEL=1
ports:
- 11435:11434
ollama-webui: ollama-webui:
image: ghcr.io/open-webui/open-webui image: ghcr.io/open-webui/open-webui:latest
container_name: ollama-webui container_name: ollama-webui
volumes: volumes:
- ollama-webui:/app/backend/data - ollama-webui:/app/backend/data
@@ -27,6 +44,18 @@ services:
- ${OLLAMA_WEBUI_PORT-3000}:8080 - ${OLLAMA_WEBUI_PORT-3000}:8080
environment: environment:
- OLLAMA_BASE_URL=http://ollama-intel-gpu:11434 - OLLAMA_BASE_URL=http://ollama-intel-gpu:11434
- OLLAMA_DEFAULT_KEEPALIVE="6h"
#- OPENAI_API_BASE_URL=
#- OPENAI_API_KEY=
#
# AUTOMATIC1111_BASE_URL="http://localhost:7860"
- ENABLE_RAG_WEB_SEARCH=True
# DO NOT TRACK
- SCARF_NO_ANALYTICS=true
- DO_NOT_TRACK=true
- ANONYMIZED_TELEMETRY=false
extra_hosts: extra_hosts:
- host.docker.internal:host-gateway - host.docker.internal:host-gateway
restart: unless-stopped restart: unless-stopped