From 17592946fabee3d59830f05c3dd2339cf3acc6c4 Mon Sep 17 00:00:00 2001 From: Andriy Oblivantsev Date: Tue, 22 Apr 2025 17:56:04 +0100 Subject: [PATCH] Update Docker configurations for deployment improvements Revised `IPEXLLM_RELEASE_REPO` value and adjusted file and path references for consistency. Updated `docker-compose.yml` with refined environment variables, device mapping, restart policies, and added necessary port bindings for better functionality and maintainability. --- Dockerfile | 2 +- docker-compose.yml | 39 ++++++++++++++++++++++++++++++++++----- 2 files changed, 35 insertions(+), 6 deletions(-) diff --git a/Dockerfile b/Dockerfile index a95627e..0b0f2bd 100644 --- a/Dockerfile +++ b/Dockerfile @@ -24,7 +24,7 @@ RUN mkdir -p /tmp/gpu && \ rm *.deb # Install Ollama Portable Zip -ARG IPEXLLM_RELEASE_REPO=intel/ipex-llm +ARG IPEXLLM_RELEASE_REPO=ipex-llm/ipex-llm ARG IPEXLLM_RELEASE_VERSON=v2.2.0 ARG IPEXLLM_PORTABLE_ZIP_FILENAME=ollama-ipex-llm-2.2.0-ubuntu.tgz RUN cd / && \ diff --git a/docker-compose.yml b/docker-compose.yml index d525736..bd1e216 100644 --- a/docker-compose.yml +++ b/docker-compose.yml @@ -4,20 +4,37 @@ services: context: . dockerfile: Dockerfile args: - IPEXLLM_RELEASE_REPO: intel/ipex-llm + IPEXLLM_RELEASE_REPO: ipex-llm/ipex-llm IPEXLLM_RELEASE_VERSON: v2.2.0 - IPEXLLM_PORTABLE_ZIP_FILENAME: ollama-ipex-llm-2.2.0-ubuntu.tar.gz + IPEXLLM_PORTABLE_ZIP_FILENAME: ollama-ipex-llm-2.2.0-ubuntu.tgz container_name: ollama-intel-gpu - restart: always + restart: unless-stopped devices: - - /dev/dri:/dev/dri + # - /dev/dri:/dev/dri + - /dev/dri/renderD128:/dev/dri/renderD128 volumes: + - /tmp/.X11-unix:/tmp/.X11-unix - ollama-intel-gpu:/root/.ollama environment: - ONEAPI_DEVICE_SELECTOR=level_zero:0 - IPEX_LLM_NUM_CTX=16384 + - LD_LIBRARY_PATH:/opt/intel/oneapi/compiler/2024.2/lib + - DISPLAY=${DISPLAY} + - OLLAMA_DEFAULT_KEEPALIVE="6h" + - OLLAMA_HOST=0.0.0.0 + - OLLAMA_KEEP_ALIVE=24h + - OLLAMA_MAX_LOADED_MODELS=1 + - OLLAMA_MAX_QUEUE=512 + - OLLAMA_MAX_VRAM=0 + #- OLLAMA_NOHISTORY=false + #- OLLAMA_NOPRUNE=false + - OLLAMA_NUM_PARALLEL=1 + + ports: + - 11435:11434 + ollama-webui: - image: ghcr.io/open-webui/open-webui + image: ghcr.io/open-webui/open-webui:latest container_name: ollama-webui volumes: - ollama-webui:/app/backend/data @@ -27,6 +44,18 @@ services: - ${OLLAMA_WEBUI_PORT-3000}:8080 environment: - OLLAMA_BASE_URL=http://ollama-intel-gpu:11434 + - OLLAMA_DEFAULT_KEEPALIVE="6h" + #- OPENAI_API_BASE_URL= + #- OPENAI_API_KEY= + # + # AUTOMATIC1111_BASE_URL="http://localhost:7860" + + - ENABLE_RAG_WEB_SEARCH=True + + # DO NOT TRACK + - SCARF_NO_ANALYTICS=true + - DO_NOT_TRACK=true + - ANONYMIZED_TELEMETRY=false extra_hosts: - host.docker.internal:host-gateway restart: unless-stopped