Use the docker image provided by Intel directly, removing the Dockerfile.

This commit is contained in:
eleiton
2025-03-12 19:51:51 +01:00
parent 4456691b41
commit 9d4b38885b
4 changed files with 20 additions and 45 deletions

View File

@@ -1,8 +1,7 @@
version: '3'
services:
ollama-intel-arc:
build: .
image: ollama-intel-arc:latest
image: intelanalytics/ipex-llm-inference-cpp-xpu:latest
container_name: ollama-intel-arc
restart: unless-stopped
devices:
@@ -11,6 +10,15 @@ services:
- ollama-volume:/root/.ollama
ports:
- 11434:11434
environment:
- no_proxy=localhost,127.0.0.1
- OLLAMA_HOST=0.0.0.0
- DEVICE=Arc
- OLLAMA_INTEL_GPU=true
- OLLAMA_NUM_GPU=999
- ZES_ENABLE_SYSMAN=1
command: sh -c 'mkdir -p /llm/ollama && cd /llm/ollama && init-ollama && exec ./ollama serve'
open-webui:
image: ghcr.io/open-webui/open-webui:latest
container_name: open-webui