Add GitHub Actions CI to build and push Docker image to GHCR
Workflow triggers on push to main/release branches, tags, PRs, and manual dispatch. Uses Docker Buildx with GHA cache for faster rebuilds. Tags images with ollama version, git SHA, and branch/tag names. Co-authored-by: Cursor <cursoragent@cursor.com>
This commit is contained in:
@@ -6,22 +6,34 @@ A Docker-based setup that pairs [Ollama](https://github.com/ollama/ollama) **v0.
|
||||
|
||||
**Why this exists:** Ollama's official release ships only a Vulkan backend for Intel GPUs, leaving significant performance on the table. This repo builds the `ggml-sycl` backend from source with Intel oneAPI, unlocking oneMKL, oneDNN, and Level-Zero direct GPU access.
|
||||
|
||||

|
||||
|
||||
---
|
||||
|
||||
## Quick start
|
||||
|
||||
### Option A: Build from source
|
||||
|
||||
```shell
|
||||
git clone https://github.com/mattcurf/ollama-intel-gpu
|
||||
cd ollama-intel-gpu
|
||||
docker compose up
|
||||
```
|
||||
|
||||
Open **http://localhost:3000** — pull a model and start chatting.
|
||||
|
||||
The first `docker compose up` builds the SYCL backend from source (~2 min on a modern CPU). Subsequent starts are instant.
|
||||
|
||||
### Option B: Use the pre-built image
|
||||
|
||||
```shell
|
||||
docker run -d \
|
||||
--device /dev/dri:/dev/dri \
|
||||
--shm-size 16G \
|
||||
-p 11434:11434 \
|
||||
-v ollama-data:/root/.ollama \
|
||||
ghcr.io/mattcurf/ollama-intel-gpu:latest
|
||||
```
|
||||
|
||||
Open **http://localhost:3000** (with WebUI) or use the API directly at `http://localhost:11434`.
|
||||
|
||||
> **Multiple GPUs?** Set `ONEAPI_DEVICE_SELECTOR=level_zero:0` in `docker-compose.yml` to pick the right device.
|
||||
|
||||
---
|
||||
|
||||
Reference in New Issue
Block a user