Commit Graph

24 Commits

Author SHA1 Message Date
971852d3af Rework README for better GitHub presentation
Rewrite README with clear value proposition, architecture diagram,
troubleshooting section, and streamlined structure. Update CHANGELOG
to reflect full history of Vulkan-to-SYCL migration.

Co-authored-by: Cursor <cursoragent@cursor.com>
2026-02-12 17:31:33 +00:00
c56646e7e7 Switch GPU backend from Vulkan to SYCL for ~2x inference performance on Intel GPUs
Build ggml-sycl from upstream llama.cpp (commit a5bb8ba4, matching ollama's
vendored ggml) using Intel oneAPI 2025.1.1 in a multi-stage Docker build.
Patch two ollama-specific API divergences via patch-sycl.py: added batch_size
parameter to graph_compute, removed GGML_TENSOR_FLAG_COMPUTE skip-check that
caused all compute nodes to be bypassed.

Tested: gemma3:1b — 27/27 layers on GPU, 10.2 tok/s gen, 65.3 tok/s prompt eval.
Co-authored-by: Cursor <cursoragent@cursor.com>
2026-02-12 17:28:23 +00:00
8debf2010b Fix ollama not reachable from host due to hardcoded OLLAMA_HOST in entrypoint
The IPEX-LLM bundled start-ollama.sh hardcodes OLLAMA_HOST=127.0.0.1 and
OLLAMA_KEEP_ALIVE=10m, overriding docker-compose environment variables and
preventing external connections through Docker port mapping.

- Add custom start-ollama.sh that honours env vars with sensible defaults
- Mount it read-only into the container
- Fix LD_LIBRARY_PATH env var syntax (: -> =)
- Add .gitignore for IDE/swap/webui data files
- Update CHANGELOG and README with fix documentation

Co-authored-by: Cursor <cursoragent@cursor.com>
2026-02-12 15:18:37 +00:00
96913a2a18 Update Intel GPU stack and ipex-llm to latest available versions
- level-zero v1.22.4 -> v1.28.0
- IGC v2.11.7 -> v2.28.4
- compute-runtime 25.18.33578.6 -> 26.05.37020.3
- libigdgmm 22.7.0 -> 22.9.0
- ipex-llm ollama nightly 2.3.0b20250612 -> 2.3.0b20250725
- Docker compose: disable webui auth, stateless webui volume
- README formatting and GPU model update

Co-authored-by: Cursor <cursoragent@cursor.com>
2026-02-12 15:00:53 +00:00
Adam Gibson
504a1d388f Update default to ipex-llm v2.2.0 (guide for v2.3.0-nightly in docs) 2025-04-16 21:25:38 +08:00
Matt Curfman
61288f5f6c Update to ipex-llm-2.2.0b20250313 2025-03-17 10:44:00 -07:00
Adam Gibson
451f91080c Revert compose to cached .tgz by default. 2025-03-17 19:29:53 +08:00
Adam Gibson
b33c01f1f0 Updated README.md for Dockerfile args. 2025-03-16 16:56:50 +08:00
Matt Curfman
fed3cf9ba0 Update README.md 2025-02-19 15:07:23 -08:00
Matt Curfman
dd84c202a7 Minor fixes 2025-02-19 15:00:46 -08:00
Matt Curfman
2fc526511f Update to use new ipex portable .zip packages 2025-02-19 14:56:56 -08:00
Ikko Eltociear Ashimine
ddd565fc8d docs: update README.md
illlustrates -> illustrates
2025-01-31 02:09:53 +09:00
Matt Curfman
c230c4518e Update README.md 2025-01-26 15:17:30 -08:00
Matt Curfman
8e69333f3f Update README.md 2025-01-26 15:14:29 -08:00
Matt Curfman
b74bab0b6a Update to latest open-webui releases 2025-01-23 17:28:33 -08:00
Matt Curfman
c74f6f2216 Update Dockerfile to use Intel public ipex container 2024-11-14 13:45:57 -08:00
Matt Curfman
40313a7364 Revert to ipex-llm version of ollama for gpu acceleration 2024-08-16 22:41:15 -07:00
Matt Curfman
ab863141ca Update README.md 2024-07-31 21:50:48 -07:00
Matt Curfman
164881bc71 Cleanups 2024-05-23 20:59:10 -07:00
mattcurf
c93791e9ba Add WSL2 support 2024-05-02 10:50:08 -07:00
mattcurf
025b1b0fc9 Misc. cleanups 2024-04-30 14:38:23 -07:00
Matt Curfman
70de05872e Update README.md 2024-04-30 10:28:06 -07:00
mattcurf
2daa02e8f4 Initial version 2024-04-29 17:19:07 -07:00
Matt Curfman
aa3be9a0ab Initial commit 2024-04-29 16:49:25 -07:00