- Improve documentation
- Clean dependencies
- Use the official oneAPI Base Toolkit
Breaking change:
- Container name changed to ollama-intel-arc
This commit is contained in:
eleiton
2025-03-09 23:55:10 +01:00
parent 5dd7e9a28c
commit 8f6f5ee0fd
5 changed files with 98 additions and 53 deletions
+7 -2
View File
@@ -1,10 +1,15 @@
#!/bin/sh
source /opt/intel/oneapi/setvars.sh
export USE_XETLA=OFF
# Reference: https://github.com/intel/ipex-llm/blob/main/docs/mddocs/Quickstart/ollama_quickstart.md#3-run-ollama-serve
export OLLAMA_NUM_GPU=999
export no_proxy=localhost,127.0.0.1
export ZES_ENABLE_SYSMAN=1
source /opt/intel/oneapi/setvars.sh
export SYCL_CACHE_PERSISTENT=1
# [optional] under most circumstances, the following environment variable may improve performance, but sometimes this may also cause performance degradation
export SYCL_PI_LEVEL_ZERO_USE_IMMEDIATE_COMMANDLISTS=1
# [optional] if you want to run on single GPU, use below command to limit GPU may improve performance
export ONEAPI_DEVICE_SELECTOR=level_zero:0
/usr/local/lib/python3.12/dist-packages/bigdl/cpp/libs/ollama serve