Due to changing priorities, I no longer have time to dedicate maintaining this project. Hopefully the ipex team will support this natively in the future to make it easier for customers to use LLM technology on Intel products. See other repos like https://github.com/eleiton/ollama-intel-arc for alternate support.
This repo illustrates the use of Ollama with support for Intel ARC GPU based via ipex-llm and Ollama Portable ZIP support. Run the recently released deepseek-r1 model on your local Intel ARC GPU based PC using Linux
All Ollama based ipex-llm defects should be reported directly to the ipex-llm project at https://github.com/intel/ipex-llm
- Ubuntu 24.04 or newer (for Intel ARC GPU kernel driver support. Tested with Ubuntu 24.04.02
- Installed Docker and Docker-compose tools
- Intel ARC series GPU (tested with Intel ARC A770 16GB and Intel(R) Core(TM) Ultra 5 125H integrated GPU)
The following will build the Ollama with Intel ARC GPU support, and compose those with the public docker image based on OpenWEB UI from https://github.com/open-webui/open-webui
Linux:
$ git clone https://github.com/mattcurf/ollama-intel-gpu
$ cd ollama-intel-gpu
$ docker compose up Note: On some platforms or versions of docker, the command above may instead be 'docker-compose'.
Note 2: If you have multiple GPU's installed (like integrated and discrete), set the ONEAPI_DEVICE_DELECTOR environment variable in the docker compose file to select the intended device to use.
Then launch your web browser to http://localhost:3000 to launch the web ui. Create a local OpenWeb UI credential, then click the settings icon in the top right of the screen, then select 'Models', then click 'Show', then download a model like 'llama3.1:8b-instruct-q8_0' for Intel ARC A770 16GB VRAM
To update to the latest portable zip version of IPEX-LLM's Ollama, update the compose file with the build arguments shown below, using the latest ollama-*.tgz release from https://github.com/ipex-llm/ipex-llm/releases/tag/v2.3.0-nightly , then rebuild the image.
ollama-intel-gpu:
build:
context: .
dockerfile: Dockerfile
args:
IPEXLLM_RELEASE_REPO: ipex-llm/ipex-llm
IPEXLLM_RELEASE_VERSON: v2.3.0-nightly
IPEXLLM_PORTABLE_ZIP_FILENAME: ollama-ipex-llm-2.3.0b20250415-ubuntu.tgz