Skip to content
This repository was archived by the owner on Jun 4, 2025. It is now read-only.

mattcurf/ollama-intel-gpu

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

64 Commits
 
 
 
 
 
 
 
 

Repository files navigation

Update

Due to changing priorities, I no longer have time to dedicate maintaining this project. Hopefully the ipex team will support this natively in the future to make it easier for customers to use LLM technology on Intel products. See other repos like https://github.com/eleiton/ollama-intel-arc for alternate support.


ollama-intel-gpu

This repo illustrates the use of Ollama with support for Intel ARC GPU based via ipex-llm and Ollama Portable ZIP support. Run the recently released deepseek-r1 model on your local Intel ARC GPU based PC using Linux

Important Note

All Ollama based ipex-llm defects should be reported directly to the ipex-llm project at https://github.com/intel/ipex-llm

Screenshot

screenshot

Prerequisites

  • Ubuntu 24.04 or newer (for Intel ARC GPU kernel driver support. Tested with Ubuntu 24.04.02
  • Installed Docker and Docker-compose tools
  • Intel ARC series GPU (tested with Intel ARC A770 16GB and Intel(R) Core(TM) Ultra 5 125H integrated GPU)

Usage

The following will build the Ollama with Intel ARC GPU support, and compose those with the public docker image based on OpenWEB UI from https://github.com/open-webui/open-webui

Linux:

$ git clone https://github.com/mattcurf/ollama-intel-gpu
$ cd ollama-intel-gpu
$ docker compose up 

Note: On some platforms or versions of docker, the command above may instead be 'docker-compose'.

Note 2: If you have multiple GPU's installed (like integrated and discrete), set the ONEAPI_DEVICE_DELECTOR environment variable in the docker compose file to select the intended device to use.

Then launch your web browser to http://localhost:3000 to launch the web ui. Create a local OpenWeb UI credential, then click the settings icon in the top right of the screen, then select 'Models', then click 'Show', then download a model like 'llama3.1:8b-instruct-q8_0' for Intel ARC A770 16GB VRAM

Update to the latest IPEX-LLM Portable Zip Version

To update to the latest portable zip version of IPEX-LLM's Ollama, update the compose file with the build arguments shown below, using the latest ollama-*.tgz release from https://github.com/ipex-llm/ipex-llm/releases/tag/v2.3.0-nightly , then rebuild the image.

ollama-intel-gpu:
    build:
      context: .
      dockerfile: Dockerfile
      args:
        IPEXLLM_RELEASE_REPO: ipex-llm/ipex-llm
        IPEXLLM_RELEASE_VERSON: v2.3.0-nightly
        IPEXLLM_PORTABLE_ZIP_FILENAME: ollama-ipex-llm-2.3.0b20250415-ubuntu.tgz

References

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Packages

No packages published