This starter is a rewrite of my Semantic Kernel practice project.
For more details on my journey to make my first AI project and tips see:
My LinkedIn article: "Multi agent Hangman game with Semantic Kernel Practice project"
This example uses a portable Ollama build with IPEX-LLM integration for fast, stable local inference on Intel GPUs (Arc / Xe Graphics).
Why this build?
- Standard Ollama can lag (> 60s responses on 32B models) or reboot frequently without full optimization.
- IPEX-LLM version: Smooth, high-token-rate runs on Intel hardware — no NVIDIA needed!
Tested on: Windows with Intel Arc iGPU + latest drivers.
Quick Setup Steps (Portable — No Full Install Needed):
-
Update Intel Drivers (essential for stability!):
Download latest from intel.com/support (search your GPU model). -
Download IPEX-LLM Ollama Portable Zip:
Get the latest Windows version from Intel's guide:
GitHub Quickstart - Ollama Portable Zip
(Your ollama-ipex-llm-2.2.0-win is solid; upgrade if needed for newer features.) -
Unzip & Start:
- Extract the zip to a folder.
- Run
start-ollama.bat(or equivalent in latest zip).
-
Pull a Model:
In a command prompt:
ollama pull qwen2.5:32b-instruct-q5_K_M
(Quantized GGUF models recommended for max speed.) -
Run the Server:
Ollama serves automatically on startup — accessible athttp://localhost:11434.
NVIDIA/Standard Ollama users: Use official Ollama download + CUDA for equivalent performance.
Feedback welcome if issues on your hardware!
This repo is a living learning project: I'm gradually evolving a multi-agent Hangman game to explore modern patterns in Microsoft Agent Framework (the successor direction after AutoGen/Semantic Kernel agent chat experiments).
- Basic multi-agent setup with Ollama IPEX-LLM for local Intel GPU inference
- Game logic via plugin/tools
- Manual chat loop
- Replace manual chat loop by GroupChat
- Add Coordinator agent who does game start and turn selection
- Dynamic guesser count (2–4 players randomized per game)
- Let Coordinator start new game with random players after game ends
I'm adding one meaningful improvement roughly every 1–2 weeks when time allows.
Feel free to ⭐ watch the repo if you're interested in following along, or open an issue/discussion with suggestions/questions — feedback on Intel GPU setups or prompt ideas is especially welcome!
Previous versions will get tags (e.g. v0.1-initial, v0.2-groupchat, etc.) so you can checkout exactly how it looked at each stage.