Skip to content

Releases: Daemoniorum-LLC/infernum-framework

Infernum v0.1.0

03 Dec 20:51

Choose a tag to compare

Initial open-source release of Infernum LLM inference framework.

Features

  • Blazingly fast local LLM inference
  • OpenAI API compatibility (drop-in replacement)
  • Streaming support with real-time token output
  • Multi-backend: CPU, CUDA (NVIDIA), Metal (Apple Silicon)
  • Interactive chat with history and session management
  • Model caching via HuggingFace Hub

Quick Start

cargo install --path crates/infernum
infernum config set-model TinyLlama/TinyLlama-1.1B-Chat-v1.0
infernum chat

License

Dual-licensed under MIT and Apache 2.0

Copyright (c) 2024-2025 Daemoniorum, LLC