Run N LiveKit voice agents in one worker. Pay the model-load cost once.
PyPI package name: openrtc.
Table of Contents
- The problem
- What openrtc does
- Installation
- Quick start: explicit registration with add()
- Quick start: one Python file per agent with discover()
- Memory: before and after
- Routing
- Greetings and session options
- Provider configuration
- CLI and TUI
- Public API at a glance
- Project structure
- Contributing
- License
You already ship three voice agents with livekit-agents. Each agent is its own worker on the same VPS. Every worker process loads the same shared stack: Python runtime, Silero VAD, and the turn-detection model. You are not loading three different models. You are loading the same stack three times because the process boundary forces it. On a 1–2 GB instance, that shows up as duplicate resident set for every idle worker. You pay RAM for copies you do not need.
openrtc gives you one AgentPool in one worker: prewarm runs once, each incoming call still gets its own AgentSession, and you register multiple Agent subclasses on the pool so dispatch can pick one per session from metadata or fallbacks. This package does not replace your agent code. It does not sit between you and livekit.agents.Agent, @function_tool, RunContext, on_enter, on_exit, llm_node, stt_node, or tts_node. You keep your subclasses and tools as they are. You change how many workers you run, not how you write an agent.
OpenRTC requires Python 3.11 or newer. The LiveKit Silero / turn-detector
plugins depend on onnxruntime, which does not ship supported wheels for
Python 3.10 in current releases—use 3.11+ to avoid install failures.
pip install openrtcThe base install pulls in livekit-agents[openai,silero,turn-detector] so shared prewarm has the plugins it expects. The package ships a PEP 561 py.typed marker for downstream type checkers.
With uv (recommended in CONTRIBUTING.md):
uv add openrtc
uv add "openrtc[cli,tui]"pip install 'openrtc[cli]'Optional Textual sidecar for live metrics:
pip install 'openrtc[cli,tui]'Set the same variables you use for any LiveKit worker:
export LIVEKIT_URL=ws://localhost:7880
export LIVEKIT_API_KEY=devkey
export LIVEKIT_API_SECRET=secretFor OpenAI-backed plugins, set OPENAI_API_KEY as you already do.
Use this when you want every agent registered in one place with explicit names and providers.
from livekit.agents import Agent
from livekit.plugins import openai
from openrtc import AgentPool
class RestaurantAgent(Agent):
def __init__(self) -> None:
super().__init__(instructions="You help callers make restaurant bookings.")
class DentalAgent(Agent):
def __init__(self) -> None:
super().__init__(instructions="You help callers manage dental appointments.")
pool = AgentPool()
pool.add(
"restaurant",
RestaurantAgent,
stt=openai.STT(model="gpt-4o-mini-transcribe"),
llm=openai.responses.LLM(model="gpt-4.1-mini"),
tts=openai.TTS(model="gpt-4o-mini-tts"),
greeting="Welcome to reservations.",
)
pool.add(
"dental",
DentalAgent,
stt=openai.STT(model="gpt-4o-mini-transcribe"),
llm=openai.responses.LLM(model="gpt-4.1-mini"),
tts=openai.TTS(model="gpt-4o-mini-tts"),
)
pool.run()Use this when you prefer one module per agent and optional @agent_config(...) on each class.
Create a directory (for example agents/) and add one .py file per agent. Then:
from pathlib import Path
from livekit.plugins import openai
from openrtc import AgentPool
pool = AgentPool(
default_stt=openai.STT(model="gpt-4o-mini-transcribe"),
default_llm=openai.responses.LLM(model="gpt-4.1-mini"),
default_tts=openai.TTS(model="gpt-4o-mini-tts"),
)
pool.discover(Path("./agents"))
pool.run()Example file agents/restaurant.py:
from livekit.agents import Agent
from openrtc import agent_config
@agent_config(name="restaurant", greeting="Welcome to reservations.")
class RestaurantAgent(Agent):
def __init__(self) -> None:
super().__init__(instructions="You help callers make restaurant bookings.")If a module has no @agent_config, the agent name defaults to the filename stem. STT, LLM, TTS, and greeting fall back to the pool defaults.
Discovered agents work with livekit dev and spawn-based workers on macOS. For add(), define agent classes at module scope so worker reload can import them.
Assume an illustrative ~400 MB idle baseline per worker for the shared stack (VAD, turn detector, and similar). Your measured RSS will differ by provider, model, and OS.
| Before openrtc | After openrtc | |
|---|---|---|
| Three workers, same stack | about 3 × 400 MB ≈ 1.2 GB idle baseline (three loads) | — |
| One worker, three registered agents | — | about one × 400 MB idle baseline (one load) plus per-session overhead |
Exact numbers depend on your providers, concurrency, and call patterns. The win is not loading that stack once per agent worker.
One process hosts several agent classes, so each session must resolve to a single registered name. AgentPool resolves the agent in this order:
ctx.job.metadata["agent"]ctx.job.metadata["demo"]ctx.room.metadata["agent"]ctx.room.metadata["demo"]- room name prefix match, such as
restaurant-call-123 - the first registered agent
If metadata names an agent that is not registered, you get a ValueError instead of a silent fallback.
You can pass a greeting and extra AgentSession options per registration.
pool.add(
"restaurant",
RestaurantAgent,
greeting="Welcome to reservations.",
session_kwargs={"turn_handling": {"interruption": {"enabled": False}}},
max_tool_steps=4,
preemptive_generation=True,
)Direct keyword arguments win over the same keys inside session_kwargs.
By default, OpenRTC sets explicit turn_handling with the multilingual turn detector and VAD-based interruption. To opt into adaptive interruption, pass session_kwargs={"turn_handling": {"interruption": {"mode": "adaptive"}}}.
Pass instantiated provider objects through to livekit-agents unchanged, for example:
openai.STT(model="gpt-4o-mini-transcribe")openai.responses.LLM(model="gpt-4.1-mini")openai.TTS(model="gpt-4o-mini-tts")
If you pass strings such as openai/gpt-4.1-mini, OpenRTC leaves them as-is and the LiveKit runtime interprets them for your deployment.
Install openrtc[cli] to get openrtc on your PATH. Subcommands follow the LiveKit Agents CLI shape (dev, start, console, connect, download-files), plus list and tui. For most commands you can pass the agents directory (or, for tui, the metrics JSONL file) as the first path argument instead of --agents-dir / --watch.
List what discovery would register (defaults are string passthroughs for livekit-agents, not constructed provider objects):
openrtc list \
./agents \
--default-stt openai/gpt-4o-mini-transcribe \
--default-llm openai/gpt-4.1-mini \
--default-tts openai/gpt-4o-mini-ttsRun a production worker (after exporting LIVEKIT_*):
openrtc start ./agentsRun a development worker:
openrtc dev ./agentsSame as openrtc dev --agents-dir ./agents. The metrics JSONL file is optional: add a second path only when you want JSONL output (same as --metrics-jsonl), e.g. openrtc dev ./agents ./openrtc-metrics.jsonl for openrtc tui.
Optional visibility: --dashboard prints a Rich summary in the terminal. --metrics-json-file ./runtime.json overwrites a JSON snapshot on each tick. Use that for scripts, dashboards, or CI. For JSON Lines plus a separate terminal UI, use --metrics-jsonl ./openrtc-metrics.jsonl on the worker and openrtc tui in another terminal (it tails ./openrtc-metrics.jsonl by default; override with --watch) after pip install 'openrtc[cli,tui]'.
Stable machine output: openrtc list --json and --plain. Combine --resources when you want footprint hints. OpenRTC-only flags are stripped before the handoff to LiveKit’s CLI parser.
Full flag lists live in docs/cli.md.
Everything openrtc exposes publicly is listed here. Anything else is internal and not treated as stable.
AgentPoolAgentConfigAgentDiscoveryConfigagent_config(...)ProviderValue— type alias for STT/LLM/TTS slot values (provider ID strings or LiveKit plugin instances)
On AgentPool:
add(...)discover(...)list_agents()get(name)remove(name)run()runtime_snapshot()drain_metrics_stream_events()— for JSONL export paths (mainly CLI; rare in app code)server
src/openrtc/
├── __init__.py
├── py.typed
├── cli.py # lazy console entry / missing-extra hints
├── cli_app.py # Typer commands and programmatic main()
├── cli_types.py # shared CLI option aliases
├── cli_dashboard.py # Rich dashboard and list output
├── cli_reporter.py # background metrics reporter thread
├── cli_livekit.py # LiveKit argv/env handoff, pool run
├── cli_params.py # shared worker handoff option bundles
├── metrics_stream.py # JSONL metrics schema
├── provider_types.py # ProviderValue and related typing
├── tui_app.py # optional Textual sidecar
└── pool.py # AgentPool, discovery, routing
pool.py—AgentPool, discovery, routingcli.py/cli_app.py— Typer/Rich CLI (openrtc[cli])metrics_stream.py— JSONL metrics schematui_app.py— optional Textual sidecar (openrtc[tui])
See CONTRIBUTING.md. CI runs Ruff and mypy on pull requests alongside the test suite.
MIT. See LICENSE.