RTSPanda v0.0.8
RTSPanda v0.0.8 Release Notes
Headline
RTSPanda v0.0.8 turns the project into a more serious edge-video platform: the default single-machine workflow stays intact, while Raspberry Pi deployments gain a cleaner first-run path, deterministic ONNX-only AI builds, and an easy upgrade path to distributed inference.
Highlights
- Added a Pi-first lightweight deployment mode for
rtspandawithout local AI-worker overhead. - Added cluster mode so a Pi can ingest streams locally and send frames to a remote AI worker on a second machine.
- Removed AI-worker export fallback from Docker builds. The worker now uses prebuilt ONNX models only.
- Added additive Compose profiles for
full,pi, andai-workerwithout breaking the existingdocker compose up --build -dflow. - Rewrote setup and deployment docs to clearly support Standard, Pi Standalone, and Pi + AI topologies.
- Fixed Pi standalone startup so it no longer traverses the local AI-worker build path.
- Repaired the default remote ONNX pin to a live Ultralytics asset and improved failure messages for invalid model sources.
What Changed
Platform and Deployment
docker-compose.ymlnow supports:- standard full-stack deployment
- lightweight Pi deployment
- standalone remote AI-worker deployment
- New
docker-compose.standalone.ymloverlay isolates Pi-only and AI-worker-only deployments from the unprofiled full stack. scripts/pi-up.shnow supports:PI_DEPLOYMENT_MODE=piPI_DEPLOYMENT_MODE=fullPI_DEPLOYMENT_MODE=ai-worker
- Standalone launch paths now use
buildfollowed byup --no-buildso targeted deployments only build the intended service. scripts/pi-preflight.shnow checks deployment mode and model-source expectations more accurately for Docker-first Pi workflows.
AI Runtime
ai_worker/Dockerfilenow resolves models deterministically:- local prebuilt ONNX file first
- explicit
YOLO_MODEL_URLsecond - named Ultralytics ONNX asset fallback last
- Default remote model pin now targets
yolo11non Ultralyticsv8.3.0, which is published and buildable today. - No PyTorch install path
- No
YOLO(...).export(...) - No runtime model conversion on ARM
Backend Detection Routing
- Added additive AI-target resolution using:
AI_MODE=local|remoteAI_WORKER_URL=http://<host>:8090DETECTOR_URLas the highest-precedence override
- Detection health now reports AI mode and remote worker context.
Release Quality Improvements
- Refreshed the frontend lockfile so current tests/tooling install cleanly.
- Fixed clean-checkout frontend embed compilation behavior in the backend.
Setup Paths
Standard
docker compose up --build -dPi Standalone
./scripts/pi-up.shPi + AI
AI host:
docker compose -f docker-compose.yml -f docker-compose.standalone.yml --profile ai-worker build ai-worker-standalone
docker compose -f docker-compose.yml -f docker-compose.standalone.yml --profile ai-worker up -d --no-build ai-worker-standalonePi host:
export AI_WORKER_URL="http://192.168.1.50:8090"
./scripts/pi-up.shValidation Checklist
docker compose config -qcd backend && go test ./internal/...cd frontend && npm run test -- --config vitest.config.tscd ai_worker && python -m pytest -q
Upgrade Notes
- Existing standard users can continue using
docker compose up --build -d. - Pi users who want the old all-in-one behavior can use
PI_DEPLOYMENT_MODE=full ./scripts/pi-up.sh. - For custom ONNX assets, place
model.onnxat repo root orai_worker/model/model.onnx, then setMODEL_SOURCE=local.