End-to-end AI evals orchestration platform for comparing LLM outputs across providers with transcription, structured logging, human review, and Supabase-backed decision tracking.
-
Updated
Mar 10, 2026 - TypeScript
End-to-end AI evals orchestration platform for comparing LLM outputs across providers with transcription, structured logging, human review, and Supabase-backed decision tracking.
Production-grade Safe GenAI Agent Orchestrator with intent routing, hallucination guard, tool orchestration, evaluation pipeline, and multi-screen Next.js dashboard.
Add a description, image, and links to the evaluation-pipeline topic page so that developers can more easily learn about it.
To associate your repository with the evaluation-pipeline topic, visit your repo's landing page and select "manage topics."