A Gradio-based web application that lets you query and compare responses from multiple LLM providers simultaneously, all in a single interface.
Designed for side-by-side evaluation, qualitative analysis, and prompt experimentation across models.
- Multi-LLM Comparison: Query OpenAI, Anthropic, Google, Groq, and Perplexity in parallel.
- Chairman Mode: Enable models to critique, discuss, and respond to each other’s outputs.
- Real-Time Streaming: View responses as they arrive.
- Secure by Default: API keys are stored in-memory only.
- Developer-Friendly: Ideal for prompt testing and model evaluation workflows.
- UI: Gradio
- Backend: Python
- LLM Providers:
- OpenAI
- Anthropic
- Groq
- Perplexity
Install the required dependencies:
pip install -r requirements.txtRun the application locally:
python app.pyOnce started, Gradio will provide a local URL:
Running on local URL: http://127.0.0.1:7860
Open the link in your browser to access the app.
You can provide API keys either via the UI or by creating a .env file.
OPENAI_API_KEY=sk-...
ANTHROPIC_API_KEY=sk-ant-...
GOOGLE_API_KEY=AIza...
GROQ_API_KEY=gsk_...
PERPLEXITY_API_KEY=pplx-...
- API keys entered via the UI are stored in-memory only and are cleared when the app restarts.
- For persistent configuration, use environment variables instead.
This project is experimental and under active development.
It is inspired by Andrej Karpathy’s llm-council
Expect rough edges and ongoing improvements.
The code in this repository is licensed under the MIT License.
