-
Notifications
You must be signed in to change notification settings - Fork 328
AI Model Providers
RedAmon supports five AI providers out of the box, giving you access to 400+ language models through a single, unified interface. The model selector in project settings dynamically fetches available models from each configured provider — no hardcoded lists, no manual updates.
| Provider | Models |
|---|---|
| OpenAI (Direct) | ~30 chat models — GPT-5.2, GPT-5, GPT-4.1, o3, o4-mini |
| Anthropic (Direct) | ~15 models — Claude Opus 4.6, Sonnet 4.6/4.5, Haiku 4.5 |
| OpenAI-Compatible | Any self-hosted or third-party OpenAI-compatible API |
| OpenRouter | 300+ models — Llama 4, Gemini 3, Mistral, Qwen, DeepSeek |
| AWS Bedrock | ~60 foundation models — Claude, Titan, Llama, Cohere, Mistral |
All providers are configured exclusively in Global Settings (http://localhost:3000/settings → gear icon in the header). API keys are stored per-user in the database.
- Provider configuration — Configure providers in Global Settings (http://localhost:3000/settings). All API keys are stored per-user in the database.

-
Dynamic model fetching — the agent's
/modelsendpoint fetches available models from all configured providers in parallel. OpenAI-Compatible providers appear as single custom model entries. Results are cached for 1 hour. - Searchable model selector — the project settings UI presents a searchable dropdown grouped by provider. Each model shows its name, context window size, and provider.

- OpenAI-Compatible setup — Add any local or third-party OpenAI-compatible endpoint (Ollama, vLLM, Groq, etc.) with presets for common providers.

-
Provider prefix convention — models are stored with a provider prefix (
custom/,openrouter/,bedrock/) so the agent knows which SDK to use at runtime. OpenAI and Anthropic models are detected by name pattern (no prefix needed). - Test Connection — each provider can be tested before saving using the "Test Connection" button.
Open Global Settings (http://localhost:3000/settings → gear icon), click "Add Provider", choose the type, enter your credentials, test the connection, and save.
Enter your API key from platform.openai.com/api-keys.
Enter your API key from console.anthropic.com.
Recommended — Claude Opus 4.6 is the default model and generally provides the best results for autonomous pentesting tasks.
Enter your API key from openrouter.ai/settings/keys. OpenRouter provides access to 300+ models from 50+ providers through a single API, including many free models for testing.
Enter your AWS Region, Access Key ID, and Secret Access Key. Create an IAM user with bedrock:InvokeModel and bedrock:ListFoundationModels permissions. Foundation models on Bedrock are automatically enabled across all commercial regions — no manual model access activation required.
Recommended region:
us-east-1(N. Virginia) has the widest model availability.
Add Provider → OpenAI-Compatible. Choose from presets (Ollama, vLLM, LM Studio, Groq, Together AI, Fireworks, Mistral, Deepinfra) or enter a custom base URL. Each entry configures one model.
Any backend exposing /v1/chat/completions endpoint works. The agent container includes host.docker.internal resolution, so local servers on your host machine are reachable from Docker.
The easiest way to run local LLMs:
- Install Ollama: ollama.com
- Pull a model:
ollama pull llama3.1:70b - In Global Settings, add an OpenAI-Compatible provider:
Ollama on the same machine as RedAmon: use base URL http://host.docker.internal:11434/v1
Ollama on a remote server (different machine): use http://192.168.1.50:11434/v1 (replace with your Ollama server's IP or hostname). Ensure port 11434 is reachable from the machine running RedAmon (check firewall rules).
Important — Bind Ollama to
0.0.0.0: By default, Ollama only listens onlocalhostand will reject connections from Docker containers and remote machines. You must configure it to listen on all interfaces:# If Ollama is managed by systemd (Linux): sudo mkdir -p /etc/systemd/system/ollama.service.d echo -e '[Service]\nEnvironment="OLLAMA_HOST=0.0.0.0"' | sudo tee /etc/systemd/system/ollama.service.d/override.conf sudo systemctl daemon-reload && sudo systemctl restart ollamaThis is required for all Linux setups (local or remote) and for remote access from any OS. macOS and Windows with Docker Desktop handle local container-to-host resolution automatically, but still need
OLLAMA_HOST=0.0.0.0if Ollama must be accessed from a different machine.
| Provider | Description | Example Base URL |
|---|---|---|
| vLLM | High-performance GPU inference | http://host.docker.internal:8000/v1 |
| LM Studio | Desktop app with built-in server | http://host.docker.internal:1234/v1 |
| LocalAI | Open-source OpenAI drop-in, runs on CPU | http://host.docker.internal:8080/v1 |
| Jan | Desktop app with local server mode | http://host.docker.internal:1337/v1 |
| llama.cpp server | Lightweight C++ inference | http://host.docker.internal:8080/v1 |
| OpenLLM | Run open-source LLMs with one command | http://host.docker.internal:3000/v1 |
| text-generation-webui | Gradio UI with OpenAI-compatible API | http://host.docker.internal:5000/v1 |
| Provider | Description |
|---|---|
| LiteLLM | Proxy for 100+ LLMs in OpenAI format — self-hostable via Docker |
| Provider | Description |
|---|---|
| Together AI | 200+ open-source models, serverless |
| Groq | Ultra-fast inference for Llama, Mixtral, Gemma |
| Fireworks AI | Fast open-source model hosting |
| Deepinfra | Pay-per-token open-source models |
| Mistral AI | Mistral/Mixtral via OpenAI-compatible endpoint |
| Perplexity | Sonar models via OpenAI-compatible API |
-
Global Settings only — All AI provider keys and Tavily API key are configured exclusively in the Global Settings page. They are not read from
.envor environment variables. - Multiple providers at once — you can configure all five providers simultaneously. The model selector shows all available models from all providers.
- OpenAI-Compatible — each entry in Global Settings configures a single model+endpoint. For key-based providers (OpenAI, Anthropic, OpenRouter, Bedrock), all models are auto-discovered from a single API key.
- Switching models — you can change the model per project at any time. Switch between a free Llama model on OpenRouter for testing and Claude Opus on Anthropic for production assessments.
- Project Settings Reference > Agent Behavior — configure agent parameters
- AI Agent Guide — learn how to use the agent
Getting Started
Core Workflow
Scanning & OSINT
AI & Automation
Analysis & Reporting
- Insights Dashboard
- Pentest Reports
- Attack Surface Graph
- EvoGraph — Attack Chain Evolution
- Data Export & Import
Reference & Help