Skip to content

AI Model Providers

“samuele edited this page Mar 15, 2026 · 2 revisions

AI Model Providers

RedAmon supports five AI providers out of the box, giving you access to 400+ language models through a single, unified interface. The model selector in project settings dynamically fetches available models from each configured provider — no hardcoded lists, no manual updates.


Supported Providers

Provider Models
OpenAI (Direct) ~30 chat models — GPT-5.2, GPT-5, GPT-4.1, o3, o4-mini
Anthropic (Direct) ~15 models — Claude Opus 4.6, Sonnet 4.6/4.5, Haiku 4.5
OpenAI-Compatible Any self-hosted or third-party OpenAI-compatible API
OpenRouter 300+ models — Llama 4, Gemini 3, Mistral, Qwen, DeepSeek
AWS Bedrock ~60 foundation models — Claude, Titan, Llama, Cohere, Mistral

All providers are configured exclusively in Global Settings (http://localhost:3000/settings → gear icon in the header). API keys are stored per-user in the database.


How It Works

  1. Provider configuration — Configure providers in Global Settings (http://localhost:3000/settings). All API keys are stored per-user in the database.

LLM Providers

  1. Dynamic model fetching — the agent's /models endpoint fetches available models from all configured providers in parallel. OpenAI-Compatible providers appear as single custom model entries. Results are cached for 1 hour.
  2. Searchable model selector — the project settings UI presents a searchable dropdown grouped by provider. Each model shows its name, context window size, and provider.

Model Selector

  1. OpenAI-Compatible setup — Add any local or third-party OpenAI-compatible endpoint (Ollama, vLLM, Groq, etc.) with presets for common providers.

OpenAI-Compatible Provider Form

  1. Provider prefix convention — models are stored with a provider prefix (custom/, openrouter/, bedrock/) so the agent knows which SDK to use at runtime. OpenAI and Anthropic models are detected by name pattern (no prefix needed).
  2. Test Connection — each provider can be tested before saving using the "Test Connection" button.

Provider Setup

Open Global Settings (http://localhost:3000/settings → gear icon), click "Add Provider", choose the type, enter your credentials, test the connection, and save.

OpenAI (Direct)

Enter your API key from platform.openai.com/api-keys.

Anthropic (Direct)

Enter your API key from console.anthropic.com.

Recommended — Claude Opus 4.6 is the default model and generally provides the best results for autonomous pentesting tasks.

OpenRouter

Enter your API key from openrouter.ai/settings/keys. OpenRouter provides access to 300+ models from 50+ providers through a single API, including many free models for testing.

AWS Bedrock

Enter your AWS Region, Access Key ID, and Secret Access Key. Create an IAM user with bedrock:InvokeModel and bedrock:ListFoundationModels permissions. Foundation models on Bedrock are automatically enabled across all commercial regions — no manual model access activation required.

Recommended region: us-east-1 (N. Virginia) has the widest model availability.

OpenAI-Compatible Provider

Add Provider → OpenAI-Compatible. Choose from presets (Ollama, vLLM, LM Studio, Groq, Together AI, Fireworks, Mistral, Deepinfra) or enter a custom base URL. Each entry configures one model.

Any backend exposing /v1/chat/completions endpoint works. The agent container includes host.docker.internal resolution, so local servers on your host machine are reachable from Docker.


Local Models & Self-Hosted Options

Ollama (Recommended for Local)

The easiest way to run local LLMs:

  1. Install Ollama: ollama.com
  2. Pull a model: ollama pull llama3.1:70b
  3. In Global Settings, add an OpenAI-Compatible provider:

Ollama on the same machine as RedAmon: use base URL http://host.docker.internal:11434/v1

Ollama on a remote server (different machine): use http://192.168.1.50:11434/v1 (replace with your Ollama server's IP or hostname). Ensure port 11434 is reachable from the machine running RedAmon (check firewall rules).

Important — Bind Ollama to 0.0.0.0: By default, Ollama only listens on localhost and will reject connections from Docker containers and remote machines. You must configure it to listen on all interfaces:

# If Ollama is managed by systemd (Linux):
sudo mkdir -p /etc/systemd/system/ollama.service.d
echo -e '[Service]\nEnvironment="OLLAMA_HOST=0.0.0.0"' | sudo tee /etc/systemd/system/ollama.service.d/override.conf
sudo systemctl daemon-reload && sudo systemctl restart ollama

This is required for all Linux setups (local or remote) and for remote access from any OS. macOS and Windows with Docker Desktop handle local container-to-host resolution automatically, but still need OLLAMA_HOST=0.0.0.0 if Ollama must be accessed from a different machine.

Other Self-Hosted Options

Provider Description Example Base URL
vLLM High-performance GPU inference http://host.docker.internal:8000/v1
LM Studio Desktop app with built-in server http://host.docker.internal:1234/v1
LocalAI Open-source OpenAI drop-in, runs on CPU http://host.docker.internal:8080/v1
Jan Desktop app with local server mode http://host.docker.internal:1337/v1
llama.cpp server Lightweight C++ inference http://host.docker.internal:8080/v1
OpenLLM Run open-source LLMs with one command http://host.docker.internal:3000/v1
text-generation-webui Gradio UI with OpenAI-compatible API http://host.docker.internal:5000/v1

Gateway / Proxy

Provider Description
LiteLLM Proxy for 100+ LLMs in OpenAI format — self-hostable via Docker

Cloud Providers with OpenAI-Compatible API

Provider Description
Together AI 200+ open-source models, serverless
Groq Ultra-fast inference for Llama, Mixtral, Gemma
Fireworks AI Fast open-source model hosting
Deepinfra Pay-per-token open-source models
Mistral AI Mistral/Mixtral via OpenAI-compatible endpoint
Perplexity Sonar models via OpenAI-compatible API

Important Notes

  • Global Settings only — All AI provider keys and Tavily API key are configured exclusively in the Global Settings page. They are not read from .env or environment variables.
  • Multiple providers at once — you can configure all five providers simultaneously. The model selector shows all available models from all providers.
  • OpenAI-Compatible — each entry in Global Settings configures a single model+endpoint. For key-based providers (OpenAI, Anthropic, OpenRouter, Bedrock), all models are auto-discovered from a single API key.
  • Switching models — you can change the model per project at any time. Switch between a free Llama model on OpenRouter for testing and Claude Opus on Anthropic for production assessments.

Next Steps

Clone this wiki locally