OpenAgent
Core Concepts

Models

Supported LLM providers and how to configure them in OpenAgent.

Models

OpenAgent integrates with 30+ language model providers through a unified interface. You can mix and match providers across different agents without changing any agent logic.

Supported Providers

Cloud Providers

ProviderModelsNotes
OpenAIGPT-4o, GPT-4 Turbo, GPT-3.5 Turbo, o1, o3DALL-E 3 for image gen
AnthropicClaude 3.5 Sonnet, Claude 3 Opus, Claude 3 HaikuVia Anthropic API
GoogleGemini 2.0 Flash, Gemini 1.5 Pro, Gemini 1.5 FlashVertex AI supported
Azure OpenAIAll OpenAI modelsHosted in your Azure subscription
Amazon BedrockClaude, Llama, TitanAWS region support
DeepSeekDeepSeek-V3, DeepSeek-R1Strong coding & reasoning
MistralMistral Large, Mistral 7B, Mixtral 8x7BOpen weights options
GroqLlama 3.x, Mixtral, GemmaUltra-fast inference
OpenRouter200+ modelsModel aggregation gateway

Chinese Providers

ProviderModels
Alibaba QwenQwen-Max, Qwen-Plus, Qwen-Turbo
Baidu ErnieERNIE 4.0, ERNIE 3.5
Zhipu ChatGLMGLM-4, GLM-4V
BaichuanBaichuan2-Turbo
MoonshotMoonshot-v1-8k/32k/128k
MiniMaxMiniMax-Text-01
StepFunStep-1, Step-1V
HunyuanHunyuan-Pro, Hunyuan-Standard
DoubaoDoubao-Pro

Local / Self-Hosted

ProviderNotes
OllamaRun Llama, Mistral, Phi, Gemma locally
LM StudioOpenAI-compatible local API
Hugging FaceInference API + local models
LiteLLMProxy any model via unified API

Configuring a Provider

Via the Dashboard

  1. Go to Settings → Model Providers
  2. Click Add Provider
  3. Select your provider from the list
  4. Enter the required credentials (API key, base URL, etc.)
  5. Click Test Connection, then Save

Configuration Examples

{
  "provider": "openai",
  "api_key": "sk-...",
  "models": ["gpt-4o", "gpt-4o-mini", "gpt-3.5-turbo"]
}
{
  "provider": "anthropic",
  "api_key": "sk-ant-...",
  "models": ["claude-3-5-sonnet-20241022", "claude-3-haiku-20240307"]
}
{
  "provider": "ollama",
  "base_url": "http://localhost:11434",
  "models": ["llama3.2", "mistral", "phi4", "gemma2"]
}

No API key required. Ollama must be running locally with models already pulled.

{
  "provider": "azure_openai",
  "api_key": "...",
  "endpoint": "https://your-resource.openai.azure.com",
  "api_version": "2024-02-01",
  "deployment_name": "gpt-4o"
}

Embedding Models

OpenAgent uses embedding models separately from chat models. Embeddings power the knowledge base search.

ProviderEmbedding Models
OpenAItext-embedding-3-small, text-embedding-3-large, text-embedding-ada-002
Cohereembed-english-v3.0, embed-multilingual-v3.0
Alibaba Qwentext-embedding-v1, text-embedding-v2
Ollamanomic-embed-text, mxbai-embed-large
Jinajina-embeddings-v3

You can use a different provider for embeddings than for chat. For example, use OpenAI embeddings with a local Ollama chat model.

Model Selection Tips

For general-purpose assistants:

  • gpt-4o or claude-3-5-sonnet — best balance of capability and speed
  • gpt-4o-mini or claude-3-haiku — faster and cheaper for high-volume use

For document Q&A / RAG:

  • Any frontier model works well; larger context windows help
  • gemini-1.5-pro (1M token context) for very large document sets

For code generation and reasoning:

  • deepseek-r1 or o3-mini for complex reasoning tasks
  • claude-3-5-sonnet for code generation

For cost-sensitive or offline deployments:

  • Ollama with llama3.2 or mistral for fully local, free inference
  • gpt-4o-mini for the cheapest capable cloud model

Context Length Management

OpenAgent automatically manages context windows. When conversation history would exceed the model's context limit, it applies a sliding window strategy: the oldest turns are dropped while keeping the system prompt and recent history.

You can configure the context window size per agent to control memory vs. cost tradeoffs.

On this page