Supported LLM Providers & Models
Dexto supports multiple LLM providers out-of-the-box, plus the ability to use any OpenAI SDK-compatible provider. This guide lists all supported providers and their available models.
For complete LLM configuration options and YAML reference, see the agent.yml → LLM Configuration section.
Built-in Providers
OpenAI
llm:
provider: openai
model: gpt-5-mini
apiKey: $OPENAI_API_KEY
Supported models:
gpt-5,gpt-5-mini,gpt-5-nanogpt-4.1,gpt-4.1-mini,gpt-4.1-nanogpt-4o,gpt-4o-minio4-mini,o3,o3-mini,o1
Features: Function calling, streaming, vision (GPT-4o), JSON mode
Anthropic (Claude)
llm:
provider: anthropic
model: claude-sonnet-4-5-20250929
apiKey: $ANTHROPIC_API_KEY
Supported models:
claude-haiku-4-5-20251001claude-sonnet-4-5-20250929claude-4-opus-20250805,claude-4-opus-20250514claude-4-sonnet-20250514claude-3-7-sonnet-20250219claude-3-5-sonnet-20240620claude-3-5-haiku-20241022
Features: Large context (200K tokens), advanced tool use, Constitutional AI
Google Gemini
llm:
provider: google
model: gemini-2.5-pro
apiKey: $GOOGLE_GENERATIVE_AI_API_KEY
Supported models:
gemini-2.5-pro(default)gemini-2.5-flashgemini-2.5-flash-litegemini-2.0-flashgemini-2.0-flash-lite
Features: Multimodal (text/image/video/audio), large context (1M tokens), fast inference
xAI (Grok)
llm:
provider: xai
model: grok-4
apiKey: $XAI_API_KEY
Supported models:
grok-4(default)grok-3grok-3-minigrok-code-fast-1
Features: State-of-the-art reasoning, real-time knowledge, strong benchmark performance
Groq
llm:
provider: groq
model: llama-3.3-70b-versatile
apiKey: $GROQ_API_KEY
Supported models:
llama-3.3-70b-versatile(default)gemma-2-9b-itopenai/gpt-oss-20bopenai/gpt-oss-120bmoonshotai/kimi-k2-instructmeta-llama/llama-4-scout-17b-16e-instructmeta-llama/llama-4-maverick-17b-128e-instructdeepseek-r1-distill-llama-70bqwen/qwen3-32b
Features: Ultra-fast inference, cost-effective, open source models
Cohere
llm:
provider: cohere
model: command-a-03-2025
apiKey: $COHERE_API_KEY
Supported models:
command-a-03-2025(default, 256k context window)command-r-pluscommand-rcommand-r7b
Features: RAG optimization, tool use, multilingual, conversational AI
OpenAI-Compatible Providers
Use any provider that implements the OpenAI SDK interface:
llm:
provider: openai-compatible
model: your-custom-model
apiKey: $YOUR_API_KEY
baseURL: https://api.your-provider.com/v1
maxInputTokens: 100000
Local Models
Run models locally using Ollama, LM Studio, or similar:
llm:
provider: openai-compatible
model: gemma3n:e2b
apiKey: dummy
baseURL: http://localhost:11434/v1
maxInputTokens: 8000
Popular options:
- Ollama - Easy local model hosting
- LM Studio - User-friendly interface
- vLLM - High-performance serving
- TGI - Hugging Face serving
Azure OpenAI
llm:
provider: openai-compatible
model: gpt-5
apiKey: $AZURE_OPENAI_API_KEY
baseURL: https://your-resource.openai.azure.com/openai/deployments/gpt-5
maxInputTokens: 128000
Notes: Replace your-resource with your Azure resource name. Supports all OpenAI models available in Azure.
OpenRouter
Access 100+ models through one API:
llm:
provider: openai-compatible
model: anthropic/claude-sonnet-4-5-20250929
apiKey: $OPENROUTER_API_KEY
baseURL: https://openrouter.ai/api/v1
maxInputTokens: 200000
Popular models:
anthropic/claude-sonnet-4-5-20250929meta-llama/llama-3.1-405b-instructgoogle/gemini-pro-1.5mistralai/mistral-large
Together.ai
llm:
provider: openai-compatible
model: meta-llama/Llama-3.2-90B-Vision-Instruct-Turbo
apiKey: $TOGETHER_API_KEY
baseURL: https://api.together.xyz/v1
maxInputTokens: 8000
Perplexity
llm:
provider: openai-compatible
model: llama-3.1-sonar-huge-128k-online
apiKey: $PERPLEXITY_API_KEY
baseURL: https://api.perplexity.ai
maxInputTokens: 128000
Special feature: Online models with real-time web search
Choosing the Right Provider
For Development
- OpenAI - Best developer experience and documentation
- Local models - Free, private, great for experimentation
For Production
- OpenAI - Reliable, extensive model selection
- Anthropic - Safety-critical applications
- Google - Multimodal and large context needs
For Cost Optimization
- Groq - Fastest and often cheapest
- OpenRouter - Compare prices across providers
- Local hosting - No per-token costs
For Privacy
- Local models - Complete data privacy
- Azure OpenAI - Enterprise security and compliance
Environment Variables
Set API keys in your ~/.dexto/.env file:
# Built-in providers
OPENAI_API_KEY=your_openai_key
ANTHROPIC_API_KEY=your_anthropic_key
GOOGLE_GENERATIVE_AI_API_KEY=your_google_key
GROQ_API_KEY=your_groq_key
XAI_API_KEY=your_xai_key
COHERE_API_KEY=your_cohere_key
# Custom providers
OPENROUTER_API_KEY=your_openrouter_key
TOGETHER_API_KEY=your_together_key
AZURE_OPENAI_API_KEY=your_azure_key
PERPLEXITY_API_KEY=your_perplexity_key