Supported Providers
Tambo works with the following LLM providers:OpenAI
Supports GPT-4, GPT-4 Turbo, GPT-3.5 Turbo, and other OpenAI models.gpt-4o- Latest GPT-4 Omni modelgpt-4-turbo- Fast GPT-4 with 128k contextgpt-4- Standard GPT-4gpt-3.5-turbo- Fast and cost-effective
Anthropic
Supports Claude 3 family models including Opus, Sonnet, and Haiku.claude-3-opus- Most capable Claude modelclaude-3-sonnet- Balanced performance and speedclaude-3-haiku- Fast and cost-effectiveclaude-3-5-sonnet- Latest Sonnet with enhanced capabilities
Google Gemini
Supports Gemini Pro and other Google AI models.gemini-pro- Google’s most capable modelgemini-pro-vision- Multimodal with image supportgemini-1.5-pro- Extended context window
Mistral AI
Supports Mistral’s open and commercial models.mistral-large- Most capable Mistral modelmistral-medium- Balanced modelmistral-small- Fast and efficientmixtral-8x7b- Open-source mixture of experts
Cerebras
Ultra-fast inference with Cerebras hardware acceleration.cerebras-gpt- Fast inference on specialized hardware
OpenAI-Compatible Providers
Tambo supports any provider with an OpenAI-compatible API, including:- Together AI - Fast inference and fine-tuning
- Anyscale - Ray-based LLM serving
- Replicate - Cloud-based model hosting
- Local models via Ollama, LM Studio, etc.
Configuration
Tambo Cloud
When using Tambo Cloud, configure providers in your project settings:- Navigate to your project in the Tambo dashboard
- Go to Settings → LLM Providers
- Add your API keys for each provider
- Select default model for your project
Self-Hosted
When self-hosting, set environment variables in your.env file:
Model Selection
Per-Request Configuration
Specify model per request using the API:Default Model
Set a default model in your project configuration or environment:Best Practices
Model Selection
- GPT-4o / Claude 3.5 Sonnet - Best for complex reasoning and component generation
- GPT-3.5 Turbo / Claude Haiku - Fast and cost-effective for simple tasks
- Gemini Pro - Great for long context and multimodal tasks
- Mistral - Good balance of performance and cost
- Cerebras - Ultra-fast inference when speed is critical
API Key Management
- Store keys securely in environment variables
- Never commit keys to version control
- Use different keys for development and production
- Rotate keys regularly
- Monitor usage and set spending limits
Error Handling
Rate Limits
Each provider has different rate limits:- OpenAI: Varies by tier (free, pay-as-you-go, enterprise)
- Anthropic: Based on API tier and usage
- Google: Requests per minute limit
- Mistral: Based on subscription plan
- Cerebras: Contact for limits