Skip to main content

LLM Providers

The LLM (Large Language Model) provider powers the conversational intelligence of your avatar.

Supported Providers

ProviderModelsContext WindowBest For
Google Geminigemini-3.0-flash, gemini-2.5-pro1M tokensRAG, long documents
OpenAIgpt-4o, gpt-4o-mini128K tokensGeneral purpose
Anthropicclaude-sonnet-4, claude-opus-4200K tokensComplex reasoning

Feature Comparison

FeatureGeminiOpenAIAnthropic
Streaming
Function calling
Vision/Images
Native RAG✅ File Search
Max context1M tokens128K tokens200K tokens

Pricing Comparison

Cost per 1 million tokens (approximate):

ProviderModelInputOutput
Gemini3.0-flash$0.50$3.00
Gemini2.5-pro$1.25$5.00
OpenAIgpt-4o$2.50$10.00
OpenAIgpt-4o-mini$0.15$0.60
Anthropicclaude-sonnet-4$3.00$15.00
Anthropicclaude-haiku-3.5$0.80$4.00

Per-Session Cost Estimate

Based on ~500 input tokens + ~300 output tokens per interaction, 4 interactions per 3-minute session:

Provider/ModelCost per Session
Gemini Flash~$0.01
GPT-4o Mini~$0.01
GPT-4o~$0.05
Claude Sonnet~$0.08

Choosing a Provider

Choose Gemini if:

  • You want the best RAG support (native File Search)
  • You need to process long documents (1M context)
  • Cost efficiency is important
  • You're starting fresh without existing infrastructure

Choose OpenAI if:

  • You already have OpenAI infrastructure
  • You need the GPT-4o ecosystem
  • You want broad model selection

Choose Anthropic if:

  • You need Claude's reasoning capabilities
  • You have complex, nuanced conversations
  • You prefer Anthropic's approach to safety

Configuration

See the individual provider pages for detailed setup: