ESC
Start typing to search...

Multi-LLM Providers

Teleton supports multiple LLM providers for flexibility and cost optimization.

Supported Providers

ProviderModelsBest For
anthropicClaude 3.5, Claude 3Complex reasoning, tool use
openaiGPT-4o, GPT-4General tasks
groqLlama 3, MixtralFast inference, cost-effective
ollamaAny local modelPrivacy, offline use

Anthropic (Default)

config.yaml
llm:
  provider: anthropic
  model: claude-sonnet-4-20250514
  apiKey: ${ANTHROPIC_API_KEY}
  maxTokens: 4096

OpenAI

config.yaml
llm:
  provider: openai
  model: gpt-4o
  apiKey: ${OPENAI_API_KEY}

Groq

Ultra-fast inference with open-source models:

config.yaml
llm:
  provider: groq
  model: llama-3.1-70b-versatile
  apiKey: ${GROQ_API_KEY}

Ollama (Local)

Run models locally for privacy and offline use:

config.yaml
llm:
  provider: ollama
  model: llama3.1
  baseUrl: http://localhost:11434
Hardware requirements

Local models require sufficient hardware. Recommended: 16GB+ RAM for 7B models, 32GB+ for larger models.

Fallback Configuration

Configure fallback providers for reliability:

config.yaml
llm:
  provider: anthropic
  model: claude-sonnet-4-20250514
  apiKey: ${ANTHROPIC_API_KEY}
  fallback:
    provider: groq
    model: llama-3.1-70b-versatile
    apiKey: ${GROQ_API_KEY}