Skip to main content
Version: 0.3.0

Configuring LLM Providers

Version: 0.1.0

DecisionBox supports six LLM providers. This guide covers setting up each one.

Provider Comparison

ProviderModelsAuthBest For
Claude (Anthropic)Claude Sonnet 4, Opus 4, Haiku 4.5API keyBest quality. Direct access, simple setup.
OpenAIGPT-4o, GPT-4o-miniAPI keyWidely used. Good alternative.
OllamaLlama 3.1, Qwen 2.5, Mistral, any GGUFNone (local)Free, private, no API key needed.
Vertex AIClaude + Gemini (via Google)GCP ADCGCP users. Managed billing, IAM auth.
AWS BedrockClaude + Llama + Mistral (via AWS)AWS credentialsAWS users. Managed billing, IAM auth.
Azure AI FoundryClaude + OpenAI GPT (via Azure)API keyAzure users. Managed billing, Azure RBAC.

Claude (Direct Anthropic API)

The simplest setup and highest quality results.

1. Get an API Key

Sign up at console.anthropic.com and create an API key.

2. Configure in Dashboard

  1. Create a project (or edit existing) → select Claude (Anthropic) as LLM provider
  2. Enter model name: claude-sonnet-4-20250514 (recommended) or claude-opus-4-20250514 (most capable)
  3. Go to Settings → AI Provider → set API Key to your sk-ant-... key

3. Model Options

ModelQualitySpeedCost
claude-opus-4-20250514HighestSlow$15/$75 per million tokens
claude-sonnet-4-20250514HighFast$3/$15 per million tokens
claude-haiku-4-5-20251001GoodFastest$0.80/$4 per million tokens

Recommendation: Start with Sonnet for a balance of quality and cost. Use Opus for complex datasets.

OpenAI

1. Get an API Key

Sign up at platform.openai.com and create an API key.

2. Configure in Dashboard

  1. Select OpenAI as LLM provider
  2. Enter model name: gpt-4o (recommended) or gpt-4o-mini (cheaper)
  3. Go to Settings → AI Provider → set API Key to your sk-... key

Ollama (Local Models)

Run models locally — free, private, no API key needed. Good for testing and development.

1. Install Ollama

# macOS/Linux
curl -fsSL https://ollama.com/install.sh | sh

# Pull a model
ollama pull llama3.1:70b # Large, high quality
ollama pull qwen2.5:32b # Good alternative
ollama pull llama3.1:8b # Small, fast, for testing

2. Configure in Dashboard

  1. Select Ollama as LLM provider
  2. Enter model name: llama3.1:70b
  3. No API key needed

Note: Ollama runs on http://localhost:11434 by default. If running in Docker, use http://host.docker.internal:11434 or the host network.

Quality Considerations

Local models are significantly less capable than Claude or GPT-4o for complex data analysis. They work for:

  • Testing your setup
  • Privacy-sensitive environments
  • Development and prompt iteration

For production discoveries, use Claude or GPT-4o.

Vertex AI (Google Cloud)

Access Claude and Gemini through Google's managed platform. Uses GCP IAM for authentication (no API keys).

1. Prerequisites

  • GCP project with Vertex AI API enabled
  • Claude and/or Gemini models enabled in Model Garden
  • Application Default Credentials configured:
gcloud auth application-default login
# Or use a service account with Vertex AI User role

2. Configure in Dashboard

  1. Select Vertex AI as LLM provider
  2. Enter model name:
    • Claude: claude-sonnet-4-20250514 or claude-haiku-4-5@20251001
    • Gemini: gemini-2.5-pro or gemini-2.5-flash
  3. Set provider-specific config:
    • Project ID: Your GCP project ID
    • Location: Region where the model is enabled (e.g., us-east5 for Claude, us-central1 for Gemini)

3. No API Key Needed

Vertex AI uses GCP Application Default Credentials (ADC). No LLM API key secret is needed.

Model Name Format

  • Claude on Vertex: claude-sonnet-4-20250514 or claude-haiku-4-5@20251001 (with @ for versioned models)
  • Gemini on Vertex: gemini-2.5-pro, gemini-2.5-flash

The provider automatically routes to the correct API format based on model name prefix (claude-* → Anthropic rawPredict, gemini-* → Google generateContent).

AWS Bedrock

Access Claude, Llama, and Mistral through AWS's managed platform. Uses AWS IAM for authentication.

1. Prerequisites

  • AWS account with Bedrock access
  • Claude model access enabled in Bedrock Model Access
  • AWS credentials configured:
aws configure
# Or use IAM role / instance profile

2. Configure in Dashboard

  1. Select AWS Bedrock as LLM provider
  2. Enter model name: us.anthropic.claude-sonnet-4-20250514-v1:0
  3. Set provider-specific config:
    • Region: AWS region (e.g., us-east-1)

3. No API Key Needed

Bedrock uses AWS credentials (IAM role, env vars, or ~/.aws/credentials). No LLM API key secret is needed.

Model Name Format

Bedrock model IDs are different from direct Anthropic IDs:

ModelBedrock Model ID
Claude Sonnet 4us.anthropic.claude-sonnet-4-20250514-v1:0
Claude Opus 4us.anthropic.claude-opus-4-20250514-v1:0
Claude Haiku 4.5us.anthropic.claude-haiku-4-5-20251001-v1:0

The us. prefix is an inference profile ID required for newer models.

Timeout Configuration

The default LLM timeout is 300 seconds (5 minutes). For very large prompts (many previous insights, large schemas), you may need more time:

# In docker-compose or env
LLM_TIMEOUT=600s # 10 minutes

Or set per-project in the dashboard (not yet available — use env var for now).

Azure AI Foundry

Access Claude and OpenAI models through Microsoft Azure's managed AI platform. Billing goes through your Azure subscription via the Microsoft Marketplace.

1. Create a Foundry Resource

  1. Navigate to ai.azure.com
  2. Create a Foundry resource or select an existing one
  3. Deploy a Claude model (e.g., claude-sonnet-4-6) or OpenAI model (e.g., gpt-4o) under Models + endpoints
  4. Copy the endpoint URL and API key from Keys and Endpoint

2. Configure in Dashboard

  1. Select Azure AI Foundry as LLM provider
  2. Enter the Endpoint URL (e.g., https://my-resource.services.ai.azure.com)
  3. Enter the deployment name as Model (e.g., claude-sonnet-4-6 or gpt-4o)
  4. Go to Settings → AI Provider → set API Key to your Azure API key

3. Available Models

ModelDeployment NameType
Claude Opus 4.6claude-opus-4-6Claude
Claude Sonnet 4.6claude-sonnet-4-6Claude
Claude Sonnet 4.5claude-sonnet-4-5Claude
Claude Haiku 4.5claude-haiku-4-5Claude
GPT-4ogpt-4oOpenAI
GPT-4o minigpt-4o-miniOpenAI

The provider automatically routes to the correct API based on the model name:

  • claude-* models → Anthropic Messages API
  • All other models → OpenAI Chat Completions API

4. Authentication

Azure AI Foundry supports API key authentication. The API key is set per-project via the dashboard's AI Provider settings tab.

For production on AKS, you can also use Entra ID (Azure AD) with managed identity, but this requires custom configuration outside DecisionBox.

Next Steps