Skip to main content
Version: Next

Discovery Lifecycle

Version: 0.3.0

A discovery run is the core operation of DecisionBox. The agent autonomously explores your data warehouse, finds patterns, validates them, and generates recommendations. This page explains each phase in detail.

Project state machine

Every project carries a lifecycle state. Most projects spend their entire life in ready; only projects created via the "Generate one for me" wizard cycle through the pack-generation states.

StateMeaningDiscovery allowed?
pack_generation_pendingDraft project; user is filling the wizard (sources, warehouse, providers) before launching pack-gen.No
pack_generationAgent is running --mode=pack-gen to synthesize the domain pack.No
pack_generation_donePack is generated; awaiting the user's "Start discovery" gate.No
readyNormal runtime state.Yes

A project with no state field set is treated as ready (legacy behavior for projects created before pack generation existed).

Transitions:

  • New project → ready (default for "use a built-in pack" flow).
  • New project with generate_pack.enabled=truepack_generation_pending.
  • pack_generation_pendingpack_generation when the user posts to /api/v1/projects/{id}/pack-generate.
  • pack_generationpack_generation_done when the agent finishes successfully.
  • pack_generation_doneready when the user accepts the draft pack (PUT /api/v1/projects/{id} with state: "ready").

Phases Overview

PhaseWhat happensDuration
1. InitializationLoad project config, secrets, providers~2s
2. Context LoadingFetch previous discoveries + feedback~1s
3. Schema DiscoveryList tables, read schemas5-30s
4. ExplorationAI writes + executes SQL queries2-30 min
5. AnalysisGenerate insights per analysis area1-5 min
6. ValidationVerify insights against warehouse30s-2 min
7. RecommendationsGenerate actionable advice from insights1-3 min
8. SavingWrite results to MongoDB~1s

Total time depends on exploration steps, LLM speed, and warehouse query time. A typical 100-step run with Claude Sonnet takes 5-15 minutes.

Phase 1: Initialization

The agent starts with a project ID and loads everything it needs. The entry point is agentserver.Run(), which parses CLI flags, initializes providers, and runs the discovery pipeline. Custom builds can import agentserver and register plugins (e.g., warehouse middleware) via init() blank imports before calling Run().

Agent receives: --project-id=abc123 --run-id=run456 --max-steps=100

Loads project from MongoDB (name, domain, category, warehouse, llm, profile)

Sets project ID in context (warehouse.WithProjectID) for middleware

Initializes secret provider (reads LLM API key, warehouse credentials)

Initializes warehouse provider (BigQuery/Redshift with credentials)

Applies warehouse middleware (warehouse.ApplyMiddleware — e.g., governance)

Initializes LLM provider (Claude/OpenAI/etc. with API key from secrets)

Loads domain pack (e.g., gaming/match3 or social/content_sharing → analysis areas, prompts, profile schema)

Loads project-level prompt overrides from MongoDB (if any)

Secret loading order:

  1. Read llm-api-key from secret provider (per-project)
  2. Read warehouse-credentials from secret provider (optional, for cross-cloud)
  3. These credentials are passed to the LLM/warehouse provider constructors

Phase 2: Context Loading

The agent loads context from previous runs to avoid repetition:

Fetch last 5 discoveries for this project

Fetch all feedback (likes/dislikes with comments)

Build previous context:
- Previously found insights (names, areas, severity, dates)
- Disliked insights with user comments → "AVOID similar conclusions"
- Liked insights → "MONITOR for changes"
- Previous recommendations → "Don't repeat unless changed"

This context is injected into all prompts via the {{PREVIOUS_CONTEXT}} template variable in base_context.md.

Phase 3: Schema Discovery

The agent reads your warehouse structure:

For each dataset in project.warehouse.datasets:

List all tables (excluding system tables like pg_*, stl_*, svv_*)

For each table:
Get column names, types, nullable flags
Get approximate row count

Cache schemas for the exploration phase

A compact Level-0 catalog (one line per table: name, column count, row count, keyword hints, joins) is injected into the exploration prompt via {{SCHEMA_INFO}}. The agent fetches per-table column lists and sample rows on demand during exploration via lookup_schema (up to 10 tables per call, 30 calls per run) or search_tables (semantic query against the per-project Qdrant index, 30 calls per run). See On-Demand Schema for the rationale (the previous "always inject L1 detail" approach exhausted the Bedrock 1M-token context on long runs).

Phase 4: Exploration

The core phase. The AI writes SQL queries, executes them, analyzes results, and decides what to query next.

Send to LLM:
- System prompt (exploration.md + category context)
- Schema information
- Profile (game mechanics, monetization model, etc.)
- Previous context (past insights, feedback)
- Filter rules (WHERE clause for multi-tenant)

LLM responds with JSON:
{
"thinking": "I want to check retention rates...",
"query": "SELECT cohort_date, retention_d1 FROM ..."
}

Agent executes query against warehouse

Send results back to LLM

LLM writes next query based on results

Repeat for max_steps (default: 100)

Each step is written to the discovery_runs collection in real-time, so the dashboard can show live progress.

Self-healing SQL: If a query fails, the agent sends the error message to the LLM and asks it to fix the SQL. This uses the warehouse provider's SQLFixPrompt() (BigQuery-specific SQL fix instructions, Redshift-specific, etc.).

Strict action parsing: Exploration continues only when the LLM returns a JSON object containing a recognized action key — query (run a query), done (terminate), or the legacy action field. Responses that are pure prose, malformed JSON, or JSON with unrelated keys (e.g., a reasoning-model preamble like {"plan": "...", "thinking": "..."}) are rejected and the agent re-prompts the LLM with a reformat nudge (up to 3 retries per step). This prevents silent early termination on reasoning models whose prose often mentions words like "done" or "complete".

Early-termination guard (--min-steps): Models biased toward early completion (Qwen3, DeepSeek-R1, GPT-OSS) sometimes return {"done": true} after just 1–2 steps even with a 100-step budget. Setting --min-steps=N rejects done signals until step N, records a complete_rejected exploration step, and injects a nudge telling the model how many steps remain. Default is 0 (no floor) for backwards compatibility.

Step types reported to the dashboard:

  • query — SQL query executed (with thinking, SQL, row count, timing)
  • lookup_schema — Agent fetched L1 detail (columns + sample rows) for one or more tables from the cache (no warehouse traffic)
  • search_tables — Agent ran a semantic search against the per-project Qdrant index for tables not surfaced by the catalog
  • complete_rejected — LLM signalled done before --min-steps; rejected and exploration continued
  • insight — The AI identified a pattern (name, severity)
  • analysis — Analysis phase started for an area
  • validation — Insight validation result
  • error — Something went wrong (with error message)

Phase 5: Analysis

For each analysis area defined by the domain pack (e.g., churn, engagement, monetization for gaming; growth, engagement, retention for social), the agent:

Load area-specific prompt (e.g., analysis_churn.md)

Pick relevant exploration steps:
- vector-rank against the per-run Qdrant index of every step
(embedding text = step purpose + SQL)
- promote any step whose text contains a verbatim area keyword
- drop the lowest-scored steps until the rendered prompt fits
the per-area token budget

Render each picked step as a compact digest (per-column
statistics + head/tail rows + small-result inlining), instead
of inlining the raw row blob

Prepend base context (profile + previous context)

Substitute template variables:
{{DATASET}} → dataset names
{{TOTAL_QUERIES}} → number of picked steps
{{QUERY_RESULTS}} → JSON array of compact digests for this area

Send to LLM

LLM responds with JSON:
{
"insights": [
{
"name": "Day 0-to-Day 1 Drop: 67% Never Return",
"description": "...",
"severity": "critical",
"affected_count": 8298,
"risk_score": 0.67,
"confidence": 0.85,
"indicators": ["...", "..."],
"source_steps": [1, 3, 5]
}
]
}

Agent parses insights, assigns IDs (e.g., "churn-1", "churn-2")

Insight IDs: The agent generates deterministic IDs in the format {area}-{index} (e.g., churn-1, monetization-3). These IDs are used by recommendations to reference which insights they address.

If analysis fails (e.g., LLM timeout), the error is recorded in analysis_log and the area is skipped. If ALL areas fail, the run is marked as run_type: "failed". If some fail, it's run_type: "partial". The errors are surfaced in the dashboard as a red banner.

Phase 6: Validation

Each insight with an affected_count is verified:

For each insight with affected_count > 0:

Generate a verification SQL query
Grounded on the SQL of the exploration steps the analyst LLM cited as
`source_steps` — those queries already executed successfully against the
warehouse, so their column references are authoritative. The verifier is
instructed to adapt one of them into a SELECT COUNT(...) AS count rather
than invent column names from the table catalog. The compact catalog
(one line per table) is supplied as supplementary context for tables the
source steps didn't touch.

Execute against warehouse (self-healing retry on dialect errors)

Compare claimed count vs verified count

Status:
- "confirmed" — within 20% tolerance
- "adjusted" — count differs, validator records the verified value
- "rejected" — verification query returned 0
- "error" — verification query failed even after retries

Validation results are stored on each insight and shown in the dashboard.

The validator does NOT overwrite the original affected_count — only the Validation block (verified_count, original_count, status, query, reasoning) is written. The dashboard surfaces the verification badge alongside the analysis-claimed count.

Phase 7: Recommendations

All validated insights are fed to the recommendations prompt:

Load recommendations.md prompt

Prepend base context (profile + previous context)

Substitute:
{{DISCOVERY_DATE}} → current date
{{INSIGHTS_SUMMARY}} → "Total: 7 insights (churn: 3, engagement: 2, monetization: 2)"
{{INSIGHTS_DATA}} → full JSON array of all insights (with IDs)

Send to LLM

LLM responds with JSON:
{
"recommendations": [
{
"title": "Send Extra Lives After 3 Failures on Level 42",
"description": "...",
"priority": 1,
"target_segment": "Players who failed level 42 3+ times",
"segment_size": 642,
"expected_impact": {
"metric": "retention_rate",
"estimated_improvement": "+15-20%"
},
"actions": ["Step 1...", "Step 2...", "Step 3..."],
"related_insight_ids": ["churn-1", "levels-2"],
"confidence": 0.85
}
]
}

Related insight IDs: Each recommendation references the insights it addresses via related_insight_ids. These are the IDs assigned in Phase 5. The dashboard shows bidirectional links — recommendations show which insights they address, and insight detail pages show related recommendations.

Phase 8: Saving

The agent writes the complete DiscoveryResult to MongoDB:

DiscoveryResult:
- project_id, domain, category
- run_type: "full" | "partial" | "failed"
- areas_requested (if selective run)
- total_steps, duration
- insights[] (with validation results)
- recommendations[] (with related_insight_ids)
- summary (totals, errors)
- exploration_log[] (every SQL query + result)
- analysis_log[] (full LLM dialog per area)
- recommendation_log (full LLM dialog)
- validation_log[] (verification queries + results)

The run status is updated to completed (or failed if critical errors occurred).

Error Handling

ErrorWhat happens
Invalid API keyAgent fails immediately. Run marked "failed" with error message.
LLM timeoutThe specific area is skipped. Other areas continue. Run marked "partial".
All areas timeoutRun marked "failed". Error banner shown in dashboard.
SQL query errorAgent asks LLM to fix the SQL. If still fails, step is skipped.
Warehouse unreachableAgent fails during schema discovery. Run marked "failed".
Agent process crashSubprocess runner detects exit code, updates run to "failed" with error from stderr.
K8s Job failureK8s runner polls Job status, detects failure, updates run.

Cost

Each run costs:

  • LLM tokens — Exploration (many small calls) + Analysis (few large calls) + Recommendations (one call)
  • Warehouse queries — Each exploration step executes one SQL query

Use the cost estimation feature (POST /api/v1/projects/{id}/discover/estimate) to preview costs before running. The dashboard has a checkbox: "Estimate cost before running."

Next Steps