General-purpose deep research with multi-source synthesis, confidence scoring, and anti-hallucination verification. Adopts SOTA patterns from OpenAI Deep Research (multi-agent triage pipeline), Google Gemini Deep Research (user-reviewable plans), STORM (perspective-guided conversations), Perplexity (source confidence ratings), and LangChain ODR (supervisor-researcher with reflection).
| query | The user's research question or topic; the unit of investigation | | claim | A discrete assertion to be verified; extracted from sources or user input | | source | A specific origin of information: URL, document, database record, or API response | | evidence | A source-backed datum supporting or contradicting a claim; always has provenance |
| provenance | The chain from evidence to source: tool used, URL, access timestamp, excerpt | | confidence | Score 0.0-1.0 per claim; based on evidence strength and cross-validation | | cross-validation | Verifying a claim across 2+ independent sources; the core anti-hallucination mechanism |