QuantQ AI Stock Analysis Background

QuantQ: AI Stock Analysis Agent

In-depth analysis of an intelligent stock analysis agent using AI

Back to Portfolio

QuantQ - Product Requirements Document

Author: J Sankpal · Version: 3.0 · March 2026 · Status: Implementation-Ready


Executive Summary

The gap: Financial professionals spend 10-15 minutes per question cross-referencing dashboards with 200-page SEC filings. The tools that solve this cost $10K-25K/year. General-purpose LLMs offer conversational access but hallucinate numbers and can't cite primary sources.

QuantQ is the first financial intelligence platform where every answer is simultaneously:

PropertyWhat it meansWhy it matters
Computationally groundedNumbers parsed from SEC XBRL machine-readable tags - not scraped, not generatedEliminates wrong-period, rounded, and typo errors that plague web-scraping approaches
Contextually intelligentAgentic orchestrator connects data (what happened) with filing narrative (why) in multi-turn conversationUsers get the number AND the management explanation in one response
Fully auditableEvery claim links to exact filing section and XBRL tag; full provenance chainOutput can go directly into client memos, compliance files, and pitch decks

No single competitor delivers all three. Bloomberg has the numbers. AlphaSense has the text search. QuantQ welds them together into production-ready research output at 1/200th the price.

Launch: Russell 3000 (tiered quality) · $49/mo Pro · $199/seat Teams · 10-week build

Target: 1,500 MAU and $5.4K MRR by Month 3 · $15K+ MRR by Month 6


The Problem

The 10-minute tax on every financial question

See interactive diagram below: The 10-Minute Tax on Every Financial Question

Why nothing on the market solves this

SolutionPriceGets rightGets wrong
Bloomberg / Capital IQ$24K+/yrComprehensive dataNumbers and narratives on separate screens; priced for institutions
AlphaSense$10K+/yrBest transcript searchText only - no structured metrics, no charts, no calculations
Dashboards (Koyfin etc.)$39-299/moClean charts, affordableStatic - can't explain why metrics changed
FinChat$29-79/moConversational AILimited source verification, no audit trail
ChatGPT / Claude$0-20/moNatural conversationHallucinate numbers, cite web articles not filings
SEC EDGARFreeAuthoritative source200-page PDFs, no search, no visualization

The white space: Institutional-grade intelligence at prosumer prices. Nobody occupies the $50-200/month range with filing-grounded AI.

See interactive diagram below: The White Space Nobody Occupies


Who We're Building For

Primary: Independent RIA / Financial Advisor ($10-50M AUM)

Trigger: Client calls after earnings - "Should I be worried about my Apple position?" Advisor needs cited analysis within the hour.

Job: "I need filing-grounded insights I can reference in client communications - institutional-quality advice without institutional-cost tools."

Current pain: 30-45 min per company to cross-reference dashboard + EDGAR + Excel. No single tool connects the number → the explanation → the source citation.

QuantQ value: Same analysis in 2-3 minutes. Client email includes "per Apple's FY2024 10-K, Item 7" - not "I read somewhere that..."

WTP: $49/mo < one billable hour. Saves 10+ hours per quarterly review.

Secondary: Equity Research Analyst at Boutique Firm

Job: "Pull filing-verified data with provenance I can cite in pitch decks and comp tables."

Pain: Terminal access rationed. Building a 10-company comp table takes hours of manual extraction. Errors are career-limiting.

WTP: $199/seat replaces $50K-100K/yr in supplemental terminal licenses.

Tertiary: Sophisticated Retail Investor (Dividend/Value Focus)

Job: "Complete my quarterly portfolio review in 2 hours instead of 10."

WTP: Free tier for casual use; $49/mo during earnings season.


The Aha Moment

The aha is NOT "that was fast." It's: "I could put my name on this output."

See interactive diagram below: The 3-Turn Aha Moment Sequence


Solution Overview

System Architecture

See interactive diagrams below: System Architecture - Layer by Layer, and System Workflow - Use Case Walkthrough

What goes where (storage routing)

QuestionStoreWhy
Has an XBRL tag? Single value with a unit?PostgreSQL (~60% of queries)Deterministic lookup, <200ms
Requires understanding prose meaning?Pinecone (~30% of queries)Semantic retrieval over filing text
Needs both number AND explanation?Both (~10% of queries)PostgreSQL for data, Pinecone for context

Data Coverage: Russell 3000, Tiered Quality

TierCompaniesAccuracyUser indicatorValidation
1S&P 500 (~500)99.5%+Green: Fully ValidatedDaily automated benchmark + manual spot checks
2Russell 1000 ex-S&P (~500)98%+Blue: ValidatedDaily automated benchmark + exception review
3Remaining Russell 3000 (~2,000)95%+Yellow: Best Effort - verify with sourceAutomated parsing with confidence scoring

Coverage: 50 core metrics · 5-year history (10-year for Teams) · 10-K, 10-Q, 8-K filing types


Functional Requirements (MVP)

FeatureDescriptionWhy it matters
Conversational Q&ANatural language → charts + narrative + citationsReplaces the 10-min multi-tool workflow
Auto-generated chartsBar, line, comparison tables based on query intentProduction-ready visuals for deliverables
Source attributionClickable link to exact filing section on every claimAudit trail for professional use - "no link, no fact"
Multi-turn conversationContext preserved; "compare that to MSFT" worksEnables research sessions, not just single lookups
Multi-company comparison (Pro)Up to 5 companies side-by-side, all sourcedComp tables that would take 30-45 min manually
Financial synonym resolution"top line" → "revenue" → "net sales" → XBRL tagEliminates missed retrievals from embedding gaps
Structured export (Pro)Excel, formatted tables with provenance metadataOutput fits directly into professional workflows
Tiered quality indicatorsGreen/blue/yellow badges on every valueUsers calibrate trust based on data quality
Confidence-based responseHigh → state directly · Medium → caveat · Low → declinePrevents silent errors; maintains accuracy bar

Query Types

TypeExampleSpeed
Single metric"Apple's FY2024 revenue?"<200ms (fast path)
Trend"MSFT margin trend, 5 years"<2s
Comparison"AAPL vs MSFT vs GOOG margins"<2s
Explanation ("why")"Why did Tesla margin decline?"<4s
Calculation"5-year CAGR for GOOGL revenue"<1s
Follow-up"How does that compare to TSLA?"<2s

Out of Scope (MVP)

Stock recommendations · Earnings call transcripts (V1.1) · International equities · Portfolio tracking/alerts (V1.1) · Form 4/13F/DEF 14A (V1.1) · Real-time prices (delayed Yahoo Finance quotes for valuation metrics only)


Activation & Conversion Design

Free-to-Pro wall: gate workflows, not quality

FeatureFreePro ($49/mo)Teams ($199/seat)
Queries/day10UnlimitedUnlimited
Company coverageRussell 3000Russell 3000Russell 3000
MetricsAll 50 (single company)All 50 + derivedAll 50 + derived + custom
ComparisonsNot availableUp to 5 companiesUp to 10 companies
HistoryCurrent year5 years10 years
ExportNot availableExcel, formatted tablesExcel, PDF, API
Multi-turn3 follow-upsUnlimitedUnlimited
Saved sessionsNot available10/monthUnlimited
Audit trailNot availableNot availableFull provenance export
SSONot availableNot availableSAML

Conversion trigger: User hits the wall during professional work - needs to compare companies (gated), pull 5-year history (gated), or export a table for a presentation (gated). The wall is felt during the workflow that justifies $49/mo, not during casual browsing.


Success Metrics

North Star

Weekly Active Research Sessions (3+ queries on a topic): 200+ by Month 3

Primary Metrics

MetricTargetIf missed
D7 Retention12%+<8%: retention mechanics failing
Free-to-Pro Conversion (30-day)5-8%<3%: value prop not landing with professionals
Queries per Session4.0+<2.5: conversation UX failing
Source Link CTR15%+<8%: core trust value prop undermined
Export Rate (Pro)15%+ of sessions<8%: output quality not meeting professional bar

Guardrails (must not degrade)

MetricThreshold
Tier 1 accuracy99.5%+
Tier 2 accuracy98%+
Tier 3 accuracy95%+ with confidence flags
P95 latency<2s structured, <4s narrative
Source attribution rate100%

Revenue Targets

Month 1Month 3Month 6
MAU3001,5005,000
Pro users2090300
Team seats0525
MRR$980$5,405$19,675

Risk Assessment

Component Risks

ComponentRiskMitigation
XBRL PipelineHIGH - non-standard filings cause silent errors harder to detect than hallucinationsTiered quality; daily automated accuracy checks per tier; confidence flags; Russell 3000 only via tiered expansion
Intent RouterMEDIUM - misclassification sends queries to wrong pathConfidence scoring + clarification fallback; 8-10 few-shot examples per query type
Narrative RAGMEDIUM - financial synonym gaps cause missed retrievalsSmart synonym expansion; 0.6 relevance threshold; re-query loop on low confidence
Period AlignmentMEDIUM - fiscal year-end differences cause subtle wrong-period errorsAutomated FY-end cross-reference; special handling for non-standard FY (MSFT Jun, NKE May)
Analytics ToolLOW - deterministic mathUnit tests; formula validation
Conversation MemoryMEDIUM - wrong company reference in multi-turnLIFO resolution + explicit clarification prompts

Top Business Risks

RiskMitigation
Professional users don't trust AI for financial dataFull XBRL-tag audit trail; 99.5% Tier 1 accuracy; export with provenance; "verify in 10-K" CTA
Pricing in no-man's-land (too expensive for retail, too cheap for credibility)Free tier for retail; Pro anchored against AlphaSense ($800+/mo), not ChatGPT ($20/mo)
General-purpose LLMs close the gapCompete on computational grounding (XBRL tags) vs textual grounding (web scraping) - different error class
XBRL errors compound at Russell 3000 scaleTiered quality with user-visible badges; never present uncertain data as certain
Low engagement - users ask 1-2 questions and leaveSuggested follow-ups; multi-turn context; quarterly review mode (V1.1); earnings alerts

Kill Criteria

GateWhenMust-meetIf missed
Tier 1 XBRL accuracyWeek 3S&P 500 at 99.5%+Add 2 weeks; do not proceed
Orchestrator worksWeek 520 test queries route correctlyRedesign intent classification
End-to-end demoWeek 83-turn aha sequence works for 5 companiesSimplify to single-turn; defer
Beta launchWeek 1025+ users, 4.0+/5 satisfactionFix and relaunch in 2 weeks
PMF signalMonth 2D7 >8%, weekly sessions >50Interview churned users; iterate
Revenue validationMonth 4$5K MRR, Pro conversion >3%Reassess pricing and persona fit

Pre-Mortem: Why This Fails at Month 6

#Failure modePrevention
1Trust gap never closes. Analysts verify manually, conclude QuantQ doesn't save time.Full XBRL-tag provenance chain. Export with citations. Aha moment in first 60 seconds. 99.5%+ accuracy - one wrong number kills trust permanently.
2Pricing wrong. $49/mo too expensive for retail, too cheap for professional credibility.Free tier generous for retail. Pro anchored against AlphaSense ($800+/mo). Teams ROI case: $597/mo for 3 seats vs $6K+/mo for terminal licenses.
3Tier 3 data erodes trust in Tier 1. Wrong numbers for small-caps make users doubt all data.User-visible tier badges. Tier 3 shows: "XBRL data not fully validated - verify with source." Never present uncertain data as certain.
4No retention. Episodic use = no return without triggers.Earnings alerts, weekly portfolio digests, session resume, quarterly review mode (V1.1).
5Incumbents respond. AlphaSense launches a cheaper tier.Speed to market. XBRL parsing + provenance chain + intervention gate = 6-12 months engineering lead.

Why Agentic AI?

"Why did this metric change?" requires multi-step reasoning no single tool handles:

  1. Identify metric - Structured data (XBRL parsing)
  2. Retrieve value - Deterministic tool (PostgreSQL)
  3. Search narrative - Semantic retrieval (Pinecone)
  4. Synthesize answer - LLM reasoning (Claude Sonnet 4)
  5. Validate output - Intervention gate (deterministic)
  6. Maintain context - Conversation memory (PostgreSQL)

The LLM reasons about documents, routes execution, and explains findings - it never fabricates raw financial figures.

LLM doesLLM does NOT do
Classify intentGenerate financial numbers
Plan tool callsMake investment recommendations
Generate explanations from retrieved textParaphrase filing text (verbatim excerpts only)
Compute derived metrics from verified inputsRecall data from training weights

AI/ML Decisions

DecisionChoiceWhyRejected
OrchestratorClaude Sonnet 4Best tool-use reliability; strong structured outputGPT-4o (weaker tool-use), Llama 3 (ops burden)
EmbeddingsOpenAI text-embedding-3-largeHighest quality on financial text similarityCohere (slightly lower), BGE (5-8% lower on domain tasks)
ApproachRAG + prompt engineeringPreserves source attribution; ships in 10 weeksFine-tuning (no labeled data, loses citations)
ArchitectureDual-speed orchestrator60% of queries don't need LLM; fast path <200msSingle orchestrator (wastes cost/latency on lookups)
StoragePostgreSQL + Pinecone (2 stores)Clear ownership; no graph DB sync complexityNeo4j (overkill for FK-based provenance chain)
Response handlingConfidence-based framingHigh→state, Medium→caveat, Low→decline; scalesHITL review queue (doesn't scale, adds latency)

Evaluation Plan

WhatMethodCadencePass
Tier 1 metric accuracyAutomated benchmark: 200+ metrics vs EDGARDaily99.5%+
Tier 2 metric accuracyAutomated benchmark: 100+ metrics vs EDGARDaily98%+
Period alignmentFY-end cross-reference; non-standard FY test casesWeekly99.5%+ (Tier 1)
Narrative relevanceEmbedding similarity + LLM-as-judgeWeekly≥0.6 threshold
Comparative accuracy50+ known comparison pairs, both-sides-correctWeekly95%+, zero inverted
Source attributionAutomated: ≥1 EDGAR link per factual claimEvery response100%
LatencyP50/P95/P99 by query typeContinuousP95 <2s structured, <4s narrative
User satisfactionThumbs up/down + issue categorizationContinuous>80% positive

HHH Launch Criteria

PhaseHelpfulHonestHarmless
Beta (Wk 10-12)>70% positive; 3.0+ queries/session100% attribution; 99.5%+ Tier 1; zero false confidenceZero investment advice; zero hallucinated numbers
Launch (Wk 14+)>80% positive; 4.0+ queries/session; D7 >8%Same at 500+ MAU; zero user-reported wrong Tier 1 numbersMonthly compliance audit passing
Scale (Mo 4+)>85% positive; D7 >12%Same at 5K+ MAU; <1% user-disputed accuracySOC 2 Type I initiated

Gate rule: Any dimension failing blocks progression. Honest failures trigger immediate pause.


Responsible AI

PrincipleHow we implement it
AccountabilityPM owns accuracy standards (99.5%+ Tier 1). Engineering owns pipeline integrity. Rollback: disable chat in 5 min; suppress individual company data in 1 min. User feedback reviewed within 24h.
TransparencyEvery claim cites exact filing section + XBRL tag. Data tier badge on every response. Confidence reflected in framing. Users know they're interacting with AI. Filing date and fiscal period disclosed.
FairnessAll Russell 3000 covered. Large-cap quality advantage acknowledged via tier badges. Free tier ensures substantive access. No personalization - same question = same answer for every user.
Reliability99.5%+ Tier 1 is a hard requirement. System declines below 0.6 confidence. No investment advice. Stale data (>30 days) flagged.

GTM & Launch Plan

Phased Launch

PhaseWhenAudienceSuccess criteria
Closed BetaWeeks 10-12Waitlist RIAs, boutique analysts, Reddit financeD7 >8%, 25+ users, 4.0+/5 satisfaction
Public LaunchWeek 14+Professional + sophisticated retail500 MAU, 5%+ Pro conversion, $980 MRR
Teams LaunchMonth 4+Boutique firms, RIA practices5 Teams accounts, SSO + audit trail working

Acquisition Bets

ChannelTacticExpected outcome
RIA communities (Kitces, NAPFA)"How I research client holdings in 2 min with filing citations"20-30 high-intent signups/post
r/dividends (185K)Weekly QuantQ vs manual analysis comparisons50-100 upvotes, 20-30 signups/post
Product Hunt"Institutional-grade financial intelligence at 1/200th the price"300-500 upvotes, 150-300 signups
Finance Twitter/X"QuantQ vs ChatGPT vs AlphaSense on the same 10 questions"10K-50K impressions, 50-100 signups
RIA conferences (T3, Orion)Demo: "AI research that meets compliance standards"10-20 Teams leads/event ($2-5K cost)

Retention Mechanics

MechanicTriggerValue
Earnings alertWatched company files 10-Q/10-K"AAPL just filed Q3 10-Q. Revenue +3.2%, margins flat."
Weekly digest (Pro)Every Monday"Your 12 holdings: 2 declining FCF, 1 filed 8-K this week."
Suggested follow-upsAfter every response"Compare to MSFT? 5-year trend? What did MD&A say?"
Session resumeReturn within 7 days"You were researching TSLA margins. Continue or start new?"
Quarterly review mode (V1.1)Manual triggerGuided workflow: review → flag changes → compare → summary

Competitive Moat

Three properties that reinforce each other - a competitor must replicate all three simultaneously:

See interactive diagram below: Competitive Moat - Three Reinforcing Properties


Roadmap

PhaseTimelineScopeGate
MVP10 weeksRussell 3000 (tiered), 50 metrics, comparisons, export, citations25+ beta users, 99.5%+ Tier 1, 4.0+/5
V1.1Mo 3-4Transcripts, Form 4/13F, segments, quarterly review mode, alertsD7 >12%, Pro >5%, Teams >3 accounts
V2Mo 5-6API, Excel plugin, PDF reports, mobile, SOC 25K MAU, $15K MRR, 25+ Teams seats
V3Mo 7-12International (UK, EU), custom metrics, portfolio analysis$50K MRR, path to $600K ARR

Pricing & Unit Economics

Pricing

TierPriceTarget
Free$0Trial + retail investors
Pro$49/mo ($490/yr)Independent advisors, analysts, sophisticated investors
Teams$199/seat/mo (annual)Boutique firms, RIA practices, corp dev teams

Rationale: $49/mo = less than one billable hour. Anchored against AlphaSense ($800+/mo) at 1/16th the price, not against ChatGPT ($20/mo). Teams at $199/seat replaces $2K+/seat terminal licenses.

Unit Economics

Amount
Cost per query~$0.02
Cost per Pro user/month (40 queries avg)~$0.98
Revenue per Pro user/month$49.00
Gross margin per Pro user$48.02 (98%)
Fixed infrastructure/month$460-660
Break-even10 Pro users

Revenue Scenarios (Month 12)

ScenarioPro usersTeam seatsMRRARR
Conservative10010$6,890$82K
Moderate30030$20,670$248K
Optimistic50050$34,450$413K

Open Questions

QuestionOwnerWhenImpact
SEC/FINRA compliance reviewLegalPre-launch (Wk 8)Launch blocker
SOC 2 timelineEng/PMMonth 4Teams adoption
Data retention policy (GDPR/CCPA)Legal/EngPre-launchUser trust + compliance
Earnings transcript quality for V1.1EngMonth 2V1.1 scope
International data sources for V3PM/EngMonth 6Roadmap commitment

Key Decisions Made

DecisionChoiceWhy
PositioningProsumer ($49-199), not consumer ($19) or enterprise ($500+)Targets the underserved gap between AlphaSense and dashboards
CoverageRussell 3000 with tiered quality, not 50 companiesProfessional credibility requires broad coverage; tiers manage risk
Accuracy bar99.5% Tier 1 (up from 98%)Professional users face career risk from wrong numbers
Architecture2 stores (PostgreSQL + Pinecone), not 3No Neo4j needed; FK provenance chain handles relationships
OrchestratorDual-speed (fast path + full path)60% of queries don't need LLM reasoning
Low confidenceConfidence-based framing, not HITL queueScales without human bottleneck

QuantQ - Grounded Financial Intelligence Every number parsed. Every claim cited. Every answer audit-ready.

The 10-Minute Tax on Every Financial Question

"Why did Tesla's gross margin decline?"

📊
Dashboard
Find metric
~2 min
🏛️
SEC EDGAR
Find filing
~2 min
📄
200-page PDF
Find section
~3 min
🔍
Read & Parse
Extract data
~5 min
🔗
Connect
Metric to why
~2 min
10-15 min per question
x 5-10 questions per session = entire workdays
vs
45 seconds with QuantQ
Sourced answer + chart + citation in one response

The White Space Nobody Occupies

Annual price per seat - institutional intelligence at prosumer prices

Bloomberg / Capital IQ
Full platform$25K+/yr
AlphaSense
AI + search$10K+/yr
QuantQ ProTHE GAP
AI + filing-grounded$600-2.4K/yr
FinChat / Seeking Alpha
AI chat$300-950/yr
Koyfin / dashboards
Data dashboards$470-3.6K/yr
ChatGPT / Claude
General AI$0-240/yr

$50-200/month range: Institutional-grade intelligence at prosumer prices. Nobody occupies this space with filing-grounded AI. QuantQ is 1/200th the price of Bloomberg with XBRL-verified accuracy.

The Aha Moment

Not "that was fast" - it's "I could put my name on this output"

U

Why did Tesla's gross margin decline last year?

Q
  • Margin chart: 25.6% → 17.9% (FY2022 → FY2024)
  • Exact MD&A paragraph explaining pricing strategy shift
  • Clickable link to Item 7, page 47 of the 10-K
💭

OK, sources are real. But ChatGPT kinda does this too.

Query Workflow

1
💬

User Query

Natural language question about stocks

2
🧠

Agent Reasoning

Claude Sonnet 4 plans tool strategy

3
🔍

RAG Retrieval

Parallel search across Pinecone + live APIs

4

Source Verified

Every fact cited with EDGAR link

5
📊

Feedback Loop

Charts delivered, user rating stored for RLHF

<2s
P95 Response Latency
0
Unsourced Financial Claims
7
Parallel RAG Tools

Architecture

Query Source

💬
Next.js 14
React 18 + Recharts
SSE Streaming / Chat UI

Agent Orchestration

Claude Sonnet 4
Intent + Tool Planning
Tool Execution
7 Concurrent Tool Calls
GUARDRAILS: Source Attribution · Hallucination Prevention · Transparency · Explainability

Intelligence

🗄️
PostgreSQL
XBRL Structured Metrics
💎
Pinecone
Narrative Filings (3 namespaces)
📈
Yahoo + EDGAR
Live Quotes + Filing URLs
Retrieval Strategy
Semantic search across 3 Pinecone namespaces
Inference Engine
Claude Sonnet 4 - reasoning, routing & derived calculations
Feedback Loop
User ratings + confidence calibration via JSONL
Data Pipeline
Java ETL: SEC EDGAR → Normalize → PostgreSQL + Pinecone

System Architecture -Layer by Layer

6-layer architecture: from user query to SEC-verified response

2 data storesPostgreSQL + Pinecone2 query speedsFast path + Full path3-layer guardrailsInput → Retrieval → Output
1

Conversational Interface

Chat UINext.js 14 + React 18

Natural language input with streaming responses

Auto-Generated ChartsRecharts

Bar, line, and comparison charts from structured data

Source Citations

Clickable links to exact SEC filing sections on every factual claim

Export ActionsPro tier feature

CSV/PDF export for tables, charts, and full analysis sessions

SSE streaming - users see reasoning + charts as they generate

Every response surfaces the filing date and fiscal period

Design philosophy: The LLM reasons about documents, routes execution, and explains -it never fabricates financial figures. All numbers come from deterministic tools backed by SEC XBRL data. The architecture enforces this at every layer.

System Workflow -Use Case Walkthrough

Trace a real query through every layer of the system

Fast Pathstructured

"What was Apple's revenue in FY 2024?"

1
Intent RouterOrchestrator

Classifies → structured metric lookup (AAPL, revenue, FY2024)

2
Fast PathOrchestrator

Direct PostgreSQL query - no LLM reasoning needed. Sub-200ms.

3
XBRL LookupData Layer

SELECT value FROM metrics WHERE ticker=AAPL AND metric=revenue AND period=FY2024

4
Intervention GateGuardrails

Source exists ✓ Value matches XBRL ✓ Filing date attached ✓

Full OrchestratorOrchestrator

Skipped -structured query doesn't need narrative retrieval

6
Response + CitationInterface

$394.3B -Source: Apple 10-K FY2024, Item 6 [EDGAR link]

<200ms
Latency
1 (routing only)
LLM Calls
PostgreSQL (XBRL)
Data Source
Query distribution:~60% structured (fast path)~30% narrative (full path)~10% hybrid (full path)

Competitive Moat

Three properties that reinforce each other - a competitor must replicate all three simultaneously

What does NOT constitute a moat
"Free SEC data" - anyone can access EDGAR"Conversational AI" - anyone can bolt on a chatbot"Source links" - general LLMs are learning to cite

The moat is the intersection and the depth. Individual pieces are replicable. The integration of all three into a single verified output is not.