The problem: AI is transforming market research — but who verifies AI’s work?
Market research firms like CB Insights, PitchBook, and Gartner are increasingly using AI to process data, identify trends, and generate insights at scale. Their reports shape billions in capital allocation. But AI introduces two risks that compound each other: the data it pulls may conflict across sources, and the analysis it produces has blind spots it can’t see.
This creates a problem for both sides. For research firms: if a single AI model is helping build your analysis, it may miss cross-source conflicts in the data it aggregates, and it won’t flag the sectors, risks, or context it doesn’t think to include. Your report goes out with gaps you didn’t know existed. For users of that research: you’re treating the report as ground truth without knowing which figures conflict with other sources, or what the analysis leaves out that would change your strategy.
Multi-model AI verification solves both sides. Research firms can use it as a QA layer before publication — catching data conflicts and blind spots before clients do. Users can run any report through it to understand what’s verified, what’s contested, and what’s missing. Either way, the question isn’t whether the report is “wrong” — it’s whether it’s complete enough to act on.
The experiment: verifying the industry’s benchmark AI report
We tested this with CB Insights’ State of AI 2025 — the definitive annual report on AI funding, M&A, and unicorn formation. It reported $225.8B in global AI funding, 75 new unicorns, and 782 M&A exits. We ran two separate analyses through TruVerifAI: one cross-referencing the data against other sources, and one checking for critical omissions.
What we did
Source Report
Took CB Insights’ State of AI 2025 covering funding, deals, M&A, and unicorn trends.
Multi-Model Audit
Ran the report through TruVerifAI — GPT, Claude, Gemini, and Grok cross-referencing data and checking completeness in Justify mode.
Two-Part Analysis
Checked for cross-source data conflicts (do other reports agree?) and critical omissions (what’s missing that changes the investment thesis?).
Where the data conflicts with other sources
CB Insights’ headline figure — $225.8B in global AI funding — is the most-cited number in the industry. But when TruVerifAI cross-referenced it against Crunchbase, Stanford HAI, PitchBook, and HumanX, the models found meaningful discrepancies across multiple metrics:
| CB Insights figure | Other source | Gap | Why it matters |
|---|---|---|---|
| $225.8B global AI funding | $202–211B (Crunchbase) | 7–10% | $15–24B difference driven by how broadly “AI company” is defined. CB Insights includes AI-adjacent firms; Crunchbase uses stricter criteria. |
| $67B generative AI funding | $89B (Stanford HAI) | 33% gap | Stanford includes corporate R&D and infrastructure investments. CB Insights tracks only VC/PE deals. Neither is “wrong” — but the $22B gap changes how you size the gen AI market. |
| 97 AI unicorns globally | 89 (HumanX AI Index) | 9% | Eight unicorns in dispute — likely due to different definitions of “AI company” versus “AI-enabled company.” Matters for portfolio construction. |
| 40% of Fortune 500 using gen AI | 65% (Ropes & Gray survey) | 25pp gap | CB Insights measures production deployment; Ropes & Gray includes pilots. The gap reveals the “pilot purgatory” problem — most enterprise AI hasn’t reached production. |
The most telling moment in the analysis: GPT found zero discrepancies in Round 1 and stated it couldn’t produce a comparison table without “fabricating data.” Claude, Gemini, and Grok each independently found different comparison sources — Crunchbase, Stanford HAI, PitchBook, HumanX. After deliberation, GPT revised completely, incorporating the others’ evidence. This is the single best argument for multi-model verification: if GPT were the only reviewer, the report would have been marked “verified” with no flags.
What the report doesn’t cover — and why it matters
The data conflicts are important, but the blind spots are more consequential. CB Insights tracks funding, deals, and exits brilliantly. But its report omits the context that determines whether those investments will actually pay off. All three models converged on the same 6 omissions — each adding different angles that the others underweighted:
| What’s missing | Why it changes the investment thesis |
|---|---|
| Revenue, profitability, and unit economics | The report tracks $225.8B in funding and 75 new unicorns but zero data on ARR, burn rates, CAC/LTV, or path to profitability. Investors can’t distinguish between real businesses and cash furnaces. Would shift the narrative from “AI boom” to “AI funding boom with uncertain business viability.” |
| Compute infrastructure constraints | GPU scarcity, data center power limits, and skyrocketing training costs create hard caps on who can compete. Mega-rounds may reflect infrastructure cost inflation rather than business traction. Gemini uniquely emphasized energy grid capacity as the binding constraint. |
| Open-source model deflationary pressure | The report focuses on closed-source LLMs that captured $93.1B in funding. No mention of Llama, Mistral, or the open-source ecosystem commoditizing capabilities these companies are spending billions to build. Challenges the sustainability of current valuations. |
| Regulatory and antitrust enforcement risk | The report celebrates 782 M&A exits (1.5x 2024) but ignores that FTC/DOJ antitrust scrutiny, the EU AI Act, and AI copyright litigation could block deals and freeze liquidity. Gemini specifically flagged that Big Tech acquisitions face potential blocking. |
| Pilot-to-production conversion reality | The report implies demand through funding but provides zero data on whether enterprises are converting AI pilots to production at scale. The 40% vs 65% adoption gap (above) reveals the disconnect. Grok uniquely added failure rates and down rounds as missing counter-data. |
| Geopolitical constraints and market fragmentation | The report includes China data without explaining how US chip export controls reshape competitive dynamics. Gemini provided the clearest causal explanation: China’s focus on robotics/applied AI isn’t a market preference — it’s a forced pivot due to hardware sanctions. |
Rare multi-model consensus with different angles: All three models in the blind spots analysis converged on the same 6 omissions — unusual in TruVerifAI’s deliberation process. But each model emphasized different facets: Gemini focused on energy grid constraints and antitrust deal-blocking; Grok highlighted failure rates and down rounds as missing counter-data; GPT emphasized distribution moats and public market validation gaps. The convergence confirms these are genuine blind spots. The different angles make the analysis richer than any single model could produce.
How it works: multi-model research verification
TruVerifAI queries multiple AI models simultaneously and synthesizes their responses through structured deliberation. For market research, each model independently cross-references data, identifies methodological differences, and surfaces missing context — then challenges the other models’ findings across two rounds. The result is a verification layer that catches what even the best single-model review misses.
TruVerifAI Report — Justify Mode
Two separate analyses were run: one cross-referencing data against other sources, and one identifying critical omissions. In the data verification pass, GPT initially found zero comparison data while three other models each found different sources — the strongest possible case for multi-model review. In the blind spots pass, all models converged on 6 omissions but contributed unique angles.
Note: TruVerifAI was asked to flag a limited number of issues per analysis. Without those constraints, the multi-model process would surface additional discrepancies and omissions.
Download the full reports
See the original CB Insights report and the complete multi-model analyses:
Original Report
CB Insights: State of AI 2025 — funding, M&A, and unicorn formation
Verification Report
Cross-source data conflicts in funding figures, unicorn counts, and adoption metrics
Blind Spot Report
6 omissions that shift the narrative from “AI boom” to “uncertain viability”
Build multi-model verification into your research process
Whether you produce market research or act on it, we’re selecting design partners who’ll shape TruVerifAI for research intelligence. Free access. Direct input on the roadmap.
Who this is for
Research Firms & Data Providers
Add a multi-model QA layer before publication. Catch cross-source data conflicts, surface missing context, and strengthen your analysis before clients find the gaps. Protect the credibility your brand depends on.
Investors & Analysts Using Research
Verify the reports you base decisions on. Know which figures conflict across sources, what methodological differences explain the gaps, and what the analysis leaves out that would change your thesis.
Strategy & Corporate Development
Before presenting market research to leadership, run it through multi-model verification. Surface the risks, constraints, and competing data points that single-source reports don’t cover.