Day 2: State of AI in 2025
Day 2 was fun because I was able to review how players are shaping and shaking the 2025 AI market. After going through three different lists (Forbes, TechCrunch, and CB Insights) as well as a stack of seminars and lectures, I could step back for a comprehensive industry scan while zooming in on a handful of market leaders that have risen to the very top. With that context in hand, I've turned the patterns I spotted today into concrete strategies and lessons, and foods for thought.
Summary
Below is a quick-reference table of the 24 U.S. AI startups that TechCrunch lists as having raised ≥ $100 million so far in 2025, with their latest-round details and an official landing-page link for each.
When I scroll through LinkedIn, I see a lot of AI start-ups announcing new $5 million seed rounds. Some will turn into solid businesses, but it is hard to know which ones. The companies raising over $100 million are different - investors only write checks that big when something substantial is in place, whether it be technology, data, or a clear market lock-in. Therefore, I decided to dig deeper. I watched a few industry reports and seminars, and compared a wide mix of AI start-ups and larger players to see what really sets them apart. The work left me with 15 foods for thought from today's session.
1. Capital has hardened into a true bar-bell structure
At the heavy end, multi-billion-dollar rounds are feeding compute-hungry organizations such as OpenAI, Anthropic, Lambda, Celestial AI and Nvidia cloud partners. Those dollars finance new fabs, Blackwell-class GPUs, and sovereign LLM projects that burn megawatts per week. At the other end, a dense swarm of sub-$40 million seed and Series A checks flows into agent frameworks, evaluation tooling, photonic or analog accelerators, and highly specialized vertical apps. A founder must decide, with brutal honesty, whether the idea belongs in the cap-ex tier, where the primary advantage is scale and access to silicon, or in the lean tier, where the right play is to exploit compliance, proprietary data and speed. Sitting in the lukewarm middle, without a data asset or hardware differentiator, means struggling for oxygen because investors can already buy better risk-adjusted exposure in the two extremes.
2. A wrapper is a feature, not a company, unless it owns something uncopyable
The most common mistake in 2024 and 2025 fundraising decks is pitching a "ChatGPT for X" that simply forwards user text to an API, packages the response in a web view and adds Stripe for billing. A junior engineer can replicate that path in an afternoon, and Microsoft Copilot or Apple's on-device LLM will subsume the use case as soon as it hits mainstream demand curves. By contrast, Harvey persists because it connects to millions of archived contracts under confidentiality, performs line-level red-lining and is licensed under matter-by-matter billing that maps to existing firm economics. Ship fast, yes, but every sprint must deposit an asset that cannot be cloned with Tailwind CSS.
3. Trust and data pipelines have eclipsed model glamour in enterprise purchasing
Between January and June Gartner repositioned AI-Ready Data and Responsible AI from supporting roles to the absolute crest of the hype cycle. Boards now start every vendor call with three questions: How do you guarantee lineage, where is the watermark and what is your real-time cost telemetry? Companies responding with citation panels, policy-aware access controls and model-ops dashboards are closing contracts that were frozen in proof-of-concept limbo throughout 2023. Snorkel weak-supervision slices, Perplexity in-line citations and Runway C2PA provenance show that front-line revenue has shifted from the LLM itself to the reliability envelope around it.
4. Infrastructure efficiency is the new gold-rush lever
Celestial AI 25-terabyte-per-second optical interconnect, EnCharge AI twenty-fold TOPS-per-watt analog chiplets and TensorWave liquid-cooled MI300X super-nodes all aim at the same pain point: token-level cost curves that have flattened at roughly one cent per thousand tokens for GPT-4o but remain five to ten cents once orchestration, retrieval calls and guardrails are included. An enterprise that logs two-hundred-million tool-invocation tokens per month can save or lose millions from a mere 30 percent efficiency delta. Infra start-ups that promise concrete percentage savings over Nvidia H100 baseline win purchase orders before the product is fully mature, which explains why those rounds push past $200 million despite thin revenue.
5. Vertical agents consistently out-compete horizontal chatbots
Harvey embeds deeply inside the document life cycle of Am Law 100 firms: drafting, red-lining, matter-code metering and multi-jurisdiction citation styles. Abridge injects an LLM straight into the Epic and Cerner EHR flow, converting patient speech into ICD-10 coded notes with peer-reviewed accuracy deltas. Runway locks creative studios by integrating with Adobe After Effects project files and preserving chain-of-ownership through cryptographic provenance. These products are defensible even if open-weight LLM performance catches up, because their workflows and regulatory approvals replicate the moat that SAP and Epic Systems built in past decades.
6. Governance and observability are profit centers, not box-checking exercises
Snorkel Flow labels and slices data so data scientists can see where the model stalls. LM-Arena crowdsources Elo scores that instantly penalize over-claimed releases, and Robust Intelligence injects adversarial fuzzing to surface injection vectors before users do. These services land at five- to six-figure annual contracts because insurers, risk committees and soon the SEC will demand continuous attestation that the model customers rely on has not drifted or gone off policy. Selling dashboards that compress ten hours of compliance paperwork into a single JSON payload is poised to become as large as the DevOps monitoring market that birthed Datadog.
7. Ninety-day product cycles beat multi-year roadmaps in this field
Perplexity's leaders freely admit they do not plan further than one quarter because median capability jumps occur inside that window. They killed a traffic-generation "Pages" bet when Google throttled SEO rankings and recycled the crawler into the personalized Discover feed that now drives time-on-site. The lesson: tie every quarter to a ship-and-learn goal, write a mandatory post-mortem for all killed features and salvage anything reusable. Start-ups that allocate ten percent of sprint points to composting cancelled code into new initiatives move materially faster than those that let dead features rot in Git branches.
8. Citation-first user experience is a strategic wedge, not just etiquette
Perplexity's answer card shows web sources prominently and lets users click through, which simultaneously calms hallucination anxiety, satisfies publishers and differentiates the search feel from Google SERP packed with ads. In the creative realm, Runway's provenance watermark assures studios that generated footage can survive in a rights chain. Showing your chain of thought, or at minimum your inputs, is now a purchasing criterion for banks and law firms. If your UI hides the evidence, you will bleed deals to someone who exposes it.
9. Open-weight momentum keeps monopoly fears in check and lowers COGS
DeepSeek-R1 and Llama-3 70B closed the test-score gap to Claude-3 within single-digit points, and can be quantized to eight gigabytes of VRAM. A founder willing to fine-tune, LoRA-inject and distill one of these checkpoints can own the entire stack, avoid API lock-in and deploy in data-sovereign regions. Even larger incumbents are hedging: Snowflake Arctic ships under an open license, and Salesforce Yellowstone is in private beta precisely to maintain price leverage against closed-weight vendors.
10. Edge inference will trigger the next SDK land grab
ElevenLabs achieves sub-100-millisecond latency by running distilled speech-to-speech on-device, and EnCharge AI aims to ship PCIe cards that deliver twenty times the in-memory compute efficiency of general-purpose GPUs. As 5- to 7-billion-parameter models generate chat-quality outputs on a smartphone NPU, developers will need tooling for quantization, local RAG caches and federated learning updates. The companies that perfect edge deployment workflows may own a market as lucrative as early mobile analytics and crash-reporting SDKs.
11. Regulation is turning into a market-making tailwind
Instead of fearing every AI bill, founders should read them as a feature wish-list. The White House Safety Institute scorecards, the EU AI Act risk tiers and the OECD classification guidelines all call for specific artifacts: lineage proofs, bias audits and data-provenance logs. Products that emit those artifacts automatically convert a compliance burden into a purchasing reason. That dynamic explains why Abridge advertises HIPAA-grade BAA signatures and why Shield AI markets Pentagon certifications.
12. Crowd benchmarking serves as an honesty thermostat
LM-Arena's 3.5 million pairwise votes demolished several lab marketing claims within days, pushing model developers to evidence fine print with transparent logs. Publishing your model to an arena instantly exposes jailbreak robustness, non-English coverage and long-context proficiency. Labs that refuse public benchmarks now appear evasive, a perception that can haunt fundraising or procurement. Participating in crowd evaluation is becoming the battlefield equivalent of open-sourcing unit tests.
13. Compute-power brittleness forces contingency plans
Every modern LLM stack leans on Nvidia silicon that rolls off a single TSMC campus in Hsinchu. Export controls or a natural disaster create a non-zero existential threat. Mission-critical builders therefore adopt a two-level hedge: first by distilling frontier checkpoints to run on older or alternative GPUs, and second by testing clusters built on MI300X or Gaudi-3, and increasingly on photonic or analog boards. If your architecture cannot execute in at least two silicon ecosystems, investors will mark you as a higher risk than peers who can.
14. Geopolitics will splinter deployment topologies
The European Union push for sovereign AI, India DPDP data-locality rules, and China CSL export filters mean a single U.S. cloud footprint no longer suffices. OpenAI GPT-4o is still absent in several jurisdictions because of local rules on child protection and SDG alignment. Start-ups that build region-switching scripts, local-weights fallback and multi-cloud compliance matrices will win RFPs from global conglomerates that cannot accept data-flow ambiguity.
15. Survivorship equals velocity multiplied by moat depth
A simplified formula captures every insight above: Survival Score = Quarterly ship rate × (Data moat + Infra moat + Distribution moat). If the ship rate is zero, your moat ossifies under faster rivals. If moat depth is zero, a GPU price hike, a Copilot feature launch or an open-weight model will wipe you out. The companies that appear on multiple "best of" lists all execute fast while growing one hard-to-copy asset every sprint. Emulate that playbook, and you stand a chance of being listed in 2026 rather than becoming part of the 99 percent casualty statistic.
Early Access
We share our learnings and insights on our newsletter. We also provide weekly insights on AI progress and new tools.
By signing up, you agree to our privacy policy.