Disclaimer: This content is for informational and educational purposes only and should not be construed as financial or investment advice. Always do your own research and consult a licensed financial advisor before making investment decisions.
Disclosure: The author holds a long position in NVDA.
Platform strength endures as AI spend broadens to networks and software. Upside remains, yet today’s scale and valuation temper 2030 gains.
Analysis
Thesis
The default AI compute and networking stack keeps compounding into a platform (CUDA/NIMs + Blackwell/GB200 + Spectrum/Quantum) with energy and sovereign buildouts, but sheer size caps upside to a disciplined 1.5–2.0× by 2030 unless software/compute-as-a-service inflects.
Last Economy Alignment
Owns attention-era compute rails; network capital with hyperscalers/sovereigns; flywheel reinvests cash into models/GPUs/energy.
Growth Outlook
Average Implied Multiple (to 2030)
1.8x (from 4 most recent periods)
Reasoning
Dominance persists as AI infra expands, but share moderates and mix shifts to systems/networking/software. Even with strong execution, today’s size and premium compress upside; platformization can still drive ~1.6× EV by 2030.
Risk Assessment
Overall Risk Summary
Key swings: China licensing, hyperscaler capex cadence, AI Ethernet vs. InfiniBand standards, HBM/packaging capacity, grid power and siting, and rising custom silicon. Any stumble hits a premium multiple.
Trends
Key Changes
Q2 FY26 revenue $46.7B; Q3 FY26 guide $54B with China H20 excluded.
Blackwell cadence and GB200 systems broadened; Spectrum‑XGS launched.
Europe sovereign/industrial AI cloud partnerships expanded.
AI networking momentum shifts toward open Ethernet initiatives.
HBM capacity and packaging scale improving; power/siting is gating.
Thesis update
No change to core: compute platform dominance. Update: more open Ethernet and custom silicon headwinds; power/export limits temper share.
Forecast update
Valuation expanded and growth normalized; size now constrains 2–5× scenarios, so expectations trimmed.