Startup ProjectBuild the future
← All episodes
Inside Story of Building the World’s Largest AI Inference Chip | Cerebras CEO & Co-Founder Andrew Feldman
116Episode 116

Inside Story of Building the World’s Largest AI Inference Chip | Cerebras CEO & Co-Founder Andrew Feldman

with Andrew FeldmanCEO

2026-01-21
Podcast Episode

Andrew Feldman, Co-Founder & CEO of Cerebras Systems, reveals the inside story of building the world's largest AI chip. Learn how their wafer-scale architecture challenges NVIDIA by delivering unprecedented speed for AI inference.

5 Things You'll Learn from This Episode

  1. Understand why a new compute workload, like AI, creates massive opportunities for new computer architectures to emerge and challenge incumbents.
  2. Learn how Cerebras's wafer-scale chip overcomes GPU memory bottlenecks by placing massive amounts of fast SRAM memory directly on-chip.
  3. Discover why AI inference is exploding and becoming the dominant cost and performance challenge, shifting focus away from just training.
  4. Analyze the strategy of selling a complete, integrated system rather than just a component chip to capture more value and solve customer problems.
  5. Recognize that in deep tech, aiming to be 10-20x faster is a better strategy than being 30-40% cheaper when competing with giants.

About the Episode

Andrew Feldman shares the origin story of Cerebras Systems, from their 2015 thesis that a new AI workload demanded a new chip architecture to building the largest chip ever made. He breaks down the technical advantages of their wafer-scale design for AI inference, the business strategy required to compete with giants like NVIDIA, and why the explosion in AI usage makes inference, not training, the new frontier.

Timestamps

  • 0:00 — Introduction: The $3 Trillion Race for AI Compute
  • 3:45 — The Thesis for Starting a Chip Company in 2015
  • 6:21 — When a New Workload Creates New Opportunity
  • 9:05 — Building the World's Largest Chip
  • 14:15 — How Wafer-Scale Architecture Solves the Memory Bottleneck
  • 18:01 — Why AI Inference is Exploding
  • 21:20 — Competing with Giants like NVIDIA and AMD
  • 26:48 — The Myth of Compute as a Commodity
  • 33:05 — Cerebras's Product Strategy: Systems, Cloud, and On-Prem
  • 40:45 — The Rise of Specialized 'Neo-Cloud' Providers
  • 46:30 — Why Cerebras Sells Systems, Not Just Chips
  • 52:00 — How the Rise of Inference Weakens NVIDIA's Cuda Moat
  • 57:15 — The Strategic Partnership with G42
  • 59:50 — Discovering the Power of Inference

About Andrew Feldman

Andrew Feldman is the Co-Founder and CEO of Cerebras Systems, a company dedicated to accelerating artificial intelligence compute. A serial entrepreneur, Andrew previously founded SeaMicro, a pioneer in low-power server technology, which was successfully acquired by AMD.

Co-Founder & CEO at Cerebras Systems

About the Host

Nataraj Sindam is the creator of The Startup Project, a podcast featuring founders, investors, and operators building the future.

Twitter

Newsletter

Website


Subscribe to Startup Project for more founder conversations!

Spotify

Apple Podcasts

Newsletter

#StartupProject #AndrewFeldman #Cerebras #AIChips #Semiconductors #NVIDIA #AIInference #DeepTech #VentureCapital #WaferScale #AICompute #Hardware

Share:PostLinkedIn