Invest Like the Best
GPUs, TPUs, & The Economics of AI Explained | Gavin Baker Interview

Episode Summary
AI-generated · Mar 2026AI-generated summary — may contain inaccuracies. Not a substitute for the full episode or professional advice.
This episode features Gavin Baker, a renowned technology investor with an encyclopedic understanding of the tech landscape, known for his infectious curiosity about markets. Baker and host Patrick O'Shaughnessy dive deep into the competitive dynamics of the artificial intelligence (AI) ecosystem, focusing on the intense battle between Nvidia's GPUs and Google's TPUs, and the profound economic shifts unfolding in the AI industry. The central thesis is that AI's underlying economics, particularly cost per token and infrastructure, are poised for dramatic changes, creating existential challenges and opportunities for every major player.
👤 Who Should Listen
- Technology investors and venture capitalists interested in the competitive landscape of AI infrastructure.
- Executives and strategists at major tech companies (Google, Nvidia, Microsoft, Meta) looking to understand market dynamics and potential shifts.
- SaaS company founders and leaders grappling with AI integration and its impact on business models and gross margins.
- Engineers and researchers in the AI and semiconductor industries seeking insights into chip development, scaling laws, and future compute architectures.
- Entrepreneurs and innovators exploring novel solutions for AI power and cooling, including space-based data centers.
- Anyone curious about the long-term economic and societal implications of artificial intelligence and its accelerating progress.
🔑 Key Takeaways
- 1.To truly understand AI's capabilities, investors and researchers must use the highest paid tiers of frontier models like Gemini Ultra or Super Grock, as free versions are analogous to judging an adult's potential based on a 10-year-old's abilities.
- 2.Scaling laws for AI pre-training are empirically intact, as reaffirmed by Gemini 3, but post-training progress has been driven by new scaling laws: reinforcement learning with verified rewards (RLVR) and test-time compute, which bridged an 18-month gap in hardware development.
- 3.Google currently holds a temporary advantage as the "lowest cost producer of tokens" due to its advanced TPUs (v6/v7) and vertically integrated design process, allowing it to strategically undercut competitors and "suck the economic oxygen" out of the AI ecosystem.
- 4.Nvidia's next-generation Blackwell chips, particularly the GB300, are anticipated to shift the cost advantage, making companies utilizing them (especially XAI, which builds data centers fastest) the new low-cost producers of tokens by early 2026.
- 5.Many large tech companies like Meta, Microsoft, and Amazon have struggled to build competitive frontier models, indicating that creating and maintaining a leading AI lab is far more complex than widely perceived, requiring not just capital but also sophisticated infrastructure management and research "taste."
- 6.The "flywheel effect" of user data feeding back into model improvement, absent in early AI, is now beginning to spin with reasoning models, creating more separation among leading labs (OpenAI, Gemini, Anthropic, XAI) that possess advanced internal checkpoints.
- 7.Traditional enterprise software (SaaS) companies are making a fundamental mistake by resisting lower gross margins for AI products, akin to brick-and-mortar retailers' slow adoption of e-commerce, which could lead to their platforms being displaced by AI-native competitors.
- 8.Data centers in space, leveraging continuous solar power and free cooling (near absolute zero on the dark side of a satellite), offer a first-principles superior solution for AI compute, potentially solving Earth-bound power and cooling constraints long-term.
💡 Key Concepts Explained
Scaling Laws for Pre-training
These are empirical observations that predict how model performance improves with increased compute, data, and model size during the initial training phase. Gemini 3 notably confirmed these laws remain intact, despite researchers not fully understanding the underlying 'how' or 'why' they work.
Scaling Laws for Post-training
Two new scaling laws driving recent AI progress: Reinforcement Learning with Verified Rewards (RLVR) and test-time compute. RLVR involves training AI models using outcomes that can be objectively verified (e.g., did a sale convert, did the model balance the books), while test-time compute refers to allowing models to 'think' or process for longer during inference. These laws enabled significant progress even when pre-training hardware was stalled.
Low-Cost Producer of Tokens
In the AI industry, this refers to the entity that can generate AI output (tokens) at the lowest computational cost. Gavin Baker highlights that Google's TPUs have given them this advantage, allowing them to exert economic pressure on competitors. This metric is uniquely important in AI, unlike most traditional tech industries where low-cost production hasn't been the primary driver of market value.
Reasoning (in AI)
Refers to an AI model's ability to 'think' or chain together logical steps to solve problems or generate more sophisticated outputs. The advent of reasoning models (like the first from OpenAI) significantly accelerated AI intelligence levels and enabled a 'flywheel' where user interactions and verifiable outcomes can be fed back to continuously improve the model.
Data Centers in Space
A visionary concept for AI infrastructure, proposing orbiting satellites equipped with chips for compute. From first principles, they offer superior conditions: constant, intense solar power (eliminating batteries) and free cooling to near absolute zero in the vacuum of space (reducing complex cooling systems), potentially leading to faster, lower-cost inference and training via laser communication.
SaaS AI Margin Mistake
The error made by incumbent Software-as-a-Service (SaaS) companies who are reluctant to adopt AI agents because it would reduce their traditional 70-90% gross margins to 35-40%. This is compared to brick-and-mortar retailers ignoring e-commerce. Gavin argues this will leave them vulnerable to AI-native startups willing to operate at these lower margins, eventually leading to their displacement.
⚡ Actionable Takeaways
- →Subscribe to the highest paid tiers of frontier AI models (e.g., Gemini Ultra, Super Grock) to accurately assess their capabilities and track progress.
- →Follow leading AI researchers and engineers on platforms like X (formerly Twitter) and read their papers (e.g., Andrej Karpathy's work) to stay at the cutting edge of AI developments.
- →Listen to podcasts featuring engineers and researchers from the four leading AI labs (OpenAI, Gemini, Anthropic, XAI) to gain direct insights into frontier model progress.
- →Utilize AI tools to manage information overload from the rapidly evolving AI landscape, like asking an AI to summarize podcasts or research papers you've consumed.
- →For SaaS companies: immediately embrace lower gross margins (e.g., 35-40%) for AI-driven agent strategies to avoid obsolescence and capitalize on existing customer data and distribution, as Microsoft has done with Copilot.
- →Keep an eye on companies that excel at applying AI-driven productivity in traditional industries (like C.H. Robinson matching freight loads), as these indicate tangible ROI beyond the immediate tech giants.
- →Consider the long-term implications of power constraints for compute on Earth, recognizing that advancements in natural gas, solar, and potentially data centers in space are critical drivers for AI growth.
⏱ Timeline Breakdown
💬 Notable Quotes
“"The free tier is like you're dealing with a 10-year-old and you're making conclusions about the 10-year-old's capabilities as an adult. And you could just pay and I do think actually you do need to pay for the highest tier whether it's Gemini Ultra, you know, um, Super Grock, whatever it is, you have to pay the $200 per per month ti whereas those are like a fullyfledged 30 35y old." – Gavin Baker [02:01]”
“"With software, anything you can specify, you can automate. With AI, anything you can verify, you can automate. It's such an important concept and I think an important distinction." – Gavin Baker [09:07]”
“"AI is the first time in my career as a tech investor that being the lowcost producer has ever mattered. Apple is not worth trillions because they're the lowcost producer of phones... It's never mattered." – Gavin Baker [11:10]”
“"This is a life ordeath decision. And essentially everyone except Microsoft is failing it. To quote that memo from that um Noia guy long ago, like their their platforms are burning." – Gavin Baker [75:19]”
More from this guest
Gavin Baker
Listen to Full Episode
📬 Get weekly summaries like this one
No spam. Unsubscribe anytime. By subscribing you agree to our Privacy Policy.
Continue Exploring





