Bg2 Pod
October 31, 2025

All things AI w @altcap @sama & @satyanadella. A Halloween Special. 🎃🔥BG2 w/ Brad Gerstner

Brad Gerstner of Altimeter Capital sits down with OpenAI CEO Sam Altman and Microsoft CEO Satya Nadella to dissect their landmark partnership, the economics of the great AI buildout, and the path to AGI.

The Microsoft x OpenAI Alliance

  • "I think this is one of the great tech partnerships ever and without Microsoft and particularly Satya's early conviction, we would not have been able to do this." - Sam Altman
  • "It's kind of like having a frontier model for free, in some sense, if you're an MSFT shareholder." - Satya Nadella
  • The partnership gives Microsoft exclusive rights to OpenAI's "stateless APIs" on Azure until 2030 (or verified AGI), a key driver for migrating enterprise customers from competing clouds.
  • The deal's restructuring created a massive non-profit capitalized with $130 billion in OpenAI stock, initially targeting health and AI resilience.
  • Beyond equity, Microsoft gains a 7-year royalty-free license to OpenAI’s IP, allowing it to embed frontier models directly into its product suite, from M365 to GitHub.

The Great Compute Buildout

  • "The biggest issue we are now having is not a compute glut, but it's power... It's not a supply issue of chips. It's actually the fact that I don't have warm shells to plug into." - Satya Nadella
  • The primary bottleneck for AI expansion has shifted from chip supply to power and physical data center infrastructure ("warm shells"). Microsoft’s Azure growth, currently 39%, would be even higher if not for these constraints.
  • Demand for intelligence is highly elastic. Sam Altman notes that if the cost per unit of intelligence dropped 100x, usage would increase by more than 100x, unlocking currently uneconomical applications.

The Business of Intelligence

  • "High ARPU low usage, then you have a little bit of a problem. But if you are... low ARPU, high usage... use this AI as in fact an accelerant." - Satya Nadella
  • The architecture of SaaS is changing. High-usage platforms like Microsoft 365 and GitHub are well-positioned, as they generate the vast data needed to ground AI agents.
  • Enterprise monetization for AI is clear ("agents are the new seats"), but consumer monetization is "murky." The unit economics of ad-supported search don't easily translate to the higher compute costs of chat agents.

Key Takeaways

  • The AI revolution is entering a new phase defined less by model breakthroughs and more by the physical and economic realities of scaled deployment. For investors and builders, the core challenge is shifting from algorithmic discovery to securing power, building infrastructure, and redesigning business models around a new cost structure for intelligence.
  • Power is the New Bottleneck: The critical constraint in the AI race is no longer chips, but energy and the physical infrastructure to deploy them at scale.
  • Microsoft's Multi-Layered Bet: Microsoft's value from the OpenAI partnership extends far beyond its equity, encompassing exclusive Azure services, massive IP advantages, and a halo effect that pulls enterprise workloads from competitors.
  • The Golden Age of Margin Expansion: AI will enable a new productivity curve where companies grow revenue far faster than headcount, fundamentally reshaping workflows and corporate economics.

For further insights and detailed discussions, watch the full podcast: Link

This episode reveals the intricate financial and strategic architecture of the OpenAI-Microsoft partnership, detailing how the world’s largest software company and leading AI lab are structuring their multi-trillion-dollar bet on the future of intelligence.

The Restructured OpenAI-Microsoft Partnership

  • Microsoft's Investment: Microsoft has invested approximately $134 billion in OpenAI, securing a 27% fully diluted ownership stake in the for-profit entity.
  • A Unique Structure: Nadella emphasizes the novelty of the structure, highlighting the creation of what will become one of the world's largest nonprofits. He notes, "We are very proud of the fact that we were associated with the two of the largest nonprofits, the Gates Foundation and now the OpenAI Foundation."
  • A High-Conviction Bet: Altman reflects on the partnership's origins, acknowledging that Microsoft's early conviction was essential. He states that few others would have taken such a significant risk when the technology's trajectory was entirely unknown.

The OpenAI Foundation: A $130 Billion Mission

  • Initial Capitalization: The foundation is capitalized with $130 billion of OpenAI stock, making it one of the largest foundations in the world from its inception.
  • Strategic Focus: Altman explains the foundation's initial $25 billion will be directed toward health, AI security, and resilience. He clarifies that while capitalism is effective for distributing tools, the foundation will address areas where market forces are insufficient.
  • Key Initiatives: The foundation will fund work in areas like cyber defense, AI safety research, and economic studies to help society navigate the transition to a world with advanced AI. Altman also points to the potential for AI to automate scientific discovery as a primary goal.

Defining the Terms: Exclusivity and Revenue Share

  • Model Exclusivity: OpenAI will keep its "stateless APIs"—which process requests independently without storing user data between sessions—exclusively on Microsoft Azure through 2032. This means flagship models like GPT-6 will not be available on competing clouds like AWS or Google Cloud.
  • Open Distribution: Other OpenAI products, including open-source models, Sora, agents, and wearables, can be distributed on any platform, which Nadella notes is also in Microsoft's interest.
  • Revenue Share: OpenAI pays Microsoft a revenue share on all its revenues, which also runs until 2032. This revenue is recognized by Microsoft, though Nadella was unsure if it's booked directly to Azure.

The AGI Verification Clause

  • The "Jury": The agreement establishes a process where a panel of experts would be convened to determine if OpenAI has achieved AGI.
  • Diverging Timelines: Nadella expresses a more conservative view, stating on his earnings call that "nobody's even close to getting to AGI." Altman, historically more bullish, takes a pragmatic stance on the partnership's durability.
  • Strategic Alignment: Altman dismisses any potential conflict, emphasizing their continued need for Microsoft's distribution capabilities even in a post-AGI world. He states, "To say the obvious, if we had superintelligence tomorrow, we would still want Microsoft's help getting this product out into people's hands."

The $1.4 Trillion Compute Commitment

  • Aggressive Growth Projections: Altman pushes back firmly on the revenue figures and the concern, stating, "First of all, we're doing well more revenue than that... If you want to sell your shares, I'll find you a buyer." He confirms OpenAI is making a significant forward bet on steep, continued revenue growth from its core products and new ventures in AI cloud services and consumer devices.
  • Execution Track Record: Nadella provides a strong endorsement from an investor's perspective, noting, "There has not been a single business plan that I've seen from OpenAI that they have put in and not beaten it."
  • Compute as a Prerequisite for Revenue: Both leaders agree that securing compute is a necessary risk. Without it, they cannot build the next generation of models required to generate future revenue.

The Economics of Compute Scarcity

  • The Real Bottleneck: Nadella reveals that the primary constraint is no longer the supply of chips but the availability of power and data center infrastructure. "It's not a supply issue of chips. It's actually the fact that I don't have warm shells to plug into."
  • Demand and Price Elasticity: Altman explains that demand for compute is not absolute but is tied to price. A significant drop in the cost per unit of intelligence would unlock exponentially more usage and new applications, a concept related to the Jevans paradox, where increased efficiency leads to increased consumption.
  • An Inevitable Glut: Despite current scarcity, Altman predicts a future compute glut is inevitable, though the timing is uncertain. He warns that technological breakthroughs, like cheap energy or radical improvements in model efficiency, could leave those with massive, long-term infrastructure contracts severely burned.

Navigating the Regulatory Patchwork

  • The Problem with State-Level Laws: Altman expresses deep concern over laws like the Colorado AI Act, stating, "I don't know how we're supposed to comply with that... I'm very worried about a 50-state patchwork."
  • Impact on Startups: Nadella argues that while Microsoft and OpenAI can navigate the complexity, a patchwork of 50 different state laws creates an insurmountable compliance burden for startups, stifling competition.
  • A Call for Federal Preemption: Both advocate for a unified federal regulatory framework to ensure safety and provide clarity, arguing it is essential for maintaining U.S. leadership in AI.

The Future of AI: 2026 and Beyond

  • From Hours to Days: Altman is particularly excited about the evolution of Codex, OpenAI's code-generation model. He predicts that by 2026, these agents will handle multi-day coding tasks, fundamentally changing software development.
  • AI-Driven Scientific Discovery: Altman makes a bold prediction: "AI is going to make a novel scientific discovery in 2026. Even a very small one. This is a wildly important thing to be talking about."
  • A New Human-Computer Interface: Nadella describes a shift toward "macro delegation and micro steering," where users assign complex tasks to AI agents and provide high-level guidance. This will require new computing form factors beyond the smartphone.
  • The Universal Personal Assistant: The ultimate consumer use case is a personal assistant available to billions, capable of managing daily life—from ordering supplies to managing calendars—through intuitive interfaces like an earbud, freeing users from screens.

Microsoft's Strategic View: Beyond the Equity

  • The Value of Exclusivity: The exclusive access to OpenAI's stateless APIs on Azure is a powerful driver for cloud migration, attracting enterprise customers from competitors like AWS who want to build applications on top of the leading AI models.
  • Royalty-Free IP: Nadella highlights a crucial, often overlooked benefit: "Having a royalty-free access all the way till seven more years gives us a lot of flexibility business model wise. It's kind of like having a frontier model for free if you're an MSFT shareholder."
  • A Fungible Fleet: Microsoft's infrastructure strategy is not just about scaling but about building a "fungible fleet"—a flexible, efficient network of compute that can be dynamically allocated across training, inference, different geographies, and multiple chip generations to maximize utilization and returns.

The Future of Software and Unit Economics

  • The New SaaS Architecture: He argues that the old model of tightly coupled data, business logic, and UI is being replaced by an "agent tier." Success in this new paradigm depends on data gravity. Companies with high-usage, low-cost-per-user platforms (like Microsoft 365) are best positioned because they generate the vast data needed for grounding AI agents.
  • Search vs. Chat Economics: The conversation contrasts the highly profitable unit economics of traditional search (amortizing a fixed-cost index) with AI chat (high marginal GPU cost per query). This shift suggests the consumer internet's economic model, built on search advertising, is facing a fundamental disruption.
  • The Golden Age of Margin Expansion: Nadella believes AI will usher in a new era of productivity. For Microsoft, this means headcount will grow much slower than revenue, as AI agents automate workflows and provide employees with unprecedented leverage.

This discussion underscores that the OpenAI-Microsoft alliance is a deeply integrated effort to build and control the entire AI value chain. For investors and researchers, the key takeaway is that future opportunities lie not just in model performance but in the complex interplay of compute infrastructure, energy, data gravity, and the evolving unit economics of intelligent software.

Others You May Like