a16z
October 29, 2025

Building the Real-World Infrastructure for AI, with Google, Cisco & a16z

This is not your father’s infrastructure cycle. Leaders from Google and Cisco get real about the unprecedented physical buildout powering the AI revolution, revealing the hard constraints and architectural shifts that will define the next decade of computing.

The 100x Infrastructure Boom

  • "The internet in the late 90s, early 2000s was big... This makes it, I mean 10x is an understatement. It's 100x what the internet was."
  • "This is like the combination of the buildout of the internet, the space race, and the Manhattan project all put into one, where there's a geopolitical implication, an economic implication, and a national security implication."
  • The current AI infrastructure buildout is an order of magnitude larger than the dot-com boom, with demand far outstripping even the most aggressive projections.
  • Unlike previous cycles, this boom is driven by geopolitical and national security pressures, making it a global strategic imperative.
  • Despite bubble concerns, the reality on the ground is a severe underestimation of need. Even Google’s 7- and 8-year-old TPUs have 100% utilization, signaling insatiable demand.

Power: The Ultimate Bottleneck

  • "We're limited by power, we're limited by transforming land, we're limited by permitting... I worry that the supply isn't actually going to catch up to the demand as quickly as we'd all like."
  • "Data centers are being built where the power is available rather than power being brought to where the data centers are."
  • The primary constraints on AI are no longer just silicon; they are physical-world realities like power, land, and permits. This supply crunch could extend for another 3-5 years.
  • This power scarcity is forcing a fundamental architectural shift. Instead of centralizing data centers, companies are building distributed "scale-across" networks where facilities up to 900km apart can function as a single logical unit.
  • Every watt matters. If power is the constraint and compute is the asset, networking has become the force multiplier—every kilowatt saved moving a packet is a kilowatt that can go to a GPU.

Reinventing the Stack: From Silicon to Software

  • "We're really seeing the golden age of specialization... A TPU... is somewhere between 10 and 100 times more efficient per watt than a CPU. That's hard to walk away from."
  • "Assume that these tools are going to get infinitely better within 6 months... get your mental model to where that tool is going to be in six months rather than assessing it for where it is today."
  • We are entering a "golden age of specialization," moving away from general-purpose hardware. The entire computing stack, from custom silicon to networking and software, will be unrecognizable in five years.
  • At major tech companies, AI is already delivering massive productivity gains. Google used AI-assist to accelerate its massive TensorFlow-to-JAX migration by "integer factors faster."
  • The pace of AI improvement requires a cultural reset. Engineers can no longer dismiss a tool that doesn’t work; they must re-evaluate it every four weeks, not every six months.

Key Takeaways

  • The Physical World is AI's Final Boss: The speed of AI progress is now governed by the speed of transformers, permits, and power plants. The biggest opportunities are in solving these hard, physical-world bottlenecks.
  • Specialization is the Only Game in Town: General-purpose is dead. Lasting value will be created through specialized hardware, co-designed software, and tightly integrated systems that optimize for performance-per-watt.
  • Founders, Ditch the Thin Wrappers: The most durable businesses will not be built on other companies' models. Instead, they will create deep, proprietary feedback loops where the product and the model improve each other.

For further insights and detailed discussions, watch the full podcast: Link

This episode reveals the unprecedented scale of the AI infrastructure buildout, a convergence of geopolitical, economic, and technological forces dwarfing the original internet boom by orders of magnitude.

The Unprecedented Scale of the AI Infrastructure Boom

  • The current AI infrastructure cycle is unlike any previous technological shift. Amin Vahdat of Google frames the scale as immense, stating, "The internet in the late 90s, early 2000s was big... This makes it I mean 10x is an understatement. It's 100x what the internet was."
  • Jeetu Patel from Cisco adds a geopolitical and national security dimension, comparing the current moment to a combination of three historic undertakings.
    • A Unique Convergence: Jeetu describes the situation as "the combination of the buildout of the internet, the space race, and the Manhattan project all put into one." This highlights the profound economic, national security, and speed implications driving the buildout.
    • Underestimated Demand: Despite concerns about a bubble, both speakers agree that the market is grossly underestimating the long-term infrastructure demand. The need for compute, power, and networking will far exceed current projections.

Capex Cycle and Demand Signals

  • The conversation confirms we are still in the early stages of the capital expenditure cycle, with demand far outstripping supply.
  • Insatiable Demand: Amin notes that even Google's 7- and 8-year-old TPUs (Tensor Processing Units)—Google's custom-designed accelerators for AI workloads—are running at 100% utilization. This signals a desperate need for any available compute, regardless of its generation.
  • Supply Chain as the Bottleneck: The primary constraint is not capital but the physical world. Amin highlights that the industry's ability to spend is limited by power availability, land acquisition, permitting, and supply chain delivery for physical components. He projects this supply-demand imbalance will persist for the next 3-5 years.
  • Strategic Implication: For investors, this signals that the core bottleneck—and thus a key area for opportunity—lies in solving physical constraints like power generation, data center construction, and advanced cooling, rather than just in silicon or models.

The Evolving Data Center Architecture

  • Power scarcity is fundamentally reshaping data center strategy and network architecture.
  • Power-Driven Placement: Jeetu explains that data centers are now being built where power is available, rather than bringing power to ideal locations. This geographic distribution is creating new architectural challenges.
  • New Networking Paradigms: This decentralization drives demand for three distinct networking types:
    • Scale-up: Increasing networking capacity within a single rack.
    • Scale-out: Connecting multiple racks and clusters within a data center.
    • Scale-across: Creating a single logical data center from two physical sites that could be up to 900 kilometers apart, a necessity driven by power constraints.

The Future of Computing: Mainframes vs. Scale-Out

  • The speakers debate whether the industry is returning to a mainframe-like, integrated system model (popularized by Nvidia) or continuing the scale-out commodity hardware approach pioneered by Google.
  • Software and Hardware Co-Design: Amin argues that the future is not a simple return to mainframes. He emphasizes that the next five years will see a complete reinvention of the computing stack, driven by the tight co-design of hardware and software. He states, "5 years from now whatever the computing stack is from the hardware to the software right is going to be unrecognizable."
  • The Need for an Open Ecosystem: Jeetu stresses the importance of deep design partnerships between multiple companies. To avoid vendor lock-in and loss of efficiency, the industry must "work like one company even though we might actually be multiple companies." This points to a future of integrated systems built on open, collaborative standards.

The Golden Age of Specialized Processors

  • The discussion moves beyond the current market leader to a future defined by architectural diversity and specialization.
  • Efficiency as the Driver: Amin highlights that a TPU can be 10-100x more power-efficient than a CPU for specific AI tasks. This dramatic efficiency gain makes specialization necessary to manage power consumption and cost.
  • Accelerating Design Cycles: A major bottleneck is the 2.5-year "speed of light" cycle to bring a new specialized chip from concept to production. Shortening this cycle is a critical area for innovation.
  • Geopolitical Architectures: Jeetu introduces a fascinating geopolitical angle. China, with access to 7nm chips but abundant power, may optimize with engineering resources. The U.S., with access to 2nm chips but power constraints, will pursue a different architectural path focused on extreme power efficiency. This could lead to regionally divergent AI hardware ecosystems.

Reinventing Networking for AI Workloads

  • Networking is becoming a primary performance bottleneck, requiring a fundamental redesign to handle the unique demands of AI.
  • Known Communication Patterns: Amin points out a massive opportunity: unlike general-purpose networking, AI training workloads have predictable communication patterns. This allows for network optimization beyond what a standard packet-switched network can offer, potentially using designs closer to circuit switching.
  • The Challenge of Bursty Workloads: AI workloads are incredibly bursty, creating massive power spikes followed by idle periods. Building a network that can operate at 100% capacity for a short duration and then efficiently power down is an unsolved and critical engineering problem.
  • Network as a Force Multiplier: Jeetu frames the network as a "force multiplier." Every kilowatt of power saved in moving a packet is a kilowatt that can be allocated to a GPU, directly improving computational output.

The Architectural Split: Training vs. Inference

  • The speakers clarify that training and inference have different architectural needs, which will lead to more specialized infrastructure.
  • Different Optimization Goals: Training workloads are often optimized for latency, while inference workloads are optimized for memory.
  • Inference-Native Infrastructure: Jeetu predicts the rise of "inferencing native infrastructure" rather than simply repurposing training hardware. This includes specialized hardware configurations and software stacks.
  • Prefill vs. Decode: Amin adds another layer of complexity, noting that the two phases of inference—prefill (processing the initial prompt) and decode (generating the response token by token)—have different hardware balance points, creating an opportunity for further hardware specialization.

Internal AI Adoption: Wins and Challenges

  • Google and Cisco are using AI extensively to boost internal productivity, offering a blueprint for other enterprises.
  • Code Migration and Debugging: Both companies report significant success using AI for code migration and debugging. Amin shares a powerful example: a migration from Google's Bigtable to Spanner was estimated to take "seven staff millennia" manually, but similar migrations are now being done with AI assistance in a fraction of the time.
  • The Cultural Reset: The biggest challenge is not technology but culture. Jeetu emphasizes the need for a mental shift, urging his engineers to re-evaluate AI tools every four weeks, not every six months. He advises, "Assume that these tools are going to get infinitely better within 6 months... and make sure that you get your mental model to where that tool is going to be."
  • Broad Applications: Beyond coding, Cisco is seeing strong results in sales preparation, legal contract review, and product marketing, where AI provides a superior starting point than a blank slate.

Forward-Looking Insights for Founders

  • The episode concludes with actionable advice for founders and researchers building in the AI space.
  • Avoid Thin Wrappers: Jeetu warns startups not to build simple wrappers around third-party foundation models. Durable businesses will require a tight feedback loop where the model and product improve together, necessitating proprietary models or deep, defensible integrations.
  • The Rise of Agents and Multimodality: Amin predicts that in the next 12 months, agents built on top of models will become "scary good," enabling complex, long-running tasks. He also forecasts that image and video modalities will mature rapidly, becoming transformative productivity and educational tools, not just creative toys.

Conclusion

This conversation underscores that the AI revolution is a full-stack infrastructure challenge. Investors and researchers must look beyond models to the foundational layers of power, silicon, and networking, where geopolitical pressures and the drive for specialized, power-efficient architectures will define the next wave of innovation and value creation.

Others You May Like