Latent Space
August 15, 2025

Greg Brockman on OpenAI's Road to AGI

OpenAI President and Co-founder Greg Brockman dives into the paradigm shifts behind GPT-5, the central role of compute in the quest for AGI, and how AI agents are poised to become our ultimate coworkers.

The Reasoning Leap

  • "I remember we had this question... why is this not AGI? This model clearly is not AGI, but it's really hard to describe why... And the most obvious thing you need to do is actually have it test out its ideas in the world, right? Actually do reinforcement learning."
  • "With GPT-4, the ideas that it produced were not very deep... [With GPT-5] we've seen physicists... say that this model was able to re-derive an insight that took me many months worth of research to produce."
  • The key gap between GPT-4’s impressive capabilities and true AGI was its inability to test hypotheses and learn from feedback. The move to a reinforcement learning (RL) paradigm, where models try things out, get feedback, and become more reliable, was the crucial next step.
  • GPT-5 marks a shift from a broadly capable tool to a genuine intellectual partner. It can solve problems requiring deep intelligence, demonstrated by its ability to replicate months of human research in moments and achieve gold-medal performance in both Math (IMO) and Informatics (IOI) Olympiads.

Compute as Crystallized Intelligence

  • "The bottleneck is always compute, right? And I mean that in a real way... If you give us a lot of compute... we will find ways to iterate that actually make the most of that compute."
  • "I think of compute as this refining process. Ultimately you start with energy, it turns into compute, it turns into intelligence. It's almost crystallizing that compute into potential energy that can be converted into the model doing something useful."
  • Compute is the fundamental fuel for intelligence. Brockman frames it as a process of turning kinetic energy into the "potential energy" stored within a trained model's weights, which can then be unleashed on countless tasks.
  • The journey to more capable AI is a story of scaling. OpenAI found that classic RL algorithms, once thought unscalable, performed exceptionally well when given massive compute, proving that engineering and scaling existing methods is often the most direct path to breakthroughs.

The Coworker Paradigm

  • "I tend to think about AI productization by analogy to a coworker... You want the pair [programming] form factor. You also want the remote async form factor. And you want it to be one entity that has knowledge and memory across all of this."
  • The future of AI interaction isn't just a better chatbot; it's an AI coworker. This entity will be available for both real-time, over-the-shoulder collaboration and asynchronous, remote tasks, all while maintaining a persistent memory and context.
  • AGI won't be a single, monolithic model. The emerging architecture is a "manager of models" that intelligently orchestrates a fleet of agents—combining small, fast, local models with large, powerful, cloud-based reasoners to achieve adaptive compute.

Key Takeaways:

  • The path to AGI is paved with reinforcement learning and an insatiable demand for compute. GPT-5 represents a step-change from a useful tool to an intellectual partner, and the future of human-AI interaction will be through versatile agents that act as persistent, multi-modal coworkers.
  • AGI is a Compute Game. The primary bottleneck is compute. The process is one of "crystallizing" energy into compute, then into the potential energy of a trained model. More compute means more intelligence.
  • The Future is a "Manager of Models." AGI won't be a single entity. It will be an orchestrator that delegates tasks to a fleet of specialized models, from fast local agents to powerful cloud reasoners.
  • Build for Your AI Coworker. To maximize leverage, structure codebases for AI. This means self-contained modules, robust unit tests, and clear documentation—treating the AI as a team member, not just a tool.

For further insights, watch the full discussion: Link

This episode reveals the strategic shift from static, pre-trained models to dynamic, reasoning-driven agents, detailing how OpenAI is architecting the future of AI by tackling the compute bottleneck and embedding models into the real world.

The Dawn of the Reasoning Paradigm

  • Greg Brockman, President of OpenAI, kicks off the discussion by contextualizing the recent releases of GPT-OSS and GPT-5. He frames GPT-5 as the first hybrid model, a culmination of years of work that began after the launch of GPT-4. The central question driving this evolution was: "Why is this not AGI?"
  • The team, including key figures like Ilya Sutskever and Wojciech Zaremba, identified a critical gap: the model's inability to test ideas in the real world and learn from feedback. This realization marked the beginning of a concerted push toward a reasoning paradigm, leveraging reinforcement learning to build more reliable and capable systems.
  • Reinforcement Learning (RL): A type of machine learning where an agent learns to make decisions by performing actions in an environment to achieve the maximum cumulative reward. This is a shift from simply predicting the next token based on static data.
  • Brockman emphasizes that this journey was built on conviction, as the first ten attempts at new methods typically fail. The success came from persistent iteration and identifying small "signs of life" in their experiments.
  • He notes that while GPT-4 could chat, it lacked reliability. The goal was to infuse it with the correctness and sophistication seen in earlier RL projects like the Dota agent, which learned complex behaviors from a randomly initialized state.

From Offline Pre-Training to Online Learning

  • The conversation explores the fundamental shift in how AI models learn, moving from a static, offline pre-training phase to a more dynamic, online learning loop. Greg Brockman highlights that while current models are not yet fully "online" in the way humans are, they are increasingly learning from the data they generate during inference.
  • Ilya Sutskever's insight is shared: as models become more capable, the value of each token they generate increases significantly. RL capitalizes on this by training the model on its own high-value, reality-tested outputs.
  • This new paradigm changes the scale of data required. While pre-training needs hundreds of thousands of examples for a behavior, RL allows models to learn sophisticated behaviors from just 10 or 100 human-curated tasks, creating immense leverage.
  • Actionable Insight: The move toward online learning signals a future where AI systems continuously adapt and improve from real-world interaction. Investors should track companies developing infrastructure for efficient RL and data feedback loops, as this will be a critical component of next-generation AI.

Compute: The Unwavering Bottleneck

  • When asked about the primary bottleneck in AI development, Brockman's answer is unequivocal: compute. He argues that with more compute, OpenAI can always find ways to iterate and improve model performance.
  • He draws a parallel to the Dota project, where the team scaled up the existing PPO (Proximal Policy Optimization) algorithm, an RL technique, against the common belief that it wouldn't scale. They kept doubling the cores, and performance consistently improved, proving that many perceived algorithmic "walls" are actually engineering bugs or solvable issues.
  • "The journey of that scaling that is the interesting stuff," Brockman states, emphasizing that the engineering challenges of scaling are where true progress is made.
  • He describes compute as a fundamental fuel being crystallized into intelligence, turning energy into potential energy stored within the model's weights. This potential can then be amortized across countless applications.

Generalization and the Path to Real-World Application

  • The discussion highlights the remarkable and sometimes "unreasonable" generalization capabilities of current models. The same foundational techniques that achieved a gold medal in the International Mathematical Olympiad (IMO) also secured a gold in the International Olympiad in Informatics (IOI) with minimal adaptation.
  • This demonstrates that learning to solve hard problems is a highly transferable skill for AI.
  • However, Brockman acknowledges the limits of generalization. A model without experience in a domain, like running a physics experiment, won't magically become an expert. It needs real-world interaction and data from that domain.
  • He shares an anecdote about wet lab scientists using GPT-4o, who found that one out of five AI-generated hypotheses for an experiment would work, leading to results publishable in a mid-tier journal. This underscores the current utility and the clear path for improvement.

Characterizing the GPT-5 Era: The Rise of Intelligent Agents

  • Greg Brockman defines the GPT-5 era by one word: "smart." He describes its intelligence as "almost indescribable," capable of performing great intellectual feats that were previously out of reach.
  • While GPT-4 was commercially useful, its ideas were not particularly deep. GPT-5, in contrast, can re-derive insights that took human researchers months to produce, positioning it as a true intellectual partner.
  • Strategic Implication: The primary value of GPT-5 lies in tackling complex, high-stakes problems. Researchers and developers should focus on applications that require deep reasoning, not simple chat. The model's performance saturates on easy tasks but excels when pushed with difficult intellectual challenges.
  • To extract maximum value, Brockman advises users to develop a skill of "tenacity," testing the model's limits, breaking down tasks, and managing multiple instances of the model like a team of agents.

The Future of AI-Powered Development

  • The conversation shifts to the practical application of AI in software engineering, envisioning a future where AI agents are seamlessly integrated into developer workflows.
  • Brockman uses the analogy of a human coworker: you want an AI that can work asynchronously in the background (like a remote agent) but also pair-program over your shoulder (like an in-IDE agent).
  • Crucially, this should be a single, persistent entity with memory, not a "junior programmer who shows up every day being like, 'Okay, I forgot everything.'"
  • Agent Robustness: OpenAI is tackling agent safety through a "defense in depth" strategy. This includes techniques like Instruction Hierarchy, where the model learns to prioritize system and developer commands over potentially malicious user inputs, functioning similarly to how operating systems use security rings.

The GPT-5 Router and the "Manager of Models"

  • Brockman confirms that GPT-5 uses a router to switch between a powerful reasoning model and a faster, non-reasoning model. This is a practical implementation of the "manager of models" concept, where different specialized models are orchestrated to handle tasks efficiently.
  • The router considers factors like conversation complexity, tool usage, and rate limits to make its decision. This internalizes complexity, simplifying the user experience.
  • Strategic Insight: The future of AGI is unlikely to be a single monolithic model. Instead, it will be a system of composable, specialized models. This architecture allows for "adaptive compute," using the most efficient resource for each specific task. This trend is critical for researchers to understand when designing their own AI systems.

Open Source Strategy and American Leadership

  • The release of GPT-OSS is framed as a strategic move to establish an American tech stack in the open-source AI ecosystem. By providing powerful, open models, OpenAI encourages developers to build on their technology.
  • Brockman argues that this creates a positive dependence, ensuring that the ecosystem evolves on models that reflect American values and can integrate with American cloud infrastructure and hardware.
  • The architectural choices for GPT-OSS, such as a fine-grained Mixture of Experts (MoE)—a technique where only a subset of the model's parameters are used for any given input, increasing efficiency—were driven by practical engineering constraints for single-machine deployment.

Engineering in the Age of AI

  • The discussion concludes with reflections on how AI is changing software engineering and the nature of research.
  • To maximize leverage from AI, codebases should be structured with self-contained modules, strong unit tests, and clear documentation, playing to the models' strengths.
  • Brockman believes the value of engineers is increasing. AI tools don't just make existing work more efficient; they enable teams to "do 100x more things," from cleaning up tech debt to tackling previously impossible projects.
  • He describes the current AI development effort as a project whose scale dwarfs historical undertakings like the Apollo program, driven by the immense economic and societal potential.

Conclusion

This episode underscores that the road to AGI is paved with dynamic, real-world learning, not just static pre-training. For investors and researchers, the key takeaway is to focus on the infrastructure enabling this shift—compute, RL frameworks, and robust agentic systems—as these will define the next frontier of AI development.

Others You May Like