Machine Learning Street Talk
September 4, 2025

The Day AI Solves My Puzzles Is The Day I Worry (Prof. Cristopher Moore)

Professor Cristopher Moore of the Santa Fe Institute, a physicist turned computer scientist, explores the deep chasm between the theoretical hardness of problems and the messy, structured reality where AI operates, questioning whether current models can ever truly replicate human insight.

The Real World's Unfair Advantage

  • "Real-world data is not designed by an adversary to be as tricky as possible. And it's very far from random. It has all kinds of structure that both human intelligence and animal intelligence and artificial intelligence can exploit."
  • "What's fascinating is that that real-world structure seems very hard to mathematize. How do we talk about that structure?"
  • While computer science proves problems are hard using adversarial examples, real-world data is rich with exploitable patterns. This structure is why machine learning works at all, yet we lack the mathematical language to formally describe it.
  • Drawing from physics, Moore describes a middle ground of problems where an answer exists but is hidden behind an "energy barrier." Our best algorithms, like gradient descent, get stuck in a "mush of inaccurate fits," unable to find the truth without an exhaustive search.

The Puzzle-Solver's Mind

  • "Humans are quite good at designing on the fly different kinds of partial knowledge or partial solution to a problem...inventing on the fly what kind of variable you should use to address the problem."
  • "I'm proud to say eight of my puzzles are in that data set. So, I'm waiting to see if AI can solve my puzzles."
  • AI currently fails at complex human-designed puzzles (like modern Sudoku variants) because it can’t fluidly absorb new rules and apply them in a 2D spatial context. It lacks our intuitive “sensorium.”
  • The essence of human problem-solving is inventing new ways to see the problem. We don’t just work with given variables; we create new ones, like coloring cells that must be the same, a process Moore calls a fascinating frontier for AI.

From Universal Machines to Societal Trust

  • "I am also a finite state machine... The difference is that I am also a tool-using and tool-making entity. If there's a problem which is difficult for me to do in my head... I can then build things... that extends my workspace, extends, if you will, the tape of my Turing machine."
  • "I don't think we should be using opaque proprietary tools to make decisions that affect people's fundamental human rights."
  • The debate over whether AI needs to be "Turing complete" misses the point. Humans are finite, but we achieve universality by extending our minds with external tools like pen and paper. The key is knowing when to build a new tool.
  • When AI is deployed in society, especially in the justice system, accuracy is not the only goal. We explicitly value procedural fairness. Demanding transparency for these systems isn't just a technical preference; it's a defense of our relationship with the state.

Key Takeaways:

  • The next leap in AI won't come from bigger models, but from systems that can fluidly change their own problem representations and leverage external tools. True intelligence is about transforming hard problems into simpler ones, not just crunching the numbers on the original.
  • Stop Confusing Hardness with Reality. Theoretical computer science focuses on worst-case scenarios. Real-world success hinges on exploiting messy, latent structure that we can’t even formally define yet.
  • Intelligence is Tool-Making. Humans aren't just powerful processors; we're tool-users who extend our cognitive workspace. AI will remain limited until it can recognize its own limitations and build the tools it needs to overcome them.
  • Demand Transparency Over Explainability. For high-stakes decisions like criminal justice or medical diagnoses, proprietary black boxes are unacceptable. The right to confront your accuser extends to the algorithms that judge you.

For further insights and detailed discussions, watch the full podcast: Link

This episode reveals that AI's path to true intelligence isn't about more compute, but about mastering abstraction and structured reasoning—a critical insight for investors betting on the next frontier of AI architectures.

Introduction to Christopher Moore & Computational Hardness

  • Professor Christopher Moore of the Santa Fe Institute opens by explaining his background at the intersection of physics and theoretical computer science, where he focuses on proving the inherent difficulty of certain problems. This perspective, which he describes as being a "frog" who enjoys concrete examples rather than a "bird" with a high-level view, frames the entire discussion. Moore's expertise lies in understanding computational limits, a crucial lens for evaluating the capabilities and hype surrounding modern AI.
  • He enjoys the two-sided nature of computer science: you can either search for an efficient algorithm to solve a problem or, if you fail, try to prove that the problem is fundamentally hard.
  • This concept of hardness is central to fields like cryptography, where the difficulty of a problem can be leveraged to build secure systems.
  • Moore's approach is grounded in tangible examples and rigorous proofs, providing a sharp contrast to the often purely empirical and black-box nature of large-scale AI development.

"I really like concrete examples. I like things I can visualize that I can hold in my hand... I'm constantly touching down and measuring the steps of that proof against my favorite examples to understand why they work."

The Limits of Transformers and Real-World Data

  • Moore argues that traditional computer science proofs of "hardness" often don't apply to real-world AI applications. These proofs typically rely on adversarial examples designed to be as difficult as possible, whereas real-world data is highly structured, not random or adversarial. This structure is what allows AI models to succeed, often in ways that theory alone cannot predict.
  • He introduces phase transitions, a concept from physics, as a more relevant framework for understanding AI performance. Just as a magnet loses its magnetism above a critical temperature, an algorithm can fail to find an underlying pattern in data if the noise-to-signal ratio is too high.
  • There exists a challenging middle ground where a solution is theoretically findable through exhaustive search, but common algorithms like gradient descent get stuck for an exponential amount of time.
  • Strategic Insight: This suggests that simply scaling models may not overcome certain fundamental problem structures. Investors should look for research into architectures that can better navigate these complex, "glassy" landscapes rather than relying solely on brute-force computation.

The Challenge of Mathematizing Real-World Structure

  • While AI models are adept at exploiting the structure in real-world data, Moore points out that we lack the mathematical language to formally describe this structure. The world is not just correlations; it's a rich hierarchy of objects, parts, and relationships. He posits that LLMs and transformers may ultimately serve as tools to help us create this new mathematics.
  • The success of current architectures is less about the specific model and more about their ability to capture the inherent, yet unformalized, structure of the world.
  • Moore believes that in the future, we will learn as much about the world's structure from successful AI models as we learn about the models themselves.

AI, Coding, and the Puzzle Benchmark

  • The conversation shifts to a concrete benchmark for AI reasoning: complex, human-designed puzzles. Moore discusses his work designing modern Sudoku variants for the YouTube channel Cracking the Cryptic. These puzzles require absorbing novel rules and applying multi-step, non-obvious logic—a task where current AI performs poorly.
  • He mentions that the AI company Sakana AI is compiling a dataset of these puzzles to test AI's ability to read rules in English and solve them. So far, the results have been "pitiful."
  • Moore attributes this failure to the models' reliance on one-dimensional text, which struggles with the two-dimensional, spatial reasoning humans use to scan rows, columns, and boxes.
  • Actionable Insight: For researchers, puzzle-solving represents a powerful benchmark for evaluating progress in genuine reasoning beyond pattern recognition. The day an AI can solve Moore's custom puzzles is a significant milestone to watch for.

The Art of Insight and Creative Problem-Solving

  • Moore delves into the human process of solving these puzzles, which is not about brute-force search but about discovering "aha moments." This involves a fluid, creative process of inventing new forms of partial knowledge on the fly—a capability that current AI lacks.
  • Humans don't just work with predefined variables; they invent new ways to represent the problem, such as identifying that two unknown cells must be the same and coloring them identically.
  • He compares this process to scientific discovery and mathematical modeling, where 90% of the work is the mathematization—figuring out the right way to formalize the problem.
  • Strategic Implication: The next leap in AI may not come from bigger models, but from systems that can autonomously re-frame problems and invent their own variables and constraints, mimicking human creativity.

Grounding, Meaning, and the Limits of Statistical Models

  • The discussion addresses a core weakness of LLMs: their lack of grounding in real-world meaning. Moore recounts an anecdote where an LLM summarized an essay by reducing it to the most common arguments on the topic, completely missing the author's unique, subtle point.
  • This "blandification" occurs because the models operate on the statistical distribution of text, not on an understanding of the concepts the text refers to.
  • He contrasts this with human cognition, where we share a perceived reality. When two people imagine a cube, they are perceiving the same platonic object with 8 corners and 12 edges.
  • Moore expects AI capabilities to improve significantly once they can utilize specialized "workspaces"—like a 3D modeling environment or a formal proof checker—to ground their reasoning, much like a human uses a piece of paper to work through a math problem.

Creativity, Cliche, and the Artistic Process

  • Drawing on novelist Martin Amis's concept of "The War Against Cliche," Moore argues that true creativity involves deliberately avoiding high-probability word combinations. This is the opposite of how LLMs, which are trained to predict the most likely next token, fundamentally operate.
  • Creative professionals, like artists and writers, don't just generate content; they engage in a rigorous loop of creation, reflection, and refinement, holding their work to a high standard.
  • This process requires a deep mental model of the audience, carefully guiding them along a path of "progressive disclosure of complexity" to create a satisfying experience without being boring or overwhelming.

Computational Irreducibility and Undecidability

  • The conversation explores computational irreducibility, a concept popularized by Stephen Wolfram. This principle states that for certain complex systems, there are no analytical shortcuts to predict their future state; the only way to know what happens next is to run the simulation step-by-step.
  • This is analogous to chaotic systems like the three-body problem, where you have to "do the work" of computation rather than plugging values into a simple formula.
  • Moore explains that our only method for proving a system is computationally irreducible is to demonstrate that you can build a universal computer (like a Turing machine) out of its components, as was done for Wolfram's Rule 110.
  • Many natural systems appear irreducible but are too chaotic to build a controlled computer from, placing them in a fascinating middle ground between simple predictability and universal computation.

Are Transformers Turing Complete? The Debate on AI Architectures

  • Moore addresses the critical debate on whether current AI models, often described as Finite State Automata (FSAs)—systems with a fixed number of states and no capacity for unbounded memory—are fundamentally limited. The counterargument is that true general intelligence requires Turing completeness, the ability to perform any computation, including recursion.
  • Moore offers a nuanced view: while he is technically a finite state machine (with a finite number of neurons), his intelligence comes from his ability to use and create tools (like a pencil and paper) to extend his memory and workspace, effectively giving him access to recursion.
  • He argues that the key is not the intrinsic architecture but this extensibility. The next generation of AI will need to learn to identify when a problem requires a recursive data structure (like a stack) and then be able to create and use one.
  • Strategic Insight: This shifts the research focus from building inherently Turing-complete models to creating AIs that can dynamically augment their own capabilities by interfacing with external computational tools and data structures.

The Computational Lens and Pan-computationalism

  • When asked if he is a pan-computationalist—one who believes the universe is fundamentally a computer—Moore prefers to view computation as one of several "lenses" for understanding a system. This lens focuses on the storage, transmission, and transformation of information.
  • He uses the origin of life as an example: is life a "replicator-first" phenomenon (driven by information in DNA) or a "metabolism-first" one (driven by the physical need to channel energy)? Both lenses offer valuable, non-exclusive insights.
  • While you can say the planets are "computing" their orbits, this lens is less informative than when applied to a cell's ribosome, which is clearly an information-processing device.
  • Ultimately, he is inclined to believe the physical Church-Turing thesis: that any process in the universe can be simulated by a computer, implying a finite amount of computation occurs in any given volume of spacetime.

Algorithmic Justice and the Need for Transparency

  • In the final segment, Moore makes a powerful case for transparency in AI systems, especially those making consequential decisions about people's lives and rights. He argues that inscrutable, proprietary black boxes are unacceptable in domains like criminal justice.
  • He draws a distinction: a black-box movie recommender is acceptable, but an AI used for pre-trial detention or DNA analysis demands full transparency. The legal system values procedural fairness, not just accuracy.
  • Moore advocates for a standard where citizens can "confront their accuser," which means being able to inspect, contest, and cross-examine the evidence and logic used by an AI system.
  • Actionable Insight: Investors and developers in the Crypto AI space must anticipate growing regulatory pressure for transparency and auditability. Building systems that are interpretable by design is not just an ethical imperative but a crucial strategy for mitigating future legal and market risk.

Conclusion

This discussion highlights that AI's next frontier lies in mastering abstraction and structured reasoning, not just scaling compute. For investors and researchers, the key takeaway is to prioritize architectures that can move beyond statistical pattern-matching toward genuine, creative problem-solving and formalization, as these will define true artificial intelligence.

Others You May Like