Delphi Digital
March 31, 2025

Ben Fielding: Gensyn’s Fueling an AI-Native Internet, Open vs. Closed Source AI and RL Swarm

Ben Fielding, co-founder of Gensyn, dives deep into the future of AI infrastructure, arguing that the current centralized, vertical scaling approach is hitting diminishing returns and paving the way for a decentralized, horizontally scaled, AI-native internet.

Gensyn's Agnostic Infrastructure Bet

  • "We made a decision really early on to build low-level infrastructure technology that was agnostic to the changes and whims and trends of machine learning."
  • "We focused really, really low level... as low as we can go, agnostic to the whims and changes of the market."
  • Gensyn avoids betting on specific AI models or tasks, instead focusing on fundamental, unchanging operations like matrix multiplications.
  • This low-level focus avoids the sunk-cost fallacy plaguing hardware companies tied to specific architectures (like Transformers).
  • Gensyn provides three core components: consistent execution across diverse hardware, standardized peer-to-peer communication for models, and trustless verification of computation via cryptographic proofs.

Vertical vs. Horizontal Scaling: AI's Awkward Teen Phase

  • "We've hit that diminishing returns point [with vertical scaling]... It's mostly a marketing reason at this point."
  • "[AI is] in that weird redesign period where we vertically scaled... diminishing returns, super expensive, we need to move to that new redesign [horizontal scaling]."
  • Current hyperscale data center buildouts (vertical scaling) face mounting constraints: power delivery, cooling, land availability, geographical limitations, and specific hardware dependencies.
  • Gensyn facilitates horizontal scaling, allowing compute resources (from laptops to data centers) to be aggregated efficiently, analogous to how Bitcoin mining opened up energy logistics globally.
  • This shift unlocks potentially 100x scale compared to the incremental gains of expensive vertical scaling, though it requires a redesign period.

The AI-Native Internet: Parameters as the Base Layer

  • "Our view is that the internet itself is changing; its base data type is becoming parameters."
  • "Every single interaction you have with technology... if that itself is able to be controlled and has biases in it, you will very quickly be influenced as a person by those biases without even realizing it's happening."
  • The future internet won't just serve static data; it will be dynamic, with interfaces and experiences generated on-the-fly by models understanding user context, moving from databases to parameter space.
  • This necessitates constant, ubiquitous ML execution, which Gensyn's infrastructure enables across all devices.
  • Centralized control over these foundational models risks embedding hidden biases, influencing users subtly but pervasively, mirroring and amplifying the issues seen with centralized social media.

RL Swarm: Decentralized Learning & Communication

  • "RL Swarm is taking that concept [reinforcement learning post-training] and allowing models to also communicate with each other and critique each other's answers with the goal of improving together as a swarm."
  • RL Swarm allows AI models to engage in peer-to-peer reinforcement learning, using explain/critique mechanisms to learn communication strategies and collectively improve.
  • Models fine-tune their weights based on swarm interactions, enabling continuous, decentralized improvement beyond waiting for centralized model updates.
  • Gensyn's testnet adds persistent on-chain identities for models participating in swarms, enabling tracking, reputation, and future market mechanisms.

Key Takeaways:

  • The AI race isn't just about bigger models; it's about the underlying infrastructure and who controls it. Decentralized networks like Gensyn offer a path to a more open, efficient, and less biased AI future.
  • AI scaling hits physical limits: Centralized hyperscalers face diminishing returns; the future needs horizontally scalable, decentralized compute enabled by protocols like Gensyn.
  • The Internet gets personal (and probabilistic): Expect a shift from static databases to dynamic, parameter-based experiences, requiring ubiquitous, verified ML execution.
  • Open beats closed (eventually): Open-source models and decentralized learning (like RL Swarm) will likely outpace closed systems by leveraging global compute and diverse data, mitigating centralized bias risks.

Link: Link

This episode dissects the strategic shift from centralized AI compute towards decentralized networks, exploring Jensen Network's vision for a future where AI operations are ubiquitous, trustless, and horizontally scalable.

Navigating AI Volatility: Jensen's Foundational Strategy

  • Ben, co-founder of Jensen Network, explains their deliberate strategy, developed over five years, to build foundational, low-level infrastructure agnostic to the rapid shifts in AI research and application trends. Witnessing the evolution from early computer vision models post-AlexNet to RNNs (Recurrent Neural Networks, once common for sequential data like text) and now Transformers (the dominant architecture for large language models), Jensen anticipated continued volatility.
  • Instead of betting on specific model architectures or tasks, which risks obsolescence (akin to the "graveyard of companies" building specialized ASICs like Transformer-specific chips), Jensen focused on the enduring core operations. Ben notes, "we looked at that whole space and we said okay Matrix multiplications haven't changed they're not really going anywhere... at the end of the day what's operating on Parallel processes is going to be very consistent."
  • This first-principles approach involves tackling harder, lower-level problems like verifying computation on untrusted hardware without relying on model-specific tricks. While more complex initially, it provides long-term resilience against market whims.
  • Strategic Implication: Jensen's focus on fundamental, verifiable compute operations positions it as durable infrastructure, less susceptible to the hype cycles affecting specific AI models or applications. Investors should note the long-term value proposition of foundational layers versus potentially transient application-level plays.

The Scaling Debate: Centralized Limits vs. Decentralized Potential

  • The conversation addresses the "velocity question": Can decentralized networks tapping into latent compute (like GPUs in laptops/phones) realistically compete with hyperscalers continuously deploying faster, centralized hardware?
  • Ben frames this as a transition from vertical scaling (making single systems bigger/faster) to horizontal scaling (distributing workload across many systems). Vertical scaling, dominant in traditional data centers, offers rapid initial performance gains but faces diminishing returns and escalating costs. Ben argues AI compute has hit this inflection point.
  • He posits that current massive investments by hyperscalers are partly a short-term "marketing reason" and user-capture strategy, not a sustainable long-term scaling approach. The future lies in horizontal scaling, analogous to historical shifts like the emergence of MapReduce (a programming model for processing large datasets across clusters) for big data.
  • AI is currently in a "weird redesign period," rediscovering and reapplying techniques from niche fields like Federated Learning (training models across devices without centralizing data, initially for privacy) for communication efficiency in horizontally scaled systems.
  • Strategic Implication: Investors should recognize the potential paradigm shift from purely centralized AI infrastructure. While hyperscalers currently dominate, the economic and technical limits of vertical scaling create opportunities for horizontally scalable, decentralized networks like Jensen.

Deconstructing Data Center Constraints: Why Vertical Scaling Hits a Wall

  • Ben elaborates on the real-world constraints limiting endless vertical scaling of centralized data centers:
    • Power: Securing sufficient gigawatt-level electricity supply in one location is a major bottleneck.
    • Cooling: Managing heat dissipation for massive GPU clusters is complex and resource-intensive.
    • Logistics: Noise pollution, land availability, local authority approvals, and geographical suitability create significant hurdles. There's a hidden "fight going on between the hyperscalers" for prime locations.
    • Hardware: Sourcing compatible high-end GPUs (e.g., H100s) and ensuring ultra-fast interconnects like InfiniBand (a high-speed networking standard) work seamlessly adds complexity and cost.
    • Data Gathering: Centralized training often overlooks the massive, ongoing communication cost already paid to aggregate data into one place. Future continual learning makes this centralized aggregation less efficient.
  • Strategic Implication: These compounding physical, logistical, and economic barriers make purely centralized scaling increasingly difficult and expensive, strengthening the case for decentralized alternatives that can leverage geographically distributed resources more effectively.

Jensen's Vision: An Infrastructure Layer, Not Just Latent Compute

  • Jensen positions itself not merely as an aggregator of underutilized hardware, but as crucial infrastructure software. It acts as a neutral layer abstracting the complexities of hardware deployment (finding locations, power, cooling) from the complexities of running ML software (training, inference, post-training).
  • Ben draws an analogy to Bitcoin mining, which opened the logistics of energy sourcing and hardware deployment to a global audience, decoupling it from needing downstream applications. Jensen aims to do the same for AI compute.
  • Another analogy used is oil refining: Energy (the raw resource) is refined into ML operations (the valuable product) by GPUs (the refinement hardware). Jensen provides the open market access, allowing hardware providers to sell compute capability almost like a commodity without needing to become full-fledged AI companies themselves.
  • Strategic Implication: Jensen aims to create a more efficient, open market for AI compute, potentially lowering barriers to entry for both hardware providers and AI developers, and disintermediating vertically integrated players.

The Future Internet: Powered by Parameters, Executed by Jensen

  • Ben outlines a vision where the internet's fundamental data type shifts from raw data (text, images) stored in databases to parameters stored in models. This implies a move towards compressed, learned representations of information.
  • Accessing this "parameter space" requires constant ML execution on potentially every device, moving beyond static data retrieval to dynamic, model-mediated interaction.
  • To enable this, Jensen builds three core, open-source components:
    • Consistent Execution: A compiler and libraries ensuring ML operations run compatibly across diverse hardware (Nvidia GPUs, MacBooks, iPhones, AMD, Intel CPUs, etc.), enabling interoperability.
    • Peer-to-Peer Communication: Standardized protocols (analogous to TCP/IP but peer-to-peer) for devices to exchange tensors (multi-dimensional data arrays used in ML) and coordinate distributed tasks like model or data parallelism without central servers.
    • Verification: Cryptographic, probabilistic, and game-theoretic techniques allowing devices to trustlessly verify computations performed by others, removing the need for human contracts or central authorities.
  • Strategic Implication: This foundational stack aims to create a universal, trustless fabric for distributed ML, enabling complex AI workflows across heterogeneous, untrusted devices – a prerequisite for a truly decentralized AI ecosystem.

From Infrastructure to Experience: The Personalized, Probabilistic Web

  • The user-facing consequence of Jensen's infrastructure is a shift towards a more dynamic, personalized, and probabilistic internet experience. This moves beyond naive chatbot interfaces to UIs generated on-the-fly by models understanding user context.
  • Ben compares this shift to the NoSQL (database systems deviating from traditional relational models) revolution, which prioritized flexibility and developer experience over rigid database schemas. The future web, powered by parameters, embraces probabilistic interactions and even "hallucinations" where appropriate.
  • He philosophically suggests this aligns better with inherent human cognition, which evolved in probabilistic environments, arguing that the era of strict technological determinism might be a temporary phase.
  • Strategic Implication: Jensen's infrastructure isn't just about compute efficiency; it enables fundamentally different, potentially more intuitive and personalized AI-native applications and user experiences.

Introducing RL Swarm: Collaborative AI Improvement in Action

  • RL Swarm is presented as a concrete product/demonstration built on Jensen's infrastructure. It's a system for Reinforcement Learning (RL - a type of ML where agents learn by trial-and-error receiving rewards or penalties) post-training.
  • Models within the swarm communicate and critique each other's reasoning (using specific communication tags like "explain" and "critique") to collectively improve on tasks, such as generating verified code or proofs. This allows individual models to learn from the diverse knowledge and reasoning processes of others.
  • Ben uses an analogy to human learning progression: learning basic rules (like pre-training), learning how to think/reason (like RL fine-tuning), and finally learning discourse/dialogue to refine ideas collectively (like RL Swarm).
  • This approach is highly horizontally scalable, as models can form sub-swarms or meta-swarms, mimicking human discussion structures.
  • Strategic Implication: RL Swarm showcases a novel, decentralized approach to AI model improvement that leverages collective intelligence, potentially accelerating progress beyond what single models or centralized training can achieve. It highlights the potential for emergent capabilities in distributed AI systems.

RL Swarm, Testnet, and User Control: Decentralized Identity and Customization

  • RL Swarm itself is open-source software runnable on any network (local or internet). The upcoming Jensen Testnet introduces a crucial layer: persistent, on-chain identity for participating devices/models.
  • Testnet allows tracking contributions, establishing leaderboards, and potentially facilitating future features like trustlessly outsourcing computation when a local device is insufficient. A user could seamlessly push parts of their model's workload to the network, verified by Jensen's protocols.
  • Users retain full control: they can choose which model checkpoints (saved versions of model parameters) to accept, run multiple model versions simultaneously (e.g., a base model and one learning "Kung Fu"), or even just contribute ("seed") to the swarm without updating their local model.
  • The potential for specialized swarms (e.g., a "Rust language coding swarm") allows for targeted model improvement.
  • Strategic Implication: The Testnet bridges Jensen's core tech with crypto-economic mechanisms (identity, potential future value transfer), enabling coordination and incentivization in the decentralized network while preserving user sovereignty over their AI models.

The Philosophical Divide: Open vs. Closed AI Ecosystems and Bias

  • Ben contrasts Jensen's open, cross-platform approach with "walled garden" systems like Apple's or Google's Federated Learning, which operate only within their own hardware/OS ecosystems.
  • A critical point is model bias. All models have biases, reflecting their training data and design. Centralized models inevitably export the biases of their creators to potentially billions of users. Ben argues this is dangerous, drawing parallels to the negative consequences of centralized social media platforms shaping global discourse.
  • He advocates for an open AI ecosystem where a diversity of models with different biases can coexist, interact, and evolve, mirroring the diversity of human thought. "We should allow the machines to have that same degree or even more of difference of opinion and allow them to just work it out..."
  • Personalized models, refined through systems like RL Swarm and controlled by the user, offer an alternative to outsourcing thinking to a few dominant, centrally controlled AIs.
  • Strategic Implication: The debate over open vs. closed AI has profound long-term societal implications. Jensen's infrastructure supports an open model, which may be more resilient, adaptable, and less prone to centralized control and bias risks – a crucial factor for investors considering systemic risks.

The Future of AI Models: Open Source vs. Proprietary Value Capture

  • Ben expresses skepticism about the term AGI (Artificial General Intelligence), finding it ill-defined. He analyzes the AI value chain, arguing that defensible value lies at the extremes: core infrastructure (compute, energy, hardware coordination) and user capture (network effects, personalization moats).
  • In his view, the models themselves occupy a "valley of decaying value." As models become easier to build, replicate, and distill, their proprietary value diminishes. This inevitably favors open source for the model layer itself.
  • Meta's open-sourcing of Llama and the rise of powerful open-source models from China (like DeepSeek) exemplify this trend. He believes Western labs, despite attempts at regulatory capture based on safety concerns, will likely release more open-source models to capture developer attention and infrastructure investment, albeit lagging their state-of-the-art.
  • Strategic Implication: Investors should be wary of business models solely reliant on proprietary model access. Long-term value is more likely to accrue to those controlling essential infrastructure (like decentralized compute networks) or those building strong user relationships and applications on top of increasingly open model foundations.

Conclusion

Jensen Network is architecting foundational infrastructure for a decentralized AI future, enabling trustless, scalable compute across diverse hardware. This challenges centralized dominance, potentially mitigating bias risks and fostering an open ecosystem where AI models can evolve collaboratively and reflect individual user needs, representing a critical trend for Crypto AI investors and researchers to monitor.

Others You May Like