The memory aspect of semiconductors today has gotten so extreme. Stuff is so expensive that people are simply not able to make lower-end equipment or like devices anymore. And this is like killing everything, right?
AI chips make like 65% operating margins and gaming does like 40%. So obviously from a business perspective it doesn't really make sense to put too much effort into GPUs which is kind of sad you know because what happened to the rest of us you know everything is like AI.
Meta's platform of apps has 3.5 billion daily active users, and they make something like I think it's like $200 a year off of each user in advertising, which just goes to show that like for every person in the world, there's a lot of companies that want to sell them something.
The AI era is fundamentally reorienting the semiconductor industry from consumer-driven volume to enterprise-driven performance and specialized memory. This means sustained, massive capital expenditure from hyperscalers will continue to be the primary growth engine.
Invest in companies providing specialized memory (HBM, high-density NAND) and custom silicon solutions for AI workloads. These components are the bottlenecks and profit centers for hyperscalers.
The AI infrastructure buildout is far from over. Expect continued, accelerating investment in compute and memory through 2027 and beyond, creating a "rising tide" for the entire semiconductor supply chain.
AI's insatiable demand for compute and memory is fundamentally re-prioritizing semiconductor manufacturing, shifting capacity and R&D from consumer products to high-margin data center components. This creates a new economic reality where memory is the bottleneck and a strategic asset.
Invest in companies positioned to supply high-performance memory (HBM, advanced DRAM, NAND) or those hyperscalers with clear, high-margin internal monetization paths for their AI capex (e.g., advertising-driven models).
The AI infrastructure buildout is far from over, with hyperscalers projecting continued, accelerating capex into 2027 and beyond. This sustained investment will keep memory prices elevated and drive innovation in optical interconnects and custom silicon, creating both challenges for consumers and immense opportunities for strategic investors and builders.
AI's pervasive influence is fundamentally re-architecting the semiconductor supply chain, shifting investment from consumer-grade components to high-margin, specialized AI memory and compute, creating a sustained demand cycle.
Invest in companies positioned to capitalize on the broad memory demand, from HBM manufacturers to NAND suppliers, and those hyperscalers with clear, high-margin monetization paths for their AI infrastructure.
The AI infrastructure buildout is far from over, with hyperscalers committing hundreds of billions annually. This sustained investment will continue to drive semiconductor prices and innovation, making memory and specialized compute the critical bottlenecks and opportunities for the next 3-5 years.
Skyrocketing Costs: GDDR7 prices have quadrupled in the last year, with DRAM contract prices doubling in a single quarter. This means the memory (VRAM) now accounts for 80% of a gaming GPU's bill of materials, making consumer GPU manufacturing increasingly unprofitable.
AI's Profitability: AI chips offer significantly higher operating margins (65%) compared to gaming GPUs (40%). This incentivizes companies like NVIDIA to focus on data center AI, meaning less investment in consumer products and a clear business rationale for the current market dynamics.
Enterprise Skepticism: Wall Street is wary of Microsoft's AI capex due to longer enterprise sales cycles and less immediate ROI compared to advertising-driven models. This suggests investors are prioritizing quick, high-margin returns in the current AI gold rush.
The memory aspect of semiconductors today has gotten so extreme. Stuff is so expensive that people are simply not able to make lower-end equipment or like devices anymore. And this is like killing everything, right?
Capex Surge: Google, Meta, Amazon, and Microsoft are collectively committing over $600 billion in capex for 2026, a 70% average increase. This massive investment is primarily directed at building out AI data centers, compute, memory, and networking infrastructure.
NAND's Moment: Nvidia's Vera Rubin platform will feature over 1,152 terabytes of NAND per rack, with Morgan Stanley estimating Reuben alone will consume 13% of global NAND supply by 2027. This highlights the critical role of massive, cheaper storage for context memory and KV cache in scaling AI.
The memory aspect of semiconductors today has gotten so extreme. Stuff is so expensive that people are simply not able to make lower-end equipment or like devices anymore. And this is like killing everything, right?
We're in an era of finding a use case for something that just requires so much memory. This I I don't see it changing in the immediate future.
AI chips make like 65% operating margins and gaming does like 40%.
AI's integration into core business models is driving hyperscalers to commit unprecedented capital to infrastructure, shifting semiconductor demand from consumer-driven cycles to enterprise-grade, high-margin AI components.
Investigate memory manufacturers and specialized AI silicon providers, as their products are becoming the foundational bottleneck and highest-margin components in the AI infrastructure buildout.
The AI capex spend, projected to exceed $600 billion in upcoming years, is a rising tide lifting all semiconductor boats. Understanding where this capital flows—from HBM to NAND and custom silicon—is crucial for positioning your portfolio and product roadmap for the next half-decade.
AI's computational hunger is fundamentally re-architecting the semiconductor industry, shifting focus from consumer-driven volume to high-margin, specialized memory and compute for hyperscalers. This means a sustained, elevated demand for advanced silicon, with traditional consumer markets becoming a secondary concern.
Invest in companies providing core AI infrastructure components—HBM, advanced NAND, and custom silicon design capabilities—or those hyperscalers with clear, high-margin monetization paths for AI, like advertising.
The AI infrastructure buildout is far from over, with hyperscalers projecting continued, accelerating capex into 2027 and beyond. This sustained investment will keep memory prices high and demand for specialized AI hardware robust, creating a new economic reality for tech investors and builders.
The Macro Shift: The Unification. Legacy finance is unbundling into onchain modules where yield is derived from real-world economic activity rather than token emissions.
The Tactical Edge: Audit your yield. Move capital toward protocols like RE that bridge to non-self-referential markets.
The Bottom Line: The next 12 months belong to "Neo-Finance" players who dominate the boring work of regulatory compliance and fiat integration.
The Macro Transition: Vertical Liquidity. Exchanges are evolving from passive pools into active revenue collectors that capture MEV and launch fees to subsidize liquidity.
The Tactical Edge: Monitor Aero. Watch the Metadex03 launch in Q2 to see if liquidity migrates from Uniswap to the higher-yield Aero pools on Ethereum Mainnet.
The Bottom Line: Aero is betting that better economics for liquidity providers will always win the war for volume. If they successfully export their Base dominance to Mainnet, the decentralized exchange hierarchy will be permanently altered over the next 12 months.
The transition from DeFi to Neo-Finance where on-chain liquidity meets institutional payment rails.
Prioritize assets that are integrated with payment processors like Stripe or Bridge.
2026 is the year of the exponential. The winners won't be the high-float L1s but the protocols that function as the economic engine for both lenders and shoppers.