The memory aspect of semiconductors today has gotten so extreme. Stuff is so expensive that people are simply not able to make lower-end equipment or like devices anymore. And this is like killing everything, right?
AI chips make like 65% operating margins and gaming does like 40%. So obviously from a business perspective it doesn't really make sense to put too much effort into GPUs which is kind of sad you know because what happened to the rest of us you know everything is like AI.
Meta's platform of apps has 3.5 billion daily active users, and they make something like I think it's like $200 a year off of each user in advertising, which just goes to show that like for every person in the world, there's a lot of companies that want to sell them something.
The AI era is fundamentally reorienting the semiconductor industry from consumer-driven volume to enterprise-driven performance and specialized memory. This means sustained, massive capital expenditure from hyperscalers will continue to be the primary growth engine.
Invest in companies providing specialized memory (HBM, high-density NAND) and custom silicon solutions for AI workloads. These components are the bottlenecks and profit centers for hyperscalers.
The AI infrastructure buildout is far from over. Expect continued, accelerating investment in compute and memory through 2027 and beyond, creating a "rising tide" for the entire semiconductor supply chain.
AI's insatiable demand for compute and memory is fundamentally re-prioritizing semiconductor manufacturing, shifting capacity and R&D from consumer products to high-margin data center components. This creates a new economic reality where memory is the bottleneck and a strategic asset.
Invest in companies positioned to supply high-performance memory (HBM, advanced DRAM, NAND) or those hyperscalers with clear, high-margin internal monetization paths for their AI capex (e.g., advertising-driven models).
The AI infrastructure buildout is far from over, with hyperscalers projecting continued, accelerating capex into 2027 and beyond. This sustained investment will keep memory prices elevated and drive innovation in optical interconnects and custom silicon, creating both challenges for consumers and immense opportunities for strategic investors and builders.
AI's pervasive influence is fundamentally re-architecting the semiconductor supply chain, shifting investment from consumer-grade components to high-margin, specialized AI memory and compute, creating a sustained demand cycle.
Invest in companies positioned to capitalize on the broad memory demand, from HBM manufacturers to NAND suppliers, and those hyperscalers with clear, high-margin monetization paths for their AI infrastructure.
The AI infrastructure buildout is far from over, with hyperscalers committing hundreds of billions annually. This sustained investment will continue to drive semiconductor prices and innovation, making memory and specialized compute the critical bottlenecks and opportunities for the next 3-5 years.
Skyrocketing Costs: GDDR7 prices have quadrupled in the last year, with DRAM contract prices doubling in a single quarter. This means the memory (VRAM) now accounts for 80% of a gaming GPU's bill of materials, making consumer GPU manufacturing increasingly unprofitable.
AI's Profitability: AI chips offer significantly higher operating margins (65%) compared to gaming GPUs (40%). This incentivizes companies like NVIDIA to focus on data center AI, meaning less investment in consumer products and a clear business rationale for the current market dynamics.
Enterprise Skepticism: Wall Street is wary of Microsoft's AI capex due to longer enterprise sales cycles and less immediate ROI compared to advertising-driven models. This suggests investors are prioritizing quick, high-margin returns in the current AI gold rush.
The memory aspect of semiconductors today has gotten so extreme. Stuff is so expensive that people are simply not able to make lower-end equipment or like devices anymore. And this is like killing everything, right?
Capex Surge: Google, Meta, Amazon, and Microsoft are collectively committing over $600 billion in capex for 2026, a 70% average increase. This massive investment is primarily directed at building out AI data centers, compute, memory, and networking infrastructure.
NAND's Moment: Nvidia's Vera Rubin platform will feature over 1,152 terabytes of NAND per rack, with Morgan Stanley estimating Reuben alone will consume 13% of global NAND supply by 2027. This highlights the critical role of massive, cheaper storage for context memory and KV cache in scaling AI.
The memory aspect of semiconductors today has gotten so extreme. Stuff is so expensive that people are simply not able to make lower-end equipment or like devices anymore. And this is like killing everything, right?
We're in an era of finding a use case for something that just requires so much memory. This I I don't see it changing in the immediate future.
AI chips make like 65% operating margins and gaming does like 40%.
AI's integration into core business models is driving hyperscalers to commit unprecedented capital to infrastructure, shifting semiconductor demand from consumer-driven cycles to enterprise-grade, high-margin AI components.
Investigate memory manufacturers and specialized AI silicon providers, as their products are becoming the foundational bottleneck and highest-margin components in the AI infrastructure buildout.
The AI capex spend, projected to exceed $600 billion in upcoming years, is a rising tide lifting all semiconductor boats. Understanding where this capital flows—from HBM to NAND and custom silicon—is crucial for positioning your portfolio and product roadmap for the next half-decade.
AI's computational hunger is fundamentally re-architecting the semiconductor industry, shifting focus from consumer-driven volume to high-margin, specialized memory and compute for hyperscalers. This means a sustained, elevated demand for advanced silicon, with traditional consumer markets becoming a secondary concern.
Invest in companies providing core AI infrastructure components—HBM, advanced NAND, and custom silicon design capabilities—or those hyperscalers with clear, high-margin monetization paths for AI, like advertising.
The AI infrastructure buildout is far from over, with hyperscalers projecting continued, accelerating capex into 2027 and beyond. This sustained investment will keep memory prices high and demand for specialized AI hardware robust, creating a new economic reality for tech investors and builders.
The Macro Transition: Privacy-First Infrastructure. As the novelty of public ledgers fades, the market is moving toward selective transparency where institutions control data visibility.
The Tactical Edge: Audit Canton. Builders should evaluate the Canton Network for any application involving sensitive corporate data or institutional capital flows.
The Bottom Line: Institutional adoption won't happen on public chains as they exist today. The next phase of growth belongs to networks that treat privacy as a foundational requirement for compliance and scale.
The Macro Transition: The move from growth at any price to hard assets for a new order is being fueled by a combination of US political shifts and Japanese monetary instability.
The Tactical Edge: Accumulate GDX and XME on pullbacks while avoiding the retail cheerleading traps in silver handles.
The Bottom Line: The next 12 months will reward those who trade breakouts in physical production and energy rather than those clinging to the 2023 tech playbook.
The Macro Transition: Institutional Convergence. Crypto is shedding its speculative skin to become a fundamental asset class. This transition mirrors the 2002 post-bubble internet era where utility replaced hype.
The Tactical Edge: Identify the Compounders. Focus on protocols with durable income and deep moats. Avoid the "L1 rotation" and prioritize DeFi entities integrating with real-world credit markets.
The Bottom Line: 2026 is about survival and positioning. The winners will be those who build sustainable equity value rather than chasing the next speculative token flip.