
Prefer to listen instead? Here’s the podcast version of this article.
The artificial intelligence revolution is often measured in terms of algorithms and compute power, but there’s another critical factor shaping its trajectory: memory chips. As AI models grow larger and more complex, their need for rapid access to vast amounts of data has placed unprecedented demands on memory technology. High-Bandwidth Memory (HBM) and emerging compute-in-memory architectures are stepping into the spotlight, not just as supporting components, but as foundational technologies driving the next wave of AI innovation. Understanding this shift is essential for businesses, technologists, and policymakers alike, as memory is poised to redefine the limits of what AI systems can achieve.
AI models, especially LLMs, demand massive memory bandwidth. As reported by the Financial Times, AI compute has exploded—while DRAM bandwidth has barely budged, leading to a “memory wall” that throttles performance. HBM changes that by delivering significantly higher bandwidth via stacked memory architectures, lowering latency and boosting efficiency. [Financial Times]
HBM is no longer just memory—it’s mission-critical. HBM (1 → 4) has evolved rapidly, offering multi-TB/s bandwidth and dense stacking. The JEDEC standard HBM4, released in April 2025, supports up to 2 TB/s per stack and 64 GB capacity—a leap from earlier generations.
SK Hynix has surged ahead of Samsung in DRAM revenue, thanks to its prowess in HBM. By Q1 2025, HBM accounted for over 40% of Hynix’s DRAM revenue, up from just 5% in late 2022. SB experts attribute this to HBM’s pricing leverage—strong margins and tailored, long-term contracts—making memory companies more strategic players in AI infrastructure. [Financial Times]
HBM4 isn’t just faster—it’s smarter. With logic dies powered by advanced processors (via TSMC or Samsung’s foundry) and hybrid bonding techniques, HBM4 promises even greater bandwidth, efficiency, and task-specific tuning. This positions memory chips as not just passive storage but optimized co-processors.
Real innovation lies even deeper. Compute-in-memory (PIM) and memory-centric computing integrate compute capabilities directly into memory chips.
The future of AI will not be determined by algorithms alone—it will be shaped by the memory systems that enable those algorithms to run at scale. From High-Bandwidth Memory (HBM) driving efficiency in today’s data centers to compute-in-memory architectures promising revolutionary gains in performance and energy use, memory chips are evolving from silent enablers into central players in the AI ecosystem.
WEBINAR