Custom HBM Compute Architecture

Transforming AI accelerators with unparalleled compute and memory density

Thriving Organization
The Marvell custom HBM compute architecture sets a new benchmark for AI accelerators by enabling up to 25% more area for compute, 33% greater memory capacity, and a 70% reduction in memory interface power. Designed to optimize the performance, and total cost of ownership (TCO) of custom XPUs, this groundbreaking architecture reflects the Marvell commitment to innovation and collaboration with industry leaders like Micron, Samsung, and SK hynix.

What is the Marvell Custom HBM Compute Architecture?

 

The Marvell custom HBM compute architecture enables the design of high-bandwidth memory (HBM) systems tailored specifically for AI accelerators (XPUs). By integrating advanced 2.5D packaging technology and custom interfaces, this architecture overcomes the limitations of standard interface-based designs, enabling higher compute and memory density with improved power efficiency.

Key features:

  • Tailored HBM interfaces: Optimized interfaces improve performance, reduce die size, and lower costs by customizing the memory subsystem.
  • Silicon real-estate optimization: High bandwidth IO interfaces reduce the required silicon real estate in each die by up to 25%, allowing the integration of additional compute logic or support for more HBM stacks.
  • Collaborative innovation: Developed in partnership with leading HBM manufacturers like Micron, Samsung, and SK hynix to meet the specific needs of hyperscale cloud operators.

 


Will Chu, SVP and GM of Custom Compute and Storage at Marvell, In Dong Kim, VP of Product Planning and Business Enabling at Samsung Semiconductor and Sunny Kang, VP of DRAM Technology at SK hynix discuss the how HBM will become the mainstream solution for hyperscalers on the Six Five.

Collaboration with Industry Leaders

Marvell custom HBM compute architecture is the result of strategic partnerships with industry-leading memory manufacturers to redefine the future of AI acceleration:

Core Behaviors

 

"Increased memory capacity and bandwidth will help cloud operators efficiently scale their infrastructure for the AI era."
Raj Narasimhan, SVP and GM Compute and Networking Business Unit, Micron

Core Behaviors

 

"Collaborating with Marvell allows us to help customers produce optimized solutions for their workloads and infrastructure."
Sunny Kang, VP of DRAM Technology, SK hynix

Core Behaviors

 

"Optimizing HBM for specific XPUs and software environments will greatly improve the performance of cloud operators’ infrastructure."
Harry Yoon, EVP, Samsung Electronics

HBM Compute Architecture Diagram

The diagram illustrates the evolution of HBM integration within XPUs, showcasing Marvell innovations that optimize the I/O interface to save up to 25% silicon real estate while boosting compute and memory capacity. The simplified I/O architecture enables efficient memory access, lower power consumption, and improved scalability for AI workloads. It emphasizes the collaboration between compute logic, HBM stacks, and advanced packaging technologies to create a holistic, efficient design.

Role of Marvell Products in Society

Custom HBM Compute - Why It Matters

 

Today’s XPUs can contain four or more vertically integrated stacks of HBM with a cumulative capacity of 192GB that are connected to processing units through thousands of individual channels.

Next-generation XPUs will demand even more capacity and higher transfer rates, creating challenges in thermal management, power delivery, cost and interface performance.

The Marvell custom HBM compute architecture provides hyperscalers with a roadmap to better performance and lower TCO by dramatically reducing the size and power consumption of HBM interfaces and controllers. Benefits include:

  • Enhanced scalability: Supports seamless scaling of cloud data centers to meet the needs of next-generation AI workloads.
  • Power efficiency: Reduces interface power consumption by up to 70%, promoting sustainable energy use.
  • Optimized cost: Minimizes silicon real estate usage and reduces overall production costs.
  • Flexibility for AI infrastructure: Customization ensures optimized performance for diverse AI applications.

 

Marvell Optimizing Power and Performance Through Innovation
Sustainable Materials

Custom Compute - The Road Ahead

 

The future of computing is custom. With Moore’s Law nearing its limits, performance and efficiency gains will increasingly be achieved through breakthrough ideas in core design, materials, packaging, interfaces, and other aspects of semiconductor design. By 2028, it is estimated that 25% of AI processors will be based around custom designs and technologies.

These custom processors will power systems designed around a unique combination of devices that, in concert, will deliver unprecedented throughput, efficiency, and TCO. Every cloud will be unique because the underlying silicon will be unique.

Marvell is at the forefront of the custom revolution. We have developed a portfolio of technologies and services to help an expanding spectrum of customers create infrastructure that will enable them to achieve their goals for growth, sustainability, performance and customer service.

Product Sustainability

Marvell Custom Portfolio

Marvell custom devices are designed to meet the high-speed, high-performance silicon needs of AI and cloud infrastructure, 5G and enterprise networks, and automotive applications. With a growing IP portfolio at advanced nodes, including 2nm, Marvell delivers optimized solutions in compute, networking, security, storage, and high-speed SerDes, ensuring faster time-to-market and maximum performance, power, and area efficiency. Over the past 25 years, Marvell has delivered over 2,000 custom ASICs, driving innovation for future-ready solutions.

Newsroom

Latest News

Contact Us

We believe better partnerships help to build better technologies. Let’s connect and see what we can design together!

Thank You for Your Interest

We will be in touch with you soon!