How the latest version of the High Bandwidth Memory standard is keeping up with increasingly demanding applications.
With the publishing of the HBM3 update to the High Bandwidth Memory (HBM) standard, a new king of bandwidth is crowned. The torrid performance demands of advanced workloads, with AI/ML training leading the pack, drive the need for ever faster delivery of bits. Memory bandwidth is a critical enabler of computing performance, thus the need for the accelerated evolution of the standard with HBM3 representing the new benchmark.
Here’s what HBM3 offers:
Let’s roll it all up in a potential use case. A future AI accelerator implementation has six (6) HBM3 DRAM devices. Total aggregate memory bandwidth at 6.4 Gb/s is 4.9 Terabytes per second (TB/s). Each 12 x 32Gb HBM3 DRAM device has a 48GB capacity, so the AI accelerator can access 288 GB of direct-attached HBM3 memory.
That’s tremendous capability. HBM3 extends the track record of bandwidth performance set by what was originally dubbed the “slow and wide” HBM memory architecture. While the interface is still wide, HBM3 operating at 6.4 Gb/s is now really quite fast. All things being equal, higher speeds mean higher power. The motivation of the wide interface (which necessitated the higher complexity 2.5D architecture) was to run at low data rates delivering high bandwidth at low power. To compensate, HBM3 drops the operating voltage (the last bullet in our list above) for higher power efficiency.
But there is no free lunch, and lower voltage means lower design margin for what is already a challenging 2.5D design. Fortunately, Rambus has your back with our 8.4 Gb/s HBM3 Memory Subsystem that provides plenty of design headroom plus room to scale. To help you successfully harness the full potential of HBM3 memory, Rambus provides both interposer and package reference designs.
The Rambus memory subsystem includes a modular and highly configurable memory controller. The controller is optimized to maximize throughput and minimize latency, and its memory parameters are run-time programmable. With a pedigree of over 50 HBM2 and HBM2E customer implementations, it has demonstrated efficiency over a wide variety of configurations and data traffic scenarios.
While the road to higher performance is a journey and not a destination, the latest generation of HBM promises to deliver some very extraordinary capabilities. All hail the new king of memory bandwidth, HBM3.
Additional resources:
Leave a Reply