SK hynix develops World’s Best Performing HBM3E

  • Product to drive AI tech innovation with industry’s top performance to be produced in volume from 1H24
  • Launch of HBM3E to solidify SK hynix’s unrivaled leadership in AI memory market following success of HBM3
  • Expansion of HBM3E supply following industry’s largest scale of mass production of HBM to help accelerate business turnaround

SK hynix Inc. (or “the company”, announced today that it successfully developed HBM3E, the next-generation of the highest-specification DRAM for AI applications currently available, and said a customer’s evaluation of samples is underway.

* HBM (High Bandwidth Memory): A high-value, high-performance memory that vertically interconnects multiple DRAM chips, enabling a dramatic increase in data processing speed in comparison to earlier DRAM products. HBM3E is the extended version of the HBM3 and the 5th generation of its kind, succeeding the previous generations HBM, HBM2, HBM2E and HBM3.

The company said that the successful development of HBM3E, the extended version of HBM3 which delivers the world’s best specifications, comes on top of its experience as the industry’s sole mass provider of HBM3. With its experience as the supplier of the industry’s largest volume of HBM products and the mass-production readiness level, SK hynix plans to mass produce HBM3E from the first half of next year and solidify its unrivaled leadership in AI memory market.

According to the company, the latest product not only meets the industry’s highest standards of speed, the key specification for AI memory products, but all categories including capacity, heat dissipation and user-friendliness.

In terms of speed, the HBM3E can process data up to 1.15 terabytes(TB) a second, which is equivalent to processing more than 230 Full-HD movies of 5GB-size each in a second.

In addition, the product comes with a 10% improvement in heat dissipation by adopting the cutting-edge technology of the Advanced Mass Reflow Molded Underfill, or MR-MUF**, onto the latest product. It also provides backward compatibility*** that enables the adoption of the latest product even onto the systems that have been prepared for the HBM3 without a design or structure modification.

** MR-MUF: a process of attaching chips to circuits and filling the space between chips with a liquid material when stacking chips instead of laying a film to improve efficiency and heat dissipation

*** Backward Compatibility: an ability to allow interoperability between an older and a new system without modification to the design, especially in information technology and computing spaces. A new memory product with backward compatibility allows continued use of the existing CPUs and GPUs without modifications to design

“We have a long history of working with SK hynix on High Bandwidth Memory for leading edge accelerated computing solutions,” said Ian Buck, Vice President of Hyperscale and HPC Computing at NVIDIA. “We look forward to continuing our collaboration with HBM3E to deliver the next generation of AI computing.”

Sungsoo Ryu, Head of DRAM Product Planning at SK hynix, said that the company, through the development of HBM3E, has strengthened its market leadership by further enhancing the completeness of HBM product lineup, which is in the spotlight amid the development of AI technology. “By increasing the supply share of the high-value HBM products, SK hynix will also seek a fast business turnaround.”

Visit AITechPark for cutting-edge Tech Trends around AI, ML, Cybersecurity, along with AITech News, and timely updates from industry professionals!

Related posts

Observe.AI Leads Talkdesk AppConnect Marketplace Downloads

Business Wire

Joe Farhat Joins Great Hill Partners as Director of Technology

Business Wire

Retina Launches the 1st AI-Driven Platform – Retina Insight

Business Wire