SK hynix begins mass production of 12-layer HBM3E memory: 36GB capacity per module @ 9.6Gbps
09/26/2024, 11:22 AM UTC
三星电子开始量产12层HBM3E内存:每模块容量高达36GB,带宽9.6GbpsSK hynix starts mass production of 12-layer HBM3E memory: 36GB capacity per module @ 9.6Gbps
➀ 三星电子开始量产全球首款12层HBM3E内存,容量高达36GB,带宽9.6Gbps;➁ 新内存专为AI GPU设计,计划在未来12个月内供应给NVIDIA;➂ 三星电子通过推出这项新技术,旨在保持其在AI内存领域的领导地位。➀ SK hynix has begun mass production of the world's first 12-layer HBM3E memory with a capacity of up to 36GB and a bandwidth of 9.6Gbps; ➁ The new memory is designed for AI GPUs and is set to be supplied to NVIDIA within 12 months; ➂ SK hynix aims to maintain its leadership in AI memory with the introduction of this new technology.SK hynix has announced the volume production of its new 12-layer HBM3E memory, boasting a capacity of up to 36GB and a bandwidth of 9.6Gbps. This marks a significant advancement in memory technology, as it is the largest capacity of existing HBM to date.
The new memory is designed for AI GPUs, with SK hynix planning to supply the memory chips to NVIDIA within the next 12 months. This comes just six months after the company launched the 8-layer HBM3E memory, which was the first of its kind in the industry.
SK hynix's HBM3 and HBM3E memory has been a key component in NVIDIA's Hopper H100 and H200 AI GPUs, as well as its new Blackwell AI GPUs. The new 12-layer HBM3E memory chips offer a bandwidth of 9.6Gbps, which is the highest memory speed on the market.
SK hynix's President of AI infrastructure, Justin Kim, emphasized the company's commitment to leading in AI memory, stating that they will continue to be the number one global AI memory provider as they prepare for the next generation of memory products to overcome the challenges of the AI era.
---
本文由大语言模型(LLM)生成,旨在为读者提供半导体新闻内容的知识扩展(Beta)。