Understanding Storage Class Memory (SCM) and NVMe: Bridging the Gap Between DRAM and SSD
The article explains the concept of Storage Class Memory (SCM), its role as a fast, non‑volatile memory tier between DRAM and NAND flash, and how NVMe and related standards enable new storage hierarchies that improve performance for AI, big data, and emerging technologies.
SCM (Storage Class Memory) was first proposed by IBM and other vendors in 2008, but 2019 is considered the true "SCM year" as the technology began to mature.
SCM sits between DRAM and NAND flash, offering two main usage models: as a low‑level storage device (or hybrid SSD) and as a DRAM‑assist cache that accelerates data access.
Ideally, SCM provides DRAM‑like speed at a cost approaching that of traditional hard drives, though currently only read speeds match DRAM while write speeds lag, and its price‑performance ratio is not yet sufficient for widespread adoption as primary storage.
Several emerging memory technologies compete to become SCM, including PCM, MRAM, and RRAM, but only a few commercial products exist today, such as Intel‑Micron 3D XPoint and Samsung’s Z‑NAND.
SCM aims to close the speed gap between DRAM and SSD, reducing system I/O bottlenecks and power consumption; integrating SCM as a memory buffer or SSD cache can alleviate these issues.
In‑memory computing, which moves computation closer to the data stored in memory, benefits greatly from SCM, enabling faster data analysis for latency‑sensitive workloads.
Although SCM will not fully replace SSDs soon, it revitalizes tiered storage architectures, especially when combined with NVMe, which provides low‑latency, PCIe‑based communication between host and storage.
NVMe has evolved beyond a simple protocol to an ecosystem that includes NVMe‑MI (management interface) and NVMe‑oF (NVMe over Fabrics), supporting fabrics such as Fibre Channel, RoCE, iWARP, and InfiniBand.
These NVMe extensions lower latency in distributed systems, expand in‑memory processing, and are attractive to machine‑learning engineers building high‑performance neural networks that require rapid data movement.
Consequently, NVMe‑SCM solutions are becoming a key component in AI accelerators, 5G communications, and cloud services, with major vendors like Dell EMC, IBM, Hitachi, HPE, Fujitsu, and Huawei offering enterprise‑grade products.
Overall, NVMe‑SCM is not just a hardware upgrade; it influences big data, AI, 5G, and cloud ecosystems, driving a shift toward low‑latency, high‑throughput system architectures.
Architects' Tech Alliance
Sharing project experiences, insights into cutting-edge architectures, focusing on cloud computing, microservices, big data, hyper-convergence, storage, data protection, artificial intelligence, industry practices and solutions.
How this landed with the community
Was this worth your time?
0 Comments
Thoughtful readers leave field notes, pushback, and hard-won operational detail here.