Accelerate the transition from pilot to production with benchmark-proven performance, automated data pipelines, and a flexible consumption model.
Discover how to unlock the true value of your data.
Unified, automated, and ready to turn data into intelligence.
Storage protocols built for spinning hard drives have been limiting flash performance for over a decade. While SSDs deliver hardware capable of microsecond response times, legacy interfaces like SATA and SAS add hundreds of microseconds of unnecessary overhead through single-queue architectures and protocol translation layers.
NVMe (non-volatile memory express) is a storage access and transport protocol built specifically for solid-state drives (SSDs) that connects directly through the PCIe bus, bypassing the bottlenecks of disk-era protocols. Instead of funneling commands through a single queue like SATA, NVMe supports up to 64,000 queues, each with 64,000 commands—a fundamental shift in how storage communicates with modern multi-core processors.
This guide covers NVMe’s architecture, real-world performance advantages, NVMe over Fabrics transport options, and why end-to-end implementation matters for enterprise workloads.
The story of NVMe starts with a mismatch. For two decades, SATA and SAS protocols assumed storage devices needed time to physically seek data, building in command overhead that made sense when disk platters had to rotate into position. These protocols funnel all commands through a single queue—adequate for mechanical seeks, but a serious constraint for flash memory.
The NVM Express consortium was formed in 2008 to address this problem. Intel, Samsung, Dell, and other major IT providers collaborated on a specification designed from the ground up for non-volatile memory, with no assumptions about mechanical components.
The first NVMe specification (1.0) arrived in 2011. Adoption moved quickly. By 2014, NVMe SSDs were shipping from multiple manufacturers, and enterprise adoption accelerated through the mid-2010s as PCIe Gen 3 provided sufficient bandwidth for the protocol’s parallel architecture. The NVMe 2.0 specification, released in 2021, reorganized the standard into a modular library, separating base specifications from command sets and transport specifications. This restructuring opened NVMe to new device types, including computational storage drives, Zoned Namespace (ZNS) SSDs, and even rotational media.
Enterprise SSD capacity on NVMe continues to grow at a compound annual rate exceeding 43%, according to IDC projections.
NVMe is a storage transfer protocol for accessing data quickly from flash memory storage devices such as SSDs. It enables flash memory to communicate directly with a computer via a high-speed peripheral component interconnect express (PCIe) bus, offering fast, high-throughput, and massively parallel data transfer.
Here’s how it works:
The end result is significantly lower overhead when compared with traditional transfer protocols such as serial attached SCSI (SAS) and serial ATA (SATA).
Also, NVMe is optimised for non-uniform memory access (NUMA), meaning it was designed to allow for multiple CPU cores to manage queues. Modern processors contain dozens of cores, yet SATA and SAS funnel them all through a single I/O queue. NVMe assigns dedicated queue pairs to each CPU core, eliminating lock contention and enabling true parallel processing.
NVMe devices connect via PCIe lanes, with each lane providing bidirectional bandwidth. A typical NVMe SSD uses four PCIe lanes, delivering about 2GB/s per lane with PCIe Gen 4. Enterprise arrays aggregate multiple devices for even higher throughput. With PCIe Gen 5 doubling per-lane bandwidth to 4GB/s, NVMe performance headroom continues to grow.
But bandwidth alone doesn’t determine performance. Latency—the time between request and response—often matters more for transactional workloads. NVMe’s direct PCIe connection eliminates multiple bus transitions and protocol conversions that add overhead in SATA implementations.
The NVMe command set is intentionally streamlined. Where SCSI requires complex command parsing and multiple protocol translation layers, NVMe uses a simplified set of commands that map directly to flash operations. The result: fewer CPU instructions per I/O request, lower host stack latency, and more efficient use of processor cycles.
NVMe also introduces features absent from legacy protocols, including multi-path I/O with namespace sharing across multiple controllers, end-to-end data protection via metadata fields, and sanitize commands that make data recovery impossible when hardware is retired.
NVMe’s main benefits include:
NVMe leverages the high-speed PCIe bus for latency, allowing for significantly faster data transfer rates compared to older interfaces. This results in substantially lower latency and higher input/output operations per second (IOPS).
NVMe's ability to deliver high-speed data transfer and low latency significantly improves the performance of storage systems. NVMe uses parallel data paths through multiple queues, each capable of handling up to 64,000 commands, to support multi-core processors used for applications requiring fast data access, such as databases, virtualised environments, and high-performance computing (HPC).
NVMe's architecture supports scalable performance as SSD technology evolves, ensuring compatibility with future advancements in storage. It easily scales with advancements in PCIe technology, supporting newer versions like PCIe 5.0 and beyond.
In enterprise environments, NVMe is essential for handling large-scale data operations and supporting demanding workloads, enabling faster data analytics, reduced processing times, and improved overall efficiency.
Although NVMe may initially be more expensive than traditional HDDs, its superior performance and durability usually leads to long-term cost savings through increased productivity and reduced downtime.
For individuals, NVMe offers faster boot times, quicker file transfers, and a more responsive computing experience. This is particularly beneficial for gamers, content creators, and professionals working with large files.
NVMe’s efficiency compounds its performance advantages. By eliminating legacy protocol overhead and reducing CPU utilization per I/O operation, NVMe SSDs deliver significantly higher work per watt than SATA. With PCIe Gen5, NVMe drives can achieve up to ~25× higher sequential read throughput and ~20× higher sequential write throughput compared to SATA SSDs, while maintaining low CPU overhead. This combination of higher throughput and better efficiency per I/O means fewer drives and less CPU time are required to achieve the same workload, resulting in measurable reductions in power consumption and cooling requirements at scale.
NVMe is very extendable and works well with all emerging persistent memory technologies. It’s also much more flexible than Transmission Control Protocol (TCP), as it can be deployed on any TCP network without special hardware, making it attractive for both on-premises and cloud environments.
NVMe offers numerous advantages over traditional storage protocols like SAS and SATA, particularly in terms of performance, scalability, and efficiency.
Here are some specific metrics and examples illustrating these differences:
The vast number of command queues and the depth supported by NVMe allows systems to minimize latency and maximise throughput. This is especially beneficial in multi-core systems where multiple processors can issue commands simultaneously.
An NVMe SSD connected via PCIe 5.0 can achieve transfer speeds up to 64 GB/s with 8 lanes, which is significantly faster than the 600MB/s limit of SATA III or the 1.5GB/s per lane of SAS-3. This makes NVMe particularly advantageous for data-intensive applications such as real-time data processing and large-scale database management.
Lower latency in NVMe drives means faster response times for applications, which is crucial for performance-critical tasks like high-frequency trading, real-time analytics, and interactive gaming.
The NVMe protocol reduces unnecessary command translations and utilizes fewer CPU cycles, which directly contributes to faster read and write operations. This efficiency is noticeable in enterprise environments where high transaction rates and minimal latency are critical.
Businesses adopting NVMe can future-proof their storage infrastructure, ensuring long-term performance gains and compatibility with cutting-edge storage technologies.
Understanding NVMe’s advantages requires comparing it directly against the protocols it replaces.
According to Architecting IT, raw NAND flash reads take approximately 100 microseconds. SATA can add 50–100 microseconds of protocol overhead on top of that. NVMe keeps protocol overhead below 10 microseconds, which means the protocol itself is no longer the bottleneck—the flash media is.
For organisations currently running SATA-attached SSDs, the performance gap is significant. A single NVMe device delivers over 1 million IOPS for 4 KB random reads—performance that requires dozens of SATA SSDs. In enterprise database workloads, this translates to more transactions per second, lower query response times, and fewer storage-related wait events.
NVMe-oF is the practice of connecting NVMe storage systems with hosts over a network or data fabric. A data fabric simply refers to the network architecture, transfer protocol, and other technologies and services that allow data to be accessed and managed seamlessly across this network. It’s about extending the low latency and performance capabilities of NVMe over PCIe to storage area networks (SANs) through NVMe-friendly standards for popular transfer protocols such as Ethernet, Fibre Channel, and TCP.
The NVMe over Fabrics (NVMe-oF) specification was created and is currently maintained by NVM Express, an open collection of standards for non-volatile memory technologies. Let’s take a closer look at NVMe transfer protocols supported by this standard.
NVMe over Fibre Channel (also known as NVMe/FC or NVMe-FC) is a high-speed transfer protocol for connecting NVMe storage systems to host devices over fabrics. It supports the fast, in-order, lossless transfer of raw block data between NVMe storage devices in a network.
The original Fibre Channel Protocol (FCP) was designed to transport SCSI commands over Fibre Channel networks. It has become the dominant protocol used to connect servers with shared storage systems. While traditional FCP can be used to connect servers with NVMe storage devices, there’s an inherent performance penalty incurred when translating SCSI commands into NVMe commands for the NVMe array.
NVMe/FC supports the transfer of native NVMe commands, eliminating this translation bottleneck. This unlocks the full potential of FCP as a transport technology for end-to-end NVMe storage solutions, including parallelism, deep queues, multi-queueing, and high-speed data transfer.
NVMe over TCP (NVMe/TCP) is a low-latency transfer protocol that allows you to use standard Ethernet TCP/IP networking equipment natively with NVMe storage.
TCP/IP is the default transfer protocol used by the internet by which messages are broken up into packets to avoid having to resend an entire message in the event of a disruption of service. As an extension of the NVMe-oF specification, NVMe/TCP allows you to send NVMe commands using the same TCP/IP protocol transfer packets you use to transmit other types of data.
The plug-and-play ease and lower cost of standard Ethernet makes it an economical solution for connecting your NVMe storage devices over a data fabric. Modern implementations achieve 200–250 microsecond latency—faster than SATA SSDs despite crossing the network. NVMe/TCP works with existing switches, standard NICs, and cloud provider networks, making it especially attractive for hybrid cloud deployments.
A dual-protocol approach supporting both FC-NVMe and NVMe/TCP gives organisations flexibility to choose the transport that best fits each workload while maintaining a consistent NVMe command model throughout.
RoCE (RDMA over Converged Ethernet) promises the lowest network latency through kernel bypass—RDMA operations complete in approximately a microsecond. But RoCE requires lossless Ethernet with Priority Flow Control configured across every switch and adapter. A single misconfigured port can cause performance collapse. When properly deployed, RoCE delivers sub-microsecond storage latency.
Simply installing NVMe drives rarely delivers expected benefits. The entire storage stack must support end-to-end NVMe operations for the protocol’s advantages to reach applications.
Many organisations buy NVMe SSDs for existing arrays and expect a transformation. The drives communicate via NVMe, but the array controller translates everything to SCSI internally for compatibility. This translation adds microseconds at every layer, negating much of what NVMe is designed to deliver.
Successful NVMe implementation requires attention across the stack:
NVMe technology is transforming many industries, including:
Basically, any industry that relies on quick data access and retrieval can benefit from NVMe.
NVMe’s evolution extends well beyond raw speed. Several developments are shaping the protocol’s trajectory:
Big data is no longer enough to maintain a competitive edge—it must also be fast.
How do you make big data fast?
You start in the server room. Transitioning from HDDs to SSDs is a good place to start, but it’s only one piece of the SAN puzzle. The transfer protocol, interconnects, and networking architecture also play important roles in the overall speed of your storage system. That means replacing legacy technologies like SAS and SATA with NVMe, which offers clear and significant advantages in throughput and latency.
Everpure leverages the increased transfer speeds of native NVMe transfer protocols to provide performant all-flash storage solutions.
The secret to Everpure performance is DirectFlash®:
All Everpure solutions leverage NVMe storage to unlock the full potential of flash memory. Everpure offers on-premises all-flash solutions for all your block, file, and object storage needs:
Additionally, Everpure offers a suite of software solutions that can simplify and unify storage management across your hybrid cloud:
Get ready for the most valuable event you’ll attend this year.
Access on-demand videos and demos to see what Everpure can do.
Charlie Giancarlo on why managing data—not storage—is the future. Discover how a unified approach transforms enterprise IT operations.
For nine consecutive years, Everpure has maintained a Net Promoter Score of over 80. Find out how we did it and what it means for our customers.