Skip to main content

NVMe over RoCE

Terms related to simplyblock

NVMe over RoCE (Remote Direct Memory Access over Converged Ethernet) is a transport protocol under the NVMe over Fabrics family. It allows NVMe commands to be transmitted over high-speed Ethernet via RDMA, enabling extremely low latency and high throughput.

By bypassing the traditional TCP/IP stack and leveraging direct memory access, NVMe over RoCE achieves near-local performance over Ethernet. This makes it highly effective for use cases that demand maximum storage IOPS and minimal latency.

How NVMe over RoCE Works

In an NVMe/RoCE deployment, NVMe commands are encapsulated within RDMA messages transmitted across Ethernet networks. RoCE operates at Layer 2 or Layer 3, depending on whether it’s RoCE v1 or v2. Both versions require a lossless network fabric, typically achieved through Data Center Bridging (DCB) and Priority Flow Control (PFC).

This architecture minimizes CPU utilization by offloading data movement to RDMA-enabled NICs, enabling efficient parallel access to remote NVMe storage targets. To better understand how this differs from TCP-based approaches, review our NVMe over TCP vs iSCSI benchmark.

🚀 NVMe/RoCE Delivers Speed — But NVMe/TCP Wins in Kubernetes
Skip the hardware complexity of RoCE. Get sub-millisecond performance with Simplyblock’s software-defined NVMe/TCP stack.
👉 Try Simplyblock for Databases on Kubernetes →

Benefits of NVMe over RoCE

Key advantages of NVMe/RoCE for high-performance environments include:

  • Microsecond latency: Ideal for real-time applications and fast storage access.
  • Maximum throughput: Leverages NVMe’s multi-queue parallelism over fast 25/50/100/200 GbE.
  • Offloaded processing: Frees up host CPU cycles by offloading transfers to the NIC.
  • Optimized for flash: Works best in environments using high-speed NVMe SSDs.
  • Supports disaggregated storage: Integrates with modular, adaptive storage architectures like those used by simplyblock.
NVMe over RoCE infographics

NVMe over RoCE vs Other Fabrics

While NVMe/RoCE is the lowest-latency NVMe-oF transport, it has stricter infrastructure requirements. For organizations needing simpler deployment and better compatibility, NVMe over TCP is often a better fit.

FeatureNVMe over RoCENVMe over TCPNVMe over FC
Transport MediumRDMA over EthernetTCP/IP over EthernetFibre Channel
Hardware RequirementsRDMA-capable NICsStandard NICsFC HBAs and switches
Latency<100µs~300–400µs~100–300µs
Network Loss HandlingRequires lossless fabricTolerant of packet lossInherent reliability
ScalabilityHigh with careful tuningHigh, simple to deployModerate
Operational ComplexityHighLowModerate

For detailed comparison of NVMe/TCP vs NVMe/RoCE (NVMe/RDMA) read our blog post on this topic.

Use Cases for NVMe over RoCE

Industries using NVMe over RoCE include:

  • AI/ML workloads requiring extreme throughput and parallelism
  • Financial trading platforms where latency affects transaction outcomes
  • Real-time video analytics and telemetry
  • Tier-0 storage in HPC or disaggregated storage environments
  • Edge computing nodes with specialized RDMA support

These workloads often also benefit from erasure-coded storage for space-efficient fault tolerance.

Easily add shared NVMe storage to your Proxmox cluster.

NVMe over RoCE and Simplyblock

At simplyblock, the focus is on high-performance NVMe storage over TCP, optimized for Kubernetes, hybrid cloud, and edge deployments.

Our software-defined storage platform delivers sub-millisecond latency without requiring RDMA-capable hardware, allowing broader compatibility and lower total cost of ownership. Features like multi-tenancy with QoS and copy-on-write snapshots make simplyblock suitable for dynamic, container-native environments.

Teams often review these glossary pages alongside NVMe over RoCE when they’re building a stable RDMA fabric on Ethernet, including routing choices, congestion controls, and end-to-end latency budgets.

RoCEv2
RDMA (Remote Direct Memory Access)
NVMe/RDMA
InfiniBand

External Resources

Questions and Answers

Why use NVMe over RoCE in low-latency environments?

NVMe over RoCE enables ultra-low latency and high-throughput storage access using RDMA over Ethernet. It’s ideal for data-intensive workloads like high-frequency trading, AI/ML pipelines, and databases that require near-real-time performance on dedicated, lossless networks.

How does NVMe over RoCE compare to NVMe over TCP?

NVMe/RoCE delivers slightly better latency and CPU efficiency than NVMe over TCP, but it requires RDMA-capable NICs and a lossless Ethernet fabric. NVMe/TCP is more flexible, easier to manage, and better suited for Kubernetes and cloud-native workloads.

Can NVMe over RoCE be used in Kubernetes environments?

Technically yes, but RoCE is complex to configure and maintain in dynamic environments like Kubernetes. Kubernetes-native NVMe storage over TCP offers a more operationally friendly alternative for scalable and persistent storage.

Does NVMe over RoCE support encryption at rest?

The NVMe/RoCE protocol does not include encryption. To ensure secure deployments, especially in multi-tenant or regulated environments, encryption at rest should be implemented at the volume level with per-tenant key isolation.

What are the limitations of NVMe over RoCE?

NVMe/RoCE provides excellent performance but requires specialized hardware, strict networking configurations, and lossless Ethernet. It’s less portable and cloud-native than software-defined storage options that use NVMe/TCP for broader compatibility.