Nvidia BlueField-3 User Manual page 13

Networking platform
Hide thumbs Also See for BlueField-3:
Table of Contents

Advertisement

Feature
Overlay Networks In order to better scale their networks, data center operators often create overlay
networks that carry traffic from individual virtual machines over logical tunnels in
encapsulated formats such as NVGRE and VXLAN. While this solves network scalability
issues, it hides the TCP packet from the hardware offloading engines, placing higher loads
on the host CPU. NVIDIA DPU effectively addresses this by providing advanced NVGRE and
VXLAN hardware offloading engines that encapsulate and de-capsulate the overlay
protocol.
RDMA and RDMA
NVIDIA DPU, utilizing IBTA RDMA (Remote Data Memory Access) and RoCE (RDMA over
over Converged
Converged InfiniBand/Ethernet) technology, delivers low-latency and high-performance
InfiniBand/
over InfiniBand/Ethernet networks. Leveraging data center bridging (DCB) capabilities as
Ethernet (RoCE)
well as advanced congestion control hardware mechanisms, RoCE provides efficient low-
latency RDMA services over Layer 2 and Layer 3 networks.
Quality of Service
Support for port-based Quality of Service enabling various application requirements for
(QoS)
latency and SLA.
Storage
Acceleration
NVMe-oF
Non-volatile Memory Express (NVMe) over Fabrics is a protocol for communicating block
storage IO requests over RDMA to transfer data between a host computer and a target
solid-state storage device or system over a network. NVIDIA BlueField-3 DPU may operate
as a co-processor offloading specific storage tasks from the host using its powerful NVMe
over Fabrics Offload accelerator.
SR-IOV
NVIDIA DPU SR-IOV technology provides dedicated adapter resources and guaranteed
isolation and protection for virtual machines (VM) within the server.
High-
Performance Acc
elerations
GPU Direct
GPUDirect RDMA is a technology that provides a direct P2P (Peer-to-Peer) data path
between the GPU Memory directly to/from the NVIDIA HCA devices. This provides a
significant decrease in GPU-GPU communication latency and completely offloads the CPU,
removing it from all GPU-GPU communications across the network. NVIDIA DPU uses high-
speed DMA transfers to copy data between P2P devices resulting in more efficient system
applications
Isolation
BlueField-3 DPU functions as a "computer-in-front-of-a-computer," unlocking unlimited
opportunities for custom security applications on its Arm processors, fully isolated from
the host's CPU. In the event of a compromised host, BlueField-3 may detect/block
malicious activities in real-time and at wire speed to prevent the attack from spreading
further.
Cryptography
From IPsec and TLS data-in-motion inline encryption to AES-XTS block-level data-at-rest
Accelerations
encryption and public key acceleration, BlueField-3 DPU hardware-based accelerations
offload the crypto operations and free up the CPU, reducing latency and enabling scalable
crypto solutions. BlueField-3 "host-unaware" solutions may transmit and receive data,
while BlueField-3 acts as a bump-in-the-wire for crypto.
A consolidated compute and storage network achieves significant cost-performance
advantages over multi-fabric networks. Standard block and file access protocols can
leverage RDMA for high-performance storage access: NVMe over Fabric offloads for the
target machine
BlueField-3 DPU may operate as a co-processor offloading specific storage tasks from
the host, isolating part of the storage media from the host, or enabling abstraction of
software-defined storage logic using the NVIDIA BlueField-3 Arm cores. On the storage
initiator side, NVIDIA BlueField-3 DPU can prove an efficient solution for hyper-
converged systems to enable the host CPU to focus on compute while all the storage
interface is handled through the Arm cores.
Tag Matching and Rendezvous Offloads
Adaptive Routing on Reliable Transport
Burst Buffer Offloads for Background Checkpointing
Description
13

Advertisement

Table of Contents
loading

Table of Contents