WEKA previewed the
industry's first high-performance storage solution for the
NVIDIA Grace CPU Superchip. The solution will run on a
powerful new storage server from
Supermicro powered by
WEKA Data Platform software and
Arm Neoverse V2 cores using the
NVIDIA Grace CPU Superchip and
NVIDIA ConnectX-7 and
NVIDIA BlueField-3 networking to accelerate enterprise AI
workloads with unmatched performance density and power efficiency.
Fueling the Next Generation of AI Innovation
Today's AI and high-performance computing (HPC) workloads demand lightning-fast
data access, but most data centers face increasing space and power
constraints.
NVIDIA Grace integrates the level of performance offered by
a flagship x86-64 two-socket workstation or server platform into a single
module. Grace CPU Superchips are powered by 144 high-performance Arm Neoverse
V2 cores that deliver 2x the energy efficiency of traditional x86 servers.
NVIDIA ConnectX-7 NICs and BlueField-3 SuperNICs feature purpose-built
RDMA/RoCE acceleration, delivering high-throughput, low-latency network
connectivity at up to 400Gb/s speeds. The combination of the WEKA Data Platform's
revolutionary zero-copy software architecture running on the Supermicro
Petascale storage server minimizes I/O bottlenecks and reduces AI pipeline
latency to significantly enhance GPU utilization and accelerate AI model
training and inference to dramatically improve time to first token,
discoveries, and insights while reducing power consumption and associated
costs.
Key benefits of the solution include:
- Extreme Speed and
Scalability for Enterprise AI: The NVIDIA Grace CPU Superchip, with 144
high-performance Arm Neoverse V2 cores connected by a high-performance
custom-designed NVIDIA Scalable Coherency Fabric, delivers the performance
of a dual-socket x86 CPU server at half the power. The NVIDIA ConnectX-7
NICs and NVIDIA BlueField-3 SuperNICs provide high-performance networking,
essential for enterprise AI workloads. Paired with the WEKA Data
Platform's AI-native architecture, which accelerates time to first token
by up to 10x, the solution ensures optimal performance across AI data
pipelines at virtually any scale.
- Optimal Resource
Utilization: The
high-performance WEKA Data Platform, combined with Grace CPUs' LPDDR5X
memory architecture, ensures up to 1 TB/s of memory bandwidth and seamless
data flow, eliminating bottlenecks. Integrating WEKA's distributed
architecture and kernel-bypass technology, organizations can achieve
faster AI model training, reduced epoch times, and higher inference
speeds, making it the ideal solution for scaling AI workloads
efficiently.
- Exceptional Energy and
Space Efficiency:
The WEKA Data Platform delivers 10-50x increased GPU stack efficiency to
seamlessly handle large-scale AI and HPC workloads. Additionally, through
data copy reduction and cloud elasticity, the WEKA platform can shrink
data infrastructure footprints by 4-7x and reduce carbon output-avoiding
up to 260 tons of CO2e per PB stored annually and lowering energy costs by
10x. Paired with the Grace CPU Superchip's 2x energy efficiency compared
to leading x86 servers, customers can do more with less, meeting sustainability
goals while boosting AI performance.
"AI is transforming how enterprises around the world
innovate, create, and operate, but the sharp increase in its adoption has
drastically increased data center energy consumption, which is expected to
double by 2026, according to the International Atomic Agency," said Nilesh
Patel, chief product officer at WEKA. "WEKA is excited to partner with NVIDIA,
Arm, and Supermicro to develop high-performance, energy-efficient solutions for
next-generation data centers that drive enterprise AI and high-performance workloads
while accelerating the processing of large amounts of data and reducing time to
actionable insights."
"WEKA has developed a powerful storage solution with
Supermicro that integrates seamlessly with the NVIDIA Grace CPU Superchip to
improve the efficiency of at-scale, data-intensive AI workloads. The solution
will provide fast data access while reducing energy consumption, enabling
data-driven organizations to turbocharge their AI infrastructure," said Ivan
Goldwasser, director of data center CPUs at NVIDIA.
"Supermicro's upcoming ARS-121L-NE316R Petascale storage
server is the first storage optimized server using the NVIDIA Grace Superchip
CPU," said Patrick Chiu, Senior Director, Storage Product Management,
Supermicro. "The system design features 16 high-performance Gen5 E3.S NVMe SSD
bays along with three PCIe Gen 5 networking slots, which support up to two
NVIDIA ConnectX 7 or BlueField-3 SuperNIC networking adapters and one OCP 3.0
network adapter. The system is ideal for high-performance storage workloads
like AI, data analytics, and hyperscale cloud applications. Our collaboration
with NVIDIA and WEKA has resulted in a data platform enabling customers to make
their data centers more power efficient while adding new AI processing
capabilities."
"AI innovation requires a new approach to silicon and system
design that balances performance with power efficiency. Arm is proud to be
working with NVIDIA, WEKA and Supermicro to deliver a highly performant
enterprise AI solution that delivers exceptional value and uncompromising
energy efficiency," said David Lecomber, director for HPC at Arm.
The
storage solution from WEKA and Supermicro using NVIDIA Grace CPU Superchips
will be commercially available in early 2025.