NVIDIA announced its first Arm Neoverse-based
discrete data center CPU designed for AI infrastructure and high
performance computing, providing the highest performance and twice the
memory bandwidth and energy-efficiency compared to today's leading
server chips.
The NVIDIA Grace CPU Superchip comprises two CPU chips connected, coherently, over NVLink-C2C, a new high-speed, low-latency, chip-to-chip interconnect.
The Grace CPU Superchip complements NVIDIA's first CPU-GPU integrated module, the Grace Hopper Superchip,
announced last year, which is designed to serve giant-scale HPC and AI
applications in conjunction with an NVIDIA Hopper architecture-based
GPU. Both superchips share the same underlying CPU architecture, as well
as the NVLink-C2C interconnect.
"A new type of data center has emerged - AI factories that process
and refine mountains of data to produce intelligence," said Jensen
Huang, founder and CEO of NVIDIA. "The Grace CPU Superchip offers the
highest performance, memory bandwidth and NVIDIA software platforms in
one chip and will shine as the CPU of the world's AI infrastructure."
Introducing NVIDIA's CPU Platform
Created to provide the highest performance, Grace CPU Superchip packs
144 Arm cores in a single socket, offering industry-leading estimated
performance of 740 on the SPECrate2017_int_base benchmark.
This is more than 1.5x higher compared to the dual-CPU shipping with
the DGX A100 today, as estimated in NVIDIA's labs with the same class
of compilers.
Grace CPU Superchip also provides industry-leading energy efficiency
and memory bandwidth with its innovative memory subsystem consisting of
LPDDR5x memory with Error Correction Code for the best balance of speed
and power consumption. The LPDDR5x memory subsystem offers double the
bandwidth of traditional DDR5 designs at 1 terabyte per second while
consuming dramatically less power with the entire CPU including the
memory consuming just 500 watts.
The Grace CPU Superchip is based on the latest data center architecture, Arm v9.
Combining the highest single-threaded core performance with support for
Arm's new generation of vector extensions, the Grace CPU Superchip will
bring immediate benefits to many applications.
The Grace CPU Superchip will run all of NVIDIA's computing software
stacks, including NVIDIA RTX, NVIDIA HPC, NVIDIA AI and Omniverse. The
Grace CPU Superchip along with NVIDIA ConnectX-7 NICs offer
the flexibility to be configured into servers as standalone CPU-only
systems or as GPU-accelerated servers with one, two, four or eight
Hopper-based GPUs, allowing customers to optimize performance for their
specific workloads while maintaining a single software stack.
Designed for AI, HPC, Cloud and Hyperscale Applications
The Grace CPU Superchip will excel at the most demanding HPC, AI, data
analytics, scientific computing and hyperscale computing applications
with its highest performance, memory bandwidth, energy efficiency and
configurability.
The Grace CPU Superchip's 144 cores and 1TB/s of memory bandwidth
will provide unprecedented performance for CPU-based high performance
computing applications. HPC applications are compute-intensive,
demanding the highest performing cores, highest memory bandwidth and the
right memory capacity per core to speed outcomes.
NVIDIA is working with leading HPC, supercomputing, hyperscale and
cloud customers for the Grace CPU Superchip. Both it and the Grace
Hopper Superchip are expected to be available in the first half of 2023.
To learn more about the Grace CPU Superchip, watch Huang's GTC 2022 keynote.