VMware,
Inc. announced a collaboration with Intel to extend
the companies' more than two decades of innovation to help customers
accelerate the adoption of artificial intelligence (AI) and enable
private AI everywhere - across data centers, public clouds, and edge
environments. VMware and Intel are working to deliver a jointly
validated AI stack that will enable customers to use their existing
general-purpose VMware and Intel infrastructure and open source software
to simplify building and deploying AI models. The combination of VMware
Cloud Foundation and Intel's AI software suite, Intel® Xeon® processors
with built-in AI accelerators, and Intel® Max Series GPUs, will deliver
a validated and benchmarked AI stack for data preparation, model
training, fine-tuning and inferencing to accelerate scientific discovery
and enrich business and consumer services.
More
than 300,000 customers deploy VMware Cloud globally, and VMware
virtualization software is deployed nearly everywhere in the enterprise
where data is created, processed, or consumed. This makes VMware Cloud a
fast means to bring AI-accelerated compute and models to wherever
business gets done. Similarly, Intel offers open, scalable and trusted
solutions to hundreds of thousands of customers. The ubiquity of VMware
and Intel products in the enterprise is a powerful combination which
will increase the accessibility of data science, and enable
organizations globally to adopt Private AI, an architectural approach
that aims to balance the business gains from AI with the practical
privacy and compliance needs.
"When
it comes to AI, there is no longer any reason to debate trade-offs in
choice, privacy, and control. Private AI empowers customers with all
three, enabling them to accelerate AI adoption while future-proofing
their AI infrastructure," said Chris Wolf, vice president of VMware AI
Labs. "VMware Private AI with Intel will help our mutual customers
dramatically increase worker productivity, ignite transformation across
major business functions, and drive economic impact."
"For
decades, Intel and VMware have delivered next-generation data
center-to-cloud capabilities that enable customers to move faster,
innovate more, and operate efficiently," said Sandra Rivera, executive
vice president and general manager of the Data Center and AI Group
(DCAI) at Intel. "With the potential of artificial intelligence to
unlock powerful new possibilities and improve the life of every person
on the planet, Intel and VMware are well equipped to lead enterprises
into this new era of AI, powered by silicon and software."
Boost AI Performance and get a More Secure AI Infrastructure with Lower TCO
VMware
Private AI brings compute capacity and AI models to where enterprise
data is created, processed, and consumed, whether in a public cloud,
enterprise data center, or at the edge, in support of traditional AI/ML
workloads and generative AI. VMware and Intel are enabling the
fine-tuning of task specific models in minutes to hours and the
inferencing of large language models at faster than human communication
using the customer's private corporate data. VMware and Intel now make
it possible to fine-tune smaller, economical state of the art models
which are easier to update and maintain on shared virtual systems, which
can then be delivered back to the IT resource pool when the batch AI
jobs are complete. Use cases such as AI-assisted code generation,
experiential customer service centers recommendation systems, and
classical machine statistical analytics can now be co-located on the
same general purpose servers running the application.
VMware
and Intel are designing a reference architecture that combines Intel's
AI software suite, Intel® Xeon® processors, and Data Center GPUs with
VMware Cloud Foundation to enable customers to build and deploy private
AI models on the infrastructure they have, thereby reducing total cost
of ownership and addressing concerns of environmental sustainability.
This VMware Private AI reference architecture with Intel AI will
include:
- 4th Gen Intel® Xeon® processors with Intel® Advanced Matrix Extensions (Intel® AMX) deliver
up to 10x significant out-of-box performance improvements using
standard industry frameworks and libraries, end-to-end data science
productivity tools, and optimized AI models.
- Intel® Data Center GPU Max contains
up to 128 Xe cores and is Intel's foundational GPU compute building
block targeted at the most demanding AI workloads. Intel Max Series GPUs
will be available in several form factors to address different customer
needs.
- Intel's AI software suite is
packaged with end-to-end open source software and optional licensing
components to enable developers to run full AI pipeline workflows from
data preparation to fine-tuning to inference, accelerate building
multi-node scaling and deploying AI on enterprise IT infrastructure. The
open oneAPI framework enables processors and hardware
accelerator-agnostic software development, allowing developers to write
code once and run it across different architectures, eliminating the
need for multiple code bases and specialized languages. Intel's
Transformer Extensions and PyTorch Extension's deep integration with
developer favorite open source Hugging Face libraries provide automated
optimization recipes for fine-tuning and compressing models for
efficient inference.
VMware
Cloud Foundation brings consistent enterprise-class infrastructure,
operational simplicity, and enhanced security to VMware Private AI
through capabilities such as:
- VMware vSAN Express Storage Architecture which offers up to 30% greater performance and minimal overhead for important capabilities such as encryption for I/O intensive AI/ML tuning and inferencing workloads.
- vSphere
Distributed Resources Scheduler helps efficiently use spare compute
capacity for AI model training and inferencing, leveraging the same
clusters other AI applications use, thereby maximizing capacity and
reducing TCO.
- VMware NSX for building micro segmentation and advanced threat protection capabilities.
- Secure boot and Virtual TPM to enable model and data confidentiality.
VMware
Private AI will be supported by servers from Dell Technologies, Hewlett
Packard Enterprise and Lenovo running 4th Gen Xeon CPUs with Intel®
Advanced Matrix Extensions (Intel® AMX) and Intel Max Series GPUs.
For more information about VMware Private AI with Intel, visit: https://www.vmware.com/products/vsphere/ai-ml.html