VMware, NVIDIA develop integrated generative AI platform

VMware, a multi-cloud services provider, and NVIDIA, a manufacturer of high-end graphics processing units (GPUs), have jointly developed a platform tailored to harness artificial intelligence (AI) capabilities, specifically designed for enterprise applications.

The VMware Private AI Foundation platform, built on VMware Cloud Foundation and NVIDIA AI Enterprise software, will enable enterprises to customize models and run generative AI applications, including intelligent chatbots, virtual assistants, search engines, and summarization tools.

“Our expanded collaboration with VMware will offer hundreds of thousands of customers — across financial services, healthcare, manufacturing, and more — the full-stack software and computing they need to unlock the potential of generative AI using custom applications built with their own data,” said Jensen Huang, founder and CEO, NVIDIA.

VMware launches multi-cloud management with VMware Aria
VMware enhances cloud platforms to drive developer productivity

The fully integrated Generative AI platform is expected to incorporate integrated AI tools that enable enterprises to execute proven models trained on their private data in a cost-effective manner.

“Customer data is everywhere — in their data centers, at the edge, and in their clouds,” said Raghu Raghuram, CEO, VMware. “Together with NVIDIA, we will empower enterprises to run their generative AI workloads adjacent to their data with confidence while addressing their corporate data privacy, security, and control concerns.”

End-to-end cloud native framework

The platform will include NVIDIA NeMo, an end-to-end, cloud-native framework integrated into NVIDIA AI Enterprise, the operating system of the NVIDIA AI platform. NeMo combines customization frameworks, guardrail toolkits, data curation tools, and pre-trained models to provide enterprises with a straightforward, cost-effective, and quick way to adopt generative AI.

For deploying generative AI in production, NeMo leverages TensorRT for Large Language Models (TRT-LLM), optimizing and accelerating inference performance on the latest LLMs on NVIDIA GPUs. With NeMo, VMware Private AI Foundation in partnership with NVIDIA will enable enterprises to incorporate their own data for constructing and executing custom generative AI models on VMware’s hybrid cloud infrastructure.

VMware Private AI Foundation with NVIDIA will be supported by Dell Technologies, Hewlett Packard Enterprise, and Lenovo, which will be among the first to offer systems that supercharge enterprise LLM customization and inference workloads with NVIDIA L40S GPUs, NVIDIA BlueField-3 DPUs, and NVIDIA ConnectX-7 SmartNICs. 

VMware intends to release VMware Private AI Foundation with NVIDIA in early 2024.