VMware and NVIDIA have introduced VMware Private AI Foundation with NVIDIA to empower enterprises to customise models and run generative AI applications such as intelligent chatbots, assistants, search, and summarisation.
To be release in early 2024, the fully integrated solution will feature generative AI software and accelerated computing from NVIDIA, built on VMware Cloud Foundation and optimised for AI.
This is a leg up for enterprises that want to adopt generative AI to boost productivity, improve efficient and transform business processes, among many other benefits.
“Generative AI and multi-cloud are the perfect match. Customer data is everywhere — in their data centres, at the edge, and in their clouds. Together with NVIDIA, we’ll empower enterprises to run their generative AI workloads adjacent to their data with confidence while addressing their corporate data privacy, security and control concerns,” said Raghu Raghuram, CEO of VMware.
“Our expanded collaboration with VMware will offer hundreds of thousands of customers — across financial services, healthcare, manufacturing and more — the full-stack software and computing they need to unlock the potential of generative AI using custom applications built with their own data,” said Jensen Huang, Founder and CEO of NVIDIA.
VMware Private AI Foundation with NVIDIA will enable enterprises to harness the capabilities of generative AI and offer the technology as a service to their users. It is expected to include integrated AI tools to empower enterprises to run proven models trained on their private data in a cost-efficient manner.
The platform will feature the NVIDIA NeMo end-to-end, cloud-native framework included in NVIDIA AI Enterprise — the operating system of the NVIDIA AI platform — that allows enterprises to build, customise and deploy generative AI models virtually anywhere. NeMo combines customisation frameworks, guardrail toolkits, data curation tools and pretrained models to offer enterprises an easy, cost-effective and fast way to adopt generative AI.
For deploying generative AI in production, NeMo uses TensorRT for Large Language Models, which accelerates and optimises inference performance on the latest LLMs on NVIDIA GPUs. With NeMo, VMware Private AI Foundation with NVIDIA will enable enterprises to pull in their own data to build and run custom generative AI models on VMware’s hybrid cloud infrastructure.
VMware Private AI Foundation with NVIDIA will be supported by Dell Technologies, Hewlett Packard Enterprise and Lenovo — which will be among the first to offer systems that supercharge enterprise LLM customisation and inference workloads with NVIDIA L40S GPUs, NVIDIA BlueField-3 DPUs and NVIDIA ConnectX-7 SmartNICs.
