As a multi-component solution, you can use VMware Private AI Foundation with NVIDIA to run generative AI workloads by using accelerated computing from NVIDIA, and virtual infrastructure management and cloud management from VMware Cloud Foundation.
VMware Private AI Foundation with NVIDIA provides a platform for provisioning AI workloads on ESXi hosts with NVIDIA GPUs. In addition, running AI workloads based on NVIDIA GPU Cloud (NGC) containers is specifically validated by VMware.
VMware Private AI Foundation with NVIDIA supports two use cases:
- Development use case
- Cloud administrators and DevOps engineers can provision AI workloads, including Retrieval-Augmented Generation (RAG), in the form of deep learning virtual machines. Data scientists and AI application developers can use these deep learning virtual machines for AI development. See About Deep Learning VM Images in VMware Private AI Foundation with NVIDIA.
- Production use case
- Cloud administrators can provide DevOps engineers with a VMware Private AI Foundation with NVIDIA environment for provisioning production-ready AI workloads on Tanzu Kubernetes Grid (TKG) clusters on vSphere IaaS control plane.
For information about the components that are part of the VMware Private AI Foundation with NVIDIA solution and their architecture on top of VMware Cloud Foundation, see System Architecture of VMware Private AI Foundation with NVIDIA.