In VMware Private AI Foundation with NVIDIA, as a DevOps engineer, by using the Kubernetes API, you provision a TKG cluster that uses NVIDIA GPUs. In a disconnected environment, you must additionally set up a local Ubuntu package repository and use the Harbor Registry for the Supervisor.

Prerequisites

Verify with the cloud administrator that the following prerequisites are in place for the AI-ready infrastructure.

Procedure

  1. Provision a TKG cluster on the vSphere namespace configured by the cloud administrator.

    See Provision a TKGS Cluster for NVIDIA vGPU.

  2. Complete the TKG cluster setup.

    See Installing VMware vSphere with VMware Tanzu (Air-gapped).

    1. Provide a local Ubuntu package repository and upload the container images in the NVIDIA GPU Operator package to the Harbor Registry for the Supervisor.
    2. Update the Helm chart definitions of the NVIDIA GPU Operator to use the local Ubuntu package repository and private Harbor Registry.
    3. Provide NVIDIA license information.
    4. Install the NVIDIA GPU Operator.

What to do next

Deploy an AI container image from the Harbor Registry for the Supervisor.