Home > AI Solutions > Artificial Intelligence > Guides > Implementation Guide—Virtualizing GPUs for AI with VMware and NVIDIA Based on Dell Infrastructure > Deploy the solution for running AI Workloads as VMs
This section provides guidelines for deploying a GPU cluster with vSphere 8.0 with A100 or A30 GPUs on PowerEdge servers. We focus on enabling MIG and configuring the VMs with access to a virtual GPU (vGPU).
See the NVIDIA AI Enterprise platform deployment guide for detailed instructions about installing and configuring the NVIDIA Virtual GPU Manager for vSphere, which includes these high-level steps:
Follow these steps to configure a VM to use MIG:
Figure 6. Creating a VM and assigning a MIG profile
Figure 7. VM boot options
pciPassthru.use64bitMMIO: TRUE
pciPassthru.allowP2P: TRUE
pciPassthru.64bitMMIOSizeGB: 64
For detailed instructions about installing the NVIDIA driver and licensing the GPU, see the Virtual GPU Software User Guide.
Use the following steps to mount AI datasets from PowerScale to the VMs:
sudo mkdir /mnt/dataset
sudo mount -t nfs <NFS IP addr>:<path> /mnt/dataset
df -h
The AI and data science applications and frameworks are distributed as NGC container images through the NVIDIA NGC Enterprise Catalog. Each container image contains the entire user-space software stack that is required to run the application or framework such as the CUDA libraries, cuDNN, any required Magnum IO components, TensorRT, and the framework. See the Installing AI and data science applications and Frameworks chapter in the NVIDIA AI Enterprise documentation.