Users that do not have access to expensive GPUs or do not wish to spend money on GPUs for learning and testing can use the GPU Operator Simulator software. This was originally created by the Run.AI team (acquired by Nvidia). This recipe describes how to configure, install and use the GPU Simulator Operator to reduce costs and simulate the presence of GPUs within a cluster.
Note
In our testing, we have noticed a number of issues with the GPU Simulator Software. Please open a GitHub issue directly on the project's repo to report issues or request enhancements.
You have provisioned or imported one or more Kubernetes clusters into a Project in your Rafay Org.
Ensure that you have not already deployed the Nvidia GPU Operator on the cluster.
Note
In this recipe, you will notice that we are configuring and deploying the GPU Simulator software as a workload and not a Cluster Blueprint. Users can also deploy the simulator software to 100s of clusters in their Org by using a blueprint if required.
You can assign GPUs to nodes by applying a label to specific nodes. Run the following command being sure to update the node name and the node pool name.
The node pool names are defined in the values.yaml file used when deploying the workload. The options in the provided values.yaml file are A100, H100 and T400. Additional node pool groups can be added to the values.yaml file as needed.
Create a Workload to deploy the GPU Simulator Operator onto the cluster.
Under Applications, select Workloads and create a new workload with the name gpu-simulator.
Ensure that you select Helm 3 for Type, Pull files from repository for Artifact Sync, Helm for Repository Type and select the namespace gpu-operator-resources.
Select the gpu-simulatorRepository
Enter fake-gpu-operator for the Chart Name
Create a YAML file named values.yaml with the following YAML:
topology:# nodePools is a map of node pool name to node pool configuration.# Nodes are assigned to node pools based on the node pool label's value (key is configurable via nodePoolLabelKey).# # For example, nodes that have the label "run.ai/simulated-gpu-node-pool: default"# will be assigned to the "default" node pool.nodePools:A100:gpuProduct:NVIDIA-A100gpuCount:4H100:gpuProduct:NVIDIA-H100gpuCount:4T400:gpuProduct:NVIDIA-T400gpuCount:4
Upload the values.yaml in the workload and Save and Go To Placement.
Select the target cluster from the list of available clusters and click Save and go to publish.
Publish the workload and make sure that it gets published successfully in the target cluster before moving to the next step.