Fake GPU Operator (Development / Testing)

View as Markdown

For development, staging, load testing, or CI environments that lack physical NVIDIA GPUs, you can install a fake GPU operator to simulate GPU resources on cluster nodes. This allows the NVCA agent to discover GPUs and manage function deployments without actual GPU hardware.

The fake GPU operator is for non-production use only. For production deployments with real GPUs, install the NVIDIA GPU Operator.

Prerequisites

  • A running Kubernetes cluster with kubectl access
  • helm >= 3.12

Install KWOK

The fake GPU operator depends on KWOK (Kubernetes Without Kubelet) to manage simulated GPU device plugins on nodes. Install KWOK before the fake GPU operator:

$kubectl apply -f https://github.com/kubernetes-sigs/kwok/releases/download/v0.7.0/kwok.yaml

Verify the KWOK controller is running:

$kubectl get pods -n kube-system -l app=kwok-controller
$# Expected: kwok-controller-... 1/1 Running

The KWOK install may produce a FlowSchema error (creation or update of FlowSchema object ... is not allowed). This is non-critical and can be safely ignored.

Installation

Add the RunAI helm repository and install the fake GPU operator:

$helm repo add fake-gpu-operator \
> https://runai.jfrog.io/artifactory/api/helm/fake-gpu-operator-charts-prod \
> --force-update
$
$helm upgrade -i gpu-operator fake-gpu-operator/fake-gpu-operator \
> -n gpu-operator --create-namespace \
> --set 'topology.nodePools.default.gpuCount=8' \
> --set 'topology.nodePools.default.gpuProduct=NVIDIA-H100-80GB-HBM3'

This configures one node pool named default with 8 simulated H100 GPUs per node.

topology.nodePools must be a map, not an array.

Using array index syntax (--set 'topology.nodePools[0].gpuCount=8') will create a YAML array instead of a map and cause the status-updater to fail with:

yaml: unmarshal errors: cannot unmarshal !!seq into map[string]topology.NodePoolTopology

Always use named keys: topology.nodePools.default.gpuCount=8.

Node Labeling

The fake GPU operator watches for nodes with the label run.ai/simulated-gpu-node-pool=<pool-name> and patches their status to advertise fake nvidia.com/gpu extended resources. You must label the nodes that should receive simulated GPUs:

$kubectl label node <node-name> run.ai/simulated-gpu-node-pool=default

The pool name (default) must match a key in topology.nodePools from the helm install.

GPU Metadata Labels (Optional)

The NVCA agent uses several GPU metadata labels for dynamic discovery. On real GPU nodes these are set by the NVIDIA GPU Operator. To suppress warnings from NVCA on fake GPU nodes, add the following labels:

$kubectl label node <node-name> \
> nvidia.com/gpu.family=hopper \
> nvidia.com/gpu.machine=NVIDIA-DGX-H100 \
> nvidia.com/cuda.driver.major=535 \
> --overwrite

Adjust the values to match the GPU product you configured (e.g., ampere for A100, ada for L40S).

Verification

Check that the fake GPU operator pods are running:

$kubectl get pods -n gpu-operator
$# Expected: 3 pods Running (topology-server, status-updater, kwok-gpu-device-plugin)

Confirm that labeled nodes now advertise GPU resources:

$kubectl get nodes -o custom-columns="NAME:.metadata.name,GPU:.status.allocatable.nvidia\.com/gpu"
$# Labeled nodes should show the configured GPU count (e.g., 8)

If GPUs do not appear, verify the node has the run.ai/simulated-gpu-node-pool=default label and that the status-updater pod is not in an error state.

Integration with NVCF

For the smoothest experience, install the fake GPU operator before running helmfile sync. This way the NVCA agent discovers GPUs on its first boot and no re-registration is needed.

The recommended sequence is:

  1. Install KWOK
  2. Install fake-gpu-operator and label target nodes
  3. Verify nvidia.com/gpu appears in node allocatable resources
  4. Proceed with the control-plane installation

If Installed After the Control Plane

If you add the fake GPU operator to a cluster that already has NVCF deployed, the NVCA agent will be crash-looping because it cannot find GPUs. After installing the fake GPU operator and verifying GPUs appear on nodes, re-register the cluster and restart the operator:

$# Re-run the cluster bootstrap
$kubectl exec -n nvca-operator deploy/nvca-operator -c nvca-operator -- \
> /usr/bin/nvca-self-managed bootstrap --system-namespace nvca-operator
$
$# Restart the operator (it caches cluster IDs at startup)
$kubectl rollout restart deployment nvca-operator -n nvca-operator
$kubectl rollout status deployment nvca-operator -n nvca-operator --timeout=120s

The operator restart will re-run the bootstrap init container, recreate the NVCFBackend resource, and spawn a fresh NVCA agent pod that discovers the simulated GPUs.

For details on the bootstrap process, see self-managed-clusters (Manual Cluster Registration).

Customization

GPU Count and Product

Adjust the GPU count, product name, and memory per node pool:

$helm upgrade gpu-operator fake-gpu-operator/fake-gpu-operator \
> -n gpu-operator \
> --set 'topology.nodePools.default.gpuCount=4' \
> --set 'topology.nodePools.default.gpuProduct=NVIDIA-A100-SXM4-80GB' \
> --set 'topology.nodePools.default.gpuMemory=81920'

Multiple Node Pools

Define multiple pools with different GPU configurations by using different map keys:

$helm upgrade gpu-operator fake-gpu-operator/fake-gpu-operator \
> -n gpu-operator \
> --set 'topology.nodePools.h100-pool.gpuCount=8' \
> --set 'topology.nodePools.h100-pool.gpuProduct=NVIDIA-H100-80GB-HBM3' \
> --set 'topology.nodePools.a100-pool.gpuCount=4' \
> --set 'topology.nodePools.a100-pool.gpuProduct=NVIDIA-A100-SXM4-80GB'

Then label nodes with the corresponding pool name:

$kubectl label node <h100-node> run.ai/simulated-gpu-node-pool=h100-pool
$kubectl label node <a100-node> run.ai/simulated-gpu-node-pool=a100-pool

Teardown

To remove the fake GPU operator and all simulated GPU resources:

$# Remove the fake GPU operator
$helm uninstall gpu-operator -n gpu-operator
$kubectl delete namespace gpu-operator --ignore-not-found
$
$# Remove KWOK
$kubectl delete -f https://github.com/kubernetes-sigs/kwok/releases/download/v0.7.0/kwok.yaml
$
$# Remove the node labels (for each labeled node)
$kubectl label node <node-name> run.ai/simulated-gpu-node-pool-
$kubectl label node <node-name> nvidia.com/gpu.product-
$kubectl label node <node-name> nvidia.com/gpu.family-
$kubectl label node <node-name> nvidia.com/gpu.machine-
$kubectl label node <node-name> nvidia.com/cuda.driver.major-

After removing the fake GPU operator, the NVCA agent will lose GPU visibility and begin crash-looping. Either install a real GPU Operator with physical GPUs or uninstall the NVCA operator.