Service Discovery#
Dynamo components (frontends, workers, planner) need to be able to discover each other and their capabilities at runtime. We refer to this as service discovery. There are 2 kinds of service discovery backends supported on Kubernetes.
Discovery Backends#
Backend |
Default |
Dependencies |
Use Case |
|---|---|---|---|
Kubernetes |
✅ Yes |
None (native K8s) |
Recommended for all Kubernetes deployments |
KV Store (etcd) |
No |
etcd cluster |
Legacy deployments |
Kubernetes Discovery (Default)#
Kubernetes discovery is the default and recommended backend when running on Kubernetes. It uses native Kubernetes primitives to facilitate discovery of components:
DynamoWorkerMetadata CRD: Each worker stores its registered endpoints and model cards in a Custom Resource
EndpointSlices: EndpointSlices signal each component’s readiness status
Implementation Details#
Each pod runs a discovery daemon that watches both EndpointSlices and DynamoWorkerMetadata CRs. A pod is only discoverable when it appears as “ready” in an EndpointSlice AND has a corresponding DynamoWorkerMetadata CR. This correlation ensures pods aren’t discoverable until they’re ready, metadata is immediately available, and stale entries are cleaned up when pods terminate.
DynamoWorkerMetadata CRD#
Each worker pod creates a DynamoWorkerMetadata CR that stores its discovery metadata:
apiVersion: nvidia.com/v1alpha1
kind: DynamoWorkerMetadata
metadata:
name: my-worker-pod-abc123
namespace: dynamo-system
ownerReferences:
- apiVersion: v1
kind: Pod
name: my-worker-pod-abc123
uid: <pod-uid>
controller: true
spec:
data:
endpoints:
"dynamo/backend/generate":
type: Endpoint
namespace: dynamo
component: backend
endpoint: generate
instance_id: 12345678901234567890
transport:
nats_tcp: "dynamo_backend.generate-abc123"
model_cards: {}
The CR is named after the pod and includes an owner reference for automatic garbage collection when the pod is deleted.
EndpointSlices#
While DynamoWorkerMetadata resources provide an up-to-date snapshot of a component’s capabilities, EndpointSlices give a snapshot of health of the various Dynamo components.
The operator creates a Kubernetes Service targeting the Dynamo components. The Kubernetes controller in turn creates and maintains EndpointSlice resources that keep track of the readiness of the pods targeted by the Service. Watching these slices gives us an up-to-date snapshot of which Dynamo components are ready to serve traffic.
Readiness Probes#
A pod is marked ready if the readiness probe succeeds. On Dynamo workers, this is when the generate endpoint is available and healthy. These probes are configured by the Dynamo operator for each pod/component.
RBAC#
Each Dynamo component pod is automatically given a ServiceAccount that allows it to watch EndpointSlice and DynamoWorkerMetadata resources within its namespace.
Environment Variables#
The following environment variables are automatically injected into pods by the operator to facilitate service discovery:
Variable |
Description |
|---|---|
|
Set to |
|
Pod name (via downward API) |
|
Pod namespace (via downward API) |
|
Pod UID (via downward API) |
The pod’s instance ID is deterministically generated by hashing the pod name, ensuring consistent identity and correlation between EndpointSlices and CRs.
KV Store Discovery (etcd)#
To use etcd-based discovery instead of Kubernetes-native discovery, add the annotation to your DynamoGraphDeployment:
apiVersion: nvidia.com/v1alpha1
kind: DynamoGraphDeployment
metadata:
name: my-deployment
annotations:
nvidia.com/dynamo-discovery-backend: etcd
spec:
services:
# ...
This requires an etcd cluster to be available. The etcd connection is configured via the platform Helm chart.