Observability Configuration
This page provides guidance on configuring observability for self-hosted NVCF control-plane, including metrics, logging, and tracing.
Overview
Self-hosted NVCF control-plane observability enables users to monitor the health and performance of their NVCF deployment. The observability solution is designed to be:
- Cloud-agnostic: Works in any Kubernetes environment (cloud provider, on-premises, or air-gapped)
- Offline-capable: Fully functional in isolated networks without external dependencies
- Bring-Your-Own (BYO): Integrates with your existing observability platforms
- No vendor lock-in: Uses open standards (Prometheus, OpenTelemetry, OTLP)
The observability solution currently provides:
- [Metrics Collection]: Prometheus-compatible metrics from all control-plane services
- [Logging]: Logs emitted to stdout/stderr for easy collection
- [Tracing]: Distributed tracing via OTLP to your collector
- [Dashboards]: Reference Grafana dashboards for key metrics
Looking for a quick start? If you want to quickly deploy example observability components to explore metrics, logs, and dashboards, see self-hosted-example-dashboards.
The example deployments are designed for development and testing only, and are not suitable for production use. For production deployments, follow the guidance on this page to integrate with your own observability infrastructure.
Early Access Phase
NVCF self-hosted observability is currently in Early Access (EA). During EA, NVCF provides interfaces and documentation for you to integrate with your own observability backend:
What’s Provided:
- Documented metrics for critical control-plane services
- Example scrape targets for prometheus-operator ServiceMonitor configuration
- Metrics exposed via Prometheus-compatible endpoints
- Logs emitted to stdout/stderr for easy collection
- Configuration and deployment documentation
- Example dashboards for key metrics
Your Responsibility:
- Deploy and manage your own observability backend (Prometheus, Grafana, Loki, Elasticsearch, etc.)
- Configure metrics scraping from control-plane services
- Deploy log collectors (e.g., Fluentd, Promtail, OTel Collector) to aggregate logs
- Set up your preferred visualization and alerting tools
Control-Plane Services
The following control-plane services expose metrics and logs for monitoring:
Core NVCF Services:
- NVCF API: Main API for function management and invocation
- Invocation Service: Handles function invocation requests
- SPOT Instance Service (SIS): Manages worker pod and cluster state
- State Metrics Service: (Available in GA) Aggregates and exports NVCF-specific metrics
Supporting Services:
- Cassandra (C*): Primary database for control-plane state
- OpenBao/Vault: Secret management and S2S authentication
- Encrypted Secrets Service (ESS): Function and account secrets
- NATS Core: Pub/sub messaging
- NATS JetStream: Persistent messaging
Worker Pod Components:
- Utils Container: Proxy to NATS from user applications
- Init Container: Setup and resource loading
- Inference Container: Inference workload
Architecture
Metrics Collection
All control-plane services expose Prometheus-compatible metrics endpoints. You can scrape these metrics using:
- Prometheus Operator: Create ServiceMonitor resources based on the provided scrape targets
- Prometheus: Configure scrape targets manually
- OpenTelemetry Collector: Use the Prometheus receiver
Metrics Documentation:
Detailed metrics documentation is available for each service, including metric names,
types, labels, and descriptions. See the per-service metrics reference under the
Metrics section.
Logging
Log Format:
- All services emit logs to stdout/stderr (standard for Kubernetes)
- Sensitive data redaction must be configured by the log collector
Log Collection:
You can collect logs using any Kubernetes-compatible log aggregator:
- Fluentd or Fluent Bit
- Promtail (for Loki)
- Filebeat (for Elasticsearch)
- OpenTelemetry Collector (filelog receiver)
System Logs:
System logs are available at standard UNIX locations and from the systemd journal.
Tracing (Available in GA)
Distributed tracing support via OpenTelemetry Protocol (OTLP) is planned for a future release:
- Key flows will be instrumented with OpenTelemetry SDK
- Traces will be exportable via OTLP (HTTP or gRPC)
- Configurable sampling strategies
- Support for any OTLP-compatible backend (Jaeger, Tempo, Zipkin, etc.)
- Tracing is configurable via Helm values under
global.observability.tracing
Configuration
You configure observability by integrating with your own backend:
Metrics Scraping
Use Prometheus Operator with the provided ServiceMonitor examples:
Or configure Prometheus scrape targets manually in your prometheus.yml.
Log Collection
Deploy a log collector as a DaemonSet to ship logs to your backend:
Configure your log collector to:
- Tail logs from all namespaces
- Add metadata labels (pod name, namespace, service)
- Forward to your log aggregation backend (Loki, Elasticsearch, etc.)
Tracing Configuration
Enable distributed tracing by setting Helm values under
global.observability.tracing. The control-plane exports traces via OTLP
to your own OTLP-compatible collector. Set collectorEndpoint,
collectorPort, and collectorProtocol to match your collector’s
address and protocol.
Helm overrides example:
Configuration fields:
- enabled: Set to
trueto enable OTLP trace export from control-plane services. - collectorEndpoint: DNS name or address of your OTLP collector (e.g.,
OpenTelemetry Collector, Jaeger collector). Use a Kubernetes service DNS name
such as
<service>.<namespace>.svc.cluster.localwhen the collector runs in-cluster. - collectorPort: Port on which the collector accepts OTLP traffic (e.g., 4317 for HTTP, 4318 for gRPC depending on your collector setup).
- collectorProtocol: OTLP protocol:
httporgrpc. Must match how your collector is configured.
Ensure your collector is deployed and reachable from the NVCF control-plane namespace, and that it forwards traces to your backend (Jaeger, Tempo, Zipkin, or another OTLP-compatible system).
Dashboards
Reference Grafana dashboards are provided for control-plane services showing critical metrics for key services:
-
ESS (Encrypted Secrets Service)
-
Cassandra
-
Vault
-
Invocation Service
-
NVCF API
-
SIS (SPOT Instance Service)
-
Worker Pods (Utils Container, Init Container, Inference Container)
- Note: Worker Pods are deployed in the backend cluster, not the control-plane cluster, but their configuration is globally controlled as part of the control-plane
-
State Metrics Service (Available in GA)
Dashboard Location:
Dashboards are provided in native Grafana JSON format for file-provisioning.
Load dashboards into Grafana by placing them in /etc/grafana/provisioning/dashboards/ on startup.
Published dashboards will be available in the nv-cloud-function-helpers public GitHub repository.
Troubleshooting
For troubleshooting common observability issues:
Metrics not appearing:
-
Verify the service is exposing metrics:
-
Check ServiceMonitor or scrape configuration:
-
Verify network policies allow scraping:
-
Check service logs for errors:
Logs not being collected:
-
Verify log collector DaemonSet is running:
-
Check collector can access pod logs:
-
Verify log backend is reachable:
-
Check for log redaction or filtering rules:
Security
Metrics Endpoints:
-
Metrics endpoints should be accessed over HTTP in-cluster only
- Any external access should be SSL/TLS or mTLS secured with a reverse proxy or other ingress controller, or
- Aggregated locally and exposed via a secured otel-collector
-
All sensitive log data should be redacted by the log collector (currently, this is the responsibility of the log collector, not the service)
- Example implementation by OTEL Collector: Log Redaction
-
User-provided observability backend should be properly secured with RBAC, TLS/SSL, and other security best practices.
Related Documentation
Version Compatibility
NVCF self-hosted control-plane observability is compatible with:
- Any official supported Kubernetes version
- Any Prometheus-compatible metrics collection system
- Any log aggregation system that can collect from Kubernetes stdout/stderr or read from the filesystem (depending on K8s cluster configuration)
For the latest compatibility information, see the release notes.