Executors#

Executors run evaluations by orchestrating containerized benchmarks in different environments. They handle resource management, IO paths, and job scheduling across various execution backends, from local development to large-scale cluster deployments.

Core concepts:

  • Your model is separate from the evaluation container; communication is via an OpenAI‑compatible API

  • Each benchmark runs in a Docker container pulled from the NVIDIA NGC catalog

  • Execution backends can optionally manage model deployment

Choosing an Executor#

Select the executor that best matches your environment and requirements:

Local Executor

Run evaluations on your local machine using Docker for rapid iteration and development workflows.

Local Executor
Slurm Executor

Execute large-scale evaluations on Slurm-managed high-performance computing clusters with optional model deployment.

Slurm Executor
Lepton Executor

Run evaluations on Lepton AI’s hosted infrastructure with automatic model deployment and scaling.

Lepton Executor