Deployment Configuration#
Deployment configurations define how to provision and host model endpoints for evaluation.
Note
For an overview of all deployment strategies and when to use launcher-orchestrated vs. bring-your-own-endpoint approaches, see Serve and Deploy Models.
Deployment Types#
Choose the deployment type for your evaluation:
None (External)
Use existing API endpoints. No model deployment needed.
vLLM
Deploy models using the vLLM serving framework.
SGLang
Deploy models using the SGLang serving framework.
NIM
Deploy models using NVIDIA Inference Microservices.
Quick Reference#
deployment:
type: vllm # or sglang, nim, none
# ... deployment-specific settings