Deployment Configuration#
Deployment configurations define how to provision and host model endpoints for evaluation.
Deployment Types#
Choose the deployment type for your evaluation:
None (External)
Use existing API endpoints. No model deployment needed.
vLLM
Deploy models using the vLLM serving framework.
SGLang
Deploy models using the SGLang serving framework.
NIM
Deploy models using NVIDIA Inference Microservices.
TRT-LLM
Deploy models using NVIDIA TensorRT LLM.
Generic
Deploy models using a fully custom setup.
Quick Reference#
deployment:
type: vllm # or sglang, nim, none
# ... deployment-specific settings