NVIDIA Inference Reference ArchitectureCommon Component CombinationsAsk a question|Copy page|View as Markdown|More actionsThese are typical adoption patterns showing how components complement each other: Use CaseCore ComponentsOptional AdditionsBasic ML InferenceTensorRT + TritonDALI, GPU OperatorSpeech/NLP PipelineRiva SDK + TritonDALI, TensorRTSingle-Node LLMTensorRT-LLM + DynamoModel OptimizerDistributed LLMDynamo + KV Block Manager + NIXL + Router + TensorRT-LLMPlanner, Model Express, Model OptimizerKubernetes DeploymentGPU Operator + KAI SchedulerNetwork Operator, GroveFull GenAI StackDynamo + NIXL + KV Block Manager + Router + Grove + KAI Scheduler + PlannerAIConfigurator, AIPerf Architecture Overview