nemo_curator.core.serve.server
nemo_curator.core.serve.server
Module Contents
Classes
Functions
Data
API
Dataclass
Serve one or more models behind a typed backend config.
_backend_impl
_host
_started
backend
endpoint
OpenAI-compatible base URL for the served models.
health_check_timeout_s
models
name
port
verbose
Check every model is accepted by the backend and that all models share one concrete type.
Poll /v1/models until all expected models appear in the response.
Deploy all models and wait for them to become healthy.
Shut down the active inference backend and release resources.
Check whether any inference server is currently running in this process.