Overview#
NeMo AutoModel integrates with Hugging Face transformers. As a result, any LLM or VLM that can be instantiated through transformers can also be used via NeMo AutoModel, subject to runtime, third-party software dependencies and feature compatibility.
Version compatibility and Day‑0 support#
AutoModel closely tracks the latest
transformersversion and updates its dependency on a regular basis.New models released on the Hugging Face Hub may require the latest
transformersversion, necessitating a package upgrade.We are working on introducing a continuous integration (CI) pipeline that will automatically bump the supported
transformersversion as soon as a new release is detected. This will enable even faster support for the newest Hugging Face models.
Note: To use newly released models, you may need to upgrade your NeMo AutoModel installation. This process is similar to upgrading transformers itself to access the latest model support. In practice, the upgrade behavior and familiar Auto* APIs mirror transformers, with AutoModel additionally providing optional performance accelerations and distributed training features.
Extended model support with NeMo AutoModel’s custom model registry#
NeMo AutoModel includes a custom model registry that allows teams to:
Add custom implementations to extend support to models not yet covered upstream.
Provide optimized, extended or faster implementations for specific models while retaining the same AutoModel interface.
Supported Hugging Face Auto classes#
Auto class |
Task |
Status |
Notes |
|---|---|---|---|
|
Text Generation (LLM) |
Supported |
|
|
Image-Text-to-Text (VLM) |
Supported |
|
|
Sequence Classification |
WIP |
Early support; interfaces may change. |
When a model listed on Hugging Face Hub may not be supported#
Sometimes a model listed on the Hugging Face Hub may not support finetuning in NeMo AutoModel. If you encounter any such model, please open a GitHub issue requesting support by sharing the model-id of interest as well as any stack trace you see. We summarize here some cases (non-exhaustive):
Issue |
Example Error Message |
Solution |
|---|---|---|
Model has explicitly disabled training functionality in the model-definition code. |
— |
Make the model available via our custom registry. Please open a new GitHub issue, requesting support. |
Model requires newer transformers version |
The checkpoint you are trying to load has model type |
Upgrade the transformers version you use, and/or open a new GitHub issue. |
Model upper-bounds transformer version, requiring older version |
— |
Open a new GitHub issue. |
Unsupported checkpoint format |
OSError: |
Open a new GitHub issue or request from the model publisher to share a safetensors checkpoint. |
These cases typically stem from upstream packaging or dependency constraints and would surface similarly when using transformers directly. AutoModel mirrors the familiar load/finetune semantics.
If you encounter any issue, you can try:
Upgrade to a NeMo AutoModel release that supports the required
transformersversion.If the model uses custom code, set
trust_remote_code=Truewhen loading.Open a GitHub issue with the model-id and error for us to prioritize support or add a registry-backed implementation.