Checkpoints#
There are two main ways to load pretrained checkpoints in NeMo:
Using the
restore_from()
method to load a local checkpoint file (.nemo), orUsing the
from_pretrained()
method to download and set up a checkpoint from NGC.
See the following sections for instructions and examples for each.
Note that these instructions are for loading fully trained checkpoints for evaluation or fine-tuning.
For resuming an unfinished training experiment, please use the experiment manager to do so by setting the
resume_if_exists
flag to True.
Loading Local Checkpoints#
NeMo will automatically save checkpoints of a model you are training in a .nemo format.
You can also manually save your models at any point using model.save_to(<checkpoint_path>.nemo)
.
If you have a local .nemo
checkpoint that you’d like to load, simply use the restore_from()
method:
import nemo.collections.asr as nemo_asr
model = nemo_asr.models.<MODEL_BASE_CLASS>.restore_from(restore_path="<path/to/checkpoint/file.nemo>")
Where the model base class is the ASR model class of the original checkpoint, or the general ASRModel class.
Inference#
The audio files should be 16KHz monochannel wav files.
Transcribe Audios to Semantics:
You may perform inference on a sample of speech after loading the model by using its ‘transcribe()’ method:
slu_model = nemo_asr.models.SLUIntentSlotBPEModel.from_pretrained(model_name="<MODEL_NAME>")
predictions = slu_model.transcribe([list of audio files], batch_size="<BATCH_SIZE>")
SLU Models#
Below is a list of all the Speech Intent Classification and Slot Filling models that are available in NeMo.
Model Name |
Model Base Class |
Model Card |
---|---|---|
slu_conformer_transformer_large_slurp |
SLUIntentSlotBPEModel |
https://ngc.nvidia.com/catalog/models/nvidia:nemo:slu_conformer_transformer_large_slurp |