NVIDIA’s main deep learning framework for NLP, Speech, Audio, and Vision. It supports end-to-end development and deployment of large language models on-premnises, data center or public cloud. Nemo supports execution on Slurm or Kubernetes environment.
It provides:
pip install nemo-toolkit[all]
Use NeMo-Run or Python script to:
You can use NeMo’s built-in preprocessors or write a simple script — NeMo provides utilities.
With NeMo 2.0 + Megatron-Bridge: Example with NeMo’s T5 training recipe:
nemo-run \
--config-path /path/to/configs \
--config-name nemo_t5_training_config.yaml \
model.train_ds.data_prefix=/path/to/bin_prefix \
model.validation_ds.data_prefix=/path/to/bin_prefix
NeMo-Run wraps torchrun behind the scenes.
After training completes:
nemo-run \
--config-path eval_configs \
--config-name t5_eval.yaml \
model.restore_path=/path/to/checkpoint.ckpt
You get BLEU / accuracy / perplexity metrics.
You can run:
from nemo.collections.nlp.models import TextGenerationModel
model = TextGenerationModel.restore_from("my_model.nemo")
response = model.generate(["translate Akkadian to English: ana ilu"])
Or export to ONNX / TensorRT for deployment.