SkillAgentSearch skills...

Automodel

Pytorch Distributed native training library for LLMs/VLMs with OOTB Hugging Face support

Install / Use

/learn @NVIDIA-NeMo/Automodel

README

<div align="center">

🚀 NeMo AutoModel

</div> <div align="center"> <!-- [![License](https://img.shields.io/badge/License-Apache%202.0-blue.svg)](https://opensource.org/licenses/Apache-2.0) -->

codecov CICD NeMo Python 3.10+ Contributions Welcome GitHub Stars

<!-- **Day-0 integration with Hugging Face models automating fine-tuning and pretraining with pytorch-native parallelism, custom-kernels and optimized recipes** **Pytorch DTensor‑native SPMD library for large‑scale training**-->

📖 Documentation🔥 Ready-to-Use Recipes💡 ExamplesModel CoveragePerformance🤝 Contributing

</div>

📣 News and Discussions

Overview

Nemo AutoModel is a Pytorch DTensor‑native SPMD open-source training library under NVIDIA NeMo Framework, designed to streamline and scale training and finetuning for LLMs and VLMs. Designed for flexibility, reproducibility, and scale, NeMo AutoModel enables both small-scale experiments and massive multi-GPU, multi-node deployments for fast experimentation in research and production environments.

<p align="center"> <a href="https://github.com/NVIDIA-NeMo/Automodel"><picture> <source media="(prefers-color-scheme: light)" srcset="https://raw.githubusercontent.com/NVIDIA-NeMo/Automodel/refs/heads/main/docs/automodel_diagram.png"> <img alt="AutoModel Logo" src="https://raw.githubusercontent.com/NVIDIA-NeMo/Automodel/refs/heads/main/docs/automodel_diagram.png"> </picture></a> </p>

What you can expect:

  • Hackable with a modular design that allows easy integration, customization, and quick research prototypes.
  • Minimal ceremony: YAML-driven recipes; override any field using CLI.
  • High performance and flexibility with custom kernels and DTensor support.
  • Seamless integration with Hugging Face for day-0 model support, ease of use, and wide range of supported models.
  • Efficient resource management using Kubernetes and Slurm, enabling scalable and flexible deployment across configurations.
  • Documentation with step-by-step guides and runnable examples.
<!-- Please refer to our design documents for more details on the architecture and design philosophy. --> <!-- NeMo Framework is NVIDIA's GPU accelerated, end-to-end training framework for large language models (LLMs), multi-modal models and speech models. It enables seamless scaling of training (both pretraining and post-training) workloads from single GPU to thousand-node clusters for both 🤗Hugging Face/PyTorch and Megatron models. It includes a suite of libraries and recipe collections to help users train models from end to end. The **AutoModel library ("NeMo AutoModel")** provides GPU-accelerated PyTorch training for 🤗Hugging Face models on **Day-0**. Users can start training and fine-tuning models instantly without conversion delays, scale effortlessly with PyTorch-native parallelisms, optimized custom kernels, and memory-efficient recipes-all while preserving the original checkpoint format for seamless use across the Hugging Face ecosystem. -->

⚠️ Note: NeMo AutoModel is under active development. New features, improvements, and documentation updates are released regularly. We are w

View on GitHub
GitHub Stars395
CategoryCustomer
Updated26m ago
Forks102

Languages

Python

Security Score

100/100

Audited on Mar 28, 2026

No findings