NeMo Megatron is an end-to-end framework for training and deploying LLMs with billions or trillions of parameters.
The containerized framework delivers high training efficiency across thousands of GPUs and makes it practical for enterprises to build and deploy large-scale models. It provides capabilities to curate training data, train large-scale models up to trillions of parameters, customize using prompt learning, and deploy using the NVIDIA Triton™ Inference Server to run large-scale models on multiple GPUs and multiple nodes.
NeMo Megaton is optimized to run on NVIDIA DGX™ Foundry, NVIDIA DGX SuperPOD™, Amazon Web Services, Microsoft Azure, and Oracle Cloud Infrastructure.