Instructor-Led Workshop
Very large deep neural networks (DNNs), whether applied to natural language processing (e.g., GPT-3), computer vision (e.g., huge Vision Transformers), or speech AI (e.g., Wave2Vec 2) have certain properties that set them apart from their smaller counterparts. As DNNs become larger and are trained on progressively larger datasets, they can adapt to new tasks with just a handful of training examples, accelerating the route toward general artificial intelligence. Training models that contain tens to hundreds of billions of parameters on vast datasets isn’t trivial and requires a unique combination of AI, high-performance computing (HPC), and systems knowledge. The goal of this course is to demonstrate how to train the largest of neural networks and deploy them to production.
By participating in this workshop, you’ll learn how to::
Introduction (15 mins) |
|
Introduction to Training of Large Models (120 mins) |
|
Model Parallelism: Advanced Topics (120 mins) |
|
Inference of Large Models (120 mins) |
|
Final Review (15 mins) |
|
Start of workshop: 9:00 am. Breaks will be allocated by the instructor, including one hour for lunch. Workshop finishes by 5:30 pm.
Duration: 8 hours
Prerequisites:
Technologies: PyTorch, Megatron-LM, DeepSpeed, Slurm, Triton Inference Server, NVIDIA Nsight™
Certificate: Upon successful completion of the assessment, participants will receive an NVIDIA DLI certificate to recognize their subject matter competency and support professional career growth.
Hardware Requirements: Laptop computer capable of running the latest version of Chrome or Firefox. Each participant will be provided with dedicated access to a fully configured, GPU-accelerated workstation in the cloud.
Language: English