Instructor-Led Workshop
Rapid Application Development Using Large Language Models

Recent advancements in both the techniques and accessibility of large language models (LLMs) have opened up unprecedented opportunities for businesses to streamline their operations, decrease expenses, and increase productivity at scale. Enterprises can also use LLM-powered apps to provide innovative and improved services to clients or strengthen customer relationships. For example, enterprises could provide customer support via AI virtual assistants or use sentiment analysis apps to extract valuable customer insights.

In this course, you’ll gain a strong understanding and practical knowledge of LLM application development by exploring the open-sourced ecosystem, including pretrained LLMs, that can help you get started quickly developing LLM-based applications.

 

Learning Objectives
 

By participating in this workshop, you’ll learn how to:
  • Find, pull in, and experiment with the HuggingFace model repository and the associated transformers API
  • Use encoder models for tasks like semantic analysis, embedding, question-answering, and zero-shot classification
  • Use decoder models to generate sequences like code, unbounded answers, and conversations
  • Use state management and composition techniques to guide LLMs for safe, effective, and accurate conversation

Datasheet (PDF 90 KB)

Workshop Outline

Introduction
(15 mins)
From Deep Learning to Large Language Models
(75 mins)

    Learn how large language models are structured and how to use them:

  • Review deep learning- and class-based reasoning, and see how language modeling falls out of it.
  • Discuss transformer architectures, interfaces, and intuitions, as well as how they scale up and alter to make state-of-the-art LLM solutions.
Break (15 mins)
Specialized Encoder Models
(45 mins)

    Learn how to look at the different task specifications:

  • Explore cutting-edge HuggingFace encoder models.
  • Use already-tuned models for interesting tasks such as token classification, sequence classification, range prediction, and zero-shot classification.
Break (60 mins)
Encoder-Decoder Models for Seq2Seq
(75 mins)

    Learn about forecasting LLMs for predicting unbounded sequences:

  • Introduce a decoder component for autoregressive text generation.
  • Discuss cross-attention for sequence-as-context formulations.
  • Discuss general approaches for multi-task, zero-shot reasoning.
  • Introduce multimodal formulation for sequences, and explore some examples.
Decoder Models for Text Generation
(45 mins)

    Learn about decoder-only GPT-style models and how they can be specified and used:

  • Explore when decoder-only is good, and talk about issues with the formation.
  • Discuss model size, special deployment techniques, and considerations.
  • Pull in some large text-generation models, and see how they work.
Break (15 mins)
Stateful LLMs
(60 mins)

    Learn how to elevate language models above stochastic parrots via context injection:

  • Show off modern LLM composition techniques for history and state management.
  • Discuss retrieval-augmented generation (RAG) for external environment access.
Assessment and Q&A
(60 mins)
  • Review key learnings.
  • Take a code-based assessment to earn a certificate.
 

Workshop Details

Duration: 8 hours

Price: Contact us for pricing.

Prerequisites:

Technologies: Python, PyTorch, HuggingFace, transformers, LangChain, LlamaIndex

Hardware Requirements: Desktop or laptop computer capable of running the latest version of Chrome or Firefox. Each participant will be provided with dedicated access to a fully configured, GPU-accelerated workstation in the cloud.

Certificate: Upon successful completion of the assessment, participants will receive an NVIDIA DLI certificate to recognize their subject matter competency and support professional career growth.

Languages: English

Upcoming Public Workshops

If your organization is interested in boosting and developing key skills in AI, accelerated data science, or accelerated computing, you can request instructor-led training from the NVIDIA DLI.

Continue Your Learning with These DLI Trainings

Getting Started with Image Segmentation

Modeling Time-Series Data with Recurrent Neural Networks in Keras

Building Transformer-Based Natural Language Processing Applications

Building Intelligent Recommender Systems

Data Parallelism: How to Train Deep Learning Models on Multiple GPUs

Questions?