Free Hands-On Labs

AI Chatbot With Retrieval-Augmented Generation

Use an NVIDIA AI workflow to adapt an existing foundation model to accurately generate responses based on your enterprise data.

In This Free Hands-On Lab, Experience:

Deploying generative AI models—anywhere—using NVIDIA® NeMo™, and inflight batching with NVIDIA® TensorRT™-LLM to achieve low-latency and high-performance inference.

Pulling information and generating responses by populating a document ingestion pipeline that creates embeddings to knowledge in a GPU-accelerated vector database.

Using a large language model (LLM) inference pipeline with a real-time query router service and NVIDIA Triton™ Inference Server.

Build Production-Grade RAG

With an NVIDIA AI Enterprise subscription, unlock your business data with generative AI and enable better business insights in real time with enterprise-ready RAG.

This Lab Is a Collaboration Between: