NVIDIA DGX Cloud Lepton

Connecting developers to a global network of GPU compute.

Overview

Build and Deploy AI Across Your Compute

Meet NVIDIA DGX™ Cloud Lepton, an AI platform that unifies your compute so you can build, train, and deploy without thinking about infrastructure. Designed for AI natives, model builders, and teams that iterate quickly, DGX Cloud Lepton brings compute environments together into one consistent experience for development, training, and inference. You keep a single workflow, avoid rearchitecting when compute changes, and move from prototype to production faster across the regions and providers you choose.

DGX Cloud Lepton Simplifies Multi-Cloud AI Development

NVIDIA DGX Cloud Lepton bridges the gap between AI demand and global GPU supply.

DGX Cloud Lepton Enables AI-First Neuroscience Research

Learn how Prima Mente built Pleiades, the world’s first whole-genome epigenetic foundation model, using NVIDIA DGX Cloud Lepton.

Demo

Your Virtual Global AI Factory

NVIDIA DGX Cloud Lepton is an AI platform that connects developers to global GPU compute across a network of cloud providers. This platform offers developers a unified experience and unmatched flexibility to develop and scale AI apps across multi-cloud environments. 

How It Works

What Is NVIDIA DGX Cloud Lepton?

DGX Cloud Lepton brings your AI compute together to provide a unified experience for development, training, and inference. The platform includes integrated tools that streamline the path from prototype to production without the complexity of managing underlying infrastructure.

DGX Cloud Lepton brings together a global network of NVIDIA Cloud Partners (NCPs), GPU marketplaces, cloud providers, and local environments to streamline discovery, development, and deployment of AI workloads in a single, developer-friendly platform.

Features and Benefits

Go From Prototype to Production Faster

Start building with instant access to NVIDIA’s accelerated APIs at build.nvidia.com—including serverless endpoints, prebuilt NVIDIA NIM™ microservices, and GPU-backed compute. When it’s time to scale, NVIDIA DGX Cloud Lepton powers seamless customization and deployment across a global network of GPU cloud providers.

Frictionless Deployment Across Any GPU Cloud

Decouple the AI PaaS platform from the underlying infrastructure to deploy AI applications across multi-cloud environments with minimal operational burden, leveraging integrated services for inference, testing, and training workloads.

Run Where Your Data Lives

Bring compute from specific regions, achieving compliance with data sovereignty regulations and meeting low-latency requirements for sensitive workloads.

Unmatched Flexibility and Productivity

Boost productivity with a unified experience delivered across development, training, and inferencing, including the ability to bring best-fit GPUs to DGX Cloud Lepton.

Predictable Performance

Through cloud partners in the DGX Cloud Lepton marketplace, experience enterprise-grade performance, reliability, and security.

Ecosystem

DGX Cloud Lepton Partners

Access NVIDIA accelerated computing from your choice of regions through a vast network of cloud providers.

Next Steps

Discover NVIDIA Cloud Accelerator

Operate AI clouds reliably and efficiently at scale with a portfolio of open, modular infrastructure software components.

Measure What Matters With Performance Benchmarking

Optimize AI workload performance on any NVIDIA infrastructure with NVIDIA performance benchmarking.

Scale AI Inference Workloads With NVIDIA Cloud Functions

Deploy and scale agentic AI, physical AI, and simulation workloads using a unified API layer on build.nvidia.com.