The AI factory foundation for AI reasoning.
NVIDIA DGX™ B300 is the powerhouse for AI innovators, delivering the hyperscaler performance needed to build a modern AI factory. Powered by NVIDIA Blackwell Ultra GPUs, DGX B300 boosts dense FP4 performance by 1.5x and attention performance by 2x over DGX B200, all in a new form factor designed to fit seamlessly into the modern data center. Compatible with NVIDIA MGX™ and traditional enterprise racks and with full-stack software, it simplifies and streamlines AI deployment, enabling any enterprise to run like a hyperscaler.
Powered by NVIDIA Blackwell Ultra GPUs, DGX B300 provides enterprises with a single platform to accelerate large language model (LLM) inference and training. Delivering 144 petaFLOPS of inference performance, the system enables every business to operate like a hyperscaler.
With multiple power options to choose from, NVIDIA DGX B300 is designed to be the most energy-efficient AI supercomputer, delivering unmatched energy efficiency and performance per watt.
NVIDIA DGX B300 has been redesigned for the modern data center, deployable in NVIDIA MGX racks for the first time. This new industry standard is powering a shift in data center engineering, making it easier than ever to obtain breakthrough performance and efficiency.
Explore the new features and capabilities, including AC and DC power options, that make DGX B300 easy to integrate into any modern data center, with greater deployment flexibility than ever before.
| GPUs | 8x NVIDIA Blackwell Ultra SXM |
| CPU | Intel® Xeon® 6776P Processors |
| Total GPU Memory | 2.1 TB |
| Performance | FP4 Tensor Core: 144 PFLOPS | 108 PFLOPS* FP8 Tensor Core: 72 PFLOPS** |
| NVIDIA NVLink™ Switch System | 2x |
| NVIDIA NVLink Bandwidth | 14.4 TB/s aggregate bandwidth |
| Networking | 8x OSFP ports serving 8x single-port NVIDIA ConnectX-8 VPI
|
| Management Network | 1GbE onboard NIC with RJ45 1GbE RJ45 Host baseboard management controller (BMC) |
| Storage | OS: 2x 1.9 TB NVMe M.2 Internal storage: 8x 3.84 TB NVMe E1.S |
| Power Consumption | ~14 kW |
| Software | NVIDIA AI Enterprise—Optimized AI software NVIDIA Mission Control—AI data center operations and orchestration with NVIDIA Run:ai technology NVIDIA DGX OS—Operating system Supports Red Hat Enterprise Linux/Rocky/Ubuntu |
| Rack Units | 10U |
| Support | Three-year business-standard hardware and software support |
|
*Specification shown in sparse | dense |
|
NVIDIA Blackwell Ultra is the GPU architecture for the most powerful systems for AI inference available today, including NVIDIA DGX B300 and DGX GB300. Blackwell Ultra systems deliver up to 50x higher throughput per megawatt and up to 35x lower cost per token than NVIDIA Hopper™ for low-latency agentic workloads, through hardware–software codesign, according to SemiAnalysis InferenceX benchmarks (Q1 2026).
Yes. The large memory spaces available with Blackwell Ultra-powered systems enable DeepSeek-R1 (671B MoE) inference on fewer GPUs with lower tensor parallelism overhead. In MLPerf Inference v6.0 (April 2026), systems powered by NVIDIA Blackwell Ultra GPUs delivered the highest throughput across the widest range of models and scenarios. On DeepSeek-R1, Blackwell Ultra systems delivered 2.5 million tokens per second—up to 2.7x higher token throughput compared to Blackwell Ultra debut submissions just six months prior, as a result of NVIDIA TensorRT™-LLM software updates.
NVIDIA Blackwell Ultra delivers AI Inference at $0.24 per million tokens at 102 TPS/user on DeepSeek-R1 using NVIDIA Dynamo TensorRT-LLM and MTP, according to SemiAnalysis InferenceX benchmarks as of April 2026.
Deploy NVIDIA DGX B300 today on premises, in a colocation facility, or in the cloud through one of our partners.
The NVIDIA DGX platform is the proven standard on which enterprise AI is built.
Reach out to an NVIDIA product specialist about your professional needs.