NVIDIA T4

Flexible Design, Extraordinary Performance

Next-Level Acceleration Has Arrived

We’re racing toward the future where every customer interaction, every product, and every service offering will be touched and improved by AI. Realizing that the future requires a computing platform that can accelerate the full diversity of modern AI, enabling businesses to create new customer experiences, reimagine how they meet—and exceed—customer demands, and cost-effectively scale their AI-based products and services.

The NVIDIA® T4 GPU accelerates diverse cloud workloads, including high-performance computing, deep learning training and inference, machine learning, data analytics, and graphics. Based on the new NVIDIA Turing architecture and packaged in an energy-efficient 70-watt, small PCIe form factor, T4 is optimized for mainstream computing environments and features multi-precision Turing Tensor Cores and new RT Cores. Combined with accelerated containerized software stacks from NGC, T4 delivers revolutionary performance at scale.

Virtual Compute-Intensive Server Workloads

Virtual Compute-Intensive Server Workloads

NVIDIA Virtual Compute Server (vCS) provides the ability to virtualize GPUs and accelerate compute-intensive server workloads, including AI, Deep Learning, and Data Science. With this, automotive manufacturers can use the latest in simulation and compute technologies to create the most fuel efficient and stylish designs and researchers can analyze the function of genes to develop medical treatments more quickly.

A new partnership is advancing the hybrid cloud to power modern enterprise workloads.

Extraordinary Performance

T4 introduces the revolutionary Turing Tensor Core technology with multi-precision computing to handle diverse workloads. Powering extraordinary performance from FP32 to FP16 to INT8, as well as INT4 precisions, T4 delivers up to 40X higher performance than CPUs.

Breakthrough Inference Performance

See How You Can Accelerate Your AI Models With Mixed Precision on Tensor Cores

State-of-the-art Inference in Real-time

Responsiveness is key to user engagement for services such as conversational AI, recommender systems, and visual search. As models increase in accuracy and complexity, delivering the right answer right now requires exponentially larger compute capability. T4 delivers up to 40X times better throughput, so more requests can be served in real time.

T4 Inference Performance

Resnet50

DeepSpeech2

GNMT

Video Transcoding Performance

As the volume of online videos continues to grow exponentially, demand for solutions to efficiently search and gain insights from video continues to grow as well. T4 delivers extraordinary performance for AI video applications, with dedicated hardware transcoding engines that bring twice the decoding performance of prior-generation GPUs. T4 can decode up to 38 full-HD video streams, making it easy to integrate scalable deep learning into video pipelines to deliver innovative, smart video services.

NVIDIA T4 Specifications

 

Performance

Turing Tensor Cores
320

NVIDIA CUDA® cores
2,560

Single Precision Performance (FP32)
8.1 TFLOPS

Mixed Precision (FP16/FP32)
65 FP16 TFLOPS

INT8 Precision
130 INT8 TOPS

INT4 Precision
260 INT4 TOPS

 
 

Interconnect

Gen3
x16 PCIe

 
 

Memory

Capacity
16 GB GDDR6

Bandwidth
320+ GB/s

 
 

Power

70 watts

 

NVIDIA AI Inference Platform

Explore the World's Most Advanced Inference Platform.