There is no product that matches the search criteria.
Your cart is empty.
Powering Scale-Out AI Training and Inference
Supercharge any server with NVIDIA® T4 GPU, the world’s most performant scale-out accelerator. Its low-profile, 70W design is powered by NVIDIA Turing™ Tensor Cores, delivering revolutionary multi-precision performance to accelerate a wide range of modern applications. This advanced GPU is packaged in an energy-efficient 70-watt, small PCIe form factor, optimized for scale-out servers and purpose-built to deliver state-of-the-art AI.
2560 / 320
CUDA / Tensor Cores
16GB
GDDR6 Memory
8.1 Tflops
Single Precision Performance
Breakthrough Performance
T4 introduces the revolutionary Turing Tensor Core technology with multi-precision computing to handle diverse workloads. Powering breakthrough performance from FP32 to FP16 to INT8, as well as INT4 precisions, T4 delivers up to 40X higher performance than CPUs.
Small form factor 70-watt (W) design
Makes T4 optimised for scale-out servers, providing an incredible 50X higher energy efficiency compared to CPUs, drastically reducing operational costs. In the last two years, NVIDIA’s Inference Platform has increased efficiency by over 10X, and remains the most energy-efficient solution for distributed AI training and inference.
NVIDIA T4 data center GPU
is the ideal universal accelerator for distributed computing environments. Revolutionary multi-precision performance accelerates deep learning and machine learning training and inference, video transcoding, and virtual desktops. T4 supports all AI frameworks and network types, delivering dramatic performance and efficiency that maximise the utility of at-scale deployments.
Turing Tensor Core technology
with multi-precision computing for AI powers breakthrough performance from FP32 to FP16 to INT8, as well as INT4 precisions. It delivers up to 9.3X higher performance than CPUs on training and up to 36X on inference.