NVIDIA Tesla P40 24GB GDDR5 900-2G610-0000-000 Accelerator is purpose-built to deliver maximum throughput for deep learning deployment. With 47 TOPS (Tera-Operations Per Second) of inference performance and INT8 operations per GPU, a single server with 8 Tesla P40s delivers the performance of over 140 CPU servers. As models increase in accuracy and complexity, CPUs are no longer capable of delivering interactive user experience. The Tesla P40 delivers over 30X lower latency than a CPU for real-time responsiveness in even the most complex models.
The world's fastest processor for inference workloads
47 TOPS of INT8 for maximum inference throughput and responsiveness
Hardware-decode engine capable of transcoding and inferencing 45 HD video streams in real time