New Update to the NVIDIA Deep Learning SDK Now Help Accelerate Inference

The latest update to the NVIDIA Deep Learning SDK includes the NVIDIA TensorRT deep learning inference engine (formerly GIE) and the new NVIDIA Deep Stream SDK.

TensorRT delivers high performance inference for production deployment of deep learning applications. The latest release delivers up to 3x more throughput, using 61% less memory with new INT8 optimized precision.

TensorRT charts

NVIDIA DeepStream SDK simplifies development of high performance video analytics applications powered by deep learning. Using a high-level C++ API and high performance runtime, developers can use the SDK to rapidly integrate advanced video inference capabilities including optimized precision and GPU-accelerated transcoding to deliver faster, more responsive AI-powered services such as real-time video categorization.

Learn more about the NVIDIA Deep Learning SDK >