New AI Technologies Announced at GTC 2020 Keynote

At GTC 2020, NVIDIA announced updates to 80 SDKs, including tools to help developers build AI-powered video streaming solutions, conversational AI, recommendation systems and more.

Announcing NVIDIA Maxine

Today we announced NVIDIA Maxine – a cloud native video streaming AI platform for services such as video conferencing. It includes state-of-the-art AI models and optimized pipelines that can run several features in real time in the cloud.

Sign up for Early Access to Maxine here.


Announcing TensorRT 7.2

Today NVIDIA announced TensorRT 7.2,  the latest version of its high-performance deep learning inference SDK. 

Highlights include:

  • 30x faster AI-effects vs CPU for video-based workloads enable super-resolution, noise removal and virtual backgrounds to run in real time.
  • 2.5x faster recommenders with optimizations for fully connected layers used in MLPs
  • 2x lower latency for RNNs vs earlier, enables apps such as real time Fraud and Anomaly detection

TensorRT 7.2 will be available in Q4, 2020 from the TensorRT page. The latest version of samples, parsers and notebooks are always available in the TensorRT open source repo.

Add this GTC session to your calendar to learn more: Get the Highest Inference Performance using TensorRT


Announcing NVIDIA Jarvis Open Beta – Application Framework for Multimodal Conversational AI 

Today NVIDIA announced Jarvis Open Beta, a fully accelerated application framework for enterprises to build multimodal conversational AI services that run in real-time on GPUs.  It includes state-of-the-art DL models, tools for composing new models, transfer learning, deployment, as well as optimized services that run under 300 ms latency. Jarvis cuts end-to-end conversational AI latency to half and offers 7x higher throughput vs CPUs on the SpeechSquad benchmark.

As part of Jarvis, we also announced NeMo 1.0 Beta. NeMo is an  open-Source toolkit to develop state-of-the-art conversational AI models in three lines of code. In the latest version you get:

  • Simplified APIs based on Pytorch Lightning
  • Built on PyTorch, interoperable with PyTorch modules
  • Easy customization of models with popular Hydra framework integration.

This version of NeMo is optimized on A100 as well as earlier architectures with Tensor Cores. You can get the new version of NeMo from here.

Resources:

Add these GTC sessions to your calendar to learn more:


Announcing NVIDIA Merlin Open Beta – Democratizing Large-Scale Deep Learning Recommenders

Today NVIDIA announced the latest release of NVIDIA Merlin, an open beta application framework that enables the end-to-end development of deep learning recommender systems, from data preprocessing to model training and inference, all accelerated on NVIDIA GPUs. With this release, Merlin addresses common pain points around optimization and interoperability. During preliminary testing, Merlin provides faster training on GPU in TensorFlow and PyTorch using our data loaders than default data loaders. This latest release reaffirms the NVIDIA commitment to accelerating the workflow of researchers, data scientists, and machine learning engineers and democratizing the development of large-scale deep learning recommender systems.

Add this GTC session to your calendar to learn more:

Resources:

Register for GTC this week for more on the latest GPU-accelerated AI technologies.