Skip to content

Latest commit

 

History

History
17 lines (8 loc) · 828 Bytes

README.md

File metadata and controls

17 lines (8 loc) · 828 Bytes

Tutorial on inference optimization using TensorRT

This repository contains code to accompany a tutorial given at NCSA on inference optimization for deep learning models using NVIDIA TensorRT.

Description:

In many applications of deep learning models, we would benefit from reduced latency (time taken for inference). This tutorial will introduce NVIDIA TensorRT, an SDK for high-performance deep learning inference. We will go through all the steps necessary to convert a trained deep learning model to an inference-optimized model.

Webinar Date: April 13, 2022 Speakers: Nikil Ravi and Pranshu Chaturvedi, UIUC

Watch the tutorial on YouTube

Watch the video