Webinar: How to Accelerate DL Inference on NVIDIA® Jetson Orin™

Learn how to optimize your deep learning models for maximum speed and efficiency.

Deci and NVIDIA experts share practical tips and best practices to help you leverage the full potential of the NVIDIA® Jetson Orin™ platform, covering topics such as:

  • Hardware selection: How to select the optimal Jetson Orin model for your application
  • Quantization: How to reduce the precision of your neural network weights and activations to speed up inference while maintaining accuracy
  • TensorRT: How to use this NVIDIA library for optimizing deep learning models to achieve faster inference times
  • Batch size tuning: How to optimize the batch size for your model to improve inference performance
  • Multi-stream inference: How to process multiple input streams simultaneously on Jetson Orin
  • Asynchronous inference: How to maximize hardware utilization and performance with concurrent inference
  • Neural architecture search: How to accelerate inference with NAS

Through real-world examples and practical demonstrations, we’ll show you how to implement such techniques in your own machine learning projects to achieve faster processing speeds and unlock new possibilities.

Watch now, and if you want to learn more about accelerating the performance of your deep learning models, book a demo here.

You May Also Like

GTC Talk: How to Accelerate NLP Performance on GPU with Neural Architecture Search

Webinar: 5 Factors to Consider in Developing Deep Learning Projects

Deci 2023 Winter Release – Product Event

The Ultimate Guide to Inference Acceleration of Deep Learning-Based Applications

Learn 12 inference acceleration techniques that you can immediately implement to improve the speed, efficiency, and accuracy of your existing AI models.