Ai4 Talk: The Importance of Model Design in Building Production-Grade CV Applications

As AI applications power a growing number of use cases and industries, they are setting new and demanding requirements for inference performance. However, running successful inference at scale requires meeting various criteria such as accuracy, latency, throughput, and model size, among others.

In this talk, we’ll cover the various approaches to computer vision model design, the common mistakes made, and how these impact inference performance. We’ll outline a new algorithmic optimization approach that is based on Neural Architecture Search (NAS) technology. You will gain insight into how NAS can be leveraged to build production-grade models, accelerate time to market, and reduce inference compute costs.

You May Also Like

Webinar: How to Accelerate DL Inference on NVIDIA® Jetson Orin™

GTC Talk: How to Accelerate NLP Performance on GPU with Neural Architecture Search

Webinar: 5 Factors to Consider in Developing Deep Learning Projects

The Ultimate Guide to Inference Acceleration of Deep Learning-Based Applications

Learn 12 inference acceleration techniques that you can immediately implement to improve the speed, efficiency, and accuracy of your existing AI models.