YoVDO

Deploying Deep Learning Models for Inference at Production Scale

Offered By: Applied Singularity via YouTube

Tags

TensorRT Courses Deep Learning Courses Neural Networks Courses Model Deployment Courses GPU Acceleration Courses Data Centers Courses

Course Description

Overview

Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Explore a comprehensive session from NVIDIA Discovery Bengaluru focused on deploying AI models at production scale. Learn about two key NVIDIA resources: TensorRT, a deep learning platform optimizing neural network models and accelerating inference across GPU platforms, and Triton Inference Server, an open-source software providing a standardized inference platform for various infrastructures. Access the accompanying PowerPoint presentation for detailed insights. Gain valuable knowledge on the latest advancements in AI, machine learning, deep learning, and generative AI by joining the Applied Singularity Meetup group and downloading their free mobile app available on iOS and Android.

Syllabus

Deploying Deep Learning Models for Inference at Production Scale - at NVIDIA


Taught by

Applied Singularity

Related Courses

Optimize TensorFlow Models For Deployment with TensorRT
Coursera Project Network via Coursera
Jetson Xavier NX Developer Kit - Edge AI Supercomputer Features and Applications
Nvidia via YouTube
NVIDIA Jetson: Enabling AI-Powered Autonomous Machines at Scale
Nvidia via YouTube
Jetson AGX Xavier: Architecture and Applications for Autonomous Machines
Nvidia via YouTube
Streamline Deep Learning for Video Analytics with DeepStream SDK 2.0
Nvidia via YouTube