Interpretable Representation Learning for Visual Intelligence
Offered By: Bolei Zhou via YouTube
Course Description
Overview
Explore a comprehensive thesis defense presentation on interpretable representation learning for visual intelligence. Delve into deep neural networks for object classification, network visualization techniques, and interpretable representations for objects and scenes. Learn about class activation mapping for explaining deep neural network predictions, weakly-supervised localization, and temporal relational networks for event recognition. Gain insights into the interpretability of medical models and understand the contributions made to the field of visual intelligence.
Syllabus
Intro
Deep Neural Networks for Object Classification
Interpretability of Deep Neural Networks
Thesis Outline
Object Classification vs. Scene Recognition
Visualizing Units
Related Work on Network Visualization
Annotating the Interpretation of Units
Interpretable Representations for Objects and Scenes
Evaluate Unit for Semantic Segmentation
IMAGENET Pretrained Network
Class Activation Mapping: Explain Prediction of Deep Neural Network
Evaluation on Weakly-Supervised Localization
Explaining the Failure Cases in Video
Interpreting Medical Models
Summary of Contributions
Temporal Relational Networks for Event Recognition
Acknowledgement
Taught by
Bolei Zhou
Related Courses
2D image processingHigher School of Economics via Coursera 3D Reconstruction - Multiple Viewpoints
Columbia University via Coursera 3D Reconstruction - Single Viewpoint
Columbia University via Coursera AI-900: Microsoft Certified Azure AI Fundamentals
A Cloud Guru TensorFlow Developer Certificate Exam Prep
A Cloud Guru