YoVDO

Model Orthogonalization - Class Distance Hardening in Neural Networks for Better Security

Offered By: IEEE via YouTube

Tags

IEEE Symposium on Security and Privacy Courses Cybersecurity Courses Machine Learning Courses Neural Networks Courses

Course Description

Overview

Explore a 19-minute IEEE conference talk on enhancing neural network security through model orthogonalization and class distance hardening. Delve into the research conducted by a team from Purdue University, including Guanhong Tao, Yingqi Liu, Guangyu Shen, Qiuling Xu, Shengwei An, Zhuo Zhang, and Xiangyu Zhang. Learn about innovative techniques to improve the robustness and reliability of neural networks in the face of security challenges, and gain insights into the potential applications of these methods in various fields of artificial intelligence and machine learning.

Syllabus

Model Orthogonalization: Class Distance Hardening in Neural Networks for Better Security


Taught by

IEEE Symposium on Security and Privacy

Tags

Related Courses

Sensor Security
IEEE via YouTube
Tracking Ransomware End-to-end
IEEE via YouTube
Cinderella - Turning Shabby X.509 Certificates into Elegant Anonymous Credentials with the Magic of Verifiable Computation
IEEE via YouTube
Algorithmic Transparency via Quantitative Input Influence - Theory and Experiments with Learning Systems
IEEE via YouTube
Bitcoin Over Tor Isn't a Good Idea
IEEE via YouTube