YoVDO

Non-negative Gauss-Newton Methods for Empirical Risk Minimization

Offered By: Paul G. Allen School via YouTube

Tags

Machine Learning Courses Deep Learning Courses Reinforcement Learning Courses Stochastic Optimization Courses

Course Description

Overview

Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Explore a distinguished seminar on optimization and data featuring Lin Xiao from Facebook AI Research. Delve into non-negative Gauss-Newton methods for empirical risk minimization, focusing on minimizing the average of numerous smooth but potentially non-convex functions. Learn how reformulating non-negative loss functions allows for the application of Gauss-Newton or Levenberg-Marquardt methods, resulting in highly adaptive algorithms. Discover the convergence analysis of these methods in convex, non-convex, and stochastic settings, comparing their performance to classical gradient methods. Gain insights from Lin Xiao's extensive experience in optimization theory and algorithms for deep learning and reinforcement learning, drawing from his work at Meta's Fundamental AI Research team and previous roles at Microsoft Research and top academic institutions.

Syllabus

Distinguished Seminar in Optimization and Data: Lin Xiao (Facebook AI Research)


Taught by

Paul G. Allen School

Related Courses

Introduction to Artificial Intelligence
Stanford University via Udacity
Natural Language Processing
Columbia University via Coursera
Probabilistic Graphical Models 1: Representation
Stanford University via Coursera
Computer Vision: The Fundamentals
University of California, Berkeley via Coursera
Learning from Data (Introductory Machine Learning course)
California Institute of Technology via Independent