YoVDO

Sharp Oracle Inequalities for Non-Convex Loss - Lecture 1

Offered By: Georgia Tech Research via YouTube

Tags

Statistics & Probability Courses Least Squares Courses Regularization Courses

Course Description

Overview

Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Explore the first lecture in the TRIAD Distinguished Lecture Series featuring Sara van de Geer from ETH Zurich, focusing on Sharp Oracle Inequalities for Non-Convex Loss. Delve into the concept of M-estimation, a statistical procedure aimed at minimizing risk functions to achieve optimal data fit. Examine the role of regularization in preventing overfitting, with particular emphasis on the l₁-penalty and its application in the Lasso estimation procedure. Investigate the conditions necessary for penalties to yield sparsity oracle inequalities and their compatibility with risk functions. Gain insights into the theoretical foundations of statistical learning and optimization techniques in this 58-minute lecture presented by Georgia Tech Research.

Syllabus

TRIAD Distinguished Lecture Series | Sara van de Geer | ETH Zurich| Lecture 1 (of 3)


Taught by

Georgia Tech Research

Related Courses

Data Analysis and Visualization
Georgia Institute of Technology via Udacity
Improving Deep Neural Networks: Hyperparameter Tuning, Regularization and Optimization
DeepLearning.AI via Coursera
機器學習基石下 (Machine Learning Foundations)---Algorithmic Foundations
National Taiwan University via Coursera
Data Science: Machine Learning
Harvard University via edX
Art and Science of Machine Learning auf Deutsch
Google Cloud via Coursera