YoVDO

Proximal Gradient Methods for Nonsmooth Nonconvex Minimax - A Unified Convergence Analysis

Offered By: Erwin Schrödinger International Institute for Mathematics and Physics (ESI) via YouTube

Tags

Nonconvex Optimization Courses Complexity Theory Courses

Course Description

Overview

Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Explore a 26-minute conference talk on proximal gradient methods for nonsmooth nonconvex minimax problems, delivered at the "One World Optimization Seminar in Vienna" workshop held at the Erwin Schrödinger International Institute for Mathematics and Physics. Delve into the analysis of parallel and alternating proximal gradient schemes within a unified framework, expanding on general convergence mechanisms for nonconvex nonsmooth optimization. Discover pointwise global convergence results and refined complexity analyses, departing from the common focus on nearly approximate stationary solutions. Learn how this approach broadens the scope of addressable minimax problems through Non Euclidean proximal steps, extending convergence and complexity results to a wider setting. Gain insights from this joint work with Eyal Cohen, which advances beyond typical weakly convex/concave and smooth models in recent literature.

Syllabus

Marc Teboulle - Proximal Gradient Methods for Nonsmooth Nonconvex Minimax: A Unified Convergence...


Taught by

Erwin Schrödinger International Institute for Mathematics and Physics (ESI)

Related Courses

The Next Generation of Infrastructure
Delft University of Technology via edX
The Beauty and Joy of Computing - AP® CS Principles Part 2
University of California, Berkeley via edX
Advanced Data Structures in Java
University of California, San Diego via Coursera
Theory of Computation
Indian Institute of Technology Kanpur via Swayam
离散数学
Shanghai Jiao Tong University via Coursera