Academics

Efficient natural gradient method for large-scale optimization problems

Time:10:00-11:00 am, Wednesday March 8th, 2023

Venue:Conference Room 1, Floor 1 Jin Chun Yuan West Building

Organizer:应用与计算数学团队

Speaker:Yunan Yang (ETH Zürich, Switzerland)

Abstract

First-order methods are workhorses for large-scale optimization problems, but they are often agnostic to the structural properties of the problem under consideration and suffer from slow convergence, being trapped in bad local minima, etc. Natural gradient descent is an acceleration technique in optimization that takes advantage of the problem's geometric structure and preconditions the objective function's gradient by a suitable "natural" metric. Despite its success in machine learning, the natural gradient descent method is far from a mainstream computational technique due to the computational complexity of calculating and inverting the preconditioning matrix. This work aims at a unified computational framework and streamlining the computation of a general natural gradient flow via efficient tools from numerical linear algebra. We obtain robust numerical methods for natural gradient flows without directly calculating, storing, or inverting the dense preconditioning matrix. We treat various natural gradients in a unified framework for any loss function.


Speaker

Yunan Yang is an applied mathematician working in inverse problems, optimization, and applied optimal transport. Currently, Yunan is an Advanced Fellow at the Institute for Theoretical Studies at ETH Zurich. She will be a Tenure-Track Assistant Professor in the Department of Mathematics at Cornell University starting in July 2023. Yunan Yang earned a Ph.D. degree in mathematics from the University of Texas at Austin in 2018, supervised by Prof. Bjorn Engquist. From September 2018 to August 2021, Yunan was a Courant Instructor at the Courant Institute of Mathematical Sciences, New York University.

DATEMarch 8, 2023
SHARE
Related News
    • 0

      AdaBB: A Parameter-Free Gradient Method for Convex Optimization

      Organizer:Chenglong Bao 包承龙Speaker:Shiqian Ma (Rice University)Time:Thur., 11:00 am -12:00, Nov. 14, 2024Online:Tencent Meeting:127-784-846Title:AdaBB: A Parameter-Free Gradient Method for Convex OptimizationAbstract:We propose AdaBB, an adaptive gradient method based on the Barzilai-Borwein stepsize. The algorithm is line-search-free and parameter-free, and essentially provides a conv...

    • 1

      Towards Provably Efficient Quantum Algorithms for Nonlinear Dynamics and Large-scale Machine Learning Models

      AbstractNonlinear dynamics play a prominent role in many domains and are notoriously difficult to solve. Whereas previous quantum algorithms for general nonlinear equations have been severely limited due to the linearity of quantum mechanics, we gave the first efficient quantum algorithm for nonlinear differential equations with sufficiently strong dissipation. This is an exponential improvemen...