清华主页 EN
导航菜单

Efficient natural gradient method for large-scale optimization problems

来源: 03-08

时间:10:00-11:00 am, Wednesday March 8th, 2023

地点:Conference Room 1, Floor 1 Jin Chun Yuan West Building

组织者:应用与计算数学团队

主讲人:Yunan Yang (ETH Zürich, Switzerland)

Abstract

First-order methods are workhorses for large-scale optimization problems, but they are often agnostic to the structural properties of the problem under consideration and suffer from slow convergence, being trapped in bad local minima, etc. Natural gradient descent is an acceleration technique in optimization that takes advantage of the problem's geometric structure and preconditions the objective function's gradient by a suitable "natural" metric. Despite its success in machine learning, the natural gradient descent method is far from a mainstream computational technique due to the computational complexity of calculating and inverting the preconditioning matrix. This work aims at a unified computational framework and streamlining the computation of a general natural gradient flow via efficient tools from numerical linear algebra. We obtain robust numerical methods for natural gradient flows without directly calculating, storing, or inverting the dense preconditioning matrix. We treat various natural gradients in a unified framework for any loss function.


Speaker

Yunan Yang is an applied mathematician working in inverse problems, optimization, and applied optimal transport. Currently, Yunan is an Advanced Fellow at the Institute for Theoretical Studies at ETH Zurich. She will be a Tenure-Track Assistant Professor in the Department of Mathematics at Cornell University starting in July 2023. Yunan Yang earned a Ph.D. degree in mathematics from the University of Texas at Austin in 2018, supervised by Prof. Bjorn Engquist. From September 2018 to August 2021, Yunan was a Courant Instructor at the Courant Institute of Mathematical Sciences, New York University.

返回顶部
相关文章