清华主页 EN
导航菜单

Adaptive Gradient Methods with Energy for Optimization Problems | Applied and Computational Math Colloquium

来源: 07-04

时间:Tues., 15:20-16:20, July 4, 2023

地点:Lecture Hall, Floor 3 Jin Chun Yuan West Building

组织者:应用与计算数学团队

主讲人:刘海亮 Hailiang Liu Iowa State University

Abstract

We propose AEGD, a new algorithm for gradient-based optimization of stochastic objective functions, based on adaptive updates of quadratic energy. The method is shown to be unconditionally energy stable, irrespective of the step size. In addition, AEGD enjoys tight convergence rates, yet allows a large step size. The method is straightforward to implement and requires little tuning of hyper-parameters. Experimental results demonstrate that AEGD works well for various optimization problems: it is robust with respect to initial data, capable of making rapid initial progress, shows comparable and most times better generalization performance than SGD with momentum for deep neural networks.


Speaker

Hailiang Liu is a Professor of Mathematics and Computer Science at the Iowa State University (ISU). He earned his Bachelor degree from Henan Normal University, Master degree from Tsinghua University, and Ph.D. degree from the Chinese Academy of Sciences, all in Mathematics. His research interests include analysis of partial differential equations, the development of high order numerical algorithms for solving these PDE problems, with diverse applications. He is the author of over 160 peer reviewed papers, and the recipient of many awards and honors, including the Alexander von Humboldt-Research Fellow, and the inaugural Holl Chair in Applied Mathematics at Iowa State University.

返回顶部
相关文章
  • Applied and Computational Math Colloquium

    Talk 1 9:00-10:00 Responsible Machine Learning and Machine Learning for Science 本报告分三部分:(一) 如何建立高效和可靠的机器学习系统;(二) 如何设计机器学习算法来解决科学问题,比如:药物开发和量子计算;(三) 对北卡罗来纳大学教堂山分校的简单分享。About the speaker 陈天龙博士将于 2024 年秋季加入北卡罗来纳大学教堂山分校计算机系担任助理教授。在这之前 (2023 - 2024),他会加入麻省理工和哈佛大学担任博士...

  • Efficient natural gradient method for large-scale optimization problems

    AbstractFirst-order methods are workhorses for large-scale optimization problems, but they are often agnostic to the structural properties of the problem under consideration and suffer from slow convergence, being trapped in bad local minima, etc. Natural gradient descent is an acceleration technique in optimization that takes advantage of the problem's geometric structure and preconditions the...