清华主页 EN
导航菜单

Neural Networks: A Perspective from Numerical Analysis

来源: 04-10

时间:Thur., 15:00-16:00 April 11, 2024

地点:腾讯会议:815-642-712

主讲人:Juncai He (King Abdullah University of Science and Technology)

Speaker

y research focuses on mathematical analysis, algorithm development, and their applications in machine learning and scientific computing, spanning both data and physical sciences. My Ph.D. training was grounded in classical numerical methods for partial differential equations (PDEs), with a particular emphasis on finite element methods (FEM) and multigrid methods. Armed with this solid foundation in numerical PDEs and scientific computing, my primary research objective is to explore deep learning models and algorithms through the lens of numerical PDEs and geometry. This approach aims to foster a comprehensive understanding and innovative advancement of these models, covering theoretical foundations, algorithmic strategies, and practical applications. From my Ph.D. program to the present, and continuing into the foreseeable future, my research efforts and aspirations are principally centered on three interrelated themes:


· Mathematical analysis of deep neural networks (DNNs) from a finite element perspective;


· Development of theories, algorithms, and applications for convolutional neural networks (CNNs) and Transformers, drawing inspiration from multigrid structures;


· Investigation into the learning of data with low-dimensional structures.



Abstract:

In this talk, we will present recent results on the theories, algorithms, and applications of deep neural networks (DNNs) from a numerical analysis perspective. First, we will illustrate the connections between linear finite elements and ReLU DNNs, as well as spectral methods and ReLU^k DNNs. Second, we will show our latest findings regarding the open question of whether DNNs can precisely recover piecewise polynomials of arbitrary order on any simplicial mesh in any dimension. Then, inspired by the multigrid structure in numerical PDEs, we will discuss a unified framework for convolutional neural networks (CNNs) and multigrid methods, known as MgNet. Additionally, we will showcase recent advancements in the theories and applications of MgNet, particularly the first approximation result for CNNs with 2D inputs and an efficient operator learning framework, MgNO.

返回顶部
相关文章
  • A Neural Networks Solver for R Matrices

    Abstractwe describe how neural networks may be used to learn solutions to the Yang-Baxter Equation using 2d spin chains of difference form as a concrete example. the talk is based on arxiv:2304.07247, which is joint work with Suvajit Majumder and Evgeny Skvortsov.Speaker IntroDr Shailesh Lal received his PhD from the Harish-Chandra Research Institute. His research interests are applications of ...

  • Physics-informed neural networks for solving differential equations

    PrerequisiteKnowledges on mathematical physics equations, deep neural networks and the Python language.AbstractDifferential equations can describe various natural and social phenomena. The physics-informed neural networks (PINNs), as a deep learning framework, is a powerful and effective way in solving forward and inverse problems involving differential equations. The course will review the pub...