Speaker
y research focuses on mathematical analysis, algorithm development, and their applications in machine learning and scientific computing, spanning both data and physical sciences. My Ph.D. training was grounded in classical numerical methods for partial differential equations (PDEs), with a particular emphasis on finite element methods (FEM) and multigrid methods. Armed with this solid foundation in numerical PDEs and scientific computing, my primary research objective is to explore deep learning models and algorithms through the lens of numerical PDEs and geometry. This approach aims to foster a comprehensive understanding and innovative advancement of these models, covering theoretical foundations, algorithmic strategies, and practical applications. From my Ph.D. program to the present, and continuing into the foreseeable future, my research efforts and aspirations are principally centered on three interrelated themes:
· Mathematical analysis of deep neural networks (DNNs) from a finite element perspective;
· Development of theories, algorithms, and applications for convolutional neural networks (CNNs) and Transformers, drawing inspiration from multigrid structures;
· Investigation into the learning of data with low-dimensional structures.
Abstract:
In this talk, we will present recent results on the theories, algorithms, and applications of deep neural networks (DNNs) from a numerical analysis perspective. First, we will illustrate the connections between linear finite elements and ReLU DNNs, as well as spectral methods and ReLU^k DNNs. Second, we will show our latest findings regarding the open question of whether DNNs can precisely recover piecewise polynomials of arbitrary order on any simplicial mesh in any dimension. Then, inspired by the multigrid structure in numerical PDEs, we will discuss a unified framework for convolutional neural networks (CNNs) and multigrid methods, known as MgNet. Additionally, we will showcase recent advancements in the theories and applications of MgNet, particularly the first approximation result for CNNs with 2D inputs and an efficient operator learning framework, MgNO.