Award Abstract # 2309778
Collaborative Research: Theory and Applications of Structure-Conforming Deep Operator Learning

NSF Org: DMS
Division Of Mathematical Sciences
Recipient: THE CURATORS OF THE UNIVERSITY OF MISSOURI
Initial Amendment Date: May 30, 2023
Latest Amendment Date: May 30, 2023
Award Number: 2309778
Award Instrument: Standard Grant
Program Manager: Jodi Mead
jmead@nsf.gov
 (703)292-7212
DMS
 Division Of Mathematical Sciences
MPS
 Directorate for Mathematical and Physical Sciences
Start Date: June 1, 2023
End Date: May 31, 2026 (Estimated)
Total Intended Award Amount: $156,046.00
Total Awarded Amount to Date: $156,046.00
Funds Obligated to Date: FY 2023 = $156,046.00
History of Investigator:
  • Shuhao Cao (Principal Investigator)
    scao@umkc.edu
Recipient Sponsored Research Office: University of Missouri-Kansas City
118 UNIVERSITY HALL
COLUMBIA
MO  US  65211-3020
(816)235-5839
Sponsor Congressional District: 03
Primary Place of Performance: University of Missouri-Kansas City
5100 ROCKHILL RD
KANSAS CITY
MO  US  64110-2446
Primary Place of Performance
Congressional District:
05
Unique Entity Identifier (UEI): J9CDGR596MN3
Parent UEI:
NSF Program(s): COMPUTATIONAL MATHEMATICS
Primary Program Source: 01002324DB NSF RESEARCH & RELATED ACTIVIT
Program Reference Code(s): 079Z, 9263
Program Element Code(s): 127100
Award Agency Code: 4900
Fund Agency Code: 4900
Assistance Listing Number(s): 47.049

ABSTRACT

The first-principle-based approach has achieved considerable success in numerous engineering and scientific disciplines, including fluid and solid mechanics, electromagnetism, and more. Among its most significant applications are partial differential equations (PDEs) which, in conjunction with their analysis and numerical algorithms, represent some of the most powerful tools humanity has ever developed for understanding the material world. However, increasingly complex mathematical models arising from physics, biology, and chemistry challenge the efficacy of first-principle-based approaches for solving practical problems, such as those in fluid turbulence, molecular dynamics, and large-scale inverse problems. A major obstacle for numerical algorithms is the so-called curse of dimensionality. Fueled by advances in Graphics Processing Unit and Tensor Processing Unit general-purpose computing, deep neural networks (DNNs) and deep learning approaches excel in combating the curse of dimensionality and demonstrate immense potential for solving complex problems in science and engineering. This project aims to investigate how mathematical structures within a problem can inform the design and analysis of innovative DNNs, particularly in the context of inverse problems where unknown parameters are inferred from measurements, such as electrical impedance tomography. Additionally, the programming component in this project will focus on training the next generation of computational mathematicians.

The Operator Learning (OpL) framework in deep learning provides a unique perspective for tackling challenging and potentially ill-posed PDE-based problems. This project will explore the potential of OpL to mitigate the ill-posedness of many inverse problems, as its powerful approximation capability combined with offline training and online prediction properties lead to high-quality, rapid reconstructions. The project seeks to bridge OpL and classical methodologies by integrating mathematical structures from classical problem-solving approaches into DNN architectures. In particular, the project will shed light on the mathematical properties of the attention mechanism, the backbone of state-of-the-art DNN Transformers, such as those in GPT and AlphaFold 2. Furthermore, the project will examine the flexibility of attention neural architectures, enabling the fusion of attention mechanisms with important methodologies in applied mathematics, such as Galerkin projection or Fredholm integral equations, in accordance with the a priori mathematical structure of a problem. This project will also delve into the mathematical foundations of attention through the lens of spectral theory in Hilbert spaces, seeking to understand how the emblematic query-key-value architecture contributes to the rich representational power and diverse approximation capabilities of Transformers.

This award reflects NSF's statutory mission and has been deemed worthy of support through evaluation using the Foundation's intellectual merit and broader impacts review criteria.

PUBLICATIONS PRODUCED AS A RESULT OF THIS RESEARCH

Note:  When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external site maintained by the publisher. Some full text articles may not yet be available without a charge during the embargo (administrative interval).

Some links on this page may take you to non-federal websites. Their policies may differ from this site.

Barker, Mary and Cao, Shuhao and Stern, Ari "A nonconforming primal hybrid finite element method for the two-dimensional vector Laplacian" The SMAI journal of computational mathematics , v.10 , 2024 https://doi.org/10.5802/smai-jcm.107 Citation Details
Cao, Shuhao and Brarda, Francesco and Li, Rui_Peng and Xi, Yuanzhe "Spectral-Refiner: Accurate Fine-Tuning of Spatiotemporal Fourier Neural Operator for Turbulent Flows" , 2025 Citation Details
Cao, Shuhao and Qin, Lizhen "A Numerical Domain Decomposition Method for Solving Elliptic Equations on Manifolds" SIAM Journal on Scientific Computing , v.46 , 2024 https://doi.org/10.1137/23M1546221 Citation Details
Guo, Ruchi and Cao, Shuhao and Chen, Long "Transformer Meets Boundary Value Inverse Problems" , 2023 Citation Details
Liu, Xinliang and Xu, Bo and Cao, Shuhao and Zhang, Lei "Mitigating spectral bias for the multiscale operator learning" Journal of Computational Physics , v.506 , 2024 https://doi.org/10.1016/j.jcp.2024.112944 Citation Details
Liu, Zhengqi and Cao, Shuhao and Li, Yuwen and ZIKATANOV, LUDMIL "Neural networks with trainable matrix activation functions" Journal of Machine Learning for Modeling and Computing , 2025 https://doi.org/10.1615/JMachLearnModelComput.2024056966 Citation Details

Please report errors in award information by writing to: awardsearch@nsf.gov.

Print this page

Back to Top of page