Back to Search Start Over

TNet: A MODEL-CONSTRAINED TIKHONOV NETWORK APPROACH FOR INVERSE PROBLEMS.

Authors :
NGUYEN, HAI V.
TAN BUI-THANH
Source :
SIAM Journal on Scientific Computing; 2024, Vol. 46 Issue 1, pC77-C100, 24p
Publication Year :
2024

Abstract

Deep learning (DL), in particular deep neural networks, by default is purely datadriven and in general does not require physics. This is the strength of DL but also one of its key limitations when applied to science and engineering problems in which underlying physical properties--such as stability, conservation, and positivity--and accuracy are required. DL methods in their original forms are often not capable of respecting the underlying mathematical models or achieving desired accuracy even in big-data regimes. On the other hand, many data-driven science and engineering problems, such as inverse problems, typically have limited experimental or observational data, and DL would overfit the data in this case. Leveraging information encoded in the underlying mathematical models, we argue, not only compensates for missing information in low data regimes but also provides opportunities to equip DL methods with the underlying physics, hence promoting better generalization. This paper develops a model-constrained DL approach and its variant TNet--a Tikhonov neural network--which are capable of learning not only information hidden in the training data but also in the underlying mathematical models to solve inverse problems governed by partial differential equations in low data regimes. We provide the constructions and some theoretical results for the proposed approaches for both linear and nonlinear inverse problems. Since TNet is designed to learn inverse solutions with Tikhonov regularization, it is interpretable: in fact it recovers Tikhonov solutions for linear cases while potentially approximating Tikhonov solutions for nonlinear inverse problems. We also prove that data randomization can enhance not only the smoothness of the networks but also their generalizations. Comprehensive numerical results confirm the theoretical findings and show that with even as little as 1 training data sample for one-dimensional (1D) deconvolution, 5 for an inverse 2D heat conductivity problem, 100 for inverse initial conditions for a time-dependent 2D Burgers's equation, and 50 for inverse initial conditions for 2D Navier--Stokes equations, TNet solutions can be as accurate as Tikhonov solutions while being several orders of magnitude faster. This is possible owing to the model-constrained term, replications, and randomization. [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
10648275
Volume :
46
Issue :
1
Database :
Complementary Index
Journal :
SIAM Journal on Scientific Computing
Publication Type :
Academic Journal
Accession number :
175928539
Full Text :
https://doi.org/10.1137/22M1526708