Abstract

Implementing deep neural networks for learning the solution maps of parametric partial differential equations (PDEs) turns out to be more efficient than using many conventional numerical methods. However, limited theoretical analyses have been conducted on this approach. In this study, we investigate the expressive power of deep rectified quadratic unit (ReQU) neural networks for approximating the solution maps of parametric PDEs. The proposed approach is motivated by the recent important work of Kutyniok et al. (Constr Approx 1–53, 2021), which uses deep rectified linear unit (ReLU) neural networks for solving parametric PDEs. In contrast to the previously established complexity-bound \(\mathcal {O}\left( d^3\log _{2}^{q}(1/ \epsilon ) \right) \) for ReLU neural networks, we derive an upper bound \(\mathcal {O}\left( d^3\log _{2}^{q}\log _{2}(1/ \epsilon ) \right) \) on the size of the deep ReQU neural network required to achieve accuracy \(\epsilon >0\), where d is the dimension of reduced basis representing the solutions. Our method takes full advantage of the inherent low-dimensionality of the solution manifolds and better approximation performance of deep ReQU neural networks. Numerical experiments are performed to verify our theoretical result.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call