Do you want to publish a course? Click here

Error Analysis of Deep Ritz Methods for Elliptic Equations

176   0   0.0 ( 0 )
 Added by Yuling Jiao
 Publication date 2021
and research's language is English




Ask ChatGPT about the research

Using deep neural networks to solve PDEs has attracted a lot of attentions recently. However, why the deep learning method works is falling far behind its empirical success. In this paper, we provide a rigorous numerical analysis on deep Ritz method (DRM) cite{Weinan2017The} for second order elliptic equations with Drichilet, Neumann and Robin boundary condition, respectively. We establish the first nonasymptotic convergence rate in $H^1$ norm for DRM using deep networks with smooth activation functions including logistic and hyperbolic tangent functions. Our results show how to set the hyper-parameter of depth and width to achieve the desired convergence rate in terms of number of training samples.



rate research

Read More

Using deep neural networks to solve PDEs has attracted a lot of attentions recently. However, why the deep learning method works is falling far behind its empirical success. In this paper, we provide a rigorous numerical analysis on deep Ritz method (DRM) cite{wan11} for second order elliptic equations with Neumann boundary conditions. We establish the first nonasymptotic convergence rate in $H^1$ norm for DRM using deep networks with $mathrm{ReLU}^2$ activation functions. In addition to providing a theoretical justification of DRM, our study also shed light on how to set the hyper-parameter of depth and width to achieve the desired convergence rate in terms of number of training samples. Technically, we derive bounds on the approximation error of deep $mathrm{ReLU}^2$ network in $H^1$ norm and on the Rademacher complexity of the non-Lipschitz composition of gradient norm and $mathrm{ReLU}^2$ network, both of which are of independent interest.
In this paper, an important discovery has been found for nonconforming immersed finite element (IFE) methods using integral-value degrees of freedom for solving elliptic interface problems. We show that those IFE methods can only achieve suboptimal convergence rates (i.e., $O(h^{1/2})$ in the $H^1$ norm and $O(h)$ in the $L^2$ norm) if the tangential derivative of the exact solution and the jump of the coefficient are not zero on the interface. A nontrivial counter example is also provided to support our theoretical analysis. To recover the optimal convergence rates, we develop a new nonconforming IFE method with additional terms locally on interface edges. The unisolvence of IFE basis functions is proved on arbitrary triangles. Furthermore, we derive the optimal approximation capabilities of both the Crouzeix-Raviart and the rotated-$Q_1$ IFE spaces for interface problems with variable coefficients via a unified approach different from multipoint Taylor expansions. Finally, optimal error estimates in both $H^1$- and $L^2$- norms are proved and confirmed with numerical experiments.
180 - Yulei Liao , Pingbing Ming 2019
We propose a new method to deal with the essential boundary conditions encountered in the deep learning-based numerical solvers for partial differential equations. The trial functions representing by deep neural networks are non-interpolatory, which makes the enforcement of the essential boundary conditions a nontrivial matter. Our method resorts to Nitsches variational formulation to deal with this difficulty, which is consistent, and does not require significant extra computational costs. We prove the error estimate in the energy norm and illustrate the method on several representative problems posed in at most 100 dimension.
119 - Min Liu , Zhiqiang Cai 2021
In this paper, we study adaptive neuron enhancement (ANE) method for solving self-adjoint second-order elliptic partial differential equations (PDEs). The ANE method is a self-adaptive method generating a two-layer spline NN and a numerical integration mesh such that the approximation accuracy is within the prescribed tolerance. Moreover, the ANE method provides a natural process for obtaining a good initialization which is crucial for training nonlinear optimization problem. The underlying PDE is discretized by the Ritz method using a two-layer spline neural network based on either the primal or dual formulations that minimize the respective energy or complimentary functionals. Essential boundary conditions are imposed weakly through the functionals with proper norms. It is proved that the Ritz approximation is the best approximation in the energy norm; moreover, effect of numerical integration for the Ritz approximation is analyzed as well. Two estimators for adaptive neuron enhancement method are introduced, one is the so-called recovery estimator and the other is the least-squares estimator. Finally, numerical results for diffusion problems with either corner or intersecting interface singularities are presented.
73 - Brendan Keith 2020
A number of non-standard finite element methods have been proposed in recent years, each of which derives from a specific class of PDE-constrained norm minimization problems. The most notable examples are $mathcal{L}mathcal{L}^*$ methods. In this work, we argue that all high-order methods in this class should be expected to deliver substandard uniform h-refinement convergence rates. In fact, one may not even see rates proportional to the polynomial order $p > 1$ when the exact solution is a constant function. We show that the convergence rate is limited by the regularity of an extraneous Lagrange multiplier variable which naturally appears via a saddle-point analysis. In turn, limited convergence rates appear because the regularity of this Lagrange multiplier is determined, in part, by the geometry of the domain. Numerical experiments support our conclusions.
comments
Fetching comments Fetching comments
mircosoft-partner

هل ترغب بارسال اشعارات عن اخر التحديثات في شمرا-اكاديميا