No Arabic abstract
In emph{Guo et al, arXiv:2005.08288}, we propose a decoupled form of the structure-preserving doubling algorithm (dSDA). The method decouples the original two to four coupled recursions, enabling it to solve large-scale algebraic Riccati equations and other related problems. In this paper, we consider the numerical computations of the novel dSDA for solving large-scale continuous-time algebraic Riccati equations with low-rank structures (thus possessing numerically low-rank solutions). With the help of a new truncation strategy, the rank of the approximate solution is controlled. Consequently, large-scale problems can be treated efficiently. Illustrative numerical examples are presented to demonstrate and confirm our claims.
The structure-preserving doubling algorithm (SDA) is a fairly efficient method for solving problems closely related to Hamiltonian (or Hamiltonian-like) matrices, such as computing the required solutions to algebraic Riccati equations. However, for large-scale problems in $mathbb{C}^n$ (also $mathbb{R}^n$), the SDA with an $O(n^3)$ computational complexity does not work well. In this paper, we propose a new decoupled form of the SDA (we name it as dSDA), building on the associated Krylov subspaces thus leading to the inherent low-rank structures. Importantly, the approach decouples the original two to four iteration formulae. The resulting dSDA is much more efficient since only one quantity (instead of the original two to four) is computed iteratively. For large-scale problems, further efficiency is gained from the low-rank structures. This paper presents the theoretical aspects of the dSDA. A practical algorithm dSDA t with truncation and many illustrative numerical results will appear in a second paper.
We review a family of algorithms for Lyapunov- and Riccati-type equations which are all related to each other by the idea of emph{doubling}: they construct the iterate $Q_k = X_{2^k}$ of another naturally-arising fixed-point iteration $(X_h)$ via a sort of repeated squaring. The equations we consider are Stein equations $X - A^*XA=Q$, Lyapunov equations $A^*X+XA+Q=0$, discrete-time algebraic Riccati equations $X=Q+A^*X(I+GX)^{-1}A$, continuous-time algebraic Riccati equations $Q+A^*X+XA-XGX=0$, palindromic quadratic matrix equations $A+QY+A^*Y^2=0$, and nonlinear matrix equations $X+A^*X^{-1}A=Q$. We draw comparisons among these algorithms, highlight the connections between them and to other algorithms such as subspace iteration, and discuss open issues in their theory.
We study a family of structure-preserving deterministic numerical schemes for Lindblad equations, and carry out detailed error analysis and absolute stability analysis. Both error and absolute stability analysis are validated by numerical examples.
In this work we consider an extension of a recently proposed structure preserving numerical scheme for nonlinear Fokker-Planck-type equations to the case of nonconstant full diffusion matrices. While in existing works the schemes are formulated in a one-dimensional setting, here we consider exclusively the two-dimensional case. We prove that the proposed schemes preserve fundamental structural properties like nonnegativity of the solution without restriction on the size of the mesh and entropy dissipation. Moreover, all the methods presented here are at least second order accurate in the transient regimes and arbitrarily high order for large times in the hypothesis in which the flux vanishes at the stationary state. Suitable numerical tests will confirm the theoretical results.
The worst situation in computing the minimal nonnegative solution of a nonsymmetric algebraic Riccati equation associated with an M-matrix occurs when the corresponding linearizing matrix has two very small eigenvalues, one with positive and one with negative real part. When both these eigenvalues are exactly zero, the problem is called critical or null recurrent. While in this case the problem is ill-conditioned and the convergence of the algorithms based on matrix iterations is slow, there exist some techniques to remove the singularity and transform the problem to a well-behaved one. Ill-conditioning and slow convergence appear also in close-to-critical problems, but when none of the eigenvalues is exactly zero the techniques used for the critical case cannot be applied. In this paper, we introduce a new method to accelerate the convergence properties of the iterations also in close-to-critical cases, by working on the invariant subspace associated with the problematic eigenvalues as a whole. We present a theoretical analysis and several numerical experiments which confirm the efficiency of the new method.