We study M-matrix algebraic Riccati equation (MARE) of the form
where $ A $, $ B $, $ C $, $ D $ are real matrices of sizes $ m\times m $, $ m\times n $, $ n\times m $, $ n\times n $, respectively, and
is an M-matrix. M-matrix algebraic Riccati equation arises from many branches of applied mathematics, such as transport theory, Wiener-Hopf factorization of Markov chains, stochastic process, and so on [1-5]. Research on the theories and the efficient numerical methods of MARE became a hot topic in recent years [6-19].
The following are some notations and definitions we will use later.
For any matrices $ A = (a_{ij}), B = (b_{ij})\in \mathbb{R}^{m\times n} $, we write $ A\geq B(A> B) $, if $ a_{ij}\geq b_{ij}(a_{ij}>b_{ij}) $ for all $ i, j $. $ A $ is called a Z-matrix if $ a_{ij}\leq0 $ for all $ i\neq j $. Z-matrix $ A $ is called an M-matrix if there exists a nonnegative matrix $ B $ such that $ A = sI-B $ and $ s\geq\rho(B), $ where $ \rho(B) $ is the spectral radius of $ B $. In particular, $ A $ is called a nonsingular M-matrix if $ s> \rho(B) $ and singular M-matrix if $ s = \rho(B) $.
We first review some basic results of M-matrix. The following lemmas can be found in [20, Chapter 6].
Lemma 1.1 (see [20]) Let $ A $ be a Z-matrix, then the following statements are equivalent
(1) $ A $ is a nonsingular M-matrix;
(2) $ A^{-1}\geq0 $;
(3) $ Av > 0 $ for some vectors $ v>0 $;
(4) all eigenvalues of $ A $ have positive real part.
Lemma 1.2 (see [20]) Let $ A $, $ B $ be Z-matrices. If $ A $ is a nonsingular M-matrix and $ A\leq B $, then $ B $ is also a nonsingular M-matrix. In particular, for any nonnegative real number $ \alpha $, $ B = \alpha I+A $ is a nonsingular M-matrix.
Lemma 1.3 (see [20]) Let $ A $ be an M-matrix, $ B\ge A $ a Z-matrices. If $ A $ is nonsingular or if $ A $ is singular and irreducible with $ A\not = B $, then $ B $ is also a nonsingular M-matrix.
Lemma 1.4 (see [20]) Let $ A $ be a nonsingular M-matrix or an irreducible singular M-matrix. $ A $ is partitioned as
where $ A_{11} $ and $ A_{22} $ are square matrices, then $ A_{11} $ and $ A_{22} $ are nonsingular M-matrices.
Lemma 1.5 (see [20]) Let $ A $, $ B $ be nonsingular M-matrices and $ A\leq B $, then $ A^{-1}\geq B^{-1} $.
Lemma 1.6 (see [13]) Let $ A $ be an M-matrix and $ \lambda_{\min\limits}(A) $ be the eigenvalue of $ A $ with smallest absolute value. Then $ \lambda_{\min\limits}(A) $ is nonnegative and satisfies
For the minimal nonnegative solution of the MARE, we have the following important result.
Lemma 1.7 (see [3, 5, 6]) If $ K $ in (1.2) is a nonsingular M-matrix or an irreducible singular M-matrix, then (1.1) has a minimal nonnegative solution $ S $. If $ K $ is a nonsingular M-matrix, then $ A-SC $ and $ D-CS $ are also nonsingular M-matrices. If $ K $ is irreducible M-matrix, then $ S>0 $ and $ A-SC $ and $ D-CS $ are also irreducible M-matrices.
Lemma 1.8 (see [3, 5]) If $ K $ in (1.2) is an irreducible singular M-matrix, then there exist unique, up to a multiplicative constant, $ u>0 $ and $ v>0 $ such that $ u^TK = 0 $, $ Kv = 0 $ and $ u^Tv = 1 $.
Partition the vectors $ u $ and $ v $ in the above lemma according to the block structure of the matrix $ K $ as $ u^T = [u_1^T, u_2^T] $, $ v^T = [v_1^T, v_2^T] $. Through $ u $ and $ v $ we can define the drift $ \mu $ as the real number
The term "drift" originates in the context of Markov chains and describes the different physical behaviors of the chain in the cases where $ \mu $ is positive, null, or negative. In this context the terms positive recurrent, null recurrent, and transient are used to denote the caess $ \mu<0 $, $ \mu = 0 $, and $ \mu>0 $, respectively.
In terms of the definition, we have the following result.
Lemma 1.9 (see [3]) If $ K $ defined in (1.2) is an irreducible singular M-matrix and $ S $ is the minimal nonnegative solution of the MARE (1.1). Then
(ⅰ) when $ \mu<0 $, $ D-CS $ is singular and $ A-SC $ is nonsingular;
(ⅱ) when $ \mu>0 $, $ D-CS $ is nonsingular and $ A-SC $ is singular;
(ⅲ) when $ \mu = 0 $, both $ D-CS $ and $ A-SC $ are singular.
Efficient numerical methods for MARE include fixed-point iterative methods, Newton method, SDA, and so on. For details see [3, 5, 8, 9, 11, 13].
Recently, Bai et al.[8] proposed an alternately linearized implicit iteration method (ALI) for the MARE, which is very simple and efficient than the fixed-point iterative methods and Newton method, since at each iteration only two linear matrix equations are needed to solve.
However, there still has room to improve the ALI iteration method. In this paper, to fully improve the effectiveness of the ALI iteration method, we propose a modified alternately linearized implicit iteration method (MALI) which has two parameters and includes the ALI iteration method as special cases.
The rest of this paper is organized as follows. In next section, we propose a modified alternately linearized implicit iteration method and give its convergence analysis. In Section 3, we analysis the convergence rate and the optimal parameters of the MALI iteration method. In Section 4, we use some numerical examples to show the feasibility of the MALI iteration method. Conclusion is given in the last section.
In the following, we propose a modified alternately linearized implicit iteration method.
Compared with the ALI iteration method, there are two parameters $ \alpha $ and $ \beta $ in the MALI iteration method, which will reduce to the ALI iteration method when $ \alpha = \beta $. Hence the ALI iteration method is a special case of the MALI iteration method.
In the following, we give convergence analysis of the MALI iteration method. First we need several lemmas.
Lemma 2.1 Let $ \{X_k\} $ be the matrix sequence generated by the MALI iteration method, $ \mathcal{R}(X) = XCX-XD-AX+B $, and $ S $ be the minimal nonnegative solution to (1.1). Then for any $ k\ge 0 $, the following equalities hold
(1) $ (X_{k+1/2}-S)(\alpha I+(D-CX_k)) = (\alpha I -(A-SC))(X_k-S) $;
(2) $ (X_{k+1/2}-X_k)(\alpha I+(D-CX_k)) = \mathcal{R}(X_k) $;
(3) $ \mathcal{R}(X_{k+1/2}) = (\alpha I-(A-X_{k+1/2}C))(X_{k+1/2}-X_k) $;
(4) $ (\beta I +(A-X_{k+1/2}C))(X_{k+1}-S) = (X_{k+1/2}-S)(\beta I-(D-CS)) $;
(5) $ (\beta I +(A-X_{k+1/2}C))(X_{k+1}-X_{k+1/2}) = \mathcal{R}(X_{k+1/2}) $;
(6) $ \mathcal{R}(X_{k+1}) = (X_{k+1}-X_{k+1/2})(\beta I-(D-CX_{k+1})) $.
Proof The proof is similar to that of in [8], so we omit here.
Lemma 2.2 For the MARE (1.1), if the matrix $ K $ in (1.2) is a nonsingular M-matrix or an irreducible singular M-matrix, $ S $ is the minimal nonnegative solution to (1.1), then for any $ \delta>0 $ and $ 0\le Z\le S $, the matrices $ \delta I+A-ZC $ and $ \delta I+D-CZ $ are nonsingular M-matrices.
Proof First, from $ 0\le Z\le S $, we have $ A-SC\le A-ZC $ and $ D-CS\le D-CZ $.
If $ K $ is a nonsingular M-matrix, then $ A-SC $ and $ D-CS $ are also nonsingular M-matrices by Lemma 1.7. Thus $ \delta I+A-ZC $ and $ \delta I+D-CZ $ are nonsingular M-matrices by Lemma 1.2.
If $ K $ is an irreducible M-matrix, then $ A-SC $ and $ D-CS $ are also irreducible M-matrices by Lemma 1.7. Since $ \delta>0 $, $ A-SC\not = \delta I+ A-ZC $ and $ D-CS\not = \delta I+D-CZ $. By Lemma 1.3, $ \delta I+A-ZC $ and $ \delta I+D-CZ $ are nonsingular M-matrices.
Lemma 2.3 For the MARE (1.1), suppose that the matrix $ K $ in (1.2) is a nonsingular M-matrix or an irreducible singular M-matrix, $ S $ is the minimal nonnegative solution to (1.1), and the parameters $ \alpha, \beta $ satisfy
where $ a_{i, i} $ and $ d_{j, j} $ are diagonal entries of $ A $ and $ D $, respectively, then for any $ k\ge 0 $,
(ⅰ) $ \{X_{k+1/2}\} $ and $ \{X_{k+1}\} $ are well defined and bounded
(ⅱ) $ \beta I+A-X_{k+1/2}C $ and $ \alpha I+D-CX_{k+1} $ are nonsingular M-matrices.
Proof If $ K $ is a nonsingular M-matrix or an irreducible singular M-matrix, from Lemma 1.4, $ A $ and $ D $ are nonsingular M-matrices. Thus, when $ \alpha $ and $ \beta $ satisfy (2.2), both $ \alpha I -A $ and $ \beta I -D $ are nonnegative matrices.
We prove this lemma by induction. When $ k = 0 $, we have $ X_{1/2}(\alpha I+D) = B $ since $ X_0 = 0 $ from the MALI iteration. As $ D $ is a nonsingular M-matrix, by Lemma 1.2, $ \alpha I+D $ is also a nonsingular M-matrix. Thus from Lemma 1.1 we have $ (\alpha I+D)^{-1}\geq 0 $. Hence
On the other hand, from Lemma 2.1(1), we have
Thus since $ C\ge 0 $ and $ S\ge 0 $, we have
This show that $ 0 \leq X_{1/2}\leq S $.
On the other hand, since $ C\geq 0 $, we have $ A-SC\leq A-X_{1/2}C $. Since that $ D $ is a nonsingular M-matrix, $ \beta $ must be positive from (2.2). Thus $ \beta I+A-X_{1/2}C $ is a nonsingular M-matrix by Lemma 2.2.
Similarly, from the MALI iteration and Lemma 2.1 (4), we have
and
Hence
This show that $ 0 \leq X_1\leq S $.
On the other hand, since $ C\geq 0 $, we have $ D-CS\leq D-CX_1. $ Note that $ A $ is a nonsingular M-matrix, $ \alpha $ must be positive. By Lemma 2.2, $ \alpha I+D-CX_1 $ is a nonsingular M-matrix.
Thus (ⅰ) and (ⅱ) hold true for $ k = 0 $.
Assume that (ⅰ) and (ⅱ) hold true for $ k = l-1 $. From the MALI iteration, we have
thus
From Lemma 2.1 (1), we have $ (X_{l+1/2}-S)(\alpha I+(D-CX_l)) = (\alpha I -(A-SC))(X_l-S), $ thus
This show that $ 0 \leq X_{l+1/2}\leq S $.
Again, since $ C\geq 0 $ and $ \beta>0 $, by Lemma 2.2, $ \beta I+A-X_{l+1/2}C $ is a nonsingular M-matrix.
This show that $ 0 \leq X_{l+1}\leq S $.
Again, since $ C\geq 0 $ and $ \alpha>0 $, by Lemma 2.2, $ \alpha I+D-CX_{l+1} $ is a nonsingular M-matrix.
Thus we prove by induction that (i) and (ii) hold true for all $ k\geq0 $.
Lemma 2.4 Under the assumption of Lemma 2.2, the following inequalities hold for all $ k\geq 0 $,
where $ \mathcal{R}(X) $ is defined in Lemma 2.1.
Proof We prove this lemma by induction.
In fact, when $ k = 0 $, we have $ \mathcal{R}(X_{0}) = B\geq 0 $. From Lemma 2.1(2), we have
Since $ \alpha I+D $ is a nonsingular M-matrix,
From Lemma 2.1(3), we have
From Lemma 2.1(5), we have $ (\beta I +(A-X_{1/2}C))(X_{1}-X_{1/2}) = \mathcal{R}(X_{1/2}) $. By Lemma 2.2, $ \beta I +(A-X_{1/2}C) $ is a nonsingular M-matrix, we have
From Lemma 2.1(6), we have
This show that (2.4) holds for $ k = 0 $.
Now we suppose that (2.4) holds for $ k = l-1 $. Then from Lemma 2.1(2) and Lemma 2.2, we have
From Lemma 2.1(5) and Lemma 2.2, we have
Thus we show by induction that (2.4) holds for all $ k\geq 0 $.
By Lemmas 2.3 and 2.4, we can prove the following convergence theorem of the MALI iteration method.
Theorem 2.1 For the MARE (1.1), if $ K $ in (1.2) is a nonsingular M-matrix or an irreducible singular M-matrix, $ S $ is the minimal nonnegative solution to (1.1), and the parameters $ \alpha, \beta $ satisfy (2.2), then $ \{X_{k}\} $ is well defined, monotonically increasing and converges to $ S $.
Proof Combining Lemma 2.3 with Lemma 2.4, we show that $ \{X_{k}\} $ is nonnegative, monotonically increasing and bounded from above. Thus there is a nonnegative matrix $ S^* $ such that $ \lim\limits_{k\rightarrow\infty}X_{k} = S^* $. It also holds that $ \lim\limits_{k \rightarrow\infty}X_{k+1/2} = S^* $. From Lemma 2.3, we have $ S^*\leq S $. On the other hand, take the limit in the MALI iteration, we have that $ S^* $ is a solution of the MARE, thus $ S\leq S^* $. Hence $ S = S^* $.
In this section, we analyse the convergence rate and give optimal parameters of the MALI iteration method.
Theorem 3.1 Under the assumption of Theorem 2.1, for the sequence $ \{X_{k}\} $ generated by the MALI iteration method, we have
where
Proof From Lemma 2.1(1) and (4), we have
Thus
Because $ \beta I +(A-SC)\leq \beta I +(A-X_{k+1/2}C) $, by Lemma 1.5, we have
Similarly, we have
By induction we have
Take limit on both side and, note that $ \rho(A) = \lim\limits_{k\rightarrow\infty}\|A^k\|_2^{1/k} $, we have
It's easy to verify that
where $ \lambda_{\min\limits} = \lambda_{\min\limits}(A-SC) $ and $ \mu_{\min\limits} = \mu_{\min\limits}(D-CS) $. Hence
Corollary 3.1 When $ K $ in (2.2) is a nonsingular M-matrix or an irreducible singular M-matrix with nonzero drift, for any $ \alpha, \beta $ satisfy (2.2), we have $ \rho(\alpha, \beta)<1 $. In this case, the MALI iteration method has linear convergence rate. When $ K $ is an irreducible singular M-matrix with zero drift, for any $ \alpha, \beta $ satisfy (2.2, we have $ \rho(\alpha, \beta) = 1 $. In this case, the the MALI iteration method has sub-linear convergence rate.
Proof When $ K $ is a nonsingular or an irreducible singular M-matrix with nonzero drift, we know that $ A-SC $ and $ D-CS $ have at least one which is nonsingular by Lemma 1.9. By Lemma 1.6,
where $ (A-SC)_{ii} $ is the diagonal entries of $ A-SC $ and $ (D-CS)_{ii} $ is the diagonal entries of $ D-CS $. Thus $ \rho(\alpha, \beta)<1 $. In this case, the MALI iteration method has linear convergence rate.
When $ K $ is an irreducible singular M-matrix with zero drift, we know that both $ A-SC $ and $ D-CS $ are singular M-matrices by Lemma 1.9. Thus $ \lambda_{\min\limits}(A-SC) = 0 $, $ \mu_{\min\limits}(D-CS) = 0 $. Hence $ \rho(\alpha, \beta) = 1 $. In this case, the the MALI iteration method has sub-linear convergence rate.
Corollary 3.2 The optimal parameters of the MALI iteration method are
Proof It's easy to verify that $ \rho(\alpha, \beta) $ is an increasing function with respect to $ \alpha, \beta $. Thus the minimum of $ \rho(\alpha, \beta) $ is attained at $ \alpha = \max \{a_{ii}\} $, $ \beta = \max \{d_{jj}\} $.
Note that, in Corollary 3.2, the optimal parameters only minimize the upper bound of the contraction factor.
In this section we use several examples to show the feasibility of the MALI iteration method. We compare the MALI iteration method with the ALI iteration method in [8], Newton method in [3, 11] and present computational results in terms of the numbers of iterations (IT), CPU time (CPU) in seconds and the residue (RES), where
In our implementations all iterations are performed in MATLAB (version 2012a) on a personal computer and are terminated when the current iterate satisfies $ RES<10^{-6} $ or the number of iterations is more than 9000, which will be denoted by '-'.
Example 1 Consider the MARE (1.1) with
where $ E_{m\times n} $ is the $ m\times n $ matrix with all ones and $ I_m $ is the identity matrix of size $ m $ with $ m = 2, n = 18 $. This example is from [5], where the corresponding $ K $ is an irreducible singular M-matrix. The computational results are summarized in Table 1.
Example 2 Consider the MARE (1.1) with
This example is taken from [5], where we choose $ \alpha = 2 $ and the corresponding $ K $ is an irreducible singular M-matrix with $ \mu > 0 $. The computational results are summarized in Table 2 for different sizes of $ n $.
From the two numerical experiments, we can observe that the MALI iteration method needs the least CPU time than the ALI iteration method and Newton method. So it is feasible.
We propose a modified alternately linearized implicit iteration method (MALI) for computing the the minimal nonnegative solution of MARE. Theoretical analysis and numerical experiments show that the MALI iteration method is feasible in some cases. However, since the MALI iteration method only has linear convergence rate in general, it will be very slowly as compared with the SDA algorithm in [9, 15]. So, in general, the SDA algorithm will be more preferable.