A generalized projection iterative methods for solving non-singular linear systems
Abstract
In this paper, we propose and analyze iterative method based on projection techniques to solve a non-singular linear system . In particular, for a given positive integer , -dimensional successive projection method (mD-SPM) for symmetric definite matrix , is generalized for non-singular matrix . Moreover, it is proved that D-SPM gives better result for large values of . Numerical experiments are carried out to demonstrate the superiority of the proposed method in comparison with other schemes in the scientific literature.
keywords:
Symmetric linear systems , Iterative techniques , Petrov-Galerkin condition , Orthogonal projection method , Oblique projection method1 Introduction
Consider the linear system of equations
(1.1) |
where , and is an unknown vector. In [1], Ujević introduced a new iterative method for solving (1.1). The method was considered as a generalization of Gauss-Seidel methods. In [2], Jing and Huang interpreted Ujević’s method as one dimensional double successive projection method (1D-DSPM), whereas Gauss-Seidel method as one dimensional single successive projection method (1D-SSPM). They established a different approach and called it as two dimensional double successive projection method (2D-DSPM). In [3], Salkuyeh improved this method and gave a generalization of it, thereby calling it D-SPM. In an independent work [4], Hou and Wang explained 2D-DSPM as two-dimensional orthogonal projection method (2D-OPM), and generalized it to three dimensional orthogonal projection method (3D-OPM). All these works address systems in which the matrix in 1.1 is symmetric positve definite(SPD). In this paper, we generalize the D-SPM and use it on any non-singular system. The proposed method is called as D-OPM, where OPM refers to ‘orthogonal’ as well as ‘oblique’ projection method.
Given an initial approximation , a typical projection method for solving (1.1), on the subspace (known as the search subspace) and orthogonal to the subspace (known as the constraint subspace), is to find an approximate solution of (1.1) by imposing the Petrov-Galerkin condition [5] that
(1.2) |
In case of orthogonal project method, search space and the constraint spaces are same, whereas in oblique projection method, they are different. The elementary Gauss-Seidel method can be considered as an one dimensional OPM with , where is the ith column of the identity matrix. In [4], authors proposed three-dimensional OPM (D-OPM) and showed both theoretically and numerically that D-OPM gives better (or atleast the same) reduction of error than 2D-OPM in [2]. In [3], author proposed a generalization of D-OPM as well as gave a way to chose the subspace . We put forward the -dimensional OPM (D-OPM) by considering -dimensional subspaces , where, for oblique projection we take . At each iterative step, are taken as -dimensional subspaces and each iteration is cycled for , until it converges.
The paper is organized as follows: In Section 2, a theoretical proof of the advantage of chosing a larger value of in D-OPM is provided and also convergence of D-OPM for an SPD system is shown, which supplements the work in [3]. Section 3 shows the application of D-OPM to more general non-singular systems. Lastly, in section 4, numerical examples are considered for illustration.
2 mD-OPM for symmetric matrices
Throughout this section, the matrix under cosideration is assumed to be SPD. From the numerical experiments provided in [3], it is observed that -OPM provides better (or at least same) result with larger value of . In this section we prove it theoretically.
In -OPM, (= ) is considered as an -dimensional space. If , and , a basic projection step for an -OPM is defined in [3, 5] as :
(2.1) | |||||
Equivalently, | (2.2) |
If is the exact solution of (1.1), the quantity is defined as the error reduction at the th iteration step of D-OPM (2.1) and is denoted by as considered in [4]. In Theorem 1 of [3], author proved that . In particular, , if is not perpendicular to . To prove the main theorem of this section, we need the following Lemma.
Lemma 2.1.
If s are defined as in (2.2), then , where and .
Proof.
Proof follows from the fact in the proof of Theorem 1 of [3]. ∎
where and . Note that , , and .
Theorem 2.2.
, when .
Proof.
For simplicity, we write . For define as and elsewhere. By Lemma 2.1, it is sufficient to show that , or equivalently, . Since is a positive definite matrix, which implies that
(2.4) |
However, .
Thus from (2.4), we get ∎
Corollary 2.3.
-OPM defined in (2.2) converges.
3 m-dimensional oblique projection method for non-singular matrices
In this section we present new -dimensional oblique projection method (mD-OPM) to solve nonsingular system (1.1). Assume that , with . Take , and so that columns of and form bases for and , respectively. If , then the oblique projection iterative steps, discussed in (1.2), are given as follows [5]:
(3.1) |
where is the residual in the th iteration step.
Choose . Then as defined in (1.2) minimizes the -norm of the residual over (see, Ch.5 in [5]). Throughout this section, represents -norm in the Euclidean space and we drop the suffix which signifies the dimension of and .
As , we may take . Then (3.1) reduces to
(3.2) |
where denotes the pseudo-inverse of so that Main goal of this section is to prove the convergence of (3.2). Following lemma will help to reach our goal.
Lemma 3.1.
If is the maximum singular valur of , and , then
Proof.
As and , we have,
(3.3) |
Using Courant-Fisher min-max principle[5], from (3.3) we achieve,
(3.4) | |||||
where , denote the maximum and minimum eigenvalues, and , denotes the maximum and minimum singular values of the corresponding matrix, respectively.
Let be the singular value decomposition of . If , and , then so that
Hence the result follows from (3.4). ∎
In Theorem 3 of [3], author provided the convergence of the method (2.2) for SPD matrices, and also gives an idea to choose the optimal vectors . Similar ideas is used to prove the convergence of (3.2). Next theorem is due to [6] (see Ch 3, Cor 3.1.1), which gives the relation between singular values of a matrix and its submatrices.
Theorem 3.2.
[6] If is an matrix and denotes a submatrix of obtained by deleting a total of rows and/or columns from , then
where the singular values ’s are arranged decreasingly.
We now prove our main theorem, in which the singular values of matrix under consideration, are assumed to be arranged in decreasing order.
Theorem 3.3.
Let be the singular values of . If , and , th column of the identity matrix, then
(3.5) |
Proof.
Remark 3.4.
4 Numerical Experiments
In this section comparison of D-OPM is established with various methods, like, CGNR, GMRES and Craig’s method [7] for any non-singular linear system.
The algorithm of the -OPM, discussed in Section 3, is as follows, which is same as proposed in [3] by considering the symmetric system .
1. Chose an initial guess and decide on , the number of times each component of is improved in each iteration.
2. Until Convergence, Do
3. .
4. For , Do
5.Select the indices of
6.
7.Solve for
8.
9.
10.End Do
11.
12.End Do.
The experiments are done on a PC-Intel(R) Core(TM) i3-7100U CPU @ 2.40 GHz, 4 GB RAM. The computations are implemented in MATLAB 9.2.0.538062. The initial guess is and the stopping criteria is .
While doing comparisions with D-OPM, we consider different values of to get various results. The theory suggests that D-OPM will have a good convergence for matrices whose singular values are closely spaced. Hence we chose the matrices accordingly.
Example 4.1.
The first matrix is a symmetric Hankel matrix with elements . The eigen values of cluster around and and the condition number is of . The matrix is of size . Comparision is done for different values of as well as with the CGNR, GMRES and Craig’s method.
Iteration Process | No of Iterations | Residual |
---|---|---|
D-OPM | 14 | |
D-OPM | 8 | |
D-OPM | 2 | |
GMRES | 10 | |
CGNR | 9 | |
Craig | 9 |
Example 4.2.
We consider a square matrix of size with singular values , . This is again a matrix with extremely good condition number. For such a well-conditioned matrix, d-dspm works like a charm and is better than the CGNR. The matrix taken here is of size .
Iteration Process | No of Iterations | Residual |
---|---|---|
D-OPM | 1 | |
CGNR | 6 | |
Craig | 6 |
5 Conclusion
D-OPM, presented in this paper, is a generalization of D-SPM [3], and can be applied to any non-singular system. Numerical experiments showed that this method is at par with other established methods. The way in which the search subspace is chosen put this method at a clear advantage over GMRES, because in GMRES, the orthogonalisation through Arnoldi process can lead to infeasible growth in storage requirements.
Reference
References
- [1] N. Ujević, A new iterative method for solving linear systems, Applied Mathematics and Computation 179 (2006) 725–730.
- [2] Y. F. Jing, T. Z. Huang, On a new iterative method for solving linear systems and comparision results, Journal of Computational and Applied Mathematics 220 (2008) 74–84.
- [3] D. K. Salkuyeh, A generalization of the 2D-DSPM for solving linear system of equations, ArXiv, 0906.1798v1 (2009). arXiv:http://arxiv.org/abs/0906.1798v1.
- [4] G. Hou, L. Wang, A generalized iterative method and comparision results using projection techniques for solving linear systems, Applied Mathematics and Computation 215 (2009) 806–817.
- [5] Y. Saad, Iterative methods for sparse linear systems,second ed., SIAM, Philadelphia.
- [6] R. A. Horn, C. R. Johnson, Topics in matrix analysis, Cambridge University Press.
- [7] N. N. Kalitkin, L. V. Kuz’mina, Convergence of Craig’s method for linear algebraic systems, Mathematical Models and Computer Simulations 4 (2012) 509–526.