A Relaxed Greedy Block Kaczmarz Method for Solving Large Consistent Linear Systems

Abstract

Many problems in science and engineering require solving large consistent linear systems. This paper presents a relaxed greedy block Kaczmarz method (RGBK) and an accelerated greedy block Kaczmarz method (AGBK) for solving large-size consistent linear systems. The RGBK algorithm extends the greedy block Kaczmarz algorithm (GBK) presented by Niu and Zheng in [1] by introducing a relaxation parameter to the iteration formulation of GBK, and the AGBK algorithm uses different iterative update rules to minimize the running time. The convergence of the RGBK is proved and a method to determine an optimal parameter is provided. Several examples are presented to show the effectiveness of the proposed methods for overdetermined and underdetermined consistent linear systems with dense and sparse coefficient matrix.

Share and Cite:

Liao, Y. , Yin, F. and Huang, G. (2021) A Relaxed Greedy Block Kaczmarz Method for Solving Large Consistent Linear Systems. Journal of Applied Mathematics and Physics, 9, 3032-3044. doi: 10.4236/jamp.2021.912196.

1. Introduction

We are concerned with the solution of the large consistent linear system

A x = b , (1)

where A m × n , and b m . The Kaczmarz method in [2] is possible one of the most popular, simple while efficient algorithms for solving (1). It was revised to be applied to image reconstruction in [3], which is called algebraic reconstruction technique, and has a large range of fields of applications such as image reconstruction in computerized tomography [4] [5] [6] and parallel computing [7].

There are many extended Kaczmarz algorithms [8] - [16] developed to solve (1). Strohmer and Vershynin in [17] introduced a randomized Kaczmarz method (RK) for consistent overdetermined systems (1). The RK method has a convergence bound with the expected exponential convergence, which was called linear convergence. Zhang [18] proposed an improved greedy Kaczmarz (GK) method for solving (1). Bai and Wu in [19] presented a greedy randomized Kaczmarz algorithm (GRK) for (1) when the system is consistent. In each step of iteration, GRK is based on a probability criterion trying to grasp larger entries of the residual vector. Bai and Wu [20] further developed a relaxed GRK method for large sparse Equations (1). Due to its fast convergence, the block method [21] [22] [23] [24] has also been extensively developed in linear or nonlinear optimization problems. Recently, Liu and Zheng in [1] presented a greedy block Kaczmarz algorithm (GBK) with the iteration

x k + 1 = x k + A J k ( b J k A J k x k ) , (2)

where the index of the selected row

J k = { i k | | b i k A ( i k ) x k | 2 ε k A ( i k ) 2 } (3)

with the parameter

ε k = η max 1 i m { | b i A ( i ) x k | 2 A ( i ) 2 2 } . (4)

A J k stands for the submatrix A ( J k , : ) of A , b i represents the ith element of the vector b, A ( i ) denotes the ithrow of A and A denotes the Moore-Penrose pseudoinverse of A .

In this paper, based on the GBK method in [1], we develop a new relaxed greedy block Kaczmarz algorithm (RGBK) for (1). RGBK extends the GBK algorithm by introducing a relaxed parameter. The convergence of the algorithm is provided and the optimal relaxed parameter is discussed. The rest of this paper is organized as follows. In Section 2, a new relaxed greedy block Kaczmarz algorithm is presented. The convergence is provided and the optimal relaxed parameter is determined. Several types of examples are shown in Section 3, including the overdetermined or underdetermined systems with dense and sparse coefficient matrices. Some conclusions are drawn in Section 4.

At the end of this section, we introduce some mathematical symbols that will be used. For a matrix Q m × n , Q ( i ) denotes the ith row vector of Q . Q J k stands for the submatrix Q ( J k , : ) of Q , where J k is an index set composed of positive integers not exceeding m, and m is the number of rows of Q . Let δ min ( Q ) and δ max ( Q ) be the maximum and minimum positive singular values of Q , respectively. Q 2 2 and Q F 2 = i = 1 m j = 1 n | q i j | 2 are the spectral norm and Frobenius norm, respectively. For any vector p m , p i represents the ith component of p.

2. The Relaxed Greedy Block Kaczmarz Algorithm

Replacing the left side x k + 1 in (2) with the combination of x k and x k + 1 in (2) by introducing a relaxed parameter λ ( 0,2 ) , we have

x k + 1 = ( 1 λ ) x k + λ ( x k + A J k ( b J k A J k x k ) ) . (5)

Thus

x k + 1 = x k + λ A J k ( b J k A J k x k ) . (6)

The method presented by (6) is called a relaxed greedy block Kaczmarz algorithm, which is abbreviated by RGBK. Algorithm 1 summarizes the RGBK algorithm.

We remark that the iteration formulation (6) reduces to (2) when λ = 1 , thus the GBK method in [1] is a special case of Algorithm 1.

The results below give the convergence of Algorithm 1.

Theorem 1. Assume the linear system (1) is consistent. The iterative sequence { x k } generated by Algorithm 1 converges to the minimum norm solution x = A b of (1). Moreover, it holds that

x k + 1 x 2 2 [ 1 ϕ k ( λ , η ) δ min 2 ( A ) A F 2 ] x k x 2 2 , k = 0,1, , (7)

where ϕ k ( λ , η ) = η λ ( 2 λ ) A J k F 2 σ max 2 ( A J k ) , A F denotes the Frobenius norm of

A , δ max ( A ) and δ min ( A ) are the maximum and minimum positive singular values of A , respectively.

Proof. Let e k = x k x , where x satisfies A J k x = b J k , then we have from (6) that

e k + 1 2 2 = e k + λ A J k ( b J k A J k x k ) 2 2 = ( I λ A J k A J k ) e k 2 2 = e k 2 2 λ ( 2 λ ) A J k A J k e k 2 2 . (8)

According to the definition of J k at Step 3 in Algorithm 1 and the fact that δ min 2 ( A J k ) = δ max 2 ( A J k ) , we have that

Algorithm 1. A relaxed greedy block Kaczmarz algorithm (RGBK).

A J k A J k e k 2 2 δ min 2 ( A J k ) A J k e k 2 2 = δ min 2 ( A J k ) i k J k | A ( i k ) e k | 2 δ min 2 ( A J k ) i k J k ε k A ( i k ) 2 = A J k F 2 δ max 2 ( A J k ) ε k . (9)

For k = 1 , 2 , , it holds that

b A x k 2 2 = i = 1 m | b i A ( i ) x k | 2 A ( i ) 2 2 A ( i ) 2 2 max 1 i m | b i A ( i ) x k | 2 A ( i ) 2 2 A F 2 ,

thus the constant ε k at step 2 of Algorithm 1 becomes

ε k = η max 1 i m | b i A ( i ) x k | 2 A ( i ) 2 2 η b A x k 2 2 A F 2 = η A x A x k 2 2 A F 2 η δ min 2 ( A ) A F 2 e k 2 2 . (10)

Thus (8) together with (9) and (10) implies (7). This completes the proof. □

We derive easily the results below from Theorem 1.

Corollary 1. Let M ( λ , η ) = max 0 j k 1 ϕ j ( λ , η ) . Under the conditions of Theorem 1, we obtain an upper bound of error

x k x 2 2 [ 1 M ( λ , η ) δ min 2 ( A ) A F 2 ] k x 0 x 2 2 . (11)

Proof. Using (7) iteratively for k = 1 , , we have (11) with the definition of M ( λ , η ) . □

The upper bound of error below is independent of J k .

Corollary 2. Under the conditions of Theorem 1, (7) becomes

x k + 1 x 2 2 [ 1 η λ ( 2 λ ) δ min 2 ( A ) A F 2 ] x k x 2 2 . (12)

Proof. We notice that

A J k F 2 = i = 1 | J k | δ i 2 ( A J k ) δ max 2 ( A J k ) ,

thus

A J k F 2 δ max 2 ( A J k ) 1. (13)

Therefore, (12) results from (12) together with (7). □

Remark 1. The RGBK method reduces to the GBK method in [1] when λ = 1 . Examples in Section 3 provide a way to determine the optimal relaxed parameter value of λ that minimizes the CPU time or the number of iteration for both overdetermined and underdetermined systems.

Remark 2. Taking into account the limitation of computer memory space, we use Gaussian Kaczmarz method defined in [22] instead of (6), which could avoid the calculation of A ,

x k + 1 = x k + λ δ k T ( b A x k ) A T δ k 2 2 A T δ k ,

where δ k = i J k ( b i A ( i ) x k ) e i , this method abbreviated as AGBK.

3. Numerical Examples

In this section, we give several examples to show the efficiency of our RGBK and AGBK methods and compare them with GBK in [1]. All experiments are carried out with the MATLAB 2020b on a computer with 3.00 GHz processing unit and 16 GB RAM.

We compute the solution of the consistent system (1) with A m × n and b m computed by b = A x * , where x * n denotes the exact solution generated by the MATLAB function randn. Denote the relative solution residuals RSE : = x k x 2 2 / x 2 2 . We define different acceleration ratios as follows,

SU-R = CPUofGBK CPUofRGBK and SU-A = CPUofGBK CPUofAGBK .

To avoid calculating the Moore-Penrose inverse A when implementing the update (6) in Algorithm 1, we use the CGLS algorithm in [21] to solve a corresponding least-squares problem. Considering the fairness of the three algorithms, all parameters involved in AGBK are the same as RGBK. We set the initial vector x 0 = 0 and the termination criterion satisfying RSE < 10−6 for GBK, RGBK and AGBK in all examples.

Example 4.1. We use this example to illustrate how to determine an optimal parameter η in Algorithm 1 for both overdetermined and underdetermined systems (1). We use different parameter η ranged from 0.05 to 0.9 to compute the number of iteration (IT) and CPU time (CPU) by Algorithm 1 for the consistent systems (1) with A 2000 × 1000 and A 1000 × 2000 , which are randomly dense matrices generated by the MATLAB function randn, respectively, then determine an optimal parameter η that minimizes the IT or CPU.

Figure 1 shows the plot of CPU versus η . From Figure 1, we can choose the optimal parameter η = 0.2 , which minimizes CPU by using Algorithm 1 for the consistent systems (1) with A 2000 × 1000 and A 1000 × 2000 .

Example 4.2. In this case, we give the same way to determine the optimal relaxation parameters λ of RGBK as shown in Figure 2, and strictly abide by the method of controlling variables. We default that GBK and RGBK have the same η .

Figure 2 shows the plot of IT and CPU versus relaxed parameter λ . From Figure 2(a) and Figure 2(b), we can choose the optimal relaxed parameter λ o p t = 1.2 and λ o p t = 1.3 , which minimizes CPU by using Algorithm 1 for the consistent systems (1) with A 2000 × 1000 and A 1000 × 2000 .

Figure 1. An optimal parameter ( η ) determined by the minimum of CPU for overdetermined 2000 × 1000 (left) and underdetermined 1000 × 2000 (right) systems in RGBK.

Figure 2. An optimal relaxed parameter ( λ ) determined by the minimum of IT (left) and CPU (right) for overdetermined and underdetermined systems. (a) IT (left) and CPU (right) versus λ with 2000 × 1000; (b) IT (left) and CPU (right) versus λ with 1000 × 2000.

Example 4.3. We compare the convergence of RGBK and AGBK with the optimal relaxed parameter λ o p t determined similarly in Example 4.2 with that of GBK algorithm for overdetermined and underdetermined systems (1) with A m × n generated by the MATLAB function randn. Table 1 lists IT and CPU(s) of the RGBK and AGBK algorithms with the optimal relaxed parameter λ o p t compared with that of GBK algorithm for different overdetermined systems (1) with Gaussian coefficient matrices A . The corresponding optimal relaxed parameter computed similarly in Example 4.2 is λ o p t = 1.2 , 1.2 , 1.3 , 1.2 , 1.2 , 1.2 , 1.25 , respectively.

Table 2 shows similar results for different underdetermined Gaussian systems (1) with A . From Table 1 and Table 2, we can see the advantages of our proposed RGBK and AGBK methods over GBK algorithm.

Example 4.4. We use the RGBK and AGBK methods and compare them with GBK to solve systems (1) with sparse coefficient matrix A from the Florida sparse matrix collection in [25]. Table 3 summarizes the different sparse systems and its density and condition number C o n d ( A ) , where the density of A means the ratio of the number of the nonzero elements of A to the total number of the elements of A .

Table 4 lists IT and CPU(s) of the RGBK and AGBK algorithms with the optimal

Table 1. Performance of GBK, RGBK, AGBK for overdetermined systems.

Table 2. Performance of GBK, RGBK, AGBK for underdetermined systems.

Table 3. The properties of different sparse matrices.

Table 4. Performance of GBK, RGBK, AGBK for overdetermined systems.

relaxed parameter λ o p t compared with that of GBK algorithm for different sparse systems (1) with coefficient matrices A . The corresponding optimal relaxed parameter computed similarly in Example 4.3 is λ o p t = 0.9 , 1.2 , 0.9 , 1.2 and 0.8, respectively. Figure 3 shows the plot of RSE versus IT (left) or RSE versus CPU (right) of Algorithm 1 applied to solve (1) with different sparse coefficient matrix A in Table 4. From Table 4, we can see that the speedup ratio SU-R can reach 1.2246 and SU-A can reach 2.5934, which shows the fast convergence of our proposed algorithm. From Figure 3, RGBK and AGBK converge much faster than GBK does on RSE versus IT (left) and RSE versus CPU (right) for sparse consistent Equation (1) with coefficient matrices A named ash958 and stat96v5 in Table 4.

Example 4.5. This example uses RGBK and AGBK to restore a computer

Figure 3. RSE versus IT (left) and RSE versus CPU (right) of RGBK and AGBK compared with GBK to solve consistent Equation (1) with sparse coefficient matrix A named ash958 (a) stat96v5 (b) in Table 3. (a) IT (left) and CPU (right) vesue ash958; (b) IT (left) and CPU (right) vesue stat96v5.

tomography (CT) image. The MATLAB function p a r a l l e l t o m o ( N , θ , p ) from Algebraic Iterative Reconstruction (ART) package in [26], which generates a large sparse matrix A and the exact solution x is used, where N = 70 , θ = 0 : 0.7 : 178 and p = 70 , then the size of A is 17,850 × 4900. We compute b by b = A x , RGBK and AGBK are used to recover x from b and compared with the GBK method.

Table 5 reports the IT and CPU(s), and RSE of RGBK and AGBK compared with GBK for overdetermined consistent sparse matrix, where RSE 10 6 . Figure 4 shows the recovered images by GBK, RGBK and AGBK together with the original image.

It can be seen from Table 5 that RGBK(1.3) and AGBK obtain lower IT and CPU(s) than GBK(0.2) for restoring CT images, which show that RGBK and AGBK converge faster than GBK dose if the parameter λ is selected appropriately.

Table 5. GBK, RGBK and AGBK for reconstruction of CT image.

Figure 4. The original “phantom” image (a), the recovered images by GBK (b), RGBK(1.3) (c) and AGBK (d).

4. Conclusion

We present a relaxed GBK algorithm abbreviated as RGBK for solving large consistent linear systems. The RGBK method extends the GBK method in [17]. The convergence is provided and a method is provided to determine an optimal relaxed parameter for the RGBK method. In addition, AGBK effectively accelerates the convergence of RGBK in running time. The examples for different cases show the advantage of the proposed RGBK and AGBK methods as long as the optimal relaxation parameter λ o p t is determined.

Acknowledgements

The authors thank the reviewers for providing some helpful comments. Research by Y.L. was partially supported by The Innovation Fund of Postgraduate, Sichuan University of Science & Engineering (grant y2021101), Research by F.Y. was partially supported by the Opening Project of Sichuan Province University Key Laboratory of Bridge Non-destruction Detecting and Engineering Computing (grant 2020QZJ03) and SUSE (grant 2019RC09, 2020RC25) and research by G.H. was supported in part by Application Fundamentals Foundation of STD of Sichuan (grant 2020YJ0366).

Conflicts of Interest

The authors declare no conflicts of interest regarding the publication of this paper.

References

[1] Niu, Y.Q. and Zheng, B. (2020) A Greedy Block Kaczmarz Algorithm for Solving Large-Scale Linear Systems. Applied Mathematics Letters, 104, Article ID: 106294.
https://doi.org/10.1016/j.aml.2020.106294
[2] Kaczmarz, S. (1937) Angenäherte auflösung von systemen linearer Gleichungen. Bulletin International de l’ Académie Polonaise des Sciences et des Lettres, 35, 355-357.
[3] Gordon, R., Bender, R. and Herman, G. (1970) Algebraic Reconstruction Techniques (ART) for Three-Dimensional Electron Microscopy and X-Ray Photography. Journal of Theoretical Biology, 29, 471-481.
https://doi.org/10.1016/0022-5193(70)90109-8
[4] Censor, Y. (1988) Parallel Application of Block-Iterative Methods in Medical Imaging and Radiation Therapy. Mathematical Programming, 42, 307-325.
https://doi.org/10.1007/BF01589408
[5] Byrne, C. (2004) A Unified Treatment of Some Iterative Algorithms in Signal Processing and Image Reconstruction. Inverse Problems, 20, 103-120.
https://doi.org/10.1088/0266-5611/20/1/006
[6] Herman, G.T. (2009) Fundamentals of Computerized Tomography: Image Reconstruction from Projections. 2nd Edition, Springer, Dordrecht.
[7] Elble, J.M. and Sahinidis, N.V. and Vouzis, P. (2010) GPU Computing with Kaczmarz and Other Iterative Algorithms for Linear Systems. Parallel Computing, 36, 215-231.
https://doi.org/10.1016/j.parco.2009.12.003
[8] Eldar, Y.C. and Needell, D. (2011) Acceleration of Randomized Kaczmarz Method via the Johnson-Lindenstrauss Lemma. Numerical Algorithms, 58, 163-177.
https://doi.org/10.1007/s11075-011-9451-z
[9] Needell, D. and Tropp, J.A. (2014) Paved with Good Intentions: Analysis of a Randomized Block Kaczmarz Method. Linear Algebra and Its Applications, 441, 199-221.
https://doi.org/10.1016/j.laa.2012.12.022
[10] Zouzias, A. and Freris, N. (2012) Randomized Extended Kaczmarz for Solving Least Squares. SIAM Journal on Matrix Analysis and Applications, 34, 773-793.
https://doi.org/10.1137/120889897
[11] Briskman, J. and Needell, D. (2015) Block Kaczmarz Method with Inequalities. Journal of Mathematical Imaging and Vision, 52, 385-396.
https://doi.org/10.1007/s10851-014-0539-7
[12] Ma, A., Needell, D. and Ramda, A.S. (2015) Convergence Properties of the Randomized Extended Gauss-Seidel and Kaczmarz Methods. SIAM Journal on Matrix Analysis and Applications, 36, 1590-1604.
https://doi.org/10.1137/15M1014425
[13] Needell, D., Zhao, R. and Zouzias A. (2015) Randomized Block Kaczmarz Method with Projection for Solving Least Squares. Linear Algebra and its Applications, 484, 322-343.
https://doi.org/10.1016/j.laa.2015.06.027
[14] Bai, Z.Z. and Wu, W.T. (2019) On Partially Randomized Extended Kaczmarz Method for Solving Large Sparse Overdetermined Inconsistent Linear Systems. Linear Algebra and Its Applications, 578, 225-250.
https://doi.org/10.1016/j.laa.2019.05.005
[15] Du, K. (2019) Tight Upper Bounds for the Convergence of the Randomized Extended Kaczmarz and Gauss-Seidel Algorithms. Numerical Linear Algebra with Applications, 26, e2233.
https://doi.org/10.1002/nla.2233
[16] Richtarik, P. and Takavc, M. (2020) Stochastic Reformulations of Linear Systems Algorithms and Convergence Theory. SIAM Journal on Matrix Analysis and Applications, 41, 487-524.
https://doi.org/10.1137/18M1179249
[17] Strohmer, T. and Vershynin, R. (2009) A Randomized Kaczmarz Algorithm with Exponential Convergence. Journal of Fourier Analysis and Applications, 15, 262-278.
https://doi.org/10.1007/s00041-008-9030-4
[18] Zhang, J.J. (2018) A New Greedy Kaczmarz Algorithm for the Solution of Very Large Linear Systems. Applied Mathematics Letters, 91, 207-212.
https://doi.org/10.1016/j.aml.2018.12.022
[19] Bai, Z.Z. and Wu, W.T. (2018) On Greedy Randomized Kaczmarz Method for Solving Large Sparse Linear Systems. SIAM Journal on Scientific Computing, 40, A592-A606.
https://doi.org/10.1137/17M1137747
[20] Bai, Z.Z. and Wu, W.T. (2018) On Relaxed Greedy Randomized Kaczmarz Methods for Solving Large Sparse Linear Systems. Applied Mathematics Letters, 83, 21-26.
https://doi.org/10.1016/j.aml.2018.03.008
[21] Björck, A. (1996) Numerical Methods for Least Squares Problems. SIAM, Philadelphia.
https://doi.org/10.1137/1.9781611971484
[22] Gower, R.M. and Richtárik, P. (2015) Randomized Iterative Methods for Linear Systems. SIAM Journal on Matrix Analysis and Applications, 36, 1660-1690.
https://doi.org/10.1137/15M1025487
[23] Kashkari, B. and Alqarni, S. (2019) Two-Step Hybrid Block Method for Solving Nonlinear Jerk Equations. Journal of Applied Mathematics and Physics, 7, 1897-1910.
https://doi.org/10.4236/jamp.2019.78130
[24] Duromola, M. and Momoh, A. (2019) Hybrid Numerical Method with Block Extension for Direct Solution of Third Order Ordinary Differential Equations. American Journal of Computational Mathematics, 9, 68-80.
https://doi.org/10.4236/ajcm.2019.92006
[25] Davis, T.A. and Hu, Y. (2011) The University of Florida Sparse Matrix Collection. ACM Transactions on Mathematical Software, 38, 1-25.
https://doi.org/10.1145/2049662.2049663
[26] Hansen, R.C. and Jorgensen, J.S (2018) AIR Tools II: Algebraic Iterative Reconstruction Methods. Numerical Algorithms, 79, 107-137.
https://doi.org/10.1007/s11075-017-0430-x

Copyright © 2024 by authors and Scientific Research Publishing Inc.

Creative Commons License

This work and the related PDF file are licensed under a Creative Commons Attribution 4.0 International License.