Abstract
In this paper we investigate the performance of four different SOR acceleration techniques on a variety of linear systems. Two of these techniques have been proposed by Dancis [1] who uses a polynomial acceleration together with a sub-optimal ω. The two other techniques discussed are vector accelerations; the ε algorithm proposed by Wynn [9] and a generalisation of Aitken’s Δ2 algorithm, proposed by Graves-Morris [3].
The experimental results show that these accelerations can reduce the amount of work required to obtain a solution and that their rates of convergence are generally less sensitive to the value of ω than the straightforward SOR method. However a poor choice of ω can result in particularly inefficient solutions and more work is required to enable cheap estimates of a effective parameter to be obtained.
Necessary conditions for the reduction in the computational work required for convergence are given for each of the accelerations, based on the number of floating-point operations.
It is shown experimentally that the reduction in the number of iterations is related to the separation between the two largest eigenvalues of the SOR iteration matrix for a given ω. This separation influences the convergence of all the acceleration techniques above.
Another important characteristic exhibited by these accelerations is that even if the number of iterations is not reduced significantly compared to the SOR method, they are competitive in terms of number of floating-point operations used and thus they reduce the overall computational workload.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
J. Dancis. The optimal w is not best for the SOR iteration method. Linear Algebra and its Applications, 154–156:819–845, 1991.
I.S. Duff, R.G. Grimes, and J.G. Lewis. Users’ guide for the Harwell-Boeing sparse matrix collection. Report TR/PA/92/86, CERFACS, October 1992.
P.R. Graves-Morris. A review of Padé methods for the acceleration of convergence of a sequence of vectors. Report No. NA 92–31, Department of Mathematics, University of Bradford, October 1992.
R.G. Grimes, D.R. Kincaid, and D.M. Young. ITPACK 2.0 user’s guide. Report No. CNA-150, Center for Numerical Analysis, University of Texas at Austin, August 1979.
L.A. Hageman and D.M. Young. Applied Iterative Methods. Academic Press, New York, 1981.
D.R. Kincaid and W. Cheney. Numerical Analysis. Brooks/Cole, Pacific Grove, 1991.
R.S. Varga. Matrix Iterative Analysis. Prentice-Hall, Englewood Cliffs, 1962.
P. Wynn. On a device for computing the em(S„) transformation. Mathematical Tables and Aids of Computation, 10:91–96, 1956.
P. Wynn. Acceleration techniques for iterated vector and matrix problems. Mathematics of Computation, 16:301–322, 1962.
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 1994 Springer Science+Business Media Dordrecht
About this chapter
Cite this chapter
Da Cunha, R.D., Hopkins, T. (1994). A Comparison of Acceleration Techniques Applied to the Sor Method. In: Cuyt, A. (eds) Nonlinear Numerical Methods and Rational Approximation II. Mathematics and Its Applications, vol 296. Springer, Dordrecht. https://doi.org/10.1007/978-94-011-0970-3_21
Download citation
DOI: https://doi.org/10.1007/978-94-011-0970-3_21
Publisher Name: Springer, Dordrecht
Print ISBN: 978-94-010-4420-2
Online ISBN: 978-94-011-0970-3
eBook Packages: Springer Book Archive