Volume 7, pp. 163-181, 1998.

Restarting techniques for the (Jacobi-)Davidson symmetric eigenvalue methods

Andreas Stathopoulos and Yousef Saad


The (Jacobi-)Davidson method, which is a popular preconditioned extension to the Arnoldi method for solving large eigenvalue problems, is often used with restarting. This has significant performance shortcomings, since important components of the invariant subspace may be discarded. One way of saving more information at restart is through “thick” restarting, a technique that involves keeping more Ritz vectors than needed. This technique and especially its dynamic version have proved very efficient for symmetric cases. A different restarting strategy for the Davidson method has been proposed in [14], motivated by the similarity between the spaces built by the Davidson and Conjugate Gradient methods. For the latter method, a three term recurrence implicitly maintains all required information. In this paper, we consider the effects of preconditioning on the dynamic thick restarting strategy, and we analyze both theoretically and experimentally the strategy based on Conjugate Gradient. Our analysis shows that, in some sense, the two schemes are complementary, and that their combination provides an even more powerful technique. We also describe a way to implement this scheme without additional orthogonalizations or matrix multiplications.

Full Text (PDF) [289 KB], BibTeX

Key words

Davidson, Jacobi-Davidson, Lanczos, Conjugate Gradient methods, eigenvalue, implicit restarting, deflation, preconditioning.

AMS subject classifications


Links to the cited ETNA articles

[1]Vol. 2 (1994), pp. 1-21 D. Calvetti, L. Reichel, and D. C. Sorensen: An implicitly restarted Lanczos method for large symmetric eigenvalue problems

< Back