We propose a new uniform framework of Compact Rational Krylov (CORK) methods for solving large-scale nonlinear eigenvalue problems: A(λ)x = 0. For many years, linearizations are used for solving polynomial and rational eigenvalue problems. On the other hand, for the general nonlinear case, A(λ) can first be approximated by a (rational) matrix polynomial and then a convenient linearization is used. However, the major disadvantage of linearization based methods is the growing memory and orthogonalization costs with the iteration count, i.e., in general they are proportional to the degree of the polynomial. Therefore, the CORK family of rational Krylov methods exploits the structure of the linearization pencils by using a generalization of the compact Arnoldi decomposition. In this way, the extra memory and orthogonalization costs due to the linearization of the original eigenvalue problem are negligible for large-scale problems. Furthermore, we prove that each CORK step breaks down into an orthogonalization step of the original problem dimension and a rational Krylov step on small matrices. We also briefly discuss implicit restarting of the CORK method and how to exploit low rank structure. The CORK method is illustrated with two large-scale examples.Keywords : linearization, matrix pencil, rational Krylov, nonlinear eigenvalue problem MSC : 65F15, 15A22.
COMPACT RATIONAL KRYLOV METHODS FOR NONLINEAR EIGENVALUE PROBLEMS *ROEL VAN BEEUMEN † , KARL MEERBERGEN † , AND WIM MICHIELS † Abstract. We propose a new uniform framework of Compact Rational Krylov (CORK) methods for solving large-scale nonlinear eigenvalue problems: A(λ)x = 0. For many years, linearizations are used for solving polynomial and rational eigenvalue problems. On the other hand, for the general nonlinear case, A(λ) can first be approximated by a (rational) matrix polynomial and then a convenient linearization is used. However, the major disadvantage of linearization based methods is the growing memory and orthogonalization costs with the iteration count, i.e., in general they are proportional to the degree of the polynomial. Therefore, the CORK family of rational Krylov methods exploits the structure of the linearization pencils by using a generalization of the compact Arnoldi decomposition. In this way, the extra memory and orthogonalization costs due to the linearization of the original eigenvalue problem are negligible for large-scale problems. Furthermore, we prove that each CORK step breaks down into an orthogonalization step of the original problem dimension and a rational Krylov step on small matrices. We also briefly discuss implicit restarting of the CORK method and how to exploit low rank structure. The CORK method is illustrated with two large-scale examples.