Skip to main content
Log in

Optimization algorithms on the Grassmann manifold with application to matrix eigenvalue problems

  • Original Paper
  • Area 2
  • Published:
Japan Journal of Industrial and Applied Mathematics Aims and scope Submit manuscript

Abstract

This article deals with the Grassmann manifold as a submanifold of the matrix Euclidean space, that is, as the set of all orthogonal projection matrices of constant rank, and sets up several optimization algorithms in terms of such matrices. Interest will center on the steepest descent and Newton’s methods together with applications to matrix eigenvalue problems. It is shown that Newton’s equation in the proposed Newton’s method applied to the Rayleigh quotient minimization problem takes the form of a Lyapunov equation, for which an existing efficient algorithm can be applied, and thereby the present Newton’s method works efficiently. It is also shown that in case of degenerate eigenvalues the optimal solutions form a submanifold diffeomorphic to a Grassmann manifold of lower dimension. Furthermore, to generate globally converging sequences, this article provides a hybrid method composed of the steepest descent and Newton’s methods on the Grassmann manifold together with convergence analysis.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6

Similar content being viewed by others

References

  1. Abraham, R., Marsden, J.E.: Foundations of Mechanics. Benjamin/Cummings Publishing, Reading (1978)

    MATH  Google Scholar 

  2. Absil, P.-A., Mahony, R., Sepulchre, R.: Optimization Algorithms on Matrix Manifolds. Princeton University Press, Princeton (2008)

  3. Absil, P.-A., Mahony, R., Sepulchre, R.: Riemannian geometry of Grassmann manifolds with a view on algorithmic computation. Acta Appl. Math. 80(2), 199–220 (2004)

    Google Scholar 

  4. Adler, R.L., Dedieu, J.P., Margulies, J.Y., Martens, M., Shub, M.: Newton’s method on Riemannian manifolds and a geometric model for the human spine. IMA J. Numer. Anal. 22(3), 359–390 (2002)

    Google Scholar 

  5. Arnold, V.I.: Mathematical Methods of Classical Mechanics, 2nd edn. Springer, Newyork (1989)

    Book  Google Scholar 

  6. Edelman, A., Arias, T.A., Smith, S.T.: The geometry of algorithms with orthogonality constraints. SIAM J. Matrix Anal. Appl. 20(2), 303–353 (1998)

    Google Scholar 

  7. Ferrer, J., García, M.I., Puerta, F.: Differentiable families of subspaces. Linear Algebra Appl. 199, 229–252 (1994)

    Google Scholar 

  8. Fujii, K.: Note on coherent states and adiabatic connections, curvatures. J. Math. Phys. 41, 4406–4412 (2000)

    Google Scholar 

  9. Gajic, Z., Qureshi, M.T.J.: Lyapunov Matrix Equation in System Stability and Control. Academic Press, Inc, New York (1995)

  10. Golub, G.H., Loan, C.F.V.: Matrix Computations. Johns Hopkins Studies in the Mathematical Sciences, 3rd edn. Johns Hopkins University Press, Baltimore (1996)

  11. Helmke, U., Moore, J.B.: Optimization and Dynamical Systems. Communications and Control Engineering Series. Springer, London (1994) (With a foreword by R. Brockett)

  12. Helmke, U., Huper, K., Trumpf, J.: Newton’s method on Grassmann manifolds. arXiv:0709.2205v2 (2007)

  13. Nocedal, J., Wright, S.J.: Numerical Optimization. Springer Series in Operations Research. Springer, New York (1999)

  14. Sato, H., Iwai, T.: A Riemannian optimization approach to the matrix singular value decomposition. SIAM. J. Optim. 23(1), 188–212 (2013)

    Google Scholar 

  15. Snyman, J.A.: Practical Mathematical Optimization: An Introduction to Basic Optimization Theory and Classical and New Gradient-Based Algorithms. Springer, New York (2005)

  16. Tanimura, S., Nakahara, M., Hayashi, D.: Exact solutions of the isoholonomic problem and the optimal control problem in holonomic quantum computation. J. Math. Phys. 46, 022101 (2005)

    Google Scholar 

  17. Trefethen, L.N., Bau, D.: Numerical Linear Algebra. SIAM (1997)

  18. Wong, Y.-C.: Differential geometry of Grassmann manifolds. Proc. Natl. Acad. Sci. USA 57, 589–594 (1967)

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Hiroyuki Sato.

Appendix A: Proof of Proposition 2.4

Appendix A: Proof of Proposition 2.4

In this section, we give the proof of Proposition 2.4 on the variational principle.

Proof

Since the geodesic equation of a Riemannian manifold is viewed as the equation of motion of a free particle on the Riemannian manifold [5], we consider the Lagrangian \(L\) of a free particle on the Grassmann manifold (2.13), which is put in the form

$$\begin{aligned} L=\frac{1}{2}\langle {\dot{X}},{\dot{X}}\rangle _X+{\mathrm{tr}}\left( \Omega \left( X^2-X\right) \right) +\lambda ({\mathrm{tr}}(X)-p), \end{aligned}$$
(6.1)

where \(\Omega \) and \(\lambda \) are Lagrange multipliers, and where \(\Omega \) should be a symmetric matrix on account of the fact that \(X^2-X\) is symmetric. The variation of the Lagrangian \(L\) is given by and calculated as

$$\begin{aligned} \delta L&={\mathrm{tr}}(\delta {\dot{X}}{\dot{X}})+{\mathrm{tr}}(\delta \Omega (X^2-X))\!+\!{\mathrm{tr}}(\Omega (2X\delta X-\delta X))\!+\!\delta \lambda ({\mathrm{tr}}(X)-p)+\lambda {\mathrm{tr}}(\delta X)\nonumber \\&=\frac{d}{dt}({\mathrm{tr}}(\delta X{\dot{X}}))+{\mathrm{tr}}(\delta X(-{\ddot{X}}+\Omega (2X-I_n)+\lambda I_n))+{\mathrm{tr}}(\delta \Omega (X^2-X))\nonumber \\&\quad +\delta \lambda ({\mathrm{tr}}(X)-p). \end{aligned}$$
(6.2)

On the variational principle, we have

$$\begin{aligned} \int _{t_1}^{t_2}\delta L\,dt=0 \end{aligned}$$
(6.3)

for any \(\delta X\) subject to the condition \(\delta X(t_1)=\delta X(t_2)=0\). From (6.2) and (6.3), we obtain

$$\begin{aligned} -{\ddot{X}}+\Omega (2X-I_n)+\lambda I_n=0,\end{aligned}$$
(6.4)
$$\begin{aligned} X^2-X=0,\end{aligned}$$
(6.5)
$$\begin{aligned} {\mathrm{tr}}(X)-p=0. \end{aligned}$$
(6.6)

Our next task is to determine \(\Omega \) and \(\lambda \). Transposing Eq. (6.4), we have

$$\begin{aligned} -\ddot{X}+(2X-I_n)\Omega +\lambda I_n=0. \end{aligned}$$
(6.7)

Equations (6.4) and (6.7) are put together to provide

$$\begin{aligned} X\Omega =\Omega X. \end{aligned}$$
(6.8)

Multiplying (6.4) by \(X\) from the right and using \(X^2=X\), we obtain

$$\begin{aligned} \Omega X={\ddot{X}} X-\lambda X. \end{aligned}$$
(6.9)

We see that \({\ddot{X}}X=X{\ddot{X}}\) from (6.8) and (6.9). Putting together (6.9) and (6.4), we have

$$\begin{aligned} \Omega =-{\ddot{X}}+2{\ddot{X}}X+\lambda (I_n-2X). \end{aligned}$$
(6.10)

On the other hand, differentiating \(X^2=X\) with respect to \(t\), we obtain

$$\begin{aligned} {\ddot{X}}=X{\ddot{X}}+2{\dot{X}}^2+{\ddot{X}}X. \end{aligned}$$
(6.11)

Since \({\ddot{X}}X=X{\ddot{X}}\), the above equation becomes

$$\begin{aligned} {\ddot{X}}=2{\ddot{X}}X+2{\dot{X}}^2. \end{aligned}$$
(6.12)

On account of (6.12), Eq. (6.10) is put in the form

$$\begin{aligned} \Omega =-2{\dot{X}}^2+\lambda (I_n-2X). \end{aligned}$$
(6.13)

Substituting (6.13) for \(\Omega \) in (6.4), we obtain

$$\begin{aligned} {\ddot{X}}+4{\dot{X}}^2 X-2{\dot{X}}^2=0. \end{aligned}$$
(6.14)

Since \({\dot{X}} X+X{\dot{X}}={\dot{X}}\), one has \({\dot{X}}^2X+{\dot{X}}X{\dot{X}}={\dot{X}}^2\), and hence Eq. (6.14) is brought into

$$\begin{aligned} {\ddot{X}}+2{\dot{X}}^2-4{\dot{X}}X{\dot{X}}=0. \end{aligned}$$
(6.15)

This completes the proof. \(\square \)

About this article

Cite this article

Sato, H., Iwai, T. Optimization algorithms on the Grassmann manifold with application to matrix eigenvalue problems. Japan J. Indust. Appl. Math. 31, 355–400 (2014). https://doi.org/10.1007/s13160-014-0141-9

Download citation

  • Received:

  • Revised:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s13160-014-0141-9

Keywords

Mathematics Subject Classification (2000)

Navigation