Derivatives of Eigenvalues and Eigenvectors

This appendix summarizes some of the results in De Leeuw [2007], De Leeuw [2008], and De Leeuw and Sorenson [2012]. We refer to those reports for more extensive calculations and applications.

Suppose and are two real symmetric matrices depending smoothly on a real parameter . The notation below suppresses the dependence on of the various quantities we talk about, but it is important to remember that all eigenvalues and eigenvectors we talk about are functions of .

The generalized eigenvalue and the corresponding generalized eigenvector are defined implicitly by . Moreover the eigenvector is identified by . We suppose that in a neighborhood of the eigenvalue is unique and is positive definite. A precise discussion of the required assumptions is, for example, in Wilkinson [1965] or Kato [1976].

Differentiating gives the equation while gives Premultiplying by gives Now suppose with . Then from , for , premultiplying by gives If we define by then and thus .


A first important special case is the ordinary eigenvalue problem, in which which obviously does not depend on , and consequently has . Then while If we use the Moore_Penrose inverse the derivatives of the eigenvector can be written as Written in a different way this expression is with , so that .


In the next important special case is the singular value problem The singular values and vectors of an rectangular , with , solve the equations and . It follows that , i.e. the right singular vectors are the eigenvectors and the singular values are the square roots of the eigenvalues of .

Now we can apply our previous results on eigenvalues and eigenvectors. If then . We have, at an isolated singular value , and thus For the singular vectors our previous results on eigenvectors give and in the same way

Now let , with and square orthonormal, and with and diagonal matrix (with positive diagonal entries in non-increasing order along the diagonal).

Also define . Then , and and Note that if is symmetric we have and is symmetric, so we recover our previous result for eigenvectors. Also note that if the parameter is actually element of , i.e. if we are computing partial derivatives, then .


The results on eigen and singular value decomposition can be applied in many different ways. mostly by simply using the product rule for derivatives, For a square symmetric or order , for example, we have and thus The generalized inverse of a rectangular is where . Summation is over the positive singular values, and for differentiability we must assume that the rank of is constant in a neighborhood of .

The Procrustus transformation of a rectangular , which is the projection of on the Stiefel manifold of orthonormal matrices, is where we assume for differentiability that is of full column rank.

The projection of on the set of all matrices of rank less than or equal to , which is of key importance in PCA and MDS, is where summation is over the largest singular values.