Article 54R56 Convex function of diagonals and eigenvalues

Convex function of diagonals and eigenvalues

by
John
from John D. Cook on (#54R56)

Sam Walters posted an elegant theorem on his Twitter account this morning. The theorem follows the pattern of an equality for linear functions generalizing to an inequality for convex functions. We'll give a little background, state the theorem, and show an example application.

Let A be a real symmetric n*n matrix, or more generally a complex n*n Hermitian matrix, with entries aij. Note that the diagonal elements aii are real numbers even if some of the other entries are complex. (A Hermitian matrix equals its conjugate transpose, which means the elements on the diagonal equal their own conjugate.)

A general theorem says that A has n eigenvalues. Denote these eigenvalues 1, 2, ..., n.

It is well known that the sum of the diagonal elements of A equals the sum of its eigenvalues.

trace_eigen0.svg

We could trivially generalize this to say that for any linear function : R R,

trace_eigen1.svg

because we could pull any shifting and scaling constants out of the sum.

The theorem Sam Walters posted says that the equality above extends to an inequality if is convex.

trace_eigen2.svg

Here's an application of this theorem. Assume the eigenvalues of A are all positive and let (x) = - log(x). Then is convex, and

trace_eigen3.svg

and so

trace_eigen4.svg

i.e. the product of the diagonals of A is an upper bound on the determinant of A.

This post illustrates two general principles:

  1. Linear equalities often generalize to convex inequalities.
  2. When you hear a new theorem about convex functions, see what it says about exp or -log.
More linear algebra postsLYeG3WLSF5c
External Content
Source RSS or Atom Feed
Feed Location http://feeds.feedburner.com/TheEndeavour?format=xml
Feed Title John D. Cook
Feed Link https://www.johndcook.com/blog
Reply 0 comments