The exponential is one of the most fundamental functions in mathematics, and especially in differential equations. Using Taylor series, it is possible to prove that
ex=k≥0∑k!xk(1)
for any x∈R. This sum is absolutely convergent for every x, so it also makes sense even in situations where x is not a real number. This is, for instance, how ez is defined when z is a complex number, giving rise to the meme-worthy formula
eπi=−1.
This sum is also how you define the exponential of a (square) matrix.
The Matrix Exponential
The exponential of a real number shows up literally everywhere when studying differential equations with a single variable, such as
dtdf(t)=t.
In the same way, the matrix exponential shows up all over the place when studying systems of differential equations, such as
dtdf0(x)=f1(x)dtdf1(x)=f0(x)
In particular, the matrix exponential shows up in linear differential equations. Take for instance the following initial value problem w(t)∈Rn, and A∈Mn×n(R).
dtdw(t)=Aw(t)
w(0)=w0
This has the (unique) solution
w(t)=eAtw0.
This solution can be verified by replacing eAt with its defining Taylor series, and differentiating term by term. Estimates of the matrix exponential help mathematicians to understand the behavior of vector-valued differential equations.
Preliminary Bounds
My goal in this post is to present a neat bound on the matrix exponential, but first I think it is prudent to go through some alternative bounds on the exponential. Because the following discussion requires us to use eigenvalues and eigenvectors, which may be complex, we naturally will naturally work with matrices in Mn×n(C). For A∈Mn×n(C), let ∥A∥ denote its spectral norm.
Within the complex numbers, exponential is very well-understood. Euler's famous formula gives that:
ez=eℜz(cos(ℑz)+isin(ℑz))
In particular,
∣ez∣=eℜz≤e∣z∣(2)
Now, consider exponentials in Mn×n(C). We get the basic bound
∥eA∥=1+1!x+2!x2…
≤1+1!∥x∥+2!∥x∥⋯≤e∥x∥
Thus,
∥eA∥≤e∥A∥(3)
This bound can be pretty lousy at times, but it is a start. For instance, if A=−t1, where 1 is the identity matrix, then ∥eA∥=e−t. In constrast, the above bound only guarantees that ∥eA∥≤et.
For certain matrices, we can compute eA exactly, and use this to get a much more exact bound. Let A be a diagonal matrix:
A=a10⋮0a2⋮……an
Then, the matrix exponential is exactly
eA=ea10⋮0ea2⋮……ean
and we have the following bound
∥eA∥=i=1…nmaxeℜai(4)
If A is diagonalizable, then we can get a bound similar to (4). Suppose that A=TDT−1, where D is a diagonal matrix. Then,
eA=eTDT−1=TeDT−1
The diagonal entries of D are exactly the eigenvalues of A. Let λ1…λn be the eigenvalues of A, and define
Λ(A)=i=1…nmaxℜλi.
It can be shown that Λ is a continuous function Mn×n(C)→R. With this notation, we can write the approximation:
∥eA∥≤∥T∥∥T−1∥eΛ(A)(5)
More generally, any square matrix A is similar to a matrix J in Jordan canonical form, A=TJT−1. The exponential eJ can be computed exactly, yielding the bound:
∥eA∥≤Cn∥T∥∥T−1∥(1+∥A∥n−1)eΛ(A)(6)
for some constant Cn>0 that depends only on n, the dimension. This bound, although very useful, is not entirely satisfying because it depends on T, which may be very hard to compute and may be arbitrarily large.
A Pretty Good Bound
It turns out that:
∥eA∥≤Cn(1+∥A∥n−1)eΛ(A)(7)
The constant Cn may be quite large, but it is at most exponential in n. Since Λ is a continuous function, this bound is actually continuous in A, which is why I like it so much.
For a period of time, I was very interested in understanding the matrix exponential, but I had trouble finding any bounds similar to (7). This is one of the reasons why I'm making this blog post. Eventually, with the help of Math Stack Exchange, I received an answer from user Anton Malyshev. The proof that I present here is a paraphrasing of his proof.
Proof: If n=0, then this bound is clearly true. Proceed by induction on n.
Let A∈Mn×n(C). Let v be a unit eigenvector of A with eigenvalue λ. Without loss of generality, suppose that Λ(A)=ℜλ. Let V be the 1-dimensional subspace of Cn spanned by v, and let W be the orthogonal subspace to V. Note that dimW=n−1. Find b∈V and D:V→V such that, for w∈W,
Aw=(b⋅w)v+Dw
Informally, we could write A in block matrix form as:
A=(λ0bD)
To compute eA, consider the initial value problem
u(0)=u0
dtdu(t)=Au(t)(8)
The (unique) solution to this differential equation is:
u(t)=eAtu0
Thus, to bound eA, it suffices to bound u(1).
We next project equation (8) onto V and W. Let u(t)=uV(t)+uW(t) where uV(t)∈V and uW(t)∈W. Then, (8) becomes the following:
Now, uW(t) satisfies a simple linear equation. We can thus solve for uW(t) exactly.
uW(t)=eDtuW(0)
By induction, we have the bound:
∥uW(t)∥≤Cn−1(1+∥D∥n−2tn−2)eΛ(D)t∥uW(0)∥
Note that ∥D∥≤∥A∥, Λ(D)≤Λ(A), and ∥uW(0)∥≤∥u0∥. This yields the following, somewhat simpler bound:
∥uW(t)∥≤Cn−1(1+∥A∥n−2tn−2)eΛ(A)t∥u0∥(10)
Bounding uV(t) is trickier, but possible. Since uV(t) is a multiple of V, we can define α:R→C by: uV(t)=α(t)v. Plugging this in to (9) gives us the differential equation:
dtdα(t)=λα(t)+b⋅uW(t)(11)
The solution to to this differential equation can be approximated using integrating factors. Define β(t)=e−λtα(t). Some routine calculations follow...