Home/Chapter 11

Stability and Lyapunov's Method

Stability criteria for linear systems. Lyapunov's direct method for general nonlinear systems. Linearization for stability analysis. Discrete-time stability.

Introduction

In many engineering applications, one wants to make sure the system behaves well in the long-run; without worrying too much about the particular path the system takes (so long as these paths are acceptable). Stability is the characterization ensuring that a system behaves well in the long-run. We will make this statement precise in the following.


Stability Criteria

Consider dxdt=f(x)\frac{dx}{dt} = f(x), with initial condition x(0)x(0). Let f(0)=0f(0) = 0. We will consider three definitions of stability:

DefinitionLocal Stability (Lyapunov Stability)

For every ϵ>0\epsilon > 0, there exists δ>0\delta > 0 such that x(0)<δ\|x(0)\| < \delta implies that x(t)<ϵ\|x(t)\| < \epsilon, t0\forall t \geq 0. This is also known as stability in the sense of Lyapunov.

Remark.

Intuition: Think of a ball resting at the bottom of a bowl. If you nudge it slightly, it stays close to the bottom -- it does not fly away. Local stability says exactly this: small perturbations produce small deviations, forever. Note that the ball does not have to return to the bottom; it just cannot wander far. A frictionless pendulum at its lowest point is locally stable (it oscillates nearby) but not asymptotically stable (it never stops oscillating).

DefinitionLocal Asymptotic Stability

δ>0\exists \delta > 0 such that x(0)<δ\|x(0)\| < \delta implies that limtx(t)=0\lim_{t \to \infty} \|x(t)\| = 0.

Remark.

Intuition: Now add friction to the ball-in-a-bowl picture. The ball not only stays near the bottom but actually returns to it over time. Local asymptotic stability means small perturbations eventually die out completely. The "local" qualifier matters: there is only a neighborhood around the equilibrium from which the system returns. Push the ball too far (over the rim of the bowl), and all bets are off.

DefinitionGlobal Asymptotic Stability

For every x(0)Rnx(0) \in \mathbb{R}^n, limtx(t)=0\lim_{t \to \infty} \|x(t)\| = 0. Hence, here, for any initial condition, the system converges to 0.

Remark.

Intuition: Global asymptotic stability is the strongest guarantee: no matter where you start in the entire state space, the system eventually returns to the equilibrium. Think of it as a bowl that extends infinitely in all directions with friction everywhere -- no matter how far you throw the ball, it always rolls back to the bottom. For linear systems, local and global asymptotic stability are equivalent, but for nonlinear systems they can differ dramatically.

It should be added that, stability does not necessarily need to be defined with regard to 0; that is stability can be defined around any zRnz \in \mathbb{R}^n with f(z)=0f(z) = 0. In this case, the above norms above should be replaced with x(t)z\|x(t) - z\| (such that, for example for the asymptotic stability case, x(t)x(t) will converge to zz).

One could consider an inverted pendulum as an example of a system which is not locally stable.

Linear Systems

Consider an initial value problem dxdt=Ax\frac{dx}{dt} = Ax, with initial conditions x(0)=x0x(0) = x_0. As studied earlier, the solution is

x(t)=eAtx0x(t) = e^{At}x_0

where

eAt=I+At+A2t22+Antnn!+e^{At} = I + At + A^2\frac{t^2}{2} + \ldots A^n\frac{t^n}{n!} + \ldots

is the matrix exponential (see Exercise). We now briefly review how to compute matrix exponentials with a focus on stability properties.

Case: A=IA = I (Identity Matrix). In this case, eAt=I+It+I2t22!++Intnn!+e^{At} = I + It + I^2\frac{t^2}{2!} + \cdots + I^n\frac{t^n}{n!} + \ldots, and

eAt=[et00et]e^{At} = \begin{bmatrix} e^t & 0 \\ 0 & e^t \end{bmatrix}

Case: AA diagonal. With similar arguments, if AA is diagonal with

A=[λ1000λ2000λ3],A = \begin{bmatrix} \lambda_1 & 0 & 0 \\ 0 & \lambda_2 & 0 \\ 0 & 0 & \lambda_3 \end{bmatrix},

we obtain

eAt=[eλ1t000eλ2t000eλ3t],e^{At} = \begin{bmatrix} e^{\lambda_1 t} & 0 & 0 \\ 0 & e^{\lambda_2 t} & 0 \\ 0 & 0 & e^{\lambda_3 t} \end{bmatrix},

Hence, it is very easy to compute the exponential when the matrix is diagonal.

Commutativity property. Note now that if AB=BAAB = BA, that is, if AA and BB commute, then (see Exercise)

e(A+B)=eAeBe^{(A+B)} = e^A e^B

Case: AA in Jordan form. We will use commutativity to compute the matrix exponential in the case where the matrix is in a Jordan form. Let us write

A=[λ1100λ1100λ1]A = \begin{bmatrix} \lambda_1 & 1 & 0 \\ 0 & \lambda_1 & 1 \\ 0 & 0 & \lambda_1 \end{bmatrix}

as B+CB + C, where

B=[λ1000λ1000λ1],C=[010001000]B = \begin{bmatrix} \lambda_1 & 0 & 0 \\ 0 & \lambda_1 & 0 \\ 0 & 0 & \lambda_1 \end{bmatrix}, \qquad C = \begin{bmatrix} 0 & 1 & 0 \\ 0 & 0 & 1 \\ 0 & 0 & 0 \end{bmatrix}

We note that BC=CBBC = CB, for BB is the identity matrix multiplied by a scalar number. Hence, eAt=eBteCte^{At} = e^{Bt}e^{Ct}. All we need to compute is eCte^{Ct}, as we have already discussed how to compute eBte^{Bt}. Here, one should note that C3=0C^3 = 0.

More generally, for a Jordan matrix where the number of 1s off the diagonal of a Jordan block is k1k - 1, the kkth power is equal to 0.

Therefore,

eCt=I+Ct+C2t22!+C3t33!+,e^{Ct} = I + Ct + C^2\frac{t^2}{2!} + C^3\frac{t^3}{3!} + \ldots,

becomes

eCt=I+Ct+C2t22!=[1tt2/201t001]e^{Ct} = I + Ct + C^2\frac{t^2}{2!} = \begin{bmatrix} 1 & t & t^2/2 \\ 0 & 1 & t \\ 0 & 0 & 1 \end{bmatrix}

Hence,

eAt=[eλ1t000eλ1t000eλ1t][1tt2/201t001]=[eλ1tteλ1tt22eλ1t0eλ1tteλ1t00eλ1t]e^{At} = \begin{bmatrix} e^{\lambda_1 t} & 0 & 0 \\ 0 & e^{\lambda_1 t} & 0 \\ 0 & 0 & e^{\lambda_1 t} \end{bmatrix}\begin{bmatrix} 1 & t & t^2/2 \\ 0 & 1 & t \\ 0 & 0 & 1 \end{bmatrix} = \begin{bmatrix} e^{\lambda_1 t} & te^{\lambda_1 t} & \frac{t^2}{2}e^{\lambda_1 t} \\ 0 & e^{\lambda_1 t} & te^{\lambda_1 t} \\ 0 & 0 & e^{\lambda_1 t} \end{bmatrix}

General matrix. Now that we know how to compute the exponential of a Jordan form, we can proceed to study a general matrix. Let A=PBP1A = PBP^{-1}, where BB is in a Jordan form. Then,

Ak=(PBP1)k=PBkPA^k = (PBP^{-1})^k = PB^kP

Finally,

eA=PeBP1e^A = Pe^BP^{-1}

and

eAt=PeBtP1e^{At} = Pe^{Bt}P^{-1}

Hence, once we obtain a diagonal matrix or a Jordan form matrix BB, we can compute the exponential eAte^{At} very efficiently.

The main insight here is that the eigenvalues determine whether the system remains bounded or not. In case, we have a repeated eigenvalue of 0, then the Jordan form determines whether the solution remains bounded or not. We state the following theorem.

TheoremGlobal Asymptotic Stability of Linear Systems

For a linear differential equation

x=Ax,x' = Ax,

the solution is locally and globally asymptotically stable if and only if

maxλi{Re{λi}}<0,\max_{\lambda_i}\{\text{Re}\{\lambda_i\}\} < 0,

where Re{}\text{Re}\{\cdot\} denotes the real part of a complex number, and λi\lambda_i denotes the iith eigenvalue of AA.

Remark.

Intuition: For linear systems, stability is entirely determined by the eigenvalues of AA. Each eigenvalue corresponds to a natural "mode" of the system, and the real part of the eigenvalue determines whether that mode grows or decays exponentially. If every mode decays (all eigenvalues in the open left half-plane), the entire system is stable -- and this works globally, not just locally. This is why pole placement is such a central concept in control design: moving eigenvalues to the left half-plane is equivalent to stabilizing the system.

This is one of the most important results in linear systems theory. It says that for a linear system, all three notions of stability coincide and reduce to a simple eigenvalue check. If all eigenvalues are in the open left half-plane, the system is globally asymptotically stable; if any eigenvalue has positive real part, the system is unstable.

TheoremLocal Stability of Linear Systems

For a linear differential equation

x=Ax,x' = Ax,

the system is locally stable if and only if the following two conditions hold:

(i) maxλi{Re{λi}}0\max_{\lambda_i}\{\text{Re}\{\lambda_i\}\} \leq 0,

(ii) if Re{λi}=0\text{Re}\{\lambda_i\} = 0, for some λi\lambda_i, the algebraic multiplicity of this eigenvalue should be the same as the geometric multiplicity.

Remark.

Intuition: This theorem relaxes the asymptotic stability condition to mere boundedness. Eigenvalues on the imaginary axis are allowed (corresponding to sustained oscillations that neither grow nor decay), but only if they have no Jordan blocks larger than 1×11 \times 1. A Jordan block for a purely imaginary eigenvalue introduces polynomial growth factors like teiωtt \cdot e^{i\omega t}, which grow without bound even though eiωt=1|e^{i\omega t}| = 1. Think of a frictionless pendulum: simple oscillation is fine (simple eigenvalue), but resonance-like behavior (repeated eigenvalue with a Jordan block) would make the amplitude grow.

Condition (ii) prevents polynomial growth from Jordan blocks. If a purely imaginary eigenvalue has a Jordan block of size greater than 1, the terms teiωtte^{i\omega t}, t2eiωtt^2e^{i\omega t}, etc. will grow without bound even though eiωt=1|e^{i\omega t}| = 1.

In practice, many systems are not linear, and hence the above theorems are not applicable.


A General Approach: Lyapunov's Method

A very versatile and effective approach to stabilization is via Lyapunov functions (this approach is often called Lyapunov's second method, the first one being an analysis based on linearization to be considered after this section). Let ΩRn\Omega \subset \mathbb{R}^n be an open set containing the equilibrium point, taken to be 0 here without any loss.

DefinitionLyapunov Function

A function V:RnRV : \mathbb{R}^n \to \mathbb{R} is called a Lyapunov function if

  1. V(x)>0V(x) > 0, x0\forall x \neq 0, xΩx \in \Omega,
  2. V(x)=0V(x) = 0, if x=0x = 0,
  3. VV is continuous, and has continuous partial derivatives.
Remark.

Intuition: A Lyapunov function is a generalized notion of "energy" for a system. Just as physical energy is always non-negative and equals zero only at rest, V(x)V(x) is positive everywhere except at the equilibrium. The key idea is that you do not need to solve the differential to prove stability -- you just need to find an appropriate energy-like function and show it never increases (or strictly decreases) along the system's trajectories. For mechanical systems, the actual energy often works as a Lyapunov function, but for abstract systems you may need to be creative.

A Lyapunov function serves as a generalized notion of "energy" for a system. The idea is that if we can find a positive-definite function that never increases along trajectories of the system, then the trajectories must remain bounded (stability). If the function is strictly decreasing, the system must converge to the equilibrium (asymptotic stability).

First we present results on local asymptotic stability. As above, let ΩRn\Omega \in \mathbb{R}^n be an open set containing 0.

TheoremLyapunov's Direct Method

a) For a given differential x(t)=f(x(t))x'(t) = f(x(t)) with f(0)=0f(0) = 0, and continuous ff, if we can find a Lyapunov function VV such that

ddtV(x(t))0,\frac{d}{dt}V(x(t)) \leq 0,

for all x(t)=xΩx(t) = x \in \Omega, then, the system is locally stable (stable in the sense of Lyapunov).

b) For a given differential x(t)=f(x(t))x'(t) = f(x(t)) with f(0)=0f(0) = 0, and continuous ff, if we can find a Lyapunov function V(x)V(x) such that

ddtV(x(t))<0,\frac{d}{dt}V(x(t)) < 0,

for x(t)=xΩ{0}x(t) = x \in \Omega \setminus \{0\}, the system is locally asymptotically stable.

c) If b) holds for VV so that limxV(x)=\lim_{\|x\| \to \infty} V(x) = \infty, and

ddtV(x(t))<0,\frac{d}{dt}V(x(t)) < 0,

for x(t)=zRn{0}x(t) = z \in \mathbb{R}^n \setminus \{0\}, then the system is globally asymptotically stable.

Remark.

Intuition: Lyapunov's direct method is like proving a ball will reach the bottom of a valley without tracking its exact path. Part (a) says: if the system's "energy" never increases, the state stays close (stability). Part (b) says: if the energy strictly decreases, the state must converge to equilibrium (asymptotic stability). Part (c) adds the global guarantee: if the energy function grows to infinity in all directions (so there are no "escape routes"), then asymptotic stability holds everywhere. The power of this method is that it avoids solving the differential entirely -- you only need a suitable energy function.

TheoremRegion of Asymptotic Stability

Let us, in addition to the conditions noted in Theorem(b), further impose that for some ll, Ωl:={x:V(x)l}\Omega_l := \{x : V(x) \leq l\} is a bounded set and ΩlΩ\Omega_l \subset \Omega where Ω\Omega satisfies Theorem(b). Then, every solution of the system with initial state x(0)Ωlx(0) \in \Omega_l converges to equilibrium.

Remark.

Intuition: This theorem gives you a concrete, computable estimate of the "basin of attraction" -- the set of initial conditions from which the system is guaranteed to converge. The sublevel set {x:V(x)l}\{x : V(x) \leq l\} acts like a fence: since VV is decreasing along trajectories, once the state is inside this set, it can never leave. In practice, you pick ll as large as possible while keeping the sublevel set within the region where V˙<0\dot{V} < 0, giving you the largest provable region of attraction.

By following (and slightly modifying) the proof of Theorem(b), we can conclude that Ωl\Omega_l is a region of attraction for the equilibrium point, which is defined as a set of initial states whose corresponding solutions converge to the equilibrium point: {x(0):limtx(t)=0}\{x(0) : \lim_{t\to\infty} x(t) = 0\}.

Remark.

For local stability, by restricting the analysis to Ω\Omega, we can allow the Lyapunov function VV to take even negative values outside Ω\Omega or not necessarily be continuous outside Ω\Omega. In Theorem, we used such properties of VV only on Ω\Omega.

ExampleCubic System Stability

Show that x=x3x' = -x^3 is locally asymptotically stable, by picking V(x)=x2V(x) = x^2 as a Lyapunov function. Is this solution globally asymptotically stable as well?

We compute ddtV(x(t))=2xx=2x(x3)=2x4<0\frac{d}{dt}V(x(t)) = 2x \cdot x' = 2x(-x^3) = -2x^4 < 0 for all x0x \neq 0. Since V(x)=x2V(x) = x^2 \to \infty as x|x| \to \infty, by Theorem(c), the system is globally asymptotically stable.

ExampleRegion of Attraction

Show that x=2x+x3x' = -2x + x^3 is locally asymptotically stable, by picking V(x)=x2V(x) = x^2 as a Lyapunov function. Is this solution globally asymptotically stable? Find a region of attraction for local stability.

We compute ddtV(x(t))=2x(2x+x3)=4x2+2x4=2x2(x22)\frac{d}{dt}V(x(t)) = 2x(-2x + x^3) = -4x^2 + 2x^4 = 2x^2(x^2 - 2). This is negative when x2<2x^2 < 2, i.e., when x<2|x| < \sqrt{2}. So the system is locally asymptotically stable with region of attraction Ωl={x:x<2}\Omega_l = \{x : |x| < \sqrt{2}\}. The system is not globally asymptotically stable since for x>2|x| > \sqrt{2}, V˙>0\dot{V} > 0.

Remark.

One should note that BIBO stability and the stability notions considered in this chapter have very different contexts; BIBO stability is concerned with the input-output behaviour of systems and the criteria considered in this chapter are with regard to the effects of initial conditions (also called internal stability). The conditions are also slightly different for the linear setup: for continuous-time linear systems, BIBO stability requires all the eigenvalues to have strictly negative real parts. Stability itself, however, may require more relaxed conditions.

Revisiting the linear case

Recall that an n×nn \times n real matrix PP is positive definite if PP is symmetric and xTPx>0x^TPx > 0 for all x0x \neq 0. It is positive semi-definite of xTPx0x^TPx \geq 0 for all xRnx \in \mathbb{R}^n. Note that being symmetric is part of the definition.

TheoremLyapunov Equation for Linear Systems

All eigenvalues of a square matrix AA have negative real parts if and only if for any given positive definite NN, the (Lyapunov) equation

ATM+MA=NA^TM + MA = -N

has a unique solution MM, where the solution is positive definite.

Remark.

Intuition: The Lyapunov converts the problem of checking eigenvalue locations (a nonlinear problem) into solving a system of linear equations for MM. If you can find a positive definite MM satisfying ATM+MA=NA^TM + MA = -N, then V(x)=xTMxV(x) = x^TMx is a Lyapunov function that certifies stability. In MATLAB, this is a single command (lyap). The theorem says this always works for stable linear systems: for any "target" energy dissipation rate NN, there is a unique energy function MM that achieves it. This makes stability verification for linear systems completely algorithmic.

The Lyapunov ATM+MA=NA^TM + MA = -N is a linear matrix equation -- given AA and NN, solving for MM is a system of linear equations. In MATLAB, the command lyap(A', N) solves this directly. This theorem converts a nonlinear eigenvalue problem (checking eigenvalue locations) into a linear algebra problem (solving a matrix and checking positive definiteness).


Non-Linear Systems and Linearization

TheoremStability via Linearization

Let f:RnRnf : \mathbb{R}^n \to \mathbb{R}^n be continuously differentiable. Consider x=f(x)x' = f(x) and let f(x)=0f(x^*) = 0 for some xRnx^* \in \mathbb{R}^n. Let A:=Df(x)A := Df(x^*) be the Jacobian of ff at xx^* (that is, with f(x)=[f1(x)fn(x)]Tf(x) = \begin{bmatrix} f^1(x) \cdots f^n(x) \end{bmatrix}^T the linearization with A(i,j)=fixj(x)A(i,j) = \frac{\partial f^i}{\partial x^j}(x^*)). Let λ1,,λn\lambda_1, \ldots, \lambda_n be the eigenvalues of AA. If Re{λi}<0\text{Re}\{\lambda_i\} < 0 for i=1,,ni = 1, \cdots, n, then xx^* is locally asymptotically stable.

Remark.

Intuition: This theorem says that near an equilibrium, a nonlinear system behaves like its linearization. If the linearized system (the Jacobian) is stable, then the nonlinear system is locally stable too -- the higher-order nonlinear terms are too small near the equilibrium to overcome the stabilizing effect of the linear part. This is why linearization is such a workhorse in engineering: you can design controllers using linear theory and trust that they will work locally around the operating point. The caveat is that if the linearization has eigenvalues on the imaginary axis, the nonlinear terms decide stability and the linearization is inconclusive.

The above shows that linearization can be a very effective method. However, when the linearization leads to a matrix with an eigenvalue having a zero real part, the analysis (above based on linearization) is inconclusive and further analysis would be required. To make this observation explicit, consider two systems

x=x5x' = -x^5

x=x5x' = x^5

which have the same linearization around 0. By a Lyapunov stability argument with taking V(x)=x2V(x) = x^2, the first system can be shown to be locally and globally stable, whereas the second one is not (which can be verified by solving the directly: show that x(t)x(t) blows up in finite time!).


Discrete-time Setup

The stability results presented for continuous-time linear systems have essentially identical generalizations for the discrete-time setup.

TheoremDiscrete-Time Global Asymptotic Stability

For a linear difference xk+1=Axkx_{k+1} = Ax_k, the solution is locally and globally asymptotically stable if and only if

maxλi{λi}<1,\max_{\lambda_i}\{|\lambda_i|\} < 1,

where λi\lambda_i denotes the iith eigenvalue of AA. That is, all eigenvalues must be strictly inside the unit disk.

TheoremDiscrete-Time Local Stability

For a linear difference xk+1=Axkx_{k+1} = Ax_k, the system is locally stable if and only if:

(i) maxλi{λi}1\max_{\lambda_i}\{|\lambda_i|\} \leq 1,

(ii) if λi=1|\lambda_i| = 1 for some λi\lambda_i, the algebraic multiplicity of this eigenvalue should be the same as the geometric multiplicity (i.e., the Jordan form corresponding to an eigenvalue on the unit circle is strictly diagonal).

Remark.

Intuition: In discrete time, the unit circle plays the role that the imaginary axis plays in continuous time. Eigenvalues inside the unit disk correspond to decaying modes (λk0|\lambda|^k \to 0), eigenvalues outside to growing modes, and eigenvalues on the unit circle to sustained oscillations. The Jordan block condition for local stability prevents polynomial growth: (nk)λnk\binom{n}{k}\lambda^{n-k} grows polynomially in nn when λ=1|\lambda| = 1 and k1k \geq 1.

In this case, we require the eigenvalues to be strictly inside the unit disk for asymptotic stability (local and global); and for local stability we additionally have the relaxation that the Jordan form corresponding to an eigenvalue on the unit circle is to be strictly diagonal: Any Jordan form block JJ of size N×NN \times N, with eigenvalue λi\lambda_i, can be written as

λiI+E\lambda_i I + E

where EE is a matrix which has all terms zero, except the super-diagonal (the points right above the diagonal), at which points the value is 1. The second term EE is such that EN=0E^N = 0. Finally, we use the power expansion and using the fact that any matrix commutes with the identity matrix:

(λiI+E)n=k=0n(nk)λinIEnk.(\lambda_i I + E)^n = \sum_{k=0}^{n}\binom{n}{k}\lambda_i^n IE^{n-k}.

Since EN=0E^N = 0, we have

(λiI+E)n=k=0N1(nk)λinkIEk(\lambda_i I + E)^n = \sum_{k=0}^{N-1}\binom{n}{k}\lambda_i^{n-k}IE^k

One can have discrete-time generalizations of Lyapunov functions.

TheoremDiscrete Lyapunov Equation

Consider

xk+1=Axk.x_{k+1} = Ax_k.

All eigenvalues of AA have magnitudes strictly less than 1 if and only if for any given positive definite matrix NN or for N=PTPN = P^TP where PP is any given m×nm \times n matrix with m<nm < n, the discrete Lyapunov equation

MATMA=NM - A^TMA = N

has a unique solution which is also positive definite.

Remark.

Intuition: This is the discrete-time counterpart of the continuous Lyapunov equation. Instead of checking that eigenvalues are in the left half-plane (continuous-time), you check that they are strictly inside the unit circle (discrete-time). The Lyapunov function V(xk)=xkTMxkV(x_k) = x_k^T M x_k decreases at each step: V(xk+1)V(xk)=xkTNxk<0V(x_{k+1}) - V(x_k) = -x_k^T N x_k < 0. Think of it as verifying that the system's "energy" drops with every discrete time step, guaranteeing that the state spirals inward to the origin.

The solution in the theorem statement is M=kZ+(AT)kNAkM = \sum_{k \in \mathbb{Z}_+}(A^T)^kNA^k.


Exercises

ExampleDerivative of Matrix Exponential

Let AA be a square matrix. Show that ddteAt=AeAt\frac{d}{dt}e^{At} = Ae^{At}.

Solution. Let t,ht, h be scalars. Since AtAt and AhAh commute so that (At)(Ah)=(Ah)(At)(At)(Ah) = (Ah)(At) we have that eA(t+h)=eAteAhe^{A(t+h)} = e^{At}e^{Ah}. Then,

ddteAt=limh0eA(t+h)eAth\frac{d}{dt}e^{At} = \lim_{h \to 0}\frac{e^{A(t+h)} - e^{At}}{h}

=eAtlimh0eAhIh= e^{At}\lim_{h \to 0}\frac{e^{Ah} - I}{h}

=eAtlimh0(k=1Akhkh(k!))= e^{At}\lim_{h \to 0}\left(\sum_{k=1}^{\infty}\frac{A^kh^k}{h(k!)}\right)

=eAtlimh0(A+hk=2Akhk2k!)= e^{At}\lim_{h \to 0}\left(A + h\sum_{k=2}^{\infty}\frac{A^kh^{k-2}}{k!}\right)

=AeAt= Ae^{At}

The final line follows from the fact that the sum converges to zero as h0h \to 0: let A~(i,j)=A(i,j)\tilde{A}(i,j) = |A(i,j)| be the matrix consisting of the absolute value of the entries of AA, then for each entry (i,j)(i,j) of the matrix

(hk=2Akhk2k!)(i,j)h(A~2k=2A~k2hk2(k2)!)(i,j)\left|\left(h\sum_{k=2}^{\infty}\frac{A^kh^{k-2}}{k!}\right)(i,j)\right| \leq |h|\left(\tilde{A}^2\sum_{k=2}^{\infty}\frac{\tilde{A}^{k-2}|h|^{k-2}}{(k-2)!}\right)(i,j)

h(A~2k=0A~kϵkk!)(i,j)\leq |h|\left(\tilde{A}^2\sum_{k=0}^{\infty}\frac{\tilde{A}^k|\epsilon|^k}{k!}\right)(i,j)

=hA~2eA~ϵ(i,j)= |h|\tilde{A}^2 e^{\tilde{A}\epsilon}(i,j)

=limh0hA~2eA~ϵ(i,j)=0= \lim_{h \to 0} |h|\tilde{A}^2e^{\tilde{A}\epsilon}(i,j) = 0

The result follows.

ExampleExponential of Commuting Matrices

Show that for square matrices AA and BB, which commute, that is AB=BAAB = BA, it follows that

e(A+B)=eAeB.e^{(A+B)} = e^Ae^B.

Solution. Recall that

eA=limTk=0TAkk!=k=0Akk!,e^A = \lim_{T\to\infty}\sum_{k=0}^{T}\frac{A^k}{k!} = \sum_{k=0}^{\infty}\frac{A^k}{k!},

with the definition that A0=IA^0 = I. It follows that

eAeB=(limTk=0TAkk!)(limTl=0TBll!)e^Ae^B = (\lim_{T\to\infty}\sum_{k=0}^{T}\frac{A^k}{k!})(\lim_{T\to\infty}\sum_{l=0}^{T}\frac{B^l}{l!})

=k=0l=0Akk!Bll!=k=0u=k1k!(uk)!AkBuk= \sum_{k=0}^{\infty}\sum_{l=0}^{\infty}\frac{A^k}{k!}\frac{B^l}{l!} = \sum_{k=0}^{\infty}\sum_{u=k}^{\infty}\frac{1}{k!(u-k)!}A^kB^{u-k}

=u=0k=0u1k!(uk)!AkBuk= \sum_{u=0}^{\infty}\sum_{k=0}^{u}\frac{1}{k!(u-k)!}A^kB^{u-k}

=u=01u!k=0uu!k!(uk)!AkBuk=u=01u!k=0u(uk)AkBuk= \sum_{u=0}^{\infty}\frac{1}{u!}\sum_{k=0}^{u}\frac{u!}{k!(u-k)!}A^kB^{u-k} = \sum_{u=0}^{\infty}\frac{1}{u!}\sum_{k=0}^{u}\binom{u}{k}A^kB^{u-k}

=u=01u!(A+B)u= \sum_{u=0}^{\infty}\frac{1}{u!}(A+B)^u

=e(A+B)= e^{(A+B)}

In the above, the substitution u=k+lu = k + l was used, the re-indexing of the double sum follows from re-expressing the summation, and the binomial theorem step uses the fact that AB=BAAB = BA to establish

(A+B)k=m=0k(km)AmBkm.(A+B)^k = \sum_{m=0}^{k}\binom{k}{m}A^mB^{k-m}.

This last statement is proved by induction. Clearly for k=1k=1, (A+B)1=(10)B+(11)A(A+B)^1 = \binom{1}{0}B + \binom{1}{1}A. Suppose this is true for kk. Then for k+1k+1:

(A+B)k(A+B)=m=0k(km)Am+1Bkm+m=0k(km)AmBk+1m.(A+B)^k(A+B) = \sum_{m=0}^{k}\binom{k}{m}A^{m+1}B^{k-m} + \sum_{m=0}^{k}\binom{k}{m}A^mB^{k+1-m}.

Let us separate out the terms involving ApBk+1pA^pB^{k+1-p} for 0pk+10 \leq p \leq k+1. It turns out that we obtain:

p=0k+1ApBk+1p((kp)+(kp1))\sum_{p=0}^{k+1}A^pB^{k+1-p}\left(\binom{k}{p} + \binom{k}{p-1}\right)

Now,

(kp)+(kp1)=k!p!(kp)!+k!(p1)!(k+1p)!\binom{k}{p} + \binom{k}{p-1} = \frac{k!}{p!(k-p)!} + \frac{k!}{(p-1)!(k+1-p)!}

=k!(p1)!(kp)!(1p+1(k+1p))=k!(p1)!(kp)!(k+1p(k+1p))= \frac{k!}{(p-1)!(k-p)!}\left(\frac{1}{p} + \frac{1}{(k+1-p)}\right) = \frac{k!}{(p-1)!(k-p)!}\left(\frac{k+1}{p(k+1-p)}\right)

=(k+1)!(p)!(k+1p)!=(k+1p)= \frac{(k+1)!}{(p)!(k+1-p)!} = \binom{k+1}{p}

This completes the proof.

ExampleStability of x' = -x^7

Let x(t)x(t) satisfy

dxdt=x7\frac{dx}{dt} = -x^7

Is x(t)x(t) (locally) asymptotically stable?

Using V(x)=x2V(x) = x^2, we get V˙=2x(x7)=2x8<0\dot{V} = 2x(-x^7) = -2x^8 < 0 for x0x \neq 0. By Theorem(b), the system is locally asymptotically stable. Since V(x)V(x) \to \infty as x|x| \to \infty, it is also globally asymptotically stable by part (c).

ExampleSecond-Order System Stability

Consider x+x+x=0x'' + x' + x = 0. Is this system asymptotically stable?

Hint: Convert this into a system of first-order differential equations, via x1=xx_1 = x and x2=x1x_2 = x_1', x2=x1x2x_2' = -x_1 - x_2. Then apply V(x1,x2)=x12+x1x2+x22V(x_1, x_2) = x_1^2 + x_1x_2 + x_2^2 as a candidate Lyapunov function.

Solution. Setting x1=xx_1 = x, x2=xx_2 = x', the system becomes:

x1=x2,x2=x1x2x_1' = x_2, \qquad x_2' = -x_1 - x_2

This is dxdt=Ax\frac{dx}{dt} = Ax with A=[0111]A = \begin{bmatrix} 0 & 1 \\ -1 & -1 \end{bmatrix}.

Consider V(x1,x2)=x12+x1x2+x22V(x_1, x_2) = x_1^2 + x_1 x_2 + x_2^2. First, note that VV is positive definite: writing V=12x12+12(x1+x2)2+12x22>0V = \frac{1}{2}x_1^2 + \frac{1}{2}(x_1 + x_2)^2 + \frac{1}{2}x_2^2 > 0 for (x1,x2)(0,0)(x_1, x_2) \neq (0,0).

Computing V˙\dot{V}:

V˙=2x1x1+x1x2+x1x2+2x2x2\dot{V} = 2x_1 x_1' + x_1' x_2 + x_1 x_2' + 2x_2 x_2'

=2x1x2+x22+x1(x1x2)+2x2(x1x2)= 2x_1 x_2 + x_2^2 + x_1(-x_1 - x_2) + 2x_2(-x_1 - x_2)

=2x1x2+x22x12x1x22x1x22x22= 2x_1 x_2 + x_2^2 - x_1^2 - x_1 x_2 - 2x_1 x_2 - 2x_2^2

=x12x1x2x22=V(x1,x2)<0= -x_1^2 - x_1 x_2 - x_2^2 = -V(x_1, x_2) < 0

for (x1,x2)(0,0)(x_1, x_2) \neq (0,0). Since VV \to \infty as (x1,x2)\|(x_1, x_2)\| \to \infty, by Theorem(c), the system is globally asymptotically stable.

Alternatively, the eigenvalues of AA are λ=1±142=1±i32\lambda = \frac{-1 \pm \sqrt{1-4}}{2} = \frac{-1 \pm i\sqrt{3}}{2}, which both have negative real part 12-\frac{1}{2}. By Theorem, the system is globally asymptotically stable.

ExampleA Further Lyapunov Stability Theorem and Barbalat's Lemma

Prove the following theorem:

TheoremLyapunov-Barbalat Stability Theorem

Consider

x=f(x),x' = f(x),

where f:RnRnf : \mathbb{R}^n \to \mathbb{R}^n is continuous and f(0)=0f(0) = 0. Let V:RnR+V : \mathbb{R}^n \to \mathbb{R}_+ be continuously differentiable. Suppose there exists a continuous function W:RnR+W : \mathbb{R}^n \to \mathbb{R}_+ such that

ddtV(x(t))W(x(t))0\frac{d}{dt}V(x(t)) \leq -W(x(t)) \leq 0

Then, provided x(t)x(t) remains bounded, W(x(t))0W(x(t)) \to 0.

Hint: Write

V(x(t))=V(x0)+0tdV(x(s))dsdsV(x0)0tW(x(s))dsV(x(t)) = V(x_0) + \int_0^t \frac{dV(x(s))}{ds} ds \leq V(x_0) - \int_0^t W(x(s)) ds

and conclude that 0tW(x(s))dsV(x0)\int_0^t W(x(s)) ds \leq V(x_0) for all t0t \geq 0 and by the non-negativity of WW, we have that 0W(x(s))dsV(x0)\int_0^{\infty} W(x(s)) ds \leq V(x_0). From here, we want to establish that W(x(t))0W(x(t)) \to 0, provided that (by hypothesis) x(t)x(t) remains bounded. Complete the proof.

Prove and use Barbalat's lemma: Let f:KR+f : K \to \mathbb{R}_+ be uniformly continuous over KK. Then, if x(t)x(t) remains in KK and if 0f(x(s))ds\int_0^{\infty} f(x(s)) ds is finite, then f(x(t))0f(x(t)) \to 0.

Note: The above result also implies an important stability theorem known as LaSalle's invariance principle.

ExampleApplication to Formation Control, Consensus Algorithms or Opinion Dynamics

Consider a network of NN agents which are connected over a graph. We say that AA is an adjacency graph if A(i,j)=1A(i,j) = 1 if Agents ii and Agent jj are connected and A(i,j)=0A(i,j) = 0 otherwise. For each agent i{1,,N}i \in \{1, \cdots, N\} define di=j=1NA(i,j)d_i = \sum_{j=1}^{N}A(i,j) to be the degree of the agent. Now define L=ADL = A - D where DD is a diagonal matrix with D(i,i)=diD(i,i) = d_i. Such a matrix LL is called a Laplacian.

Now, suppose that the agents update their states by the following equation:

dxdt=Lx\frac{dx}{dt} = -Lx

Observe that LL is a positive semi-definite matrix and if the graph is connected the only eigenvector corresponding to the zero eigenvalue is [111]T\begin{bmatrix} 1 & 1 & \cdots & 1 \end{bmatrix}^T. This you can see by noting that xTLx=12A(i,j)(xixj)2x^TLx = \frac{1}{2}\sum A(i,j)(x_i - x_j)^2.

In this case, define the following Lyapunov function:

V(x)=12xTxV(x) = \frac{1}{2}x^Tx

Then,

ddt(V(x(t)))=12((dxdt)Tx(t)+xT(t)dxdt)=xT(t)Lx(t)0\frac{d}{dt}(V(x(t))) = \frac{1}{2}((\frac{dx}{dt})^Tx(t) + x^T(t)\frac{dx}{dt}) = -x^T(t)Lx(t) \leq 0

The above ensures that x(t)x(t) remains bounded. Now, invoke Barbalat's Lemma to conclude that xT(t)Lx(t)0x^T(t)Lx(t) \to 0. Since the only eigenvalue corresponding to Lx(t)=0Lx(t) = 0 is x(t)=[111]Tx(t) = \begin{bmatrix} 1 & 1 & \cdots & 1 \end{bmatrix}^T and throughout the updates the sum 1N(x1(t)+x2(t)+xN(t))\frac{1}{N}(x_1(t) + x_2(t) + \cdots x_N(t)) is a constant (as the sum does not change), we have that x(t)1N(x1(0)+x2(0)++xN(0))x(t) \to \frac{1}{N}(x_1(0) + x_2(0) + \cdots + x_N(0)).

ExampleApplication to Adaptive Control

Consider

dxdt=ax+u\frac{dx}{dt} = ax + u

Suppose that our goal is to have limtx(t)=0\lim_{t\to\infty}x(t) = 0. We know that if we select u(t)=(a+κ)xu(t) = -(a + \kappa)x for any κ>0\kappa > 0, the system is stable. In particular, let κ=1\kappa = 1.

In many engineering applications, the value of aa is unknown.

Adaptive control theory is the sub-field of control theory studying such problems. The goal is to allow the controller to learn the system to be able to achieve the desired goal.

Suppose that the controller runs the following policy:

u(t)=(a^(t)+1)x(t),u(t) = -(\hat{a}(t) + 1)x(t),

which leads to x=(aa^(t)1)x(t)x' = (a - \hat{a}(t) - 1)x(t), where a^(t)\hat{a}(t) is an estimate of aa. Suppose that we take

a^(t)=x2(t)\hat{a}'(t) = x^2(t)

In this case, consider the Lyapunov function:

V(x,a^)=x2+(aa^)2V(x, \hat{a}) = x^2 + (a - \hat{a})^2

We compute V˙=2xx+2(aa^)(a^)=2x(aa^1)x2(aa^)x2=2x20\dot{V} = 2x \cdot x' + 2(a - \hat{a})(-\hat{a}') = 2x(a - \hat{a} - 1)x - 2(a-\hat{a})x^2 = -2x^2 \leq 0.

Since VV is non-increasing, x(t)x(t) and a^(t)\hat{a}(t) remain bounded. By Barbalat's Lemma (Theorem), x(t)0x(t) \to 0.