Home/Chapter 9

Realizability and State Space Representation

Realizations of transfer functions: controllable, observable, and modal canonical forms. Zero-state equivalence, algebraic equivalence, and discretization of continuous-time systems.

Realizations: Controllable, Observable and Modal Forms

So far in the course, we considered the input-output approach to study control systems, which in the convolution (linear time-invariant) system setup, resulted in frequency domain methods through a transfer function analysis (such as in arriving at the root locus / Nyquist stability / Bode plot methods). These are often referred to as classical control design methods.

In this chapter, we will introduce state-space based methods. Control design based on state-space methods is called modern control design.

The notion of a state. Suppose that, given tRt \in \mathbb{R} (or Z\mathbb{Z}), we wish to compute the output of a system at tt0t \geq t_0. In a general causal system, we may need to use all the past applied input terms u(s)u(s), sts \leq t and/or all the past output values y(s)y(s), s<ts < t to compute the output at tt. The state of a system summarizes all the past relevant data that is sufficient to compute the future paths in the sense that if the state at t0t_0, x(t0)x(t_0) is given, then to compute y(t)y(t), one would only need to use {u(s)  s[t0,t]}\{u(s) \; s \in [t_0,t]\}, {y(s)  s[t0,t)}\{y(s) \; s \in [t_0,t)\} and x(t0)x(t_0). In particular, the past {y(s),us;  s<t0}\{y(s), u_s; \; s < t_0\} would not be needed.

Some systems admit a finite-dimensional state representation, some do not.

Consider a linear system in state-space form:

dxdt=Ax(t)+Bu(t),y(t)=Cx(t)+Du(t)\frac{dx}{dt} = Ax(t) + Bu(t), \qquad y(t) = Cx(t) + Du(t)

We will say that such a system is given by the 4-tuple: (A,B,C,D)(A, B, C, D).

In the above xx serves as the state variable for the system.

We know that the solution to this system is given with

x(t)=eAtx(0)+0teA(ts)Bu(s)dsx(t) = e^{At}x(0) + \int_0^t e^{A(t-s)}Bu(s)ds

and

y(t)=CeAtx(0)+C0teA(ts)Bu(s)ds+Du(t)y(t) = Ce^{At}x(0) + C\int_0^t e^{A(t-s)}Bu(s)ds + Du(t)

Taking the (one-sided) Laplace transform, we obtain for ss in the ROC:

Y+(s)=C(sIA)1(x(0)+BU+(s))+DU+(s)Y_+(s) = C(sI - A)^{-1}(x(0) + BU_+(s)) + DU_+(s)

Assuming x(0)=0x(0) = 0, we have the following as the transfer function:

Y+(s)=(C(sIA)1B+D)U(s)Y_+(s) = (C(sI - A)^{-1}B + D)U(s)

A transfer function H(s)H(s) is state-space realizable if there exists finite dimensional (A,B,C,D)(A, B, C, D) so that we can write for sCs \in \mathbb{C}, whenever well-defined:

H(s)=C(sIA)1B+DH(s) = C(sI - A)^{-1}B + D

TheoremRealizability

A transfer function of a linear time-invariant system H(s)H(s) is realizable if and only if it is a proper rational fraction (that is, H(s)=P(s)Q(s)H(s) = \frac{P(s)}{Q(s)} where both the numerator PP and the denominator QQ are polynomials, and with degree of PP less than or equal to the degree of QQ).

Remark.

Intuition: This theorem draws a sharp line between systems you can build with a finite number of integrators, amplifiers, and summing junctions (realizable) and those you cannot. A proper rational transfer function means the system does not require taking pure derivatives of the input (which would amplify noise infinitely). Systems like pure time delays (esTe^{-sT}) are not rational and therefore require infinite-dimensional state spaces -- you would need to "remember" the entire input history over the delay interval.

Controllable canonical realization

Consider a continuous-time system given by:

k=0Nakdkdtky(t)=m=0Nbmdmdtmu(t),\sum_{k=0}^{N} a_k \frac{d^k}{dt^k}y(t) = \sum_{m=0}^{N} b_m \frac{d^m}{dt^m}u(t),

with aN=1a_N = 1. Taking the Laplace transform, we know that the transfer function writes as

H(s)=m=0Nbmsmk=0NakskH(s) = \frac{\sum_{m=0}^{N} b_m s^m}{\sum_{k=0}^{N} a_k s^k}

Suppose that the system is strictly proper. Such a system can be realized with the form:

ddtx(t)=Ax(t)+Bu(t),y(t)=Cx(t)\frac{d}{dt}x(t) = Ax(t) + Bu(t), \qquad y(t) = Cx(t)

with

Ac=[010000101a0a1a2aN1]A_c = \begin{bmatrix} 0 & 1 & 0 & \cdots & 0 \\ 0 & 0 & 1 & \cdots & 0 \\ \vdots & \vdots & & \ddots & 1 \\ -a_0 & -a_1 & -a_2 & \cdots & -a_{N-1} \end{bmatrix}

Bc=[0001]B_c = \begin{bmatrix} 0 \\ 0 \\ \vdots \\ 0 \\ 1 \end{bmatrix}

Cc=[b0b1bN1]C_c = \begin{bmatrix} b_0 & b_1 & \cdots & b_{N-1} \end{bmatrix}

If the system is proper, but not strictly proper, then we will also have Dc=dD_c = d, where dd is the remainder term in the partial fraction expansion,

H(s)=d+i=1K(k=1miAik(spi)k)H(s) = d + \sum_{i=1}^{K}\left(\sum_{k=1}^{m_i}\frac{A_{ik}}{(s - p_i)^k}\right)

where pip_i are the roots of sn+k=0N1aksks^n + \sum_{k=0}^{N-1}a_k s^k and mim_i is the multiplicity of pip_i.

This is called the controllable canonical form. The key insight is that the characteristic polynomial of AcA_c is exactly sN+aN1sN1++a1s+a0s^N + a_{N-1}s^{N-1} + \cdots + a_1 s + a_0, matching the denominator of the transfer function. The companion structure of AcA_c ensures that the single input uu can influence all state components through successive integrations.

Observable canonical realization

Consider the same system. This system can also be realized as

ddtx(t)=Ax(t)+Bu(t),y(t)=Cx(t)\frac{d}{dt}x(t) = Ax(t) + Bu(t), \qquad y(t) = Cx(t)

with

A=[aN1100aN2010a1001a0000]A = \begin{bmatrix} -a_{N-1} & 1 & 0 & \cdots & 0 \\ -a_{N-2} & 0 & 1 & \cdots & 0 \\ \vdots & \vdots & & \ddots & \vdots \\ -a_1 & 0 & 0 & \cdots & 1 \\ -a_0 & 0 & 0 & \cdots & 0 \end{bmatrix}

B=[bN1bN2b0]B = \begin{bmatrix} b_{N-1} \\ b_{N-2} \\ \vdots \\ b_0 \end{bmatrix}

C=[100]C = \begin{bmatrix} 1 & 0 & \cdots & 0 \end{bmatrix}

If we reverse the order of the coordinates of xx, we arrive at the standard observable canonical form:

ddtx(t)=Ax(t)+Bu(t),y(t)=Cx(t)\frac{d}{dt}x(t) = Ax(t) + Bu(t), \qquad y(t) = Cx(t)

with

Ao=[000a0100a1000aN2000aN1]A_o = \begin{bmatrix} 0 & 0 & 0 & \cdots & -a_0 \\ 1 & 0 & 0 & \cdots & -a_1 \\ \vdots & \ddots & & & \vdots \\ 0 & 0 & 0 & \cdots & -a_{N-2} \\ 0 & 0 & 0 & \cdots & -a_{N-1} \end{bmatrix}

Bo=[b0b1bN1]B_o = \begin{bmatrix} b_0 \\ b_1 \\ \vdots \\ b_{N-1} \end{bmatrix}

Co=[001]C_o = \begin{bmatrix} 0 & 0 & \cdots & 1 \end{bmatrix}

Observe that Ao=AcTA_o = A_c^T, Bo=CcTB_o = C_c^T, Co=BcTC_o = B_c^T. This is the standard observable canonical form.

The duality Ao=AcTA_o = A_c^T between controllable and observable forms is a fundamental structural property. It reflects the fact that observability and controllability are dual concepts -- a connection that will be made precise in Chapter 12.

Exercise. Show that the transfer functions under the controllable and observable realization canonical forms are equivalent directly by comparing Cc(sIAc)1BcC_c(sI - A_c)^{-1}B_c and Co(sIAo)1BoC_o(sI - A_o)^{-1}B_o.

Modal realization

Consider a partial fraction expansion with only simple poles:

H(s)=m=0Nbmsmk=0Nakss=d+iNkispiH(s) = \frac{\sum_{m=0}^{N} b_m s^m}{\sum_{k=0}^{N} a_k s^s} = d + \sum_{i}^{N} \frac{k_i}{s - p_i}

In this case, we can realize the system as the sum of decoupled modes:

ddtx(t)=Ax(t)+Bu(t),y(t)=Cx(t)\frac{d}{dt}x(t) = Ax(t) + Bu(t), \qquad y(t) = Cx(t)

with

A=[p10000p20000pN]A = \begin{bmatrix} p_1 & 0 & 0 & \cdots & 0 \\ 0 & p_2 & 0 & \cdots & \vdots \\ \vdots & \vdots & & \ddots & \vdots \\ 0 & 0 & 0 & \cdots & p_N \end{bmatrix}

B=[k1k2kN]B = \begin{bmatrix} k_1 \\ k_2 \\ \vdots \\ k_N \end{bmatrix}

C=[111]C = \begin{bmatrix} 1 & 1 & \cdots & 1 \end{bmatrix}

If in the partial fraction expansion is more general (with repeated poles), then the corresponding structure can be realized also: this will lead to a Jordan form for the matrix AA, since, e.g.,

1(spi)2=1spi1spi\frac{1}{(s - p_i)^2} = \frac{1}{s - p_i}\frac{1}{s - p_i}

will define a serial connection of two modal blocks; the first one with x2=pix2+ux_2' = p_i x_2 + u and the second one x1=pix1+x2x_1' = p_i x_1 + x_2.

The modal realization makes the system poles (eigenvalues of AA) directly visible as diagonal entries. Each mode evolves independently, making it easy to see which modes are fast, slow, stable, or unstable.

Discrete-time setup. The above also apply to the discrete-time setup. For example, a discrete-time system of the form

k=0Naky(nk)=m=1Nbmu(nm)\sum_{k=0}^{N} a_k y(n-k) = \sum_{m=1}^{N} b_m u(n-m)

with a0=1a_0 = 1, can be written in the controllable canonical form

x(n+1)=Ax(n)+Bu(n),yt=Cxtx(n+1) = Ax(n) + Bu(n), \qquad y_t = Cx_t

where xN(n)=y(n),xN1(n)=y(n1),    ,x1(n)=y(n(N1))x_N(n) = y(n), x_{N-1}(n) = y(n-1), \;\cdots\;, x_1(n) = y(n-(N-1))

A=[010000101aNaN1aN2a1]A = \begin{bmatrix} 0 & 1 & 0 & \cdots & 0 \\ 0 & 0 & 1 & \cdots & 0 \\ \vdots & \vdots & & \ddots & 1 \\ -a_N & -a_{N-1} & -a_{N-2} & \cdots & -a_1 \end{bmatrix}

B=[0001]B = \begin{bmatrix} 0 \\ 0 \\ \vdots \\ 0 \\ 1 \end{bmatrix}

C=[bNbN1b1]C = \begin{bmatrix} b_N & b_{N-1} & \cdots & b_1 \end{bmatrix}

Observable and modal canonical forms follow similarly.


Zero-State Equivalence and Algebraic Equivalence

We say that two systems (A,B,C,D)(A, B, C, D) and (A~,B~,C~,D~)(\tilde{A}, \tilde{B}, \tilde{C}, \tilde{D}) are zero-state equivalent if the induced transfer functions are equal, that is

C(sIA)1B+D=C~(sIA~)1B~+D~C(sI - A)^{-1}B + D = \tilde{C}(sI - \tilde{A})^{-1}\tilde{B} + \tilde{D}

TheoremZero-State Equivalence Condition

Two linear time-invariant state-space models (A,B,C,D)(A, B, C, D) and (A~,B~,C~,D~)(\tilde{A}, \tilde{B}, \tilde{C}, \tilde{D}) are zero-state equivalent if and only if D=D~D = \tilde{D} and CAmB=C~A~mB~CA^mB = \tilde{C}\tilde{A}^m\tilde{B} for all mZ+m \in \mathbb{Z}_+.

Remark.

Intuition: Two state-space models are zero-state equivalent when an external observer feeding inputs and measuring outputs cannot distinguish between them. The Markov parameters CAmBCA^mB are exactly the impulse response coefficients -- they describe how the output responds at time step mm after a single impulse input. If these coefficients match, the two systems produce identical outputs for every possible input, even though their internal states may live in completely different coordinate systems or even have different dimensions.

The quantities CAmBCA^mB are called the Markov parameters of the system. Two systems are zero-state equivalent precisely when they share the same Markov parameters and the same feedthrough term DD.

There is an alternative notion, called algebraic equivalence: Let PP be invertible and let us define a transformation through x~=Px\tilde{x} = Px. Then, we can write the state-space model as

dx~dt=A~x~(t)+B~u(t),y~(t)=C~x~(t)+D~u(t),\frac{d\tilde{x}}{dt} = \tilde{A}\tilde{x}(t) + \tilde{B}u(t), \qquad \tilde{y}(t) = \tilde{C}\tilde{x}(t) + \tilde{D}u(t),

with A~=PAP1\tilde{A} = PAP^{-1}, B~=PB\tilde{B} = PB, C~=CP1\tilde{C} = CP^{-1}, D~=D\tilde{D} = D. In this case, we say that (A,B,C,D)(A, B, C, D) and (A~,B~,C~,D~)(\tilde{A}, \tilde{B}, \tilde{C}, \tilde{D}) are algebraically equivalent.

TheoremAlgebraic Implies Zero-State Equivalence

Algebraic equivalence implies zero-state equivalence but not vice versa.

Remark.

Intuition: Algebraic equivalence means two state-space models are really the same system viewed through different "lenses" -- a change of coordinates x~=Px\tilde{x} = Px on the state vector. This is analogous to describing the same physical object using different coordinate systems (Cartesian vs. polar). Such a transformation always preserves the input-output behavior (zero-state equivalence), but the converse fails because a system with redundant or unobservable states can have the same transfer function as a simpler system that cannot be obtained by any coordinate change.

Algebraic equivalence is a stronger condition: it says the two systems are really the same system viewed in different coordinates. Zero-state equivalence only says the input-output maps agree, but the internal structure may differ.


Discretization

Consider

ddtx(t)=Ax(t)+Bu(t),y(t)=Cx(t)\frac{d}{dt}x(t) = Ax(t) + Bu(t), \qquad y(t) = Cx(t)

Suppose we apply piece-wise constant control actions uu which are varied only at the discrete time instances given with {t:t=kT,kZ+}\{t : t = kT, k \in \mathbb{Z}_+\} so that u(t)=u(kT)u(t) = u(kT) for t[kT,(k+1)T)t \in [kT, (k+1)T).

We write

x((k+1)T)=eATx(kT)+(kT(k+1)TeA((k+1)Ts)Bu(s)ds)x((k+1)T) = e^{AT}x(kT) + \left(\int_{kT}^{(k+1)T} e^{A((k+1)T - s)}Bu(s)ds\right)

Writing τ=(k+1)Ts\tau = (k+1)T - s with dτ=dsd\tau = -ds, we arrive at

x((k+1)T)=eATx(kT)+(0TeAτBdτ)u(kT)x((k+1)T) = e^{AT}x(kT) + \left(\int_0^{T} e^{A\tau}Bd\tau\right)u(kT)

With xk:=x(kT)x_k := x(kT) and uk:=u(kT)u_k := u(kT), we arrive at

xk+1=Adxk+Bdukx_{k+1} = A_d x_k + B_d u_k

where

Ad=eATA_d = e^{AT}

Bd=0TeAτBdτB_d = \int_0^{T} e^{A\tau}Bd\tau

If AA is invertible, the integration of eAτdτ\int e^{A\tau}d\tau leads to A1(eATI)BA^{-1}(e^{AT} - I)B.

Discretization converts a continuous-time state-space model into an equivalent discrete-time model that is exact at the sampling instants. The key point is that no approximation is involved -- the discrete-time system exactly reproduces the continuous-time state at times t=kTt = kT, provided the input is held constant between samples (zero-order hold).


Exercises

ExampleCanonical Realizations

Algebraically express the realizations of the transfer function

H(s)=s2+2s+1(s+1)(s24)H(s) = \frac{s^2 + 2s + 1}{(s+1)(s^2 - 4)}

in the controllable canonical realization, observable canonical realization, and modal canonical realization forms.

ExampleNon-Rational Transfer Function

Consider the following continuous-time system:

y(t)=a(u(t1)+y(t1)),tRy(t) = a(u(t-1) + y(t-1)), \quad t \in \mathbb{R}

Find the transfer function, from uu to yy, of this system and note that this transfer function is not rational. Such systems with no rational transfer function are sometimes called distributed parameter systems.