Home/Chapter 4

Systems

System properties: linearity, time-invariance, causality. LTI (convolution) systems, BIBO stability, transfer functions, frequency response, Bode plots, feedback systems, and state-space descriptions.

An input-output system is defined by an input signal set U\mathcal{U}, an output set Y\mathcal{Y} and a subset RU×Y\mathcal{R} \subset \mathcal{U} \times \mathcal{Y}, called the rule (relation) of the system. Hence, R\mathcal{R} consists of the input-output pairs in the system. Associated with such an input-output relation R\mathcal{R} is a transformation or map T\mathcal{T} such that y=T(u)y = \mathcal{T}(u), and thus

T:UuT(u)=yY,\mathcal{T} : \mathcal{U} \ni u \mapsto \mathcal{T}(u) = y \in \mathcal{Y},

and thus

R={(u,T(u)),uU}.\mathcal{R} = \{(u, \mathcal{T}(u)),\quad u \in \mathcal{U}\}.

Let T1,T2T_1, T_2 be time-index sets; and U,YU, Y be signal range spaces such that U=UT1,Y=YT2\mathcal{U} = U^{T_1}, \mathcal{Y} = Y^{T_2}, that is:

U={f:T1U},\mathcal{U} = \{f : T_1 \to U\},

Y={f:T2Y}.\mathcal{Y} = \{f : T_2 \to Y\}.

If U\mathcal{U} and Y\mathcal{Y} consist of signals with discrete-time indices, then the system is said to be a discrete-time (DT) system. If the indices are both continuous, then the system is a continuous-time (CT) system. If one of them is discrete and the other continuous, the system is said to be hybrid. Often, we have T=T1=T2T = T_1 = T_2, which will be assumed in the following.


System Properties

DefinitionMemorylessness

Let UU be an input signal range, YY an output signal range, and TT a time index. A system is memoryless if any input-output pair (u,T(u))(u, \mathcal{T}(u)) can be written component-wise as

y(t)=Ψ(t,u(t)),tTy(t) = \Psi(t, u(t)), \quad t \in T

for some fixed map Ψ:T×UY\Psi : T \times U \to Y.

Remark.

Intuition: A memoryless system is one whose output at any time tt depends only on the input at that same instant tt -- it has no "memory" of past or future inputs. A simple resistor (V=IRV = IR) is memoryless, while a capacitor (whose voltage depends on accumulated charge) is not.

DefinitionCausality (Non-anticipativeness)

A system is causal (non-anticipative) if the output at any time tt is not dependent on the input signal values at time s>ts > t. That is, let u1={u1(t),tT}u^1 = \{u^1(t), t \in T\} and u2={u2(t),tT}u^2 = \{u^2(t), t \in T\}. Let (u1,y1)R(u^1, y^1) \in \mathcal{R} and (u2,y2)R(u^2, y^2) \in \mathcal{R}. For any tTt \in T, if it is that u1(s)=u2(s)u^1(s) = u^2(s) for sts \leq t, then for a causal system it must be that y1(t)=y2(t)y^1(t) = y^2(t).

Remark.

Intuition: Causality means the system cannot "look into the future." The output at time tt can depend on the present and past inputs, but never on inputs that have not yet occurred. All physically realizable systems are causal -- you cannot respond to a stimulus before it happens.

ExampleCausal and Memoryless Systems

Let a relation be given by yn=axn+1+xn+bxn1y_n = ax_{n+1} + x_n + bx_{n-1} where nZn \in \mathbb{Z}. Such a system is causal if a=0a = 0; it is memoryless if a=0,b=0a = 0, b = 0.

DefinitionTime-Invariance

A system is time-invariant if for every input-output pair ((u,y)R)((u, y) \in \mathcal{R}), a time-shift in the input leads to the same time-shift in the output; that is,

(σθu,σθy)R,(\sigma^\theta u, \sigma^\theta y) \in \mathcal{R},

where we define a time-shift as follows: With T=ZT = \mathbb{Z} or R\mathbb{R}, let θT\theta \in T. We define σθ:UU\sigma^\theta : \mathcal{U} \to \mathcal{U} with

(σθ(u))t=ut+θ,tT.\left(\sigma^\theta(u)\right)_t = u_{t+\theta}, \quad \forall t \in T.

Remark.

Intuition: Time-invariance means the system's behaviour does not change over time. If you delay the input by θ\theta, the output is simply delayed by the same amount. The laws governing the system are the same today as they will be tomorrow. Note that σθ\sigma^\theta pushes a signal to the left by θ\theta.


Linear Systems

Linear systems have important engineering practice. Many physical systems are locally linear, as we have seen earlier.

An input-output system is linear if U\mathcal{U}, Y\mathcal{Y}, R\mathcal{R} are all linear vector spaces. However, in the context of our course, we will have a more restrictive definition for linearity.

DefinitionLinear System (DT)

A discrete-time (DT) system is linear if the input-output relation can be written as:

y(n)=m=h(n,m)u(m).y(n) = \sum_{m=-\infty}^{\infty} h(n, m) u(m). \qquad \text{}

The function h(n,m)h(n, m) is called the kernel of the system. The value h(n,m)h(n, m) reveals the effect of an input at time mm to the output at time nn.

Remark.

Intuition: A linear system is one where the output is a weighted sum (superposition) of all the input values, with the weights given by the kernel h(n,m)h(n,m). This kernel tells you how much influence the input at time mm has on the output at time nn. The key feature is that doubling the input doubles the output, and the response to a sum of inputs is the sum of responses.

We note here that a precise characterization for linearity (for a system as in ) would require the interpretation of a system as a (bounded) linear operator from one space to another space. One can obtain a Riesz representation theorem type characterization leading to , provided that U\mathcal{U} and Y\mathcal{Y} satisfy certain properties, and the system is continuous and linear. The following discussion makes this explicit.

Let T\mathcal{T} be a linear system mapping l1(Z;R)l_1(\mathbb{Z}; \mathbb{R}) to l1(Z;R)l_1(\mathbb{Z}; \mathbb{R}). Let this system be linear and continuous; then the system can be written so that y=T(u)y = \mathcal{T}(u) where:

y(n)=mZh(n,m)u(m),nZy(n) = \sum_{m \in \mathbb{Z}} h(n, m) u(m), \quad n \in \mathbb{Z}

for some h:Z×ZRh : \mathbb{Z} \times \mathbb{Z} \to \mathbb{R}.

Building on this discussion, one takes the representation above as a definition of a linear system: In our course and in standard terminology in engineering and applied science, we generally say that a discrete-time (DT) system is linear if the input-output relation can be written as .

Remark.

Observe that in the representation argument above, one can generalize the result for any lp(Z;R)l_p(\mathbb{Z}; \mathbb{R}) as the input space with 1p<1 \leq p < \infty; and the output space can be any lq(Z;R)l_q(\mathbb{Z}; \mathbb{R}) space with 1q1 \leq q \leq \infty.

DefinitionLinear System (CT)

Likewise, we define a continuous-time (CT) system to be linear if the input-output relation can be expressed as

y(t)=τ=h(t,τ)u(τ)dτ.y(t) = \int_{\tau=-\infty}^{\infty} h(t, \tau) u(\tau)\,d\tau. \qquad \text{}

Remark.

Intuition: The continuous-time analogue replaces summation with integration. Instead of discrete weights h(n,m)h(n,m), we have a kernel function h(t,τ)h(t,\tau) that describes how the input at continuous time τ\tau influences the output at time tt.


Linear and Time-Invariant (Convolution) Systems

If, in addition to linearity, we wish to have time-invariance, then one can show that

y(n)=k=k(n,m)u(m),y(n) = \sum_{k=-\infty}^{\infty} k(n, m) u(m),

will have to be such that k(n,m)k(n, m) should be dependent only on nmn - m. This follows from the fact that a shift in the input would have to lead to the same shift in the output, implying that k(n,m)=k(n+θ,m+θ)k(n, m) = k(n + \theta, m + \theta) for any θZ\theta \in \mathbb{Z}.

Let us discuss this further. Suppose a linear system described by

y(n)=mk(n,m)u(m),y(n) = \sum_m k(n, m) u(m), \qquad \text{}

is time-invariant. Let, for some θZ\theta \in \mathbb{Z},

v=σθ(u)v = \sigma^{-\theta}(u)

so that v(m)=u(mθ)v(m) = u(m - \theta). Let the signal gg be the output of the system when the input is the discrete-time signal vv. It follows that

g(n)=mZk(n,m)v(m)=mZk(n,m)u(mθ)=mZk(n,m+θ)u(m).g(n) = \sum_{m \in \mathbb{Z}} k(n, m) v(m) = \sum_{m \in \mathbb{Z}} k(n, m) u(m - \theta) = \sum_{m' \in \mathbb{Z}} k(n, m' + \theta) u(m').

By time-invariance, it must be that g=σθ(y)g = \sigma^{-\theta}(y). That is, g(n)=y(nθ)g(n) = y(n - \theta) or g(n+θ)=y(n)g(n + \theta) = y(n). Thus,

g(n+θ)=mZk(n+θ,m+θ)u(m)=y(n).g(n + \theta) = \sum_{m' \in \mathbb{Z}} k(n + \theta, m' + \theta) u(m') = y(n). \qquad \text{}

Since the equivalence in - above has to hold for every input signal, it must be that k(n+θ,m+θ)=k(n,m)k(n + \theta, m + \theta) = k(n, m) for all n,mn, m values, and for all θ\theta values. Therefore k(n,m)k(n, m) should only be a function of the difference nmn - m. Hence, a linear system is time-invariant if and only if the input-output relation can be written as:

y(n)=m=h(nm)u(m)y(n) = \sum_{m=-\infty}^{\infty} h(n - m) u(m)

for some function h:ZRh : \mathbb{Z} \to \mathbb{R}.

DefinitionImpulse Response (DT)

The function hh is called the impulse response of the system since, if u=δ0u = \delta_0, then

y(n)=m=h(nm)δ0(m)=h(n).y(n) = \sum_{m=-\infty}^{\infty} h(n - m) \delta_0(m) = h(n).

Due to this representation, linear time-invariant systems are also called convolution systems.

Remark.

Intuition: The impulse response hh completely characterizes an LTI system. It is the output you get when you "kick" the system with a single unit pulse at time zero. Because of linearity and time-invariance, knowing this single response lets you predict the output for any input via convolution: every input is just a weighted, shifted sum of impulses.

One can show that a convolution system is non-anticipative (causal) if h(n)=0h(n) = 0 for n<0n < 0.

Similar discussions apply to continuous-time systems by replacing the summation with integrals:

y(t)=τ=h(tτ)u(τ)dτ.y(t) = \int_{\tau=-\infty}^{\infty} h(t - \tau) u(\tau)\,d\tau.

Let δ0\delta_0 be the generalized Dirac delta (impulse) function which we view as the limit of an approximate identity sequence (thus defining the Dirac delta distribution). Notably, if hh is continuous, it follows that

limnh(tτ)ψn(τ)dτ=h(t).\lim_{n \to \infty} \int h(t - \tau) \psi_n(\tau)\,d\tau = h(t).

Thus, we have that when u=δ0u = \delta_0,

y(t)=τ=h(tτ)δ0(τ)dτ=h(t).y(t) = \int_{\tau=-\infty}^{\infty} h(t - \tau)\delta_0(\tau)\,d\tau = h(t).

DefinitionImpulse Response (CT)

The function hh is the output of the system when the input is the generalized Dirac delta function. This is why hh is called the impulse response of a convolution system.

Remark.

Intuition: Just as in the discrete-time case, the continuous-time impulse response tells you the system's output when hit with an idealized instantaneous "kick" (δ0\delta_0). Through convolution, the output to any input u(t)u(t) is the integral of all these shifted, scaled impulse responses.

Exercise

Let x(t)RNx(t) \in \mathbb{R}^N and t0t \geq 0 and real-valued. Recall that the solution to the following differential equation:

x(t)=Ax(t)+Bu(t),x'(t) = Ax(t) + Bu(t),

y(t)=Cx(t),y(t) = Cx(t),

with the initial condition x(t0)=x0x(t_0) = x_0 is given by

x(t)=eA(tt0)xt0+τ=t0teA(tτ)Bu(τ)dτ,t0.x(t) = e^{A(t-t_0)}x_{t_0} + \int_{\tau=t_0}^{t} e^{A(t-\tau)}Bu(\tau)\,d\tau, \quad t \geq 0.

(a) Suppose that x(t0)=0x(t_0) = 0 and all eigenvalues of AA have their real parts as negative and u<\|u\|_\infty < \infty. Let t0t_0 \to -\infty. Show that if one is to represent x(t)=(hu)(t)x(t) = (h * u)(t), we have

h(t)=CeAtB1{t0}.h(t) = Ce^{At}B\,1_{\{t \geq 0\}}.

(b) Alternatively, we could skip the condition that the eigenvalues of AA have their real parts as negative, but require that x(0)=0x(0) = 0 and u(t)=0u(t) = 0 for t<0t < 0. Express the solution as a convolution y(t)=(hu)(t)y(t) = (h * u)(t), and find h(t)h(t).

(c) Let y(t)=Cx(t)+Du(t)y(t) = Cx(t) + Du(t). Repeat the above.

Exercise

Let x(n)RNx(n) \in \mathbb{R}^N and nZn \in \mathbb{Z}. Consider a linear system given by

x(n+1)=Ax(n)+Bu(n)x(n+1) = Ax(n) + Bu(n)

y(n)=Cx(n),n0y(n) = Cx(n), \quad n \geq 0

with the initial condition x(n0)=0x(n_0) = 0 for some n0n_0.

(a) Suppose all the eigenvalues of AA are strictly inside the unit disk in the complex plane and u<\|u\|_\infty < \infty. Let n0n_0 \to -\infty. Express the solution y(n)y(n) as a convolution y(n)=(hu)(n)y(n) = (h * u)(n), and find that

h(n)=CAn1B1{n1}.h(n) = CA^{n-1}B\,1_{\{n \geq 1\}}.

(b) Alternatively, we could skip the condition that the eigenvalues of AA are strictly inside the unit disk in the complex plane, but require that n0=0n_0 = 0 so that x(0)=0x(0) = 0 and also u(n)=0u(n) = 0 for n<0n < 0. Express the solution as a convolution y(n)=(hu)(n)y(n) = (h * u)(n), and find h(n)h(n).

(c) Let y(n)=Cx(n)+Du(n)y(n) = Cx(n) + Du(n). Repeat the above.


Bounded-Input-Bounded-Output (BIBO) Stability of Convolution Systems

DefinitionBIBO Stability (DT)

A DT system is BIBO stable if u:=supmZu(m)<\|u\|_\infty := \sup_{m \in \mathbb{Z}} |u(m)| < \infty implies that y:=supmZy(m)<\|y\|_\infty := \sup_{m \in \mathbb{Z}} |y(m)| < \infty.

Remark.

Intuition: BIBO stability asks a simple question: if the input is bounded (never blows up), is the output guaranteed to be bounded too? A system that amplifies bounded inputs into unbounded outputs is unstable and dangerous in practice -- an amplifier that produces infinite voltage from a finite input signal is clearly undesirable.

DefinitionBIBO Stability (CT)

A CT system is BIBO stable if u:=suptRu(t)<\|u\|_\infty := \sup_{t \in \mathbb{R}} |u(t)| < \infty implies that y:=suptRy(t)<\|y\|_\infty := \sup_{t \in \mathbb{R}} |y(t)| < \infty.

Remark.

Intuition: The continuous-time version of BIBO stability is identical in spirit to the discrete-time version: bounded inputs must produce bounded outputs for the system to be considered stable.

TheoremBIBO Stability of Convolution Systems

A convolution system is BIBO stable if and only if

h1<.\|h\|_1 < \infty.

In particular, as a linear map, the convolution system T:l(Z;R)l(Z;R)\mathcal{T} : l_\infty(\mathbb{Z}; \mathbb{R}) \to l_\infty(\mathbb{Z}; \mathbb{R}) satisfies

T=supul(Z;R)T(u)u=h1.\|\mathcal{T}\| = \sup_{u \in l_\infty(\mathbb{Z}; \mathbb{R})} \frac{\|\mathcal{T}(u)\|_\infty}{\|u\|_\infty} = \|h\|_1.

Remark.

Intuition: This is a beautifully clean result: an LTI system is BIBO stable if and only if its impulse response is absolutely summable (or integrable in CT). The l1l_1 norm of hh is exactly the worst-case amplification factor -- the operator norm of the system. If the impulse response decays fast enough that its total absolute area is finite, the system is stable; if not, there exists some bounded input that will drive the output to infinity.


The Frequency Response (or Transfer) Function of Linear Time-Invariant Systems

A very important property of convolution systems is that, if the input is a harmonic function, so is the output. Let

u(t)=ei2πft,u(t) = e^{i2\pi ft},

be the input to a system with impulse response hL1(R;R)h \in L_1(\mathbb{R}; \mathbb{R}). Then,

y(t)=τ=h(tτ)u(τ)dτ=τ=h(τ)u(tτ)dτ=τ=h(τ)ei2πf(tτ)dτ,y(t) = \int_{\tau=-\infty}^{\infty} h(t - \tau) u(\tau)\,d\tau = \int_{\tau=-\infty}^{\infty} h(\tau) u(t - \tau)\,d\tau = \int_{\tau=-\infty}^{\infty} h(\tau) e^{i2\pi f(t-\tau)}\,d\tau,

which leads to

y(t)=ei2πft(h(s)ei2πfsds).y(t) = e^{i2\pi ft} \left(\int_{-\infty}^{\infty} h(s) e^{-i2\pi fs}\,ds\right).

DefinitionFrequency Response (CT)

We define

h^(f):=h(s)ei2πfsds,\hat{h}(f) := \int_{-\infty}^{\infty} h(s) e^{-i2\pi fs}\,ds,

and call this value the frequency response of the system for frequency ff, whenever it exists. This expression is the Fourier Transform of hh.

Remark.

Intuition: The frequency response tells you what the system does to each pure frequency: it scales the amplitude by h^(f)|\hat{h}(f)| and shifts the phase by h^(f)\angle \hat{h}(f). Complex exponentials are eigenfunctions of LTI systems -- they pass through unchanged in shape, only modified in amplitude and phase. This is precisely why Fourier analysis is so powerful for studying LTI systems.

Later on we will consider u(t)=estu(t) = e^{st}, sCs \in \mathbb{C}, and we will generalize the frequency response above (defined for s=i2πfs = i2\pi f, fRf \in \mathbb{R}) to the notion of transfer function of a system.

A similar discussion applies for a discrete-time system. Let hl1(Z;R)h \in l_1(\mathbb{Z}; \mathbb{R}). If u(n)=ei2πfnu(n) = e^{i2\pi fn}, then

y(n)=(m=h(m)ei2πfm)ei2πfny(n) = \left(\sum_{m=-\infty}^{\infty} h(m) e^{-i2\pi fm}\right) e^{i2\pi fn}

DefinitionFrequency Response (DT)

The frequency response function for a discrete-time LTI system is

h^(f):=m=h(m)ei2πfm.\hat{h}(f) := \sum_{m=-\infty}^{\infty} h(m) e^{-i2\pi fm}.

Remark.

Intuition: The discrete-time frequency response is the DTFT of the impulse response. Just like in continuous time, it tells you the gain and phase shift the system applies to each frequency component. Convolution systems are used as filters through the characteristics of the frequency response.


Steady-State vs. Transient Solutions

Let x(t)RNx(t) \in \mathbb{R}^N. Consider a system defined with the relation:

x(t)=Ax(t)+Bu(t),y(t)=Cx(t)+Du(t),tt0,x'(t) = Ax(t) + Bu(t), \qquad y(t) = Cx(t) + Du(t), \qquad t \geq t_0,

for some fixed t0t_0. Consider an input u(t)=estu(t) = e^{st}, tt0t \geq t_0 for some sCs \in \mathbb{C} (for the time being, assume that s=i2πfs = i2\pi f for some fRf \in \mathbb{R}). Suppose that ss is not an eigenvalue of AA. Using the relation

x(t)=eA(tt0)x(t0)+eAtt0teAτBesτdτ=eA(tt0)x(t0)+eAtt0tesτeAτBdτx(t) = e^{A(t-t_0)}x(t_0) + e^{At}\int_{t_0}^{t} e^{-A\tau}Be^{s\tau}\,d\tau = e^{A(t-t_0)}x(t_0) + e^{At}\int_{t_0}^{t} e^{s\tau}e^{-A\tau}B\,d\tau

we obtain

x(t)=(eA(tt0)x(t0)eA(tt0)eAt0(sIA)1eAt0est0B)+(eAt(sIA)1eAtestB).x(t) = \left(e^{A(t-t_0)}x(t_0) - e^{A(t-t_0)}e^{At_0}(sI - A)^{-1}e^{-At_0}e^{st_0}B\right) + \left(e^{At}(sI - A)^{-1}e^{-At}e^{st}B\right).

Using the property that for any tt

eAt(sIA)1eAt=(sIA)1,e^{At}(sI - A)^{-1}e^{-At} = (sI - A)^{-1},

we obtain

y(t)=CeA(tt0)(x(t0)(sIA)1est0B)+(C(sIA)1B+D)est,tR+.y(t) = Ce^{A(t-t_0)}\left(x(t_0) - (sI - A)^{-1}e^{st_0}B\right) + \left(C(sI - A)^{-1}B + D\right)e^{st}, \qquad t \in \mathbb{R}_+.

The first term is called the transient response of the system and the second term is called the steady-state response.

If AA is a stable matrix, with all its eigenvalues inside the unit circle, the first term decays to zero as tt increases (or with fixed tt, as t0t_0 \to -\infty). Alternatively, if we set t0=0t_0 = 0 and write

x(0)=(sIA)1B,x(0) = (sI - A)^{-1}B,

the output becomes

y(t)=(C(sIA)1B+D)est,tt0.y(t) = \left(C(sI - A)^{-1}B + D\right)e^{st}, \qquad t \geq t_0.

DefinitionTransfer Function

The map

H(s)=C(sIA)1B+DH(s) = C(sI - A)^{-1}B + D

is called the transfer function of the system. When s=i2πfs = i2\pi f, this is the frequency response.

Remark.

Intuition: The transfer function H(s)H(s) generalizes the frequency response from the imaginary axis (s=iωs = i\omega) to the entire complex plane. It encodes how the system responds not just to pure sinusoids, but to exponentially growing or decaying complex exponentials. The case s=i2πfs = i2\pi f recovers the frequency response; evaluating at other complex values gives insight into transient behaviour and stability.

The case with s=i2πfts = i2\pi ft is crucial for stable systems. Later on we will investigate the more general case with sCs \in \mathbb{C}.


Bode Plots for Studying System Response to Harmonic Inputs

If we apply u(t)=ei2πftu(t) = e^{i2\pi ft} or eiωte^{i\omega t}, we observed in the above that the output would be h^(f)ei2πft\hat{h}(f)e^{i2\pi ft}.

Bode plots allow us to efficiently visualize h^(f)\hat{h}(f) by depicting the magnitude and phase, with a logarithmic scale; in the pre-digital era of mid-20th century in the absence of advanced computers, such plots were effective means to represent transfer functions with the logarithmic scale.

Observe that since h^(f)=h^(f)\hat{h}(f) = \overline{\hat{h}(-f)}, it suffices to consider only f0f \geq 0. Let ω=2πf\omega = 2\pi f. Let i=1,2,,5i = 1, 2, \cdots, 5 and si=rieiθis_i = r_i e^{i\theta_i}, where ri=sir_i = |s_i| and θi\theta_i is the phase of sis_i.

h(iω)=s1s2s3s4s5=(r1r2r3r4r5)ei(θ1+θ2θ3θ4θ5).h(i\omega) = \frac{s_1 s_2}{s_3 s_4 s_5} = \left(\frac{r_1 r_2}{r_3 r_4 r_5}\right)e^{i(\theta_1 + \theta_2 - \theta_3 - \theta_4 - \theta_5)}.

Thus,

h(iω)=r1r2r3r4r5|h(i\omega)| = \frac{r_1 r_2}{r_3 r_4 r_5}

and

log(h(iω))=log(r1)+log(r2)log(r3)log(r4)log(r5).\log(|h(i\omega)|) = \log(r_1) + \log(r_2) - \log(r_3) - \log(r_4) - \log(r_5).

Note also that

log(ei(θ1+θ2θ3θ4θ5))=i(θ1+θ2θ3θ4θ5)\log(e^{i(\theta_1 + \theta_2 - \theta_3 - \theta_4 - \theta_5)}) = i(\theta_1 + \theta_2 - \theta_3 - \theta_4 - \theta_5)

so that

h(iω)=θ1+θ2θ3θ4θ5.\angle h(i\omega) = \theta_1 + \theta_2 - \theta_3 - \theta_4 - \theta_5.

Thus, the logarithms allow us to consider the contributions of each complex number in an additive fashion both for the magnitude and the phase.

Building Blocks for Bode Plots

Now, consider

h(iω)=K(iω)n1+iωω0(iωωn)2+2ζiωωn+1.h(i\omega) = K(i\omega)^n \frac{1 + i\frac{\omega}{\omega_0}}{\left(i\frac{\omega}{\omega_n}\right)^2 + 2\zeta i\frac{\omega}{\omega_n} + 1}.

We can thus consider the contributions of K(iω)nK(i\omega)^n, 1+iωω01 + i\frac{\omega}{\omega_0}, and (iωωn)2+2ζiωωn+1\left(i\frac{\omega}{\omega_n}\right)^2 + 2\zeta i\frac{\omega}{\omega_n} + 1 separately.

For K(iω)nK(i\omega)^n, we note that

logK(iω)n=log(K)+nlog(ω)\log|K(i\omega)^n| = \log(|K|) + n\log(\omega)

and

K(iω)n=K+nπ2.\angle K(i\omega)^n = \angle K + n\frac{\pi}{2}.

For 1+iωω01 + i\frac{\omega}{\omega_0}, we use the following approximations:

  • For ω0\omega \approx 0: 1+iωω011 + i\frac{\omega}{\omega_0} \approx 1.
  • For ω=ω0\omega = \omega_0: 1+iωω0=1+i1 + i\frac{\omega}{\omega_0} = 1 + i.
  • For ωω0\omega \gg \omega_0: 1+iωω0ωω0|1 + i\frac{\omega}{\omega_0}| \approx \frac{|\omega|}{\omega_0}.

Likewise, for the angle:

  • For ω0\omega \approx 0: (1+iωω0)0\angle\left(1 + i\frac{\omega}{\omega_0}\right) \approx 0.
  • For ωω0\omega \gg \omega_0: (1+iωω0)π2\angle\left(1 + i\frac{\omega}{\omega_0}\right) \approx \frac{\pi}{2}.
  • At ω=ω0\omega = \omega_0: (1+iωω0)=π4\angle\left(1 + i\frac{\omega}{\omega_0}\right) = \frac{\pi}{4}.

For ((iωωn)2+2iζωωn+1)1\left(\left(i\frac{\omega}{\omega_n}\right)^2 + 2i\zeta\frac{\omega}{\omega_n} + 1\right)^{-1}:

  • For ω0\omega \approx 0, the magnitude is approximately 1, with its logarithm approximately 0.
  • For ω=ωn\omega = \omega_n, the magnitude is 12ζ\frac{1}{2\zeta}.
  • For ωωn\omega \gg \omega_n, the magnitude decays as 2log(ω)-2\log(|\omega|).

For the phase:

  • For ω0\omega \approx 0, the phase is approximately 0.
  • For ω=ωn\omega = \omega_n, the phase is π2-\frac{\pi}{2}.
  • For ωωn\omega \gg \omega_n, the phase is close to π\pi.

Bode plots approximate these expressions in a log-log plot (for the magnitude).


Interconnections of Systems and Feedback Control Systems

Feedback Control System
+reC(s)uP(s)y

We will discuss serial connections, parallel connections, output and error feedback connections.

Control systems are those whose input-output behaviour is shaped by control laws (typically through using system outputs to generate the control inputs -- termed, output feedback --) so that desired system properties such as stability, robustness to incorrect models, robustness (to system or measurement noise) -- also called, disturbance rejection --, tracking a given reference signal, and ultimately, optimal performance are attained. These will be made precise as control theoretic applications are investigated further.


State-Space Description of Linear Systems

We will study state-space realizations of linear time-invariant systems in further detail in Chapter 9. We provide a brief discussion in the following.

Principle of Superposition

For a linear time-invariant system, if (u,y)(u, y) is an input-output pair, then σθu,σθy\sigma_\theta u, \sigma_\theta y is also an input-output pair and thus, a1u+b1σθua_1 u + b_1 \sigma_\theta u, a1y+b1σθya_1 y + b_1 \sigma_\theta y is also such a pair.

State-Space Description of Input-Output Systems

DefinitionState of a System

The notion of a state. Suppose that we wish to compute the output of a system at tt0t \geq t_0 for some t0t_0. In a general (causal) system, we need to use all the past applied input terms u(s),st0u(s), s \leq t_0 and all the past output values y(s),s<t0y(s), s < t_0 to compute the output at t0t_0. The state of a system summarizes all the past relevant data that is sufficient to compute the future paths. Some systems admit a finite-dimensional state representation, some do not.

Remark.

Intuition: The state is a "sufficient summary" of the system's entire history. If you know the state at time t0t_0, you can predict all future outputs given future inputs, without needing any information about what happened before t0t_0. This is what makes state-space methods so powerful: they compress potentially infinite history into a finite-dimensional vector.

Continuous-Time State-Space Form

Consider a continuous-time system given by:

k=0Nakdkdtky(t)=m=0N1bmdmdtmu(t),\sum_{k=0}^{N} a_k \frac{d^k}{dt^k} y(t) = \sum_{m=0}^{N-1} b_m \frac{d^m}{dt^m} u(t),

with aN=1a_N = 1. Such a system can be written in the form:

ddtx(t)=Ax(t)+Bu(t),y(t)=Cx(t)\frac{d}{dt}x(t) = Ax(t) + Bu(t), \qquad y(t) = Cx(t)

A=[010000100a0a1a2aN1]A = \begin{bmatrix} 0 & 1 & 0 & \cdots & 0 \\ 0 & 0 & 1 & \cdots & 0 \\ \vdots & \vdots & & \cdots & 0 \\ -a_0 & -a_1 & -a_2 & \cdots & -a_{N-1} \end{bmatrix}

B=[0001]B = \begin{bmatrix} 0 \\ 0 \\ \vdots \\ 0 \\ 1 \end{bmatrix}

C=[bNbN1b1]C = \begin{bmatrix} b_N & b_{N-1} & \cdots & b_1 \end{bmatrix}

Discrete-Time State-Space Form

Likewise, consider a discrete-time system of the form:

k=0Naky(nk)=m=1Nbmu(nm)\sum_{k=0}^{N} a_k y(n - k) = \sum_{m=1}^{N} b_m u(n - m)

with a0=1a_0 = 1, can be written in the form

x(n+1)=Ax(n)+Bu(n),y(n)=Cx(n)x(n+1) = Ax(n) + Bu(n), \qquad y(n) = Cx(n)

where

xN(n)=y(n),  xN1(n)=y(n1),  ,  x1(n)=y(n(N1))x_N(n) = y(n),\; x_{N-1}(n) = y(n-1),\; \cdots,\; x_1(n) = y(n - (N-1))

A=[010000101aNaN1aN2a1]A = \begin{bmatrix} 0 & 1 & 0 & \cdots & 0 \\ 0 & 0 & 1 & \cdots & 0 \\ \vdots & \vdots & & \cdots & 1 \\ -a_N & -a_{N-1} & -a_{N-2} & \cdots & -a_1 \end{bmatrix}

B=[0001]B = \begin{bmatrix} 0 \\ 0 \\ \vdots \\ 0 \\ 1 \end{bmatrix}

C=[bNbN1b1]C = \begin{bmatrix} b_N & b_{N-1} & \cdots & b_1 \end{bmatrix}

Stability of Linear Systems Described by State Equations

TheoremBIBO Stability via Eigenvalues (CT)

Consider a system defined by the linear differential equation

x(t)=Ax(t)+u(t),tR.x'(t) = Ax(t) + u(t), \quad t \in \mathbb{R}.

This system is BIBO stable if and only if

maxλi{Re{λi}}<0,\max_{\lambda_i}\{\operatorname{Re}\{\lambda_i\}\} < 0,

where Re{}\operatorname{Re}\{\cdot\} denotes the real part of a complex number, and λi\lambda_i denotes the eigenvalues of AA.

Remark.

Intuition: For a continuous-time system, BIBO stability requires all eigenvalues of AA to have strictly negative real parts -- they must lie in the open left half of the complex plane. This ensures that all natural modes of the system decay exponentially, so no bounded input can cause the output to blow up. This is the continuous-time analogue of the "all poles inside the unit circle" condition.

TheoremBIBO Stability via Eigenvalues (DT)

Consider a system defined by the linear difference equation

x(n+1)=Ax(n)+u(n),nZ.x(n+1) = Ax(n) + u(n), \quad n \in \mathbb{Z}.

This system is BIBO stable if and only if

maxλi{λi}<1,\max_{\lambda_i}\{|\lambda_i|\} < 1,

where λi\lambda_i denotes the eigenvalues of AA.

Remark.

Intuition: For a discrete-time system, BIBO stability requires all eigenvalues of AA to lie strictly inside the unit disk in the complex plane. This is the discrete-time counterpart to the left-half-plane condition: geometric decay in discrete time corresponds to exponential decay in continuous time.


Exercises

Exercise

Consider a linear system described by the relation:

y(n)=mZh(n,m)u(m),nZy(n) = \sum_{m \in \mathbb{Z}} h(n, m)u(m), \quad n \in \mathbb{Z}

for some h:Z×ZCh : \mathbb{Z} \times \mathbb{Z} \to \mathbb{C}.

(a) When is such a system causal?

(b) Show that such a system is time-invariant if and only if it is a convolution system.

Exercise

Let x(t)RNx(t) \in \mathbb{R}^N and t0t \geq 0 and real-valued. Recall that the solution to the following differential equation:

x(t)=Ax(t)+Bu(t)x'(t) = Ax(t) + Bu(t)

with the initial condition x(0)=x0x(0) = x_0 is given by

x(t)=eA(t)x0+τ=0teA(tτ)Bu(τ)dτ,tt0.x(t) = e^{A(t)}x_0 + \int_{\tau=0}^{t} e^{A(t-\tau)}Bu(\tau)\,d\tau, \quad t \geq t_0.

Suppose x(0)=0x(0) = 0 and u(t)=0u(t) = 0 for t<0t < 0. Express the solution as a convolution x(t)=(hu)(t)x(t) = (h * u)(t), and find h(t)h(t).

Note: With the assumption that the system is stable, we can avoid the condition that u(t)=0u(t) = 0 for t<0t < 0. In this case, we are able to write

x(t)=eAtt0x(t0)+τ=t0teA(tτ)Bu(τ)dτ,x(t) = e^{At - t_0}x(t_0) + \int_{\tau=t_0}^{t} e^{A(t-\tau)}Bu(\tau)\,d\tau,

and take the limit as t0t_0 \to -\infty, leading to h(t)=eAtB1{t0}h(t) = e^{At}B\,1_{\{t \geq 0\}}.

Exercise

Let x(n)RNx(n) \in \mathbb{R}^N and nZn \in \mathbb{Z}. Consider a linear system given by

x(n+1)=Ax(n)+Bu(n),n0x(n+1) = Ax(n) + Bu(n), \quad n \geq 0

with the initial condition x(0)=0x(0) = 0. Suppose x(0)=0x(0) = 0 and u(n)=0u(n) = 0 for n<0n < 0. Express the solution x(n)x(n) as a convolution x(n)=(hu)(n)x(n) = (h * u)(n), and find h(n)h(n).

Note: With the assumption that the system is stable, we can avoid the condition that u(n)=0u(n) = 0 for n<0n < 0. In this case, we can write

x(n)=Ann0x(n0)+m=n0n1Anm1Bu(m),x(n) = A^{n-n_0}x(n_0) + \sum_{m=n_0}^{n-1} A^{n-m-1}Bu(m),

and take the limit as n0n_0 \to -\infty leading to h(n)=An1B1{n1}h(n) = A^{n-1}B\,1_{\{n \geq 1\}}.

Exercise

Consider a continuous-time system described by the equation:

dy(t)dt=ay(t)+u(t),tR,\frac{dy(t)}{dt} = ay(t) + u(t), \quad t \in \mathbb{R},

where a<0a < 0.

(a) Find the impulse response of the system. Is the system bounded-input-bounded-output (BIBO) stable?

(b) Suppose that the input to this system is given by cos(2πf0t)\cos(2\pi f_0 t). Let yf0y_{f_0} be the output of the system. Find yf0(t)y_{f_0}(t).

(c) If exists, find

limf0yf0(t),\lim_{f_0 \to \infty} y_{f_0}(t),

for all tR+t \in \mathbb{R}_+.

Exercise

Consider a discrete-time system described by the equation:

y(n+1)=a1y(n)+a2y(n1)+u(n),nZ.y(n+1) = a_1 y(n) + a_2 y(n-1) + u(n), \quad n \in \mathbb{Z}.

(a) Is this system linear? Time-invariant?

(b) For what values of a1,a2a_1, a_2 is the system BIBO (bounded-input-bounded-output) stable?

Exercise (Stability of Linear Time-Varying Systems)

Let TT be a linear system mapping with the representation;

y(n)=mZh(n,m)u(m),nZy(n) = \sum_{m \in \mathbb{Z}} h(n, m)u(m), \quad n \in \mathbb{Z}

for some h:Z×ZRh : \mathbb{Z} \times \mathbb{Z} \to \mathbb{R}. Show that this system is BIBO stable if

supnmh(n,m)<.\sup_n \sum_m |h(n, m)| < \infty.

Let us define a system to be regularly BIBO stable if for ϵ>0\epsilon > 0, δ>0\exists \delta > 0 such that uδ\|u\|_\infty \leq \delta implies yϵ\|y\|_\infty \leq \epsilon. Show that the system above is regularly BIBO stable if and only if

supnmh(n,m)<.\sup_n \sum_m |h(n, m)| < \infty.

Exercise

Let TT be a linear system mapping l1(Z;R)l_1(\mathbb{Z}; \mathbb{R}) to l1(Z;R)l_1(\mathbb{Z}; \mathbb{R}). Show that this system is linear and continuous only if the system can be written so that with y=T(u)y = T(u);

y(n)=mZh(n,m)u(m),nZy(n) = \sum_{m \in \mathbb{Z}} h(n, m)u(m), \quad n \in \mathbb{Z}

for some h:Z×ZRh : \mathbb{Z} \times \mathbb{Z} \to \mathbb{R}.