Signal Spaces: Linear, Banach and Hilbert Spaces, and Basis Expansions
Normed linear spaces, metric spaces, Banach spaces, inner product spaces, Hilbert spaces. Orthogonality, separability, and signal expansions using Fourier, Haar, and polynomial bases.
In this chapter, we present a general review of signal spaces. These spaces form the mathematical foundation for everything that follows in the course: systems, transforms, stability, and control all rely on understanding the structure of the spaces in which signals live.
Normed Linear (Vector) Spaces and Metric Spaces
Linear Spaces
A linear (vector) space is a space which is closed under addition operation and a scalar multiplication operation , such that
with the following properties (we note that we may take the scalars to be either real or complex numbers). The following are satisfied for all and scalars:
(i)
(ii)
(iii)
(iv)
(v) There is a null vector such that
(vi)
(vii)
(viii) For every , there exists an element, called the (additive) inverse of and denoted with with the property .
Intuition: A linear space is the most basic algebraic structure for working with signals: it guarantees you can add signals together, scale them, and these operations behave predictably. Think of it as the minimal "sandbox" in which superposition -- the cornerstone of linear systems theory -- makes sense.
Intuitively, a linear space is a set where you can add elements together and scale them, and these operations behave the way you would expect. This is the most basic algebraic structure we need to talk about "signals" in a mathematically precise way.
(i) The space with the pointwise addition and scalar multiplication is a linear space. The null vector is .
(ii) Consider the interval . The collection of real-valued continuous functions on , with pointwise addition and scalar multiplication is a linear space. The null element is the function which is identically . This space is called the space of real-valued continuous functions on .
(iii) With pointwise addition and scalar multiplication, the set of all infinite sequences of real numbers mapping to , each having only a finite number of elements not equal to zero is a vector space; for example, if one adds two such sequences, the sum also belongs to this space. This space is called the space of finitely many non-zero sequences.
(iv) The collection of all polynomial functions defined on an interval with complex coefficients
forms a complex linear space. Note that the sum of polynomials is another polynomial.
(v) The -dimensional integer lattice with pointwise addition and scalar multiplication is not a linear space.
A non-empty subset of a (real) linear vector space is called a subspace of if
Intuition: A subspace is a "self-contained" portion of a vector space -- any linear combination of elements in the subspace stays in the subspace. In signal processing, subspaces often represent restricted signal classes (e.g., band-limited signals or signals spanned by a finite set of basis functions) within a larger signal space.
In particular, the null element is an element of every subspace. For , two subspaces of a vector space , is also a subspace of .
Normed Linear Spaces
A normed linear space is a linear vector space on which a map from to , called its norm, is defined such that:
- , and if and only if is the null element of .
- (triangle inequality)
- .
Intuition: A norm gives us a way to measure "size" or "distance" in a vector space -- just like absolute value does for real numbers, but generalized to arbitrary signal spaces. Without a norm, we can add and scale vectors but have no notion of how large a signal is or how far apart two signals are. The choice of norm (e.g., peak value vs. total energy) determines what "closeness" means for your application.
The norm gives us a way to measure the "size" of a signal. Different norms measure size in different ways -- for instance, the maximum value of a signal vs. its total energy. The choice of norm profoundly affects which signals are "small" or "large."
Let for two linear spaces , , denote the set of all maps such that for all , .
a) The space of continuous functions from to with the norm is a normed linear space.
b) For :
is a normed linear space for all .
c) Recall that if is a set of real numbers bounded from above, then there is a smallest real number such that for all . The number is called the least upper bound or supremum of . If is not bounded from above, then the supremum is . In view of this, for , we define
d) is a normed linear space. For , we typically write: . However, for , to satisfy the condition that implies that , we need to assume that functions which are equal to zero almost everywhere are equivalent; for the definition is often revised with essential supremum instead of supremum so that
This subtle difference needs to be made explicit in some applications.
To show that defined above is a normed linear space, we need to show that .
For with ,
Intuition: Minkowski's inequality is the triangle inequality for spaces. It says the "length" of the sum of two sequences is at most the sum of their "lengths." This is the key property that confirms norms are genuine norms, and it generalizes the familiar fact that the shortest path between two points is a straight line.
See Exercise, which also studies a proof of Holder's Inequality, that is critically used in the proof of Minkowski's inequality.
Let with . Then, for , ,
Intuition: Holder's inequality bounds the "interaction" between two sequences living in dual spaces. When , this reduces to Cauchy-Schwarz. In practice, it is the workhorse behind proving that convolutions, inner products, and many signal processing operations produce finite results.
Transformations and Continuity
Let and be two normed linear spaces, and let be a subset of . A law (rule, relation) which relates with every element of an element in , is called a transformation from to with domain . The relation is often expressed as .
If for every there is an such that , the transformation is said to be onto (or surjective). If for every element of , there is at most one such that , the transformation is said to be one-to-one (or injective). If these two properties hold simultaneously, the transformation is said to be bijective.
Intuition: A transformation is simply a rule that maps inputs to outputs -- this is the abstract notion of a "system." Surjectivity means every possible output is reachable, injectivity means distinct inputs always produce distinct outputs, and bijectivity means the mapping is perfectly reversible. These properties are essential when asking whether a system can be inverted or whether information is lost.
A transformation (or ) is linear if for every and , we have .
Intuition: Linearity means the system obeys superposition: the response to a sum of inputs equals the sum of the individual responses, and scaling the input scales the output by the same factor. This is the defining property of the systems we study in this course, and it is what makes Fourier analysis so powerful -- we can decompose any input into simple components, analyze each one separately, and add the results.
A transformation for normed linear spaces is continuous at , if for every , such that implies that (here the norms depend on the vector spaces and ). is said to be continuous, if it is continuous at every .
Intuition: Continuity means that small perturbations in the input produce small perturbations in the output -- the system does not "blow up" from tiny changes. This is the mathematical formalization of robustness: a continuous system behaves predictably under small measurement errors or noise.
A transformation is sequentially continuous at , if implies that .
Intuition: Sequential continuity says the same thing as continuity but in the language of sequences: if a sequence of inputs converges to some input, then the corresponding outputs converge to the corresponding output. This formulation is often easier to verify in practice, since you can work with concrete sequences rather than abstract - arguments.
Sequential continuity and continuity are equivalent for normed linear spaces.
Intuition: This theorem tells us that in normed linear spaces, the - definition of continuity and the sequence-based definition are interchangeable. You can use whichever is more convenient for the problem at hand -- they will always give the same answer.
If the transformation is a linear one, then continuity is equivalent to being continuous at the null element.
Intuition: For a linear transformation, you only need to check continuity at the origin -- continuity there automatically guarantees continuity everywhere. This is because linearity lets you "translate" any neighborhood to the origin. In practice, this greatly simplifies verifying that a linear system is well-behaved.
For some applications, sequential continuity may be more convenient to work with as one may not need to quantify pairs to verify continuity.
Metric Spaces
A metric defined on a set , is a function such that:
- and if and only if .
- .
- .
A metric space is the set equipped with metric .
Intuition: A metric generalizes the concept of "distance" to sets that may not be vector spaces. While a norm requires linear structure (addition and scaling), a metric only requires a sensible notion of distance between pairs of points. This makes metric spaces useful for analyzing convergence and continuity in settings where addition of elements may not be defined.
A normed linear space is also a metric space, with metric . Many spaces of functions that we encounter are not linear spaces. For such spaces the concept of metric, defined above, provides an appropriate generalization of the notion of norm studied earlier.
Banach Spaces
A sequence in a normed space is Cauchy if for every , there exists an such that , for all .
Intuition: A Cauchy sequence is one whose elements get arbitrarily close to each other as you go further along -- the sequence "wants" to converge, even if we do not yet know the limit. Think of it as a sequence that is "trying to settle down." The key question is whether the space contains the point it is settling toward.
An important observation on Cauchy sequences is that every converging sequence is Cauchy, however, not all Cauchy sequences are convergent: This is because the limit might not live in the original space where the sequence elements take values from. This motivates the property of completeness.
A normed linear space is complete, if every Cauchy sequence in has a limit in . A complete normed linear space is called Banach.
Intuition: Completeness means there are "no holes" in the space -- every sequence that looks like it should converge actually does converge to something inside the space. A Banach space is the right setting for iterative algorithms and approximations, because you are guaranteed that limits of approximating sequences exist. Without completeness, a perfectly reasonable iterative procedure might converge to something that is not even in your space.
Banach spaces are important for many applications in engineering and applied mathematics: In many mathematical applications (such as existence of and numerical methods for solutions to differential equations), machine learning problems (such as iterative updates of data driven dynamics), stochastic analysis or optimization problems (for which a sequence of approximating solutions may be obtained for example via dynamic programming methods), sometimes we would like to know whether a given sequence converges, without necessarily knowing what the limit of the sequence may be. Banach spaces allow us to use Cauchy sequence arguments to ensure the existence of limits and some of their properties.
An example is the following: consider the solutions to the for a square matrix and a vector. One can identify conditions on an iteration of the form to form a Cauchy sequence and converge to a solution through the contraction principle. As noted above, existence of solutions to ordinary differential equations also follow from Cauchy sequence arguments.
A subset of a Banach space is complete if and only if it is closed. If the set is closed, every Cauchy sequence in this set has a limit in and this limit should be a member of this set, hence the set is complete. If it is not closed, one can provide a counterexample sequence which does not converge to a point in the subset.
The real space with the usual distance is a complete space.
a) Let be the space of continuous functions in with the supremum norm
This space is a Banach space.
b) Let be the space of continuous functions in with the norm
This space is not a Banach space.
Intuition: This theorem shows that completeness depends critically on the choice of norm. Under the supremum norm, the limit of continuous functions must be continuous (uniform convergence preserves continuity). But under the norm, a sequence of continuous functions can converge to a discontinuous function that is not in the original space. The lesson: the same set of functions can be "complete" or "incomplete" depending on how you measure distance.
This theorem illustrates a crucial point: the same set of functions can be complete or incomplete depending on the norm. The supremum norm is "strong enough" to prevent discontinuous limits, while the norm is not.
is a Banach space for all . The same holds for .
Intuition: The sequence spaces are complete, meaning they are safe to work in: any "convergent-looking" sequence of signals in will converge to something that is still in . This is essential because many algorithms in signal processing and control construct solutions as limits of iterative approximations, and completeness guarantees those limits are well-defined.
A brief remark on some typical notations: When the range space is , the notation denotes for a discrete-time index set and likewise for a continuous-time index set , denotes .
In a normed linear space , an infinite sequence of elements converges to an element , if the sequence converges to zero.
Intuition: Convergence in a normed space means the distance between the sequence elements and the limit shrinks to zero. Note that what "distance" means depends on the norm -- convergence in the supremum norm (uniform convergence) is a much stronger statement than convergence in the norm (energy convergence). Always be aware of which norm you are working with.
Hilbert Spaces
A pre-Hilbert space is a linear vector space on which an inner product is defined, where the following are satisfied:
- (the superscript denotes the complex conjugate) (we will also use to denote the complex conjugate)
- , equals iff is the null element.
Intuition: An inner product enriches a vector space with geometric structure: it gives us notions of "angle" and "projection" between signals, not just "size." This is what allows us to decompose a signal into orthogonal components -- the mathematical foundation for Fourier series, least-squares fitting, and matched filtering.
Thus, corresponding to each pair the inner product is a scalar. The inner product gives us a notion of "angle" between signals and allows us to define orthogonality -- concepts that are essential for signal decomposition and optimal approximation.
For ,
where equality occurs if and only if for some scalar .
Intuition: Cauchy-Schwarz is the abstract generalization of "the dot product of two vectors cannot exceed the product of their lengths." It bounds how much two signals can be "aligned" -- and equality holds precisely when one signal is a scaled copy of the other. This inequality underpins nearly every bound and estimate in signal processing and statistics.
The Cauchy-Schwarz inequality is one of the most important inequalities in all of mathematics. It tells us that the inner product of two signals is bounded by the product of their norms.
In a pre-Hilbert space, the inner-product defines a norm:
The proof for this result requires one to show that satisfies the triangle inequality, that is
which can be proven by an application of the Cauchy-Schwarz inequality.
In a given problem or application, the inner-product is to be defined. The inner product, in the special case of , is typically the usual inner vector product; hence with the usual inner-product is a pre-Hilbert space.
A complete pre-Hilbert space, is called a Hilbert space.
Intuition: A Hilbert space is the "best of all worlds" for signal analysis: it has addition and scaling (linear space), a notion of distance (norm), a notion of angle and projection (inner product), and no missing limits (completeness). The spaces and are the prototypical Hilbert spaces in signal processing -- they capture signals with finite energy.
Hence, a Hilbert space is a Banach space, endowed with an inner product, which induces its norm. For example, if we define as a pre-Hilbert space with the inner-product , this would not be a Hilbert space, since this would not be a complete space.
The space is a Hilbert space with for . Furthermore, defines a norm in .
Likewise, the space is a Hilbert space with for and its norm is .
The inner product is continuous: if , and , then for in a Hilbert space.
Intuition: Continuity of the inner product means that if two sequences of signals converge, then their "correlation" (inner product) converges to the correlation of the limits. This is practically important because it guarantees that Fourier coefficients, projections, and energy computations are stable under approximation.
Why are we interested in Hilbert Spaces?
Hilbert spaces have several very useful properties:
- Hilbert spaces allow us to have a geometric insight on a set of signals via the inner-product, orthogonality, and basis expansions.
- If a Hilbert space is separable (to be defined shortly), there exists a countably (or sometimes only finitely) many sequence of orthonormal vectors which can be used as basis to represent all the members in this space. This is used in many areas of applied mathematics and engineering, such as Fourier expansions among many others.
- A Hilbert space formulation allows us to develop approximations of signals using a finite number of basis signals.
- Also related to the item above, we can state a Projection Theorem which certifies optimality in a wide variety of applications. This geometric insight carries over to more general optimization problems in spaces which are not Hilbert (via duality analysis, to be discussed later).
Orthogonality and the Projection Theorem
In a Hilbert space , two vectors are orthogonal if . A vector is orthogonal to a set if .
Intuition: Orthogonality is the generalization of perpendicularity to abstract signal spaces. Two orthogonal signals have zero "overlap" or correlation -- they carry completely independent information. This concept is at the heart of Fourier analysis, where signals are decomposed into mutually orthogonal frequency components.
By the Pythagorean theorem, for orthogonal, we have that .
A set is closed if contains the limit of any converging sequence taking values from .
Let be a Hilbert space and a subspace of . Consider the problem:
(i) A necessary and sufficient condition for to be the minimizing element in so that
or equivalently
is that, is orthogonal to .
If exists, such an is unique.
(ii) Let be a Hilbert space and be a closed subspace of . For any vector , there is a unique vector satisfying .
Intuition: The Projection Theorem says that the best approximation to a signal from a subspace is obtained by "dropping a perpendicular" onto that subspace -- the approximation error is orthogonal to the subspace. This is exactly what happens in least-squares estimation, optimal filtering (Wiener filter), and signal compression: the optimal reconstruction is always the orthogonal projection.
The Projection Theorem is one of the most powerful results in Hilbert space theory. It says that the best approximation to a signal from a subspace is obtained by "projecting" orthogonally onto that subspace -- the error is perpendicular to the subspace. This has direct applications in optimal filtering, least squares estimation, and signal compression.
Consider the minimization of
for some , over all such that
(a) State the problem as a Projection problem by identifying the Hilbert space, and the projected subspace.
(b) Compute the solution, that is find the minimizing .
Intuition: This exercise shows that finding the best polynomial approximation of degree to on is exactly a projection problem in . The Hilbert space is , the subspace is the set of polynomials of degree at most 2, and the minimizer is the orthogonal projection of onto .
Approximations and Signal Expansions
Orthogonality
A set of vectors in a Hilbert space is orthogonal if all elements of this set are orthogonal to each other. The set is orthonormal if each vector in this set has norm equal to one.
Intuition: An orthonormal set is a collection of "perfectly independent, unit-sized" building blocks for signals. Orthogonality ensures no redundancy (each basis vector carries unique information), and normalization to unit length simplifies coefficient computation to simple inner products. The complex exponentials in Fourier analysis form the prototypical orthonormal set.
The Gram-Schmidt orthogonalization procedure can be invoked to generate a set of orthonormal sequences. This procedure states that given a sequence of linearly independent vectors, there exists an orthonormal sequence of vectors such that for every , there exists with
that is, the linear span of is equal to the linear span of for every .
Recall that a set of vectors is linearly dependent if there exists a complex-valued vector such that with at least one coefficient .
A sequence of orthonormal vectors is a linearly independent collection.
Intuition: Because orthonormal vectors point in completely "different directions," no vector in the set can be built from the others. This guarantees there is no redundancy in an orthonormal basis -- each basis function contributes something genuinely new to the representation.
We say that a sequence converges to , if for every there exists such that , for all .
Let be a sequence of orthonormal vectors in a Hilbert space . Let be a sequence of vectors in . The sequence converges to some vector if and only if
In this case, we have that .
Intuition: This theorem provides the bridge between abstract Hilbert space theory and practical signal representation: a signal can be expanded in an orthonormal basis if and only if the sum of squared coefficients is finite -- which is exactly the condition that the signal has finite energy. Moreover, the coefficients are simply inner products with the basis vectors, making them easy to compute.
This theorem provides the fundamental link between Hilbert space theory and signal representation: a signal can be expanded in an orthonormal basis if and only if the sum of squared coefficients is finite -- which is exactly the condition that the signal has finite energy.
Separable Hilbert Spaces and Countable Expansions
Given a normed linear space , a subset is dense in , if for every , and each , there exists a member such that .
Intuition: A dense subset can approximate any element in the space to arbitrary precision. Think of how rational numbers are dense in the reals -- every real number can be approximated as closely as you like by a rational. Similarly, saying that polynomials are dense in means any square-integrable signal can be approximated arbitrarily well by a polynomial.
A set is countable if either it has finitely many elements or there is a one-to-one correspondence between the set and the set of natural numbers (which would then be an enumeration / counting map).
Intuition: Countability means you can list all elements in a sequence, even if the list is infinite. This matters because computers and algorithms work with sequences of numbers. A countable basis for a signal space means you can, in principle, represent any signal by a list of coefficients -- which is exactly what digital signal processing does.
Examples of countable spaces are finite sets, the set of integers, and the set of rational numbers. An example of uncountable sets is the set of real numbers.
(a) A countable union of countable sets is countable.
(b) Cartesian product of finitely many countable sets is countable.
(c) Cartesian product of (countably) infinitely many finite sets may not be countable. The same holds even when each of the sets is finite.
(d) is not countable.
Intuition: These results establish the fundamental size distinctions between sets. Part (d) -- that is uncountable -- is especially important: it means continuous-time signals live in an inherently "larger" space than any list can capture. This is precisely why we need the machinery of dense subsets and basis expansions to represent continuous signals with countable data.
Cantor's diagonal argument and the triangular enumeration are important steps in proving the theorem above.
A space is separable, if it contains a countable dense set.
Intuition: Separability means that even though the space may contain uncountably many elements, a countable subset is "enough" to approximate everything. This is the mathematical guarantee that digital signal processing is possible: separable signal spaces can be faithfully represented using countably many basis functions (like Fourier harmonics) and their coefficients.
Separability states that it suffices to work with a countable set, when a set is uncountable, for computational purposes. Examples of separable sets are , and the set of continuous and bounded functions on a compact set metrized with the maximum distance between the functions.
Let be a separable Hilbert space. Then, every orthonormal system of vectors in has a finite or countably infinite number of elements.
Intuition: In a separable Hilbert space, you can never have an uncountable collection of mutually orthogonal directions. This limits the "dimensionality" of the space in a way that makes countable basis expansions (like Fourier series) sufficient to represent every signal. It is the reason why a countable set of frequencies captures all the information in a square-integrable signal.
An orthonormal sequence in a Hilbert space is complete if the only vector in which is orthogonal to each of the vectors in the sequence is the null vector.
Intuition: Completeness of an orthonormal sequence means there is "no signal left over" that the basis misses. If a signal is orthogonal to every basis vector, it must be zero. This is the condition that distinguishes a true basis (which can represent everything) from a partial set (which can only represent signals in a subspace).
A Hilbert space contains a complete orthonormal sequence (that is, a countable collection of such vectors) if and only if it is separable.
Intuition: This is the central structural result: a Hilbert space admits a countable basis (enabling Fourier-type expansions) if and only if it is separable. Since the signal spaces we care about (, ) are separable, this theorem guarantees that every finite-energy signal can be expressed as a countable sum of orthonormal basis functions.
The proof above also showed the following result:
In a Hilbert space , a complete orthonormal sequence defines a basis in the sense that for any , we have
Intuition: This is the master representation theorem: any signal in a separable Hilbert space can be perfectly reconstructed from its inner products with basis vectors. The coefficients are the "coordinates" of the signal, and the reconstruction converges in the energy () sense. Fourier series, wavelet decompositions, and all other orthonormal expansions are special cases of this result.
This is the fundamental representation theorem for signals in Hilbert spaces. Every signal can be decomposed into a (possibly infinite) sum of orthonormal basis vectors, with coefficients given by inner products. This is the abstract framework underlying Fourier series, wavelet expansions, and many other signal representations.
Separability of and spaces
In view of Theorem, the following result builds on the fact that the sequence of orthonormal vectors
is a countable complete orthonormal set in : Note that for any , and hence for any vector
The Hilbert space with inner product
is separable.
Intuition: Separability of means that the space of square-summable sequences -- the natural home for discrete-time finite-energy signals -- can be spanned by a countable orthonormal basis. The standard basis vectors (which are 1 at position and 0 elsewhere) serve as this basis, and every sequence is simply a list of its values at each time index.
Next, we will show that is separable for . To establish this result, we will review some useful facts.
Any function in can be approximated arbitrarily well by a polynomial under the supremum norm.
Intuition: The Weierstrass theorem says that polynomials are "universal approximators" for continuous functions on a closed interval. No matter how complicated a continuous signal is, a polynomial of sufficiently high degree can match it to any desired accuracy. This foundational result is what lets us build toward proving that spaces are separable.
The set , of continuous functions, is dense in .
Intuition: This theorem says that any square-integrable function -- even one with discontinuities -- can be approximated arbitrarily well (in the energy sense) by a continuous function. This is reassuring for engineers: continuous, "well-behaved" signals are dense enough to represent all finite-energy signals, so working with smooth approximations does not lose generality.
The space is separable.
Intuition: Separability of is the key result that justifies Fourier series: it guarantees that every finite-energy signal on can be represented by a countable orthonormal basis. The proof chains together polynomial approximation (Weierstrass) and density of continuous functions to construct such a basis -- and the complex exponentials of Fourier series are one natural choice.
is separable.
Intuition: Extending separability from a bounded interval to all of means Fourier-type expansions work for signals defined on the entire positive real line, not just finite intervals. The key idea is that any finite-energy signal has most of its energy concentrated in some finite window, so the bounded-interval result can be "stitched together" to cover the whole line.
Two further results:
The set is dense in .
Intuition: This says that any absolutely integrable signal on can be approximated by a square-integrable signal. In practice, this means results proven for signals (where we have Hilbert space tools) can often be extended to signals through approximation arguments.
Let denote the space of continuous functions with compact support. is dense in .
Intuition: This result says that any integrable signal can be approximated by a continuous function that is zero outside some finite interval. In engineering terms, every signal can be well-approximated by a "nice" signal that is both smooth and compactly supported -- a key step in proving results about Fourier transforms, since such signals are easy to integrate and transform.
Signal expansions in or : Fourier, Haar and Polynomial Bases
Fourier Signals as Basis Vectors and Fourier Series
Fourier series is a very important class of orthonormal sequences which are used to represent both discrete-time and continuous-time signals. These will be studied later on in much detail. In particular, we will soon see that in the family of complex exponentials
provides a complete orthonormal sequence.
Accordingly, for any , we can write
or by expanding the inner-product, we have
where the convergence of the infinite sum is in the sense.
This expansion is precisely the Fourier series expansion of the function in . The inner-product defines the Fourier transform.
Legendre Polynomials as Basis Vectors
We have seen, in the context of Theorems 2.3.7 and 2.3.8 (see the proof of 2.3.9), the functions can be used to construct an orthonormal collection of signals which is complete in . These complete orthonormal polynomials are called Legendre polynomials.
Haar Functions as Basis Vectors
One further practically very important basis is the class of Haar functions (known as wavelets). Define
and for , ,
The (Haar) set of vectors
is a complete orthonormal sequence in .
Intuition: The Haar system provides an alternative to Fourier series for representing signals. While Fourier basis functions are smooth sinusoids spread over the entire interval, Haar functions are localized step functions that capture information at specific locations and scales. This makes wavelets particularly effective for signals with sharp transitions or edges, which is why they are widely used in image compression (e.g., JPEG 2000).
The important observation to note here is that, different expansions might be suited for different engineering applications: for instance Haar series are occasionally used in image processing with certain edge behaviours, whereas Fourier expansion is extensively used in speech processing and communications theoretic applications.
Approximations
Approximations allow us to represent data using finitely many vectors. The basis expansions studied above can be used to obtain the best approximation of a signal up to finitely many terms to be used in an approximation: This can be posed as a projection problem, and we have seen that the best approximation is one in which the approximation error is orthogonal to all the vectors used in the approximation (defining an approximation subspace).
Exercises
Exercise
(a) The set , which is the set of all functions from to that are infinitely differentiable, together with the operations of addition and scalar multiplication defined as follows, is a vector space: For any
and for any and
(i) Now, consider to be the set of all (polynomial) functions that maps to such that any can be written as for some with . Suppose that we define the same addition and scalar multiplication operations as defined above. Is a subspace in ?
(ii) Show that the space of all functions in which map to which satisfy is a vector space with addition and multiplication defined as above.
(b) Consider the set . On , define an addition operation and a scalar multiplication operation as follows:
Show that, with these operations, is a vector space.
(c) Consider the set
On this set, define an addition operation and a scalar multiplication operation as follows:
Show that, with these operations, is a vector space. Hint: Consider a bijection between and the space with .
Exercise (Holder's and Minkowski's Inequalities)
Let with . Let and . Then,
This is known as Holder's inequality. Equality holds if and only if
for each .
To prove this, perform the following:
(a) Show that for : with equality if and only if . To show this, you may consider the function and see how it behaves for and let .
(b) Apply the inequality to the numbers:
Holder's inequality is useful to prove Minkowski's inequality which states that for ,
This proceeds as follows:
Thus, using that ,
Now, the above holds for every . Taking the limit (first on the right and then on the left), it follows that
which is the desired inequality.
Exercise
Given a normed linear space , introduce a map :
Show that is a metric: That is, it satisfies the triangle inequality:
and that iff , and finally .
Exercise
Let be a complete orthonormal sequence in a real Hilbert space . Let be a subspace of , spanned by , for some finite set . That is,
Let be given. Find which is the solution to the following:
in terms of , and .
Hint: Any vector in can be written as .
Exercise
Let be a mapping given by:
Is continuous at any given ?
Exercise
Consider an inner-product defined by:
Is the resulting inner-product (pre-Hilbert) space separable?
Exercise
Let ; that is, map to , such that and and , for all .
For (a)-(c) below, state if the following are true or false with justifications in a few sentences:
(a) is an inner-product.
(b) is an inner-product.
(c) contains a complete orthonormal sequence, where .
Exercise
Let be a Hilbert space and . Prove the following:
(a) (Parallelogram Law)
(b)
(c)
Exercise
Let be a finite dimensional Hilbert space and be two linearly independent vectors in .
Let . Show that, among all vectors , which satisfies
the vector has the minimum norm if satisfies:
with
Exercise
Let be a Hilbert space and be a dense subset of . Suppose that any element in is such that for every , there exist and so that
where is a countable sequence of orthonormal vectors in .
Is it the case that is separable?
Exercise
Let be in the real Hilbert space with the inner product
We would like to express in terms of the following two signals (which belong to the Haar signal space)
such that
is minimized, for .
(a) Using the Gram-Schmidt procedure, obtain two orthonormal vectors such that these vectors linearly span the same space spanned by .
(b) State the problem as a projection theorem problem by clearly identifying the Hilbert space and the projected subspace.
(c) Let . Find the minimizing values.
Exercise
Let denote the normed linear space of continuous functions from to under the supremum norm. We observed earlier that polynomials can be used to approximate any function in this space with arbitrary precision, under the supremum norm (Weierstrass Theorem).
Given this, repeating the arguments we made in class, argue that the family of polynomials can be used to form a complete orthonormal sequence in . This also establishes that is separable.
Exercise
Alice and Bob are approached by a generous company and asked to solve the following problem: The company wishes to store any signal in in a computer with a given error of , that is for every , there exists some signal such that (thus the error is uniform over all possible signals), where is the stored family of signals (in the computer's memory).
To achieve this, they encourage Alice or Bob to use a finite or a countable expansion to represent the signal and later store this signal in an arbitrarily large memory. Hence, they allow Alice or Bob to purchase as much memory as they would like for a given error value of .
Alice turns down the offer and says it is impossible to do that for any with a finite memory and argues then she needs infinite memory, which is impossible.
Bob accepts the offer and says he may need a very large, but finite, memory for any given ; thus, the task is possible.
Which one is the accurate assessment?
(a) If you think Alice is right, which further conditions can she impose to make this possible? Why is she right?
(b) If you think Bob is right, can you suggest a method? Why is he right?
Exercise
Prove Theorem using a probability theoretic method. Proceed as follows: The number
can be expressed as the expectation , where , where is an i.i.d. collection of Bernoulli random variables where with probability and with probability . Here, observe that the sum has a binomial distribution. Thus,
where , for each , denotes the expectation with respect to the i.i.d. Bernoulli random variables each with . Let denote the probability measure induced by these -parametrized i.i.d. sequence of Bernoulli random variables.
Since is continuous and is compact, is uniformly continuous. Thus, for every , there exists such that implies that . Thus,
The last term converges to as by the law of large numbers. The above holds for every .
Now, one needs to show that this convergence is uniform in : For this show that for all , via Markov's inequality and the independence of
establishing uniform convergence (over ), and thus complete the proof.
Exercise (A useful result on countability properties)
Let be a monotonically increasing function (that is, implies that ). Show that can have at most countably many points of discontinuity.
Hint: If a point is such that is discontinuous, then there exists with , , . Express . Let . It must be that is finite for otherwise the jump would be unbounded in the interval . Then, the countable union will be countable. Finally is also countable.
Exercise
Prove Theorem (that the Haar system is a complete orthonormal sequence in ).