Abstract

The purpose of this chapter is to introduce key structural concepts that are needed for theoretical transform analysis and are part of the common language of modern signal processing and computer vision. One of the great insights of this approach is the recognition that natural abstractions which occur in analysis, algebra and geometry help to unify the study of the principal objects which occur in modern signal processing. Everything in this book takes place in a vector space , a linear space of objects closed under associative, distributive and commutative laws. The vector spaces we study include vectors in Euclidean and complex space and spaces of functions such as polynomials, integrable functions, approximation spaces such as wavelets and images, spaces of bounded linear operators and compression operators (infinite dimensional). We also need geometrical concepts such as distance and shortest (perpendicular) distance, and sparsity. This chapter first introduces important concepts of vector space and subspace which allow for general ideas of linear independence, span and basis to be defined. Span tells us for example, that a linear space may be generated from a smaller collection of its members by linear combinations. Thereafter, we discuss Riemann integrals and introduce the notion of a normed linear space and metric space. Metric spaces are spaces, nonlinear in general, where a notion of distance and hence limit makes sense. Normed spaces are generalizations of “absolute value” spaces.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call