The monomial basis for the vector space of all polynomials , and its derivatives (Lagrange, Newton, ) allow the definition of an approximant for real functions , e.g., for smooth functions . A different approach to approximation in infinite-dimensional vector spaces such as or is to endow the vector space with a scalar product and associated norm . The availability of a norm allows definition of convergence of sequences and series.
All Hilbert spaces have orthonormal bases, and of special interest are bases that arise Sturm-Liouville problems of relevance to the approximation task.
The space of periodic, square-integrable functions is a Hilbert space ( is the only Hilbert space among the function spaces), and has a basis
that is orthonormal with respect to the scalar product
An element can be expressed as the linear combination
An alternative orthonormal basis is formed by the exponentials
with respect to the scalar product
The partial sum
has coefficients determined by projection
that can be approximated by the Darboux sum on the partition
with
denoting the root of unity. The Fourier coefficients are obtained through a linear mapping
with , and with elements
The above discrete Fourier transform can be seen as a change of basis from the basis in which the coefficients of are to the basis in which the coefficients are .
Carrying out the matrix vector product directly would require operations, but the cyclic structure of the matrix arising from the exponentiation of can be exploited to reduce the computational effort. Assume and separate even and odd indexed components of
Through the above, the matrix-vector product is reduced to two smaller matrix-vector products, each requiring operations. For , recursion of the above procedure reduces the overall operation count to , or in general for composed of a small numer of prime factors, . The overall algorithm is known as the fast Fourier transform or FFT.
One step of the FFT can be understood as a special matrix factorization
where is diagonal and is the even-odd permutation matrix. Though the matrix is full (all elements are non-zero), its factors are sparse, with many zero elements. The matrix is said to be data sparse, in the sense that its specification requires many fewer than numbers. Other examples of data sparse matrices include:
has constant diagonal terms, e.g., for
or in general the elements of can be specified in terms of numbers through .
Rank-1 updates arising in the singular value or eigenvalue decompositions have the form
and the components of are suficient to specify the matrix with components. This can be generalized to any exterior product of matrices , through
The components of are specified through only components of .
The relevance to approximation of functions typically arises due basis sets that are solutions to Sturm-Liouville problems. In the case of the Fourier transform are eigenfunctions of the Sturm-Liouville problem
with eigenvalues . The solution set to a general Sturm-Liouville problem to find
form an orthonormal basis under the scalar product
and approximations of the form
and Parseval's theorem states that
read as an equality between the energy of and that of . By analogy to the finite-dimensional case, the Fourier transform is unitary in that it preserves lengths in the norm with weight function .
The bases arising from Sturm-Liouville problems are single-indexed, giving functions of increasing resolution over the entire definition domain. For example resolves ever finer features over . When applied to a function with localized features, must be increased with increased resolution in the entire domain. This leads to uneconomical approximation series with many terms, as exemplified by the Gibbs phenomenon in approximation of a step function, for , and . The approach can be represented as the decomposition of a space of functions by the direct sum
with , for example
with for the Fourier series.
Approximation of functions with localized features is more efficiently accomplished by choosing some generating function and then defining a set of functions through translation and scaling, say
Such systems are known as wavelets, and the simplest example is the step function
with having support on the half-open interval . The set is known as an Haar orthonormal basis for since
Approximations based upon a wavelet basis
allow identification of localized features in .
The costly evaluation of scalar products in the double summation can be avoided by a reformulation of the expansion as
(1) |
with . In addition to the (“mother” wavelet), an auxilliary scaling function (“father” wavelet) is defined, for example
for the Haar wavelet system.
The above approach is known as a multiresolution representation and is based upon a hierarchical decomposition of the space of functions, e.g.,
with
The hierarchical decomposition is based upon the vector subspace inclusions
and the relations
that state that the orthogonal complement of within is . Analogous to the FFT, a fast wavelet transformation can be defined to compute coefficients of (1).