Lecture 18: Best Approximant

1.Best approximants

Interpolation of data 𝒟={(xi,yi=f(xi)),i=0,,n} by an approximant p(t) corresponds to the minimization problem

minp||f-p||,

in the discrete one-norm at the sample points xi

||f||=||𝒇||1=i=0n|f(xi)|.

Different approximants are obtained upon changing the norm.

Theorem (Existence of best approximant. For any element fF in a normed vector space =(F,S,+,), there exists a best approximant gG within a finite dimensional subspace GF that is a solution of

mingG||f-g||.

The argument underlying the above theorem is based upon constructing the closed and bounded subset of G

K={gG|.||g-f||||0-f||=||f||}G.

Since G is finite dimensional, K is compact, and the continuous mapping g||g-f|| attains is extrema.

The two main classes of approximants g of real functions f:[a,b] that arise are:

Approximants based upon sampling

The vectors 𝒇=f(𝒙),𝒈=g(𝒙) are constructed at sample points 𝒙m and the best approximant solves the problem

mingG||𝒇-𝒈||.

Note that the minimization is carried out over the members of the subset G, not over the vectors 𝒈. The norm can include information on derivatives as in the norm

||f||H=||𝒇||1+||𝒇'||1,

arising in Hermite interpolation.

Approximants over the function domain

The norm is now expressed through an integral such as the p-norms

||f||p=(ab|f(t)|pdt)1/p.

In general, the best approximant in a normed space is not unique. However, the best approximant is unique in a Hilbert space, and is further characterized by orthogonality of the residual to the approximation subspace.

Theorem (Best Approximant in Hilbert space). For any element fF in a Hilbert space =(F,S,+,), there exists a unique approximant gG within a finite dimensional subspace GF that is a solution of

mingG||f-g||,

and the residual f-g is orthogonal to G, hG

(f-g,h)=0.

Note that orthogonality of the residual (f-g,h)=0 implies (f,h)=(g,h) or that the best approximant is the projection of f onto G.

2.Two-norm approximants in Hilbert spaces

For Hilbert spaces with a norm is induced by the scalar product

||f||=(f,f)1/2,

finding the best approximant reduces to a problem within m (or m). Introduce a basis ={b1,b2,} for such that any fF has an expansion

f(t)=j=1fjbj(t),fj=(f,bj)

Since G is finite dimensional, say n=dim(G), an approximant has expansion

g(t)=j=1ngjbs(j)(t).

Note that the approximation may lie in an arbitrary finite-dimensional subspace of . Choosing the appropriate subset through the function s: is an interesting problem in itself, leading to the goal of selecting those basis functions that capture the largest components of f, i.e., the solution of

min𝒔nj=1n|(f,bs(j))|.

Approximate solutions of the basis component selection are obtained by processes such as greedy approximation or clustering algorithms. The approach typically adopted is to exploit the Bessel inequality

i=1nfs(i)2||f||2,

and select

s(1)=argmaxiSfi2,

eliminate s(1) from S, and search again. The kth-step is

s(k)=argmaxiSfi2,

with Sk=S-{s(1),,s(k-1)}.

Assuming s(j)=j, the orthogonality relation f-gG leads to a linear system

(f-g,bi)=0(j=1ngjbj,bi)=j=1n(bi,bj)gj=(f,bi)𝑩𝒈=𝒇.

If the basis is orthonormal, then 𝑩=𝑰, and the best approximant is simply given by the projection of f onto the basis elements. Note that the scalar product need not be the Euclidean discrete or continuous versions

(f,g)=i=1nfigi,(f,g)=abf(t)g(t)dt.

A weighting function may be present as in

(f,g)=𝒇T𝑾𝒈,(f,g)=abf(t)g(t)w(t)dt,

discrete and continuous versions, respectively. In essense the appropriate measure μ(t) for some specific problem

dμ(t)=w(t)dt,

arises and might not be the Euclidean measure w(t)=1.

3.Inf-norm approximants

In the vector space of continuous functions defined on a topological space X (e.g., a closed and bounded set in n), a norm can be defined by

||f||=maxxX|f(x)|,

and the best approximant is found by solving the problem

infgG||f-g||=infgGmaxxX|f(x)-g(x)|.

The fact that g is the best approximant of f can be restated as 0 being the approximant of f-g since

||f-g-0||||f-(g+h)||.

A key role is played by the points where f(x)=g(x) leading to the definition of a critical set as

crit(f)=𝒵(f)={xX:|f(x)|=||f||}.

When G=Pn-1, the space of polynomials of degree at most n-1, with dimPn-1=n, the best approximant can be charaterized by the number of sign changes of f(x)-g(x).

Theorem (Chebyshev Alternation). The polynomial pPn-1 is the best approximant of f:[a,b] in the inf-norm

||f-p||=maxaxb|f(x)-p(x)|

if and only if there exist n+1 points ax0<x1<<xnb such that

f(xi)-p(xi)=s(-1)i||f-p||,

where |s|=1.

Recall that choosing xi=cos[(2i-1)π/(2n)], the roots of the Tn(θ)=cos(nθ) Chebyshev polynomial (with x=cosθ, a=-1, b=1), leads to the optimal error bound in polynomial interpolation

|f(t)-p(t)|||f(n+1)||(n+1)!2n.

The error bound came about from consideration of the alternation of signs of p(xj)-q(xj) at the extrema of the Chebyshev polynomial Tn, xi=cos(iπ/n), i=0,1,n, with p,q monic polynomials. The Cebyshev alternation theorem generalizes this observation and allows the formulation of a general approach to finding the best inf-norm approximant known as the Remez algorithm. The idea is that rather than seeking to satisfy the interpolation conditions

𝑴𝒂=𝒚

in the monomial basis

𝑴=n-1(x)=[ 𝟏 𝒙 𝒙n-1 ]n×n,

attempt to find n alternating-sign extrema points by considering the basis set

𝑹=n(𝒙)=[ 𝟏 𝒙 𝒙n-1 ±𝟏](n+1)×(n+1)

with ±𝟏=[ +1 -1 +1 ].

Algorithm (Remez)

  1. Initialize 𝒙n+1 to Chebyshev maxima on interval [a,b]

  2. Solve 𝑹𝒄=f(𝒙) (𝒙), 𝒄T=[ 𝒂T cn+1 ], 𝒂n

  3. Find the extrema 𝒚 of p(t)-f(t) with p(t)=a0+a1t++an-1tn-1

  4. If p(yi)-f(yi) are approximately equal in absolute value and of opposite signs, return 𝒙

  5. Otherwise set 𝒙=𝒚, repeat