The null space of a linear mapping represented through matrix is defined as , the set of all points that have a null image through the mapping. The null space is a vector subspace of the domain of the linear mapping. A first step in the study of nonlinear mappings is to consider the generalization of the concept of a null set, starting with the simplest case,
(1) |
where , , , i.e., has continuous derivatives. It is assumed that a closed form analytical solution is not available, and algorithms are sought to construct an approximating sequence whose limit is a root of (1). The general approach is to replace (1) with
(2) |
where is some approximation of , and the root of (2) can be easily determined.
(3) |
The solution of (3) is
an iteration known as the secant method. The error in root approximation is
and can be estimated by Taylor series expansions around the root for which ,
where derivatives are assumed to be evaluated at . In the result
assuming , (i.e., is a simple root) gives
For small errors, to first order the above can be written as
Assuming -order convergence of ,
leads to
Since are finite while , the above asymptotic relation can only be satisfied if
hence the secant method exhibits superlinear, but subquadratic convergence.
which is given in Newton form as
with root
(4) |
an iteration known as the Newton-Raphson method. The error is given by
(5) |
Taylor series exapnsion around the root gives for small ,
The resulting expression
(6) |
states quadratic convergence for Newton's method. This faster convergence than the secant method requires however knowledge of the derivative, and the computational expense of evaluating it.
The above estimate assumes convergence of , but this is not guaranteed in general. Newton's method requires an accurate initial approximation , within a neighborhood of the root in which is increasing, , and convex, . Equivalently, since roots of are also roots of , Newton's method converges when . In both cases (6) in the prior iteration states that , hence . Since is increasing , hence (5) implies . Thus the sequence is decreasing and bounded below by zero, hence , and Newton's method converges.
An immediate extension of the above approach is to increase the accuracy of the approximant by seeking a higher-degree polynomial interpolant. The expense of the resulting algorithm increases rapidly though, and in practice linear and quadratic approximants are the most widely used. Consider the Hermite interpolant based on data
given in Newton form as
with roots
Tha above exhibits the difficulties arising in higher-order interpolants. The iteration requires evaluation of a square root, and checking for a positive discriminant.
exhibits cubic convergence.
The secant iteration
in the limit of recovers Newton's method
This suggests seeking advantageous approximations of the derivative
based upon some step-size sequence . Since , the choice suggests itself, leading to Steffensen's method
Steffensen's method exhibits quadratic convergence, just like Newton's method, but does not require knowledge of the derivative. The higher order by comparison to the secant method is a direct result of the derivative approximation
which, remarkably, utilizes a composite approximation
Such composite techniques are a prominent feature of various nonlinear approximations such as a -layer deep neural network .
The above iterative sequences have the form
and the root is a fixed point of the iteration
For example, in Newton's method
and indeed at a root . Characterization of mappings that lead to convergent approximation sequences is of interest and leads to the following definition and theorem.
The fixed point theorem is an entry point to the study of non-additive approximation sequences.
Example
is expressed recursively as
and has the limit
that is the fixed point of ,
Over the interval , is a contraction since
Example
is expressed recursively as
and has the limit
that is the fixed point of ,
Over the interval , is a contraction since