« previous | Tuesday, November 20, 2012 | next »
Least Squares Problem
given a
matrix
with rank
, the least-squares solution to the equation
is given by
and has the unique solution
the vector
in the range of
is closest to
.
Assume the columns of
constitute an orthonormal basis in
.
Theorem 5.5.6
If the column vectors of
form an orthonormal set of vectors in
, then
and the solution to the least squares problem is
Proof
We need to show that
.
If
, then
Therefore, the diagonal entries for
will be 1, and all other matrix cells will be 0.
Q.E.D.
Theorem 5.5.7
Let
be a subspace in
and let
.
Let
and
be an orthonormal basis for
.
If
where
for all
, then
.
Proof
Show that
for any
:
Theorem 5.5.8
Under hypothesis of #Theorem 5.5.7,
is the element of
that is closest to
. That is,
for all
in
.
Corollary 5.5.9
Let
be a nonzero subspace in
and let
.
If \{\vec{u}_1, \ldots, \vec{u}_k \} is an orthonormal basis for
and
, then the projection
of
onto
is given by \vec{p} = UU^T \vec{b}</math>
Proof
Let
be the projection matrix for
.
Going back to the formula for the least squares solution,
,
is unique to
regardless of basis used. However,
can only be used for orthonormal bases; other calculations must use
Example
Find best least squares approximation of
on [0,1] by a linear function (subspace of C[0,1])
- Space: C[0,1]
- Inner product:

- Subsace:
(linear functions)
Find
such that
:
Therefore
, and
forms an orthogonal basis for
.
Find orthonormal basis for
:
Therefore
forms an orthonormal basis for
Find least squares approximation:
Fourier Approximation
The last formula is the equation for harmonic motion.
Gram-Schmidt Process
For
, and
basis in
,
How do we obtain
, an orthonormal basis for
so that
?
Theorem 5.6.1
Let
be a basis for
. Let
and define
recursively by
Where
is the projection of
onto
.
The set
is an orthonormal basis for
.