« previous | Thursday, April 17, 2014 | next »
Suppose we want to find a linear approximation
for the points
:
Hence the coefficients to our control points
and
are
and our values are
We compute
and solve the following for
:
Hence our equation is:
General Case
for
,
is a subspace spanned by
, where
's are linearly independent.
We wish to approximate
by
. We do so by minimizing the value of
for all
. This is called the best least squares fit

is just the projection of

onto

.
Normal equations are:
Residual (or rejection) belongs to the subspace perpendicular to
, or
.
Proof. If
is the projection of
onto a subspace, then
is the shortest vector from
to
. The shortest vector from
to
is orthogonal to
(otherwise
would not be minimized). Hence
is orthogonal to
.
Take
and compute
Theus
minimizes the value of
, and thus
quod erat demonstrandum
Special Case
Let
be spanned by an orthonormal basis
, and let
be spanned by an orthonormal basis
. Thus
.
Now
is given by
.
To find the projection, we just take the components that are members of
: