« previous | Thursday, November 8, 2012 | next »
Least Squares Problem
Given subspace
and a vector
, find the closest approximation
.
is a vector projection, and
is the α-scalar projection.
When represented by
,
is the column space of
, and
is the vector such that
. We get
is the residual vector.
Normal Equation
Theorem 5.3.2
is a
matrix of rank
(same rank as number of columns). Then the normal equation
has a unique solution given by
And
is the unique least squares solution to the sysetm
.
Proof
Based on premise that
is nonsingular.
Assume we have some vector
. For
to be nonsingular,
must be the only solution.
We know that


- and

Therefore
.
Q.E.D.
Corollary
We already know that
, and
, so
The projection matrix
is interesting because
:
Example
Overdetermined system
Regression
Given a set of measurements
at points
, each set of values defines a point at
.
Linear Regression
Find equation such that
that approximates the system of equations
Solution for
is given by
Inner Product Spaces
Vector space
. Suppose we have a function such that for all
, the inner product of
(notation
) is a real number.
We want to have the following properties:
and is equal to 0 iff 
(commutativity)
and the same applies for the second component.
Example 1
,
is the scalar product.
Given
weights such that
,
Example 2
Given two matrices
, let
.
Example 3
Given two functions
, let
and 
- …
Properties
Given a vector space
and an inner product function
We can redefine:
- length / norm:

- Orthogonality:

- Scalar projection:

- Vector projection:

Theorem 5.4.1
The Pythagorean Law
If
, then
Proof
Orthogonality of Functions
, where the inner product is defined as
Theorem 5.4.2
The Cauchy-Schwarz Inequality
Holds iff
and
are linearly dependent.