« previous | Tuesday, November 6, 2012 | next »
Orthogonality
Subspaces 
: 
If 
, then take 
 for 
: 
 is not orthogonal to 
.
, 
 is orthogonal complement. For example, a plane and a normal vector.
Range
 for 
 matrix 
 and 
 is defined as
For transpose matrix, 
 Note: Range is nothing more than the column space of a matrix
Theorem 5.2.1
Fundamental subspaces theorem

 

Proof
Prove one, then the proof of the second follows from the first:
Let 
, then 
Example

 

Theorem 5.2.2
If 
 is a subspace of 
, then 
Furthermore, if 
 is a basis for 
 and 
 is a basis for 
, then 
 is a basis for 
Proof
If 
 and 
 is a basis for 
, then 
.
Let 
 be a 
 matrix formed by using the basis vectors as rows of 
. The rank of 
 is 
, and 
.
 by equation 1 of the previous theorem, so 
Therefore 
. This proves the first part of the theorem.
Check linear independence of 
s to determine whether it is a valid basis of 
.
In order for 
 to be true, 
 and 
 must be elements of 
 Since 
 and 
 are orthogonal subspaces, 
, so 
.
Direct Sum
If 
 are subspaces of a vector space 
, and each 
 can be written as a sum 
, where 
 and 
, then 
 is a direct sum of 
 and 
, written 
Theorem 5.2.3
If 
 is a subspcae of 
, then 
. In other words (or lack thereof):
Proof
Let 
 be a basis for 
, then
This must be unique since 
.
Theorem 5.2.4
Example
Find basis for 
, 
, 
, and 
Therefore, 
 is a basis for 
.
, so 
 is basis for 
.
Repeat above steps for 
Section 5.3: Least Squares
Find best approximation of 
 (outside of a subspace) using vector 
 (in subspace)
Theorem 5.3.1
Let 
 be a subspace.
For each 
, there is a unique element 
 of 
 that is closest to 
, i.e.
 for any 
Furthermore, 
Definition: Residual Vector
A vector 
 is a solution to the least squares problem 
 iff 
 is the vector in 
 that is closest to 
.
Thus we know that 
 is the projection of 
 onto 
, where 
 is the residual vector.
Thus 
 is a solution of the least squares problem iff 
.