« previous | Thursday, October 4, 2012 | next »
Substitute lecture
Chapter 3.4: Basis and Dimension
Definitions
We say that vectors are linearly independendent if
We say that span if any can be written as a linear combination of .
Let be a vector space.
The set is called a basis in if
- are linearly independent
- They span
The number of vectors in the basis is called the dimension of
Theorem
All bases in have the same number of elements called the dimension of .
Note: it is possible for a "finitely defined" Vector space to have infinite dimension.
Theorem 3.4.1
If is a spanning set of , , then any vectors are linearly independent.
Proof. (substitution/elimination process).
If , is a 0 vector, then there is nothing to prove:
If , then . At least one of the 's is nonzero. Without loss of generality, we assume that it is which is nonzero.
Therefore , so is a spanning set for .
Repeat the process with the above spanning set and . When we can no longer continue, we end up with as a spanning set for , and
Example
has bases , and , .
Let , , and
Note: is a spanning set of .
…
Corollary 3.4.2
If and are two bases in , then .
Proof. Assume 's are spanning set and , then from Thm. 3.4.1, we know that are linearly dependent. Not a basis (contradiction)
Flip and by same logic → contradiction. Therefore .
Infinite Dimension
If the vector space has no basis of finitely many vectors, we say that has infinite dimension.
If , we say that
Example
Find the dimension of the space of solutions to the system
Three variables, two constraints:
Note that the solutions form a vector space, and the set of solutions will be the nullspace of the coefficient matrix .
The solution is of the form , so the basis has only one vector. Therefore
Example
Find the dimension of the nullspace for the operation over
In other words, does . For a polynomial of degree ≤ 1, . If , then , and is anything. Therefore, , the basis is , and .
Example
A basis in would be:
Theorem 3.4.3
Let be a vector space, and . Then
- any linearly independent vectors span
- any vectors that span are linearly independent.
Proof.
- assume they do not span : there is a vector which is not a linear combination of 's, then would be linearly independent. this contradicts #Theorem 3.4.1 since there are vectors with only in a basis.
- assume they are linearly dependent: then , where are not all 0. This means that one vector could be written as a linear combination of other vectors, so vectors would span . This contradicts the definition of a basis, stating that vectors must be linearly independent.
Theorem 3.4.4
- No set of fewer than vectors spans
- Any linearly independent vectors could be extended (by more vectors) to form a basis.
- If , , span , we can pare them down to vectors, which would be a basis.