Download Definition of a Vector Space A collection of vectors: V , scalars for

Document related concepts

Determinant wikipedia , lookup

Bivector wikipedia , lookup

Non-negative matrix factorization wikipedia , lookup

Orthogonal matrix wikipedia , lookup

Cross product wikipedia , lookup

Perron–Frobenius theorem wikipedia , lookup

Cayley–Hamilton theorem wikipedia , lookup

Jordan normal form wikipedia , lookup

Exterior algebra wikipedia , lookup

Eigenvalues and eigenvectors wikipedia , lookup

Singular-value decomposition wikipedia , lookup

Matrix multiplication wikipedia , lookup

Laplace–Runge–Lenz vector wikipedia , lookup

System of linear equations wikipedia , lookup

Gaussian elimination wikipedia , lookup

Euclidean vector wikipedia , lookup

Vector space wikipedia , lookup

Matrix calculus wikipedia , lookup

Covariance and contravariance of vectors wikipedia , lookup

Four-vector wikipedia , lookup

Transcript
Definition of a Vector Space
A collection of vectors: V , scalars for scaling : R;
A vector addition: +, A scalar multiplication: ·
for vector addition:
(i) (closure of vector addition) u + v is always in V .
(ii) (commutative law) u + v = v + u.
(iii) (associative law) (u + v) + w = u + (v + w).
(iv) (existence of zero vector) has 0 s.t. u + 0 = u.
(v) (existence of negative vector) for each u ∈ V , there is a
vector −u s.t. u + (−u) = 0.
1
...continued for scalar multiplication:
(vi) (closure of scalar multiplication) cu is always in V .
(vii) (distributive law) c(u + v) = cu + cv.
(viii) (distributive law) (c + d)u = cu + du.
(ix) (compatibility) c(du) = (cd)u.
(x) (normalization) 1u = u.
Def: A non-empty set of vectors V with vector addition
“+” and scalar multiplication “·” satisfying all the above
properties is called a vector space over R.
2
Facts: (i) The zero vector 0 so defined is unique.
0=0+w =w
(ii) The negative vector for each u is unique.
−u = −u + 0 = −u + (u + w)
= (−u + u) + w = 0 + w = w
We define vector subtraction u − v := u + (−v).
(iii) c0 = 0.
(iv) 0u = 0.
(v) −u = (−1)u.
3
Examples of Common Vector Spaces
• {0}, zero vector space.
• Rn : ordered n-tuples of real numbers with entry-wise
vector addition and scalar multiplication.


u1 + v1
..

u+v =
,
.
un + vn
Blue-print for vector space.
4


cu1
.
cu =  .. 
cun
• S, the doubly infinite sequences of numbers:
{yk } = (. . . , y−2 , y−1 , y0 , y1 , y2 , . . .)
with component-wise addition and scalar multiplication.
{yk } + {zk } = (. . . , y−2 , y−1 , y0 , y1 , y2 , . . .)
+ (. . . , z−2 , z−1 , z0 , z1 , z2 , . . .)
:= (. . . , y−1 + z−1 , y0 + z0 , y1 + z1 , . . .)
c{yk } := (. . . , cy−2 , cy−1 , cy0 , cy1 , cy2 , . . .)
zero vector: {0}, sequence of zeros.
5
• Pn : collection of polynomials with degree at most equal
to n and coefficients chosen from R:
p(t) = p0 + p1 t + . . . + pn tn ,
with polynomial operations.
→ “vector addition”: “p(t) + q(t)” is defined as:
p(t) + q(t) = (p0 + q0 ) + (p1 + q1 )t + . . . + (pn + qn )tn ;
→ “scalar multiplication”: “cp(t)” is defined as:
cp(t) = (cp0 ) + (cp1 )t + . . . + (cpn )tn .
→ “zero vector”: the zero polynomial 0(t) ≡ 0.
• P: all polynomials.
6
• The collection of all real-valued functions on a set D, i.e.
f : D → R. (D usually is an interval.)
→ “vector addition”: the function “f + g” is defined as:
(f + g)(x) = f (x) + g(x) for all x in D.
→ “scalar multiplication”: the function “cf ” is defined
as:
(cf )(x) = cf (x) for all x in D.
→ “zero vector”: the zero function 0(x) which sends
every x in D to 0, i.e.
0(x) = 0 for all x in D.
7
• Mm×n : collection of all m × n matrices with entries in
R with matrix addition and scalar multiplication.
(
)
A + B = aij + bij ,
(
cA = caij .
zero vector: m × n zero matrix Om×n .
. . . and many more.
8
)
Subspace: A subspace H of V is a non-empty subset of V
such that H itself forms a vector space under the same vector
addition and scalar multiplication induced from V .
Checking: Following conditions are automatically valid:
(ii) u + v = v + u.
(iii) (u + v) + w = u + (v + w).
(vi) c(u + v) = cu + cv.
(viii) (a + b)u = au + bu.
(ix) (ab)u = a(bu).
(x) 1u = u.
9
Need to verify the remainings:
(i) the sum u + v is always in H.
(iv) there is a zero vector 0 in H.
(v) for each u ∈ H, there is a negative vector −u ∈ H.
(vi) the scalar multiple cu is always in H.
[Same zero vector, negative vector for H.]
.... but we know ...
• −u = (−1)u for any u.
10
Alternative Definition: H is a subspace of V if:
1. 0 of V is in H.
2. sum of two vectors in H is again in H.
3. the scalar multiple of a vector in H is again in H.
In fact, (2) & (3) can be combined together as a single checking condition:
Thm: H is a subspace of V iff H contains 0 and:
u, v ∈ H
⇒
au + bv ∈ H
11
for all scalars a, b.
Examples of Subspaces
1. V is a subspace of itself.
2. {0} is a subspace of V , called the zero subspace.
3. R2 is not a subspace of R3 , strictly speaking.
3′ . The following subset of R3 is a subspace of R3 :
 

 s

H =  t  : s, t in R


0
4. Pn is a subspace of P;
Pn is a subspace of Pm if n ≤ m.
12
5. The collection of n × n symmetric matrices H is a subspace of Mn×n .
5′ . The collection of n × n skew-symmetric matrices is also
a subspace of Mn×n .
6. The collection Ve of even functions from R to R:
f (−x) = f (x) for all x ∈ R
and the collection Vo of odd functions:
f (−x) = −f (x)
for all x ∈ R
are both subspaces of the vector space of functions.
13
Linear Combination and Span
Similar to vectors in Rn , we define:
Def: Let S = {v1 , . . . , vk } be a collection of vectors in V
and let c1 , . . . , ck be scalars. The following y is called a
linear combination (l.c.) of vectors in S:
y = c1 v1 + . . . + ck vk .
Note: When S contains infinitely many vectors, a l.c. of
vectors in S means a l.c. of a finite subset of vectors in S.
Rmk: Sum of infinite number of vectors is not defined (yet).
[We need a concept of “limit” to do this.]
14
Example: In function space, let S = {sin2 x, cos2 x}. Then
the constant function 1(x) ≡ 1 is a l.c. of “vectors” in S:
1(x) = 1 · sin2 x + 1 · cos2 x,
for all x in D.
Example: In polynomial space, let S = {1, t, t2 , t3 , . . .}.
Then any polynomial p(t) is a l.c. of “vectors” in S.
p1 (t) = 1 + t + t3 , S1 = {1, t, t3 }
p2 (t) = 3t − 5t100 , S2 = {t, t100 }
p3 (t) = 0, S3 = {1}
15
Exercise: Consider the function space.
Is sin3 x a l.c. of S = {sin x, sin 2x, sin 3x}?
Exercise: Consider the function space.
Is cos x a l.c. of S = {sin x, sin 2x, sin 3x}?
***
16
Span of a Collection of Vectors
Def: Let S = {v1 , . . . , vk }. The collection of all possible l.c.
of vectors in S is called the span of S:
Span S := {c1 v1 + . . . + ck vk | c1 , . . . , ck are scalars.}.
Note: When S contains infinitely many vectors, Span S is
the collection of all possible l.c. of any finite subset of vectors
in S.
Example: In P, consider S = {1, t2 , t4 , . . .}. Then
Span S = collection of polynomials with only even powers.
17
Thm: Let v1 , v2 be vectors in V . Then H = Span {v1 , v2 }
is a subspace of V .
Checking: (1): 0 = 0v1 + 0v2 , so 0 ∈ H.
(2): Let u, v ∈ H, i.e. both are l.c. of {v1 , v2 }:
u = s1 v1 + s2 v2 ,
v = t1 v1 + t2 v2 ,
for some suitable numbers s1 , s2 , t1 , t2 . Then:
u + v = (s1 + t1 )v1 + (s2 + t2 )v2
is again a l.c. of {v1 , v2 }, and thus collected by H.
18
(3): Let u ∈ H, c a number. Then:
cu = (cs1 )v1 + (cs2 )v2
is a l.c. of {v1 , v2 }, and hence collected by H.
As conditions (1),(2),(3) are all valid, H is a subspace of V .
The above proof allows obvious generalization to:
Thm 1 (P.210): For any v1 , . . . , vp ∈ V , the collection of
vectors H = Span {v1 , . . . , vp } is a subspace of V .
Note: We will call H to be the subspace spanned (or generated) by {v1 , . . . , vp }, and {v1 , . . . , vp } is called a spanning
set (or generating set) of H.
19
Subspaces of a Matrix
Let A be an m × n matrix. Associated to this matrix A,
we define:
1. The null space of A, denoted by Nul A;
2. The column space of A, denoted by Col A;
3. The row space of A, denoted by Row A.
Note that:
• Nul A and Row A will be subspaces of Rn ;
• Col A will be a subspace of Rm .
20
The Null Space of a Matrix
Def: Let A be an m × n matrix. The null space of A is:
Nul A := {x | x in Rn and Ax = 0}.
i.e. the solution set of the homogeneous system Ax = 0.
• size of each solution: no. of columns in A.
so if A is of size m × n, Nul A is inside Rn .
Example: Does v belong to Nul A?
[
1 1
A=
0 −3
]
−1
,
2
21
 
1
v = 2
3
Thm 2 (P.215): Nul A is a subspace of Rn .
Checking: (1) A0 = 0, so 0 is in Nul A.
(2) Let Au = 0 = Av. Then consider u + v:
A(u + v) = Au + Av = 0 + 0 = 0,
so u + v ∈ Nul A.
(3) Let Au = 0, c any number. Then consider cu:
A(cu) = c(Au) = c0 = 0,
so cu ∈ Nul A.
22
Exercise: Describe the null space of A:


1 2 2 1
A =  2 5 10 3  .
1 3 8 2
***
From the above exercise, obviously we will have:
Thm: Nul A = {0} when Ax = 0 has unique solution. When
Ax = 0 has {x1 , . . . , xk } as a set of basic solutions, we have
Nul A = Span{x1 , . . . , xk }.
23
The Column Space of a Matrix
Def: Let A = [ a1 . . .
column space of A is:
an ] be an m × n matrix. Then the
Col A := Span {a1 , . . . , an }.
Fact: Col A is a subspace of Rm .
Exercises: Check if v ∈ Col A:


 
1 2 1 2
1
A =  2 4 0 6  (i) v =  2 
1 2 2 1
1
***
24
 
1
(ii) v =  4  .
7
Thm: v ∈ Col A ⇔ [ A | v ] is consistent.
Example: Find a condition

1
2
 −1 −1
A=
3
2
4
0
Sol: Consider the

1
2
5
 −1 −1 −3

3
2
7
4
0
4
on v such that v ∈ Col A.:

 
5
y1
−3 
 y2 
 v =  .
7
y3
4
y4
augmented matrix [ A | v ]:



| y1
1 2
5
|
y1
| y2 
2
| y2 + y1 
0 1
→



| y3
0 −4 −8 | y3 − 3y1
| y4
0 −8 −16 | y4 − 4y1
25

1
 −1

3
4


2
5 | y1
1 2 5
−1 −3 | y2 
0 1 2
→
2
7 | y3
0 0 0
0
4 | y4
0 0 0

|
y1
|
y1 + y2


| y1 + 4y2 + y3
| 4y1 + 8y2 + y4
For the system to be consistent, we need:
{
y1 + 4y2 + y3 = 0
4y1 + 8y2 + y4 = 0
26
Remark: When A is representing a matrix transformation
T : Rn → Rm , we have:
v ∈ range of T
↔
can find x ∈ Rn such that T (x) = v
↔
[ A | v ] is consistent.
Therefore, Col A is the same as the range of T in this case,
as:
Range of T := {v ∈ Rm | T (x) = v for some x ∈ Rn }.
Thm: Col A = Rm iff every row of A has a pivot position.
(i.e. T : x 7→ Ax is onto.)
27
The Row Space of a Matrix
Identify:
Row ↔ Column by taking transpose


r1
.
Def: Let A =  ..  be an m × n matrix. The row space of
rm
A is:
Row A := Span{rT1 , . . . , rTm }.
i.e. Row A = Col AT .
Note: Row A is a subspace of Rn .
28
Let A → B by an ERO, e.g.




r1
−3r2 + r1
r2
=
A =  r2  → 
..
..
.
.

r′1
 r′2

..
.


 = B.
Then r′1 , r′2 , . . . ∈ Row A. So:
T
T
Row B ⊆ Row A.
But B → A by the reverse ERO, then we also have:
Row A ⊆ Row B.
29
Thm: Let A → B by EROs. Then Row A = Row B.
Recall: A → B = P A where P (size: m × m) is invertible.
Thm: When P is invertible, Row A = Row (P A).
[
]
1 1 3
Example: Describe Row A when A =
.
2 2 4
Sol: By definition:
   
1
2
Row A = Span { 1  ,  2 }.
3
4
30
[
1
Example: Describe Row A when A =
2
1
2
3
.
4
Sol: By EROs:
[
1 1
A=
2 2
]
[
]
3
1 1 0
→
= B.
4
0 0 1
As Row A = Row B, we have:
   
1
0
Row A = Span { 1  ,  0 }.
0
1
31
]
Linear Transformations in general
Def: Let V , W be two vector spaces over R. A transformation T : V → W is called linear if
(a) T (u + v) = T (u) + T (v)
(b) T (cu) = cT (u)
for any choices of u, v in V and any choice of scalar c.
In the special case that V = W , i.e. T : V → V is linear, we
will call T to be a linear operator on V .
32
Def: Let T : V → W be linear. We define:
1. The kernel of T , denoted by ker T , to be:
ker T := {v ∈ V : T (v) = 0} .
(solution set of T (v) = 0.)
2. The image/range of T , denoted by Im T , to be:
Im T := {w ∈ W : T (v) = w for some v ∈ V } .
(which is the same as T (V ).)
Thm: ker T is a subspace of V . Im T is a subspace of W .
33
Remark: When V = Rn , W = Rm , i.e. T is given by a
matrix transformation: x 7→ Ax. Then:
ker T = Nul A and Im T = Col A.
Thm: T is one-to-one iff ker T = {0}.
T is onto iff Im T = W .
Pf: (1-1) As T is linear, we have:
T (x1 ) = T (x2 ) ⇔ T (x1 − x2 ) = 0.
So T is one-to-one ⇔ always have x1 = x2 ⇔ ker T = {0}.
For onto property, directly from definition.
34
Example: Consider V = W = P(R). The differential operator D will be a linear transformation:
d
D(p(t)) = p(t).
dt
We have ker D = Span {1} and Im D = P.
• Thus D is onto but NOT one-to-one.
Example: The integral operator Ia : P → P with lower
limit a:
∫ t
p(x)dx.
Ia (p(t)) =
a
will be one-to-one but NOT onto.
35
Linearly Independent/Dependent Sets
Def: An indexed set of vectors S = {v1 , . . . , vp } is called
linearly independent (l.i.) if the vector equation:
c1 v1 + . . . + cp vp = 0
has unique (zero) solution:
c1 = . . . = cp = 0.
Def: S is called linearly dependent (l.d.) if it is not l.i., i.e.
exist d1 , . . . , dp , not all zeros, such that:
d1 v1 + . . . + dp vp = 0.
36
Note: When S contains infinitely many vectors:
(i) S is said to be l.i. if every finite subset of S is always l.i.
(ii) S is said to be l.d. if there exists a finite subset of S
being l.d.
Example: In P2 (R), S = {1, 2t, t2 } is l.i.
Sol: Consider the “vector” equation:
c1 · 1 + c2 · 2t + c3 · t2 = 0(t) (as polynomials).
By equating coefficients of 1, t, t2 , we get c1 = c2 = c3 = 0.
Rmk: The above equation is a polynomial “identity”.
37
Example: In P2 , S = {1 + t, 1 − t2 , t + t2 } is l.d. since:
c1 (1 + t) + c2 (1 − t2 ) + c3 (t + t2 ) = 0(t)
has a non-zero solution (c1 , c2 , c3 ) = (1, −1, −1).
Example: In P, S = {1, t, t2 , t3 , . . .} is l.i.
Sol: Take any finite subset of S:
S ′ = {ti1 , ti2 , . . . , tik },
0 ≤ i1 < i2 < . . . < ik .
Obviously S ′ is always l.i., so S will be l.i.
38
Example: In function space, S = {sin2 x, cos2 x} l.i.
Sol: Consider the “vector” equation:
c1 sin2 x + c2 cos2 x = 0(x) (as functions).
Put x = 0, π2 , we obtain “number” equations:
{
c1 · 0 + c2 · 1 = 0
c1 · 1 + c2 · 0 = 0
from which we get c1 = c2 = 0 already. So S must be l.i.
39
Example: In function space, S = {1, sin2 x, cos2 x} is l.d.
Sol: We have non-trivial solution to the “vector” equation:
(−1) · 1 + (1) · sin2 x + (1) · cos2 x = 0(x),
which is true for all x ∈ D.
Exercise: Let c1 , c2 , c3 be distinct numbers. In function
space, is {ec1 x , ec2 x , ec3 x } l.i.?
***
40
Bases, Coordinates, and Dimensions
• When {v1 , . . . , vp } is l.d., we can remove one vj from
the set without changing its span.
e.g. Consider {v1 , v2 , v3 } with v2 = 3v1 − 4v3 .
the l.c. x = a1 v1 + a2 v2 + a3 v3 can be rewritten as:
x = a1 v1 + a2 (3v1 − 4v3 ) + a3 v3
= (a1 + 3a2 )v1 + (a3 − 4a2 )v3 .
which is a vector in Span {v1 , v3 }.
• If removing any vector from {v1 , . . . , vp } will change its
span, the set must be l.i.
41
Basis for a Subspace
Def: Let H be a subspace of V . An indexed set B of vectors
in H is called a basis for H if:
(i) B is linearly independent; and
(ii) Span B = H.
When H = {0}, we choose B = ϕ as the basis.
Remarks: (i) Since Span B = H, every vector in H can be
written as a l.c. of vectors in B.
(ii) Since B is l.i., removing any vector from it will make
the span smaller. So we can regard a basis as a “minimal
collection of vectors” from H that can span H.
42
Examples: (i) {e1 , . . . , en } is a basis for Rn .
[ ] [ ]
[
] [ ]
1
0
1
1
(ii) Both {
,
} and {
,
} are bases for R2 .
0
1
−1
1
(iii) {1, t, t2 , . . . , tn } forms a basis for Pn .
P has a basis {1, t, t2 , . . .}.
(iv) The following set of matrices is a basis for the vector
space of 2 × 2 matrices.
[
] [
] [
1 0
0 1
0
B={
,
,
0 0
0 0
1
43
] [
]
0
0 0
,
}.
0
0 1
(v) Check that B is a basis for R3 .

 
 

3
−4
−2
B = { 0  ,  1  ,  1 }.
−6
7
5
Sol: Consider the equation:






3
−4
−2
c1  0  + c2  1  + c3  1  = b.
−6
7
5
44





3 −4 −2
c1
b1
↔ 0
1
1   c2  =  b2  .
−6 7
5
c3
b3
We find that the coefficient matrix is invertible. So:
• B is linearly independent.
(all basic variables, unique solution for c1 , c2 , c3 .)
• Span B = Col A = R3 .
(every row of A has a pivot position.)
Thus, B is a basis of R3 .
Thm: Let [ a1 . . . an ] be an n×n invertible matrix. Then
B = {a1 , . . . , an } will form a basis for Rn .
45
Bases for Nul A
Express the general solution of Ax = 0 in parametric form:
x = s1 x1 + . . . + sk xk ,
where s1 , . . . , sk ∈ R.
So B = {x1 , . . . , xk } can span Nul A.
We note that B must be l.i., for example:






−s − t
−1
−1
x =  s  = s 1  + t 0 .
t
0
1
x = 0 iff s = t = 0.
46
Thm: When Ax = 0 has non-zero solutions, a set of basic
solutions B = {x1 , . . . , xk } will form a basis for Nul A.
When Ax = 0 has unique zero solution, i.e. Nul A = {0}, we
say that ϕ is the basis for Nul A.
Exercises: Find a basis for Nul A where:

[
]
1 1 2 2
1 2
(i) A =
, (ii) A =  2 2 1 1
2 1
3 3 2 2
***
47

3
3.
2
Bases for Row A
Recall when A → B by EROs, we have Row A = Row B. So
let B be a REF of A. Clearly the non-zero rows of B will
span Row B = Row A, and also they form a l.i. set, e.g.:
     

1
0
0
1 1 1 1
1 2 0
0 2 3 4
B=
 ↔ B = {  ,   ,  }
0 0 0 5
1
3
0
0 0 0 0
1
4
5
 
   
 
1
0
0
0
1
2
0 0
c1   + c2   + c3   =   ,
1
3
0
0
1
4
5
0

48
unique solution: c1 = c2 = c3 = 0.
Thm 13 (P.247): Let B be a REF of A. Then the non-zero
rows of B will form a basis for Row A.
Exercise: Find a basis for Row A:

1
1
A=
2
1
4
1
5
0
***
49

6 8
3 2 
.
9 10
2 0
Bases for Col A
We use the trick: Col A = Row AT .
Example: Find a basis for

1
2
A=
1
1
Sol: Perform EROs

1
2
T
A =
1
1
on
2
3
1
2
AT :
1
3
1
3
Col A:
2
3
3
2
1
1
1
1
1
2
.
3
1


1
1
2
0
→
1
0
1
0
50

2
1
0
0
1
0
1
0

1
0

0
0
A basis for Col A will be:
     
1
0
0
2 1 0
B = {  ,   ,  }.
1
0
1
1
0
0
Warning: Row operations on A will change Col A!.
• i.e. EROs needed to be performed on AT .
(then Row AT = Col A will not change.)
51
If we want to use only the columns from A to form such a
basis, we have the following result:
Thm 6 (P.228): The pivot columns of A will form a basis
for Col A.
Exercise: Find a basis for Col A, using columns from A:

1
2
A=
1
1
2
3
3
2
***
52
1
1
1
1

1
2
.
3
1
Idea of Proof of Thm 6:
1. A l.c. of columns of A:
c1 a1 + . . . + cn an = b
corresponds to a solution of Ax = b: xi = ci
2. Row operations [ A | b ] → [ A′ | b′ ] will not change this
solution, i.e. same dependence relation for new columns:
c1 a′1 + . . . + cn a′n = b′
3. In RREF(A) the pivot columns are automatically l.i.,
spanning Col RREF(A). These two properties can be
brought back to pivot columns in A by reverse EROs.
53
Recall: Studying V using a basis B
Basis: An indexed set B of vectors that is:
(a) linearly independent;
(b) spanning the space V , i.e. Span B = V .
Thm 7 (P.232): Let B = {b1 , . . . , bp } be a basis for a subspace H (or V ). Then for each x ∈ H, there exists a unique
set of scalars (c1 , . . . , cp ) such that:
x = c1 b1 + . . . + cp bp .
54
Proof: The existence of {c1 , . . . , cp } is guaranteed by condition (b). Now suppose that d1 , . . . , dp are numbers with the
same properties, i.e.
x = d1 b1 + . . . + dp bp .
Take the difference, then:
0 = (c1 − d1 )b1 + . . . + (cp − dp )bp .
As B is l.i. (condition (a)), ci − di = 0 for each i.
Hence the numbers c1 , . . . , cp are uniquely determined.
55
Coordinate Vector Relative to a Basis
Def: Let B = {b1 , . . . , bp } be a basis for H. The coordinate
vector of x relative to B (or B-coordinate vector of x) is the
vector in Rp formed by the numbers c1 , . . . , cp :

c1
.. 

[x]B := . .
cp

x = c1 b1 + . . . + cp bp
↔
Note that x is in H ⊂ V , but [x]B is in Rp .
56
[
] [
]
2
−2
Example: Let B = {
,
} be a basis for R2 . De−1
−1
[ ]
x1
scribe geometrically the B-coordinate vector of x =
.
x2
Sol: First we find [x]B :
[
x1
x2
]
[
= c1
]
[
2
−2
+ c2
−1
−1
]
↔

x1 − 2x2

 c1 =
4

 c2 = − x1 + 2x2
4
So the B-coordinate vector of x is:
[ ] [ x1 −2x2 ]
c1
4
[x]B =
=
x1 +2x2 .
c2
− 4
57
y
y
x
x
x2
c2 (−2,−1)
x1
c1 (2,−1)
x
(−2,−1)
x
(2,−1)
B-coordinate system
[ ]
c1
[x]B =
c2
original coordinate system
[ ]
x1
x=
x2
58
Example: Let B = {1, 1 + t, 1 + t + t2 } be a basis for P2 .
Consider p(t) = t − t2 . To find the B-coordinate vector of
p(t), we express p(t) as a l.c. of vectors in B:
t − t2 = (−1) + 2(1 + t) − (1 + t + t2 ).
Then the B-coordinate vector of p is:


−1
[p]B =  2  .
−1
59
Example: Let H = Span B where B = {1, cos x, cos 2x}.
It is easy to check that B is l.i., so it is a basis for H.
Let f (x) = cos2 x. Since f (x) = 12 +
1
1
2
cos 2x, we have:
2
[f ]B =  0  .
1
2
But if we take B ′ = {cos x, 1, cos 2x}, we have:
 
0
[f ]B′ =  21  .
1
2
So the ordering of vectors in a basis is important.
60

 
 

1
1
−1
Exercise: Let B = { 1  ,  −1  ,  1 } be a basis for
−1
1
1
R3 . Find [e1 ]B , [e2 ]B , and [e3 ]B .
***
Def: The mapping x 7→ [x]B is called the coordinate mapping determined by B, or simply B-coordinate mapping.
Thm
Then
linear
linear
8 (P.235): Let B = {b1 , . . . , bp } be a basis for H.
the B-coordinate mapping is a one-to-one and onto
transformation from H to Rp . Its inverse is also a
transformation from Rp back to H.
Proof: Direct verification.
61
Under this coordinate mapping, any linear problem in V can
be translated to a corresponding problem in Rp , and then
we are equipped with the powerful matrix theory.
Example: Check that the set of vectors:
{t + t3 , 1 − t + t2 , 5 − 3t3 , 4 + 2t2 , 3 − t + t3 },
is l.d. in P3 .
Can check it directly by solving the polynomial identity.
Sol: Let B = {1, t, t2 , t3 }. By B-coordinate mapping, they
are sent to:
62
  
 
   

0
1
5
4
3




1
−1
0
0
−1
  
 
   

,
,
,
,
  
 
   

0
1
0
2
0




1
0
−3
0
1
5 vectors in R4 must be l.d., i.e. exists c1 , . . . , c5 , not all
zeros, such that:

  

 



 
0
3
4
5
1
0
 −1   0 
0
 0 
 −1 
1
c1   + c2 
 =  .
 + c4   + c5 
 + c3 
0
0
2
0
1
0
0
1
0
−3
0
1
63
Under inverse B-coordinate mapping, we obtain a dependence relation among the 5 polynomials in P3 :
c1 (t + t3 ) + c2 (1 − t + t2 ) + c3 (5 − 3t3 )
+ c4 (4 + 2t2 ) + c5 (3 − t + t3 ) = 0(t).
So the 5 polynomials are l.d. in P3 .
The technique used above generalizes to:
Thm 9 (P.241): Let B = {b1 , . . . , bp } be a basis for H.
Then any subset in H containing q > p vectors must be l.d.
Proof: Send these q vectors by B-coordinate mapping to q
vectors in Rp .
64
Thm 10 (P.242): Let B and C be two bases for V consisting
of n vectors and m vectors respectively. Then n = m.
Pf: First consider B as a basis for V .
• If m > n, C must be l.d. by previous theorem.
• Since C is l.i., we must have m ≤ n.
Now consider C as a basis for V .
• If n > m, B must be l.d. by previous theorem.
• Since B is l.i., we must have n ≤ m.
Therefore n = m.
65
Dimension of a Vector Space
Def: When V has a basis B of n vectors, V will be called
finite dimensional of dimension n or n-dimensional. We will
write dim V = n.
Def: The dimension of the zero space {0} is defined to be 0.
Def: If V cannot be spanned by any finite set of vectors, V
will be called infinite dimensional. We will write dim V = ∞.
Examples: dim Rn = n, dim Pn = n + 1, dim P = ∞.
Basis for Rn : E = {e1 , . . . , en }
Basis for Pn : {1, t, t2 , . . . , tn }
Basis for P: {1, t, t2 , . . .}
66
Let H be a subspace of V and dim V = n ≥ 1.
Thm 11 (P.243): Any l.i. set in V can be extended to a
basis for V . In particular, if H is a subspace of V , we have:
dim H ≤ dim V,
and if dim H = dim V , we have H = V .
Proof: Let S be a l.i. set in V .
1. If Span S = V , S is already a basis.
2. If not, there will be a vector u ∈
/ Span S.
3. The set S ∪ {u} will have one more element, still l.i.
67
(For the equation c1 v1 + . . . + cp vp + cp+1 u = 0; first
show cp+1 = 0 using (2), then show c1 = . . . = cp = 0.)
4. Go back to (1) with new S ′ = S ∪ {u}.
Such addition of vectors must stop somewhere since l.i.
set in V cannot contain more than n vectors.
(Process stops → the l.i. set S ′ can span V .)
If dim H = dim V , first choose any basis B for H.
a. dim H = |B| = n(= dim V ) by assumption.
b. Extend B to B′ , a basis for V , by adding suitable vectors.
c. Any l.i. set in V cannot contain more than n vectors.
d. Must have B ′ = B (adding no vector in (b)), and hence
H = Span B = Span B ′ = V .
68
The proof of previous theorem says that:
Thm 12 (P.243): Let dim V = n ≥ 1. Then
(a) Any l.i. set with n vectors is a basis for V .
(b) Any spanning set of V with n vectors is a basis for V .
Proof: (a) We cannot add vector to S as before since any
set with n + 1 vectors must be l.d., so S is a basis for V .
(b) If S is l.d., we can remove certain vector in S and obtain
a smaller set S ′ which still spans V . Keep removing vectors
until S ′ is l.i., and thus forming a basis for V . This S ′ contains dim V = n vectors, i.e. we actually remove nothing
from S.
69
So, any two of the conditions will characterize a basis S:
(a) |S| = n = dim V .
(b) S is l.i.
(c) S spans V .
0. (b) & (c): (original defintion)
1. (a) & (b): Any l.i. set of n = dim V vectors is a basis.
2. (a) & (c): Any spanning set of n = dim V vectors is a
basis.
70
Example: Describe all possible subspaces of R3 .
Sol: Let H be a subspace of R3 . Then dim H = 0, 1, 2, 3.
0. dim H = 0. H = {0} is the only possibility.
3. dim H = 3. Then H = R3 .
1. dim H = 1. Then H has a basis B = {v}.
→ every x ∈ H is a l.c. of {v};
i.e. x = cv where c ∈ R.
→ H is a line passing through origin, containing v.
2. dim H = 2. Then H has a basis B = {v1 , v2 }.
→ every x ∈ H is of the form x = c1 v1 + c2 v2 .
→ H is a plane through 0, containing both v1 , v2 .
71
Dimensions of Nul A, Row A, and Col A
Def: The dimension of Nul A is called the nullity of A.
Def: The dimension of Row A is called the row rank of A.
Def: The dimension of Col A is called the column rank of A.
Let A be an m × n matrix with p pivot positions. Then
A has n − p free variables, and so the general solution of
Ax = 0 can be written as:
x = s1 x1 + . . . + sn−p xn−p ,
where s1 , . . . , sn−p ∈ R.
Thm: nullity of A = dim Nul A = n − p.
72
Let A be an m × n matrix with p pivot positions. Then
there are p non-zero rows in a REF of A.
Thm: row rank of A = dim Row A = p.
Let A be an m × n matrix with p pivot positions. Then
there are p pivot columns of A:
Thm: column rank of A = dim Col A = p.
Recall the definition that rank A = no. of pivot positions
in A. So:
Thm: row rank of A = column rank of A = rank A.
73
In summary, we have the following “rank theorem”:
Thm 14 (P.249): Let A be an m × n matrix. Then:
dim Row A = dim Col A = rank A
rank A + dim Nul A = n.
In the language of matrix transformations, we have:
Thm: Let T : Rn → Rm be a matrix transformation. Then:
dim Im T + dim ker T = n = dim Rn .
74