Download SOME REMARKS ON BASES Let V denote a vector space. We say

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project

Document related concepts

Exterior algebra wikipedia , lookup

Laplace–Runge–Lenz vector wikipedia , lookup

System of linear equations wikipedia , lookup

Euclidean vector wikipedia , lookup

Matrix calculus wikipedia , lookup

Vector space wikipedia , lookup

Lp space wikipedia , lookup

Four-vector wikipedia , lookup

Covariance and contravariance of vectors wikipedia , lookup

Transcript
SOME REMARKS ON BASES
Let V denote a vector space. We say that a set of vectors {A1 , A2 , · · · } spans V if
V = Span {A1 , A2 , · · · }.
We may also say that {A1 , A2 , · · · } is a spanning set for V. Notice we are not requiring the set of
vectors to be finite here.
Now, suppose we have a vector space V and we have a spanning set {A1 , A2 , · · · }. What does
the spanning set do for us? The spanning set should be thought of as providing us basic building
blocks for our vector space, in kind of1 the same way the atoms from the periodic table form the
basic building blocks for all other molecules and larger chemical complexes, or how the primary
colors red, blue, and yellow combine to give us all other colors.
Typically, vector spaces are very large complicated things. Even the ”easy” vector spaces, like
R, R2 , R3 , are very large and complicated objects. However, most questions in matrix or linear
algebra decompose in a nice way with respect to linear combinations. This usually means that if
you want to answer some question about a vector space, then it will be enough to be able to answer
the question when restricting to a spanning set. Essentially, the spanning vectors describe all the
different directions in V, and each direction just looks like a copy of R.
x
1 2
, then
and you want to understand the effect of multiplying A by
For example, if A =
y
0 1
one way2 is to use the linearity properties
of matrix multiplication and the fact that {e1 , e2 } is a
1
0
spanning set for R2 , where e1 =
, e2 =
. Then
0
1
x
A
= A(xe1 + ye2 ) = xAe1 + yAe2 .
y
We only have to compute the products of A with the two spanning vectors. We find
1
2
Ae1 =
, Ae2 =
.
0
1
We therefore find that
x
1
2
x + 2y
A
=x
+y
=
.
y
0
1
y
1The
chemistry and painting examples are only metaphorical, and should not be taken too literally.
course, another way is to just to carry out the multiplication process. However, the point of view we develop
here turns out to be a very useful one, despite its apparent inefficiency.
2Of
1
2
SOME REMARKS ON BASES
{A1 , A2 , · · · } being a spanning set for V says that every vector in V can be built up from
{A1 , A2 , · · · }. Working with a set that does not span the space V would be like trying to do
chemistry without carbon, or trying to paint only using red and blue, and the purplish colors we
get by mixing these two. We can do some chemistry without carbon, but we will be in trouble if
we need graphite or carbon dioxide. We can paint some paintings without yellow, but we will be
in trouble if we need yellow or green or orange. We can describe some chunk of our vector space V
with a non-spanning set. However, if we only knew
what Ae1 was in the above example we would
2
be in trouble if we wanted to multiply A by, say,
, because this vector does not sit in the chunk
5
of R2 given by the span of e1 .
Recall that we say V is finite dimensional if we can write
V = Span {A1 , A2 , · · · , Ak }
for some finite collection of vectors {A1 , A2 , · · · , Ak } in V. In other words, V is finite dimensional
provided that V has a finite spanning set. This means that in some sense there are only a finite
number of different directions in our vector space. In the plane we only need two directions, say
North and East. There are plenty of other directions, but they can all be described in terms of
North and East.
Finite dimensional spaces have the advantage that we really only need to understand what happens in a finite number of directions. There doesn’t seem to be much consolation in knowing
that to understand the space of all polynomials we only need to understand all of the monomials
1, x, x2 , · · · . There are just too many directions here! Understanding what happens for the three
standard basis vectors of R3 seems like a much more manageable task.
Some spanning sets are better than others. For
0, 0], [0, 1, 0], [0, 0, 1]} is a much
√ instance,2 {[1,
√
3
nicer spanning set for R than {[1, 2, 3], [3, −1627, 157], [π, π , π], [2−19 , 9, 3]}. The second is bad
for two reasons. First, given an arbitrary vector in our space, say [1, 36, −7], we would like to be
able to find coefficients expressing our vector in terms of the spanning set. For the first basis this is
trivial; [1, 36, −7] = 1[1, 0, 0] + 36[0, 1, 0] + (−7)[0, 0, 1]. I wouldn’t want to have to find coefficients
for [1, 36, −7] in terms of the second spanning set. Of course, if we absolutely had to we could find
them, it just would be a pain. There is a rather precise reason why the first basis is better than
the second (It has to do with dot products) but we won’t discuss this until chapter 6. Suffice it to
say for now, it is easy to read off the coefficients in the first case. The second reason that the first
spanning set is better than the second is much easier to pin down: The second set is too big! We
have 4 vectors in a 3dimensional space, so there is redundant information in the second spanning
set. More precisely, there must be some linear dependency in the second set of vectors. (What
the dependency is is not so clear due to the ugliness of the vectors, but never the less, there is a
dependency.) Linear dependencies in a spanning set means that the spanning set is not as efficient
as it could be. This leads to the important definition of a basis for a vector space.
SOME REMARKS ON BASES
3
A basis for V is a linearly independent spanning set {A1 , A2 , · · · } for V. Spanning says that
everything in V can be built from {A1 , A2 , · · · }. Independence says we can only build things up
in a single way. Going back to the chemistry example, we know that water can be built up from
things in the periodic table, namely hydrogen and oxygen. We also know that this is the only way.
We can’t build up water from, say, uranium, cobalt and chlorine. We know that yellow and blue
make green. We also know we can’t get green by combining, say purple and orange.
Now, why should we be interested in finding bases for R2 if we already have the standard basis,
which seems to be the best kind of basis we could use? The reason is that the standard basis is not
always the best basis to use. There is no best basis for any vector space. Sometimes we may wish
to choose a basis for our vector space which is adapted to a particular problem we are studying.
Think back to calculus when we study polar coordinates. (This does not quite correspond to finding
a new basis for R2 because polar coordinates are what are called curvilinear coordinates, but the
metaphor is still helpful.) Certain integrals are nasty, if not impossible in cartesian coordinates,
but become almost trivial when we pass to polar coordinates. On the other hand, certain nearly
trivial integrals become impossible when we change to polar coordinates. The general rule is that
if the integrand has some sort of circular structure to it then it is probably advantageous to change
to polar. In matrix algebra, it is usually very helpful to find a basis that is engineered to help solve
whatever problem we are working on. In hindsight, we see that the whole row reduction algorithm
we do when solving a linear system is really just a construction of a nice basis for the null space of
the coefficient matrix.
One example of finding a basis for Rn different from the standard basis arises in solving systems
of equations. Suppose A is some m × n matrix and we want to study the equation Ax = b for
all possible b ∈ Rm . In studying this problem, it is helpful to find a basis adapted to A. Refer
back to the handout I gave on Thursday for a detailed example, but roughly the idea is to start
by finding bases for the column space of A (which is a subspace of Rm ) and the null space of A
(which is a subspace of Rn . Say the column space basis is {b1 , b2 , · · · , bk } and the null space basis
is {y1 , y2 , · · · , yn−k }. Next, for each basis vector for the column space, we find a vector xj ∈ Rn so
that Axj = bj . Then the set {y1 , y2 , · · · , yn−k , x1 , x2 , · · · , xk } is a basis for Rn and is in some sense
much better than the standard basis, as far as working with systems of equations with coefficient
matrix A is concerned. Given b in the column space of A, we want to find all solutions to Ax = b.
To do this, we simply write b in terms of the column space basis,
b = c 1 b1 + c 2 b2 + · · · + c k bk .
Then a particular solution is
x = c1 x1 + c2 x2 + · · · + ck xk .
(Look at how nice this is! All we did is take the coefficients from the column space basis and moved
them over to a different basis!) The general solution is then obtained by adding things from the null
space to this particular solution. This basis is nice for solving systems of equations with coefficient
matrix A because we chose basis elements which had particularly nice vectors when multiplied by
A.