Download VECTOR SPACES: FIRST EXAMPLES 1. Definition So far in the

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project

Document related concepts

Matrix multiplication wikipedia , lookup

Exterior algebra wikipedia , lookup

Laplace–Runge–Lenz vector wikipedia , lookup

Cayley–Hamilton theorem wikipedia , lookup

Euclidean vector wikipedia , lookup

Eigenvalues and eigenvectors wikipedia , lookup

Covariance and contravariance of vectors wikipedia , lookup

Four-vector wikipedia , lookup

System of linear equations wikipedia , lookup

Matrix calculus wikipedia , lookup

Vector space wikipedia , lookup

Transcript
VECTOR SPACES: FIRST EXAMPLES
PIETER HOFSTRA
1. Definition
So far in the course, we have been studying the spaces Rn and their subspaces, bases, and so on. In
mathematics, it is common practice to ask what features of the objects you study are essential for the
theory to work, and which are coincidental. We then separate the essential from the inessential, and leap
into abstraction: we turn the essential features into a definition, so that we can develop the theory on a
more general level. In the case of linear algebra, it turns out that the essential features that make things
work for Rn are the fact that you can add vectors and that you can do scalar multiplication (and that
they behave properly). For example, the notions of linear combination, linear independence, and basis
only refer to addition and scalar multiplication. This leads to the following definition:
Definition 1.1 (Abstract Vector Space). A vector space consists of three things:
• A set V (to whose elements we sometimes refer as vectors).
• An addition operation V × V → V , written, for v, w ∈ V , as (v, w) 7→ v + w.
• A scalar multiplication operation R × V → V , written, for v ∈ V, c ∈ R, as (c, v) 7→ cv.
These operations are required to satisfy the following axioms:
(1)
(2)
(3)
(4)
(5)
(6)
(7)
(8)
v + w = w + v for all v, w ∈ V (commutativity of + )
(v + w) + z = v + (w + z) for all v, w, z ∈ V (associativity of + )
There is an element 0 ∈ V such that 0 + v = v = v + 0 for all v ∈ V
For all v ∈ V there is a unique w with v + w = 0
(cd)v = c(dv) for all c, d ∈ R and v ∈ V
1v = v for all v ∈ V
(c + d)v = cv + dv for all c, d ∈ R and v ∈ V (distributivity 1)
c(v + w) = cv + cw for all c ∈ R and v, w ∈ V (distributivity 2)
2. Examples
Of course, the motivating example of a vector space is Rn , with the usual addition of vectors and
scalar multiplication. However, there are many other examples. Several of those will be closely related
to Rn , but especially the last one (function spaces) is quite different in nature.
Example 2.1.
Let V be C, the set of complex numbers. When discussing the algebra of complex numbers, we noted
that we could define addition of complex numbers by
(a + bi) + (c + di) = (a + c) + (b + d)i.
Also, given a complex number z = a + bi and a scalar (real number) c, we can multiply them to get
cz = c(a + bi) = (ca) + (cb)i.
With these notions of addition and scalar multiplication, it is easily verified that C forms a vector space.
When we define complex numbers, we make use of the real numbers: given z = a + bi, the two parts
a and b are just real numbers. That is, a complex number z = a + bi is completely determined by the
two real numbers a and b, and conversely, any two real numbers a and b uniquely determine a complex
number a + bi. We might say that C is like R2 , but with a different notation: in R2 we write elements as
(a, b), while in C we write them as a + bi. Technically, however, they are different sets; in mathematics,
2
PIETER HOFSTRA
such a situation is made precise by saying that the two sets, while different, are in one-one correspondence
with each other. Schematically:
/ R2
C
a + bi
/ (a, b)
But more is true: addition of complex numbers (a + bi) + (c + di) agrees with that of addition of
vectors in R2 . On the one hand we have (a + bi) + (c + di) = (a + c) + (b + d)i, while on the other hand
we have (a, b) + (c, d) = (a + c, b + d). This tells us that not only can we “go back and forth” between
the complex numbers and R2 , but this passage respects the addition: first adding two complex numbers
and then turning it into a vector in R2 gives the same result as first turning both complex numbers into
vectors in R2 and then adding them there. Diagrammatically:
/ (a, b) , (c, d)
(a + bi) , (b + di)
_
_
add individually
add
(a + c) + (b + d)i / (a + c, b + d)
The same holds for scalar multiplication: given z = a + bi and a scalar c, first turning z into (a, b) ∈ R2
and then forming c(a, b) = (ca, cb) gives the same result as first doing scalar multiplication in C and then
turning the result into a vector in R2 :
c
,
multiply
(a + bi)
_
(ca) + (cb)i /c
,
_
(a, b)
multiply
/ (ca, cb)
We may summarize all this informally by saying that not only is C just R2 in disguise as a set, but C is
also R2 in disguise as a vector space!
√
Example 2.2. A polynomial is an expression like p(x) = 2x2 − x + 3, or q(x) = x4 − 5, or r(x) =
1 3
2 x − x + 5. Formally:
Definition 2.3. A polynomial (of degree n) is an expression of the form
an xn + an−1 xn−1 + · · · + a1 x + a0
where x is a variable and where the a0 , . . . , an are real numbers, called the coefficients of the polynomial.
Given two polynomials p(x) = an xn + · · · + a1 x + a0 and q(x) = bn xn + · · · + b1 x + b0 , we have
p(x) = q(x) precisely when ai = bi for all i = 0, . . . , n.
The set of all polynomials of degree n is denoted Pn [x], and the set of all polynomials (of arbitrary
degree) P[x].
We now explain how to add polynomials and how to define scalar multiplication (although you already
know this from calculus!). Given two polynomials p(x) = an xn + · · · + a1 x + a0 and q(x) = bn xn + · · · +
b1 x + b0 , we define
p(x) + q(x) = (an + bn )x + · · · + (a1 + b1 )x + (a0 + b0 ).
That is, we add polynomials by adding their coefficients.
For scalar multiplication, consider Given two polynomials p(x) = an xn + · · · + a1 x + a0 and c ∈ R,
and define
cp(x) = (can )xn + · · · + (ca1 )x + (ca0 ).
With these definitions, Pn [x] becomes a vector space. Moreover, using the fact that we can always
regard a polynomial of degree n as one of higher degree by adding coefficients of zero, we also get that
P[x] is a vector space as well.
Just as with complex numbers, a polynomial p(x) = an xn + · · · a1 x + a0 is completely determined by
its coefficients (an , . . . , a0 ), which form an element of Rn+1 . Moreover, addition and scalar multiplication
are defined coefficient-wise; this means that Pn [x] is just Rn+1 in disguise!
VECTOR SPACES: FIRST EXAMPLES
3
Example 2.4. Matrices. We know from matrix algebra that matrices of the same size can be added
entry-wise. Also, we can multiply any matrix by a scalar. This means that when we fix m, n, the set
Mm,n of all mxn matrices is a vector space.
Since an mxn matrix can be regarded as n column vectors in Rm (or, of course, as m row vectors in
R ), we find that Mm,n is Rmn in disguise. For example, we can translate a 3x2 matrix to a vector in
R6 as follows:
a b c
7−→ (a, b, c, d, e, f ) ∈ R6
d e f
n
Example 2.5. Our last example is different from the previous ones, in the sense that it is not a disguised
form of Rn . Consider F[R], the set of all functions from the reals to the reals. So an element of F[R] is
a function f : R → R. (Remember, a function has to give a unique output f (x) for every input x, so
something like f (x) = 1/x is not an element of F[R], since it is not defined for x = 0.)
From calculus, we know that it is possible to add functions (pointwise, as it is sometimes called): given
two functions f, g : R → R, define a new function f + g : R → R by
(f + g)(x) = f (x) + g(x)
Note that we’re defining this new function f + g by specifying what it’s output is given an arbitrary input
value x: to know a function h is to know h(x), for all possible values of x. Note also that this equation
uses the + symbol twice: on the left hand side, it means addition of functions (which is what we’re trying
to define) and on the right hand side it’s ordinary addition of real numbers (something we assume we
understand already).
For scalar multiplication, consider f : R → R and a real number c ∈ R. Define a new function
cf : R → R by
(cf )(x) = c · f (x)
That is, on input x, the function cf returns the output value c times f (x).
With these definitions of addition and scalar multiplication, F[R] is a vector space. Note that the
“origin” in this vector space is the constant function z(x) = 0.
3. A bit of linear algebra
We now proceed to examine the concepts of linear combinations and linear (in)dependence in the
examples given above. Recall first the definition of linear combination: it makes sense in any vector
space.
Definition 3.1 (Linear Combination). Let V be a vector space, and let v1 , . . . , vk be elements of V . A
linear combination of v1 , . . . , vk is a vector v of the form
v = a1 v1 + . . . + ak vk
for some real numbers a1 , . . . , ak .
Here are some examples:
Example 3.2. We work in P2 [x]. The polynomial p(x) = 4x2 −4x is a linear combination of q(x) = x2 +2
and r(x) = 2x2 + x + 3 because
p(x) = 12q(x) − 4r(x).
Example 3.3. Is p(x) = 4x2 − 4x a linear combination of x2 , x2 − x + 1 and 2x − 2? The answer is no.
To see this, suppose we had
4x2 − 4x = a(x2 ) + b(x2 − x + 1) + c(2x − 2)
for some numbers a, b, c. Collecting powers of x in the right hand side gives
4x2 − 4x = (a + b)x2 + (2c − b)x + (b − 2c).
For this equation to hold, we need
a + b = 4;
(2c − b) = −4;
b − 2c = 0.
4
PIETER HOFSTRA
This is a system of linear equations in the unknowns a, b, c, which is readily seen to be inconsistent.
Hence it is not possible to find numbers a, b, c with the desired property, and therefore p(x) is not a linear
combination of the given polynomials.
Exercise 1. Is x3 − 3 a linear combination of 2x3 + 1, x2 − x + 1, x + 2, x3 + 1 and x2 − 1?
√
Example 3.4. This example is more a warning than anything else: while 5x, 41 x, 2x and 0 are all linear
combinations (scalar multiples) of x, p(x) = x2 is not. Yes, x2 = x · x, but this involves multiplying x
with itself, which is not a linear operation.
Note that each time we try to answer a question about Pn [x], we end up solving a linear system. This
is not a coincidence, and can be made more explicit by “translating” the question into one about Rn+1 .
This is an important idea: it allows us to use all the old techniques and results. For example, consider
again the problem in Ex. 3.3. We solved this by reasoning directly about polynomials, then setting up a
system, then solving that system and then drawing a conclusion from that. But what’s much nicer and
quicker is to translate the problem into one about R3 . The key is that that a polynomial ax2 + bx + c
can be regarded as a vector (a, b, c) in R3 . So we translate all the polynomials involved, and we get the
following question about vectors in R3 :
Example 3.5 (Translated version of Ex. 3.3). Is the vector (4, −4, 0) a linear combination of (1, 0, 0), (1, −1, 1)
and (0, 2, −2)?
Here, we can use any method we like, for example row reduction:

 

1 1
0 | 4
1 1 0 | 4
0 −1 2 | −4 ∼ 0 1 −2 | 4
0 1 −2 | 0
0 0 0 | 4
leading to the immediate conclusion that the system is inconsistent.
Schematically, here’s what’s happening:
Problem about P3 [x] 99K
translate ↓
Problem about R3
−→
Solution to the original problem
↑ translate back
Solution to problem inR3
Exercise 2. Try the same approach for Ex. 3.2.
For our next example, we work in M2,2 .
1 2
2 0
0 1
0
Example 3.6. The matrix
is a linear combination of
,
and
3 4
0 2
1 1
1
1 2 0
0 1
0 0
1 2
+2
+
=
0
2
1
1
1 1
3 4
2
1 2
2 0
0 1
0 0
However,
is not a linear combination of
,
and
.
3 4
0 2
1 1
0 1
Exercise 3. Verify the last claim by considering the equation
1 2
2 0
0 1
0
=a
+b
+c
3 4
0 2
1 1
0
0
, because
1
0
1
and then solving for a, b, c.
Example 3.7. Is the set
1
3
2
2
,
4
0
0
0
,
2
1
1
0
,
1
1
0
1
linearly independent?
Solution. No, from Ex. 3.6 we know that the first is a linear combination of the other three.
We’ll now indicate how to solve questions like these using the idea that M2,2 is R4 in disguise.
VECTOR SPACES: FIRST EXAMPLES
5
1 2
2 0
0 1
Example 3.8 (Translation of Ex. 3.6). Is the matrix
is a linear combination of
,
3 4
0 2
1 1
0 0
and
?
1 1
Translate to R4 : is the vector (1, 2, 3, 4) a linear combination of (2, 0, 0, 2), (0, 1, 1, 1) and (0, 0, 1, 1)?


 
2 0 0 | 1
1 0 0 | 12
0 1 0 | 2 0 1 0 | 2 


 
0 1 1 | 3 ∼ 0 0 1 | 1 
2 1 1 | 4
0 0 0 | 0
Thus the answer to the translated problem is Yes, and hence the answer to the original problem is Yes,
too.
Finally, some examples in F[R].
Example 3.9. f (x) = 2x+1 + x2 is a linear combination of g(x) = 2x − 1 and h(x) = x2 + 2, because
2g(x) + h(x) = 2(2x − 1) + x2 + 2 = 2 · 2x − 2 + x2 + 2 = 2 · 2x + x2 = 2x+1 + x2 .
Example 3.10. g(x) = sin2 (x) is a linear combination of h(x) = cos2 (x) and k(x) = 1 because of the
Pythagorean identity sin2 (x) + cos2 (x) = 1, from which we get
g(x) = k(x) − h(x)
Exercise 4. Use the same idea to show that m(x) =
3 sin2 (x) + 5.
1
2
cos2 (x) is a linear combination of 2 sin2 (x) − 3 and
Example 3.11. Recall the double angle formula sin(2x) = 2 sin(x) cos(x). One might be tempted to
jump to the conclusion that sin(2x) is therefore a linear combination of sin(x) and cos(x), but that would
be incorrect. Indeed, it is impossible to find a, b ∈ R with
2 sin(x) cos(x) = a sin(x) + b cos(x).
To see why, keep in mind that this equality would have to hold for all x ∈ R. So if it were true, it should
be true for x = 0, giving
2 sin(0) cos(0) = a sin(0) + b cos(0)
which, because of sin(0) = 0, becomes 0 = b cos(0) = b. This forces b = 0, reducing the original equation
to
2 sin(x) cos(x) = a sin(x).
However, this should in particular be true of x = π/2, giving
π
π
π
2 sin( ) cos( ) = a sin( ).
2
2
2
Because of cos(π/2) = 0, this becomes 0 = a sin(π/2) = a, forcing a = 0 as well. Contradiction, and
therefore such a, b cannot exist.
4. Subspaces
n
The notion of subspace of R only referred to linear independence and to spanning. Therefore it makes
sense for general vector spaces:
Definition 4.1. Let V be a vector space, and let U ⊆ V be a subset of V . Then U is a subspace of V
provided the following three conditions hold.
(1) 0 ∈ U
(2) v, w ∈ U implies v + w ∈ U
(3) v ∈ U, c ∈ R implies cv ∈ U .
Thus, a subspace of V is a collection of elements of V containing the zero vector, and is closed under
addition and scalar multiplication. All three conditions have to hold; as soon as one of them fails, you
no longer have a subspace.
6
PIETER HOFSTRA
Example 4.2. Consider the subset U = {ax2 | a ∈ R} of P2 [x]. Show that U is a subspace, give a basis
and find its dimension.
Solution 1. We verify the three axioms. Throughout, keep in mind that to say that a polynomial p(x)
is in U , is exactly to say that is of the form p(x) = ax2 + 0x + 0 for some a.
(1) We must show that the zero vector is in U . The zero vector in P2 [x] is the polynomial 0, or, more
formally, 0x2 + 0x + 0. Hence this polynomial is in U , as required.
(2) Consider p(x), q(x) ∈ U . This means that p(x) = ax2 for some number a, and q(x) = bx2 for
some number b. Now to show that p + q is in U , note that
p(x) + q(x) = ax2 + bx2 = (a + b)x2 .
Thus p + q is also a multiple of x2 , and hence is in U .
(3) Consider p(x) ∈ U and c ∈ R. Then we have p(x) = ax2 for some number a. Now
cp(x) = (ca)x2 ,
and therefore cp(x) is again a multiple of x2 and hence in U .
For a basis, observe that ax2 is a scalar multiple of x2 ; thus {x2 } is a basis and dim(U ) = 1.
Solution 2. We can write U = Span{x2 }. Since the span of any set of vectors is a subspace, we’re
done; it is clear that {x2 } is linearly independent (having only one vector in it) and hence is a basis,
resulting in dim(U ) = 1.
Solution 3. We can translate this into a problem about subspaces of R3 . To do so, recall that an
arbitrary polynomial ax2 + bx + c is translated as the vector (a, b, c). Thus a polynomial ax2 is translated
as (a, 0, 0). We thus get: is the subset U 0 = {(a, 0, 0)|a ∈ R} of R3 a subspace? We can now use any
method we like to answer this question, e.g. we can regard it as the span of {(1, 0, 0)}.
Example 4.3. Is U = {A | A = AT } a subspace of M2,2 ? If so, find a basis and find its dimension.
Solution. First note that U is the set of all symmetric 2x2 matrices. We could easily verify the three
4
axioms
fora subspace here, but we’ll instead translate the problem to R . Observe that for a matrix
a b
A=
, we have A = AT precisely when b = c. This means that the translated subset is:
c d
U 0 = {(a, b, c, d) ∈ R4 | b = c}
which can also be written as U 0 = {(a, b, c, d) ∈ R4 | b − c = 0}. Therefore it is
homogeneous system, hence is a subspace. For a basis, we solve this system
(keep
has four variables a, b, c, d and one equation b − c = 0, so has as matrix 0 1 −1
parameters, say a = t, c = s, d = r, and one leading variable b, which becomes b =
general solution is
(a, b, c, d) = t(1, 0, 0, 0) + s(0, 1, 1, 0) + r(0, 0, 0, 1).
the solution set of a
in mind, this system
0 ) . We get three
c = s. Therefore the
This means that we have three basic solutions forming the basis for U 0 , namely {(1, 0, 0, 0), (0, 1, 1, 0), (0, 0, 0, 1)}.
Consequently, dim(U ) = 3.
Translating this back to M2,2 , we see that a basis for U is
1 0
0 1
0 0
,
,
0 0
1 0
0 1
and that dim(U ) = 3.
Example 4.4. Which of the following sets are subspaces of F[R]?
•
•
•
•
•
A = {f | f (1) = 0}
B = {f | f (x) ≥ 0 for all x}
C = {f | f (4) − 2f (1) = f (−2)}
D = {f | f (x) = 2f (x + 1) for all x}
E = {f | f (0)f (1) = 1}
VECTOR SPACES: FIRST EXAMPLES
7
Solution. As a rough guideline, we use the idea that subsets defined by homogeneous linear equations
are subspaces, but subsets defined by nonlinear equations aren’t. For example, A is defined by the
equation f (1) = 0, which is linear; similarly, C and D are defined by linear equations, but E is not,
because (i) the equation uses a product of two values of f and (ii) the right hand side is 1, making it
non-homogeneous. We’ll prove that B and E are not subspaces first.
The set B contains the zero vector, is closed under addition but not under scalar multiplication.
Consider for example f (x) = ex . This function is positive, hence is in B. However, −3ex (a scalar
multiple of f ) is negative, hence not in B.
The set E doesn’t contain the zero vector: recalling that the zero vector is z(x) = 0 for all x, we get
z(0)z(1) = 0 6= 1.
We now show that D is a subspace. First of all, the zero vector satisfies z(x) = 2z(x + 1) because both
sides are 0. Next, consider f, g ∈ D. That is, f (x) = 2f (x + 1) and g(x) = 2g(x + 1) for all x. Then to
show that f + g ∈ D, consider
(f + g)(x) = f (x) + g(x) = 2f (x + 1) + 2g(x + 1) = 2(f + g)(x + 1)
so that f + g ∈ D. Finally, if f ∈ D and c ∈ R we get
(cf )(x) = cf (x) = c(2f (x + 1)) = 2cf (x + 1)
so that cf ∈ D.
The items A and C are left as an exercise.