March 19, 2019

Bases and Dimension of Vector Spaces

Bases and Dimension of Vector Spaces
  1. Bases
  2. Dimension
  3. Some Useful Theorems

Bases

Previously, I discussed span and linear independence. Essentially, the span of a list of vectors is the set of all possible linear combinations of those vectors, and a list of vectors is linearly independent if none of them can be expressed as a linear combination of the others.

Now I'll use these ideas to introduce one of the most important concepts in linear algebra:

Definition. A basis for a vector space V is a linearly independent list of vectors which spans V.

This definition, combined with what we have already discussed, tells us three things:

  1. Every vector in V can be expressed as a linear combination of basis vectors, since a basis must span V.
  2. This representation is unique, since the basis elements are linearly independent.
  3. V is the direct sum of the spans of the basis vectors.

Stated more explictly, if (e1,,en) is a basis for a vector space V, then for every vector vV there is a unique set of scalars aiF for which

v=i=1naiei.

Bases for vector spaces are similar to bases for topological spaces. The idea is that a basis is a small, easy to understand subset of vectors from which it is possible to extrapolate pretty much everything about the vector space as a whole.

Here are some examples and non-examples.

Example. The list ((1,0),(0,1)) is a basis for the vector space R2, since the list is linearly independent and for any vector (x,y)R2, we may write

(x,y)=x(1,0)+y(0,1).

Example. The list ((1,0),(1,1)) is also a basis for R2. To see that it is linearly independent, we must show that

(1)a(1,0)+b(1,1)=(0,0)

only if a=b=0. Since in this vector space addition and scalar multiplication are done component-wise, equation (1) reduces to the system of equations:

(2)a+b=0(3)a+b=0.

Equation (2) implies that a=b, from which equation (3) implies that a=b=0, as desired.

To see that our list spans R2, we note that any vector (x,y)R2 can be written

(x,y)=(xy)(1,0)+y(1,1).

Example. The list ((1,0),(1,0)) is not a basis for R2 because it is linearly dependent and it does not span R2. To see that it is linearly dependent, note that

(1,0)+(1,0)=(0,0).

That is, we have written the zero vector as a linear combination of vectors in the list with nonzero coefficients.

To see that it does not span R2, note that no vector in R2 with a nonzero second component can ever be written as a scalar multiple of these vectors, since both have a second component of zero.

Example. The list ((1,0),(0,1),(1,1)) is not a basis for R2 because it is linearly dependent. We can see this immediately from the linear dependence lemma, since it has length three and we have already exhibited a linearly dependent list of length two which spans R2.

Example. The list (1,x,x2,x3) is a basis for P3(F), the vector space of polynomials over a field F with degree at most three. To justify this claim, note that this list is certainly linearly independent, and any polynomial of degree less than or equal three can be written

p(x)=a+b+cx2+dx3

for some choice of a,b,c,dF.

The next theorem is fairly obvious, but without it we would be pretty much lost. Recall that a vector space is finite-dimensional if it is spanned by a finite list of vectors. For this next proof we will use this definition, as well the linear dependence lemma.

Theorem. Every finite-dimensional vector space has a basis.

Let V denote a finite-dimensional vector space. Then V is spanned by some finite list of vectors, not all zero, (v1,,vn). If this list is linearly independent, then we are done. Otherwise, we can use the linear dependence lemma to remove a vector from this list and produce a new list of length n1 which still spans V. We continue this process until we are left with a linearly independent list, which will take at most n1 steps, since any list containing one nonzero vector is automatically linearly independent. This resulting list is, by definition, a basis for V.

This argument only works for finite-dimensional vector spaces, since it relies on the fact that we only have to apply the linear dependence lemma a finite number of times. Infinite-dimensional vector spaces, as I've mentioned before, are a lot grosser. It is possible to show that infinite-dimensional vector spaces are guaranteed to have bases if we accept the axiom of choice. Since most mathematicians much prefer to have bases for their vector spaces, this is just one more point in favor of accepting the axiom of choice. Luckily for us, we aren't currently interested in infinite-dimensional vector spaces and so we can simply ignore this icky business.

Dimension

Next comes a super important result which will finally settle the question I posed last time about how to define the dimension of a vector space. Its proof will employ the theorem I proved at the end of my last post, that the length of a list of spanning vectors is never shorter than any linearly independent list.

Theorem. All bases of a finite-dimensional vector space have the same length.

Proof. Let V denote a finite dimensional vector space, and suppose (e1,,en) and (f1,,fm) are both bases for V. Since (e1,,en) is linearly independent and (f1,,fm) spans V, we have that nm. Similarly, since (f1,,fm) is linearly independent and (e1,,en) spans V, we have that mn. It follows that m=n, as desired.

What we have really shown is that the length of a basis for a finite-dimensional vector space is an invariant of that space! And it's a particularly special invariant:

Definition. The dimension of a finite-dimensional vector space is the length of any basis for that space.

If the dimension of a vector space V is n, we write

dimV=n.

As a special case, recall that we defined span ()={0}. That means that dim{0}=0.

We've already shown that dimension is well-defined, since all bases for a vector space have the same length. But does this definition coincide with what we expect? For instance, we would certainly expect that dimRn=n for any positive integer n. And it turns out that this always the case, as we can see by the following definition.

Definition. Given a positive integer n, the standard basis for Fn is the list containing the vectors

e1=(1,0,,0),e2=(0,1,,0),en=(0,0,,1).

That is, ei is the vector whose ith component is 1 and every other component is zero.

It should be fairly obvious that for any n, the standard basis for Fn is, in fact, a basis. It is certainly linearly independent and it is easy to write any vector in Fn as a linear combination of basis vectors, implying it also spans Fn. Since there are n basis vectors in the standard basis, it follows that dimFn=n, just like we would expect.

Some Useful Theorems

This first result tells use how to calculate the dimension of a sum of two subspaces. Recall that we previously showed that the intersection of two subspaces is itself a subspace.

Theorem. If V1 and V2 are subspaces of a finite-dimensional vector space, then

dim(V1+V2)=dimV1+dimV2dim(V1V2).

Proof. Since V1V2 is a subspace of a finite-dimensional vector space, it is also finite-dimensional. Thus it has a basis, which we will denote (e1,,en).

This basis for V1V2 is linearly independent, and thus we may extend it to a basis for V1 by adding m1=dimV1n new linearly independent vectors. That is, there is some basis (e1,,en,f1,,fm1) for V1, and certainly dimV1=m1+n.

We may similarly extend (e1,,en) to a basis for V2 by adding m2=dimV2n new linearly independent vectors. That is, there is some basis (e1,,en,g1,,gm2) for V2, and certainly dimV2=m2+n.

We argue that (e1,,en,f1,,fm1,g1,,gm2) is a basis for V1+V2. Certainly

V1,V2span (e1,,en,f1,,fm1,g1,,gm2)

and thus

V1+V2=span (e1,,en,f1,,fm1,g1,,gm2).

To see that this list is linearly independent, suppose that

(4)i=1naiei+i=1m1bifi+i=1m2cigi=0

for some scalars {ai}i=1n, {bi}i=1m1 and {ci}i=1m2. We can rewrite (4) as

i=1m2cigi=i=1naieii=1m1bifi,

from which we see that i=1m2cigiV1V2. Since (e1,,en) is a basis for V1V2, it follows that we may write it uniquely as a linear combination of basis vectors,

(5)i=1m2cigi=i=1ndiei,

for some scalars dii=1n. But (e1,,en,g1,,gm2) is linearly independent, so it follows from equation (5) that ci=0 for all i. This means we may rewrite equation (4) as

i=1naiei+i=1m1bifi=0.

Since (e1,,en,f1,,fm1) is linearly independent, it follows that ai=0 for all 1in and bi=0 for all 1im1. Since all coefficients in equation (4) have been shown to be zero, it follows that (e1,,en,f1,,fm1,g1,,gm2) is a basis for V1+V2.

It follows then that

dim(V1+V2)=n+m1+m2=(n+m1)+(n+m2)n=dimV1+dimV2dim(V1V2),

completing the proof.

If you think that proof was gross, try doing it for sums of three or more subspaces. Or don't, because as far as I know there is no general formula for such a thing.

The next result ties together our work from last post regarding direct sums, and follows immediately from the theorem we just proved. Recall that if a sum of two subspaces is direct, their intersection is trivial.

Theorem. If U1,,Un are subspaces of a finite-dimensional vector space V for which

V=i=1nUi,

then

dimV=i=1ndimUi.

Proof. We first rewrite the sum as

V=(((U1U2)U3)Un).

We work outward, first considering the subspace W2=U1U2 and then the subspace W3=W2U3, then the subspace W4=W3U4, etc. This takes n1 iterations, but eventually we reach Wn=V.

But U1U2={0} since the sum is direct, and so by the previous theorem we see that

dimW1=dim(U1+U2)=dimU1+dimU2dim(U1U2)=dimU1+dimU2dim{0}=dimU1+dimU20=dimU1+dimU2.

And in general, since Wj=Wj1Uj,

dimWj=dim(Wj1+Uj)=dimWj1+dimUj=i=1j1dimUi+dimUj.

Setting j=n, the result follows.

Next time I will introduce linear maps, which are (besides vector spaces themselves) the primary objects of study in linear algebra.

Menu