Fehler: Aktuelle Seite wurde in der Sitemap nicht gefunden. Deswegen kann keine Navigation angezeigt werden
In the last two articles we got to know the two concepts generator and linear independence. In this articles we combine the two concepts and introduce the notion of basis of a vector space.
Motivation
Via linear independence
We want to work towards the concept of dimension. Intuitively, we can think of it as the maximum number of linearly independent directions in a space. If we stick to this intuition, we can give the following preliminary definition of dimension: The dimension of a vector space is the maximum number of linearly independent vectors that we can simultaneously choose in a vector space. To find this, we need to find a maximal system of linearly independent vectors.
To see if this preliminary definition makes sense, let's try to apply it in the
: Clearly, the
should be three-dimensional. This means we have to ask ourselves two questions: Do three linearly independent vectors fit into the
and is there no fourth vector being linearly independent to any of such three independent vectors? Let us now try this out: We take any first vector, e.g.
. Now we want to find a linearly independent vector, i.e. a vector that is not a multiple of
or does not lie in
. An example for this is
.
If we want to follow the intuition that
is three-dimensional, we still have to find a third vector that is linearly independent to the system
. Now
and
span a plane in which the last component is always zero. Thus
is a vector linearly independent of
.
Now the question is whether with these three vectors the maximum number of linearly independent vectors has already been reached. To answer this, let us first consider the vector
as an example. We want to check whether we can add this vector to
and
and still obtain a system of linearly independent vectors. First we note that
So we can write the above vector as
So
. Now let us ask the same question for any vector
with
. For this we first get
With this consideration we obtain for
the representation
Thus
. Since this vector was arbitrarily chosen, every vector from
is representable as a linear combination of the linearly independent vectors
and
. Thus
is a generator of
. Therefore, we cannot add another vector to
and
, so the system remains linearly independent, since every other vector from
can be represented as a linear combination of
and
. In other words,
and
form a maximal system of linearly independent vectors.
In summary, we proceeded as follows to find a maximal system of linearly independent vectors: We start with a vector that is not the zero vector. That means we should not consider the null vector space here. Then we proceed step by step: Once we have found linearly independent vectors
, we form the span of these vectors. If this is the entire vector space, we have found a generator and are done. We choose a vector
that is not in
. This contributes a new direction and the system
is linearly independent again. Then we do the same again until we find a generator. We thus obtain the characterisation that a maximal system of linearly independent vectors is a generator of linearly independent vectors.
Via generators
So far we have started with a system of linearly independent vectors (which is not yet a generator) and extended it until it became a maximal system of linearly independent vectors. Now we want to investigate what happens when we reverse the direction. That is, we start with a generator (which is not linearly independent) and reduce it until we find a minimum (i.e., linearly independent) generator.
Let us consider
First, we establish that
is a generator of
by following calculation: For a vector
with
we have that
So we can represent any vector
as a linear combination of vectors from the generator.
Now we ask ourselves whether we can reduce the above generator without losing the property of it being a generator. The vector
is a multiple of
. That is, the direction represented by
is the same as the direction represented by
. Hence, we can remove this vector from
and obtain a new generator
Can we reduce the size of this generator, as well? Yes, since
So
adds no new direction that is not already spanned by
and
. We thus obtain a smaller generator
Now we cannot reduce the generator any further without losing the property of it being a generator. For if we remove any of the three vectors in
, it is no longer in the span of the remaining two vectors. For
, for example, we see this as follows: Suppose we had some
, so that
Then
would have to hold because the second component must be zero on both sides. Because of the third component,
must be true. Thus we obtain the contradiction
So
is not in
.
So we have found a generator containing linearly independent vectors (a minimal generator).
In summary, we proceeded as follows: We started with a generator of vectors
and reduced it according to the following algorithm: If
is a system of linearly independent vectors, we cannot remove any vector without losing the property of having a generator. That means we are done and we have found a minimal generator. In the converse case, we find a vector
with
that is in
. This vector can be omitted and we obtain a new generator consisting of
vectors. With this generator, we do the same steps again until we have found a system of linearly independent vectors.
We thus obtain the characterisation that a minimal generator is a generator consisting of linearly independent vectors.
Motivation - Conclusion
We have found the characterisations of linearly independent generators as minimal generators and as maximal linearly independent subsets. Thus the property of being a linearly independent generator is a special property of both linearly independent sets and generators. A set with both properties is called a basis in Linear Algebra.
Since bases are generators, every vector has a representation as a linear combination of basis vectors. This representation is unambiguous because bases are linearly independent. So we found another way of describing bases:
Important!
A basis is a subset such that every vector has a representation as a unique linear combination of basis vectors
.
Definition: Basis of a vector space
Hint
We have defined the base as a set of vectors. This does not determine the order of the vectors. Alternatively, one can define the base as a tuple of vectors. In this case, the order of the vectors is fixed. Changing the order in this case results in a different basis.
Equivalent definitions of basis
Proof (Equivalent definitions of basis)
We prove the equivalence of these statements by a "circle of implications" of the kind
:
Proof step: 
We show this using a proof by contradiction. We assume that statement 1 is true and statement 2 is false. Then we show that it follows that statement 1 must also be false (which is a contradiction). So statement 2 cannot be false, i.e., it must be true.
So let us assume that
is a linearly independent generator but not a maximal linearly independent subset of
. Thus there is a
such that the vectors of
are linearly independent. (This is exactly the opposite of statement 2.) However, since
is a generating system according to statement 1, we can write
as a linear combination of elements from
:
with
. At least one
with
, because
is linearly independent and therefore
. The above equation can be transformed to:
Thus the set
is linearly dependence. This is a contradiction to our assumption that
is linearly independent. That means, we verified
by contradiction.
Proof step: 
The proof is divided into two steps. In the first step we show that under the assumption of statement 2 a linear combination exists for every element from
with elements from
. In the second step we show the uniqueness of this linear combination. Both steps must be shown, because statement 3 requires the existence of such a linear combination as well as its uniqueness.
Let
be a maximal linearly independent subset of
. Let
be any element from our vector space. We will now show that we can write
as a linear combination of elements from
. We now distinguish two cases. It is important to cover all cases. Here it is obvious, because we first consider a subset of
and then its complement.
Thus we have finished the consideration of both cases and get a linear combination of
using elements from
. Since
was arbitrarily chosen from
, we have now already shown that every element from
can be represented as a linear combination of vectors from
. This shows the existence of the linear combination.
In the next and final step we have to prove the uniqueness of this linear combination. We show this again via a proof by contradiction. We assume that there is a
which can be represented by two different linear combinations. A contradiction to statement 2 will then be inferred from this representation ambiguity. Let
with
. Since both linear combinations are different, there is at least one
for which
. If this were not true, all coefficients would be identical and thus both linear combinations would be identical. We now subtract these two equations and get the following representation of the zero vector:
Let us define the difference vector
. Since
, we have that also
. Thus we can rewrite the above equation as:
With
there is at least one non-zero coefficient. So this is a non-trivial linear combination of zero with elements from
. It follows that
is linearly dependent. This is a contradiction to statement 2, according to which
is linearly independent. So our ambiguity assumption was false and the linear combination must be unique. Therefore we have that
.
Proof step: 
We carry out the proof in two steps. First we show that, assuming statement 3,
is a generator. Then we show that it is minimal. By the definition of a generator,
must be a subset of
and it must span
(i.e.
). By statement 3,
. Since we can represent every element from
as a linear combination using elements from
, we also have that
spans the vector space
. Thus
is a generator.
In the next step we want to show that
is minimal. We again perform a proof by contradiction and assume that
is not a minimal generator. This leads us to a contradiction to statement 3. If
is not a minimal generator, there exists a real subset
which is also a generator. Let now
, so
. Then there exist two different linear combinations for
constructed by vectors in
. Now,
can be "trivially" represented:
or we use the representation of
by elements from
(which can be done because
forms a generator):
with
. Since, according to statement 3, for every element from
there is a unique linear combination of vectors from
, the existence of the two linear combinations contradicts statement 3. Thus
must be a minimal generator. This concludes the proof of contradiction and we have that
.
Proof step: 
We show this again using a proof by contradiction. So we assume that statement 1 is false. This means that
is not a linearly independents generator. That is, there exists a
that can be represented as a linear combination of vectors
from
:
Here, for at least one coefficient,
. We now lead this to contradict statement 4 by showing that
is then no longer minimal.
Let
be any vector. Since
is a generator, there is a linear combination of
with vectors
from
:
Now we plug the above linear combination of
into the linear combination of
:
We have now found a linear combination for the arbitrary vector
. Thus
is a generator of
. This contradicts statement 4, according to which
is a minimal generator. So we also verified
by contradiction.
Examples
Canonical basis in coordinate space
The vector space
of the
tuples over the field
has a so-called standard basis or canonical basis
For instance,
has the canonical basis
and the
the canonical basis
.
Exercise (Basis of the plane)
Show that the set
is a basis of
.
Summary of proof (Basis of the plane)
We first show that any vector can be represented as a linear combination of the two given canonical basis vectors, so they form a generator. Then we show that they are linearly independent.
Solution (Basis of the plane)
Proof step:
is generator of 
Let
be arbitrary. Then,
So
is a generator of
.
Proof step:
is linearly independent
We assume
if follows
So
and
. Hence, the vectors of
are linearly independent.
A different basis of 
Exercise
Show that the vectors
provide a Basis of
.
Summary of proof
In a first step, one shows that
is a generator of
. In a second step, linear independence of these vectors is established.
Solution
Proof step: generator
Let
be any vector of
. Then,
is a linear combination of
using the vectors
so these vectors are a generator of
.
Proof step: Linear independence
We represent the zero as a linear combination of the three vectors and check for which coefficients, a zero can be obtained:
This results in the following system of equations:
The solution of this system of equations is
. Thus the vectors are linearly independent.
Example: complex numbers
The set of complex numbers
is a vector space over
, with multiplication by real numbers:
Then
has as an
-vector space the basis
, because for
we have unique
with
.
If we consider
as a
-vector space,
is no longer a basis of
, since.
For
as a
-vector space we have the one-element basis
. As
-vector space the complex numbers have a two-element basis (dimension is
) and as
-vector space a one-element basis (dimension is
). So be cautious: it is important over which field we take a fixed vector space!
Abstract example
The vector space
of the polynomials with coefficients from
has a basis with infinitely many elements. An example for a basis are the powers of
:
This is a generator, because for a polynomial
of degree
we have a representation
Where
for all
. Thus every polynomial is a finite linear combination of elements from
. Consequently,
is a generator.
For linear independence we consider the following: With
let:
We can also write the zero polynomial as
. A comparison of coefficients yields that
So
is a Basis.
Bases are not unique
Example: Basis of the plane is not unique
We will show by using the plane
that the basis of a vector space is not unique. Let us look at the (canonical) basis for
consisting of the unit vectors:
These vectors obviously form a generator:
They are also linearly independent because it is impossible to find a linear combination of zero with non-trivial coefficients. Thus
is a basis. However, for the plane there are also a lot of other bases. An example is the following set:
We can generate all vectors with these two vectors:
These vectors are linearly independent because one vector is not a multiple of the other vector (two vectors are linearly dependent if one vector is a multiple of the other vector). Thus
is also a basis. These two examples show that the basis for
is not unique. And one can indeed find a lot of other bases, for instance by stretching and rotating.
Building a further basis from an existing one
In general, for every
-vector space with a basis, we can construct a second, different basis: Consider the two-dimensional vector space
with a basis
. Then
is also a basis of
. The same argument of "substituting the last vector" also works in higher dimensions. We first show that
is a generator and then that the basis vectors are linearly independent.
Let
be any vector and
a linear combination of it using elements of the basis
. Then a linear combination of
can also be found via vectors from
:
Thus
is a generator because the vector
was arbitrarily chosen.
We show that the basis vectors are linearly independent via proof by contradiction. To do this, we prove that if
is linearly dependent, then
must also be linearly dependent. By contraposition, it thus follows that
is linearly independent if
is linearly independent (which it must be the case it is a basis). If
is linearly dependent, there is a representation of zero:
Here
or
. Now we also find a representation of the zero with the basis
:
We still have to show that one of the coefficients
or
is not equal to zero. As a premise we have
or
. The case
leads directly to the non-trivial representation of the zero, since this factor also appears in the second equation.
If
holds, we have to distinguish again. If
and
, then
and hence one of the coefficients is not zero. If
and
hold, then
and hence the first coefficient is non-zero.
It follows that one of the coefficients is always non-zero. Thus the vectors of the basis
are also linearly dependent. It follows (by contradiction) that
is linearly independent if
is linearly independent.
is thus a new basis constructed from the first basis.
This principle can also be applied to larger bases and shows: The basis of a vector space is (usually) not unique. A vector space with dimension equal to or larger than 2 has several bases.
Proving existence and constructing a basis
Existence of a basis
We have not yet answered the question of whether there is a basis for every n vector space at all. Attention, this is not self-evident, as bases may be infinitely large! Nevertheless, you can rejoice, because the answer is: Yes, every vector space has (at least) one basis.
Of course, we still have to justify this answer. For the case of finitely generated vector spaces, i.e. all vector spaces that have a finite generator, we will prove this in a moment. For infinitely generated vector spaces, i.e. vector spaces that do not have a finite generating system, the proof is much more complicated and uses the axiom of choice.
Theorem (Existence of a basis)
Every finitely generated vector space has at least one Basis.
Proof (Existence of a basis)
We take a finite generator. With the basis theorem, there is a subset of it that is a basis of the vector space. In particular, the vector space has at least one basis.
Construction of a basis by removing vectors
Now we know that every vector space has a basis, but how can you find a basis for a given vector space?
For finitely generated vector spaces, the proof of the theorem on the existence of a basis gives you a procedure for constructing a basis in finitely many steps (it is not applicable for infinitely generated vector spaces). According to the basis completion theorem, we can proceed as follows:
- Find a finite generator
of the vector space.
- Check whether
is linearly independent.
- If yes: We are done and the generator is a basis.
- If no: Find a smaller generator
of the vector space and repeat step 2.
We now need an explicit way to get a smaller generator
from a finite generator
, which is not a minimal generator. Since
is not a minimal generator,
is linearly dependent. So we find
so that not all
and we have that
Now we choose an
with
and set
We now want to show that
is also a generator. In the article linear independence of vectors we have proven that
Since
is a generator of
, we find for a vector
scalars
such that
Thus
. Since
was arbitrarily chosen, it follows that
is a generator. With the proof of the basis theorem, we now get the following procedure for determining a basis:
How to get to the proof? (Construction of a basis for finite vector spaces)
- Find a finite generator of the given vector space.
- Try to find a non-trivial linear combination of the zero vector from this generator. (For this, you need to solve a linear system of equations.) If no non-trivial linear combination exists, the generator is also linearly independent, and we are done.
- If a non-trivial linear combination exists, remove one of the vectors from the generator whose coefficient in the linear combination is not zero. Then, go back to Step 2.
Construction of a basis by adding vectors
Alternatively, we can proceed as in the section Motivation via linear independence; that is, we start with a linearly independent set and extend it until it is maximal, i.e., a basis.
Proof (Basis completion theorem)
Let
be any vector space and
be any independent subset of
. Since
is finitely generated, we find a finite generator
of
. We now want to add elements from
to the set
until this new set is a generator. When adding the vectors, we want to keep the linear independence, so that the new set is a linearly independent generator and thus a basis.
If
is a generator, then
is already a basis of
. Else we find that since
, there is some
with
. Now we set
. To show the linear independence of
, we assume
would be linearly dependent. Then a non-trivial linear combination would exist:
with
. Since
is linearly independent,
must hold. So we get
This is a contradiction to
. Thus
is linearly independent.
We now inductively construct linearly independent sets
by adding elements from
according to the above procedure until
has become a basis:
- start with

- and inductively set
for some
with
, if
is not yet a generator of
.
Since we can only add finitely many vectors from
to
, there is an
at which we cannot add a vector from
to
. Then
is a generator and hence also a basis of
.
This proof gives you another method to determine a basis of a finitely generated vector space in finitely many steps (this method is also only applicable for finitely generated vector spaces):
- Choose a finite generator and start with the empty set as your first linearly independent set.
- Try to find a vector from your generator that is not in the span of your previous linearly independent set. If you don't find one, you're done.
- Add the vector you found to your linearly independent set and go back to Step 2.
In the next chapter we will see that every two bases of the same finitely generated vector space have the same cardinality. We get that every linearly independent set that has as many elements as a basis is automatically already a maximally linearly independent subset. Therefore, we can change step 2 in the above procedure as follows: "Try to find a vector from your vector space that is not in the span of your previous linearly independent set. If you don't find one, you're done." In this variant of the procedure, you do not have to choose a generator in Step 1.
Examples: Construction of a basis by removing vectors
Example (Construction of a basis by removing vectors 1)
Let
. Then, we have
, since for all
we have that
That means,
is a generator. Now,
is not a basis of
because we have the non-trivial representation of the zero vector
Since in this linear combination the prefactor of
is not zero, according to the above procedure for constructing a basis from a generator
is also a generator. We now need to check that
is linearly dependent. So we look at how the zero vector can be combined using vectors from this set:
If we set up a linear system of equations and solve it, it follows that
. Thus B' is linearly independent and hence a basis of
.
Example (Abstract example: Construction of a basis by removing vectors)
Let
be the two-element field. We consider the
-vector space
of all mappings from
to
. This is a function space (missing) consisting of
elements. Thus
. Here
is the constant zero mapping,
is the constant one mapping,
is the identity and
is the mapping that interchanges
and
:
As initial generating system
we choose the entire vector space, i.e.
. This is linearly dependent, because we have the following non-trivial representation of the zero vector
So we get a new generator
of
since the prefactor of
is non-zero. Specifically,
. Now
is also linearly dependent because we have
Since the prefactor of
is non-zero, we choose as a new generator
. This set is linearly independent. From
we obtain (by inserting the two possible arguments):
So
and thus
is linearly independent. With
we have found a basis of our vector space.
Example: Construction of a basis by adding vectors
Example (Construction of a basis by adding vectors)
We consider the space
and start with the linearly independent set
. Then
. That is, we get the linearly independent set
. Further
That means
is linearly independent. Now
Thus
and
is linearly independent. Further,
, because
Thus
is a basis of
.
Example (Abstract example: Construction of a basis by adding vectors)
We consider the two-element field
and the
-vector space
of all mappings from
to
. As seen above,
consists of 4 elements, that is, in the above notation
. Let again
. Then
. Thus
linearly independent. Now
This means
is linearly independent. Further we have that
Thus
is a basis of
.
Fehler: Aktuelle Seite wurde in der Sitemap nicht gefunden. Deswegen kann keine Navigation angezeigt werden