Serlo: EN: Linear independence
{{#invoke:Mathe für Nicht-Freaks/Seite|oben}}
Motivation Vorlage:Anker
Basic motivation
Maybe, you learned about vectors in school, where they were drawn as arrows in the plane or in space. Both a plane and space are vector spaces. But how do they differ?
A spontaneous answer could be: "The plane is two-dimensional and the space is three-dimensional". But this brings us immediately to further questions:
- What is the dimension of a vector space?
- How can we define it?
In the definition of the vector space the term "dimension" does not occur...
Intuition of a dimension

The term "dimension" describes in how many independent directions geometric objects can be extended in a space. The objects can also move in just as many independent directions in space ("degrees of freedom of motion").
The plane has two dimensions - the width and the length. It is flat, no object of the plane can reach out of it "into height". A sphere as a three-dimensional object cannot be part of the plane. In contrast, the space with length, width and height has three dimensions. A sphere can thus be part of space.
We summarize: The dimension intuitively corresponds to the number of independent directions into which a geometric object can expand or move. So, for the definition of dimension, we need to answer the following questions:
- What is a direction in a vector space?
- When are two directions independent?
- How can the number of independent directions be determined?
Derivation of the Definition
What is a direction within a vector space?
Let's take the vector space of the plane as an example. We can represent a direction with an arrow:

Now an arrow is nothing but a vector. So with the help of vectors, directions can be represented. Here we must not use the zero vector. As an arrow of length zero it has no direction. We can generalize this to arbitrary vector spaces:
The direction in which the vector points is , that is, the span of the vector . To this span belong all extensions of the direction vector and thus describes the straight line, which is spanned by :

From the line to the plane
To get from the straight line to the plane, we need not only one vector but several, more precisely at least two vectors (). This is intuitively obvious, because a plane can be spanned unambiguously only with two vectors. Therefore we need a further, linearly independent vector. What does "independent" mean in this case? First, we notice that the new vector must not be the zero vector. This vector does not give any direction. Furthermore, the new vector must also not be a multiple of the original vector, i.e. . This also holds for reflections of straight line vectors, represented by multiplicatioin with a negative factor.
-
The vector is a stretched version of the vector with a positive stretching factor. This vector is not pointing in a direction independent of .
-
The vector is a stretched version of the vector with a positive stretching factor (which includes a reflection). This vector is not pointing in a direction independent of .
-
The direction of is independent of . Both vectors together span a plane.
We conclude: The new vector is independent of the direction vector exactly when the latter is not on the straight line. So we need for all real numbers . Hence, the new vector must not be in the span of the other one. The two spans have only the zero point as intersection.
From the plane to space
We have just seen that we can characterize a plane by two independent vectors. Now we want to go from the plane to space. Here, we also have to add an independent direction. But what is a direction independent of the plane?
The new vector must not be the zero vector, because this vector does not indicate a direction. The new vector must also not lie in the plane, because in that case, no new direction would be described. Only if the new vector does not lie in the plane, it will point in a new, independent direction:
-
The vector lies in the plane spanned by and . Hence, does not point into a direction spanned by and .
-
The vector does not lie in the plane spanned by and . All three vectors span the entire space, which means that points into a direction independent of and .
How can we formulate this insight mathematically? Let and be the two direction vectors spanning the plane. This plane is then equal to the set . Hence, the plane is the set of all sums for real numbers . In order for the new vector not to be in the plane, it must be for all . Thus, in independent of <ma
Mathe für Nicht-Freaks: Vorlage:Frage
Mathe für Nicht-Freaks: Vorlage:Frage
A first criterion for linear independence
Let's summarize: To describe a straight line we needed a vector not being the zero vector. In the transition from the straight line to the plane, we had to add a vector independent of . Independence of from means that does not lie in the line described by . So we need to have for all .
In the second step, we added a new direction to the plane, which is independent of the two vectors and . Here independence manifests itself in the fact that is not in the plane spanned by and . Hence, we need for all real numbers and . We can generalise this to an arbitrary number of vectors (but it is not so easy to visualize anymore):
In the above description, the sum appears. Such a sum is called linear combination of the vectors to . We may also say that is linearly independent if . The description can be changed to:
Here we have clarified when a vector is independent of other vectors. Is this sufficient to describe the independence of vectors?! Take the following three vectors , and as an example:

Since no vector is a multiple of another vector, the three vectors, seen in pairs, point in independent directions. For example is independent of and is independent of . So the three vectors are not independent of each other because they all lie in one plane. We have and so is independent of and . Accordingly, we have to impose linear independence between , and :
- is independent of and : We have for all .
- is independent of and : We have for all .
- is independent of and : We have for all .
It should be emphasised at this point that it is necessary to require all three conditions. If we were to waive the last two conditions, the first requirement would guarantee that the vector is linearly independent of the vectors and , but it is not clear from this requirement that and are linearly independent of each other. This does not have to be fulfilled, which would mean that the three vectors would again not be linearly independent of each other.
Therefore, none of the three vectors must be able to be represented as a linear combination of the other two vectors. Otherwise at least one of the vectors is dependent on the other vectors. We can generalise this to any number of vectors:
Mathe für Nicht-Freaks: Vorlage:Definition
From the first criterion to the formal definition
With our first criterion, which we found above, we have already found a suitable definition for linear independence of vectors. In the following, we will try to find a more concise equivalent criterion, with which we can examine the linear independence of vectors more easily.
Vectors are independent if no vector can be represented as a linear combination of the other vectors. From this we will derive another criterion for linear independence, which is less computationally demanding. Let us take vectors , to from a vector space that are not independent. So there is one vector that can be represented by the others. Let be this vector. There are thus stretching factors (scalars) to , such that Vorlage:Einrücken We can transform this equation by computing on both sides ( is the zero vector of the vector space ): Vorlage:Einrücken This is a so-called nontrivial linear combination of the zero vector. A nontrivial linear combination of the zero vector is a linear combination with the result where at least one coefficient is not equal to . For we would trivially . This is the so-called trivial linear combination of the zero vector, where all coefficients are equal to . You can always form this trivial linear combination, no matter which vectors to you choose. So it does not carry information. If to are dependent, there is at least one non-trivial linear combination of the zero vector (as we saw above) in addition to the trivial linear combination. So:
In other words: Vorlage:Important Now we can apply the principle of contraposition: holds if and only if . So:
With this we have found a criterion for linear independence. If the zero vector can only be represented trivially by a linear combination of to , then these vectors are linearly independent. However, this criterion can also be used as a definition of linear independence. To do this, we need to show the converse direction of the above implication. If there is a non-trivial linear combination of the zero vector, then the vectors under consideration are linearly dependent.
So let to be vectors for which there exists a non-trivial linear combination of the zero vector. This means, there are coefficients (scalars) to , such that where at least one of the coefficients to is not . Let be this coefficient. Then
Since we can multiply both sides by . Then,
On both sides we can now add :
Thus can be represented as a linear combination of the other vectors and hence the vectors to are linearly dependent. This proves taht the following definition of linear independence is equivalent to the first one:
Mathe für Nicht-Freaks: Vorlage:Definition
Definition of a family
We have talked above about a several vectors being linearly independent. But what is this "collection" of vectors from a mathematical point of view? We already know the notion of a set. So it is obvious to understand also as a set. Does this view intuitively fit linear independence? Actually, it turns out problematic, if we have two equal vectors with . Both point in the same direction and span no two independent directions. Thus they are intuitively linearly dependent. And indeed, one can be written as a linear combination of the other as . Thus the vectors are also strictly mathematically linearly dependent. However, a set may only contain different elements. That is, the set containing and is . So the set contains only one element and does not capture duplications of vectors.
So we need a new mathematical term that also captures duplications. This is the concept of family:
Mathe für Nicht-Freaks: Vorlage:Definition
Formally, a family can be seen as a mapping of the index set into the set . In contrast to sets, elements may occur more than once in families, namely if they belong to different indices.
If the set is countable, the elements of the family can be numbered: . However, the index set may also be overcountable, e.g. . In this case cannot be written as a sequence . The term family thus contains all sequences, and includes even larger "collections" of mathematical objects.
So when we say the vectors and are linearly dependent we can express it by saying that the family with is linearly dependent.
Often one writes (with slight abuse of notation) if the are elements of and it is clear from the context what the index set looks like. Similarly, means that there is an with .
With this we can rewrite the second definition of linear independence:
Mathe für Nicht-Freaks: Vorlage:Definition
General definition of linear independence
Motivation
We have learned above two definitions for the fact that finitely many vectors are linearly independent:
- A somewhat unwieldy: vectors are independent if no vector can be written as a linear combination of the others. So must not occur.
- A somewhat more compact one: The zero vector can only be represented as a trivial linear combination. So implies .
So far we have only considered finitely many vectors. What happens with infinitely many vectors? Can there even be an infinite number of linearly independent vectors? We would need a vector space that has infinitely many linearly independent directions. We know intuitively that the vector space has at most two and the at most three independent directions. So we need a much "bigger" vector space to get infinitely many independent directions. So we consider a vector space where every vector has infinitely many coordinates: with . Accordingly, corresponds to a real sequence and is the sequences vector space, or sequence space.
In we have the linearly independent unit vectors . We can continue this construction and obtain for the vectors with the at the -th place and otherwise .
The infinitely many vectors form a family . This family intuitively represents "infinitely many different directions" in and is thus intuitively linearly independent. So it makes sense to define linear independence for infinitely many vectors in such a way that is a linearly independent family. The "somewhat unwieldy definition 1." above would be suitable for this in principle: We could simply copy it and say "a family of vectors is linearly independent if no can be written as a linear combination of the others". In fact, in none of the can be written as a linear combination of the other vectors. Therefore, the definition already makes sense at this point. However, there are infinitely many and thus infinitely many conditions!
We prefer to consider the "somewhat more compact definition 2.": "Vectors are linearly independent if can only be represented by the trivial linear combination." What does this formulation mean explicitly in this example? We are given a linear combination of . Linear combinations are finite, that is, we have finitely many vectors and such that
We now have to show that all , since then the linear combination of above is trivial. This works in exactly the same way as in , except that here we have to compare infinitely many entries.
What do we have to do now to get a general definition for general families and general vector spaces? The "somewhat more compact definition 2." carries over almost literally: "A family of vectors is linearly independent if can only be represented by the trivial linear combination." For the written out implication, we can make use of our language of families: We replace the double indices by the word "sub-family".
Definition
Mathe für Nicht-Freaks: Vorlage:Definition
Mathe für Nicht-Freaks: Vorlage:Warnung Mathe für Nicht-Freaks: Vorlage:Hinweis Mathe für Nicht-Freaks: Vorlage:Hinweis
Implications of the definition
Re-formulating the definition for finite sub-families Vorlage:Anker
We have a definition of linear independence for arbitrary subfamilies of a vector space . Does this agree with our old definition for finite subfamilies? Intuitively, they should agree for finite subfamilies, since we derived the general definition from our old definition. The following theorem actually proves this:
Mathe für Nicht-Freaks: Vorlage:Satz
Reducing the definition to finite sub-families
We have defined linear independence for any family of vectors, so also for infinitely many vectors. But in the definition we only need to show a statement for finite subfamilies : For all with we need the following: Vorlage:Einrücken In the previous theorem we have seen that this statement is exactly linear independence of .
Mathe für Nicht-Freaks: Vorlage:Satz
Overview
The following properties can be derived from the definition of linear independence with a few proof steps. Let be a field and a -vector space:
- Every sub-family of a family of linearly independent vectors is linearly independent. Conversely, every super-family of a family of linearly dependent vectors is again linearly dependent.
- Let be a single vector. Then is linearly independent if and only if . So "almost always". Conversely, every family (no matter how large) is linearly dependent as soon as it contains the zero vector.
- Let . The vectors and are linearly dependent if and only if there is a with the property or .
- If a family of vectors is linearly dependent, one of them can be represented as a linear combination of the others.
Sub-families of linear independent vectors are linearly independent
A linearly independent family remains linearly independent if you take away vectors. Linear dependence, on the other hand, is preserved if you add more vectors. Intuitively, the addition of vectors tends to "destroy" linear independence and cannot be restored by adding more vectors.
Mathe für Nicht-Freaks: Vorlage:Satz
Families including the zero vector are linearly independent
When is a family with exactly one vector linearly independent? This question is easy to answer: whenever this vector is not the zero vector. Conversely, every family with the zero vector is linearly dependent. Including the one that contains only the zero vector itself.
Mathe für Nicht-Freaks: Vorlage:Satz
Two vectors are linearly dependent if one is a stretched version of the other
When is a family with two vectors linearly independent? We can answer the question by saying when the opposite is the case. So when are two vectors linearly dependent? Linear dependence of two vectors holds if and only if both "lie on a straight line", i.e. one vector is a stretched version of the other.
Mathe für Nicht-Freaks: Vorlage:Satz Vorlage:Anker
With linear dependence, one vector is a linear combination of the others
For finitely many vectors, we started with the definition that vectors are linearly dependent if one of the vectors can be written as a linear combination of the others (first definition). We have already seen that this definition is equivalent to the null vector being able to be written as a linear combination of the vectors (second definition). For the general definition with possibly infinitely many vectors, we have used the version with the zero vector (the second) as our definition. And one can indeed show that even in the general case the first definition is equivalent to it:
Mathe für Nicht-Freaks: Vorlage:Satz
Linear independence and unique linear combinations Vorlage:Anker
In this section, we will take a closer look at the connection between linear independence and linear combinations. To do this, we recall what it means that the vectors are linearly dependent or independent. Suppose the vectors are linearly dependent. From our definition of linear independence, we know that there must then be a non-trivial zero representation, since at least one scalar for some . We illustrate this with the following example
Mathe für Nicht-Freaks: Vorlage:Beispiel
Regardless of whether the considered family of vectors is linearly independent or not, there is always the trivial zero representation, in which all scalars have the value :
In case of linear dependence of the vectors, the representation of the zero is no longer unambiguous. We can summarise our results so far in a theorem and generalise them:
Mathe für Nicht-Freaks: Vorlage:Satz
Exercises
Exercise 1
Mathe für Nicht-Freaks: Vorlage:Aufgabe
Exercise 2
Mathe für Nicht-Freaks: Vorlage:Aufgabe
Mathe für Nicht-Freaks: Vorlage:Lösung
Exercise 3
Mathe für Nicht-Freaks: Vorlage:Aufgabe
Exercise 4
Mathe für Nicht-Freaks: Vorlage:Aufgabe
Exercise 5
Mathe für Nicht-Freaks: Vorlage:Aufgabe
Exercise 6
Mathe für Nicht-Freaks: Vorlage:Aufgabe
{{#invoke:Mathe für Nicht-Freaks/Seite|unten}}