Serlo: EN: Kernel of a linear map

Aus testwiki
Version vom 10. Juni 2022, 18:14 Uhr von imported>Sascha Lill 95 (Exercises)
(Unterschied) ← Nächstältere Version | Aktuelle Version (Unterschied) | Nächstjüngere Version → (Unterschied)
Zur Navigation springen Zur Suche springen

{{#invoke:Mathe für Nicht-Freaks/Seite|oben}} The kernel of a linear map intuitively contains the information that is "deleted" when applying the linear map. Further, the kernel can be used to characterize the injectivity of linear maps. It also plays a central role in solving systems of linear equations.

Introduction

We have learned about special mappings between vector spaces, called linear maps. Those are structure-preserving; that is, they are compatible with addition and scalar multiplication of a vector space. We can therefore think of a linear map from V to W as something that transports the vector space structure from V to W.

Introductory examples

We consider two accounts, each with the account balance x and y respectively. We can describe this information with a vector (x,y)T2. The total account balance is the sum of the two account balances. We can calculate it by using the map Vorlage:Einrücken This map is linear and therefore transports the vector space structure from 2 to . In the process, information is lost: one no longer knows how the money is distributed among the accounts. For example, one can no longer distinguish the individual account balances (500,0)T and (200,300)T because they both map to the same total account balance 500+0=200+300=500. In particular, the mapping is not injective. However, we get the information about how much money is in the accounts in total.

Rotation of the real plane by 90° against the clockwise direction

Next, we consider the map Vorlage:Einrücken Visually, this corresponds to a counterclockwise rotation of 2 by 90 degrees. By undoing this rotation, one can recover the original vector from any rotated vector in 2. Formally speaking, this mapping is an isomorphism and no information is lost. In particular, the image of linearly independent vectors is linearly independent again (because an isomorphism is injective, see the article monomorphism) and the image of a generator of 2 is again a generator of 2 (because an isomorphism is surjective, see the article epimorphism).

Finally, we consider a rotation again, but then embed the rotated plane into the 3: Vorlage:Einrücken Although this mapping is no longer bijective, no information is lost here when transporting the vector space structure of the 2 into the 3: As in the previous example, different vectors in the 2 are mapped to different vectors in the 3 because of injectivity. Linear independence of vectors is also preserved. However, a generating system of 2 is not mapped to a generator of 3. For example, the linear map sends the standard basis {(1,0)T,(0,1)T} to {(0,1,0)T,(1,0,0)T}, which is not a generator of 3. The property of a set of vectors to be a generator depends on the ambient space. This is not the case with linear independence; it is an "intrinsic" property of sets of vectors.

Derivation Vorlage:Anker

We have seen various examples of linear maps that transport a K-vector space into another K-vector space, while preserving the structure. In the process, varying amounts of "intrinsic" information from the original vector space (such as differences of vectors or linear independence) were lost. The last example suggests that injective mappings preserve such intrinsic properties. On the other hand, we see: If f:VW is not injective, then there are vectors v,vV with f(v)=f(v). So in that case, f "eliminates" the difference vv of v and v. The difference vv is again an element in V. Since f is linear, we can reformulate: Vorlage:Einrücken Intuitively, f is injective if and only if differences vv of vectors under f are not eliminated (i.e., mapped to zero). Because f is structure-preserving, we have that for all v,vV and λK, that f(vv)=0 implies Vorlage:Einrücken If the difference of v and v is eliminated under f, so is that of λv and λv. In the same way, if v,v,w,wV: if f(vv)=0 and f(ww)=0, then also Vorlage:Einrücken So the difference of v+w and v+w is also eliminated. The differences eliminated by f are themselves vectors in V. These are send by f to the zero element 0W of W and thus, the eliminated vectors are in the preimage f1({0W}). Conversely, any vector vf1({0W}) can be written as a difference v=v0; that is, the difference v0 between v and the zero vector is eliminated by f. The preimage f1({0W}) measures exactly what differences of vectors (how much "information") is lost in the transport from V to W. Our considerations show that f1({0W}) is even a subspace of V. We give a name to this subspace: the kernel of f.

Definition

The kernel of a linear map intuitively measures how much "intrinsic" information about vectors from V (differences of vectors or linear independence) is lost when applying the map. Mathematically, the kernel is the preimage of the zero vector. Mathe für Nicht-Freaks: Vorlage:Definition

In the derivation we claimed that the kernel of a linear map from V to W is a subspace of V. We will now prove this in detail.

Mathe für Nicht-Freaks: Vorlage:Satz

Examples

We determine the kernel of the examples from the introduction.

Vector is mapped to the sum of entries

We consider the mapping Vorlage:Einrücken The kernel of f is made up by the vectors (x,y)T2 with 0=f((x,y)T)=x+y, so y=x. In other words Vorlage:Einrücken Thus the kernel of f is a one-dimensional subspace of 2. More generally, for n we can consider the mapping Vorlage:Einrücken Again, by definition, a vector (x1,,xn)T,xn lies in the kernel of g if and only if 0=g((x1,,xn))=x1++xn holds. So we can freely choose x1,,xn1 and then set xn=x1xn1. Thus Vorlage:Einrücken Hence, the kernel of g is a (n1)-dimensional subspace of n. It is also called a hyperplane in n.

Rotation in 2

We consider the rotation Vorlage:Einrücken Suppose (x,y)T lies in the kernel of f, i.e. it holds that Vorlage:Einrücken From this we obtain x=y=0. So only the zero vector lies in the kernel of f and we have that kerf={(0,0)T}.

2 is rotated and embedded into the 3 Vorlage:Anker

Next we consider Vorlage:Einrücken As in the previous example, we determine the kernel by choosing any vector (x,y)Tkerf. Thus it holds that Vorlage:Einrücken Again it follows that x=y=0, so that also for this mapping kerf={(0,0)T} holds.

Derivatives of polynomials Vorlage:Anker

Finally, we consider a linear map that did not appear in the introduction: Vorlage:Einrücken which maps a real polynomial to its derivative. That is, a polynomial Vorlage:Einrücken with coefficients a0,,an is mapped to the polynomial Vorlage:Einrücken Graphically, we associate with p a polynomial p that indicates the gradient of p at each point. From this information, we still learn what the shape of the polynomial is (just as if we were given a stencil). However, we no longer know where it is positioned on the y-axis, because the information about the constant part of the polynomial is lost when taking the derivative. Polynomials that just differ by a displacement along the y-axis can no longer be distinguished after derivation. For example, both p=x2x+1 and q=x2x+42 have the derivative p=q=2x1. So the mapping f maps them to the same polynomial.

The kernel of f thus contains exactly the constant polynomials: Vorlage:Einrücken The inclusion "" is clear, because the derivative of a constant polynomial is always the zero polynomial. For the converse inclusion "", we consider any polynomial pkerf and show that it is constant. We can always write such a polynomial as p=i=1naiXi for some n and certain coefficients a0,,an. Because of pkerf it holds that Vorlage:Einrücken and by comparison of the coefficients, we obtain a1=a2==an=0. So p is constant. Vorlage:Todo

Kernel and injectivity

In the derivation above, we saw that a linear map preserves all differences of vectors (i.e., no vector is eliminated) if and only if the kernel consists only of the zero vector. We also saw there that linearity implies: A linear map is injective if and only if no difference of vectors is eliminated. So we have the following theorem:

<section begin="InjektivitätSatz" />Mathe für Nicht-Freaks: Vorlage:Satz <section end="InjektivitätSatz" />

The larger the kernel is, the more differences between vectors are "eliminated" and the more the mapping "fails to be injective". The kernel is thus a measure of the "non-injectivity" of a linear map.

Injective maps and subspaces

In the introductory examples we conjectured that injective linear maps preserve "intrinsic" properties of vector spaces. By this, we mean properties that do not depend on the ambient vector space, such as the linear independence of vectors or vectors being distinct. The property of being a generator can be lost in injective linear maps, as we have seen in the example of the twisted embedding of 2 into 3: The mapping is injective, but the standard basis of 2 is not mapped to a generator of 3.

What exactly does it mean that a property of a family N=(vi)iIV of vectors does not depend on the ambient space V? Often, properties of vectors from V (for example, linear independence) depend on the vector space structure of V, that is, addition and scalar multiplication. To make dependences as small as possible, we restrict our attention to the smallest subspace of V containing N, that is, we restrict to span(N). Now, we call a property of N intrinsic if it depends only on span(N) but not on V.

Mathe für Nicht-Freaks: Vorlage:Beispiel

What do intrinsic properties of a family of vectors have to do with injectivity? Let f:VW be a linear map. Suppose f preserves intrinsic properties of vectors, that is, if a family N=(vi)iIV has some intrinsic property, then its image f(N)=(f(vi))iI under f also has this property. Then f also preserves the property of vectors being different, since this is an intrinsic property. That means, if v,vV are different, i.e., vv, then their image under f is also different, i.e., f(v)f(v). So f is injective.

Conversely, if f is injective, then V is isomorphic to the subspace f(V) of W: If we restrict the target space of f to its image, we obtain an injective and surjective linear map f:Vf(V), that is, an isomorphism. In particular, for any family N in V, it holds that the subspace span(N) of V is isomorphic to f(span(N)). Thus, the latter has the same properties as span(N) and hence, f preserves intrinsic properties of subsets of V.

So we have seen that f:VW is injective if and only if f preserves intrinsic properties of subsets of V.

Kernel and linear independence

In the previous section we have seen that injective linear maps VW are exactly those linear maps which preserve intrinsic properties of V. The linear independence of a family of vectors is such an intrinsic property, as they either hold for any choice of an ambient space or do not hold for any choice of an ambient space.

So, injective linear maps should preserve linear independence of vectors, i.e., the image of linearly independent vectors is again linearly independent. Conversely, a linear map cannot be injective if it does not preserve the linear independence of vectors, since the intrinsic information of "being linearly independent" is lost.

Overall, we get the following theorem, which has already been proved in the article on monomorphisms:

Mathe für Nicht-Freaks: Vorlage:Satz

In particular, for any linear map f:VW, the vector space f(V) is a dim(V)-dimensional subspace of W. In the finite-dimensional case, there cannot exist an injective linear map from V to W if dim(W)<dim(V). This has also already been shown in the article on monomorphisms.

Kernel and linear systems Vorlage:Anker

The kernel of a linear map is an important concept in the study of systems of linear equations.

Let K be a field and let m,n. We consider a linear system of equations Vorlage:Einrücken with n variables x1,,xn and m rows. We have aij,biK, where i{1,,m} and j{1,,n}. We can also write this system of equations using matrix multiplication: Vorlage:Einrücken where AKm×n, xKn and bKm. We denote the set of solutions by Vorlage:Einrücken

Determining a solution to the linear system of equations Ax=b for a given right-hand side b is the same as finding a preimage of b under the linear map Vorlage:Einrücken

Vorlage:Todo The system of equations Ax=b has solutions if the preimage fA1(b) is not empty. In this case, we may ask whether there are multiple solutions, that is, whether the solution is not unique. In other words, we are interested in how many preimages a b has under fA.

By definition of injectivity, every point bKm has at most one element in its preimage if and only if fA is injective. This means that the linear system of equations Ax=b has at most one solution for each bKm, that is, |L(A,b)|1. Because fA is linear, injectivity is equivalent to ker(fA)={0}. So we can already state: Mathe für Nicht-Freaks: Vorlage:Satz Mathe für Nicht-Freaks: Vorlage:Hinweis

Even if fA is not injective, i.e., ker(fA){0} holds, we can still say more about the set of solutions by exploiting the kernel: The difference of two vectors x and x, which fA maps to the same vector, lies in the kernel of fA. Therefore, the preimage of some bKm under fA can be written as Vorlage:Einrücken where x^ is any element of fA1(b). This is shown by the following theorem: Mathe für Nicht-Freaks: Vorlage:Satz We have thus even extended the statement of the theorem above. The larger the kernel of fA is, that is, the "less injective" the mapping xAx is, the "less unique" are solutions of Ax=b, if any exist. The set of solutions of a linear system of equations Ax=b is the kernel of the induced linear map fA shifted by a particular solution x^. Furthermore, Vorlage:Einrücken The set of solutions of the homogeneous system of equations Ax=0 (that is, with right-hand side zero) is exactly the kernel of fA. Mathe für Nicht-Freaks: Vorlage:Hinweis

Exercises

<section begin=injektivität_und_dimension /> Mathe für Nicht-Freaks: Vorlage:Aufgabe<section end=injektivität_und_dimension /> <section begin=aufgabe_kern_bestimmen /> Mathe für Nicht-Freaks: Vorlage:Aufgabe

Mathe für Nicht-Freaks: Vorlage:Frage<section end=aufgabe_kern_bestimmen /> <section begin=kern_nilpotenter_endo /> Mathe für Nicht-Freaks: Vorlage:Aufgabe<section end=kern_nilpotenter_endo />

{{#invoke:Mathe für Nicht-Freaks/Seite|unten}}