3 - Linear Maps

Vector space

A set V and a field F with (V,+) being an abelian group, and a scalar multiplication (,):F×VV That is associative, with 1F producing the identity map 1v=v, and distributive laws both ways between + on V and + on F, and the scalar multiplication of a product in F is the composition of the scalar multiplications.
Elements in V are called vectors, and elements of F are called scalars.
UV is a subspace of V if U is closed under vector addition and scalar multiplication, inherited from V, and also contains the 0 vector.

I don't know why I wrote the definition like this, maybe I am lazy to type

Linear combination

Let V over F be a vector space, <vi> a list of vectors, and <ai> a list of scalars. The linear combination of these lists is a vector u=i=1naivi.
The set of all linear combinations of <vi> is called Span(<vi>) and is a subspace of V, as choosing all scalars to be zero gives the zero vector in the span, and also by definition, the span is closed under both operations.
if there exists a list of scalars <ai> not all zero for which 0=i=1aivi, then the list of vectors <vi> is called linearly dependent.
On the other hand if 0=i=1aivi if and only if each ai=0, then the list of vectors is linearly independent.
The dimension of a vector space is the length of the smallest list of vectors that spans it. The same applies for subspaces.
for a linearly dependent list of vectors, the dimension of the subspace induced by its span is at most the length of the list minus one, as one of these vectors is already in the span, so we need not include it in the list, and we can still span this subspace.
if V is a vector space, any smallest list of linearly independent vectors that span V are called "basis" vectors.

Linear map

if U,V are vector spaces, (on fields FU,FV) then a function f:UV is a linear map if:

  • f(u1+u2)=f(u1)+f(u2)
  • FU=FV (upto isomorphism).
  • f(λu)=λf(u)
    The linear map is a homomorphism between two vector spaces. let <ei> be a basis list of U. Then for any vector vRANGE(f) we have \mathbf{u} = \sum_{i=1}{#na_}{i}\mathbf{e_{i}} such that f(u)=v
    Therefore, f(i=1naiei)=i=1naif(ei).
    This allows us to write a matrix $$
    M = \begin{bmatrix} \uparrow \ f(\mathbf{e_{1}} ) \dots f(\mathbf{e_{n}}) \ \downarrow
    \end{bmatrix}$$ Whose columns are the image of each basis, and when we write u as a column vector [a1,a2,,an]T, with components wrt the same basis, The application f(u) is equivalent to the matmul Mu.
  • Also f(U) (also called RANGE(f)) is a subspace of V, simple due to the structure preserving nature of the linear map: if f(u)f(U), then λf(u)=f(λu)f(U). Similiary, if f(u1),f(u2)f(U), then f(u1)+f(u2)=f(u1+u2)f(U). Zero vector is in there too.
    Hence, we are concerned about the dimension of f(U)dim(V). To see if our linear map f squashes the space U into a lower dimension than U.
Rank of a linear map

Let U be a vector space with dim(U)=n, and <ei> a basis of U, and let V also be a vector space, both over a scalar field F.
then the rank of the linear map f:UV is the dimension of f(U).
By definition, the list <f(ei)> spans f(U) and has length n. Therefore, dim(f(U)) is at most n ( We may find smaller spanning lists for f(U) ).
In other words, rank(f)dim(U).

Kernel of a linear map

for linear map f:UV, the kernel of f, is the subset of U that maps to the zero vector in V.
Kern(f)U such that for all uKern(f), f(u)=0V.
We notice that Kern(f) is a subspace of U. if f(u)=0 then, f(λu)=λf(u)=0 . Similarly, if f(u1)=0 and f(u2)=0, then f(u1+u2)=f(u1)+f(u2)=0. Finally, notice that f(0+0)=f(0)+f(0) hence f(0)=f(0)+f(0), but f(0) has some additive inverse in V, therefore 0V=f(0).

We need some nice spanning list and linear independent list lemmas from LADR for this one

2.19 Linear Dependence Lemma

Suppose v1,,vm is a linearly dependent list in V. Then there exists k{1,2,,m} such that

vkspan(v1,,vk1).

Furthermore, if k satisfies the condition above and the kth term is removed from v1,,vm, then the span of the remaining list equals span(v1,,vm).

Proof

Because the list v1,,vm is linearly dependent, there exist numbers a1,,amF, not all 0, such that

a1v1++amvm=0.

Let k be the largest element of {1,,m} such that ak0. Then

vk=a1akv1ak1akvk1,

which proves that vkspan(v1,,vk1), as desired.

Length of any linearly independent list is at most length of any spanning list

let U be a vector space, <ui> a linearly independent list of length m in U, <wi> a spanning list of length n in U.
Then, mn.
proof:
Consider the following process:

  • step 1: create the list L1=u1,w1,,wn of length n+1. This list is linearly dependent because u1Span<wi>, but u10 as it comes from a linearly independent list. Hence there exist some wjSpan L1, which we can remove to get L1 which is still a spanning list of U
  • step k: Repeating this process to get Lk1 which contains uk1,u1 and the rest are the remaining ws. make Lk by adding uk to Lk1. Lk is a list of length n+1 and is linearly dependent as (by induction) Lk1 was a spanning list. Yet again, there is some removable wjspan Lk[1..j1] (linear dependence lemma) and it can't be any one of the us as they are all preceding by other us which are all linearly independent. Remove wj to obtain Lk.
  • At the end of this process, we obtain Lm, which is a spanning list, obtained by removing (at least) m number of ws. This means than mn.

All basis have the same length, allowing the definition of a dimension (finite dimension vector space obviously)

if V is a vector space, and L=<ui> and L=<vi> and L,L are both linearly independent and spanning lists (they're both basis), then they have the same length
Moreover, any linearly independent list of this length is a basis of V.

proof: let |L|=m, |L|=n. Apply the above lemma, treating L as a linearly independent list and L as a spanning list, to get mn. Now do it the other way around to get mn and hence m=n:=dim(V).
For the second part, let L=<ei> be a linearly independent list with length n=dim(V). Do the replacement process, treating L are linearly independent, and a basis L as the spanning list, both of the same length. This allows us to add an element of L to L and remove an element which was in L, and still maintaining that the list is spanning throughout the process. so we modify L to become L while maintaning "spanning Ness" at each step. Therefore L is spanning. Hence L is a basis.


Rank Nullity theorem

Let U,V be vector spaces over F and f:UV a linear map. Then, dim(U)=rank(f)+dim(kern(f)).

proof: let K=<ki> be a basis for Kern(f) of length m. let E=<ei> be a basis of U of length n=dim(U). Using the process described in the above lemma, treating K as the linearly independent list, E as the spanning list, create L=k1,,km,em+1en. This list spans U. moreover, since E and K are both linearly independent, at each step, after adding a ki when we remove an ej we have a linearly independent list (this can be again shown by induction on the same process). Therefore L is a linearly independent list. Since it is also of length n=dim(U), L is a basis of U. Hence f(L):=f(k1),f(km),f(em+1),f(en) spans V (refer to def of linear map). Since any ktKern(f) each f(ki=0). Therefore, fL:=f(em+1),f(en) still spans V. We claim fL is linearly independent.
Otherwise we have 0V=j=m+1najf(ej)=f(j=m+1najej). This means that t=j=m+1najejKern(f), and hence, tspan(K).If t0U, then t can be written as a linear combination in two ways, one using just K and other one using just E[m+1..n]. Subtracting the two, we get 0=j=1mλjkjj=m+1najej. so if even one aj0F, then we contradict the linear independence of L (and even E for that matter). Hence, bubbling back, fL is linearly independent, as each aj has to be zero whenever 0V=j=m+1najf(ej).


In pratice, if M is a matrix representing a linear map from U to V, then by solving for the kernel/nullspace of U by Mx=0 and finding the dimension of this nullspace, we can say that the rank/dimension of the image of this linear map M is equal to the dimension of U minus the dimension of the nullspace/kernel of M.

Invertible linear maps

Linear maps from smaller to larger dimensions are not surjective

Let U,V be vector spaces, dim(U)=m<dim(V)=n, f:UV a linear map. Since dim(f(U))=rank(f)dim(U)=m<dim(V)=n, f(U) cannot span V. Hence f is not surjective. (refer definition of rank of linear map, or use rank-nullity theorem to conclude dim(f(U))dim(U)).

Linear maps from larger to smaller dimensions are not injective

Let U,V be vector spaces, dim(U)>dim(V) and f:UV a linear map. Using rank-nullity theorem, we have dim(U)=dim(f(U))+dim(kern(f)) Hence, dim(kern(f))=dim(U)dim(f(U))>dim(V)dim(f(U))0 (the rightmost inequality owing to the fact that f(U) lives in V so its dimension is at most V), Finally, dim(kern(f))>0, which means that the kernel of f is not the trivial subspace {0}. Therefore, there are at least two vectors in the kernel, both of which (by definition) map to zero in V

Combining the two above ideas, and stating with a little flair, we have:

Invertible linear maps

A linear map f:UV is bijective if and only if dim(U)=dim(V) and Kern(f)={0}.

proof: Necessity of the two conditions are already clear from the above two theorems. We just have to show sufficiency.
Well, given Kern(f)={0} and dim(U)=dim(V), suppose f is not injective. then there exists vectors uv such that f(u)=f(v), since the two vectors are not equal, at least one of them is non zero. However, it follows that f(uv)=0V. hence uv0Kern(f), contradicting that kern(f) = {0}. Moreover, using rank nullity, along with dim(Kern(f))=0, We see that dim(f(U))=dim(U)=dim(V). It is a trivial fact that the 0 and highest dimensional subspaces, are uniquely the trivial subspace and the original space itself respectively. Therefore f(U)=V. showing surjectivity.


Maps on the same vector space.

Eigenvectors of a linear map

Let f be a linear map on a vector space V (to itself). Then the Eigenvectors of f : Ef={vV:f(v)=λv,λF}. The λ here is called an eigenvalue of f. The set of all eigenvalues of f is called the spectrum of f and denoted Spec(f). If we define Eλf as the set that contains all vectors v for which f(v)=λv ofr that particular lambda, then we have Ef=λspec(f)Eλf.
For any particular λspec(f) and a linear map f, Eλf{0} is a subspace of V: Notice that f(v1)=λv1, f(v2)=λv2f(v1+v2)=λ(v1+v2), and f(v)=λvf(cv)=c(λv)=λ(cv).

The set of eigenvectors of f,Ef itself, without fixing those eigenvectors with a particular eigenvalue IS NOT ALWAYS A SUBSPACE of V!!!

For example just think about f(v1)=λ1v1,f(v2)=λ2v2 then, f(v1+v2)=λ1v1+λ2v2.

This is pretty cool, so when we transform a vector space V, there are subspaces of V which contain vectors that each just get scaled by a particular amount λ under the transform.
The eigenvalue and eigenvector sections are pure gold. Read them :) (LADR)

The idea is to have Invariant subspaces :

Operator, invariant subspace

If V is a vector space, a linear map f on V (to V) is called an operator. A subspace U is invariant under f if f(u)U for each uU. In other words, the restriction f|U is an operator on U.

Let's talk about one dimensional Invariant subspaces, pick a particular vV. Let U={λv:λF} . Then, if f is an invariant on U, then for any uU , fuU. But any u=λv for some λF. since f is linear, fλv=λf(v). This means that the map f is defined by the image of v itself, and that image is in U therefore there has to be some unique scalar κ for which f(v)=κv. (note that U=span(v)).
Conversely if there exists an operator f such that for some vV, f(v)=κv for some scalar κF, then span(v) is an invariant subspace under f.

eigenvalue

if f is an operator on V, then λ is an eigenvalue of f if there exists a non zero vV such that f(v)=λv. Writing f as a matrix T, We have (TλI)v=0. Since v is non zero, the linear map (TλI)fλι has a non trivial kernel/null space. As we know, (from the discussion on invertibility), This is equivalent to saying, fλι is not injective. Using rank nullity theorem, Since the kernel is non trivial, the dimension of (fλι)(V) is smaller than the dimension V which means that this map is also not surjective. Hence finally this map is not invertible.

Linearly independent eigen vectors

If f is an operator on V, if we pick a list of distinct eigenvectors, each associated with it's own distinct eigen value, then that list is linearly independent. That is if L=<vi> such that viEλif, then L is linearly independent.

proof: since each vi is non zero (by definition), pick the smallest subset of L, L=<vi> of length m2, such that there exists a linear combination G = \sum_{i=1}{#ma_}{i}\mathbf{v^*_{i}} = \mathbf{0} such that each ai is non zero (if an ai is zero, just drop it).
Using matrix notation apply Tλm to G, since (Tλm)vm=0, we obtain a smaller sublist G (of length m-1) that is also linearly dependent, contradicting the minimality of G.


The number of eigenvalues is at most the dimension of the space

if f is an operator on V, then the number of distinct eigenvalues of f is at most dim(V)

proof: Pick a list L of eigenvectors, one for each distinct eigenvalue. Then since this list is linearly independent (form the theorem above), it's length is at most dim(V).