Revision Notes - MA2101
Revision Notes - MA2101
Revision Notes - MA2101
Ma Hongqiang
April 27, 2017
Contents
1 Vector Spaces over a Field 2
2 Vector Subspaces 6
11 Problems 59
1
1 Vector Spaces over a Field
Definition 1.1 (Field, Rings, Groups).
Let F be a set containing at least two elements and equipped with the following two binary
operations +(the addition, or plus) and ×(the multiplication, or times), where F × F :=
{(x, y) | x, y ∈ F} is the product set of F with iteself:
+ :F × F → F
(x, y) 7→ x + y;
× :F × F → F
(x, y) 7→ x × y;
Axiom(0) Addition and multiplication are well defined on F in the sense that:
∀x ∈ F, ∀y ∈ F ⇒ x + y ∈ F
∀x ∈ F, ∀y ∈ F ⇒ xy ∈ F
Namely, the operation map +(resp. ×) takes element (x, y) in the domain F × F to some ele-
ment x + y(resp. xy) in the codomain F. The quintuple (F, +, 0; ×, 1) with two distinguished
elements 0 (the additive identity) and 1 (the multiplicative identity), is called a field if the
following Eight Axioms (and also Axiom (0)) are satisfied.
(1) Existence of an additive identity 0F or simply 0:
x + 0 = x = 0 + x, ∀x ∈ F
(x + y) + z = x + (y + z), ∀x, y, z ∈ F
x + (−x) = 0 = (−x) + x
x1 = x = 1x, ∀x ∈ F
xx−1 = 1 = x−1 x.
2
(7) Distributive Laww:
(x + y)z = xz + yz, ∀x, y, z ∈ F
z(x + y) = zx + zy, ∀x, y, z ∈ F
x + y = y + x, ∀x, y ∈ F
xy = yx, ∀x, y ∈ F
The triplet (F, +, ×) with only the Axioms (1)—(5) and (7)—(8) satisfied is called a (com-
mutative) ring.
The pair (F, +) with only Axioms (1)—(3) satisfied by its binary operation +, is called an
(additive) group.
Notation 1.1 (about F× ). For a field (F, +, 0; ×, 1), we use F× to denote the set of nonzero
elements in F:
F× := F \ {0}
b+x=b+y ⇒x=y
3
(2) (Killing Power of 0)
0x = 0 = x0, ∀x ∈ F
(3) In F, we have
0F 6= 1F
x−1 ∈ F×
(6) If xx00 = 1 then x and x00 are multiplicative inverse to each other.
+:V ×V →V
(v1 , v2 ) 7→ v1 + v2
×:V ×V →V
(c, v) 7→ cv
Axiom(0) These two operations are well defined in the sense that
∀vi ∈ V ⇒ v1 + v2 ∈ V
∀c ∈ F, ∀v ∈ V ⇒ cv ∈ V
(V, +) is a vector space over the field F if the following Seven Axioms are satisfied.
v + 0 = v = v + 0, ∀v ∈ V
(u + v) + w = u + (v + w), ∀u, v, w ∈ V
v + (−v) = 0 = (−v) + v
4
(5) (Multiplicative) Associativity:
u + v = v + u, ∀u, v ∈ V
5
2 Vector Subspaces
Definition 2.1 (Subspace).
Let V be a vector space over a field F. A non-empty subset W ⊆ V is called a vector
subspace of V if the following two conditions are satisfied:
∀wi ∈ W ⇒ w1 + w2 ∈ W
∀a ∈ F, ∀w ∈ W ⇒ aw ∈ W
(i) W is a vector subspace of V , i.e. W is closed under vector addition and scalar multi-
plication, in the sense of Definition of Subspace.
∀ao ∈ F, ∀wi ∈ W ⇒ a1 w1 + a2 w2 ∈ W
(iii) W together with the vector addition + and the scalar multiplication ×, becomes a
vector space.
Remark: Union of subspaces may not be a subspace. However, union of subspaces is closed
under scalar multiplication.
6
3 Linear Spans and Direct Sums of Subspaces
Definition 3.1 (Linear combination and Linear Span).
Let V be a vector space over a field F. A vector v ∈ V is called a linear combination of
some vectors vi ∈ V (1 ≤ i ≤ s) if
v = a1 v1 + a2 v2 + · · · + as vs
1. U + W = Span(U ∪ W )
Note 1: Let U and W be two vector subspaces of a vector space V over a field F. Then the
following are equivalent.
2. Either U ⊆ W or W ⊆ U .
7
Definition 3.3 (Sum of many subspaces).
Let V be a vector space over a field F and let Wi (1 ≤ i < s) be vector subspaces of V . The
subset
= W1 + · · · + Ws
s s
X
X
Wi = { wi | wi ∈ W }
i=1 i=1
= {w1 + · · · + ws | wi ∈ Wi }
W1 ∩ W2 = {0}
1. We have
W1 + W2 = W1 ⊕ W2
i.e., W1 + W2 is a direct sum of W1 , W2 ,i.e., W1 ∩ W2 = {0}
w = w1 + w2
8
Theorem 3.5 (Equivalent Direct Multiple Sum Definition).
i (1 ≤ i ≤ s, s ≥ 2) be vector subspaces of a vector space V over a field F. Set
Let WP
W := si=1 Wi . Then the following are equivalent.
1. We have
W1 + · · · + Ws = W1 ⊕ · · · ⊕ Ws
Ps
i.e., i=1 Wi is a direct sum of Wi .
2. !
X
Wi ∩ Wl = {0} (∀1 ≤ l ≤ s)
i6=j
w = w1 + · · · + ws
9
4 Linear Independence, Basis and Dimension
Definition 4.1 (Linear (in)dependence).
Let V be a vector space over a field F. Let T be a (not necessarily finite) subset of V and
let
S = {v1 , . . . , vn }
be a finite subset of V .
(1) We call S a linear independent set or L.I., if the vector equation below
x1 v1 + · · · + xm vm = 0
(x1 , . . . , xm ) = (0, . . . , 0)
(2) We call S a linear dependent set or L.D. if there are scalars a1 , . . . , am in F which
are not all zero (i.e. (a1 , . . . , am ) 6= (0, . . . , 0)) such that
a1 v1 + · · · + am vm = 0
(3) The set T is a linearly independent set if every non-empty finite subset of T is
linearly independent. The set T is a linearly dependent set if at least one non-
empty finite subset of T is linearly dependent.
Theorem 4.1 (L.D./L.I. Inheritance).
(1) Let S1 ⊆ S2 . If the smaller set S1 is linearly dependent then so is the larger set S2 .
Equivalently, if the larger set S2 is linearly independent then so is the smaller set S1 .
a1 , . . . , ak−1 , ak+1 , . . . , am
10
(3) Suppose that S = {v1 } (a single vector). Then S is linearly dependent if and only if
v1 = 0. Equivalently, S is linearly independent if and only if v1 6= 0.
(4) Suppose that S = {v1 , v2 } (two vectors). Then S is linearly dependent if and only if
one of v1 , v2 is a scalar multiple of the other. Equivalently, S is linearly independent
if and only if neither one of v1 , v2 is a scalar multiple of the other.
Definition 4.3 (Basis, (in)finite demension).
Let V be a nonzero vector space over a field F. A subset B of V is called a basis if the
following two conditions are satisfied.
(1) (Span) V is spanned by B: V = Span(B)
dimF V := |B|
dim{0} = 0
v = a1 v1 + · · · + an vn
= Span{v1 } ⊕ · · · ⊕ Span{vn }
V
= Fv1 ⊕ · · · ⊕ Fvn
dimF V = B1 ≤ B
11
(2) Let B2 be a maximal linearly independent subset of B: first B2 is L.I. and secondly
every subsets B3 of B larger than B2 is L.D. Then B2 is a basis of V = Span(B).
(1) Suppose that S is a subset of V with |S| > n = |B|. Then S is L.D.
(2) Suppose that T is a subset of V with |T | < n. Then T does not span V .
(3) Suppose that B 0 is another basis of V . Then |B 0 | = |B|. So the dimension dimF V (=
|B|) of V depends only on V , but not on the choice of its basis.
In other words, dimF V is well defined.
B ∪ {w}
is a L.I. subset of V .
w|B|+1 , · · · , wn
(1) B is a basis of V .
12
(1) Suppose that B is a basis of V . Decompose it as a disjoint union
a a
B = B1 B2 · · · Bs
V = W1 ⊕ · · · ⊕ Ws
(3) In particular, if
V = W1 ⊕ · · · ⊕ Ws
is a direct sum, then
s
X
dimm athbbF V = dimF Wi
i=1
13
5 Row Space and Column Space
Definition 5.1 (Column/Row Space, Nullspace, Nullity, Range of A).
Let
a11 · · · a1n
A = (ai j) = ... .. ..
. .
am1 · · · amn
be an m × n matrix with entries in a field F. Let
a11 a1n
Col(A) := Span{c1 := ... , . . . , cn := ... }
am1 amn
be the column space of A, and let
R(A) := Span{r1 := (a11 , . . . , a1n ), . . . , rm := (am1 , . . . , amn )}
be the row space of A so that we can write
r1
A = (c1 , . . . , cm ) = ...
rm
The range of A is defined as
R(A) = {AX | X ∈ Fnc }
The nullity of A is defined as the dimension of the nullspace or kernel
Ker(A) := Null(A) := {X ∈ Fnc | AX = 0}
i.e.
nullity(A) = dim Null(A)
Theorem 5.1 (Rank of Matrix, Matrix Dimension Theorem).
14
Theorem 5.2 (L.I. vs L.D.).
In previous theorem, suppose that m = n so that A is a square matrix of order n. Then the
following are equivalent.
(1) A is an invertible matrix, i.e. A has a so called inverse A−1 ∈ Mn (F) such that
AA−1 = In = A−1 A
15
(2) The column vectors
ai1 , . . . , ais , aj1 , . . . , ajt
of A are linearly dependent if and only if the corresponding column vectors
Col(A) = Span{a1 , . . . , an }
bi1 , . . . , bis
Col(B) = Span{b1 , . . . , bn }
(5) If
B1 := {ai1 , . . . , ais }
is a maximal L.I. subset of the set
C = {a1 , . . . , an }
i1 , . . . , i s
Then
ai1 , . . . , ais
forms a basis of the column space Col(A) of A.
16
(7) The row space of A and B are identical
R(A) = R(B)
j1 , . . . t
Then
b0j1 , . . . , b0jt
form a basis of the row space R(A) = R(B) of A.
17
6 Quotient Spaces and Linear Transformations
Definition 6.1 (Sum of subsets of a space).
Let V be a vector space over a field F, and let S and T be subsets (which are not necessarily
subspaces) of V . Define the sum of S and T as
S + T := {s + t | s ∈ S, t ∈ T }
(S1 + S2 ) + S3 = S1 + (S2 + S3 )
(2) Commutativity
S1 + S2 = S2 + S1
W + {0} = W, W +W =W
S+W =W ⇔S ⊆W
v̄ := v + W
(1) v + W = W ,i.e.,v̄ = 0̄
(2) v ∈ W
18
(3) v + W ⊆ W
(4) W ⊆ v + W
v¯1 = v¯2 ⇔ v1 − v2 ∈ W
f : U → V /W
u 7→ ū = u + W
V /W := {v̄ = v + W | v ∈ V }
+ : V /W × V /W → V /W
¯ v2
(v¯1 , v¯2 ) 7→ v¯1 + v¯2 := v1 +
× : F × V /W → V /W
(a, v¯1 ) 7→ av¯1 := av
¯1
(1) The binary addition operation and scalar multiplication operation on V /W is well
defined.
(2) V /W together with these binary addition and scalar multiplication operations, becomes
a vector space over the same field F, with the zero vector
0V /W = 0¯V = w̄
for any w ∈ W .
Definition 6.4 (Linear transformation, and its Kernel and Image; Isomorphism).
Let Vi be two vector spaces over the same field F. A map
ϕ : V1 → V2
19
is called a linear transformation from V1 to V2 if ϕ is compatible with the vector addition
and scalar multiplication on V1 and V2 in the sense below:
ϕ(v1 + v2 ) = ϕ(v1 ) + ϕ(v2 )
ϕ(av) = aϕ(v)
When ϕ : V → V is a linear transformation from V to itself, we call ϕ a linear operator on
V.
A linear transformation is called an isomorphism if it is a bijection. In this case, we denote
V1 ' V2
Remark: If T : V → W is a linear transformation, then T (0V ) = 0W .
Remark:(Direct sum vs quotient space)
Let V = U ⊕ W , where U, W are subspaces of V . Then the map below is an isomorphism.
f : U → V /W
u 7→ ū = u + W
Theorem 6.5 (Equivalent Linear Transformation definition).
Let ϕ : V1 → V2 be a map between two vector spaces Vi over the same field F. The the
following are equivalent.
(1) ϕ is a linear transformation.
(2) ϕ is compatible with taking linear combination in the sense below:
ϕ(a1 v1 + a2 v2 ) = a1 ϕ(v1 ) + a2 ϕ(v2 )
for all ai ∈ F, vi ∈ V .
Theorem 6.6 (Evaluate T at a basis).
Let V be a vector space over a field F and with a basis B = {u1 , u2 , . . .}. Let T : V → W
be a linear transformation.
Then T is uniquely determined by its valuations T (ui ) (i = 1, 2, . . .) at the basis B.
Namely, if T 0 : V → W is another linear transformation such that T 0 (ui ) = T (ui )∀i, then
they are equal: T 0 = T .
Theorem 6.7 (Quotient map).
Let V be a vector space over a field F. Let W be a subspace V and V /W . One verifies that
γ is surjective and
ker(γ) = W
Theorem 6.8 (Image being a vector subspace).
Let
ϕ:V →W
be a linear transformation between two vector spaces over the same field F. Let V1 be a
vector subspace of V . Then the image of V1 :
T (V1 ) = {T (u | u ∈ V1 }
is a vector subspace of W .
In particular, T (V ) is a vector subspace of W .
20
Theorem 6.9 (Subspace vs. Kernel).
Let V be a vector space over a field F.
(1) Suppose that
ϕ:V →U
is a linear transformation. Then the kernel ker(ϕ) is a vector subspace of V .
(2) Conversely, suppose W is a vector subspace of V . Then there is a linear transformation
ϕ:V →U
such that
W = ker(ϕ)
ϕ:V →W
ϕ̄ : V / ker(ϕ) ' U
21
(3) a
B1 {ur+1 , . . . , un }
is a basis of V if and only if the cosets
{ur+1
¯ , . . . , u¯n }
is a basis of V /W .
Theorem 6.13 (Goodies of Isomorphism).
Let ϕ : V → W be an isomorphism and let B be a subset of V . Then we have:
(1) If there is a relation
r
X s
X
ai vi = a i vi
i=1 i=r+1
(4) We have
ϕ(Span(B)) = Span(ϕ(B))
In particular,
ϕ(Span{v1 , . . . , vs }) = Span{ϕ(v1 ), . . . , ϕ(vs )}
dim V = dim W
22
Theorem 6.15 (Dimension Theorem).
Let ϕ : V → W be a linear transformation between vector spaces over a field F. Then
(1) ϕ is an isomorphism
(2) ϕ is an injection
(3) ϕ is an surjection
23
7 Representation Matrices of Linear Transformations
Definition 7.1 (Coordinate vector).
Let V be vector space of dimension n ≥ 1 over a field F. Let
B = BV = (v1 , . . . , vn )
be a basis of V . Every vector v ∈ V can be expressed as a linear combination
v = c1 v1 + · · · + cn vn
and this expression is unique. We gather the coefficients ci and form a column vector
c1
..
[v]B := . ∈ Fnc
cn
which is called the coordinate vector of v related to basis B.
One can recover v from its coordinate vector [v]B :
v = B[v]B
Theorem 7.1 (Isomorphism V → Fnc ).
Let V be an n-dimensional vector space over a field F and with a basis B = {v1 , . . . , vn }.
Show that the map
ϕ : V → Fnc
v 7→ [v]B
is an isomorphism between the vector space V and Fnc .
Theorem 7.2 (Representation matrix).
Let
T :V →W
be a linear transformation between vector spaces over a field F. Let
B = {v1 , . . . , vn }
be a basis of V , and
BW = {w1 , . . . , wm }
be a basis of W . Let A ∈ Mm×n (F). Then the following three conditions on A are equivalent.
[T (v)]BW = A[v]B
A = ([T (v1 )]BW , . . . , [T (vn )]BW )
(T (v1 ), . . . , T (vn )) = (w1 , . . . , wm )A
We denote the above matrix A as
[T ]B,BW := A = ([T (v1 )]BW , . . . , [T (vn )]BW )
and call it the representation matrix of T relative to B and BW .
24
Theorem 7.3 (Linear transformation theory = Matrix theory).
For every matrix
A ∈ Mm×n (F)
there is a unique linear transformation
T :V →W
such that the representation matrix
[T ]B,BW = A
Consequently, the map
ϕ : HomF (V, W ) → Mm×n (F)
T 7→ [T ]B,BW
is an isomorphism of vector spaces over F.
Theorem 7.4 (Close relation between the space of vectors and space of their coordinates).
Let
T :V →W
be a linear transformation between the vector spaces V and W over the same field F, of
dimensions n and m, respectively. Let
BV := (v1 , . . . , vn )
be a basis of V , and
BW := (w1 , . . . , wm )
be a basis of W . Let
Mm×n (F) 3 A := [T ]BV ,BW = (a1 , . . . , an )
Then the following are isomorphisms:
ϕ : Ker(T ) → Null(A)
v 7→ [v]BV ,
ψ : Null(A) → Ker(T )
X 7→ BV X,
ξ : R(T ) → R(TA ) = col.sp. of A = Span{a1 , . . . , an }
η : R(TA ) → R(T )
Y 7→ BW Y = (w1 , . . . , wm )Y
Below are some consequences of the isomorphism above
(1) The subset
{X1 , . . . , Xs }
of Fnc is a basis of Null(A) if and only if the vectors
BV X1 , . . . , BV XS
of V forms a basis of Ker(T ).
25
(2) The subset
{Y1 , . . . , Yt }
of Fm
c is a basis of R(TA ) if and only if the vectors
BW Y1 , . . . , BW Yt
R(T ) = Span{BW a1 , . . . , BW an }
T −1 : W → V
26
Theorem 7.8 (Equivalent transition matrix definition).
Let V be a vector space over a field F and of finite dimension n ≥ 1. Let
B = (v1 , . . . , vn )
and
B 0 := (v10 , . . . , vn0 )
be two bases of V . Let P ∈ Mn (F). Then the following are equivalent.
(1)
P = ([v10 ]B , . . . , [vn0 ]B )
(2)
B 0 = BP
B = (v1 , . . . , vn )
and
B 0 := (v10 , . . . , vn0 )
be two bases of V . Then
[T ]B 0 = P −1 [T ]B P
where
P = PB 0 →B
A1 ∼ A2 ⇒ |A1 | = |A2 |
27
Definition 7.3 (Determinant/Trace of a linear operator).
Let
T :V →V
be a linear operator on a finite-dimensional vector space V .
We define the determinant det(T ) of T as
det(T ) := det([T ]B )
pA (x) : = |xIn − A|
= xn + bn−1 xn−1 + · · · + b1 x + b0
(2) Let
T :V →V
be a linear operator on an n-dimensional vector space V . Set
A := [T ]B
pT (x) : = |xIn − A|
= xn + bn−1 xn−1 + · · · + b1 x + b0
Theorem 7.12.
For A ∈ Mn (F), we have
Tr(A) = −bn−1
det(A) = (−1)n pA (0)
28
8 Eigenvalue and Cayley-Hamilton Theorem
Definition 8.1 (Eigenvalue, eigenvector).
Assume that
λ∈F
T :V →V
T (v) = λv
(arabic*) For an n × n matrix A in Mn (F), a nonzero column vector u in Fnc is called an eigen-
vector of A corresponding to the eigenvalue λ ∈ F of A if
Au = λu
T :V →V
0 6= v ∈ Ker(λIV − T )
(4) The matrix λIn − [T ]B is not invertiblem i.e. the matrix equation
(λIn − [T ]B )X = 0
29
(5) λ is a zero of the characteristic polynomial pT (x) of T
pT (λ) = |λIn − [T ]B | = 0
p(x) = (x − λ1 ) · · · (x − λn )
T :V →V
on an n-dimensional vector space V over the field F. The subspace (of all the eigenvectors
corresponding to the eigenvalue λ, plus 0V ):
Vλ : = Vλ (T )
: = Ker(λIV − T )
= {v ∈ V | T (v) = λv}
(2) The algebraic multiplicity of the eigenvalue λ of T is defined to be the largest positive
integer k such that (x − λ)k is a factor of the characteristic polynomial pT (x),i.e.
30
Theorem 8.2 (Eigenspace of T and [T ]B ).
Let
T :V →V
be a linear operator on an n-dimensional vector space V with a basis B. Set
A := [T ]B
The map
f : Ker(T − λIV ) → Null(A − λIn )
w 7→ [w]B
gives an isomorphism. In particular,
1. The subset
{u1 , · · · , us }
of V is a basis of the eigenspace Vλ (T ) of T .
2. THe subset
{[u1 ]B , cdots, [us ]B }
of Fnc is a basis of the eigenspace Vλ ([T ]B ) of the representation matrix [T ]B of T
relative to a basis B of V .
1. The subset
{X1 , · · · , Xs }
of Fnc is a basis of the eigenspace Vλ ([T ]B ) of the representation matrix [T ]B of T
relative to a basis B of V .
2. The subset
{BX1 , · · · , BXs }
of V is a basis of the eigenspace Vλ (T ) of T .
31
be some distinct eigenvalues of a linear operator T on a vector space V over a field F. Then
the sum of eigenspaces
Xk
W := Vλi (T )
i=1
= Vλi (T ) + · · · + Vλk (T )
is a direct sum:
W = ⊕ki=1 Vλi (T )
= Vλ1 (T ) ⊕ · · · ⊕ Vλk (T )
T s := T ◦ · · · ◦ T (s times)
Define r
X
f (T ) = ai T i
i=0
f (T ) : V &toV
v 7→ f (T )(v)
32
be a linear operator on a vector space V over a field F and with a basis
B = {u1 , . . . , un }
Let
f (x), g(x) ∈ F[x]
be polynomials. We have
(1)
[f (T )]B = f ([T ]B )
(3) Commutativity:
f (T )g(T ) = g(T )f (T )
f (P −1 AP ) = P −1 f (A)P
(5) If
S:V →V
is an isomorphism with inverse isomorphism
S −1 : V → V
Then
f (S −1 T S) = S −1 f (T )S
T (W ) := {T (w) | w ∈ W }
i.e.,
T (w) ∈ W ∀w ∈ W
In this case, define the restriction of T on W as:
T |W :W →W
w 7→ T (w)
33
Theorem 8.6 (Kernels and Images of Commutative Operators).
Let Ti : V → V be two linear operator commutative to each other, i.e.
T1 ◦ T2 = T2 ◦ T1
as maps. Namely,
T1 (T2 (v)) = T2 (T1 (v)) (∀v ∈ V )
Both
Ker(T2 ), ima(T2 )
are T1 -invariant subspaces of V .
Theorem 8.7 (Evaluate T on a basis of a subspace).
Let
T :V →V
be a linear operator on a vector space V over a field F. Let W be a subspace of V with a
basis BW = {w1 , w2 , · · · }. Then W is T -invariant, if and only if
T (BW ) ⊆ W
Theorem 8.8 (T -cyclic subspace).
Let
T :V →V
be a linear operator on a vector space V over a field F. Fix a vector
0 6= w1 ∈ V
1. The subspace
W := Span{T s (w1 | s ≥ 0}
of V is T -invariant.
W is called the T -cyclic subspace of V generated by w1 .
2. Suppose that V is finite-dimensional. Let s be the smallest positive integer such that
T s (w1 ) ∈ Span{w1 , T (w1 ), . . . , T s−1 (w1 )}
We have
dimF W = s
and
B := {w1 , T (w1 ), . . . , T s−1 (w1 )}
is a basis of W .
3. In (2), if
T s (w1 ) = c0 w1 + c1 T (w1 ) + · · · + cs−1 T s−1 (w1 )
for some scalars ci ∈ F, then the characteristic polynomial of the restiction operator
T | W on W is
pT |W (x) = −c0 − c1 x − · · · − cs−1 xs−1 + xs
34
Theorem 8.9 (Characteristic Polynomial of the Restriction Operator).
Let
T :V →V
be a linear operator on a vector space V over a field F and of dimension n ≥ 1. Let W be
a T -invariant subspace of V . Then the characteristic polynomial pT |W (x) of the restriction
operator T | W on W is a factor of the characteristic polynomial pT (x) of T , i.e.
B = BW ∪ B2
for some λi ∈ F. Then there is an basis B of V such that the representation matrix [T ]B is
upper triangular.
Theorem 8.12 (Characteristic Polynomials of Direct Sums).
Let
T :V →V
be a linear operator on an n-dimensional vector space V over a field F. Suppose that there
are T -invariant subspaces
Wi (1 ≤ i ≤ r)
of V such that V is the direct sum
V = ⊕ri=1 Wi
35
of Wi .
Then the characteristic polynomial pT (x) of T is the product:
r
Y
pT (x) = pT |Wi (x)
i=1
V = ⊕ri=1 Wi
of some T -invariant subspaces Wi , if and only if every set of bases Bi of Wi gives rises to a
basis a a
B = B1 ··· Br
of V such that the representation matrix of T relative to B is in the form
A1 0 · · · 0
0 A2 · · · 0
[T ]B = ..
.. . . ..
. . . .
0 0 0 Ar
T :V →V
on an n-dimensional vector space V over a field F and with a basis B. Then T satisfies the
equation pT (x) = 0, i.e.
pT (T ) = 0IV
which is the zero map on V .
36
9 Minimal Polynomial and Jordan Canonical Form
Definition 9.1 (Minimal Polynomial).
Let
T :V →V
be a linear operator on an n-dimensional vector space over a field F. A nonzero polynomial
m(x) ∈ F[x]
is a minimal polynomial of T if it satisfies:
1. m(x) is monic,
2. Vanishing condition:
m(T ) = 0IV
3. Minimality degree condition:
Whenever f (x) ∈ F[x] is another nonzero polynomial such that f (T ) = 0IV , we have
deg(f (x)) ≥ deg(m(x))
37
Theorem 9.4.
The set of zeros of pT (x) and that of mT (x) are identical.
Definition 9.2 (Jordan Block).
Let λ be a scalar in a field F. The matrix below
λ 1 0 0 ···
0
λ 1 0 ···
J := Js (λ) = 0
0 λ 0 ···
∈ Ms (F)
.. .. ....
..
. . . . .
0 0 0 ··· λ
The eigenspace
Vλ (J) = Span{e1 }
has dimension 1, i.e. the geometric multiplicity of λ is 1, but the algebraic multiplicity of λ
is s.
Definition 9.3 (Jordan Canonical Form).
Let λ be a nonzero scalar in a field F. Let
s1 ≤ s2 ≤ · · · ≤ se
38
has dimension equal to e.
The geometric multiplicity of the eigenvalue λ of A(λ) is dim Vλ (A(λ)) = e. And we have,
e≤s
se (λi )
Jordan blocks in J.
Now we have
k
Y
pJ (x) = (x − λi )s(λi )
i=1
Yk
mJ (x) = (x − λi )se (λi )
i=1
39
The eigenspace Vλi (J) has dimension e(λi ) and is spanned by the e(λi ) ectors corresponding
to the first columns of the e(λi ) Jordan blocks in A(λi ).
We also have
dim Vλi (J) = e(λi ) ≤ s(λi )
Sometimes, a block diagonal J below
Js1 (λ1 ) 0 0 ··· 0
0
Js2 (λ2 ) 0 ··· 0
J =
0 0 Js3 (λ3 ) ··· 0
.. .. .. ... ..
. . . .
0 0 0 · · · Jsr (λr )
is also called a Jordan Canonical Form, where each Jsi (λi ) is a Jordan block with eigen-
value λi ∈ F, but these λi ’s may not be distinct.
Assume that there are exactly k distinct elements in the set
{λ1 , . . . , λr }
e(λmi )
be the number of Jordan blocks (among the r such in J) with eigenvalue of the same λmi ;
among these e(λmi ) Jordan blocks, the largest is of order say
se (λmi )
The eigenspace Vλmi (J) has dimension e(λmi ) and is spanned by e(λmi ) vectors corresponding
to the first columns of these e(λmi ) Jordan blocks.
Also,
Yk
pJ (x) = (x − λi )s(λmi )
i=1
Yk
mJ (x) = (x − λi )se (λmi )
i=1
40
Theorem 9.5 (Jordan Canonical Form of a Linear Operator).
Let V be a vector space of dimension n over a field F and
T :V →V
a linear operator with characteristic polynomial pT (x) and minimal polynomial mT (x) as
follows
pT (x) = (x − λ1 )n1 · · · (x − λk )nk
mT (x) = (x − λ1 )m1 · · · (x − λk )mk
where
λ1 , . . . , λ k
are distinct scalars in F.
Then there is a basis B of V such that the representative matrix [T ]B equals a Jordan
canonical form J ∈ Mn (F), with
ni = s(λi ), mi = se (λi )
T :V →V
B 0 = (v1 , . . . , vn )
P −1 AP = J
(2) THere is an invertible matrix P ∈ Mn (F) such that the representation matrix [T ]B
relative to the new basis
B = B0P
is J, i.e.
[T ]B = J
41
(1) A has a Jordan canonical form J ∈ Mn (F).
In particular, if F is so called algebraically closed, then every matrix A ∈ Mn (F) and every
T on an n-dimensional vector space V over F have a Jordan canonical form j ∈ Mn (F).
1. The characteristic polynomial p(x) and the minimal polynomial m(x) have the same
zero sets.
{α ∈ F | p(α) = 0} = {α ∈ F | m(α) = 0}
Also, the multiplicity ni and mi of a zero λi of p(x) and m(x) satisfy
ni ≥ mi ≥ 1
42
exhaust all zeros of pT (x) and the characteristic polynomial of T is
pT (x) = (x − λ1 ) · · · (x − λn )
i.e.
(Ap1 , . . . , Apn ) = (λ1 p1 , . . . , λn pn )
i.e. each pi is an eigenvector of A corresponding to the eigenvalue λi .
Suppose that A = [T ]B 0 . Then the condition above is equivalent to
where vi = B 0 pi ∈ V with
[vi ]B 0 = pi
i.e.
T (vi ) = λi vi
i.e.
λ1 0 0 ··· 0
0 λ2 0
··· 0
T (v1 , . . . , vn ) = (v1 , . . . , vn ) 0 0 λ3
··· 0
.. .. .. .. ..
. . . . .
0 0 0 · · · λn
43
i.e.
λ1 0 0 ··· 0
0 λ2 0
··· 0
[T ]B = 0 0 λ3
··· 0
.. .. .. .. ..
. . . . .
0 0 0 · · · λn
where
B = (v1 , . . . , vn ) = B 0 P
is a basis of V since
([v1 ]B 0 , . . . , [vn ]B 0 ) = (p1 , . . . , pn )
is a basis of column vector space Fnc .
Theorem 9.10.
(1) T is diagonalisable if and only if the representation matrix [T ]B 0 relative to every basis
B 0 is diagonalisable.
T :V →V
3. A basis
B = (v1 , . . . , vn )
of V is formed by eigenvectors vi of T .
44
4. There are n linearly independent eigenvectors vi of T .
P = (p1 , . . . , pn )
6. For the representation matrix [T ]B 0 relative to every basis B 0 of V , there are n linearly
independent eigenvectors pi of [T ]B 0 .
7. Let
λm1 , . . . , λmk
be the only distinct eigenvalues of T and let Bi be a basis of the eigenspace Vλi (T ).
Then
B = (B1 , . . . , Bk )
is a basis of V , automatically with
λm1 I|B1 | 0 0 ··· 0
0
λm2 I|B2 | 0 ··· 0
0
[T ]B = 0 λm3 I|B3 | ··· 0
.. .. .. ... ..
. . . .
0 0 0 · · · λmk I|Bk |
8. Let
λm1 , . . . , λmk
be the only distinct eigenvalues of T . Then V is a direct sum of the eigenspaces
V = Vλmi (T ) ⊕ · · · ⊕ Vλmk (T )
9. Let
λm1 , . . . , λmk
be the only distinct eigenvalues of T . Then
k
X
dim Vλmi (T ) = dim V
i=1
45
(2) The minimal polynomial m(x) is a product of distinct linear polynomials in F[x].
m(x) = (x − λ1 ) · · · (x − λk )
m(x) = (x − λ1 ) · · · (x − λk )
dim Vλi = ni
Theorem 9.13.
Let V be an n-dimensional vector space over a field F.
A linear operator
T :V →V
on V is nilpotent if
T m = 0IV
for some positive integer m.
Suppose that T has a Jordan canonical form J ∈ Mn (F). The following are equivalent.
(1) T is nilpotent.
(2) J equals some A(λ) with λ = 0.
(3) Every eigenvalue of T is zero/
(4) The characteristic polynomial of T is pT (x) = xn .
(5) The minimal polynomial of T is mT (x) = xs for some s ≥ 1.
Theorem 9.14 (Additive Jordan Decomposition).
Suppose a linera operator
T :V →V
has a Jordan canonical form in Mn (F). There are linear operators
TS : V → V
and
Tn : V → V
satisfying the following:
46
(1) A decomposition
T = Ts + Tn
(2) Ts is semi-simple.
(3) Tn is nilpotent.
(4) Commutativity
Ts ◦ Tn = Tn ◦ Ts
Ts = f (T ) Tn = g(T )
47
10 Quadratic Forms, Inner Product Spaces and Conics
Definition 10.1 (Bilinear forms).
Let V be a vector space over a field F. Consider the map H below:
H :V ×V →F
(x, y) 7→ H(x, y)
(1) H is called a bilinear form on V if H is linear in both variables, i.e., for all
xi , yj , x, y ∈ V, ai , bi ∈ F
we have
H(a1 x1 + a2 x2 , y) = a1 H(x1 , y) + a2 H(x2 , y)
H(x, b1 y1 + b2 y2 ) = b1 H(x, y1 ) + b2 H(x, y2 )
where n n
X X
HA ( x i vi , yi vj )
i=1 j=1
n X
X n
:= aij xi yj
i=1 j=1
a11 · · · a1n y1
.. . . . .
=(x1 , . . . , xn ) . . .. ..
an1 · · · ann yn
=X t AY
48
(1) Then HA is a bilinear form on V and called the bilinear form associated with A
(and relative to the basis B of V ).
(2) Conversely, every bilinear form H on V is of the form HA for some A in Mn (F). Indeed,
just set
aij = H(vi , vj ), A := (aij )
Then one can use the bilinearity of H, show that H = HA .
The matrix A is called the representation matrix of H relative to the basis of B¿
H(x, y0 ) = 0(∀x ∈ V ) ⇒ y0 = 0
P t AP = diag[d1 , . . . , dn ] =: D
49
i.e. A is congruent to a diagonal matrix D.
In this case, the bilinear form
n X
X n
H(X1 , X2 ) = aij xi yj
i=1 j=1
= X1t AX2
= Y1t DY2
d1 · · · 0 y2 1
= (y11 , . . . , y1n ) 0 . . 0 ...
.
0 · · · dn y2 n
Xn
= dj y1j y2j
j=1
Xi = P Y i
hx, yi := H(x, y)
xi , yj , x, y ∈ V, ai , bi ∈ R
we have
ha1 x1 + a2 x2 , yi = a1 hx1 , yi + a2 hx2 , yi
hx, b1 y1 + b2 y2 i = b1 hx, y1 i + b2 hx, y2 i
hx, yi = hy, xi
(3) Positivity:
For all 0 6= x ∈ V , we have
hx, xi > 0
50
Next is the complex version. Consider a function H:
V ×V →C
(x, y) 7→ H(x, y)
hx, yi := H(x, y)
xi , yj , x, y ∈ V, ai , bi ∈ R
we have
h< a1 x1 + a2 x2 , yi = a1 hx1 , yi + a2 hx2 , yi
h< x, b1 y1 + b2 y2 i = b̄1 hx, y1 i + b̄2 hx, y2 i
hx, yi = hy, xi
(3) Positivity:
For all 0 6= x ∈ V , we have
hx, xi > 0
We have
kxk ≥ 0
and
kxk == 0 ⇔ x = 0V
x⊥y
51
Definition 10.5 (Non-degenerate Inner Product).
Let (V, h, i) be an inner product space over a field F with F = R or F = C. Then the product
h, i is non-degenerate in the sense:
for every u0 ∈ V
hu0 , yi = 0(∀y ∈ V ) ⇒ u0 = 0V
and for every v0 ∈ V ,
hx, v0 i = 0(∀x ∈ V ) ⇒ v0 = 0V
Definition 10.6 (Orthonormal basis).
Let (V, h, i) be a real or complex inner product space. A basis B = (v1 , . . . , vn is called an
orthonormal basis of the inner product space V , if it satisfies the following two conditions:
(1) Orthogonality:
for all i 6= j, we have:
vi ⊥ vj i.e., hvi , vj i = 0
(2) Normalised:
for all i, we have
kvi k = 1
Namely, vi is a unit vector.
Theorem 10.3 (Gram-Schmidt Process).
Let V = Fnc with F = R or C. Employ the standard inner product h, i for V .
Let
(u1 , . . . , ur )
be a basis of a subspace W of V . Then one can apply the following Gram-Schmidt process
to get an orthonormal basis
(v1 , . . . , vr )
of W .
v10 = u1
hu2 , v10 i 0
v20 = u2 − v1
kv10 k2
k−1
X huk , v0 i
vk0 = uk − 0 2
i
i=1
kvi k
vj0
vj =
v0
j
A∗ = (A)t = (aij )t
A∗ = (At )
52
Theorem 10.4 (Adjoint matrix A∗ and inner product).
Let V = F)nc with F = R or C. Employ the standard inner product h, i for V . For a matrix
A ∈ Mn (F), we have
hAX, Y i = hX, A∗ Y i
Theorem 10.5 (Adjoint Linear Operator).
Let T : V → V be a linear operator on an n-dimensional inner product space V over a field
F. Then we have:
(1) There is a unique linear operator
T∗ : V → V
on V such that
hT (u), vi = hu, T ∗ (v)i
Such T ∗ is called the adjoint linear operator of T .
(2) Let B = (w1 , . . . , wn ) be an orthonormal basis of the inner product space V . Then
[T ∗ ]B = ([T ]B )∗
T ∗ = αIV
(2)
(a1 T1 + a2 T2 )∗ = a1 T1∗ + a2 T2∗
(3)
(T1 ◦ T2 )∗ = T2∗ ◦ T1∗
T T ∗ = IV
AAt = In
53
(3) A linear operator T over a complex inner product space is unitary if
T T ∗ = IV
AA∗ = In
T = T∗
When the field F = R, a self adjoint operator is also called a symmetric operator.
(6) A complex matrix A ∈ Mn (C) is self-adjoint if the adjoint matrix of A equals itself:
A∗ = A
T T ∗ = T ∗T
AA∗ = A∗ A
hP X, P Y i = hX, Y i
54
(4) The matrix transformation TP preserves the distance, i.e. for all X, Y in Cnc , we have
kP X − P Y k = kX − Y k
(5) The matrix transformation TP preserves the norm, i.e. for all X in Cnc , we have
kP Xk = kXk
B = (v1 , . . . , vn )
(1) If a real matrix P ∈ Mn (R) is orthogonal, then every zero of pP (x) has modulus equal
to 1. In particular, the determinant
|P | = ±1
of P has modulus q
|λ| = r12 + r22 = 1
In particular, the determinant |P | ∈ C has modulus 1.
Theorem 10.11 (Eigenvalue of self-adjoint linear operators).
(1) Suppose that a real matrix A ∈ Mn (R) is symmetric. Every zero of pA (x) is a real
number.
(2) Suppose a complex matrix A ∈ Mn (C) is self-adjoint. Every zero of pA (x) is a real
number.
(3) More generally, suppose that T is a self adjoint linear operator. Every zero of pT (x) is
a real number.
hv1 , v2 i = 0
55
Definition 10.9 (Positive/Negative definite linear operators).
Let A ∈ Mn (C) (resp. let V be an n-dimensional inner product space which is over a field
F = R or C).
(1) T is positive definite if T is self-adjoint and
hT (v), vi > 0
hT (v), vi < 0
(AX)t X = X t At X > 0
(AX)t X = X t At X < 0
A := [T ]B ∈ Mn (F)
56
Theorem 10.13. Let A ∈ Mn (C). The function H on V := Cnc
H :V ×V →C
(X, Y ) 7→ hX, Y i := (AX)t Y
P −1 AP = P t AP
W ⊥ := {x ∈ V | hx, vi = 0, ∀w ∈ W }
(3)
V = W ⊕ W⊥
57
Definition 10.10 (Quadratic Form).
Let V be a vector space over a field F. A function
K:V →F
H :V ×V →F
such that
K(x) = H(x, x)
58
11 Problems
1 Let A ∈ Mn (C) be a complex matrix of order n ≥ 9 and let
Suppose that A is self-adjoint and f (A) = 0. Find all possible minimal polynomials
mA (x) of A.
5 Let V be a finite dimensional vector space over a field F and let T be a linear operator
on V . Suppose there exists v ∈ V such that {v, T (v), . . . , T n−1 (v)} is a basis for V
where n = dim(V ).
(a) Prove that the linear operators IV , T, . . . , T n−1 are linearly independent. (Done)
(b) Let S be a linear operator on V such that S ◦ T = T ◦ S. Write
where a0 , a1 , · · · , an−1 ∈ F.
Prove that S = p(T ) where p(x) = a0 + a1 x + · · · + an−1 xn−1 .
(c) Suppose pT (x) = (x−λ1 )r1 (x−λ2 )r2 · · · (x−λk )rk where λ1 , λ2 , . . . , λk are distinct
eigenvalues of T . Find mT (x).
59