Lecture Notes
Lecture Notes
June 6, 2025
Contents
1 Algebraic structures 3
1.1 Binary operations and groups . . . . . . . . . . . . . . . . . . . . . . 3
1.2 Fields . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10
1.3 Working with complex numbers . . . . . . . . . . . . . . . . . . . . 19
1.4 Working with modular arithmetic . . . . . . . . . . . . . . . . . . . 31
2 Vector spaces 42
2.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 42
2.2 Elementary vector-space theory . . . . . . . . . . . . . . . . . . . . . 47
2.3 Subspaces . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 56
1 Algebraic structures
Definition
Examples
1. Addition and multiplication are binary operations on the set N: for all n1 ,
n2 ∈ N the quantities n1 + n2 and n1 .n2 are also elements of N. They are
also binary operations on the sets Z, Q and R.
Definitions
Examples
3
1.1 Binary operations and groups
Definition
A group is a non-empty set X equipped with a binary operation · which has the
following properties.
Examples
4
1.1 Binary operations and groups
If X is a finite set we refer to this group as the symmetric group S(X ) and its
elements as permutations. For a permutation σ of the set { x1 ,x2 , . . . ,xn } we
often use the notation
x1 x2 · · · x n
,
σ ( x1 ) σ ( x2 ) · · · σ ( x n )
so that e.g.
1 2 3
1 3 2
denotes the permutation 1 7→ 1, 2 7→ 3, 3 7→ 2 of the set {1,2,3}. This permu-
tation can also be written in cyclic notation as (1)(23) or simply (23). The
notation S({1, . . . ,n}) is often abbreviated to Sn ab, so that e.g.
S3 = {(),(123),(321),(23),(13),(12)}.
The diagram shows the six symmetries of an equiliateral triangle: the iden-
tity I, rotations R±2π/3 through ± 2π
3 and reflections T1 , T2 , T3 in the angle
bisectors.
5
1.1 Binary operations and groups
2 3
I 1 R2π/3 3 R−2π/3 2
2 3 1 2 3 1
T1 T2 T3 2
1 3
3 2 2 1 1 3
−1
One finds that I −1 = I, R2π/3 = R−2π/3 , R− 1 −1
−2π/3 = R2π/3 , T1 = T1 ,
T2−1 = T2 and T3−1 = T3 .
6
1.1 Binary operations and groups
a11 a12 · · · a1n
a21 a22 · · · a2n
A = (aij ) = .. .. .. m rows,
. . .
am1 am2 · · · amn
| {z }
n columns
i.e. we arrange the mn-tuple (aij ) i=1,...,m, in a rectangular array with m rows
j =1,...,n
and n columns. The coefficient aij is found in the i-th row and j-th column.
since
the neutral element is the zero matrix (0) whose coefficients are all zero,
Proposition
Proof
x = x · i1 , y = i2 · y
i2 = i2 · i1 , i1 = i2 · i1 .
Hence i2 = i1 .
7
1.1 Binary operations and groups
y · x = x · y = i, z · x = x · z = i.
It follows that
y = i · y = (z · x ) · y = z · ( x · y) = z · i = z.
Mappings between groups which preserve the group structure are particularly
important.
Definition
Proposition
Proof
We write both · G and · H as ·, since it is clear from the context which is meant.
8
1.1 Binary operations and groups
It follows that
i H = f (i G ) − 1 · f (i G )
= f (i G ) − 1 · f (i G ) · f (i G )
= f (i G ) − 1 · f (i G ) · f (i G )
= i H · f (i G )
= f (i G ).
Examples
Definition
9
1.2 Fields
Proposition
1. i ∈ Y,
Example
1.2 Fields
Definition
10
1.2 Fields
Remarks
1. (A1)-(A4) state that (X,+) is an abelian group, while (A5)-(A8) state that
(X \ {0},.) is an abelian group.
2. These properties, which we also call the axioms of arithmetic, state simply
that the ‘usual’ rules of arithemetic apply to a field.
1. x + y = x + z ⇒ y = z
3. x.(−y) = −( x.y)
4. x.y = x.z, x 6= 0 ⇒ y = z
5. x.y = 0 ⇒ x = 0 or y = 0
11
1.2 Fields
The set of real numbers forms a field with respect to addition and multiplication.
We usually write a.b, a.b−1 and a + (−b) as respectively ab, a/b and a − b.
Definition
The set nm o
Q= : m ∈ Z, n ∈ Z \{0}
n
of rational numbers forms a field with respect to addition and multiplication. We
establish this fact using the following definition.
Definition
A subfield (Y, + ,.) of a field (X, + ,.) is a subset Y of X which is itself a field with
respect to the binary operations + and ..
Proposition
2. 0 ∈ Y and 1 ∈ Y;
a ∈ Y ⇒ − a ∈ Y, a ∈ Y \{0} ⇒ a−1 ∈ Y.
12
1.2 Fields
Theorem
Q is a field with respect to the addition and multiplication of the real numbers.
Proof
a c a.d + b.c
+ = ∈ Q,
b d b.d
a c a.c
. = ∈ Q.
b d b.d
0 n
0= , 1= ,
n n
so that 0, 1 ∈ Q.
• Prime fields
13
1.2 Fields
on Z. The set
Z p := {[0], . . . ,[ p − 1]}
of equivalence classes forms a field (Z p , ⊕ ,⊙) with
[ a ] ⊕ [ b ] : = [ a + b ], [ a] ⊙ [b] := [ ab].
(We usually abbreviate [ a] ⊕ [b] and [ a] ⊙ [b] to respectively [ a] + [b] and [ a][b].)
a − c = q1 p, b − d = q2 p
Axioms (A1)-(A7) and (A9) are readily verified (the additive and multiplica-
tive identities are respectively [0] and [1] and the additive inverse to [ a] is
−[ a] := [− a]).
• Complex numbers
Definitions
A complex number is an ordered pair (a,b), where a and b are real numbers.
14
1.2 Fields
Lemma
The formulae
define binary operations on C with respect to which it is a field. The additive and
multiplicative identities are given by respectively (0,0) and (1,0), and
−1 a b
−(a,b) = (− a, − b), (a,b) = ,− 2 .
a2 + b2 a + b2
Lemma
Proof
1. We show that:
−(a,0) = (− a,0),
(a,0)−1 = (a−1 ,0), a 6= 0.
15
1.2 Fields
Remark
The above notation for complex numbers is a little unwieldy and is usual simpli-
fied as follows.
i.i = (0,1).(0,1)
= (−1,0)
= −1.
It follows that
We therefore write (a,b) as a + ib and work with the usual rules of arithmetic and
the additional identity i2 = −1.
Example
z1 z2 = ( x1 + iy1 )( x2 + iy2 )
= x 1 x 2 + x 1 y 2 i + x 2 y 1 i + y 1 y 2 i2
= x1 x2 − y1 y2 + ( x1 y2 + x2 y1 )i.
16
1.2 Fields
Definition
A field (K, + ,·) is called ordered if there is a relation ≺ on K with the following
properties:
Remark
(O1) and (O2) are equivalent to the statement that a total order, i.e. it is reflex-
ive, antisymmetric, transitive and connex. (O3) and (O4) assert that the order is
compatible with the axioms of arithmetic (A1)–(A9).
Notation
1. If 0 ≺ c we say: “c is positive”.
If c ≺ 0 we say: “c is negative”.
Lemma
1. x + z ≺ y + z ⇒ x ≺ y
2. x ≺ y, z w ⇒ x + z ≺ y + w
3. x ≻ 0, y 0 ⇒ x + y ≻ 0
17
1.2 Fields
4. x ≺ y ⇒ −y ≺ − x
5. x ≺ y, z w ⇒ x + (−w) ≺ y + (−z)
6. x ≺ y, z ≺ 0 ⇒ y.z ≺ x.z
7. 0 ≺ x ≺ y ⇒ 0 ≺ y−1 ≺ x −1
8. x2 ≻ 0 for each x 6= 0
9. 1 ≻ 0
11. 0 ≺ x ≺ 1 ⇒ x2 ≺ x
12. x ≻ 1 ⇒ x2 ≻ x
Examples
18
1.3 Working with complex numbers
Definitions
Proposition
1. z + z̄ = 2Re z,
2. z − z̄ = 2i Im z,
3. zz̄ = |z|2 .
Proposition
1. z1 + z2 = z̄1 + z̄2 ,
2. z1 z2 = z̄1 z̄2 ,
1 1
3. = if z1 6= 0,
z1 z̄1
4. z̄1 = z1 ,
5. |z̄1 | = |z1 |.
19
1.3 Working with complex numbers
Since a complex number is an ordered pair ( x,y) of real numbers we can represent
it as a point in the coordinate plane (the complex plane):
Im
y z = x + iy
Im z
z̄ is related to z
by a reflection in
x Re the real axis.
Re z
−y z̄ = x − iy
Im
Here
y z = x + iy x = r cos θ,
y = r sin θ,
r q
r = x 2 + y2 ,
y
θ tan θ = .
x
x Re
20
1.3 Working with complex numbers
Im Im
z z
|z| arg z
|z|
arg z
Re Re
Im Im
Re Re
arg z |z|
arg z
|z|
z z
Example
(i) {z ∈ C : |z − (1 + i)| = 3}
21
1.3 Working with complex numbers
Solution
Im
1+ i
1
3
1 Re
(ii) {z ∈ C : 1 < |z| ≤ 3} is the set of all complex numbers whose distance to the
origin is bigger than 1 or smaller than or equal to 3. It is therefore an annulus
with centre 0 and radii 1 and 3:
Im
1
Re
3
22
1.3 Working with complex numbers
(iii) {z ∈ C : |z − i| < |z + i|} is the set of all complex numbers which are closer
to i than −i. It is therefore the upper half-plane:
Im
Re
−1
(iv) {z ∈ C : −π/4 < arg z ≤ π/4} is the set of all complex numbers which make
an angle with the real axis between −π/4 and π/4. It is therefore a sector:
Im
π
4
π Re
4
Lemma
23
1.3 Working with complex numbers
z1n = r1n cos nθ1 + ir1n sin nθ1 , n = 1,2,3, . . . . (de Moivre’s theorem) (ii)
Proof
Remark
We can interpret (i) geometrically: the distance from the point z1 z2 to the origin
is the product of the distances from the points z1 and z2 to the origin, while the
angle with the real axis made by the point z1 z2 is the sum of the angles with the
real axis made by the points z1 and z2 :
Im
z1 z 2
z2
r1 r2 r2
z1
θ1+θ2 r1
θ2
θ1
Re
24
1.3 Working with complex numbers
Definitions
For z ∈ C we define
Further special functions such as tan(·), cot(·), etc. are defined from the above
functions in the usual fashion.
Proposition
We have that
cosh z = cos(iz), z ∈ C,
sinh z = −i sin(iz), z ∈ C,
eiθ = cos θ + i sin θ, θ ∈ R, (Euler’s formula)
and the basic identities for the relationships between the exponential, trigono-
metric and hypberolic functions remain valid.
Remark
25
1.3 Working with complex numbers
Remark (zeros)
1. The calculation
2. We have that
(see below).
Example
Solution
We have that
cos z = 0
1
⇔ (eiz + e−iz ) = 0
2
⇔ eiz = −e−iz
⇔ e2iz = −1.
26
1.3 Working with complex numbers
Im
π | − 1| = 1
−1 arg(−1) = π
1 Re
It follows that
e2iz = −1
⇔ e2i (x+iy) = −1
⇔ e−2y e2ix = 1eiπ
⇔ e−2y = 1, 2x = π + 2nπ, n = 0, ± 1, ± 2, . . .
π
⇔ y = 0, x = + nπ, n = 0, ± 1, ± 2, . . . .
2
We conclude that
π 3π 5π
cos z = 0 ⇔ z = ± ,± ,± ,....
2 2 2
Definition
Let n ∈ N. Every complex polynomial of degree n has at least one complex zero.
Corollary
27
1.3 Working with complex numbers
of degree n has n complex zeros z1 ,. . . ,zn (which are not necessarily distinct) and
can therefore be factorised as
Remark
If the coefficients a0 , . . . , an are real, all zeros occur in complex-conjugate pairs (as
a special case they can be real). Suppose that z⋆ is a zero. It follows that
a n ( z ⋆ ) n + a n − 1 ( z ⋆ ) n − 1 + . . . + a1 ( z ⋆ ) + a0 = 0
⇒ a n ( z ⋆ ) n + a n − 1 ( z ⋆ ) n − 1 + . . . + a1 ( z ⋆ ) + a0 = 0
⇒ ān (z̄⋆ )n + ān−1 (z̄⋆ )n−1 + . . . + ā1 z̄⋆ + ā0 = 0
⇒ an (z̄⋆ )n + an−1 (z̄⋆ )n−1 + . . . + a1 z̄⋆ + a0 = 0, (since a0 , . . . , an are real)
Definition
zn = a.
Example
Find all fifth roots of unity and interpret the result geometrically.
Solution
z5 = 1.
28
1.3 Working with complex numbers
Im
|1| = 1
1 arg(1) = 0
1 Re
It follows that
z5 = 1
⇔ r5 e5iθ = 1ei0
⇔ r5 = 1, 5θ = 0 + 2nπ, n = 0, ± 1, ± 2, . . .
2nπ
⇔ r = 1, θ= , n = 0, ± 1, ± 2, . . . .
| {z 5 }
The five values of this expression
in (−π,π ] are 0, 2π 4π 2π 4π
5 , 5 ,− 5 ,− 5
Im
2πi
e 5
4πi
e 5
2πi
e− 5
The following lemma is proved using the method in the above example.
29
1.3 Working with complex numbers
Lemma
For n ≥ 2 they are uniformly distributed points on a circle with centre 0 and radius
| a|1/n . They are the corners of a regular n-gon.
Proposition
az2 + bz + c
has
Proof
Remark
Since z2 = −z1 we often abuse the notation and summarise the above proposition
as r
b b2 − 4ac
z=− ± ,
2a 4a2
p
where (b2 − 4ac)/4a2 stands for either z1 or z2 .
30
1.4 Working with modular arithmetic
Example
Solution
z2 − 2z + 2
z − 1 z3 − 3z2 + 4z − 2
z3 − z2
−2z2 + 4z − 2
−2z2 + 2z
2z − 2
2z − 2
0
shows that
z3 − 3z2 + 4z − 2 = (z − 1)(z2 − 2z + 2).
The zeros of the polynomial
z2 − 2z + 2
are found using the formula from the previous proposition. They are
r
2 4 − 4.1.2 √
± = 1 ± −1 = 1 ± i.
2 4
It follows that
z3 − 3z2 + 4z − 2 = (z − 1)(z − 1 − i)(z − 1 + i).
Definition
31
1.4 Working with modular arithmetic
(ii) Any other natural number c with c| a and c|b satisfies c|d.
Remark
The greatest common denominator of a and b is unique. Supose that d1 and d2 are
greatest common denominators of a and b. By part (ii) of the definition we have
d1 |d2 and d2 |d1 , so that d1 = d2 .
Proposition
Proof
Suppose c ∈ N satisfies c| a and c|b. It follows that c|b and c| (a − bq), so that
| {z }
=r
c|d (since d = (b,r )).
Example
Calculate (2406,654).
32
1.4 Working with modular arithmetic
Solution
Remark
This is the euclidian algorithm for computing (a,b), which can be summarised
as follows:
Proposition
Let a and b be natural numbers and d = (a,b). There are integers u, v such that
d = ua + vb.
Proof
33
1.4 Working with modular arithmetic
Let u0 = 1, u1 = 0 and v0 = 0, v1 = 1.
For i = 1, . . . , k let
ui + 1 = ui − 1 − q i ui , v i + 1 = v i − 1 − qi v i .
ri −1 = ui a + vi b, i = 2, . . . ,k,
d = uk a + vk b.
Remark
Example
It follows that
6 = 28.2406 − 103.654
Proposition
va ≡ 1 (mod n).
34
1.4 Working with modular arithmetic
Proof
un + va = 1.
It follows that
va ≡ 1 (mod n).
Lemma
Proof
va ≡ 1 (mod p).
It follows that
[v][ a] = [1]
in Z p , i.e [v] is the multiplicative inverse of [ a].
Example
35
1.4 Working with modular arithmetic
Solution
We have that
581.7039 − 626.6533 = 1,
so that
[6533]−1 = [−626] = [6413]
in Z7039 .
x ≡ a (mod n),
x ≡ b (mod m)
x = avm + bun,
un + vm = 1,
Proof
un + vm = 1.
It follows that
x = avm + bun
is a solution to the given equations because
36
1.4 Working with modular arithmetic
i.e. n|( x1 − x2 ) und m|( x1 − x2 ). Since n and m have no common divisors this
implies that nm|( x1 − x2 ) (see below).
Example
Two cog wheels with respectively 13 and 17 teeth fit together. The teeth and gaps
are numbered from respectively 1 to 13 and 1 to 17 (in the direction of rotation).
Tooth 1 sits in gap 1. After how many clicks does tooth 4 of the small wheel sit in
gap 6 of the larger wheel?
Solution
x ≡ 4 (mod 13),
x ≡ 6 (mod 17)
4.13 − 3.17 = 1,
so that
x = −4.3.17 + 6.4.13 = 108
solves the above equations.
so that the wheels are in the shown positions after 108-1, 329-1, 550-1, . . . clicks.
37
1.4 Working with modular arithmetic
Prime numbers
Proposition
Suppose that a and b are natural numbers and p is a prime number such that p| ab.
It follows that p| a or p|b.
Proof
Suppose that p6 | a, so that ( p,a) = 1. There exist Bézout coefficients u and v with
1 = pu + av,
Remark
This result does not hold if p is not a prime number. For example 4|2.6 but 46 |2
and 46 |6.
Corollary
Corollary
Proof
If ab and n have a common factor, then they have a common prime factor p. But
p| ab implies p| a or p|b.
38
1.4 Working with modular arithmetic
Definition
A composite number is a natural number which is bigger than 1 and not a prime
number.
Remark
The expression “composite number" alludes to the fact that every natural num-
ber n > 1 which is not prime can be reduced to a product of (powers of) prime
numbers. For example:
360 = 3.120
= 3.30.4
= 3.3.10.2.2.
= 2.2.2.3.3.5
| {z }
prime decomposition
of 360
Proof
Note that N is not a prime number. There therefore exist natural numbers a,
b > 1 such that
N = ab.
Since a,b < N they both have prime decompositions, so that N = ab also has
a prime decomposition. This is a contradiction.
Suppose there is a natural number n > 1 with two different prime decompo-
sitions. There exists a smallest such natural number N.
39
1.4 Working with modular arithmetic
Definition
Let n ≥ 2 and φ(n) be the number of elements a in {1, . . . ,n}, which have no
common factors with n, i.e. (a,n) = 1. We call φ : {2,3, . . . } → N Euler’s totient
function.
Example
Proof
φ( pk ) = pk − pk−1 = pk (1 − 1p ).
40
1.4 Working with modular arithmetic
Proof
Consider the set S1 = {m1 ,m2 , . . . ,mφ(n) } of numbers in {1, . . . ,n} which have no
factors in common with n. The elements of the set S2 = { am1 ,am2 , . . . ,amφ(n) } also
have no factors in common with n. Calculating modulo n, we conclude that the
sets S1 und S2 are identical. It follows that
M j m j ≡ 1 (mod n),
Proof
41
2 Vector spaces
2 Vector spaces
2.1 Introduction
x
The order in which the translations
y are performed is irrelevant, so that
y
x + y = y + x.
42
2.1 Introduction
x
For every vector x there is a vector − x
which represents the opposite transla-
tion.
−x Obviously x + (−− x) = 0
− x) + x = 0).
(and (−
3
2
x The vector αx has the same
x direction as x and α times
the length.
We find the following rules for scalar multiplication. For all vectors x, y and all
scalars α and β we have that
(α + β)x = αx + βx,
(α · β)x = α( βx),
1x = x,
α(x + y) = αx + αy.
Definition
Suppose that V is a set equipped with a binary operation + which has the follow-
ing properties.
43
2.1 Introduction
(V1) (u + v) + w = u + (v + w) ∀u, v, w ∈ V,
(V2) v+ w = w+ v ∀v, w ∈ V,
(V3) There exists an element 0 ∈ V (‘the zero vector’) such that
v + 0 = 0 + v = v ∀v ∈ V,
(V4) For every v ∈ V there exists a unique
Element − v ∈ V (‘the inverse vector to ’) such that
− v) = 0.
− v + v = v + (−
Now let (K, + ,·) be a field with identities 0 and 1. We say that V is a vector space
over K if there exists a mapping
K × V → V, (α, v) 7→ αv (⋆)
with the properties
The elements of V and K are respectively vectors and scalars, the binary opera-
tion + is vector addition and the mapping (⋆) is scalar multiplication.
Proposition
Proof
We have that
0v = (0 + 0)v
= 0v + 0v (S1)
44
2.1 Introduction
and therefore
0 = − 0v + 0v (definition of − 0v)
= − 0v + (0v + 0v)
− 0v + 0v) + 0v
= (− (V1)
= 0 + 0v (definition of − 0v)
= 0v (V3)
Moreover
v + (−1)v = 1v + (−1)v (S3)
= (1 − 1 ) v (S1)
= 0v
= 0,
so that
− v = (−1)v
(because of the uniqueness of inverse vectors).
Remark
We usually write + and − as + and − since the meaning is clear from the context.
In most examples we do not write the vectors in bold.
Examples
1. The set
R n = {x = ( x1 , . . . ,xn ) : x1 , . . . , xn ∈ R }
is a real vector space with componentwise vector addition
x + y : = ( x1 + y1 , . . . , x n + y n )
αx := (αx1 , . . . , αxn ).
−x = (− x1 , . . . , − xn ).
45
2.1 Introduction
The vector x = ( x1 ,x2 ) represents
x x2 a translation through a distance
x1 in the horizontal direction and
through x2 in the vertical direc-
tion.
x1
In linear algebra we usually write the n-tuple (k1 , . . . , kn ) in the vector space
K n as a column vector
k1
..
. ,
kn
and we now adopt this notation.
3. The set {0} (with trivial vector addition and scalar multiplication) is a vec-
tor space over every field.
4. The set R m×n of all (m × n) real matrices is a real vector space with compo-
nentwise vector addition
i.e.
a11 a12 · · · a1n b11 b12 · · · b1n
a a22 · · · a2n b b22 · · · b2n
21 21
.. .. .. + .. .. ..
. . . . . .
am1 am2 · · · amn bm1 bm2 · · · bmn
a11 + b11 a12 + b12 · · · a1n + b1n
a +b
21 a22 + b22 · · · a2n + b2n
21
:= .. .. ..
. . .
am1 + bm1 am2 + bm2 · · · amn + bmn
46
2.2 Elementary vector-space theory
i.e.
a11 a12 · · · a1n αa11 αa12 · · · αa1n
a a22 · · · a2n αa αa22 · · · αa2n
21 21
α .. .. .. : = .. .. .. .
. . . . . .
am1 am2 · · · amn αam1 αam2 · · · αamn
The zero vector is the zero matrix (0) whose coefficients are all 0 and the
inverse vector to (aij ) is the matrix (− aij ).
5. The set K m×n of all (m × n) matrices with entries from the field K is a vector
space K with componentwise vector addition and scalar multiplication.
7. The set
R [ x ] = {αn x n + αn−1 x n−1 + · · · + α1 x + α0 : n ∈ N0 , α0 , . . . ,αn ∈ R }
of all real polynomials is a real vector space with pointwise vector addition
and scalar multiplication.
Definition
47
2.2 Elementary vector-space theory
Examples
1 1 0
1. In R2the vector is a linear combination of the vectors und
1 0 1
2
in R since
1 1 0
=1 +1 .
1 0 1
Definition
Example
1. In R [ x ] we have that
h1, x, x2 i = {α0 + α1 x + α2 x2 : α0 , α1 , α2 ∈ R } = P2 (R ).
v2
v1
48
2.2 Elementary vector-space theory
Remark
Definition
Examples
3. The vector space {0} (over any field) is finite-dimensional since {0} = h∅i.
Definition
α1 v1 + · · · + α n v n = 0 ⇒ α1 = 0, . . . , αn = 0.
49
2.2 Elementary vector-space theory
Remark
If we consider
α1 v1 + · · · + α n v n = 0
as an equation for α1 , . . . , αn , we always have the solution α1 = 0, . . . , αn = 0. The
linear independence of v1 , v2 , . . . , vn is the statement that this is the only solution.
Examples
1 0 0
3
1. In R the vectors 0 , 0 , 1 are linearly independent: it follows
0 1 0
from
1 0 0 0
α 1 0 + α 2 1 + α 3 0 = 0
0 0 1 0
| {z
}
α1
= α2
α3
that α1 = 0, α2 = 0, α3 = 0.
1 1
2. In R3 the vectors 1, −1 are linearly independent: it follows from
0 0
1 1 0
α 1 1 + α 2 − 1 = 0
0 0 0
| {z }
α1 + α2
= α1 − α2
0
50
2.2 Elementary vector-space theory
1 1 0
4. In R2 the vectors , , are linearly dependent since
1 0 1
1 1 0 0
1 −1 −1 = .
1 0 1 0
1( x4 + x2 + 2) − 1( x4 + 12 x2 ) − 21 ( x2 + 1) − 32 (1) = 0.
α0 = 0
Remark
Proposition
Proof
vi = k 1 v1 + · · · + k i −1 vi −1 .
It follows that
k1 v1 + · · · + ki −1 vi −1 − vi + 0vi +1 + · · · + 0vn = 0,
51
2.2 Elementary vector-space theory
Remark
Proposition
Proof
Definition
52
2.2 Elementary vector-space theory
Examples
Remark
Lemma
53
2.2 Elementary vector-space theory
Theorem
Proof
Since E0 is finite this procedure yields a minimal spanning set for V after finitely
many steps.
Remark
Next we show that all bases for a given finite-dimensional vector space at least
have the same number of elements.
54
2.2 Elementary vector-space theory
Proof
Theorem
All bases for a given finite-dimensional vector space have the same number of
elements.
Definition
Remark
Since ∅ is a basis for the trivial vector space {0} we have dim{0} = 0.
55
2.3 Subspaces
Examples
2. dim R n [ x ] = n + 1.
Corollary
2.3 Subspaces
Definition
Remark
((i) implies 0 ∈ W and −w ∈ W for all w ∈ W, so that (V1)–(V4) follow from (i),(ii).)
56
2.3 Subspaces
Examples
λ1 (αk1 ) + · · · + λn (αkn ) = 0,
so that αk ∈ E.
λ1 (k1 + ℓ1 ) + · · · + λn (kn + ℓn ) = 0,
so that k + ℓ ∈ E.
Now let K = R.
E = {( x,y) ∈ R2 : λ1 x + λ2 y = 0}
x
is a line through the origin.
57
2.3 Subspaces
Lemma
Proof
Obviously 0 ∈ hv1 , . . . , vn i.
v = k 1 v1 + · · · + k n v n .
It follows that
αv = αk1 v1 + · · · + αkn vn ∈ hv1 , . . . , vn i.
v = k 1 v1 + · · · + k n v n , w = ℓ1 v 1 + · · · + ℓ n v n .
It follows that
v + w = ( k 1 + ℓ1 ) v 1 + · · · + ( k n + ℓ n ) v n ∈ h v 1 , . . . , v n i.
Proposition
58
2.3 Subspaces
Proof
Remark
The union of two susbspaces of a vector space is in general not a vector space.
y
E2
(1,1) The sets E1 = {( x,0) : x ∈ R } and
E2 = {(0,y) : y ∈ R } are subspaces
of R2 .
E1 x
Noe that (1,0) ∈ E1 , (0,1) ∈ E2
but (1,1) = (1,0) + (0,1) 6∈ E1 ∪ E2 .
Lemma
Proof
Let n = dim V.
59
2.3 Subspaces
Example
The subspaces of dimension 1 have the form hvi for some nontrivial vector v.
They are lines through the origin.
The subspaces of dimension 2 have the form hv1 , v2 i for two linearly inde-
pendent vectors v1 , v2 . They are planes through the origin.
v2
v1
60
2.3 Subspaces
Proof
The set Tn−m = {v1 , . . . ,vn } is linearly independent and hence a basis for V.
Example
61
2.3 Subspaces
W1
W1
W2 W2
If the line W1 lies in the plane W2 , If the line W1 intersects the plane W2
then only at the origin, then
The dimension of W1 ∩ W2 therefore depends not only upon dim W1 and dim W2
ab.
Definition
W1 + W2 := {w1 + w2 : w1 ∈ W1 , w2 ∈ W2 }
of V.
Proposition
Theorem
62
2.3 Subspaces
Proof
We have that
α1 v1 + · · · + αr vr + β 1 w1 + · · · + β s ws + γ1 z1 + · · · + γt zt = 0. (⋆)
γ1 z 1 + · · · + γ t z t = − α 1 v 1 − · · · − α r v r − β 1 w 1 − · · · − β s w s
γ1 z 1 + · · · + γ t z t = λ 1 v 1 + · · · + λ r v r
of v1 , . . . , vr . It follows from
γ1 z 1 + · · · + γ t z t − λ 1 v 1 − · · · − λ r v r = 0
Hence
dim(W1 + W2 )
=r+s+t
= dim(W1 ∩ W2 ) + (dim W1 − dim(W1 ∩ W2 )) + (dim W2 − dim(W1 ∩ W2 ))
= dim W1 + dim W2 − dim(W1 ∩ W2 ).
63
2.3 Subspaces
Definition
w = w1 + w2 , w1 ∈ W1 , w2 ∈ W2 .
Lemma
Proof
y 1 + y 2 = z1 + z2 ⇒ y −z =z −y ,
| 1 {z }1 |2 {z }2
∈ W1 ∈ W2
Remarks
64
2.3 Subspaces
Definition
Lemma
Proof
Let {v1 , . . . ,vm } a basis for W1 . Extend it to a basis {v1 , . . . ,vn } for V. Note that
W1 = hv1 , . . . ,vm i and set W2 = hvm+1 , . . . ,vn i.
Obviously
v = α1 v1 + · · · + α m v m , v = α m+1 v m+1 + · · · + α n v n .
It follows that
α1 v1 + · · · + αm vm − αm+1 vm+1 − · · · − αn vn = 0,
Remark
65
2.3 Subspaces
E2
We have that
v
R2 = E1 ⊕ E2
e1 E1
with E1 = he1 i, E2 = hvi, where v is
not parallel to e1 .
66
3 Linear transformations and matrices
Mappings between vector spaces which are compatible with the vector-space
structure are particularly important.
Definitions
T ( v1 + v2 ) = T ( v1 ) + T ( v2 ) for all v1 , v2 ∈ V
and
T (αv) = αT (v) for all v ∈ V and α ∈ K.
Examples
P1 [U]
67
3.1 Linear transformations
Im
wz
z
Rr
r
θ +φ
θ
Re
Proposition
Let V and W be vector spaces over a field K. The set of all linear transformations
T : V → W is a vector space over K with pointwise addition
(T1 + T2 )(v) := T1 (v) + T2 (v), v∈V
and scalar multiplication
(αT )(v) := αT (v), v ∈ V.
(We denote this vector space by Hom(V,W ).)
Definitions
68
3.1 Linear transformations
Im T = { T (v) : v ∈ V }
of W.
Proposition
1. ker T is a subspace of V.
2. Im T is a subspace of W.
Proof
Lemma
Proof
69
3.1 Linear transformations
Note that
T ( x ) = T (y) ⇒ T ( x ) − T (y) = 0
⇒ T ( x − y) = 0
⇒ x − y ∈ ker T,
Proposition
Proof
v = α1 e1 + · · · + α n e n ,
so that
On the other hand, it follows from w ∈ hT (e1 ), . . . ,T (en )i that w can be written
as a linear combination
w = β 1 T (e1 ) + · · · + β n T (en ) = T ( β 1 e1 + · · · + β n en ) ∈ Im T.
Definitions
70
3.1 Linear transformations
Proof
Let {v1 , . . . ,vm } be a basis for ker T and extend it to a basis {v1 , . . . ,vn } for V. We
show that { T (vm+1 ), . . . ,T (vn )} is a basis for Im T. It follows that Im T is finite-
dimensional with dim Im T = n − m = dim V − dim ker T.
Note that
so that
T (αm+1 vm+1 + · · · + αn vn ) = 0.
Hence αm+1 vm+1 + · · · + αn vn ∈ ker T and can therefore be written as a linear
combination of {v1 , . . . ,vm }. There therefore exist scalars α1 , . . . , αm such that
α m +1 v m +1 + · · · + α n v n = α1 v1 + · · · + α m v m ,
α1 v1 + · · · + α m v m − α m +1 v m +1 − · · · − α n v n = 0
71
3.2 Matrix representations of linear transformations
Corollary
Proof
T ist surjective ⇔ Im T = W
⇔ rank T = dim W (Im T is a subspace of W)
⇔ null T = 0
⇔ ker T = {0}
⇔ T is injective.
In this section V und W are finite-dimensional vector spaces over a field K and T :
V → W is a linear transformation. Let {e1 , . . . , en } be a basis for V and { f 1 , . . . , f m }
be a basis for W (so that n = dim V and m = dimW).
v = α1 e1 + α2 e2 + · · · + α n e n . (1 )
It follows that
T ( v ) = α 1 T ( e 1 ) + α 2 T ( e 2 ) + · · · + α n T ( e n ). (2 )
Remark
72
3.2 Matrix representations of linear transformations
On the other hand T (e1 ), . . . , T (en ) are elements of W and can therefore be written
as unique linear combinations of the basis vectors f 1 , . . . f m : there are unique
scalars αij , i = 1, . . . , m, j = 1, . . . ,n such that
T (e1 ) = α11 f 1 + α21 f 2 + · · · + αm1 f m ,
T (e2 ) = α12 f 1 + α22 f 2 + · · · + αm2 f m ,
.. .. ..
. . .
T (en ) = α1n f 1 + α2n f 2 + · · · + αmn f m .
Remark
Definition
The m × n matrix
α11 α12 · · · α1n
α α22 · · · α2n
21
A = (αij ) = .. .. ..
. . .
αm1 αm2 · · · αmn
is the matrix of T : V → W with respect to the basis {e1 , . . . ,en } for V and { f 1 , . . . , f m }
for W.
Remark
Example
73
3.2 Matrix representations of linear transformations
with respect to
Solution
74
3.3 Matrix algebra
Example
Solution
Note that
In this section X, Y and Z are finite-dimensional vector spaces over a field K and
S : X → Y, T : Y → Z are linear transformations. Furthermore X = { x1 , . . . , xn },
Y = {y1 , . . . ,ym } and Z = {z1 , . . . , zℓ } are bases for X, Y and Z.
Denote the matrices of S and T with respect to the given bases for X, Y and Z by
respectively
75
3.3 Matrix algebra
s11 s12 · · · s1n
s21 s22 · · · s2n
X
MY (S) = (sij ) = .. .. ..
. . .
sm1 sm2 · · · smn
and
t11 t12 · · · t1m
Y
t21
t22 · · · t2m
MZ (T ) = (tij ) = .. .. .. ,
. . .
t ℓ1 t ℓ2 · · · t ℓ m
so that
i.e.
m
S( x j ) = ∑ skj yk , j = 1, . . . ,n,
k=1
and
i.e.
ℓ
T (y j ) = ∑ tkj zk , j = 1, . . . ,m.
k=1
Now we ask the following question: can we define the product of two matrices
in such a way that the formula
X Y
MZ ( T ◦ S ) = MZ (T ).MYX (S) (⋆)
holds?
76
3.3 Matrix algebra
Lemma
Proof
Let U = T ◦ S and
u11 u12 · · · u1n
u u22 · · · u2n
X 21
MZ (U ) = (uij ) = .. .. ..
. . .
u ℓ1 u ℓ2 · · · u ℓ n
be the matrix of U with respect to the given bases for X and Z, so that
ℓ
U (x j ) = ∑ uij zi , j = 1, . . . ,n. (1 )
i =1
Note that
U ( x j ) = T (S( x j ))
!
m
=T ∑ skj yk
k=1
m
= ∑ skj T (yk )
k=1
m ℓ
= ∑ skj ∑ tik zi
k=1 i =1
ℓ m
= ∑ ∑ tik skj zi . (2 )
i =1 k=1
77
3.3 Matrix algebra
Definition
Let K be a field and A ∈ K ℓ×m , B ∈ K m×n . The product AB is the matrix C = (cij )
in K ℓ×n with
m
cij = ∑ aik bkj .
k=1
Remark
↓
↓
→ → · · · →
i-th row
..
.
↓
j-th column
Example
1 2 6 5 4 3 6+4 5+2 4+0 3−2 10 7 4 1
3 4 2 1 0 −1 = 18 + 8 25 + 4 12 + 0 9 − 4 = 26 19 12 5 .
5 6 30 + 12 15 + 6 20 + 0 15 − 6 42 31 20 9
Remarks
but
5 6 1 2 23 34
= .
7 8 3 4 31 46
78
3.3 Matrix algebra
Lemma
v = a1 x 1 + · · · + a n x n ,
S(v) = b1 y1 + · · · + bm ym .
It follows that
X
b = MY (S)a,
where
a1 b1
..
b = ... .
a = . ,
an bm
Proof
Note that
!
n
Sv = S ∑ ak xk
k=1
n
= ∑ ak S( x k )
k=1
n m
= ∑ ∑ak sik yi
k=1 i =1
m n
= ∑ ∑ sik ak yi ,
i =1 k=1
so that
n
bi = ∑ sik ak , i = 1, . . . ,m.
k=1
79
3.3 Matrix algebra
Example
Using this matrix we can compute p′ for each p ∈ R n [ x ]. We represent the poly-
nomial
p = α0 + α1 x + α2 x 2 + · · · + α n −1 x n −1 + α n x n
as the column vector
α0
α1
α2
.. ,
.
α n −1
αn
and the polynomial p′ is represented by
0 1 0 0 ··· 0 α0 α1
0 0 2 0 · · · 0 α1 2α2
0 0 0 3 · · · 0 α2 3α3
.. .. .. .. .. .. = .. .
. . . . . . .
0 0 0 0 · · · nαn−1 nαn
0 0 0 0 ··· 0 αn 0
Hence
p′ = α1 + 2α2 x + 3α3 x2 + · · · + nαn x n−1.
80
3.3 Matrix algebra
Notation
Now let X ′ = { x1′ , . . . , xn′ }, Y ′ = {y1′ , . . . ,y′m } be further bases for respectively X
and Y.
Definition
The change of basis matrix BXX from X to X ′ is the matrix of the identity trans-
′
formation I : X → X with respect to X and X ′ , i.e.
X X
BX ′ = M X ′ ( I ).
Example
Compute the change of basis matrix from the basis B1 = {1,x,x2 , . . . , x n } to the
basis B2 = {1,1 + x,1 + x + x2 , . . . , 1 + x + x2 + · · · + x n } for R n [ x ].
Remark
We write the old basis vectors als linear combinations of the new basis vectors:
1 = 1,
x = − 1 + (1 + x ),
x2 = −(1 + x ) + (1 + x + x2 ),
.. ..
. .
x n = −(1 + x + · · · + x n−1 ) + (1 + x + · · · + x n ).
Hence
1 −1 0 0 ···
0
0
1 −1 0 ···
0
0 0 1 −1 ···
0
BBB21
= 0 0 0 1 ···
0
.. .. .. .. ..
. . . . .
0 0 0 0 ··· 1
81
3.3 Matrix algebra
Lemma
X ′ Y X X ′
MY ′ ( S ) = BY ′ M Y ( S ) BX .
Proof
′
X (S) and M X (S) by respectively (s ) and (s′ ), so that
Denote MY Y′ ij ij
m
S( x j ) = ∑ skj yk , j = 1, . . . ,n
k=1
and
m
S( x ′j ) = ∑ s′kj y′k , j = 1, . . . ,n. (⋆)
k=1
′
X and BY by respectively (b ) and c , so that
Denote BX Y′ ij ij
n
x ′j = ∑ bkj xk , j = 1, . . . ,n
k=1
and
m
yj = ∑ ckj y′k , j = 1, . . . ,m
k=1
Note that
!
n
S( x ′j ) = S ∑ bij xi
i =1
n
= ∑ bij S( xi )
i =1
n m
= ∑ bij ∑ ski yk
i =1 k=1
n m m
= ∑ bij ∑ ski ∑ cℓk y′ℓ
i =1 k=1 ℓ=1
m m n
= ∑ ∑ cℓk ∑ ski bij y′ℓ
ℓ=1 k=1 i =1
m m
′
= ∑ ∑ cℓk ( MYX (S)BXX )kj y′ℓ
ℓ=1 k=1
m
X′
= ∑ ( BYY ′ MYX (S) BX )ℓ j y′ℓ
ℓ=1
82
3.3 Matrix algebra
It follows that
Y ′
s′ℓ j = ( BY X X
′ M Y ( S ) BX ) ℓ j
Example
Solution
Note that
1 1 0 0
0 = 1 0 + 0 1 − 1 0 ,
−1 0 0 1
1 1 0 0
1 = 1 0 + 1 1 + 1 0 ,
1 0 0 1
0 1 0 0
0 = 0 0 + 0 1 + 1 0 ,
1 0 0 1
83
3.3 Matrix algebra
so that
1 1 0
B3′
B B3 = 0 1 0 .
−1 1 1
Note that
1 1 1
=0 +1 ,
0 1 0
0 1 1
=1 −1 ,
1 1 0
so that
0 1
BBB2′ = .
2 1 −1
Hence
0 1 1 1 −1 1 1 0
B B3′
BBB2′ MB23 (T ) B B3 = 1 −1 2 0 1 0 1 0
2
−1 1 1
2 0 1 1 1 0
= −1 1 −2 0 1 0
−1 1 1
1 3 1
= .
1 −2 −2
84
4 Matrix algebra and systems of linear equations
Definition
Let K be a field and A ∈ K m×n . Write A = (aij ) i=1,...,m, . The matrix AT ∈ K n×m
j =1,...,n
defined by the formula
aTij := a ji
is the transpose of A.
Remarks
1. We obtain AT from A by exchaning the rows and columns, so that the trans-
pose of the matrix
a11 a12 · · · a1n
a a22 · · · a2n
21
A = .. .. .. m rows,
. . .
am1 am2 · · · amn
| {z }
n columns
is the matrix
a11 a21 · · · am1
a a22 · · · am2
T 12
A = .. .. .. n rows.
. . .
a1n a2n · · · amn
| {z }
m columns
85
4.1 On the columns and rows of a matrix
Notation
Let K be a field and A ∈ K m×n . Denote the columns and rows of A by c1A , . . . , cnA
bzw. r1A , . . . , rm
A , so dass
r1A
A A A
r2A
A = c1 c2 · · · cn = .
..
.
A
rm
In this notation
a1j
a2j
c jA = . ,
j = 1, . . . ,n
..
amj
and
riA = (ai1 ,ai2 , . . . ,ain ), i = 1, . . . ,m.
Remark
and
m
riC = ∑ aik rkB , i = 1, . . . ,ℓ.
k=1
In particular one finds that
86
4.1 On the columns and rows of a matrix
Definitions
Remark
The column rank (row rank) of A is the maximum number of linearly indepen-
dent columns (rows) of A.
Example
Theorem
Let K be a field. The formula column rank A = row rank A holds for each A ∈
K m× n .
Proof
Let c = column rank A and {e1 , . . . , ec } be a basis for the column space of A. Every
column c jA of A can be written as a linear combination of e1 , . . . , ec , i.e. there are
scalars rkj such that
c
c jA = ∑ rkj ek , j = 1, . . . ,n
k=1
87
4.1 On the columns and rows of a matrix
E = e1 e2 · · · ec .
In view of the previous lemma we speak only of ‘the rank’ of a matrix. Now we
turn to the question of computing it.
Proposition
Let A be a matrix with m rows whose first r elements on the diagonal are non-
zero, but whose sub-diagonal elements and last m − r rows vanish, i.e.
a11
0 a22
.
. .. ⋆
. .
0 · · · 0 arr
A= (⋆)
0 0
88
4.1 On the columns and rows of a matrix
Definition
Proposition
Elementary row (columnn) operations do not change the row (column) space of
a matrix and hence do not change its rank.
We can therefore determine the rank of a matrix by converting it into the form (⋆)
by means of elementary row and column operations.
Example
Solution
89
4.1 On the columns and rows of a matrix
2 −1 3 1 0 1
1 0 1
2 −1 3
−→ (r 1 ↔ r 2 )
0 2 −1 0 2 −1
1 1 4 1 1 4
1 0 1
0 −1 1
−→ (r2 → r2 − 2r1 , r4 → r4 − r1 )
0 2 −1
0 1 3
1 0 1
0 −1 1
−→
0
(r3 → r3 + 2r2 , r4 → r4 + r2 )
0 1
0 0 4
1 0 1
0 −1 1
−→
0
(r4 → r4 − 4r3 )
0 1
0 0 0
The rank of the last – and hence also the first – matrix is therefore 3.
Definition
(i) The last m − r rows are identically zero, but the first r rows are not identically
zero. (Here 0 ≤ r ≤ m.)
(ii) Supposing that the first non-vanishing element in row i lies in columnn ji , we
have that
1 ≤ j1 < j2 < . . . < jr ≤ n.
0 · · · 0 a1j1 ⋆ ⋆ ⋆ ⋆ ⋆ ⋆ ⋆ ⋆
0
··· 0 0 · · · 0 a2j2 ⋆ ⋆ ⋆ ⋆ ⋆
. .. .. .. .. .. .. .. ..
.. . . . . . . . .
A=
0 ··· 0 0 ··· 0 0 · · · 0 arjr ⋆ ⋆
0
··· 0 0 ··· 0 0 ··· 0 0 ··· 0
.. .. .. .. .. .. .. ..
. . . . . . . .
0 ··· 0 0 ··· 0 0 ··· 0 0 ··· 0
90
4.1 On the columns and rows of a matrix
Proposition
The rank of a matrix in echelon form is the number of rows which are not identi-
cally zero.
Any matrix can be converted into echelon form using elementary row operations.
Proof
We begin with an m × n matrix A. The proof is complete if the columns are all
identically zero.
Otherwise let j1 be the smallest index for which the column c jA1 does not vanish.
By swapping rows we can arrange that the first entry of c jA1 is not zero. With
a misuse of notation we continue to denote the entries of c jA1 by a1j1 , . . . , amj1 .
We can convert the remaining entries of this column to zero using the elemen-
tary column operations
aij1
ci → ci − c1 , i = 2, . . . ,m.
a1j1
where A′ is an (m − 1) × (n − j1 ) matrix.
91
4.1 On the columns and rows of a matrix
Remark
We can convert a matrix in echelon form into the form (⋆) by swapping columns.
Remark
Let K be a field.
Im A = hAe1 , . . . , Aeni
| {z }
the columns of A
that dim Im A is the dimension of the column space of A. This definition
coincides with our previous definition of rank A.
92
4.2 The inverse of a matrix
Definitions
Let K be a field.
1. The n × n matrix
1
In := ..
.
1
is the identity matrix. It has the property that
MIn = In M = M
Proposition
Let K be a field.
(ii) The product of two invertible matrices A, B in K n×n is invertible with ( AB)−1 =
B −1 A −1 .
Definition
Let K be a field. The set of all invertible matrices in K n×n forms a group with
respect to matrix multiplication called the general linear group GL(n,K ).
93
4.2 The inverse of a matrix
Example
a b
GL(2,R ) = : a,b,c,d ∈ R, ad 6= bc
c d
and −1
a b 1 d −b
= , ad 6= bc.
c d ad − bc −c a
Remark
Let n ∈ N and V be an n-dimensional vector space over a field K.. The identity
matrix In ∈ K n×n is the matrix of the identity transformation V → V, x 7→ x (with
respect to an arbitrary basis for V).
Proposition
Proof
B B
Suppose that MB21 (T ) is invertible. Its inverse MB21 (T )−1 defines a linear
B
transformation S : V → V witht MBB12 (S) = MB21 (T )−1 . The matrices of the
linear transformations ST : V → V (with respect to B1 ) and TS : V → V (with
respect to B2 ) are respectively
B B B B
MB11 (ST ) = MBB12 (S) MB21 (T ) = MB21 (T )−1 MB21 (T ) = In ,
94
4.2 The inverse of a matrix
and
MBB22 (TS) = MBB21 (T ) MBB2 (S) = MBB21 (T ) MBB21 (T )−1 = In ,
1
Corollary
Let B1 and B2 be two bases for a finite-dimensional vector space V over a field
K.. It follows that BBB21 is invertible with ( BBB21 )−1 = BBB2 .
1
Proposition
1. A ist invertible.
2. A ist injective.
3. A ist surjective.
4. rank A = n.
5. AT is invertible.
Lemma
Proof
A is invertible if and only if rank A = n, i.e. if it can be converted into the echelon
form
a11
0 a22
.. ..
. .
0 · · · 0 ann
using elementary row operations.
95
4.2 The inverse of a matrix
Furthermore every matrix A can be converted into the above form using elemen-
tary row operations if and only if it can be converted into the identity matrix
using elementary row operations.
Now we present a method to compute the inverse of a matrix with maximal rank.
Definition
.. ..
1 0 . .
.. .. ..
0 . . .
.
1 .. 0
· · · · · · · · · 0 · · · · · · 0 1 0 · · · · · · ← i-th row
..
. 1 0
Vij = .
. . . .
.
. . .
..
0 1 .
· · · · · · 0 1 0 · · · · · · 0 · · · · · · · · · ← j-th row
..
0 . 1
.
.. .
.. . ..
.. ..
. . 1
↑ ↑
i-th column j-th column
The elementary operation ri → λri oder ci → λci with λ 6= 0 yields the matrix
1
..
.
1
Miλ =
λ ← i-th row
1
..
.
1
↑
i-th column
96
4.2 The inverse of a matrix
Proposition
Corollary
The elementary matrices are invertible with Vij−1 = Vij , ( Miλ )−1 = Mi1/λ and
( Aijλ )−1 = Aij−λ .
Corollary
97
4.2 The inverse of a matrix
Proof
E p E p−1 · · · E1 A = In .
Remarks
A = E1−1 · · · E−
p
1
of elementary matrices.
2. The equation
A−1 = E p E p−1 · · · E1 In ,
shows that the elementary row operations which convert A into In also con-
vert In into A−1 .
Example
98
4.3 Systems of linear equations
Solution
The calculation
2 −1 3 1 0 0 1 0 1 010
1 0 1 0 1 0 −→ 2 −1 3 1 0 0 (r 1 ↔ r 2 )
0 2 −1 0 0 1 0 2 −1 0 0 1
1 0 1 0 1 0
−→ 0 −1 1 1 −2 0 (r2 → r2 − 2r1 )
0 2 −1 0 0 1
1 0 10 1 0
−→ 0 −1 1 1 −2 0 (r3 → r3 + 2r2 )
0 0 1 2 −4 1
1 0 0 −2 5 −1
−→ 0 −1 0 −1 2 −1 (r 1 → r 1 − r 3 , r 2 → r 2 − r 3 )
0 0 1 2 −4 1
1 0 0 −2 5 −1
−→ 0 1 0 1 −2 1 (r 2 → − r 2 )
0 0 1 2 −4 1
a11 x1 + a12 x2 + · · · + a1n xn = b1 ,
a21 x1 + a22 x2 + · · · + a2n xn = b2 ,
.. .. .. (⋆)
. . .
am1 x1 + am2 x2 + · · · + amn xn = bm ,
where the coefficients aij and the right-hand sides bi are given. Do there exist
value of the unknowns x1 , . . . , xn which solve these equations? (Here we work
in a field K..) And how can we describe the set of solutions?
99
4.3 Systems of linear equations
Example
describe three planes, so that the solution set M of this system of equations con-
sists of the points of intersection of the planes. There are several possible geomet-
ric configurations.
a M = {a + λv : λ ∈ R }
100
4.3 Systems of linear equations
a M = {a + λv : λ ∈ R },
Now we consider the system (⋆) more systematically. We can write it in the form
Ax = b
101
4.3 Systems of linear equations
Definitions
Let
Ax = b
be a system of linear equations for the unknown x ∈ K n .
The matrix A ∈ K m×n is its matrix of coefficients and the matrix ( A|b) ∈ K m×(n+1)
is its extended matrix of coefficients.
Proposition
Proof
First notice that rank ( A|b) ≥ rank A with equality if and only if b lies in the
column space of A.
Definition
102
4.3 Systems of linear equations
Proposition
Ax = 0
x⋆ + ker A.
Proposition
Suppose that ( A|b) can be converted into ( A′ |b′ ) by an elementary row opera-
tion. The solution sets of the systems
Ax = b
and
A′ x = b′
are identical.
Remark
The elementary column swap ci ↔ c j in the matrix A converts the system of linear
equations
x1 c1A + · · · + xi ciA + · · · + x j c jA + · · · + xn cnA = b
into
x1 c1A + · · · + xi c jA + · · · + x j ciA + · · · + xn cnA = b
and thus corresponds to a relabelling of the variables xi and x j .
103
4.3 Systems of linear equations
Example
Suppose that ( A|b) ∈ K m×(n+1) is in echelon form with rank ( A|b) = rank A. By
relabelling the variables we may assume that ( A|b) has the form
a11 b1
0 a22 b2
.
. .. A2 ..
. . .
0 · · · 0 arr br
( A|b) =
0 0 0
Example
x + 3y + 2z = k,
2x + ky + 4z = 12,
kx + 2y − 4z = 20.
104
4.3 Systems of linear equations
Solution
We have that
13 2 k 1 3 2 k
2 k 4 12 −→ 0 k − 6 0 12 − 2k (r2 → r2 − 2r1 , r3 → r3 − kr1 )
k 2 −4 20 0 2 − 3k −4 − 2k 20 − k2
x + 3y + 2z = 6,
y+z=1
is
x 3+λ
y = 1 − λ : λ ∈ R .
z λ
105
4.4 Determinants
x + 3y + 2z = k,
y = −2,
−2(2 + k)z = 24 − 6k − k2
2 (2 + k )
4.4 Determinants
Definition
106
4.4 Determinants
Remarks
Lemma
1. Converting the matrix A into the matrix A′ by swapping two rows yields
det A′ = − det A.
Proof
107
4.4 Determinants
r1A
..
.
riA + r jA
← i-th row
0 = det ..
.
riA + r jA
← j-th row
..
.
rnA
.. ..
. .
riA r jA
= det .. ..
. + det .
riA + r jA riA + r jA
.. ..
. .
.. .. .. ..
. . . .
A
riA
r jA
r jA
ri
= det ... + det .. .. ..
. + det . + det . ,
r A
r jA
riA
r jA
i
.. .. .. ..
. . . .
| {z } | {z }
=0 =0
so that
0 = det A + det A′ .
2. This follows directly from the linearity of the determinant function in each
row.
108
4.4 Determinants
r1A
..
.
riA + λr jA
← i-th row
det ..
.
r jA
← j-th row
..
.
rnA
.. ..
. .
riA r jA
= det .. ..
. + λ det . ,
r jA r jA
.. ..
. .
| {z }
=0
so that
det A′ = det A.
Remark
Lemma
Proof
109
4.4 Determinants
Proposition
Proof
Theorem
and
a11 , n = 1,
(Expansion
cdet j A = n j = 1, . . . ,n, along the j-th
i+ j ′
∑ (−1) aij det Aij , n > 1,
column)
i =1
where
a11 · · · a1j · · · a1n
.. ..
. .
′
(n −1)×(n −1)
ai1 · · · aij · · · ain ∈ K
Aij =
.. ..
. .
an1 · · · anj · · · ann
110
4.4 Determinants
is the matrix obtained from A by deleting its i-th row and j-th column, are nor-
malised determinant functions K n×n → K (and are therefore all identical).
Remark
+ − + − + ...
− + − + − . . .
+ − + − + . . .
− + − + − . . .
.. .. .. .. ..
. . . . .
Examples
1. We have that
a b a b a b
det = a det − b det = ad − bc,
c d c d c d
so that the matrix
a b
c d
is invertible if and only if ad 6= bc.
2. We have that
a11 a12 a13
det a21 a22 a23
a31 a32 a33
a11 a12 a13 a11 a12 a13 a11 a12 a13
= a11 det a21 a22 a23 − a12 det a21 a22 a23 + a13 det a21 a22 a23
a31 a32 a33 a31 a32 a33 a31 a32 a33
a a a a a a
= a11 det 22 23 − a12 det 21 23 + a13 det 21 22
a32 a33 a31 a33 a31 a32
= a11 (a22 a33 − a32 a23 ) − a12 (a21 a33 − a31 a23 ) + a13 (a21 a32 − a31 a22 )
= a11 a22 a33 + a12 a23 a31 + a13 a21 a32 − a31 a22 a13 − a32 a23 a11 − a33 a21 a12 .
111
4.4 Determinants
4. Note that
1 a2 + bc a3
det 1 b2 + ac b3 = (a − b)(a − c)(b − c)(a2 + b2 + c2 ),
1 c2 + ab c3
because
1 a2 + bc a3
det 1 b2 + ac b3
1 c2 + ab c3
1 a2 + bc a3
= det 0 (b − a)(a + b − c) (b − a)(b2 + ba + a2 )
0 (c − a)(a + c − b) (c − a)(c2 + ca + a2 )
(r 2 → r 2 − r 1 , r 3 → r 3 − r 1 )
1 a2 + bc a3
= (b − a)(c − a) det 0 a + b − c b2 + ba + a2
0 a + c − b c2 + ca + a2
112
4.4 Determinants
a + b − c b2 + ba + a2
= (b − a)(c − a) det
a + c − b c2 + ca + a2
(Expansion along the first column)
a+b−c b2 + ba + a2
= (b − a)(c − a) det
2(c − b) (c − b)(a + b + c)
(r 2 → r 2 − r 1 )
a + b − c b2 + ba + a2
= (b − a)(c − a)(c − b) det .
2 a+b+c
| {z }
2 2 2
= −a − b − c
Lemma
Let K be a field. The formula det AB = det A det B holds for all A,B ∈ K n×n .
Proof
It follows from
r1 r1 r1 B
.. .. ..
Det . = det . B = det .
rn rn rn B
that
113
4.4 Determinants
r1 r1 B
.. ..
.
.
1 2 1 2
Det α1 r j + α2 r j = det α1 r j B + α2 r j B
.. ..
.
.
rn rn B
r1 B r1 B
.. ..
.
.
1 2
= α1 det r j B + α2 det r j B
.. ..
.
.
rn B rn B
r1 r1
.. ..
. .
1
= α1 Det r j + α2 Det r j ,2
.. ..
. .
rn rn
Hence
| {z A} = µ A Det
Det I ,
| {z n}
det A = µ A det In ,
| {z }
= det AB = det B =1
so that
det AB = det A det B.
Corollary
114
4.4 Determinants
Proof
Lemma
Proof
A = E1 · · · E p
Note that VijT = Vij and ( Miλ )T = Miλ , so that det VijT = det Vij and det( Miλ )T =
det Miλ .
Furthermore ( Aijλ )T = Aλji and det Aijλ = det In = 1, det Aλji = det In = 1, so that
det( Aijλ )T = det Aλji .
115
4.4 Determinants
Corollary
Let K be a field. The function det : K n×n → K is linear in each column of its
argument, i.e.
det c1 · · · α1 c1j + α2 c2j · · · cn = α1 det c1 · · · c1j · · · cn + α2 det c1 · · · c2j · · · cn
Corollary
1. Converting the matrix A into the matrix A′ by swapping two columns yields
det A′ = − det A.
Now we present methods for inverting a regular matrix and inverting a system
of linear equations which are of academic rather than pratical interest.
Definition
♯
Let K be a field and A ∈ K n×n . Die Matrix A♯ = (aij ) ∈ K n×n with
116
4.4 Determinants
Remark
Note that
Proposition
A♯ A = AA♯ = (det A) In .
117
4.4 Determinants
Proof
Note that
n
♯ ♯
( A A)ij = ∑ aik akj
k=1
n
= ∑ akj det c1A · · · ciA−1 ek ciA+1 · · · cnA
k=1
!
n
= det c1A ··· ciA−1 ∑ akj ek ciA+1 ··· cnA
k=1
= det ··· c1A ciA−1 c jA ciA+1 ··· cnA
det A, i = j,
=
0, i 6= j
= (det A)δij .
Corollary
1
A −1 = A♯ .
det A
118
4.4 Determinants
Proof
Note that
1
x = A −1 b = A♯ b,
det A
so that
1
xi = ( A ♯ b )i
det A
n
1
aij♯ b j
det A j∑
=
=1
n
1
= ∑ b j det c1A · · · ciA−1 e j ciA+1 · · · cnA
det A j=1
n
!
1 A A A A
= det c1 · · · ci −1 ∑ b j e j ci +1 · · · cn
det A j =1
1
= det c1A · · · ciA−1 b ciA+1 · · · cnA .
det A
119
5 Eigenvalues and eigenvectors
Definition
Remark
Examples
Definitions
Tv = λv.
Eλ = {v : Tv = λv} = ker(T − λI )
120
5 Eigenvalues and eigenvectors
Remark
Examples
v Tv
v-1 v1 θ v
Rv1 Sv v1
Tv1
Rv-1
121
5 Eigenvalues and eigenvectors
Proposition
Proof
Note that
Tv − λv = 0 for some v 6= 0
⇔(T − λI )v = 0
⇔ ker(T − λI ) 6= {0}
⇔ dim ker(T − λI ) > 0
⇔ dim Im (T − λI ) < dim V
| {z }
= rank (T − λI )
⇔ det(T − λI ) = 0.
Definition
where A = (aij ) is the matrix of T with respect to a basis for V, is called the
characteristic function of T.
Proposition
122
5 Eigenvalues and eigenvectors
Proof
Note that
A − λIn = (c1A − λe1 | · · · |cnA − λen ).
Since the determinant is linear in each colum of its argument, we find that
so that c(λ) = det( A − λIn ) is a polynomial with deg c ≤ n. Moreover, the coeffi-
cient of λn is (−1)n det(e1 | · · · |en ) = (−1)n det In = (−1)n 6= 0, so that deg c = n.
Remarks
Example
123
5 Eigenvalues and eigenvectors
Solution
= −(1 + λ)(λ2 − 7λ − 8)
i.e. λ = −1 and λ = 8.
We have that
E−1 = {x ∈ R3 : ( A + I3 )x = 0}
x 4 2 4 x 0
= y : 2 1 2 y = 0 .
z 4 2 4 z 0
It follows that
124
5 Eigenvalues and eigenvectors
x
E−1 = y : 4x + 2y + 4z = 0
z
1
−2µ − λ
= µ : λ, µ ∈ R
λ
* 1 −1 +
−2
= 1 , 0 ,
0 1
and the geometric multiplicity of this eigenvalue is
dim E−1 = dim ker( A + I3 ) = 2.
We have that
E8 = {x ∈ R3 : ( A − 8I3 )x = 0}
x −5 2 4 x 0
= y : 2 −8 2 y = 0 .
z 4 2 −5 z 0
We therefore study the extended matrix of coefficients
−5 2 4 0
2 −8 2 0
4 2 −5 0
−1 4 −1 0
→ 2 −8 2 0 (r 1 → r 1 + r 3 )
4 2 −5 0
−1 4 −1 0
→ 0 0 0 0 (r2 → r2 + 2r1 , r3 → r3 + 4r1 )
0 18 −9 0
−1 4 −1 0
→ 0 2 −1 0 (r3 → 91 r3 , r3 ↔ r2 )
0 0 0 0
It follows that
x
− x + 4y − z = 0,
E8 = y :
2y − z = 0
z
λ
= 1
λ :λ∈R
2
λ
* 1 +
= 1 ,
2
1
125
5 Eigenvalues and eigenvectors
Remark
The matrix
0 2
∈ R 2× 2
1 0
√
has eigenvalues ± 2, but
0 2
∈ Q 2× 2
1 0
has no eigenvalues
√ since the characteristic polynomial c(λ) = λ2 − 2 has the
real roots ± 2 but no rational roots.
The matrix
0 4
∈ C 2× 2
−1 0
has eigenvalues ±2i, but
0 4
∈ R 2× 2
−1 0
has no eigenvalues since the characteristic polynomial c(λ) = λ2 + 4 has the
complex roots ±2i but no real roots.
Lemma
Proof
126
5 Eigenvalues and eigenvectors
Definitions
Let K be a field.
Proposition
Proof
T is diagonalisable
There is a basis {v1 , . . . ,vn } for V with respect to which the matrix
⇔
of T has the form diag(λ1 , . . . ,λn )
There is a basis {v1 , . . . ,vn } for V such that Tvi = λi vi for i =
⇔
1, . . . , n
⇔ There is a basis for V consisting of eigenvectors of T
127
5 Eigenvalues and eigenvectors
Definition
Remark
Example
E8 = 1 .
2
1
1
−2 −1 1
The eigenvectors 1 , 0 , 21 are linearly independent and therefore
0 1 1
3
form a basis B for R .
128
5 Eigenvalues and eigenvectors
we have that
P−1 AP = diag(−1, − 1,8).
Lemma
Proof
Now suppose that the result holds for m = k. Let α1 , . . . , αk+1 be scalars such that
α1 v1 + · · · + αk+1 vk+1 = 0. (⋆)
Mutiplying (⋆) by λ1 yields
λ1 α1 v1 + λ1 α2 v2 + · · · + λ1 αk+1 vk+1 = 0.
Applying T to (⋆) yields
λ1 α1 v1 + λ2 α2 v2 + · · · + λk+1 αk+1 vk+1 = 0.
It follows that
(λ2 − λ1 )α2 v2 + · · · + (λk+1 − λ1 )αk+1 vk+1 = 0.
However by the inductive hypothesis v2 , . . . , vk+1 are linearly independent, so
that
(λ − λ ) α = 0, . . . , (λk+1 − λ1 ) αk+1 = 0,
| 2 {z 1} 2 | {z }
6= 0 6= 0
⇒ α2 = 0, . . . , αk+1 = 0.
It then follows from (⋆) that
α1 v1 = 0,
so that α1 also vanishes. Hence v1 , . . . , vk+1 are linearly independent.
Corollary
129
5 Eigenvalues and eigenvectors
Corollary
Proof
m
L
Eλi is a subspace of V, and the union of bases for Eλ1 , . . . , Eλm is a basis for
i =1
Lm
Eλi . Hence
i =1 !
m
M m
dim V ≥ dim Eλ i = ∑ dim Eλi
i =1 i =1
m
L
with equality if and only if Eλi = V.
i =1
m
L
If Eλi = V, then the union of bases for Eλ1 , . . . , Eλm is a basis for V consisting
i =1
of eigenvectors of T.
m
L
If V has a basis B consisting of eigenvectors of T, then B ⊆ Eλi implies that
i =1
m
L
Eλi = V.
i =1
Examples
130
5 Eigenvalues and eigenvectors
E8 = 12 .
1
Remark
Let V be an n-dimensional vector space over a field K. Suppose that the char-
acteristic polynomial c of a linear transformation T : V → V can be decomposed
into distinct linear factors, i.e.
c(λ) = (−1)n (λ − λ1 ) . . . (λ − λn ),
131