Geometric:Clifford Algebra For Physics
Geometric:Clifford Algebra For Physics
FOR PHYSICISTS
CHRIS DORAN
and
ANTHONY LASENBY
University of Cambridge
published by the press syndicate of the university of cambridge
The Pitt Building, Trumpington Street, Cambridge, United Kingdom
cambridge university press
The Edinburgh Building, Cambridge CB2 2RU, UK
40 West 20th Street, New York, NY 10011-4211, USA
477 Williamstown Road, Port Melbourne, VIC 3207, Australia
Ruiz de Alarcón 13, 28014 Madrid, Spain
Dock House, The Waterfront, Cape Town 8001, South Africa
http://www.cambridge.org
C Cambridge University Press, 2003
A catalogue record for this book is available from the British Library
Preface ix
Notation xiii
1 Introduction 1
1.1 Vector (linear) spaces 2
1.2 The scalar product 4
1.3 Complex numbers 6
1.4 Quaternions 7
1.5 The cross product 10
1.6 The outer product 11
1.7 Notes 17
1.8 Exercises 18
3 Classical mechanics 54
3.1 Elementary principles 55
3.2 Two-body central force interactions 59
3.3 Celestial mechanics and perturbations 64
v
CONTENTS
vi
CONTENTS
10 Geometry 340
10.1 Projective geometry 341
10.2 Conformal geometry 351
10.3 Conformal transformations 355
10.4 Geometric primitives in conformal space 360
10.5 Intersection and reflection in conformal space 365
10.6 Non-Euclidean geometry 370
10.7 Spacetime conformal geometry 383
10.8 Notes 390
10.9 Exercises 391
vii
CONTENTS
14 Gravitation 497
14.1 Solving the field equations 498
14.2 Spherically-symmetric systems 500
14.3 Schwarzschild black holes 510
14.4 Quantum mechanics in a black hole background 524
14.5 Cosmology 535
14.6 Cylindrical systems 543
14.7 Axially-symmetric systems 551
14.8 Notes 564
14.9 Exercises 565
Bibliography 568
Index 575
viii
1
Introduction
1
INTRODUCTION
1.1.1 Properties
Vector spaces are defined in terms of two objects. These are the vectors, which
can often be visualised as directions in space, and the scalars, which are usually
taken to be the real numbers. The vectors have a simple addition operation rule
with the following obvious properties:
a + b = b + a. (1.1)
a + (b + c) = (a + b) + c. (1.2)
a + 0 = a. (1.3)
a + (−a) = 0. (1.4)
For the case of directed line segments each of these properties has a clear geo-
metric equivalent. These are illustrated in figure 1.1.
Vector spaces also contain a multiplication operation between the scalars and
the vectors. This has the property that for any scalar λ and vector a, the product
λa is also a member of the vector space. Geometrically, this corresponds to the
dilation operation. The following further properties also hold for any scalars λ, µ
and vectors a and b:
2
1.1 VECTOR (LINEAR) SPACES
b b
c
a+b b+c
a a a
a+b
a+b+c
The preceding set of rules serves to define a vector space completely. Note that
the + operation connecting scalars is different from the + operation connecting
the vectors. There is no ambiguity, however, in using the same symbol for both.
The following two definitions will be useful later in this book:
(i) Two vector spaces are said to be isomorphic if their elements can be
placed in a one-to-one correspondence which preserves sums, and there
is a one-to-one correspondence between the scalars which preserves sums
and products.
(ii) If U and V are two vector spaces (sharing the same scalars) and all the
elements of U are contained in V, then U is said to form a subspace of V.
3
INTRODUCTION
λ1 a1 + · · · + λn an = 0. (1.6)
If such a set of scalars cannot be found, the vectors are said to be linearly
independent.
(iii) A set of vectors {a1 , . . . , an } is said to span a vector space V if every
element of V can be expressed as a linear combination of the set.
(iv) A set of vectors which are both linearly independent and span the space
V are said to form a basis for V.
These definitions all carry an obvious, intuitive picture if one thinks of vectors
in a plane or in three-dimensional space. For example, it is clear that two
independent vectors in a plane provide a basis for all vectors in that plane,
whereas any three vectors in the plane are linearly dependent. These axioms and
definitions are sufficient to prove the basis theorem, which states that all bases
of a vector space have the same number of elements. This number is called the
dimension of the space. Proofs of this statement can be found in any textbook
on linear algebra, and a sample proof is left to work through as an exercise. Note
that any two vector spaces of the same dimension and over the same field are
isomorphic.
The axioms for a vector space define an abstract mathematical entity which
is already well equipped for studying problems in geometry. In so doing we are
not compelled to interpret the elements of the vector space as displacements.
Often different interpretations can be attached to isomorphic spaces, leading to
different types of geometry (affine, projective, finite, etc.). For most problems
in physics, however, we need to be able to do more than just add the elements
of a vector space; we need to multiply them in various ways as well. This is
necessary to formalise concepts such as angles and lengths and to construct
higher-dimensional surfaces from simple vectors.
Constructing suitable products was a major concern of nineteenth century
mathematicians, and the concepts they introduced are integral to modern math-
ematical physics. In the following sections we study some of the basic concepts
that were successfully formulated in this period. The culmination of this work,
Clifford’s geometric product, is introduced separately in chapter 2. At various
points in this book we will see how the products defined in this section can all
be viewed as special cases of Clifford’s geometric product.
4
1.2 THE SCALAR PRODUCT
to our vector space. These are distances between points, which allow us to de-
fine a circle, and angles between vectors so that we can say that two lines are
perpendicular. The introduction of a scalar product achieves both of these goals.
Given any two vectors a, b, the scalar product a · b is a rule for obtaining a
number with the following properties:
(When we study relativity, this final property will be relaxed.) The introduction
of a scalar product allows us to define the length of a vector, |a|, by
√
|a| = (a·a). (1.7)
Here, and throughout this book, the positive square root is always implied by
√
the symbol. The fact that we now have a definition of lengths and distances
means that we have specified a metric space. Many different types of metric
space can be constructed, of which the simplest are the Euclidean spaces we
have just defined.
The fact that for Euclidean space the inner product is positive-definite means
that we have a Schwarz inequality of the form
(a + λb)·(a + λb) ≥ 0 ∀λ
2
⇒ a·a + 2λa·b + λ b·b ≥ 0 ∀λ
2
⇒ (a·b) ≤ a·a b·b, (1.9)
where the last step follows by taking the discriminant of the quadratic in λ.
Since all of the numbers in this inequality are positive we recover (1.8). We can
now define the angle θ between a and b by
Two vectors whose scalar product is zero are said to be orthogonal. It is usually
convenient to work with bases in which all of the vectors are mutually orthogonal.
If all of the basis vectors are further normalised to have unit length, they are
said to form an orthonormal basis. If the set of vectors {e1 , . . . , en } denote such
a basis, the statement that the basis is orthonormal can be summarised as
5
INTRODUCTION
where we have started to employ the Einstein summation convention that pairs
of indices in any expression are summed over. This convention will be assumed
throughout this book. The {ai } are the components of the vector a in the {ei }
basis. These are found simply by
ai = ei ·a. (1.14)
The scalar product of two vectors a = ai ei and b = bi ei can now written simply
as
a·b = (ai ei )·(bj ej ) = ai bj ei ·ej = ai bj δij = ai bi . (1.15)
In terms of vector arithmetic, neither the real nor imaginary parts of this ex-
pression have any geometric significance. A more geometrically useful product
6
1.4 QUATERNIONS
is defined instead by
zz ∗ = (x + iy)(x − iy) = x2 + y 2 , (1.17)
which returns the square of the length of the vector. A product of two vectors
in a plane, z and w = u + vi, can therefore be constructed as
zw∗ = (x + iy)(u − iv) = xu + vy + i(uy − vx). (1.18)
The real part of the right-hand side recovers the scalar product. To understand
the imaginary term consider the polar representation
z = |z|eiθ , w = |w|eiφ (1.19)
so that
zw∗ = |z||w|ei(θ − φ) . (1.20)
The imaginary term has magnitude |z||w| sin(θ − φ), where θ − φ is the angle
between the two vectors. The magnitude of this term is therefore the area of
the parallelogram defined by z and w. The sign of the term conveys information
about the handedness of the area element swept out by the two vectors. This
will be defined more carefully in section 1.6.
We thus have a satisfactory interpretation for both the real and imaginary
parts of the product zw ∗ . The surprising feature is that these are still both parts
of a complex number. We thus have a second interpretation for complex addition,
as a sum between scalar objects and objects representing plane segments. The
advantages of adding these together are precisely the advantages of working with
complex numbers as opposed to pairs of real numbers. This is a theme to which
we shall return regularly in following chapters.
1.4 Quaternions
The fact that complex arithmetic can be viewed as representing a product for
vectors in a plane carries with it a further advantage — it allows us to divide
by a vector. Generalising this to three dimensions was a major preoccupation
of the physicist W.R. Hamilton (see figure 1.2). Since a complex number x + iy
can be represented by two rectangular axes on a plane it seemed reasonable to
represent directions in space by a triplet consisting of one real and two complex
numbers. These can be written as x + iy + jz, where the third term jz represents
a third axis perpendicular to the other two. The complex numbers i and j have
the properties that i2 = j 2 = −1. The norm for such a triplet would then be
(x + iy + jz)(x − iy − jz) = (x2 + y 2 + z 2 ) − yz(ij + ji). (1.21)
The final term is problematic, as one would like to recover the scalar product
here. The obvious solution to this problem is to set ij = −ji so that the last
term vanishes.
7
INTRODUCTION
The anticommutative law ij = −ji ensures that the norm of a triplet behaves
sensibly, and also that multiplication of triplets in a plane behaves in a reasonable
manner. The same is not true for the general product of triplets, however.
Consider
8
1.4 QUATERNIONS
(a + ib + jc + kd)(a − ib − jc − kd)
= a2 + b2 + c2 + d2 (−k 2 ) − bd(ik + ki) − cd(jk + kj), (1.23)
where we have assumed that i2 = j 2 = −1 and ij = −ji. The expected norm of
the above product is a2 + b2 + c2 + d2 , which is obtained by setting k 2 = −1 and
ik = −ki and jk = −kj. So what values do we use for jk and ik? These follow
from the fact that ij = k, which gives
ik = i(ij) = (ii)j = −j (1.24)
and
kj = (ij)j = −i. (1.25)
Thus the multiplication rules for quaternions are
i2 = j 2 = k 2 = −1 (1.26)
and
ij = −ji = k, jk = −kj = i, ki = −ik = j. (1.27)
These can be summarised neatly as i2 = j 2 = k 2 = ijk = −1. It is a simple
matter to check that these multiplication laws define a closed algebra.
Hamilton was so excited by his discovery that the very same day he obtained
leave to present a paper on the quaternions to the Royal Irish Academy. The
subsequent history of the quaternions is a fascinating story which has been de-
scribed by many authors. Some suggested material for further reading is given
at the end of this chapter. In brief, despite the many advantages of working with
quaternions, their development was blighted by two major problems.
The first problem was the status of vectors in the algebra. Hamilton identified
vectors with pure quaternions, which had a null scalar part. On the surface
this seems fine — pure quaternions define a three-dimensional vector space.
Indeed, Hamilton invented the word ‘vector ’ precisely for these objects and this
is the origin of the now traditional use of i, j and k for a set of orthonormal
basis vectors. Furthermore, the full product of two pure quaternions led to the
definition of the extremely useful cross product (see section 1.5). The problem
is that the product of two pure vectors does not return a new pure vector, so
the vector part of the algebra does not close. This means that a number of ideas
in complex analysis do not extend easily to three dimensions. Some people felt
that this meant that the full quaternion product was of little use, and that the
scalar and vector parts of the product should be kept separate. This criticism
misses the point that the quaternion product is invertible, which does bring many
advantages.
The second major difficulty encountered with quaternions was their use in
9
INTRODUCTION
describing rotations. The irony here is that quaternions offer the clearest way
of handling rotations in three dimensions, once one realises that they provide
a ‘spin-1/2’ representation of the rotation group. That is, if a is a vector (a
pure quaternion) and R is a unit quaternion, a new vector is obtained by the
double-sided transformation law
a = RaR∗ , (1.28)
where the * operation reverses the sign of all three ‘imaginary’ components. A
consequence of this is that each of the basis quaternions i, j and k generates
rotations through π. Hamilton, however, was led astray by the analogy with
complex numbers and tried to impose a single-sided transformation of the form
a = Ra. This works if the axis of rotation is perpendicular to a, but otherwise
does not return a pure quaternion. More damagingly, it forces one to interpret
the basis quaternions as generators of rotations through π/2, which is simply
wrong!
Despite the problems with quaternions, it was clear to many that they were
a useful mathematical system worthy of study. Tait claimed that quaternions
‘freed the physicist from the constraints of coordinates and allowed thoughts to
run in their most natural channels’ — a theme we shall frequently meet in this
book. Quaternions also found favour with the physicist James Clerk Maxwell,
who employed them in his development of the theory of electromagnetism. De-
spite these successes, however, quaternions were weighed down by the increas-
ingly dogmatic arguments over their interpretation and were eventually displaced
by the hybrid system of vector algebra promoted by Gibbs.
a = a1 i + a2 j + a3 k, b = b1 i + b2 j + b3 k. (1.29)
ab = −ai bi + c, (1.30)
ci = ijk aj bk , (1.32)
10
1.6 THE OUTER PRODUCT
We recognise the preceding as defining the cross product of two vectors, a×b.
This has the following properties:
These properties can alternatively be viewed as defining the cross product, and
from them the algebraic definition can be recovered. This is achieved by starting
with a right-handed orthonormal frame {ei }. For these we must have
11
INTRODUCTION
12
1.6 THE OUTER PRODUCT
a∧b b∧a
b b
θ θ
Figure 1.4 The outer product. The outer or wedge product of a and b
returns a directed area element of area |a||b| sin(θ). The orientation of the
parallelogram is defined by whether the circuit a, b, −a, −b is right-handed
(anticlockwise) or left-handed (clockwise). Interchanging the order of the
vectors reverses the orientation and introduces a minus sign in the product.
the remarkable German mathematician H.G. Grassmann (see figure 1.3). His
work had its origin in the Barycentrischer Calcul of Möbius. There the author
introduced expressions like AB for the line connecting the points A and B and
ABC for the triangle defined by A, B and C. Möbius also introduced the
crucial idea that the sign of the quantity should change if any two points are
interchanged. (These oriented segments are now referred to as simplices.) It was
Grassmann’s leap of genius to realise that expressions like AB could actually be
viewed as a product between vectors. He thus introduced the outer or exterior
product which, in modern notation, we write as a ∧ b, or ‘a wedge b’.
The outer product can be defined on any vector space and, geometrically, we
are not forced to picture these vectors as displacements. Indeed, Grassmann
was motivated by a projective viewpoint, where the elements of the vector space
are interpreted as points, and the outer product of two points defines the line
through the points. For our purposes, however, it is simplest to adopt a pic-
ture in which vectors represent directed line segments. The outer product then
provides a means of encoding a plane, without relying on the notion of a vector
perpendicular to it. The result of the outer product is therefore neither a scalar
nor a vector. It is a new mathematical entity encoding an oriented plane and is
called a bivector. It can be visualised as the parallelogram obtained by sweep-
ing one vector along the other (figure 1.4). Changing the order of the vectors
reverses the orientation of the plane. The magnitude of a∧b is |a||b| sin(θ), the
same as the area of the plane segment swept out by the vectors.
The outer product of two vectors has the following algebraic properties:
13
INTRODUCTION
a∧b a∧c
b c
a a∧(b + c)
b+c
(ii) Bivectors form a linear space, the same way that vectors do. In two and
three dimensions the addition of bivectors is easy to visualise. In higher
dimensions this addition is not always so easy to visualise, because two
planes need not share a common line.
(iii) The outer product is distributive over addition:
14
1.6 THE OUTER PRODUCT
actual shape of the object is not conveyed by the result of the product. This can
be seen easily by defining a = a + λb and forming
a ∧b = a∧b + λb∧b = a∧b. (1.40)
The same bivector can therefore be generated by many different pairs of vectors.
In many ways it is better to replace the picture of a directed parallelogram with
that of a directed circle. The circle defines both the plane and a handedness,
and its area is equal to the magnitude of the bivector. This therefore conveys
all of the information one has about the bivector, though it does make bivector
addition harder to visualise.
15
INTRODUCTION
The components in this frame are therefore the same as those of the cross prod-
uct. But instead of being the components of a vector perpendicular to a and b,
they are the components of the bivector a ∧ b. It is this distinction which enables
the outer product to be defined in any dimension.
1.6.3 Handedness
We have started to employ the idea of handedness without giving a satisfactory
definition of it. The only space in which there is an unambiguous definition of
handedness is three dimensions, as this is the space we inhabit and most of us
can distinguish our left and right hands. This concept of ‘left’ and ‘right’ is
a man-made convention adopted to make our life easier, and it extends to the
concept of a frame in a straightforward way. Suppose that we are presented
with three orthogonal vectors {e1 , e2 , e3 }. We align the 3 axis with the thumb
of our right hand and then close our fist. If the direction in which our fist closes
is the same as that formed by rotating from the 1 to the 2 axis, the frame is
right-handed. If not, it is left-handed.
Swapping any pair of vectors swaps the handedness of a frame. Performing two
such swaps returns us to the original handedness. In three dimensions this corre-
sponds to a cyclic reordering, and ensures that the frames {e1 , e2 , e3 }, {e3 , e1 , e2 }
and {e2 , e3 , e1 } all have the same orientation.
There is no agreed definition of a ‘right-handed’ orientation in spaces of di-
mensions other than three. All one can do is to make sure that any convention
used is adopted consistently. In all dimensions the orientation of a set of vec-
tors is changed if any two vectors are swapped. In two dimensions one does
still tend to talk about right-handed axes, though the definition is dependent
on the idea of looking down on the plane from above. The idea of above and
below is not a feature of the plane itself, but depends on how we embed it in our
three-dimensional world. There is no definition of left or right-handed which is
intrinsic to the plane.
where the [ ] denotes antisymmetrisation. Grassmann was able to take this idea
further by defining an outer product for any number of vectors. The idea is a
simple extension of the preceding formula. Expressed in an orthonormal frame,
the components of the outer product on n vectors are the totally antisymmetrised
16
1.7 NOTES
products of the components of each vector. This definition has the useful prop-
erty that the outer product is associative,
a∧(b∧c) = (a∧b)∧c. (1.45)
For example, in three dimensions we have
a∧b∧c = (ai ei )∧(bj ej )∧(ck ek ) = ijk ai bj ck e1 ∧e2 ∧e3 , (1.46)
which represents a directed volume (see section 2.4).
A further feature of the antisymmetry of the product is that the outer product
of any set of linearly dependent vectors vanishes. This means that statements like
‘this vector lies on a given plane’, or ‘these two hypersurfaces share a common
line’ can be encoded algebraically in a simple manner. Equipped with these
ideas, Grassmann was able to construct a system capable of handling geometric
concepts in arbitrary dimensions.
Despite Grassmann’s considerable achievement, the book describing his ideas,
his Lineale Ausdehnungslehre, did not have any immediate impact. This was
no doubt due largely to his relative lack of reputation (he was still a German
schoolteacher when he wrote this work). It was over twenty years before anyone
of note referred to Grassmann’s work, and during this time Grassmann produced
a second, extended version of the Ausdehnungslehre. In the latter part of the
nineteenth century Grassmann’s work started to influence leading figures like
Gibbs and Clifford. Gibbs wrote a number of papers praising Grassmann’s work
and contrasting it favourably with the quaternion algebra. Clifford used Grass-
mann’s work as the starting point for the development of his geometric algebra,
the subject of this book.
Today, Grassmann’s ideas are recognised as the first presentation of the ab-
stract theory of vector spaces over the field of real numbers. Since his death, his
work has given rise to the influential and fashionable areas of differential forms
and Grassmann variables. The latter are anticommuting variables and are fun-
damental to the foundations of much of modern supersymmetry and superstring
theory.
1.7 Notes
Descriptions of linear algebra and vector spaces can be found in most intro-
ductory textbooks of mathematics, as can discussions of the scalar and cross
products and complex arithmetic. Quaternions, on the other hand, are much less
likely to be mentioned. There is a large specialised literature on the quaternions,
and a good starting point are the works of Altmann (1986, 1989). Altmann’s
paper on ‘Hamilton, Rodriques and the quaternion scandal’ (1989) is also a good
introduction to the history of the subject.
The outer product is covered in most modern textbooks on geometry and
17
INTRODUCTION
physics, such as those by Nakahara (1990), Schutz (1980), and Gockeler &
Schucker (1987). In most of these works, however, the exterior product is only
treated in the context of differential forms. Applications to wider topics in geom-
etry have been discussed by Hestenes (1991) and others. A useful summary in
provided in the proceedings of the conference Hermann Gunther Grassmann
(1809–1877), edited by Schubring (1996). Grassmann’s Lineale Ausdehnun-
gslehre is also finally available in English translation due to Kannenberg (1995).
For those with a deeper interest in the history of mathematics and the develop-
ment of vector algebra a good starting point is the set of books by Kline (1972).
There are also biographies available of many of the key protagonists. Perhaps
even more interesting is to return to their original papers and experience first
hand the robust and often humorous language employed at the time. The col-
lected works of J.W. Gibbs (1906) are particularly entertaining and enlightening,
and contain a good deal of valuable historical information.
1.8 Exercises
1.1 Suppose that the two sets {a1 , . . . , am } and {b1 , . . . , bn } form bases for
the same vector space, and suppose initially that m > n. By establishing
a contradiction, prove the basis theorem that all bases of a vector space
have the same number of elements.
1.2 Demonstrate that the following define vector spaces:
(a) the set of all polynomials of degree n;
(b) all solutions of a given linear ordinary differential equation;
(c) the set of all n × m matrices.
1.3 Prove that in Euclidean space |a + b| ≤ |a| + |b|. When does equality
hold?
1.4 Show that the unit quaternions {±1, ±i, ±j ± k} form a discrete group.
1.5 The unit quaternions i, j, k are generators of rotations about their re-
spective axes. Are rotations through either π or π/2 consistent with the
equation ijk = −1?
1.6 Prove the following:
(a) a·(b×c) = b·(c×a) = c·(a×b);
(b) a×(b×c) = a·c b − a·b c;
(c) |a×b| = |a| |b| sin(θ), where a·b = |a| |b| cos(θ).
1.7 Prove that the dimension of the space formed by the exterior product
of m vectors drawn from a space of dimension n is
n(n − 1) · · · (n − m) n!
= .
1 · 2···m (n − m)!m!
18
1.8 EXERCISES
1.8 Prove that the n-fold exterior product of a set of n dependent vectors is
zero.
1.9 A convex polygon in a plane is specified by the ordered set of points
{x0 , x1 , . . . , xn }. Prove that the directed area of the polygon is given by
A = 12 (x0 ∧x1 + x1 ∧x2 + · · · + xn ∧x0 ).
What is the significance of the sign? Can you extend the idea to a
triangulated surface in three dimensions?
19