Hormander Riemannian Geometry
Hormander Riemannian Geometry
1990
Document Version:
Other version
Link to publication
General rights
Unless other specific re-use rights are stated the following general rights apply:
Copyright and moral rights for the publications made accessible in the public portal are retained by the authors
and/or other copyright owners and it is a condition of accessing publications that users recognise and abide by the
legal requirements associated with these rights.
• Users may download and print one copy of any publication from the public portal for the purpose of private study
or research.
• You may not further distribute the material or use it for any profit-making activity or commercial gain
• You may freely distribute the URL identifying the publication in the public portal
LUNDUNI
VERSI
TY
PO Box117
22100L und
+4646-2220000
RIEMANNIAN GEOMETRY
Lars Hörmander
Department of Mathematics
Box 118, S-221 00 Lund, Sweden
RIEMANNIAN GEOMETRY
Lectures given during the fall of 1990 by
Lars Hörmander
Department of Mathematics
Box 118, S-221 00 Lund, Sweden
i
Lars Hörmander
Department of Mathematics
University of Lund
Box 118
S-221 00 Lund, Sweden
e-mail lvh@maths.lth.se (internet)
ii
PREFACE
The inspiration for the theory of partial differential equations has always come from
two main sources, physics and geometry. The interaction between all three areas has
become intensified in recent years. The solution of the index problem by Atiyah and
Singer in the early 60’s forced the people working on differential equations to improve
their knowledge of differential geometry. This was very useful in the subsequent de-
velopment of microlocal analysis which mainly involved symplectic geometry, a topic
which had previously mainly been cultivated in connection with ordinary differential
equations. Riemannian geometry is central in the recent development of gauge theories,
which rely on a mixture of geometry, physics and partial differential equations. There
has also been a great deal of recent activity in the general theory of relativity, that is, in
pseudo-Riemannian geometry — for example, the proof of the positive mass conjecture
and global existence theorems for the vacuum Einstein equations. The solution of some
purely geometric problems, such as the Yamabe problem and the isometric imbedding
problem, have also enriched the theory of non-linear partial differential equations. The
fundamental open problems in the theory of overdetermined systems of linear differ-
ential equations also require a strong background in geometry even to understand the
present state of affairs. All this should be sufficient reason for an analyst to study
geometry seriously.
In a half semester course it is only possible to present a brief outline of the most clas-
sical Riemannian geometry with a few glimpses of more recent developments. However,
if enough interest is manifested, I plan to continue for one or several semesters more
in order to be able to approach the research front, and these lectures should then be a
convenient platform to build on. One could continue in many different directions. One
possibility is to discuss pseudo-Riemannian manifolds with Lorentz signature (general
relativity theory). Another is to discuss relations to the theory of functions of several
complex variables; the presence of a complex structure gives a much richer structure.
Lars Hörmander
iii
iv
Contents of Lectures on Riemannian Geometry
Bibliography . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 195
Index . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 197
vi
CHAPTER I
Summary. In Section 1.1 we shall just recall the definition of the curvature of a curve in
a Euclidean vector space. The special two dimensional case where the curvature can be
given a sign is studied in somewhat greater depth in Section 1.2. We define the torsion
of a curve and prove the Frenet formulas in Section 1.3. This leads to a discussion of
moving frames in Section 1.4 which is a preparation for the Riemannian geometry in
Chapter IV.
(1.1.1) I ∋ t 7→ x(t) ∈ V
Thus the second derivative X ′′ (s) is orthogonal to the tangent X ′ (s) of the curve.
To interpret it geometrically we consider a circle with two perpendicular radii y, z of
length R and center x0 ; it is parametrized by
Since (y, y) = (z, z) = R2 and (y, z) = 0, we confirm using the second equation that
kx′ (s)k2 = 1. The second derivative
is directed toward the center of the circle and kx′′ (s)k = 1/R.
Definition 1.1.1. If s 7→ x(s) ∈ V is a C 2 curve parametrized by the arc length,
then x′ (s) is the unit tangent vector at x(s). If x′′ (s) 6= 0, then n = x′′ (s)/kx′′ (s)k
is called the principal normal of the curve at x(s), 1/kx′′ (s)k is called the radius
of curvature, and the circle with center at x(s) + x′′ (s)/kx′′ (s)k2 lying in the plane
spanned by the vectors x′ (s) and x′′ (s) at x(s) is called the osculating circle. One calls
κ = kx′′ (s)k the curvature at x(s) even if κ = 0; otherwise x′′ (s) = κn.
From the discussion above it is clear that when the curvature is not 0 then the
osculating circle is the only one which has a tangency of higher order with the curve
at x(s).
Now suppose that we have a curve t 7→ x(t) with x′ (t) 6= 0 which may not have the
arc length as parameter. If we write x(t) = X(s(t)) as above, we obtain
x′ (t) = X ′ (s)ds/dt, x′′ (t) = X ′′ (s)(ds/dt)2 + X ′ (s)d2 s/dt2 .
Since X ′′ (s) is orthogonal to X ′ (s) we have
X ′′ (s)(ds/dt)2 = x′′ (t) − (x′′ (t), X ′ (s))X ′ (s)
or equivalently,
(1.1.2) X ′′ (s) = x′′ (t)/kx′ (t)k2 − x′ (t)(x′′ (t), x′ (t))/kx′ (t)k4 .
Exercise 1.1.1. Calculate the curvature of a C 2 curve t 7→ x(t) with arbitrary
parametrization.
Exercise 1.1.2. Show that the tangential component of x′′ (t) is equal to d2 s/dt2
times the unit tangent vector, and that the normal component is (ds/dt)2 κ times the
principal normal.
1.2. The two dimensional case. For curves in the plane R2 we can attach a
sign to the curvature (which depends on the orientation of the plane). If s 7→ x(s)
is a curve with the arc length s as parameter and if x′′ (s) 6= 0, then we change the
direction of the normal n(s) so that x′ (s) and n(s) are positively oriented, that is,
det(x′ (s), n(s)) > 0 and define the curvature κ(s) so that x′′ (s) = κ(s)n(s) as before.
Thus κ > 0 means that moving along the curve one sees the curve to the left of the
tangent.
Consider now a closed simple C 2 curve
R/LZ ∋ s 7→ x(s) ∈ R2
where s is still the arc length and x(s) = x(s′ ) if and only if s − s′ ∈ LZ. We can write
x′ (s) = (cos θ(s), sin θ(s)), 0 ≤ s ≤ L,
where θ is a C 1 function uniquely determined up to an integer multiple of 2π. Since
x′′ (s) = (− sin θ, cos θ)dθ/ds = κn = κ(− sin θ, cos θ),
we have κ = dθ/ds. The integral
Z L Z L
(1.2.1) κ ds = dθ = θ(L) − θ(0)
0 0
is obviously a multiple of 2π.
THE TWO DIMENSIONAL CASE 3
Theorem 1.2.1. The integral (1.2.1) of the curvature of a simple closed curve is
±2π with the positive sign if the curve lies entirely to the left of some tangent when
one moves around it in the positive direction.
Proof. The normalized chord direction
We want to find
where the last equality follows from the fact that x(L, s) = −x(s, 0). Choose the origin
of the parametrization so that x2 (0) = mins x2 (s). Then ϕ(s, 0) only varies between 0
and π. If x′1 (0) > 0 then ϕ(0, 0) = 0 and ϕ(L, 0) = π so the integral of the curvature
becomes 2π. If x′1 (0) < 0 the sign changes, which proves the theorem.
Exercise 1.2.1. Show that when x′1 (0) = 1 in the preceding argument, then the
variation d(y) of the argument of (x(s) − y)/kx(s) − yk for 0 ≤ s ≤ L is either 0 or 2π
for all points y in the complement Ω of the curve, and that Ωe = {y ∈ Ω; d(y) = 0} is
an open connected unbounded set while Ωi = {y ∈ Ω; d(y) = 2π} is an open connected
bounded set, the interior of the curve. (Jordan’s curve theorem.) Hint: Examine first
a neighborhood of x(0), then a neighborhood of the curve.
Exercise 1.2.2. Show that if Γ : [0, T ] ∋ s 7→ x(s) ∈ R3 is a simple closed C 2
curve parametrized by the arc length, and N ∈ S 2 , then s 7→ x(s) − N (x(s), N ) is a
C 2 curve in the orthogonal plane of N for almost all N . Write an expression for its
RT
curvature and conclude that if κ(s) is the curvature of Γ, then 0 κ(s) ds ≥ 2π. (See
also Milnor [1].)
Exercise 1.2.3. Show that if
R/T Z ∋ t 7→ x(t) ∈ R2
1.3. The Frenet formulas. We shall now return to the study of a C k curve
(1.1.1) taking the arc length s as parameter. If the curvature is not equal to 0 at
x(s0 ), then the principal normal n(s) is defined for s near s0 . If k ≥ 3 we can take
its derivative n′ (s). Since (n(s), n(s)) = 1 we obtain 2(n′ (s), n(s)) = 0, and since
(n(s), x′ (s)) = 0 we have
(n′ (s), x′ (s)) + (n(s), x′′ (s)) = 0, that is, (n′ (s), x′ (s)) = −κ(s).
Hence n′ (s) + κ(s)x′ (s) is orthogonal to the plane spanned by x′ (s) and n(s). The
length τ (s) of this vector is called the torsion of the curve at x(s). If τ (s) 6= 0, then
normalization gives a unit vector b(s), called the binormal of the curve at x(s), and we
have
n′ (s) = −κ(s)x′ (s) + τ (s)b(s).
The procedure can be continued by differentiation of b with respect to s. Differentiation
of the equations
gives
If dim V = 3 the vectors x′ (s), n(s) and b(s) form a complete orthonormal system, so
we get the third of the Frenet formulas
However, the procedure becomes more illuminating in the higher dimensional case if
we use somewhat different notation.
Consider a C k+1 curve (1.1.1) with arbitrary parametrization, such that for all t ∈ I
the derivatives x′ (t), . . . , x(k) (t) are linearly independent but x′ (t), . . . , x(k+1) (t) are
linearly dependent. Clearly k ≤ dim V . For any given sufficiently differentiable curve
this is true if I is replaced by any interval in the dense open subset where the rank
of x′ (t), x′′ (t), . . . is locally maximal. The hypothesis is independent of the choice of
parametrization, and so is the linear span Ej (t) of x′ (t), . . . , x(j) (t) when j ≤ k. In
fact, if t̃ is another parameter then
j
dj x dj x
dt
− ∈ Ej−1
dt̃j dt̃ dtj
THE FRENET FORMULAS 5
where cj (t) > 0 is chosen so that kej (t)k = 1. These vectors at x(t) do not depend
on the choice of parametrization; e1 (t) is the unit tangent vector, e2 (t) is the principal
normal, e3 (t) is the binormal, and so on.
Assume now for the sake of simplicity that the parameter t is the arc length s. By
the construction and the hypothesis that Ek+1 (s) = Ek (s) we have
k
X
e′i (s) = ωij (s)ej (s), i ≤ k,
j=1
where ωij (s) = (e′i (s), ej (s)) = 0 if j > i + 1. If we differentiate the equations
0 = (e′i (s), ej (s)) + (ei (s), e′j (s)) = ωij (s) + ωji (s),
so the matrix (ωij (s)) is skew symmetric. Thus ωii (s) = 0, and since ωij (s) = 0 when
j > i + 1, this is also true when i > j + 1. If we set κj (s) = ωj,j+1 (s), 1 ≤ j < k, it
follows that the matrix (ωij ) has the special form for k = 4, say:
0 κ1 0 0
−κ1 0 κ2 0
.
0 −κ2 0 κ3
0 0 −κ3 0
Here κ1 (s) is of course the curvature, and κ2 is the torsion. The Frenet formulas can
now be written
(1.3.1) e′1 (s) = κ1 (s)e2 (s), e′j (s) = −κj−1 (s)ej−1 (s) + κj (s)ej+1 (s), 1 < j < k,
e′k (s) = −κk−1 (s)ek−1 (s).
This system of differential equations has a unique solution with given initial data. If
these are orthonormal then the solution remains orthonormal, and the solution must
belong to the space Ek (s0 ), for there is a solution with e1 (s), . . . , ek (s) contained in
this space which is equal to e1 (s0 ), . . . , ek (s0 ) at s0 , so Ek (s) is independent of s and
the curve lies in an affine subspace of dimension k.
Exercise 1.3.1. Let k ≥ 3 and determine the Taylor expansion of x(s) with error
o(s3 ) at 0 ∈ I in terms of x(0), e1 (0), e2 (0), e3 (0) and the Taylor expansions of κ1 and
κ2 at 0.
6 I. CURVES IN A EUCLIDEAN SPACE
This means that infinitesimal rotations are defined by skew symmetric matrices. Every
such infinitesimal rotation ω gives rise to a one parameter group of rotations O(t), that
is a function R ∋ t 7→ O(t) ∈ O(n) such that
The first condition implies O(0) = I, and differentiation with respect to s gives O ′ (t) =
O(t)ω when s = 0, hence
X∞
tω
O(t) = e = (tω)j /j!.
0
MOVING FRAMES 7
The properties (1.4.4) follow at once, and since t O(t) = O(−t) it follows that O(t) ∈
O(n).
Let us consider the examples of lowest dimension. If n = 2 then
0 θ tω cos(tθ) sin(tθ)
ω= , e = ,
−θ 0 − sin(tθ) cos(tθ)
Noting that ωθ = 0 it is easy to see that etω means rotation by the angle t|θ| around
θ.
Exercise 1.4.1. Show that if
0 θ3 −θ2
ω = −θ3
0 θ1 ,
θ2 −θ1 0
Show that
(1) g is a linear subspace of M(n).
(2) if g ∈ G and X ∈ g, then
getX g −1 = etY , t ∈ R,
G ∋ eXk e−Zk = eYk +Zk e−Zk = eWk , where Wk = Yk + O(|Yk ||Zk |).
Summary. In Section 2.1 we introduce the notions of first and second fundamental
forms for a submanifold of a Euclidean space. After introducing the Christoffel symbols
and geodesic curvature, we define the Riemann curvature tensor and connect it with
the second fundamental form through the Gauss equations. Section 2.2 is devoted to
the special case of a hypersurface, and in particular the Gauss map. Basic algebraic
(symmetry) properties of the curvature tensor are given in Section 2.3. The discussion
includes the first Bianchi identity and the decomposition of the curvature tensor with
special emphasis on the four dimensional case.
definition for abstract manifolds in Chapter III. We shall consistently use upper case
letters for points in V and lower case letters for points in the parameter space. It is
the latter which will survive in Chapter III.
If ν ≥ 1 then the tangent of our curve has the direction
n
dX df (x(t)) X dxj
(2.1.1) = = fj (x(t)); fj (x) = ∂f (x)/∂xj = ∂j f (x).
dt dt 1
dt
Thus the tangent vectors of all C 1 curves passing through f (x) span the plane
Tf (x) M = f ′ (x)Rn ,
One calls the quadratic form (2.1.2) the first fundamental form of M . It is a quadratic
form depending on x ∈Pω, which can invariantly be regarded as a quadratic form in
n
Tf (x) M , for Rn ∋ τ 7→ 1 τj fj (x) = X ∈ Tf (x) M is a bijection, and the form is equal
to the square of the norm of the tangent vector X in V .
Assuming from now on that µ ≥ 2 and that we have a C 2 curve, we compute the
second derivative:
n n
d2 f (x(t)) X d2 xj X dxj dxk
(2.1.3) 2
= 2
fj (x(t)) + fjk (x(t)),
dt j=1
dt dt dt
j,k=1
where fjk (x) = ∂j ∂k f (x). Second derivatives of x just occur in the first sum, and the
coefficients there are tangent vectors. If we project on the normal plane Nf (x) M of
Tf (x) M , we can eliminate this sum and obtain:
Theorem 2.1.1 (Meusnier). Let hjk (x), j, k = 1, . . . , n, be the orthogonal projec-
tion of fjk (x) = ∂j ∂k f (x) in the normal plane Nf (x) M of M at f (x). For every C 2
curve on M with the unit tangent vector f ′ (x)τ at f (x), the sum
n
X
(2.1.4) hjk (x)τj τk
j,k=1
is then equal to the curvature times the orthogonal projection of the principal normal
in Nf (x) M .
We can regard (2.1.4) as a quadratic form H in Tf (x) M with values in Nf (x) M ; by
Theorem 2.1.1 it is then independent of the choice of parametrization. One calls (2.1.4)
CURVES ON A SUBMANIFOLD OF A EUCLIDEAN SPACE 13
the second fundamental form of M . Classically it was defined for hypersurfaces (in R3 ).
In that case, an orientation of M identifies Nf (x) M with R, so one can then regard H
as a real valued quadratic form. We shall sometimes introduce an orthonormal basis in
Nf (x) M and will then be able to write H as an N − n tuple of scalar quadratic forms.
We shall now consider the tangential components of (2.1.3). Since f1 (x), . . . , fn (x)
form a basis in Tf (x) M , we can write
n
X
(2.1.5) fik (x) = Γik l (x)fl (x) + hik (x),
l=1
for the normal component is hik (x) by definition. To calculate the coefficients Γik l we
take the scalar product with fj which gives
Xn
(2.1.6) Γik l glj = (fik , fj ).
l=1
By a miracle one can compute the right-hand side by means of the coefficients of the
first fundamental form and their derivatives, for we have
∂k gij = (fik , fj ) + (fi , fjk )
∂i gjk = (fji , fk ) + (fj , fki ),
∂j gki = (fkj , fi ) + (fk , fij ).
The desired quantity occurs in the first two equations so we add them and subtract
the third, which gives
(2.1.7) Γikj = (fik , fj ) = 21 (∂k gij + ∂i gjk − ∂j gki ),
where the first equality is a new definition. Note the symmetry in the indices i, k, and
that
(2.1.7)′ ∂k gij = Γikj + Γjki .
If (g ij ) denotes the inverse of the matrix (gij ) we obtain from (2.1.6) and (2.1.7)
Xn n
X
l lj
(2.1.8) Γik = g (fik , fj ) = g lj Γikj .
j=1 j=1
Definition 2.1.2. The functions Γikj and Γik j are called Christoffel symbols of
the first and second kind; they are determined by the first fundamental form.
l
Remark. The classical notation was [ij, k] and instead of Γijk and Γij l .
ij
√
Exercise 2.1.1. Prove that if g = det(gij ), then ∂l g = 2g i Γil i , that is, ∂l g =
P
√
g i Γil i .
P
Summing up the preceding discussion, we can now write (2.1.3) in the form
n n n
d2 f (x(t)) X d2 xj X
j dxi dxk
X dxi dxk
(2.1.9) 2
= 2
+ Γik f j (x) + hik (x).
dt j=1
dt dt dt dt dt
i,k=1 i,k=1
We discussed the normal component in Theorem 2.1.1, and we can now examine the
tangential component:
14 II. CURVATURE OF SUBMANIFOLDS OF A EUCLIDEAN SPACE
Thus the vector (2.1.10) does not depend on the choice of parametrization f . The
length is called the geodesic curvature of the curve at f (x), and the direction is called
the geodesic principal normal direction.
In (2.1.10) we have just computed the derivative of the tangent vector along the
curve. However, the argument is much more general. Suppose that we have a C 1
vector field v defined in a neighborhood of the curve and that v is everywhere tangent
to M . Then we can write the vector at f (x) as
n
X
v(x) = vj (x)fj (x),
1
and we obtain
n
dv(x(t)) X
= ((∂k vj (x))fj (x) + vj (x)fjk (x))x′k .
dt
j,k=1
This is called the covariant derivative of the vector field along the curve. Note that
one does not need to know the embedding function in order to compute (2.1.11); it
suffices to know the first fundamental form. (The right-hand side of (2.1.11) is also well
defined if v is just given on the curve.) This aspect will be discussed systematically in
Chapter III.
Now we pass to determining the tangential components of the derivatives of a C 1
normal vector field n(x) at f (x). We can write
n
X
∂i n(x) = nik (x)fk (x) mod Nf (x) .
k=1
it follows that
n
X
(2.1.12) nik = − (hij , n)g kj .
j=1
P
This means that y 7→ yi ∂i n(x) mod Nf (x) regarded as a linear transformation in
Tf (x) is precisely the linear transformation defined by the scalar product (H, −n) of
the second fundamental form and −n, using the identification of quadratic forms and
linear transformations given by the first fundamental form.
It is now easy to find the tangential component of fijk = ∂i ∂j ∂k f . (The following
formulas should be understood in the sense of distribution theory if M is only in C 2 ,
but one can also assume that M ∈ C 3 and use approximation to extend P the final
2 l
formulas where no third derivatives occur to the C case.) Since fij = Γij fl + hij
we obtain
n
X n
X n
X
fkij = ∂k fij = (∂k Γij m + Γij l Γlk m − (hij , hkl )g lm )fm mod Nf (x) .
m=1 l=1 l=1
Again we note that the expression (2.1.13)′ shows that Rijkl can be computed from
the first fundamental form alone while (2.1.15) only involves the second fundamental
form. From (2.1.15) we also see that the corresponding 4-linear form
n
X
1 2 3 4
R(t , t , t , t ) = Rijkl t1i t2j t3k t4l ; t1 , . . . , t4 ∈ R n ,
i,j,k,l=1
is antisymmetric in the pair t1 , t2 and in the pair t3 , t4 but symmetric for exchange
of the pairs. Considered as a 4-linear form in f ′ (x)t1 , . . . , f ′ (x)t4 ∈ Tf (x) M the form
R is independent of the choice of parametrization since thisP is true for the symmetric
bilinear map H : Tf (x) M × Tf (x) M → Nf (x) M defined by hij t1i t2j . In fact, for the
corresponding form R e we have by (2.1.15)
In particular, the symmetric functions are given by the coefficients in this equation,
n
X n
X n
Y
ij
(2.2.2) Ki = h̃ij g , Ki = det(h̃ij )/ det(gij ).
i=1 i,j=1 i=1
Definition 2.2.1. For an oriented hypersurface the eigenvectors of the second fun-
damental form with respect to the first fundamental form, both regarded as quadratic
forms on the tangent space Tx M of M at x, are called principal curvature directions at
x. The corresponding eigenvalues are called principal curvatures. Their product and
sum, given by (2.2.2), are called the total (or Gauss) curvature and the mean curvature
respectively.
Exercise 2.2.1. Write down explicitly the formulas for the mean curvature and
the total curvature of a surface of the form xn+1 = ϕ(x1 , . . . , xn ) in Rn+1 , where
ϕ ∈ C2.
Since (n, n) = 1 we have (n, ∂i n) = 0, so (2.1.12) gives
n
X
∂i n = − h̃ij g kj fk ,
j,k=1
or equivalently
n
X n
X n
X
( ti ∂i n, s k fk ) = − h̃ij ti sj .
i=1 k=1 i,j=1
18 II. CURVATURE OF SUBMANIFOLDS OF A EUCLIDEAN SPACE
We can of course equally well sum for i < j, omitting the factor 1/2. In the particular
case where n = 2 we only have one term then and conclude that R(t e 1 , t2 , t3 , t4 ) is the
product of the total curvature K = K1 K2 and the areas of the parallelograms spanned
by t1 , t2 and by t3 , t4 . Since R is determined by the first fundamental form, we have
in particular proved:
ALGEBRAIC PROPERTIES OF THE CURVATURE TENSOR 19
When n = 2 it is the Gauss curvature times the firstP fundamental form. For any n
it has diagonal form, and the diagonal elements are j;j6=i Ki Kj , for i = 1, . . . , n.
This is the sum of the Gauss curvatures of the sections with three dimensional planes
containing n, ei and ej for some j 6= i. The scalar curvature is
X n
X X
(2.2.6) S= Ki Kj = ( Ki )2 − Ki2 ,
i6=j 1
only depends on the two plane spanned by the tangent vectors t1 and t2 ; the de-
nominator is the square of the area of the parallelogram spanned by them. Because
of Theorem 2.2.3 one calls this quotient the sectional curvature for the two plane in
Tx M .
We shall now determine if there are additionals restrictions on thePtensors which
may occur at a point. This is a simple problem in linear algebra. Let hjk xj xk be a
n r
quadratic form in R with coefficients hjk = hkj in R where r = N − n, and define
as in (2.1.15)
P 2
E is a Euclidean space with the norm square Rijkl , so every linear form on E can
be written X
R 7→ Rijkl Sijkl
where S ∈ E. If the linear form vanishes on the term in (2.2.4) with i = 1, j = 2, then
we have for the 4-linear form defined by S
(2.3.3) S(e1 , e2 , e1 , e2 ) = 0,
for (2.3.2) implies that the four terms in R̃ give the same contribution. Since
(2.3.3)′ S(e1 , e2 , e1 , e4 ) = 0, e1 , e2 , e4 ∈ Rn ,
for this is a symmetric bilinear form in e2 and e4 . Trying to polarize again we just
obtain
S(e1 , e2 , e3 , e4 ) + S(e3 , e2 , e1 , e4 ) = 0,
or if we use the symmetries (2.3.2)
S(e1 , e2 , e3 , e4 ) = S(e1 , e4 , e2 , e3 ).
Because of the symmetries (2.3.2) we can rearrange the arguments so that t1 stands
first and the permutation is positive, so the condition is equivalent to
This is called the first Bianchi identity. It is often written Ri[jkl] = 0 where the bracket
is read as summation over all circular permutations. The symmetry conditions are not
independent of each other. The condition on Rijlk in (2.3.2) is obviously a consequence
of the others, and the condition on Rklij follows from the first two equalities in (2.3.2)
and the Bianchi identity. In fact, they give
R(t1 , t2 , t3 , t4 )
= −R(t1 , t4 , t2 , t3 ) − R(t1 , t3 , t4 , t2 ) = R(t4 , t1 , t2 , t3 ) + R(t3 , t1 , t4 , t2 )
= −R(t4 , t3 , t1 , t2 ) − R(t4 , t2 , t3 , t1 ) − R(t3 , t2 , t1 , t4 ) − R(t3 , t4 , t2 , t1 )
= 2R(t3 , t4 , t1 , t2 ) + R(t2 , t4 , t3 , t1 ) + R(t2 , t3 , t1 , t4 )
= 2R(t3 , t4 , t1 , t2 ) − R(t2 , t1 , t4 , t3 ) = 2R(t3 , t4 , t1 , t2 ) − R(t1 , t2 , t3 , t4 )
R∈T, R(t1 , t2 , t1 , t2 ) = 0, if t1 , t2 ∈ Rn =⇒ R = 0.
In fact, the condition on R here is the condition (2.3.3), which means that R is orthog-
onal to T , hence equal to 0.
A symmetric bilinear form in ν variables has ν(ν + 1)/2 independent coefficients
ν
whereas an antisymmetric bilinear form has 2 = ν(ν − 1)/2 independent coefficients.
Since we can interpret E as a space of symmetric bilinear forms on the space Rn ∧ Rn
which has dimension ν = n(n−1)/2, it is clear that dim E = ν(ν+1)/2. The orthogonal
space of T in E has dimension n4 , so it follows that
Exercise 2.3.1. Use the fact that (2.3.1) is invariant if h is replaced by Oh where
O ∈ O(r) to show that (2.3.5) can be strengthened to
Classical results to be discussed in Chapter III will prove that one can always take
r = n(n − 1)/2. Using representation theory Berger, Bryant and Griffiths [1] have
n−1
proved that one can always take r = 2 + 2, but the best value of r does not seem
to be known. Such a value is an obvious lower bound for the codimension with which a
general Riemannian manifold can be locally C 2 embedded. In Chapter III we shall see
that a local embedding with high regularity is usually not possible with codimension
n
lower than 2 .
The full linear group GL(n) acts on T by
There is no invariant subspace for this operation (see Berger, Bryant and Griffiths [1]).
However, in the context where we encountered the Riemann curvature tensor only the
operation of O(n) is natural, since it corresponds to changing orthonormal basis in
the tangent space. At the end of Section 2.1 we also defined the Ricci tensor, which
belongs to the space S 2 (Rn ) of symmetric bilinear forms in Rn . The passage from
the Riemann tensor to the Ricci tensor commutes with the operation of O(n), so the
kernel W consisting of all R ∈ T such that
n
X
(2.3.6) Rijil = 0, j, l = 1, . . . , n,
i=1
is invariant under O(n). (In view of the symmetries (2.3.2) the contraction of R with
respect to any pair of indices is equal to 0 if R ∈ W.) Since the Euclidean norm we
introduced in E above is invariant under O(n), the orthogonal complement of W in T is
also invariant under O(n). We shall determine the decomposition of a general R ∈ T
in components belonging to these spaces, but first we shall make some elementary
remarks on how O(n) acts on S 2 (Rn ).
The metric scalar product (·, ·) in Rn is of course invariant under O(n), and so is
its orthogonal space consisting of symmetric forms which are traceless (with respect to
the Euclidean form). If a O(n) invariant subspace of S 2 (Rn ) contains
P a form which is
not proportional to the metric form, then it contains a form λj xj yj with λ1 6= λ2 .
Hence it also contains the form with λ1 and λ2 interchanged and therefore the form
x1 y1 − x2 y2 and so the form xj yj − xk yk for arbitrary j and k. The linear hull contains
all diagonal forms with zero trace, so the only O(n) invariant subspaces of S 2 (Rn )
are the multiples of the metric form and the forms with zero trace. If n ≥ 3, as we
assume now, it follows at once from (2.2.5) that both the metric form and other forms
can occur as Ricci tensors, so the map T (Rn ) → S 2 (Rn ) corresponding to passage
ALGEBRAIC PROPERTIES OF THE CURVATURE TENSOR 23
from the Riemann tensor to the Ricci tensor is surjective. (It is easy to show that any
symmetric tensor can in fact occur as Ricci tensor for a manifold with codimension
r = 2.)
From the surjectivity just proved it follows that
P the orthogonal space of W in T
is of dimension n(n + 1)/2. If R ∈ T and S = Rijij is the scalar curvature while
Bij = Rij − Sδij /n is the traceless Ricci tensor, then we claim that
(2.3.7) Rijkl = Wijkl
+ (Bik δjl − Bil δjk + Bjl δik − Bjk δil )/(n − 2) + S(δik δjl − δil δjk )/n(n − 1)
decomposes R in an element in W and two orthogonal ones, one which has traceless
Ricci tensor and one which has Ricci tensor proportional to the metric tensor. To
verify this it suffices to show that the last two terms are in T , for they are obviously
orthogonal to W and the contractions with respect to i, k are
(0 − Bjl + nBjl − Bjl )/(n − 2) = Bjl resp. S(nδjl − δjl )/n(n − 1) = Sδjl /n.
The symmetries (2.3.2) are obviously valid, and the Bianchi condition is then equivalent
to orthogonality to all antisymmetric 4-linear forms, which is equally obvious since
every term is symmetric in some pair of indices.
Definition 2.3.2. The component W ∈ W of R ∈ T defined by (2.3.7) is called
the Weyl tensor or the conformal curvature tensor. We have
(2.3.7)′ Rijkl = Wijkl
+ (Rik δjl − Ril δjk + Rjl δik − Rjk δil )/(n − 2) − S(δik δjl − δil δjk )/(n − 1)(n − 2)
The reason for the terminology will be clear in Chapter III, but some motivation
will be provided already by the following
Exercise 2.3.2. Calculate the Riemann curvature tensor of a sphere of radius R
in Rn+1 and decompose it according to (2.3.7).
We have proved that O(n) acts on T ⊖W with just two invariant subspaces, displayed
in the decomposition (2.3.7), but the proof that W has no O(n) invariant subspace
requires a bit of invariant theory. We must therefore refer the reader to the proof given
in Berger, Gauduchon and Mazet [1, pp. 76–78].
Exercise 2.3.3. Prove that
X X
Rijkl Riklj = − 12 Rijkl 2 .
i,j,k,l i,j,k,l
When n = 3 then n(n + 1)/2 = 6 = n2 (n2 − 1)/12 so there is no Weyl tensor. When
n > 4 it is known that W has no SO(n) invariant subspace (see references in Besse
[1, p. 49]). However, this is not true when n = 4, and since further decomposition of
W is then relevant in Yang-Mills theory, we shall discuss it briefly. (For details and
references we refer to Atiyah, Hitchin and Singer [1], Besse [1].) Since the 4-linear
forms R with the symmetries (2.3.2) can be identified with symmetric bilinear forms
on ∧2 (Rn ), we shall first discuss how SO(n) acts on ∧2 (Rn ).
24 II. CURVATURE OF SUBMANIFOLDS OF A EUCLIDEAN SPACE
Lemma 2.3.3. ∧2 (Rn ) has no proper invariant subspace under the action of SO(n)
if n 6= 4, but ∧2 (R4 ) = ∧+ ⊕ ∧− where ∧± are irreducible SO(4) invariant subspaces
of dimension 3.
Proof. If S is a skew symmetric n × n matrix, then Rn = V0 ⊕ V1 ⊕ · · · ⊕ Vj where
SV0 = 0 and Vi is of dimension 2 and SVi = Vi when i = 1, . . . , j. (See Exercise 1.4.2.)
This means that every element in ∧2 (Rn ) can be transformed by the action of SO(n)
to the form
Since e1 − e3 + e3 − e5 + e5 − e1 = 0, the rank of this two vector is just 4 and the lemma
follows also in this case.
We are now just left with the case n = 4, and then we know that ∧ must contain
either e1 ∧ e2 + e3 ∧ e4 or e1 ∧ e2 − e3 ∧ e4 = e1 ∧ e2 + e4 ∧ e3 . In the first case
∧ must contain the space ∧+ spanned by ei ∧ ej + ek ∧ el where i, j, k, l is an even
permutation of 1, 2, 3, 4, and in the other case ∧ contains the space ∧− defined using
the odd permutations. If ϕ± ∈ ∧± then
as basis in ∧± . With this basis the SO(4) action on ∧2 (R4 ) gives a homomorphism
SO(4) → SO(3)×SO(3). Show that the corresponding homomorphism of Lie algebras
so(4) → so(3) ⊕ so(3) is an isomorphism and determine it explicitly in terms of the
corresponding 4 × 4 and 3 × 3 skew symmetric matrices.
Remark. The decomposition when n = 4 will perhaps be more evident when we
have introduced the ∗ operator on forms later on.
ALGEBRAIC PROPERTIES OF THE CURVATURE TENSOR 25
with spaces of dimensions 6, 6 and 9. However, since we have an SO(3) action on the
first two spaces they decompose further into the multiples I(∧± ) of the metric form
and the tracless forms S02 (∧± ), so we have in fact an SO(4) invariant decomposition
where the spaces now have dimensions 5, 1, 5, 1, 9. Since the space of Weyl tensors
W has dimension 10, and the space of traceless Ricci tensors has dimension 9, it is
now clear that Hom(∧+ , ∧− ) is precisely the space of traceless Ricci tensors while
W = W+ ⊕ W− with W± = S02 (∧± ). The purely antisymmetric tensor and the
pure scalar curvature part S(δik δjl − δil δjk )/12 are found in the two invariant one
dimensional spaces; the antisymmetric tensors are generated by the differences of the
two metric tensors whereas the scalar curvature term corresponds to their sum. Thus
a general curvature tensor in T is represented in block matrix form by
W+ + SI/12 B
t
B W− + SI/12
where B corresponds to the traceless Ricci tensor and the traceless symmetric W±
together define the Weyl tensor. We leave the details of the verification for the energetic
reader.
The SO(n) invariant classification of the curvature tensors gives rise to important
classes of Riemannian manifolds:
Definition 2.3.4. A Riemannian manifold is called an Einstein manifold if the
traceless Ricci tensor vanishes, and it is called conformally flat if the Weyl tensor
vanishes. An oriented manifold of dimension 4 is called self-dual if the Weyl tensor is
in W+ and anti-self-dual if it is in W− .
26 II. CURVATURE OF SUBMANIFOLDS OF A EUCLIDEAN SPACE
CHAPTER III
Summary. In Section 3.1 we extend the intrinsic results proved in Chapter II to abstract
Riemannian manifolds, taking the notion of covariant differentiation as the main tool.
After introducing geodesic coordinates we also discuss various geometrical interpretations
of curvature and prove the classical Gauss-Bonnet theorem. Section 3.2 is devoted to
embedding theorems which show that abstract Riemannian manifolds are not really more
general than the submanifolds of a Euclidean space studied in Chapter II. In Section
3.3 we show that a Riemannian manifold with vanishing curvature tensor is flat, that
is, locally isomorphic to Rn ; more generally we also discuss manifolds with constant
curvature. Section 3.4 is devoted to the transformation rules for the curvature under
conformal changes of metric. For dimensions ≥ 4 it is proved that vanishing of the Weyl
tensor is necessary and sufficient for the existence of a flat conformal metric. The Yamabe
problem to find conformal metrics of constant scalar curvature is also mentioned briefly,
but we leave the study of it for another chapter.
is defined in the open set κ−1 (κ(ω) ∩ κ̃(ω̃)) ⊂ ω which is in C k since κ̃ has a C k left
inverse.
An abstract C k manifold M of dimension n is by definition a Hausdorff topological
space provided with a family of homeomorphisms, κ : ω → f (ω) ⊂ M with open range,
such that κ̃−1 ◦ κ ∈ C k (κ−1 (κ(ω) ∩ κ̃(ω̃)) if κ̃ is another member of the family. We shall
always assume that M has a countable dense subset. Then it is well known that M
can be embedded as a submanifold of R2n+1 , so the difference between submanifolds
of RN and abstract manifolds is more one of principle and attitude than of substance.
The tangent bundle T (M ) was defined for a submanifold of RN as the set of all
(x, t) with x ∈ M and t in the tangent space Tx (M ) of M in RN . It has a natural
generalization for an abstract manifold. One just takes Tx (M ) to be the n dimensional
vector space of first order differential operators at x which annihilate constants. In a
n
local coordinate
Pn patchj ω ⊂ R with local coordinates x, the first order operators can
n
be written 1 tj ∂/∂x , which identifies T (M ) with ω × R over ω. (From now on we
shall use superscripts for the coordinates to conform with the conventions of tensor
calculus.) If one switches to coordinates x̃ as above, then (x, t) is identified with (x̃, t̃)
if κ(x) = κ̃(x̃) and (κ̃−1 κ)′ t = t̃, which means that t̃ = (∂ x̃/∂x)t.
27
28 III. ABSTRACT RIEMANNIAN MANIFOLDS
which means that det(gij (x)) dx, where dx is Lebesgue measure in Rn , is an invariant
p
(2.1.11) suggests that another vectorfield, the covariant derivative of Y in the direction
X, should be invariantly defined by
n
X n
X
(3.1.2) ∇X Y = (XY j + Γik j X k Y i )∂/∂xj ,
j=1 i,k=1
COVARIANT DERIVATIVES AND CURVATURE 29
where Γik j is defined in terms of the metric by the last expression in (2.1.8), (2.1.7).
We could verify by direct computation that this definition is indeed independent of the
choice of local coordinates, but instead we shall prove that in a local coordinate patch
the linear differential operator defined by (3.1.2) has the properties
(3.1.3) ∇X Y − ∇Y X = [X, Y ],
(3.1.4) (∇Z X, Y ) + (X, ∇Z Y ) = Z(X, Y ),
where Z is a third vector field. From (3.1.3) it follows that ∇X Y is linear in X as well
as in Y . We shall also prove that ∇X is uniquely determined by (3.1.3), (3.1.4), which
implies the invariance. The proof of (3.1.3), where [X, Y ] = XY − Y X denotes the
j
commutator of the vector fields, follows immediately from the fact that P Γjik , defined
by (2.1.7), (2.1.8), is symmetric in i, k. To prove (3.1.4) we write Z = Z ∂/∂xj and
obtain
n
X n
X
j
(∇Z X, Y ) = (ZX + Γik j Z k X i )gjl Y l
j,l=1 i,k=1
n
X n
X
j l
= gjl (ZX )Y + Z k Γikl X i Y l .
j,l=1 i,k,l=1
If we exchange X and Y and add using (2.1.7)′ , it follows that (3.1.4) holds. To prove
the uniqueness means proving that if the right hand sides of (3.1.3), (3.1.4) are replaced
by 0, then ∇X Y must be 0 for arbitrary X and Y . This follows since
where we have alternated using the homogeneous forms of (3.1.3), (3.1.4). This implies
∇X Y = 0 as claimed. Thus we have proved:
Theorem 3.1.2. For smooth vector fields X, Y on M the vector field defined in
local coordinates by (3.1.2) is invariantly defined and the covariant differentiation ∇ is
characterized by (3.1.3), (3.1.4).
Note that ∇X Y (x) is defined even if Y is only defined on a curve with tangent X at
x. Taking for X and Y the unit tangent of a curve we can define the geodesic curvature
and geodesic principal normal as the norm and the direction of ∇X Y . This generalizes
the definition given after Theorem 2.1.3 for submanifolds of RN .
In the Euclidean case the differential operators ∇X and ∇Y act componentwise as
the scalar operators X and Y , so their commutator is ∇[X,Y ] . We shall now examine
to what extent this is true in the Riemannian case, so we form with three vector fields
X, Y, Z
S = (∇X ∇Y − ∇Y ∇X − ∇[X,Y ] )Z.
30 III. ABSTRACT RIEMANNIAN MANIFOLDS
the other terms −(Y ϕ)∇X Z + (Y ϕ)∇X Z will cancel. Thus S contains no derivatives
of X, hence no derivatives of Y by the skew symmetry. This is also true for Z since
We can therefore calculate S using local coordinates and assuming that the components
of X, Y, Z are constants, hence [X, Y ] = 0. Then
X X X
∇ X ∇Y Z = ∇X Γjl i Y l Z j ∂/∂xi = ∂k Γjl i + Γkν i Γjl ν X k Y l Z j ∂/∂xi .
i,j,l ijkl ν
If we subtract the analogous formula for ∇Y ∇X Z and recall the first definition (2.1.13)
of the Riemann curvature tensor, it follows that
X
(3.1.6) (∇X ∇Y − ∇Y ∇X − ∇[X,Y ] )Z = Ri jkl X k Y l Z j ∂/∂xi .
ijkl
In particular, this proves that the definition of the Riemann curvature tensor given
by (2.1.13) is coordinate independent and does define a tensor of type 1, 3. (2.1.13)′
follows as before.
If [0, a] ∋ t 7→ γ(t) ∈ M is a (piecewise) C 1 curve and X0 ∈ Tγ(0) M , then there is
a unique vector field X(t) ∈ Tγ(t) M , t ∈ [0, a], along the curve with X(0) = X0 and
∇γ ′ (t) X(t) = 0. In fact, in local coordinates this is the Cauchy problem for a linear
system of differential equations
dX j (t) X dγ k (t) i
+ Γik j (γ(t)) X (t) = 0, j = 1, . . . , n.
dt dt
i,k
e0 (ε) − X0 )/ε2 if X
Use this to find the limit as ε → 0 of (X e0 is the parallel translation
of X0 around the square from x to x + εek to x + ε(ek + el ) to x + εel to x.
COVARIANT DERIVATIVES AND CURVATURE 31
∇X ω = (∇X ω ♯ )♭ .
X(ω ♯ , Y ) = (∇X ω ♯ , Y ) + (ω ♯ , ∇X Y ),
This could also have been taken as a definition of ∇X ω. In local coordinates (3.1.7)
means that
n
X n
X n
X
h∇X ω, Y i = X ωj Y j − ωj (XY j + Γkl j X k Y l )
j=1 j=1 k,l=1
n
X n
X
j
= (Xωj )Y − ωl Γlkj X k Y j
j=1 j,k,l=1
so we obtain
n
X
(3.1.2) ′
(∇X ω)j = Xωj − Γkj l X k ωl .
k,l=1
Exercise 3.1.3. Show that for tensor fields f of type k, l, that is, sections of
T (M ) ⊗ · · · ⊗ T (M ) ⊗ T ∗ (M ) ⊗ · · · ⊗ T ∗ (M )
| {z } | {z }
k times l times
32 III. ABSTRACT RIEMANNIAN MANIFOLDS
∇X (Y1 ⊗ · · · ⊗ Yk ⊗ ω1 ⊗ · · · ⊗ ωl )
= (∇X Y1 ) ⊗ Y2 ⊗ · · · ⊗ ωl + · · · + Y1 ⊗ · · · ⊗ ωl−1 ⊗ ∇X ωl ,
where Y1 , . . . , Yk are vector fields and ω1 , . . . , ωl are one forms. Prove that the product
rule also holds for arbitrary tensor products.
To round off the definition of ∇X , we define ∇X f = Xf if f is just a function on M .
Recall that contraction of a tensor means in local coordinates that one of the upper
indices is put equal to one of the lower indices followed by summation over that index.
In particular, the contraction of Y ⊗ ω where Y is a vector and ω a one form is the
scalar product hω, Y i discussed above. For decomposable tensors as in the preceding
exercise, the contraction also means precisely taking scalar product of one of the vector
factors and one of the one form factors. Since (3.1.7) shows that for Y ⊗ ω it does
not matter if one first applies ∇X and then contracts or if one first contracts and then
applies ∇X , it follows that covariant differentiation commutes with contraction.
Let f again by a tensor field of type k, l. Since ∇X f at every point is just a linear
function of X, not depending on its derivatives, we can regard ∇X f as the contraction
of X and a tensor field ∇f of type k, l + 1,
k X
n
i ...i µiν+1 ...ik
X
(3.1.2)′′′ (∇f )ij11...i i1 ...ik
...jl ,j = ∂j fj1 ...jl +
k
Γµj iν fj11...jlν−1
ν=1 µ=1
l X
X n
− Γjν j µ fji11...j
...ik
ν−1 µjν+1 ...jl
ν=1 µ=1
Note the notation , j for the added index. Also note that ∇X f is defined at x even if f
is only defined on a curve passing through x with tangent X. A tensor f defined along
a curve t 7→ x(t) ∈ M can therefore be differentiated with respect to t as ∇x′ (t) f .
If X andP Y are two vector fields and g is the metric tensor, of type 0, 2, then
i j
(X, Y ) = gij X Y is the contraction of g ⊗ X ⊗ Y with respect to both pairs of
indices. Thus ∇(X, Y ) is the contraction of
On the other hand, by (3.1.4) it is the contraction of the last two terms only. Hence it
follows that
(3.1.8) ∇g = 0.
Exercise 3.1.4. Prove (3.1.8) directly from the definitions. Show also that ∇g ♯♯ = 0.
Prove that if X, Y, Z are vector fields then ∇X ∇Y Z is the sum of the contraction of
(∇2 Z) ⊗ Y ⊗ X and that of (∇Z) ⊗ ∇X Y . Deduce that in local coordinates
X
Z i ,kj − Z i ,jk = Ri ljk Z l .
l
Exercise 3.1.5. Let f be a tensor field of type k, l, and compute the antisymmetric
part of ∇∇f , that is, in local coordinates,
fji11...j
...ik
l ,rs
− fji11...j
...ik
l ,sr
.
with equality if and only if ds/dt is constant, that is, the parameter t is proportional
to the arc length. This can always be achieved, so the minimum problem is equivalent
to minimizing
Z 1 n
X
(3.1.9) I= gjk (x)dxj /dtdxk /dt dt
0 j,k=1
for all smooth curves from x0 to x1 . This has the advantage that one can expect a
unique solution if the points are sufficiently close. If the minimum is attained, then
the Euler equations
n n
d X dxk X dxi dxk
(3.1.10) 2 gjk (x) = ∂j gik (x) , j = 1, . . . , n,
dt dt dt dt
k=1 i,k=1
must be valid. (Replace x(t) by x(t) + εy(t) where y(0) = y(1) = 0, put the derivative
with respect to ε equal to 0 when ε = 0, and integrate by parts.) Carrying out the
differentiation in the left-hand side we obtain the equations
n n
X d 2 xk X dxi dxk
2 gjk (x) 2 + (2∂i gjk (x) − ∂j gik (x)) = 0, j = 1, . . . , n.
dt dt dt
k=1 i,k=1
These equations mean that the covariant derivative of the tangent vector dx/dt along
the curve is equal to 0, that is, that the geodesic curvature vanishes (cf. Theorem
COVARIANT DERIVATIVES AND CURVATURE 35
∂ X X
(3.1.10)′′ 2 Gjk (tX)X k = (∂j Gik )(tX)X iX k .
∂t
Conversely, if we have a coordinate system such that (3.1.11) is valid, then reversing the
preceding argument shows that the straight lines through the origin (with respect to
the parameters) are geodesics. We could make a linear change of coordinates to make
Gjk (0) equal to the identity matrix. However, with pseudo-Riemannian geometry in
view, where gjk will be non-singular but not positive definite, we prefer to leave Gjk (0)
arbitrary. The coordinate systems now obtained are called geodesic coordinates; they
are uniquely determined up to a linear transformation (up to an orthogonal one if one
has insisted on the Euclidean normal form).
36 III. ABSTRACT RIEMANNIAN MANIFOLDS
n
X 21
j k
Gjk (X)dX dX
j,k=1
n
X n
X 12 n
X 12
j k j k
≥ Gjk (0)X dX / Gjk (0)X X =d Gjk (0)X j X k ,
j,k=1 j,k=1 j,k=1
and this implies that when X is in the domain of the geodesic coordinates, then the
shortest path to the origin is indeed the geodesic ray. Thus we have proved that
geodesics do give the shortest path between any two of its points which are sufficiently
close. (The importance of the latter restriction was already clear for the sphere.)
So far in this chapter we have only discussed local properties of Riemannian man-
ifolds. A Riemannian manifold is a metric space with the distance s(x, y) between
x, y ∈ M defined as the infimum of the lengths of differentiable curves from x to y.
We add a theorem of a global nature:
Theorem 3.1.4 (Hopf-Rinow). Let M be a connected Riemannian manifold. Then
the following properties are equivalent:
(1) M is a complete metric space.
(2) Every geodesic in M can be extended indefinitely in both directions.
(3) There exists a point x ∈ M such that all geodesics starting at x can be extended
indefinitely in both directions.
(4) Every closed bounded subset of M is compact.
They imply that any two points x, y ∈ M can be joined by a geodesic of length s(x, y).
Proof. (1) =⇒ (2): Let R ⊃ (a, b) ∋ s 7→ x(s) be a geodesic with maximal interval
of definition and the arc length as parameter. Since s(x(s1 ), x(s2 )) ≤ |s1 − s2 |, the
sequence x(b − 1/k) is a Cauchy sequence if b < ∞. By the local existence theorem
for ordinary differential equations applied in a neighborhood of the limit, the geodesic
with initial data x(b − 1/k), x′ (b − 1/k) can be extended for an interval δ independent
of k for large k, which shows that (a, b) is not a maximal interval of definition and
proves (2).
The implications (4) =⇒ (1) and (2) =⇒ (3) are trivial. Assume now that (3) is
valid. This means that we have a globally defined geodesic exponential map γ : Tx M ∋
v 7→ X(1) ∈ M , giving the value at time 1 of the geodesic with initial data x, v. Since
γ is continuous and B(r) = {v ∈ Tx M ; |v|x ≤ r} is compact, it follows that γB(r) is
compact. We know already that γB(r) is equal to B(r) b = {y ∈ M ; s(x, y) ≤ r} for
small r, and it is obvious that γB(r) ⊂ B(r) for every r > 0. If we prove that equality
b
holds for every r, then (4) will be proved.
Let r1 be the supremum of all r such that γB(̺) = B(̺)b when ̺ < r. If r1 = ∞ our
claim is true so assume that r1 < ∞. Then γB(r1 ) = B(r1 ), for B(r
b b 1 ) is the closure of
∪r<r1 B(r
b 1 ) = γ ∪r<r B(r) ⊂ γB(r1 ),
1
COVARIANT DERIVATIVES AND CURVATURE 37
and γB(r1 ) is compact and therefore closed. Now we claim that γB(r) = B(r) b for
r < r1 + δ if δ is small enough. If y ∈ B(r) \ B(r1 ), we can for large ν find a path
b b
of length < s(x, y) + 1/ν from x to y. If zν is the last point in B(r b 1 ) = γB(r1 ), then
s(x, zν ) = r1 and s(zν , y) + r1 ≤ s(x, y) + 1/ν. If z is a limit point of the sequence zν
in the compact set γB(r1 ), it follows that s(z, y) ≤ s(x, y) − r1 < δ. In view of the
compactness of γB(r) we can choose δ independently of z so small that this implies
that there is a geodesic from z to y of length s(x, y) − r1 . We also have a geodesic of
length r1 from x to y. If these did not fit together to one geodesic, we could get a path
shorter than s(x, y) from x to y by smoothing out the broken geodesic near z, which
is a contradiction completing the proof of Theorem 3.1.4.
Variations of geodesics are controlled by the Jacobi differential equations:
Theorem 3.1.5. Let γ ∈ C 3 (ω, M ) where ω is an open set in R2 , and assume that
t 7→ γ(t, s) is for fixed s a geodesic with parameter proportional to the arc length, when
(t, s) ∈ ω. Let T = γ∗ ∂/∂t be the tangent of the geodesic and let X = γ∗ ∂/∂s be a vector
field along the geodesic describing the direction in which it moves. Then X satisfies
the Jacobi differential equation ∇T ∇T X = R(T, X)T , where in local coordinates
X
R(T, X)Z = Ri jkl Z j T k X l ∂/∂xi .
Proof. If γ ′ has rank 2 then the range is locally a two dimensional surface where the
vector fields T and X are defined and commute, since ∂/∂t and ∂/∂s commute. In
this surface we have ∇T T = 0 for t 7→ γ(t, s) is a geodesic. Since [T, X] = 0 we have
∇T X − ∇X T = 0, so (3.1.6) gives
which proves the Jacobi differential equation in the closure of the open subset of ω
where γ ′ has rank 2. In the complement in ω we have locally γ(t, s) = γ0 (ψ(t, s))
where γ0 (t) = γ(t, s0) for a suitable s0 . The geodesic equation means that ψ(t, s) is
linear in in t for fixed s. Hence X = aT where a = (∂ψ/∂s)/(∂ψ/∂t) is linear in t, and
Hint: This means precisely that the rays t 7→ (tx1 , . . . , txν , xν+1 , . . . , xn ) are
geodesics orthogonal to M0 . (Note that when ν = 1 the condition means that g1k (x) =
δ1k .)
38 III. ABSTRACT RIEMANNIAN MANIFOLDS
(Hint: Use Exercise 3.1.6 and make the vector fields ∂/∂xj parallel along M0 .) Con-
clude that if the differential equations
G = G0 + G1 + G2 + . . .
must therefore change the sign although we get back the same elements, which proves
the claim.
The first interesting term is therefore G2 . We shall write
We have seen in Theorem 2.3.1 that this is precisely the dimension of the space T of
curvature tensors. We can polarize G2 to a 4-linear form G2 (X1 , X2 ; Y1 , Y2 ) which is
symmetric in X1 , X2 as well as in Y1 , Y2 and has the property
G2 (X; Y ) = G2 (X, X; Y, Y ).
From (3.1.12)′ we easily obtain G2 (X, X; Y, Y ) = −2G2 (X, Y ; X, Y ), hence the sym-
metry
Since G′jk (0) = 0, the corresponding Christoffel symbols vanish at the origin. (Note
that the Christoffel symbols are not tensors, for otherwise we could not make them
equal to 0 at a point by changing coordinates. The fact that we have raised and
lowered an index in the same way as for tensors might suggest otherwise, but it is just
for linear transformations that they behave as tensors.) Thus the formulas (2.1.13)′
for the curvature tensor simplify at 0 to
after cancellation of two terms ∂j ∂k Gil . This means that for the corresponding 4-linear
forms we have
Here we have only used so far that G1 = 0. For geodesic coordinates we have the
symmetry (3.1.13), which simplifies (3.1.14) to
We have a one to one correspondence between the two 4-linear forms G2 (X1 , . . . , X4 )
and R(X1 , . . . , X4 ) with the properties:
symmetry for G2 when X1 ↔ X2 or X3 ↔ X4 or (X1 , X2 ) ↔ (X3 , X4 ); for R when
(X1 , X2 ) ↔ (X3 , X4 );
antisymmetry for R when X1 ↔ X2 or X3 ↔ X4 ;
40 III. ABSTRACT RIEMANNIAN MANIFOLDS
The preceding observations were expressed in words by Riemann [1, p. 279]: “Führt
man diese GrössenPein, so wird für unendlich kleine Werthe von x das Quadrat des
Linienelements = dx2 , das Glied der Nächsten Ordnung in demselben aber gleich
einem homogenen Ausdruck zweiten Grades der n(n − 1)/2 Grössen (x1 dx2 − x2 dx1 ),
(x1 dx3 − x3 dx1 ), . . . , also eine unendlich kleine Grösse der vierten Dimension, so
dass man eine endliche Grösse erhält wenn man sie durch das Quadrat des unendlich
kleinen Dreiecks dividirt, in dessen Eckpunkten die Werthe der Veränderlichen sind
(0, 0, 0, . . . ), (x1 , x2 , x3 , . . . ), (dx1 , dx2 , dx3 , . . . ). Diese Grösse behält denselben Werth,
so lange die Grössen x und dx in denselben binären Linearformen enthalten sind, oder
so lange die beiden kürzesten Linien von den Werthen 0 bis zu den Werthen x und
von den Werthen 0 bis zu den Werthen dx in demselben Flächenelement bleiben, und
hängt also nur von Ort und Richtung desselben ab. Sie wird offenbar = 0, wenn
die
P dargestellte Manningfaltigkeit eben, d.h. das Quadrat des Linienelements auf
2
dx reducirbar ist, und kann daher als das Mass der in diesem Punkte in dieser
Flächenrichtung stattfindenden Abweichung der Mannigfaltigkeit von der Ebenheit
angesehen werden. Multiplicirt mit −3/4 wird sie der Grösse gleich, welche Herr
Geheimer Hofrath Gauss das Krümmungsmass einer Fläche genannt hat.”
The factor 4 here comes from the fact that Riemann divided by the square of the
area of a triangle and not the corresponding parallelogram. The factor −3 is the same
as in (3.1.15) above, and we shall now show that it is also closely connected to the
denominator 3 in Legendre’s theorem.
Still with geodesic coordinates we shall consider the geodesic triangle with corners
at 0, εY ,P εZ when ε is small. For the sides from 0 the lengths are ε|Y | and ε|Z| where
|X|2 = Gjk (0)X j X k ; we denote the corresponding scalar product by (·, ·). The
third side is not as easy to determine since we do not know the geodesic. However, it
is clear that the square of its length is equal to ε2 times the square of the length of the
geodesic from Y to Z for the metric
X
Gjk (εX)dX j dX k = |dX|2 + ε2 G2 (X; dX) + O(ε3 ).
COVARIANT DERIVATIVES AND CURVATURE 41
This geodesic must differ from the straight line segment [0, 1] ∋ t 7→ Y + t(Z − Y ) by
O(ε2 ), and since the straight line is a geodesic for the metric |dX|2 , we obtain for the
geodesic distance εa between εY and εZ
Z 1
2 2 2
a = |Y − Z| + ε G2 (Y + t(Z − Y ); Z − Y ) dt + O(ε3 )
0
= |Y − Z|2 + ε2 G2 (Y ; Z) + O(ε3 ),
Since the first factor is −1/3 times the sectional curvature in the Y Z plane (cf. (3.1.15)
and Theorem 2.2.3), and the second factor is the area of the geodesic triangle +O(ε3 ),
we have proved Legendre’s theorem. Note that the total angle excess is equal to the
sectional curvature times the area +O(ε3 ).
Exercise 3.1.11. Assuming that the earth is a sphere with circumference 40000 km,
and that T is an equilateral geodesic triangle on the earth with the base equal to 100
km and angles 60◦ at the base, estimate how many seconds of arc the third angle differs
from 60◦ .
In the two dimensional case considered by Gauss we can make a subdivision of the
geodesic triangle T by geodesics joining the midpoints of each side. Repeating this
subdivision and noting that angle excess is additive when we subdivide, we conclude
in the limit that
ZZ
(3.1.16) α+β+γ = π+ K dS
T
where α, β, γ are the angles at the corners, K is the Gauss curvature, and dS is the
Riemannian area measure. This is Gauss’ part of the Gauss-Bonnet theorem; it was
extended by Bonnet to more general regions. This we shall now do with a different
more analytical proof which will introduce some ideas which will be important later
on.
Let us assume that we have a Riemannian metric in a simply connected open set
ω ⊂ R2 . We shall generalize Theorem 1.2.1 by calculating the integral of the signed
42 III. ABSTRACT RIEMANNIAN MANIFOLDS
2
X dxj (s) dxk (s)
gjk (x(s)) = 1,
ds ds
j,k=1
and we denote by e(s) the unit tangent vector e(s) = x′ (s). The covariant derivative
of e along the curve has the components
2
j dej X dxk
(∇e e) = + Γik j (x(s))ei (s) ,
ds ds
i,k=1
but we shall avoid explicit calculations using this expression. Let n be the unit vector
orthogonal to e such that e, n is positively oriented. Then the signed geodesic curvature
κg is given by
κg = (∇e e, n),
and our task is to calculate the integral
Z L
(3.1.17) κg ds.
0
2
X 2
X
(3.1.18) κ= gjl (dej + Γik j (x)ei dxk )nl
j,l=1 k,i=1
the integral being taken along the curve γ : s 7→ (x(s), e(s)). (Recall that n is uniquely
determined by e.) Here e(s) = x′ (s), but we shall simplify the problem by generalizing
it, so we allow any closed C 1 curve
R γ in S(T ω) now.
We shall first examine how γ κ depends on the choice of the unit vector e. Any
other choice ẽ can be written
Hence
(∇ẽ, ñ) = (∇e, n) + dθ,
R R RL
so κ = γ κ + 0 dθ if γ̃ is the curve s 7→ (x(s), ẽ(s)).
γ̃
If v and ṽ are two arbitrary vector fields with no zeros defined along the curve
s 7→ x(s), then we can normalize them with respect to the Riemannian metric to unit
RL
vector fields e and ẽ and introduce the variation 0 dθ of the angle from e to ẽ along the
curve. This is an integer multiple of 2π which must be independent of the Riemannian
metric. In fact, if we replace gjk by λgjk + (1 − λ)δjk , 0 ≤ λ ≤ 1, then the angle
variation depends continuously on λ so it must have the same value when λ = 0 as
when λ = 1. If for example one of the vector fields is tangent to the curve and the
other is ∂/∂x1 , then it follows from Theorem 1.2.1 that the angle variation between
them is ±2π. R
We shall now determine the integral γ κ when e is a unit vector field defined in the
whole of ω, for example the normalization of the vector field ∂/∂x1 . The advantage
of this is that we can then pull the form κ back by the R map x 7→ (x, e(x)) to a form
in ω with integral over the curve s 7→ x(s) equal to γ κ. (It also follows that γ is a
boundary in S(T ω).) The interior ω ′ of the curve is ⋐ ω since ω was assumed simply
connected, and we shall calculate the integral of the differential form over ∂ω ′ using
Stokes’ formula. If εi = ∂/∂xi and we write ∇i = ∇εi for the sake of brevity, the
differential form to integrate is
2
X
(∇i e, n) dxi .
i=1
(3.1.19) ∇1 (∇2 e, n) −∇2 (∇1 e, n) = ((∇1 ∇2 −∇2 ∇1 )e, n) +(∇2 e, ∇1 n) −(∇1 e, ∇2 n).
Here we have used (3.1.4). The last two terms are equal to 0 since ∇j e has the direction
of n and ∇k n has the direction of e. By (3.1.6) the first term, on the right is equal
to R(n, e, ε1 , ε2 ) = −R(e, n, ε1 , ε2 ). The vectors e, n are positively oriented and span
√
a parallelogram with area 1 while ε1 , ε2 span one of area g where g = det(gjk ). In
√
view of Theorem 2.2.3 we therefore conclude that (3.1.19) is equal to −K g, where
K is the total (Gaussian) curvature, and we have proved:
Theorem 3.1.6 (Gauss-Bonnet). Let ω ⊂ R2 be a simply connected open set with
Riemannian metric, and let ω ′ ⋐ ω be simply connected with C 2 boundary. Then
Z Z
(3.1.20) κg ds + K dS = 2π,
∂ω ′ ω′
44 III. ABSTRACT RIEMANNIAN MANIFOLDS
where K is the total curvature, κg is the geodesic curvature of ∂ω ′ with the orientation
making ω ′ lie to the left of ∂ω ′ , ds is the arc length of ∂ω ′ , and dS is the Riemannian
√
area element g dx in ω ′ .
In the proof we have actually proved
√
(3.1.21) dκ = π ∗ (−K gdx1 ∧ dx2 )
where αj denote the exterior angles at the corners and κg is integrated only over the
smooth part of ∂ω ′ . For a geodesic triangle, the three angles are π − αj , so the sum is
X Z
3π − αj = π + K dS.
ω′
Thus (3.1.20)′ contains (3.1.16). It is also easy to obtain (3.1.20) from (3.1.16) applied
to a triangulation of a polygonal approximation of ω ′ . We leave this also as an exercise.
We shall now discuss the case of a compact oriented Riemannian manifold M of
dimension 2 (without boundary). Suppose that M is decomposed by geodesic arcs
into a finite number ν2 of geodesic polygons ωj . Denote the interior angles of ωj by
βjk . Then (3.1.20)′ yields
X Z
(π − βjk ) + K dS = 2π.
k ωj
The number of terms in the sum is equal to the number of sides of ωj . If ν0 and ν1
denote the total number of corners and sides occurring in some ωj , we get by adding
Z
K dS + 2πν1 − 2πν0 = 2πν2 ,
M
COVARIANT DERIVATIVES AND CURVATURE 45
for there will be altogether 2ν1 terms π in the left-hand side. Thus
1
Z
(3.1.22) K dS = ν2 − ν1 + ν0 ,
2π M
This is a manifold whose boundary consists of the circles over the zeros of F , that
is, the critical points of f . For at such a point we can introduce geodesic coordinates
diagonalizing the quadratic terms in f , that is,
gjk (x) = δjk + O(|x|2 ), f (x) = f (0) + (f1 (x1 )2 + f2 (x2 )2 )/2 + O(|x|3 ).
When x winds around a small circle |x| = ε then F (x)/kF (x)k winds around the unit
circle in the same or opposite direction depending on the sign of f1 f2 . The boundary
of Γ is the limit of the image of the negatively oriented circle of radius ε as ε → 0, so
it consists of the circles in S(T M ) with orientation opposite to the sign of the Hessian
of f at the critical point. If we integrate (3.1.21) over Γ it follows now from Stokes’
formula that Z Z Z X
− K dS = dκ = κ = −2π εj
Γ ∂Γ
where εj is the sign of the Hessian of f at the critical point. Thus we have
Theorem 3.1.7. If M is a compact two dimensional
R oriented Riemannian manifold
with total curvature K and no boundary, then M K dS/2π is for every real valued
function on M with only non-degenerate critical points equal to the sum of the signs
of the Hessian of f at the critical points. The integral is also equal to the Euler
characteristic.
Instead of the vector field F = (df )♯ we could have used here any vector field with
only non-degenerate fixed points (that is, zeros). We shall later on give an extension
due to Chern of the preceding arguments which is applicable to any oriented manifold of
even dimension. The problem is to find the appropriate differential forms in S(T M ).
To do so we need a systematic approach to Riemannian geometry using differential
forms; the proof of Theorem 3.1.6 was meant to motivate the need for that.
R
We shall now supplement Theorem 3.1.7 by studying the integral M K dS for a
compact oriented hypersurface M ⊂ Rn+1 of any dimension n. We denote by γ the
46 III. ABSTRACT RIEMANNIAN MANIFOLDS
Gauss map (2.2.3); the direction of the normal n is chosen so that a positive system of
tangent vectors followed by n is a positive system in Rn+1 . We choose the orientation
of S n so that with this definition the normal at x ∈ S n is −x. Set γ̌ = −γ, which is
then the identity map if M = S n . The degree D of the map γ̌ is then defined by
Z Z
∗
γ̌ u = D u
M Sn
where dS denotes the area element in M and that in S n in the two integrals.
When n = 2 it follows in view of Theorem 3.1.7 that the degree of the mapping −γ
is equal to half the Euler characteristic. Later on we shall extend this to arbitrary n
such that (n − 2)/4 is an integer.
We shall finally prove an important result on the covariant derivative of the curvature
tensor. For a geodesic system of coordinates the Christoffel symbols are O(|x|), so the
non-linear terms in (2.1.13)′ are O(|x|2 ). Hence we obtain
at the center of the geodesic coordinates. Note that the first and the last term are
equal apart from the sign and a circular permutation of the indices klm, and that this
is also true for the middle terms. Hence we obtain the second Bianchi identity
or more briefly Rij[kl,m] = 0. Since Rijkl,m is a tensor, this is of course true for any
system of coordinates. Contraction of (3.1.24) with respect to the indices ik and the
indices jl, that is, multiplication by g ik g jl and summation gives, since contraction and
the musical isomorphisms commute with covariant differentiation
X X
S,m − g ik Rim,k − g jl Rjm,l = 0,
LOCAL ISOMETRIC EMBEDDING 47
that is,
X
(3.1.25) ∂k S = 2 g ij Rik,j .
Altogether there are n(n+1)/2 equations, so it follows from general theorems discussed
at the end of this section that smooth solutions do not exist in general unless N ≥ n(n+
1)/2. We shall now discuss a classical theorem of Janet and Cartan (see Jacobowitz
[1] and references there) which shows that a local real analytic solution always exists
when N = n(n + 1)/2 if gjk are real analytic. The idea of the proof is to argue by
induction with respect to n, and extend a local isometric embedding of Rn−1 × {0} by
solving a Cauchy problem. To do so one must cut down the number of equations to
solve:
Lemma 3.2.1. The equations (3.2.1) are valid in a ball with center at 0 if and only
if they hold for 1 ≤ j ≤ k < n when xn = 0 and in addition we have the equations
(∂n2 f , ∂j f ) = Γnnj , 1 ≤ j ≤ n,
(3.2.2)
(∂n2 f , ∂i ∂j f ) = ∂n Γijn + (∂i ∂n f, ∂j ∂n f ) − 21 ∂i ∂j gnn , 1 ≤ i ≤ j < n,
Proof. The first n equations (3.2.2) and the second set of initial conditions (3.2.3)
follow from the definition (2.1.7) of the Christoffel symbols. To prove the second set
of equations (3.2.2) we note that
so using the first boundary condition (3.2.3) with j = n we obtain (∂n f, ∂n f ) = gnn
also when xn 6= 0. From the first equation (3.2.2) with j < n we now obtain
and using the boundary condition we conclude that (∂n f, ∂j f ) = gnj . Now we have
for 1 ≤ i ≤ j < n
When xn = 0 this is equal to ∂i gjn + ∂j gin − 2Γijn = ∂n gij by the second set of initial
conditions (3.2.3). Differentiating again we obtain
where we have used (3.2.4). In view of the initial conditions it follows that
(∂i f, ∂j f ) = gij for 1 ≤ i ≤ j < n, which proves the lemma.
As already mentioned we shall start from a local isometric embedding of Rn−1 , and
we must choose it so that the equations (3.2.3) with j < n can be solved. This requires
a stronger condition on the embedding, which we formulate with Rn−1 replaced by
Rn in the following definition:
Definition 3.2.2. A C 2 map x 7→ f (x) from a neighborhood of 0 in Rn to RN is said
to be free at 0 if the derivatives
are linearly independent. Their linear hull is then called the osculating space at 0.
It is clear that f is free at every point in a neighborhood of 0 if f is free at 0. If a
free map exists then N ≥ n + n(n + 1)/2, the dimension of the osculating space. Note
that when n is replaced by n − 1, this condition becomes N ≥ (n − 1) + (n − 1)n/2 =
n(n + 1)/2 − 1.
gjk (x)dxj dxk is a real analytic Riemann-
P
Theorem 3.2.3 (Janet-Cartan). If
ian metric in a neighborhood of 0 in Rn then there is a local real analytic isometric
embedding in RN with N = n(n + 1)/2. It can be chosen free if N = n(n + 3)/2.
LOCAL ISOMETRIC EMBEDDING 49
Proof. When n = 1 the statement is obvious. We may therefore assume that n > 1
and that the theorem has already been proved for dimensions smaller than n. Without
restriction we may assume that the coordinates are geodesic or at least that
Using the inductive hypothesis we choose a free real analytic map f0 from a neighbor-
hood of 0 in Rn−1 to RN−1 , where N = n(n + 1)/2, such that for x′ in a neighborhood
of 0
(∂j f0 (x′ ), ∂k f0 (x′ )) = gjk (x′ , 0), 1 ≤ j ≤ k < n.
We want to find f satisfying (3.2.2) so that f (x′ , 0) = (f0 (x′ ), 0) and (3.2.3) is valid.
The equations
determine uniquely the component v(x′ ) of ∂n f (x′ , 0) in RN−1 , because f0 is free, and
v(0) = 0 since the right-hand sides of these equations vanish at 0. The only remaining
equation (3.2.3) can now be written
where k · k denotes the Euclidean norm. Since the right-hand side is positive when
x′ = 0 we have a unique analytic positive solution. Summing up, we have well defined
analytic boundary conditions
p
(3.2.5) f (x′ , 0) = (f0 (x′ ), 0), ∂n f (x′ , 0) = (v(x′ ), gnn (x′ , 0) − kv(x′ )k2 ).
The equations (3.2.2) still have a unique analytic solution (3.2.6) if we require that
∂n2 f shall lie in the linear hull of ∂j f , 1 ≤ j ≤ n and ∂i ∂j f , 1 ≤ i ≤ j < n. Then it is
50 III. ABSTRACT RIEMANNIAN MANIFOLDS
clear that all derivatives ∂j f and ∂i ∂j f with i ≤ j are linearly independent at 0 with
the exception of ∂n2 f , and since we have an isometric embedding
P in R
N+n−1
we could
not hope for more. If f is such an embedding for the metric gjk dx dx − 4(xn dxn )2 ,
j k
Choose ϕ ∈ C0∞ (Rn ) such that ϕ(x) = 1 when |x| ≤ 1 and ϕ(x) = 0 when |x| ≥ 2,
and set for small ε
hij (x) = ϕ(x/ε) gij (x) − (∂i f (x), ∂j f (x)) .
Then |x| ≤ 2ε if x ∈ supp hij , gij (x) = hij (x) + (∂i f (x), ∂j f (x)) if |x| < ε, and
where Tij (u) is quadratic in the second derivatives of u. We can simplify the equations
by adding just n equations which annihilate the first two terms in all these equations
and conclude that (3.2.8) is valid if
in Ω and u = 0 on ∂Ω. Let G be the operator solving the Dirichlet problem for the
Laplacian in Ω, thus ∆Gψ = ψ in Ω and Gψ = 0 on ∂Ω if ψ ∈ C(Ω). Then these
conditions are fulfilled if
The vectors ∂i f (x), ∂i ∂j f (x) with 1 ≤ i ≤ j ≤ n form a basis for RN for every x ∈ Ω
since f is free, so we have for every U ∈ RN
X X
U= ϕi (∂i f, U ) + ϕij (∂i ∂j f, U ),
1≤i≤n 1≤i≤j≤n
where ϕi , ϕij are analytic in Ω. Thus the equations (3.2.9) are equivalent to u =
H + T (u) where
X X X
H = − 12 ϕij hij , T (u) = − ϕi G(∂i u, ∆u) + 1
2 ϕij GTij (u).
1≤i≤j≤n 1≤i≤n 1≤i≤j≤n
Standard Hölder estimates in the theory of elliptic equations show that T (u) is
continuous in C 2+̺ and a contraction operator in the convex subset where
α
P
|α|≤2 sup |∂ u| is small enough. In view of (3.2.7) it follows that the equation
u = H + T (u) has a unique solution there if ε is small enough. Arguing by stan-
dard elliptic theory we conclude that u is as smooth as h if h has additional regularity.
Admitting these basic facts without proof here, we obtain:
gjk (x)dxj dxk is a Riemannian metric in a neighborhood of 0
P
Theorem 3.2.4. If
in R with coefficients in C 2+̺ for some non-integer ̺ > 0, then there is a local C 2+̺
n
Theorem 3.2.5. If for all v in an open subset of C ∞ (Rn , Rν ) one can find a function
u ∈ C ∞ (Rn , RN ) such that Φ(x, Jm u) = v in a neighborhood of 0, then N ≥ ν.
Proof. Let µ be some large integer to be chosen later. We shall not really use the full
differential equation but only the much weaker condition
If N < ν then this condition is fulfilled for large µ, which proves the theorem and even
more: If N < ν then the equation Φ(x, Jm u) = v cannot even be satisfied to order µ
at a fixed point when µ is large, unless Jµ v is exceptional there in the sense that it
belongs to a set of measure 0.
For the isometric embedding problem one can improve the estimate of µ obtained
from the preceding proof by using the special properties of the equations. (See Exercise
2.3.1 and Gromov-Rokhlin [1].) Whether local isometric embedding of low regularity
is possible in dimensions below the Janet-Cartan dimension n(n + 1)/2 does not seem
to be known except for the result of Nash [2], Kuipers [1]; they showed that even a
global C 1 isometric embedding is possible in essentially the same dimension where a
C 1 embedding exists, hence always in 2n + 1 dimensions. However, for C k embeddings
with k ≥ 2 the situation is quite different as shown by (2.3.5), because the Gauss
equations are then available, and the problem does not seem to have been studied
then.
3.3. Spaces of constant curvature. Let M be a connected Riemannian manifold
of dimension n, and recall from Section 2.3 that the sectional curvature of M at a
point x ∈ M for the two plane spanned by t1 , t2 ∈ Tx M is defined by
If this is a function f (x) independent of the direction of the two plane then
by the uniqueness statement in Theorem 2.3.1, for the right-hand side has all the
symmetry properties defining T . Thus the Weyl tensor and the traceless Ricci tensor
are equal to 0. If this is true for every point in M , then M is an Einstein manifold, and
SPACES OF CONSTANT CURVATURE 53
if n > 2 it follows from Theorem 3.1.8 that f is a constant, so the sectional curvature is
independent of x also. (This is a classical theorem of F. Schur, far older than Theorem
3.1.8.)
Definition 3.3.1. A connected Riemannian manifold M is said to have constant cur-
vature K if the curvature tensor is given by
with a constant K.
From (3.3.2) it follows that the Ricci tensor is (n − 1)Kgij and that the scalar
curvature is n(n − 1)K. Thus a manifold of constant curvature is an Einstein manifold
but the converse is not true when n > 3 since the Weyl tensor may not be equal to 0
then. Since the covariant derivative of the metric tensor is equal to 0, this is also true
for the curvature tensor (3.3.2).
An obvious example of a manifold of constant zero curvature is Rn with the standard
Euclidean metric. We shall now prove that locally there are no others.
Theorem 3.3.2. A Riemannian manifold M with curvature tensor identically equal
to 0 is flat in P
the sense that at every point one can choose local coordinates such that
the metric is (dxj )2 .
Proof. Choose first some arbitrary local coordinates x = (x1 , . . . , xn ) varying over a
ball Ω with center at the origin. Set ∇i = ∇∂i where ∂i = ∂/∂xi . Since [∂i , ∂j ] = 0 it
follows from (3.1.6) and the hypothesis that [∇i , ∇j ] = 0. For every v0 ∈ Rn we can
therefore find a unique vector field v in Ω with v(0) = v0 and ∇i v = 0, i = 1, . . . , n. In
fact, we shall prove inductively for ν = 1, . . . , n that there is such a vector field defined
in Ων = {x ∈ Ω; xj = 0, j > ν} with ∇i v = 0, i ≤ ν. This is obvious when ν = 1,
for the equation ∇1 v = 0 is just a linear system of differential equations with leading
term ∂v/∂x1 which we can solve with initial value v0 . If ν > 1 and ṽ has the required
property with ν replaced by ν − 1, then we can find v defined in Ων with ∇ν v = 0 and
v = ṽ when xν = 0. If µ < ν it follows that ∇ν ∇µ v = ∇µ ∇ν v = 0 in Ων , and since
∇µ v = ∇µ ṽ = 0 when xν = 0, it follows that ∇µ v = 0 in Ων . When ν = n, the claim
is proved.
Now choose vector fields v1 , . . . , vn in Ω such that ∇vj = 0 for j = 1, . . . , n and
v1 , . . . , vn form an orthonormal basis at 0. Since d(vj , vk ) = (∇vj , vk ) + (vj , ∇vk ) = 0,
this follows at every point in Ω. We have
where K = 1/R2 . Thus gij = δij + Kxi xj + O(|x|4 ), and we obtain when x = 0 by
(3.1.14)
i l
Rijkl = K
2 (∂j ∂k x x + ∂i ∂l xj xk − ∂i ∂k xj xl − ∂j ∂l xi xk ) = K(δjl δik − δil δkj ).
This we knew already of course, for when n = 2 the total curvature is 1/R2 = K.
For the restriction to the hyperboloid H = {x ∈ Rn+1 ; (xn+1 )2 = |x|2 + R2 , xn+1 >
0} of the hyperbolic (Lorentz) metric |dx|2 − |dxn+1 |2 we have
where K = −1/R2 < 0 now. The metric is positive definite since |x|2 < (xn+1 )2 on H.
The preceding calculation for the sphere gives that
at the origin. To prove that the curvature is constant it suffices to note that the group
of
PnLorentz transformations, that is, linear transformations preserving the Lorentz form
j 2
1 (x ) − (xn+1 )2 acts isometrically and transitively on H.
The parametrization of S n above covers only a half sphere. We can cover the whole
sphere minus one point by means of the stereographic projection. If X ∈ S n then the
stereographic projection x ∈ Rn from the point (0, . . . , 0, −R) to the tangent plane at
the antipodal point is obtained from the equations
X = (0, . . . , 0, −R) + t(x, 2R), |X| = R, that is, t = 4R2 /(4R2 + |x|2 ).
Thus
|dX|2 = |tdx + xdt|2 + 4R2 dt2 = t2 |dx|2 + (4R2 + |x|2 )dt2 + 2tdthx, dxi = t2 |dx|2 ,
The stereographic projection just fills the open ball with radius 2R now. We obtain
n
X
|dX j |2 − |dX n+1 |2 = |tdx + xdt|2 − 4R2 dt2
1
= t2 |dx|2 + (|x|2 − 4R2 )dt2 + 2tdthx, dxi = t2 |dx|2 .
With K = −1/R2 we therefore obtain the metric
(3.3.3) ds2 = (1 + K|x|2 /4)−2 |dx|2 .
For K = −1 this is the Poincaré model of non-Euclidean geometry. By an inversion
we pass to the half space model. First we move the boundary point (0, . . . , 0, −2R) to
0 by introducing
y = x + (0, . . . , 0, 2R); 4R2 − |x|2 = −|y|2 + 4Ry n .
With z = y/|y|2 we make an inversion at the origin and obtain
We shall prove in Lemma 3.3.4 below that for j = 1, . . . , n there is a unique vector
field ej such that
(3.3.5) ∇̺ ej = 0, ej (0) = ∂j .
(This is not quite obvious since the radial vector field vanishes at 0.) Since
where (·, ·) denotes scalar product in the Riemannian metric, it follows that e1 , . . . , en
are an orthonormal system at every point in Ω. Let θ 1 , . . . , θ n be the one forms
biorthogonal to e1 , . . . , en , that is,
hθ j , ek i = δkj , j, k = 1, . . . , n.
i
P
Then t = i hθ , tiei for every tangent vector, so
X X
gjk tj tk = hθ i , ti2 .
i
Writing X
θi = Aij dxj , that is, Aij = hθ i , ∂j i,
̺Aij = hθ i , ∇̺ ∂j i = hθ i , ∇j ̺ − ∂j i,
for [̺, ∂j ] = −∂j . Now hθ i , ̺i = xi , for ∇̺ ̺ = ̺ since the radial direction is parallel, so
̺hθ i , ̺i = hθ i , ̺i, which means that hθ i , ̺i is homogeneous of P degree 1. At the origin
i i
we have θ = dx , so the assertion follows. It means that ̺ = xi ei . Thus
X X
(̺ + 1)Aij = hθ i , ∇j ̺i = ∂j xi − h∇j θ i , ̺i = δji − xk h∇j θ i , ek i = δji + xk Bkj
i
.
and must now obtain a differential equation for this new quantity. By (3.1.6)
i
̺Bjk = hθ i , ∇̺ ∇k ej i = −hθ i , ∇k ej i + R(ei , ej , ̺, ∂k ),
so it follows that X
R(ei , ej , ̺, el ) = xm Rijml.
m
Summing up and introducing polar coordinates x = rω, we have the differential equa-
tions
d X
(rAij ) = δji + rBkji
ωk ,
dr
d i
X
(rBjk )= Rijml ω m rAlk
dr
l,m
where (Av)j = Γik j xk v i has coefficients vanishing of second order at the origin. If
P
we introduce polar coordinates, x = rω, the problem takes the form
X
∂v j /∂r + Γik j (rω)ω k v i = 0, v = v0 when r = 0.
v = v0 + v1 + v2 + . . .
v1 + 2v2 + · · · + A(v0 + v1 + . . . ) = 0
where f = O(r N−1 ). This implies that w = O(r N ), and since ∂/∂xν = ων ∂/∂r +
h∂ω/∂xν , ∂/∂ωi, where ∂ω/∂xν is homogeneous of degree −1, it follows that w ∈ C N−1
as a function of the original variables, the derivatives vanishing at 0. This completes
the proof of the lemma and of the theorem.
The tools introduced to prove Theorem 3.3.3 also give another useful result on the
connection between the curvature tensor and the metric tensor.
(x)dxj dxk be the metric form for a geodesic coordinate
P
Theorem 3.3.5. Let gjkP
system centered at 0, thus gjk (x)xk = xj . Any derivative ∂ α gjk (0) can then be
expressed as a polynomial in the components of the curvature tensor and its covariant
derivatives of order ≤ |α| − 2.
Proof. It suffices to prove the theorem with covariant derivatives replaced by deriva-
tives ∂ α with respect to the coordinates. In fact, the difference
Here em are the vector fields in the proof of Theorem 3.3.3, said to form a synchronous
frame for T M . If we prove that derivatives of em of order ≤ ν at 0 are polynomials in
those of R of order ≤ ν − 2, the theorem will be proved.
SPACES OF CONSTANT CURVATURE 59
which gives the following result which will be needed in Section 6.10:
Theorem 3.3.6. If e1 , . . . , en is the synchronous frame, equal to the basis vectors at
the origin of a geodesic coordinate system, then
X
∇l ej = 21 Ri jkl (0)xk ei + O(|x|2 ),
i,k
Lemma 3.4.4. If a ∈ Rn then the metric (1 + 2hx, ai + |x|2 |a|2 )−2 |dx|2 in Rn is flat
except at the singularity where |a|2 x = −a 6= 0.
Proof. If we take y = |a|2 x + a, the metric becomes |y|−4 |dy|2 , and the inversion
z = y/|y|2 reduces it to |dz|2 , so it is flat. (See also the argument leading to (3.3.4).)
Proof of Theorem 3.4.3. If ϕ is a constant, the statement is obvious, so we may assume
that ϕ(0) = 0. Next assume that g is the Euclidean metric in Rn and that
ϕ(x) = ϕa (x) = − log(1 + 2hx, ai + |x|2 |a|2 ) = −2hx, ai − |x|2 |a|2 + 2hx, ai2 + O(|x|3 )
and it follows at once that the right-hand side of (3.4.1) vanishes, so Theorem 3.4.3 is
valid in this case. In the general case we write ϕ(x) = ϕa (x) + q(x) + O(|x|3 ) where
q is quadratic, and we assume that gjk (x) = δjk + Gjk (x) + O(|x|3 ) where Gjk is a
quadratic form. Then
X
g̃ = e2ϕa g0 + G + 2qg0 + O(|x|3 ); g0 = (dxj )2 .
Only the first term contributes to the Christoffel symbols Γ eijk for the metric g̃ at 0.
′
From the linearity of the formula (2.1.13) in the second order derivatives we conclude
e is the sum of the curvature tensor for the metric e2ϕa g0 , which is 0, the curvature
that R
tensor R and the curvature tensor for the metric g0 + 2qg0 . The latter has the ijkl
component
gil ∂j ∂k q + gjk ∂i ∂l q − gjl ∂i ∂k q − gik ∂j ∂l q
which agrees with the terms in (3.4.1) of second order in ϕ. The proof is complete.
If we contract (3.4.1) in the indices jl we obtain the transformation law for the Ricci
curvature:
X
(3.4.2) Reik = Rik + (2 − n)ϕik − ( g jl ϕjl + (n − 1)|∇ϕ|2 )gik .
Another contraction gives the transformation rule for the scalar curvature:
If we multiply (3.4.3) by gij /(2 − 2n) and add to (3.4.2)′ , we eliminate ∆ϕ and obtain
R
eij + Sg̃
e ij /(2 − 2n) = Rij + Sgij /(2 − 2n) + (2 − n)Φij , Φij = ϕij + 12 |∇ϕ|2 gij .
(3.4.1)′ e−2ϕ R
eijkl = Rijkl + gil Φjk + gjk Φil − gik Φjl − gjl Φik
it follows that
−2ϕ
g̃il R
ejk + g̃jk R
eil − g̃ik R
ejl − g̃il R
eik S(g̃
e il g̃jk − g̃ik g̃jl )
e Rijkl +
e −
n−2 (n − 1)(n − 2)
gil Rjk + gjk Ril − gik Rjl − gil Rik S(gil gjk − gik gjl )
= Rijkl + − .
n−2 (n − 1)(n − 2)
(3.4.5) e−2ϕ W
fijkl = Wijkl
for the corresponding Weyl tensors. Hence we have the following corollary to Theorem
3.4.3:
Corollary 3.4.4. Under the hypotheses of Theorem 3.4.3 we have the transformation
law (3.4.5) for the corresponding Weyl tensors.
Exercise 3.4.1. Show that if g̃ = e2ϕ g then the corresponding covariant derivatives are
related by
e X Y = ∇X Y + (Xϕ)Y + (Y ϕ)X − (X, Y )(dϕ)♯ ,
∇
To clarify the role of the Weyl tensor we shall discuss these equations for the vanishing
of the Ricci P
tensor Reij without assuming at first that W = 0. Let Φ be the one form
∇ϕ = dϕ = ∂i ϕdxi . Then the preceding equations can be written
X
(3.4.6) Rij + (2 − n)(Φi,j − Φi Φj ) − ( g kl Φk,l + (n − 2)|Φ|2 )gij = 0.
k,l
Conversely, if we can find a one form satisfying these equations in an open set Ω ⊂ M ,
we obtain using (3.1.2)′
0 = Φi,j − Φj,i = ∂j Φi − ∂i Φj
or equivalently
X
2(n − 1)( g kl Φk,l + (n − 2)|Φ|2 ) = S + (n − 2)(n − 1)|Φ|2 .
k,l
(3.4.9) Rijkl = Wijkl + ωik gjl − ωil gjk + ωjl gik − ωjk gil .
Here the right-hand side can be expressed using (3.4.9), and we have
Furthermore,
X
(3.4.12) 1
∇ |Φ|2
2 k
= (∇k Φ, Φ) = g lm Φl,k Φm
l,m
X X
= g lm (Φl Φk + ωlk − 21 glk |Φ|2 )Φm = g lm ωlk Φm + 12 Φk |Φ|2 .
l,m l,m
With Φν denoting the coordinates of Φ♯ , it follows from (3.4.9) that the right-hand
side of (3.4.10) can be written
X X X
Wνijk Φν + gik ωνj Φν − gij ωνk Φν + ωik Φj − ωij Φk .
ν ν ν
If W = 0, then these conditions no longer involve Φ at all, so they are necessary for
the existence of any solution at all to (3.4.6), and they imply the local existence of a
solution with given value at a point. On the other hand, if there is a solution with
∇ϕ given at a point, then it follows that both sides of (3.4.10)′ must vanish, since the
right-hand side is independent of Φ, hence W = 0 and
We shall now show that when n > 3 there is a second miracle: the conditions
(3.4.10)′′ are always fulfilled if the Weyl tensor is equal to 0. (This is not true when
n = 3 which is not surprising since the Weyl tensor is always 0 then.) In fact,
X
(3.4.13) (n − 3)(ωjk,i − ωik,j ) = W l kij,l .
l
(I owe the following calculations to Anders Melin.) First note that the definition of
the Ricci tensor and the second Bianchi identity give
X X X X
Rjk,i − Rik,j = Rl klj,i − Rl kli,j = (Rl klj,i + Rl kil,j ) = Rl kij,l .
l l l l
CONFORMAL GEOMETRY 65
for ∇g = 0. Since
we have
Rij,l = (n − 2)ωij,l + S,l gij /(2n − 2).
If we multiply by g il and contract in i, l, recalling (3.1.25), we obtain
X X
1
S
2 ,j
= (n − 2) ω l j,l + S,j /(2n − 2), hence ω l j,l = S,j /(2n − 2).
(3.4.13) follows if we use this result and (3.4.8)′ in (3.4.14), for S drops out. Hence we
have proved (3.4.13) and the following
Theorem 3.4.5. A Riemannian manifold of dimension > 3 is conformally flat if
and only if the Weyl tensor vanishes. A Riemannian manifold of dimension 3 is
conformally flat if and only if the integrability conditions (3.4.10)′ are valid.
We shall finally discuss how the Laplace-Beltrami operator defined by (3.4.4) is
changed when one passes to a conformal metric. First we give an equivalent and often
more useful definition:
Proposition 3.4.6. For every u ∈ C 2 we have
1 1
X X
(3.4.15) g ij u,ij = g − 2 ∂j (g 2 g ij ∂i u),
i,j i,j
where g = det(gij ).
Proof. If u, v are in C 2 with support in the same coordinate patch, then
Z X
1
g ij ∂i u∂j vg 2 dx
i,j
is invariantly defined, dx denoting the Lebesgue measure in the local coordinates, for
1
g 2 dx is the invariant volume element dvol(x). Since integration by parts shows that
the integral is equal to
Z X
1 1
− v g − 2 ∂j (g 2 g ij ∂i u) dvol(x)
66 III. ABSTRACT RIEMANNIAN MANIFOLDS
we conclude that the right-hand side of (3.4.15) is invariantly defined, and so is the
left-hand side. Both are equal at the center of a geodesic coordinate system, which
proves the proposition.
Exercise 3.4.2. Prove (3.4.15) by explicit computation using Exercise 2.1.1.
Denote by ∆e the Laplace-Beltrami operator defined using the conformal metric
2ϕ
g̃ = e g. When n = 2 it follows from the second expression for ∆ in (3.4.15) that ∆ e =
e−2ϕ ∆; hence the harmonic functions satisfying the homogeneous Laplace-Beltrami
equation are the same for two conformal metrics. This is not true when n > 2, but we
shall now prove a substitute result in that case. With a constant a to be chosen later
we have
e aϕ u) = e−nϕ g − 21 1
X
∆(e ∂j (e(n−2)ϕ g 2 g jk ∂k (eaϕ u))
X
= e(a−2)ϕ (∆u + (n − 2 + 2a) g jk ∂j ϕ∂k u) + F u,
where F does not depend on u. Now we choose a = 1 − n2 so that the first order terms
disappear. If the formula is applied with u = e−aϕ , it follows that
(3.5.1) e X Y − ∇X Y = h(X, Y ),
∇
τ∇
eXY − τ∇
e Y X = τ [X, Y ] = [X, Y ],
(τ ∇
e Z X, Y ) + (X, τ ∇
e Z Y ) = (∇
e Z X, Y ) + (X, ∇
e Z Y ) = Z(X, Y ),
hence τ ∇
e X Y = ∇X Y by the uniqueness statement in Theorem 3.1.2.
From (3.1.3) applied to ∇ and ∇
e we obtain
e X Y − ∇X Y = ∇
∇ e Y X + [X, Y ] − ∇Y X − [X, Y ] = ∇
e Y X − ∇Y X,
which proves that the left-hand side of (3.5.1) is symmetric in X and Y . If ϕ ∈ C0∞
then
e X (ϕY ) − ∇X (ϕY ) = ∇
∇ e ϕY X − ∇ϕY X = ϕ(∇
e Y X − ∇Y X) = ϕ(∇
e X Y − ∇X Y ).
Hence the left-hand side of (3.5.1) at x ∈ M is a symmetric bilinear form in X(x) and
Y (x), with values in Nx (M ), which completes the proof.
Definition 3.5.2. The symmetric bilinear map h in Tx M with values in Nx M defined
by (3.5.1) is called the second fundamental form of M with respect to M
f.
In the particular case where M f = RN Theorem 3.1.5 gives back the expression
(2.1.11) for the covariant derivative. We shall now also give an extension of the Gauss
equations (2.1.13), (2.1.15). Let X, Y, Z, W be vector fields in M . By (3.1.6) we have
(∇X ∇Y − ∇Y ∇X − ∇[X,Y ] )Z, W = R(W, Z, X, Y ),
68 III. ABSTRACT RIEMANNIAN MANIFOLDS
where R is the Riemann curvature tensor of M , and we have a similar formula with ∇
and R replaced by ∇e and R.
e (We can extend X, Y, Z, W to vector fields on Mf, but on
M the result is independent of the extension.) By Theorem 3.5.1 we have
(∇X ∇Y Z, W ) = (∇
e X ∇Y Z, W ) = (∇ e Y Z, W ) − (∇
eX∇ e X h(Y, Z), W ).
Since (h(Y, Z), W ) = 0 on M , it follows from (3.1.4) and Theorem 3.5.1 that
e X h(Y, Z), W ) = (h(Y, Z), ∇
−(∇ e X W ) = h(Y, Z), h(X, W ) .
Hence
(∇X ∇Y Z, W ) = (∇ e Y Z, W ) + h(Y, Z), h(X, W ) ,
eX∇
R(W, Z, X, Y ) = R(W,
e Z, X, Y ) + h(Y, Z), h(X, W ) − h(X, Z), h(Y, W ) ,
which proves the following extension of the Gauss equations (2.1.13), (2.1.15):
Theorem 3.5.3. If M is a smooth submanifold of a Riemannian manifold M f, with
Riemann curvature tensors R and R, respectively, and t1 , t2 , t3 , t4 ∈ Tx M , then
e
(3.5.2) R(t1 , t2 , t3 , t4 ) = R(t
e 1 , t2 , t3 , t4 ) + h(t1 , t3 ), h(t2 , t4 ) − h(t1 , t4 ), h(t2 , t3 ) ,
Summary. Chapters II and III have been based on the tensor calculus of Ricci. We
shall now discuss an alternative approach due to E. Cartan using the calculus of exterior
differential forms and moving orthogonal frames systematically. In Section 4.1 we recon-
sider the study in Chapter II of submanifolds of Euclidean space, and in Section 4.2 we
discuss abstract Riemannian manifolds, as in Chapter III. Using the tools developed we
then return to the Gauss-Bonnet theorem in Section 4.3 and prove its analogue in higher
dimensions following S. S. Chern. Pontrjagin classes are then defined in Section 4.4.
ej : F0 (V ) ∋ (e1 , . . . , eN ) 7→ ej ∈ V.
The differential dej is a linear form on the tangent space of F0 (V ) with values in V ,
so it can be written
N
X
(4.1.1) dej = ωkj ek ,
k=1
we have
Since
N
X N
X N
X N
X
0 = d2 ej = (dωkj )ek − ωkj ∧ dek = (dωkj )ek − ωlj ∧ ωkl ek ,
k=1 k=1 k=1 k,l=1
69
70 IV. EXTERIOR DIFFERENTIAL CALCULUS IN RIEMANNIAN GEOMETRY
we obtain
N
X
(4.1.3) dωkj + ωkl ∧ ωlj = 0, j, k = 1, . . . , N.
l=1
The equations (4.1.2), (4.1.3) are of course just another way of writing the results of
Section 1.4 such as (1.4.1).
If W ⊂ V is a subspace of dimension n, the set of orthonormal frames in F0 (V )
such that e1 , . . . , en ∈ W , hence en+1 , . . . , eN ∈ W ⊥ , is a submanifold F0 (V, W ) which
we can identify with the group O(n) × O(N − n). It is clear that the restriction of the
form ωjk above to F0 (V, W ) vanishes if j ≤ n and k > n. When j ≤ n and k ≤ n,
it is equal to the analogous form on F0 (W ) pulled back to F0 (V, W ) by the obvious
surjective map F0 (V, W ) → F0 (W ).
Now let M be a C ∞ submanifold of V , of dimension n, and set
(4.1.4) Fx = F0 (V, Tx M ), x ∈ M,
which is the set of orthonormal frames at x such that e1 , . . . , en span the tangent space
Tx M while en+1 , . . . , eN span the normal space. We shall use the notation α, β, . . . for
indices running from 1 to n, the notation r, s, . . . for indices running from n +1, . . . , N ,
and A, B, . . . for indices running from 1 to N . It is clear that F (M ) = ∪x∈M {x} × Fx ,
as a subset of M × F0 (V ) is a C ∞ fiber space over M . Composing the projection
p : F (M ) → M with the embedding M → V we get a map F (M ) → V , which we also
denote by p. In addition we have N maps eA : F (M ) → V . Since the range of the
differential dp at a point in Fx is equal to Tx M , we can write
n
X
(4.1.5) dp = ωα eα ,
α=1
N
X
(4.1.6) deA = ωBA eB , A = 1, . . . , N,
B=1
(4.1.7) ωAB + ωBA = 0, A, B = 1, . . . , N.
Here ωα and ωAB are scalar differential forms on F (M ). From (4.1.5) we obtain
X X X X
0 = d2 p = (dωα )eα − ωα ∧ deα = (dωα )eα − ωα ∧ ωBα eB
hence
N
X
(4.1.10) dωBA + ωBC ∧ ωCA = 0, A, B = 1, . . . , N.
C=1
In particular,
(4.1.11)
n
X
dωαβ + ωαγ ∧ ωγβ = Ωαβ , α, β = 1, . . . , n, where
γ=1
N
X
(4.1.12) Ωαβ = − ωαr ∧ ωrβ , α, β = 1, . . . , n.
r=n+1
with uniquely determined coefficients λrαβ ∈ C ∞ (F (M )). Here λrαβ = λrβα according
to (4.1.9). Hence
N
X N
X n
X
Ωαβ = ωαr ∧ ωβr = λrα̺ λrβσ ω̺ ∧ ωσ
r=n+1 r=n+1 ̺,σ=1
is also a form with vanishing restriction to the fibers. The forms ωα , ωαβ and Ωαβ ,
α, β = 1, . . . , n, can be obtained by pulling back forms to F (M ) from the fiber space
P (M ) over M with fiber at x consisting of the orthonormal frames in Tx M , for the
definitions above are already applicable in P (M ) without reference to the normal
vectors.
To clarify the meaning of the preceding equations we shall express them in terms
of our earlier notation. Thus consider a subset of M with a local parametrization
x 7→ f (x), where x varies in an open subset of Rn . We can consider the coordinates
x1 , . . . , xn as functions on M and lift them to F (M ). With the notation fα = ∂f /∂xα ,
which is a vector field on M , we then obtain
n
X n
X n
X
dp = fβ (x)dxβ = ωα eα , thus ωα = (fβ , eα )dxβ .
β=1 α=1 β=1
72 IV. EXTERIOR DIFFERENTIAL CALCULUS IN RIEMANNIAN GEOMETRY
Here dxβ is a differential form on M pulled back to F (M ), and the coefficients are
functions on F (M ). Let s be a local section of F (M ) and set s∗ eα = Eα . Then we
have
n
X n
X
s∗ ωβα = (dEα , Eβ ) = (∂Eα /∂xi , Eβ )dxi = (∇fi Eα , Eβ )dxi .
i=1 i=1
n
X
(4.1.13) Ωαβ = 1
2 R(eα , eβ , fi , fj )dxi ∧ dxj .
i,j=1
fi dxi =
P P
If we use that dp = ωγ eγ , and evaluate (4.1.13) on a pair of tangent
vectors, we conclude that
n
X
′ 1
(4.1.13) Ωαβ = 2
R(eα , eβ , eγ , eδ )ωγ ∧ ωδ .
γ,δ=1
Thus the forms Ωαβ contain exactly the same information as the Riemann curvature
tensor. From (4.1.13)′ we also see again that Ωαβ is a linear combination with co-
efficients in C ∞ (P (M )) of forms pulled back from M to P (M ). By the expression
(2.1.13)′ for the curvature tensor it also follows that Ωαβ is independent of the imbed-
ding of M . In Section 4.2 we shall give an independent proof of that by reconsidering
the preceding formulas for an abstract Riemannian manifold.
We shall now also calculate the forms ωαr . With the same notation as above and
(2.1.5), we have
X X
s∗ ωrα = (dEα , Er ) = (∂Eα /∂xj , Er )dxj = (h(fj , Eα ), Er )dxj
X
= (h(Eβ , Eα ), Er )s∗ ωβ ,
β
fj dxj =
P P
since dp = eβ ωβ . Hence we obtain
n
X
(4.1.14) ωrα = (h(eβ , eα ), er )ωβ , α = 1, . . . , n, r = n + 1, . . . , N.
β=1
SUBMANIFOLDS OF A EUCLIDEAN SPACE 73
n
X X
(4.1.15) Ωαβ = ωαr ∧ ωβr = (h(eγ , eα ), h(eδ , eβ ))ωγ ∧ ωδ ,
r=1 γ,δ
which in view of (4.1.13) shows that (4.1.12) is equivalent to the Gauss equations.
We shall finally express the degree of the Gauss mapping for a hypersurface in terms
of differential forms (see (3.1.23)). Thus consider a compact oriented hypersurface
M ⊂ Rn+1 . We can then restrict ourselves to positively oriented frames in F (M ) such
that en+1 has the positive normal direction. The product
whereQKα are the principal curvatures. Hence (4.1.16) is equal to (−1)n Kω where
n
K = 1 Kα is the total curvature and ω = ω1 ∧ · · · ∧ ωn is the volume form on M
(pulled back to F (M )). If D is the degree of the reflected Gauss map γ̌, we obtain in
view of (3.1.23)
Z
n n
Dvol(S ) = (−1) ω1,n+1 ∧ · · · ∧ ωn,n+1
M
Z X
n
= (−1) εi1 ...in ωi1 ,n+1 ∧ · · · ∧ ωin ,n+1 /n!
M
where εi1 ...in is +1 for even and −1 for odd permutations of 1, . . . , n and 0 otherwise.
(The integrands here and below should be read as the forms on M which pull back to
them.) If n is even we can pair the factors using the Gauss equations (4.1.12), which
gives
Z
n
X
n
(4.1.17) Dvol(S ) = (−1) 2 εi1 ...in Ωi1 i2 ∧ Ωi3 i4 ∧ · · · ∧ Ωin−1 in /n!.
M
The forms in the right-hand side are well defined for any abstract Riemannian manifold,
and we shall take (4.1.17) as our starting point when generalizing the Gauss-Bonnet
theorem to higher dimensions.
74 IV. EXTERIOR DIFFERENTIAL CALCULUS IN RIEMANNIAN GEOMETRY
(Since all indices run from 1 to n now, we shall often omit the range.) We shall now
extend (4.1.8).
Theorem 4.2.1. There are uniquely determined one forms ωij on P (M ), i, j =
1, . . . , n, such that
n
X
(4.2.2) ωij + ωji = 0, dωj + ωk ∧ ωkj = 0, i, j = 1, . . . , n.
k=1
so a circular permutation of the indices changes the sign. After three circular permu-
tations we conclude that λijk = 0.
Existence. In view of the uniqueness it suffices to prove existence locally, so we may
assume that M is an open subset of Rn , with coordinates denoted by (x1 , . . . , xn ), and
we identify the tangent space of Rn with Rn . Since
X
dp = (dx1 , . . . , dxn ) = ωj ej ,
(The scalar product is of course taken in the Riemannian metric.) The restriction of
(dek , ej ) to a fiber of P (M ) is antisymmetric in j and k, so we can find one forms Bjk
which are also antisymmetric in j, k such that (dek , ej ) + Bjk vanishes on the fibers.
Then we obtain X X
dωi = Bik ∧ ωk + Aijk ωj ∧ ωk
Since
Akij + Aijk + Aikj + Akji = 0
it follows that ωik is antisymmetric in its indices, and
X X X
ωi ∧ ωki = ωi ∧ (Bik + Akij ωj ) = dωk ,
j
This proves (4.2.4). From the discussion of vector spaces at the beginning of Section
4.1 we know that the restriction of Ωik to any fiber must vanish. (See (4.1.3).) We can
therefore write X X
Ωik = θikj ∧ ωj + 21 Rikjl ωj ∧ ωl .
Here Rikjl are C ∞ functions on P (M ) and θikj are linear combinations with such
coefficients of forms extending ω1 , . . . , ωn to a basis for one forms at a point in P (M ).
The forms θikj are skew symmetric in i and k since Ωik are and θikj is unique. By
(4.2.4) we must have X
θikj ∧ ωk ∧ ωj = 0
k,j
and we can conclude that θikj are skew symmetric in k, j. Recalling the first part of
the proof of Theorem 4.2.1 we can therefore conclude that θikj = 0. Hence we obtain
(4.2.5), where the coefficients are uniquely determined if we require skew symmetry in
j, l. The proof is complete.
If we enter the expression (4.2.5) in (4.2.4) we find that
Ri[kjl] = 0,
To show that this is equivalent to the second Bianchi identity (3.1.24) we choose a
system of geodesic coordinates x1 , . . . , xn with center at a given point. By Lemma 3.3.4
there are unique orthonormal vector fields E1 , . . . , En near 0 such that Ej = ∂/∂xj at
the origin and the covariant derivative along the radial vector field vanishes. Thus we
obtain a section s of P (M ) which we can use to pull back (4.2.6) from P (M ) to M .
By Lemma 3.3.4 the first order derivatives of Ej vanish at 0, and since the forms s∗ ωi
are biorthogonal,
P ∗ their first derivatives also vanish at 0. Hence it follows from (4.2.2)
that s ωij ∧ s∗ ωj = 0 at 0. Since s∗ ωij is skew symmetric in i, j, it follows (see the
proof of uniqueness in Theorem 4.2.1) that s∗ ωij = 0 at 0, so (4.2.6) gives ds∗ Ωik = 0.
By (4.2.5) we have
X
s∗ Ωik = 21 R(Ei , Ek , Ej , El )s∗ ωj ∧ s∗ ωl ,
THE GAUSS-BONNET THEOREM IN HIGHER DIMENSIONS 77
at the origin, if we recall that the derivatives of s∗ ωj and Ej vanish there. At the center
of a geodesic coordinate system the Christoffel symbols vanish so covariant derivatives
are equal to the corresponding partial derivatives, and we obtain
Rik[jl,r] = 0,
We shall first verify that it is the pullback of a form on M . Since (4.2.5) shows that ∆0
is a linear combination with coefficients in C ∞ (P (M )) of pullbacks of forms on M , it
suffices to show thatP ∆0 is invariant under the map of P (M ) into itselfPreplacing P
a frame
′
e1 , . . . , en by eP
k = O e
i ik i where O ∈ SO(n) is constant. Since ω e
k k = ωk′ e′k
we have ωi = Oik ωk′ then, hence
X X X
dωi = Oik dωk′ = − ′
Oik ωkl ∧ ωl′ = − ′
Oik ωkl ∧ ωj Ojl ,
k k,l j,k,l
X X
′
so ωij = Oik Ojl ωkl , Ωij = Oik Ojl Ω′kl ,
k,l k,l
and we obtain
X
∆0 = εi1 ...in Oi1 j1 . . . Oin jn Ω′j1 j2 ∧ · · · ∧ Ω′jn−1 jn .
Here we have used that there will be no change of sign in terms where d acts on a factor
Ω. Differential forms of even degree commute, so the differential of such a factor can be
moved to the right of the others without any sign change, and this corresponds to an
even number of inversions of the indices αj . Thus we obtain k identical contributions
when differentiating on the Ω factors. When the factor ωα2k+j n is differentiated we
obtain a sign factor (−1)j−1 which is compensated by j − 1 inversions if it is moved
to the left of the other factors ω, so we get n − 2k − 1 identical contributions of this
kind, which proves (4.3.3).
We shall now calculate the right-hand side of (4.3.3) using that
X X
dΩαβ = Ωαj ∧ ωjβ − ωαj ∧ Ωjβ ,
j j
X
dωαn = Ωαn − ωαj ∧ ωjn .
j
THE GAUSS-BONNET THEOREM IN HIGHER DIMENSIONS 79
In view of Lemma 4.3.1 we may omit terms containing a factor ωαβ when simplifying
(4.3.3), so we may replace dΩαβ by Ωαn ∧ ωnβ − ωαn ∧ Ωnβ = −Ωαn ∧ ωβn + Ωβn ∧ ωαn
and dωαn by Ωαn . To handle the first sum in (4.3.3) we now introduce for 0 ≤ k < n2
the n forms
(4.3.4) X
Ψk = 2(k + 1) εα1 ...αn−1 Ωα1 α2 ∧ · · · ∧ Ωα2k−1 α2k ∧ Ωα2k+1 n ∧ ωα2k+2 n ∧ · · · ∧ ωαn−1 n .
As for the forms Φk one verifies at once that they are forms on S(M ) pulled back to
P (M ). In view of Lemma 4.3.1 we obtain
(4.3.5) dΦk = −Ψk−1 − ((n − 2k − 1)/(2k + 2))Ψk ,
for the difference between the two sides is the pullback to P (M ) of a form on S(M ),
and every term contains a factor ωαβ . For k = 0 we must interpret Ψ−1 as 0, for then
there are no factors Ω to differentiate.
When k = p − 1 where p = n2 (recall that n is even) we obtain
X
Ψp−1 = n εα1 ...αn−1 Ωα1 α2 ∧ · · · ∧ Ωαn−1 n = ∆0 ,
so we can write ∆0 as a differential by successive elimination of Ψk , k = p−1, p−2, . . . , 0
using (4.3.5). This gives
p−1
X p−1
Y
p−k
(4.3.6) ∆0 = dΦ, Φ= (−1) Φk (2j + 2)/(n − 2j − 1) .
k=0 j=k
Using (4.3.6) instead of (3.1.21) we can now proceed as in the proof of Theorem
3.1.7. Choose f ∈ C ∞ (M ) with only non-degenerate critical points, and let F =
(df )♯ = grad f be the corresponding vector field. We recall that
Γ = {(x, F (x)/kF (x)k); F (x) 6= 0} ⊂ S(M ),
is a manifold with boundary ∂Γ consisting of the fibers Sx over the finitely many
critical points of f , with positive or negative orientation when the sign of the Hessian
of f at x is negative or positive, respectively. The integral of ∆0 over M is equal to the
integral over Γ of the pullback to S(M ), which by Stokes’ formula and (4.3.6) is equal
to the sum of the integrals of Φ over Sx (M ) when f ′ (x) = 0. Since the restriction of
Ωαβ to Sx (M ) is equal to 0, we only get contributions from Φ0 ,
Z Z
p p
(−1) Φ = 2 p!/(2p − 1)!! Φ0 = −(2p p!/(2p − 1)!!)(n − 1)!vol(S n−1 )
Sx (M ) Sx (M )
This gives Z
p n
(−1) /(n!vol(S )) Φ = − 12 .
Sx (M )
Here Sx (M ) has been oriented so that e1 , . . . , en−1 is a positively oriented frame in the
tangent space at en ∈ Sx (M ) if e1 , . . . , en−1 , en , is positively, that is, en , e1 , . . . , en−1
is negatively oriented, so the sphere has been oriented as the boundary of its exterior.
Hence we obtain
Z X
p n
(4.3.7) (−1) /(n!vol(S )) ∆0 = 12 εx ,
M f ′ (x)=0
where εx is the sign of the Hessian of f at x. (We can check the sign by noting that
for M = S n , embedded in Rn+1 as the unit sphere, the right-hand side is equal to 1 if
f is one of the coordinates in Rn+1 , so the sign agrees with (4.1.17).) Using a method
of Chern we have now proved the following extension due to Allendoerfer, Fenchel and
Weil of the Gauss-Bonnet theorem:
Theorem 4.3.2. If M is a compact oriented Riemannian manifold of even dimension
n = 2p, then (4.3.7) is valid with ∆0 denoting the form on M with pullback to P (M )
defined by (4.3.1) and any f ∈ C ∞ (M ) with only non-degenerate critical points, εx
denoting the sign of the Hessian at a critical point x.
As observed after Theorem 3.1.7 we could also replace (df )♯ by any C ∞ vector field F
with only non-degenerate zeros, with the sum taken over the zeros of F and εx defined
as the index of the zero. The right-hand side of (4.3.7) is the Euler characteristic of
M , which we cannot define until Chapter VI though. Without having this concept
available we can still remark that the right-hand side is completely independent of the
Riemannian metric, so the integral in (4.3.7) can only depend on the differentiable
structure of M .
4.4. Pontrjagin forms and classes. The proof that the form ∆0 defined by (4.3.1)
is the pullback of a form on M suggests the definition of other forms on P (M ) with
the invariance properties which guarantees that they are such pullbacks. The most
important ones come from the determinant
where I = (i1 , . . . , ik) and J = (j1 , . . . , jk ) are sequences of k different indices between
1 and n, and sgn JI is defined as 0 if they are not permutations of each other and
as the sign of the permutation otherwise. The sum (4.4.2) is equal to 0 if k is odd,
I
for if we exchange I and J then sgn J does not change but the following differential
forms are multiplied by −1. It follows from the Bianchi identities (4.2.6) that the forms
(4.4.2) are closed. In fact, since the restrictions of the forms ωjk , j < k, to the fibers
of P (M ) are linearly independent, an obvious analogue of Lemma 4.3.1 shows that we
may calculate the differential as if dΩjk were equal to 0. With a normalization which
will be explained later on, we introduce:
Definition 4.4.1. The Pontrjagin form Pk of a Riemannian manifold M is the form on
M of degree 4k which pulled back to the frame bundle P (M ) is the term of degree 4k
in det(δij + Ωij /2π) where Ωij are the curvature forms (4.2.3). Thus Pk is given by
(4.4.2) with k replaced by 2k, divided by (2π)2k . The de Rham cohomology class of
the closed form Pk is called the Pontrjagin class of M .
The definition of the Pontrjagin form depends heavily on the Riemannian metric,
but we shall now prove that the Pontrjagin classPis independent of it. On any C ∞
manifold one can define a Riemannian metric by ϕj gj where {ϕj } is a partition of
unity subordinate to a covering with coordinate patches and gj is the Euclidean metric
in the local coordinates. Thus the Pontrjagin classes are well defined on arbitrary
compact C ∞ manifolds.
Theorem 4.4.2. If M1 and M2 are Riemannian manifolds with metrics P g1Mand g2 ,
respectively, and if M1 × M2 is given the metric g1 + g2 , then the sum Pk of the
∗
P M1 ∗
P M2
Pontrjagin forms of M is equal to p1 Pk ∧ p2 Pk , where pj is the projection
M1 × M2 → Mj . The Pontrjagin class of a differentiable manifold is independent of
the metric.
Proof. On P (M1 ) × P (M2) ⊂ P (M1 × M2 ) it is clear that the matrices ωjk and Ωjk
∗ 0
are block matrices where the diagonal entries are the corresponding matrices
0 ∗
for M1 and M2 , lifted to P (M1 × M2 ). This proves the first statement. Now denote by
M a C ∞ manifold with two different metrics g0 and g1 . Choose an increasing function
χ ∈ C ∞ (R) with χ = 0 in (−∞, 1/3) and χ = 1 in (2/3, ∞). On M × R we introduce
the metric
g = χ(t)g1 + (1 − χ(t))g0 + (dt)2 ,
where t is the coordinate in R. From the first part of the proof it follows that on
M × (−∞, 1/3) (resp. M × (2/3, ∞)) the Pontrjagin forms Pk on M × R are equal to
the Pontrjagin forms Pk0 (resp. Pk1 ) of (M, g0 ) (resp. (M, g1 )) pulled back to M × R.
If
it (x) = (x, t) ∈ M × R, x ∈ M, t ∈ R,
then Pkt = i∗t Pk , t = 0, 1. Since i0 and i1 are homotopic and Pk is closed, it follows
that Pk0 and Pk1 are in the same cohomology class. The proof is complete.
82 IV. EXTERIOR DIFFERENTIAL CALCULUS IN RIEMANNIAN GEOMETRY
CHAPTER V
Summary. This chapter is not restricted to Riemannian geometry but rather a discus-
sion of analogues for a general vector bundle of the study in Chapters III and IV of the
tangent bundle of a Riemannian manifold. In Section 5.1 we introduce connections as
differential operators on sections of a vector bundle and define the curvature as in Section
3.1. We also give an analogue of the arguments in Chapter IV using exterior differential
forms, for this is essential in Section 5.2 when we define Chern classes, generalizing the
Pontrjagin classes in Section 4.4. A brief introduction to Lie groups is given in Section 5.3
in preparation for a study of principal bundles and associated vector bundles in Section
5.4.
5.1. Connections in vector bundles. The essential point in the expression for
the Riemann curvature tensor in (3.1.6) was the covariant differentiation ∇. Recall
that for a vector field u on M , that is, a section of the tangent bundle T M , ∇u is a
section of T ∗ M ⊗ T M such that for ϕ ∈ C ∞ (M )
(5.1.1) ∇(ϕu) = ϕ∇u + (dϕ) ⊗ u,
and u 7→ ∇u is a linear operator. Here T M can be replaced by any other vector bundle:
Definition 5.1.1. Let M be a C ∞ manifold and E a C ∞ vector bundle over M . Then
a linear map ∇ : C ∞ (M, E) → C ∞ (M, T ∗ M ⊗ E) is called a connection if (5.1.1) is
valid for u ∈ C ∞ (M, E) and ϕ ∈ C ∞ (M ).
In this definition E may be a real or a complex vector bundle provided that ϕ is
in the same category. From (5.1.1) it follows at once that ∇u = 0 in a neighborhood
of x if u = 0 in a neighborhood of x, for we can then choose ϕ ∈ C ∞ (M ) equal to 0
in a neighborhood of x so that u = ϕu. Thus ∇ is a local operator. If e1 , . . . , eN ∈
C ∞ (M, E) form a basis in Ex for all x in an open set ω ⊂ M , then
X X X
∇( ϕj ej ) = ϕj ∇ej + dϕj ⊗ ej .
(We could also have taken (5.1.2) as our definition for it is obvious that this property
is invariant under change of local coordinates and bases for E and F .) If all Pjk are
of order m, then P is said to be of order m, and the principal part is defined as the
matrix of principal parts of Pjk , obtained when (∂1 , . . . , ∂n ) is replaced by (ξ1 , . . . , ξn )
in the terms of order m. An invariant formulation is that if 0 6= ξ ∈ Tx∗ and we choose
ϕ ∈ C ∞ with dϕ = ξ at x, then
In fact, using the local representation (5.1.2) we see immediately that this limit exists
and is linear in u(x) and a homogeneous polynomial of degree m in ξ. The invariance
of (5.1.3) shows that p(ξ) ∈ L(Ex , Fx ) is invariantly defined. In the particular case of
a first order operator, p is also linear in ξ, so p can be considered as a vector bundle
map from Tx∗ M ⊗ E to F , and then we have
A(ϕu) = ϕAu,
R∇ (X, Y )u(x) depends only on the values of X, Y and u at x, for the proof of this
fact for covariant differentiation only used that it is a connection. Thus (5.1.5) is at x
a skew symmetric bilinear form in the tangent vectors X(x), Y (x) and a linear form
in u(x) ∈ Ex , with values in Ex , so R∇ is a two form with values in Ex ⊗ Ex∗ .
CONNECTIONS IN VECTOR BUNDLES 85
Definition 5.1.3. The curvature R∇ of the connection ∇ in E is the two form with
values in E ⊗ E ∗ defined by (5.1.5).
In the spirit of Chapter IV we can also argue more elegantly by extending the
connection to forms, that is, sections of ∧p T ∗ M ⊗ E. Note that if u is such a section
and ψ is a scalar q form, then ψ ∧ u ∈ ∧q+p T ∗ M ⊗ E can be uniquely defined so that
the multiplication is bilinear and ψ ∧ (ϕ ⊗ v) = (ψ ∧ ϕ) ⊗ v, if v is a section of E and
ϕ is a scalar p form.
Proposition 5.1.4. Given a connection ∇ in the vector bundle E, there is a unique
differential operator d∇ : C ∞ (M, ∧p T ∗ M ⊗ E) → C ∞ (M, Λp+1 T ∗ M ⊗ E), for any p,
such that
(5.1.7) d∇ d∇ = R ∇ .
One can also argue as in Sections 4.1 and 4.2. Since we are only interested in forms
on the base manifold we shall study pullbacks of forms analogous to those studied
there. Note that if s is a local section of P (M ) and we write s∗ eα = Eα , which is a
local basis
P for T M , we saw in the proof of (4.1.13) that s∗ ωαβ = (∇Eα , Eβ ), that is,
∇Eα = (s∗ ωαβ )Eβ . We take this as our starting point now.
Let e1 , . . . , eN be a local basis for the vector bundle E, that is, sections of E over
an open set U which form a basis in the fiber Ex for every x ∈ U . Omitting the tensor
product sign ⊗, we can then write
N
X
(5.1.8) ∇ei = ωji ej , i = 1, . . . , N,
j=1
by repeating the proof of (4.2.6). With matrix notation this can be written more
succinctly
(5.1.11)′ dΩ = [Ω, ω],
where the matrix multiplication is made using the wedge product of course. (See also
the discussion following (5.3.8).)
The characterization of the covariant differentiation in Theorem 3.1.2 involved the
conditions (3.1.3) and (3.1.4). The first is only meaningful for the tangent bundle, but
the second has a general analogue:
Definition 5.1.5. A real (complex) vector bundle is called Euclidean (Hermitian) if in
each fiber there is given a Euclidean (Hermitian) scalar product (·, ·) such that (e, e)
is a C ∞ function on E.
CONNECTIONS IN VECTOR BUNDLES 87
when u and v are sections of E. For the corresponding curvature matrices in terms of
orthonormal bases this implies that Ωij = −Ωji , thus Ωij = −Ωji in the real case.
Proof. By the Gram-Schmidt orthogonalization procedure one can in some neighbor-
hood U of any point find an orthonormal basis for E, which means that the identifi-
cation of E over U with U × RN (resp. U × CN ) carries the Euclidean (Hermitian)
form in the bundle to the standard form in RN (in CN ). The trivial connection in
U × RN (in U × CN ) obviously has the property (5.1.12), so we have such a connection
in the restriction of the bundle E to U . Piecing together such local connections by a
partition of unity as in the proof of Proposition 5.1.2, we obtain a global connection
such that (5.1.12) holds.
If e1 , . . . , eN is a local orthonormal basis, it follows from (5.1.8) that
ωji = (∇ei , ej ).
N
X
ωij + ωji = 0, hence Ωji = −dωij − ωkj ∧ ωik = −Ωij ,
k=1
Proof. The uniqueness is obvious so it suffices to prove existence locally. Thus we may
assume that there is a global basis v1 , . . . , vN for sections of F . Every section s of
PN
E ⊗ F can then uniquely be written in the form s = 1 uj ⊗ vj , where uj is a section
of E. If (5.1.13) holds we must have
N
X N
X
E
∇s = (∇ uj ) ⊗ vj + uj ⊗ (∇F vj ),
1 1
PN
and we define ∇ now by this equation. If v = 1 ϕj vj with scalar ϕj ∈ C ∞ , and if u
is a section of E, we have by definition
X X
∇(u ⊗ v) = ∇E (ϕj u) ⊗ vj + ϕj u ⊗ (∇F vj )
X X X
= ∇E u ⊗ ϕj vj + dϕj (u ⊗ vj ) + u ⊗ ϕj (∇F vj ) = (∇E u) ⊗ v + u ⊗ (∇F v),
where we have used first that ∇E is a connection, then that ∇F is a connection. Thus
(5.1.13) is valid which shows that the construction is independent of the choices made.
That ∇ is a connection follows at once from (5.1.13) if we replace u by ϕu and use
that ∇E is a connection.
If E is a vector bundle with connection ∇ and E ∗ is the dual bundle, with fiber Ex∗
equal to the dual space of Ex , then there is a unique connection ∇∗ in E ∗ such that
for arbitrary sections u and v of E and E ∗ . We leave for the reader the simple verifi-
cation that (5.1.14) defines a connection ∇∗ .
5.2. Chern classes. If T is a linear transformation in a finite dimensional vector
space V , then the determinant of T can be defined as the determinant of the matrix
A representing T with respect to a coordinate system in V . This is independent of
the choice of coordinates, for if x′ = Bx, where B is an invertible matrix, is another
system of coordinates, then the matrix BAB −1 for T in the new coordinates has the
same determinant.
Now let ϕ be any polynomial in the vector space M(N ) of N ×N (complex) matrices
which is invariant in the sense that
for every B ∈ GL(N ), the group of invertible matrices in M(N ). This is of course
equivalent to
for all A, B ∈ M(N ). An example is ϕ(A) = det(I + λA) for an arbitrary λ, hence also
the coefficient of λk for any k. (Every invariant polynomial is in fact a polynomial in
these special ones, but we shall not give the proof here.)
CHERN CLASSES 89
Now let Ω be the matrix of curvature forms for a vector bundle E with fiber dimension
N , defined by means of a local basis. Since all two forms commute, we can form ϕ(Ω)
unambiguously, and since another choice of local basis just replaces Ω by B −1 ΩB for
some invertible matrix B (cf. Exercise 5.1.1), it follows that ϕ(Ω) is independent of
the choice of local basis, so we obtain a globally defined differential form for which we
shall also use the notation ϕ(Ω).
Theorem 5.2.2. If ϕ is an invariant polynomial in M(N ) then ϕ(Ω) is a closed
differential form, and its cohomology class does not depend on the choice of connection.
Proof. Locally we write
ϕ(Ω) = Φ(Ω, . . . , Ω)
where the curvature matrix Ω satisfies the Bianchi identity (5.1.11)′ . Since the forms
Ωij have even degree, we can apply d separately to each argument in the multilinear
form Φ. Two forms commute with arbitrary differential forms so we obtain using
(5.2.2)
dϕ(Ω) = mΦ(dΩ, Ω . . . , Ω) = mΦ([Ω, ω], Ω, . . . , Ω) = 0.
The proof that the cohomology class is independent of the choice of connection is
essentially the same as the proof of Theorem 4.4.2. We keep the notation used there.
If ∇0 and ∇1 are two connections for E over M , and Ee is the vector bundle E pulled
back to M × R by the projection M × R → M , then
is a connection for E
e on M × R. Note that if x ∈ M then the fiber of E e over (x, t) is
always equal to Ex , so the derivative of a section with respect to t is well defined. It
is now obvious that the matrices of curvature forms for the connections ∇0 and ∇1 on
M are pullbacks from M to M × R by the homotopic maps i0 and i1 of the matrices of
curvature forms for ∇,
e which proves that the cohomology class of ϕ(Ω) is independent
of the choice of connection.
Definition 5.2.3. If E is a complex vector bundle over M and Ω is the (local) matrix
of curvature forms with respect to a connection, then the cohomology class of
where I is the unit matrix, is called the total Chern class of E. It can be written
1 + c1 + c2 + . . . , where cj is of degree 2j and is called the jth Chern class.
If E is the complexification of the tangent bundle of a Riemannian manifold with
the covariant differentiation as connection, it is clear that (−1)j c2j is the Pontrjagin
class of degree 4j and that c2j+1 = 0. More generally, if E is the complexification of
a real vector bundle, it follows from Propositions 5.1.6 and 5.1.7 that we can choose
a connection such that Ωij = −Ωji is a real form. That was all we used to prove
that (4.4.2) vanishes for odd k. For a general complex vector bundle it follows from
Propositions 5.1.6 and 5.1.7 that we can choose a connection such that Ωij = −Ωji ,
which implies that complex conjugation preserves (5.2.3). Hence we obtain:
Proposition 5.2.4. The Chern classes of any complex vector bundle are real. For the
complexification of a real vector bundle the Chern classes cj vanish when j is odd.
This proposition motivates the factor i in (5.2.3). To explain the factor 2π we shall
now discuss the case of a complex line bundle L in a manifold M . We introduce a
Hermitian structure in L and choose an open covering {Uj } such that L is trivial, that
is, has a section sj with norm 1 at every point in Uj . An arbitrary section can then in
Uj be written in the form uj sj . In Uj ∩ Uk we have uj sj = uk sk then, hence writing
gjk sj = sk we obtain |gjk (x)| = 1, x ∈ Uj ∩ Uk , and
uj = gjk uk .
The functions gjk are the transition functions of L. With a connection ∇ in L satisfying
(5.1.12) we now set
∇sj = Γj sj
where iΓj is a real one form in Uj by (5.1.12). The curvature form is then
Ω = dΓj + Γj ∧ Γj = dΓj in Uj ,
so c1 is the cohomology class of the two form equal to idΓj /(2π) in Uj for every j.
Note that
Γk sk = ∇sk = ∇(gjk sj ) = gjk Γj sj + sj dgjk in Uj ∩ Uk ,
which means that
Γk = Γj + dgjk /gjk , dΓj = dΓk in Uj ∩ Uk .
If the cohomology class is equal to 0, then there is a global one form Γ in M such
that d(Γ − Γj ) = 0 in Uj for every j. If Uj has been chosen simply connected, it follows
that we can find a function ϕj in Uj such that Γ − Γj = dϕj in Uj ; we can choose Γ
and ϕj purely imaginary. Since Γk − Γj = dϕj − dϕk in Uj ∩ Uk , this means that
dgjk /gjk = dϕj − dϕk ,
so cjk = gjk eϕk −ϕj is locally constant, that is, constant in each component of Uj ∩ Uk .
If we replace the sections sj by sj eϕj , we get new transition functions which are locally
constant. But in a line bundle with locally constant transition functions we can define
a connection ∇ by ∇sj = 0 in Uj , and then all Γj become 0. Thus we have proved:
LIE GROUPS 91
Proposition 5.2.5. The Chern class of a line bundle vanishes if and only if it can be
defined by locally constant transition functions.
We shall now consider the example where M is the sphere S 2 of dimension 2. If L
is a complex line bundle on S 2 , we can find a non-zero section s1 in a neighborhood
U1 of the upper hemisphere and another s2 in a neighborhood U2 of the lower one. We
write as before ∇sj = Γj sj and recall that the Chern form is idΓj /2π in Uj ,
Γ1 = Γ2 + dg21 /g21 in U1 ∩ U2 .
the last three integrals taken over the equator. This is the winding number of g12 on
the
R equator, thus an integer. In this example the effect of the factor 2π is therefore that
c can precisely take integer values, so that the Chern class is an integer cohomology
class. It would take us too far into a discussion of cohomology theory to give a general
form of this result here; the example should suffice as a motivation.
5.3. Lie groups. In Section 4.1 we worked systematically with forms on the or-
thonormal frame bundle of a Riemannian manifold. Each fiber becomes isomorphic to
the orthogonal group O(n) when a point in it is distinguished. In Section 5.4 we shall
clarify the results by putting them in a more general context. In order to cover the
results in Sections 5.1 and 5.2 we shall also need other groups than O(n). Subgroups of
the full linear groups suffice for our purposes, but the arguments become conceptually
more transparent in an abstract setting where O(n) is replaced by an arbitrary Lie
group, so we shall give a brief introduction to this concept here.
Definition 5.3.1. A group G which has also the structure of a C ∞ manifold is called
a Lie group if
G × G ∋ (a, b) 7→ ab−1 ∈ G
is a C ∞ map.
In particular, the hypothesis implies that b 7→ b−1 and (a, b) 7→ ab are C ∞ maps. It
would suffice to make the latter assumption:
Exercise 5.3.1. Show that if G is a group which is also a C ∞ manifold, then G is a
Lie group if G × G ∋ (a, b) 7→ ab ∈ G is a C ∞ map.
Let La be the left translation La x = ax, a, x ∈ G, which maps the identity e ∈ G
to a, thus L′a : Te → Ta . If X0 ∈ Te , we define a C ∞ vector field X on G by
X(a) = L′a X0 ∈ Ta .
92 V. CONNECTIONS, CURVATURES AND CHERN CLASSES
For arbitrary a, b ∈ G we have L′a X(b) = L′a L′b X0 = L′ab X0 = X(ab) because La Lb =
Lab by the associative law. Thus X is a left invariant vector field. If ω0 ∈ Te∗ we define
similarly a C ∞ one form on G so that ω is equal to ω0 at the identity and ω(X) is a
constant if X is a left invariant vector field. Then we have L∗a ω = ω for every a ∈ G,
and this left invariance together with the value ω0 at e determine ω.
Let X1 , . . . , Xn be left invariant vector fields such that X1 (e), . . . , Xn (e) form a
basis for Te , and let ω 1 , . . . , ω n be the left invariant one forms which are biorthogonal
at e and hence at any a ∈ G. Then we have
n
X
i
(5.3.1) dω = 1
2 cijk ω j ∧ ω k ,
j,k=1
for some constants cijk with cijk = −cikj . In fact, this is obviously true at the identity,
and both sides are left invariant, hence equal everywhere. The constants cijk are called
the structural constants of the Lie group. (They depend in an obvious way on the
choice of basis in Te∗ .) We can also express (5.3.1) in terms of the left invariant vector
fields if we use that by (C.4)
Proposition 5.3.2. In the tangent space g of a Lie group G at the identity, there is
a natural bilinear antisymmetric product g × g ∋ X, Y 7→ [X, Y ] defined by taking the
commutator of the left invariant vector fields extending X and Y . We have the Jacobi
identity
This implies that ϕ(0) = e, and differentiation with respect to t gives when t = 0, if
we write ϕ(s)ϕ(t) = Lϕ(s) ϕ(t),
where X0 = ϕ′ (0) and X is the left invariant vector field with X(e) = X0 . This
system of differential equations in the local coordinates of G at e has a unique solution
with ϕ(0) = e, when |s| < 2δ, say. We claim that (5.3.3) is then valid if |s| < δ and
|t| < δ. Fix s and denote the left (right) side of (5.3.3) by ϕ1 (t) (resp. ϕ2 (t)). Then
ϕ1 (0) = ϕ2 (0) = ϕ(s), and
ϕ′1 (t) = X(ϕ1 (t)), ϕ′2 (t) = L′ϕ(s) X(ϕ(t)) = X(ϕ(s)ϕ(t)) = X(ϕ2 (t)),
so the uniqueness theorems for ordinary differential equations show that ϕ1 (t) = ϕ2 (t)
when |t| < δ. Now we can define ϕ(s) uniquely for arbitrary s by setting
ϕ(s) = ϕ(s/N )N
where N is a positive integer so large that |s/N | < δ. This does not depend on N ,
for ϕ(s/N ) = ϕ(s/N M )M for every positive integer M , hence ϕ(s/N )N = ϕ(s/M )M
if |s/M | < δ too. It is clear that (5.3.3) is valid for all s, t.
Proposition 5.3.3. There is a unique C ∞ map exp from the Lie algebra g of G to
G such that the differential at the origin is the identity g → Te (G) = g and R ∋ t 7→
exp(tX) is a one parameter subgroup for every X ∈ g. For X, Y in a neighborhood
of 0 in g we have exp X exp Y = exp ϕ(X, Y ) where ϕ(X, Y ) is a C ∞ function with
values in g defined in a neighborhood of 0 in g × g such that
(5.3.4) ϕ(X, Y ) = X + Y + 12 [X, Y ] + O(|X||Y |(|X| + |Y |)).
Proof. Let X1 , . . . , Xn be P
a basis for the left invariant vector fields. The one parameter
subgroup with derivative aj Xj (e) at 0 is the solution of the Cauchy problem
n
X
dϕ(t)/dt = aj Xj (ϕ(t)), ϕ(0) = e.
1
There is a unique solution for |t| < δ if |a| ≤ 1. It is a C ∞ function of a and t depending
only on ta which we define to be exp(ta). The extension of the map is done as before
by writing exp X = (exp(X/N ))N for large N .
By the implicit function theorem the exponential map is a diffeomorphism of a
neighborhood of the origin in g on a neighborhood of the identity in G. The function
ϕ is therefore well defined as the composition of exp X exp Y with the inverse of the
exponential map, ϕ ∈ C ∞ , and ϕ(sX, tX) = (s + t)X for small sX, tX. Hence it
follows from Taylor’s formula applied first in X, then in Y , that
ZZ
ϕ(X, Y ) − X − Y = ϕ′′XY (sX, tY )(X, Y ) ds dt
0≤s,t≤1
= B(X, Y ) + O(|X||Y |(|X| + |Y |))
94 V. CONNECTIONS, CURVATURES AND CHERN CLASSES
where B(X, Y ) is a bilinear g valued form with B(X, X) = 0, hence skew symmetric.
If X
e is the left invariant vector field equal to X at the identity, pushed back to g from
G by the inverse of the exponential map, then
X(Z)
e = L′Z (0)X, LZ = ϕ(Z, ·), hence
X(Z)
e = X + B(Z, X) + O(|Z|2 ), as Z → 0.
If Ye is a left invariant vector field equal to Y at 0, then we have a similar formula for
Ye , hence
[X,
e Ye ](0) = B(X, Y ) − B(Y, X) = 2B(X, Y ).
∞
X X
ϕ(X, Y ) = (−1)ν+1 ν −1 (ad X)α1 (ad Y )β1 . . . (ad X)αν (ad Y )βν −1 Y /cαβ ,
ν=1 αi +βi 6=0
Xν ν
Y
cαβ = (αi + βi ) αi !βi !,
i=1 i=1
and using (5.3.4) one can also show that there is a local Lie group with any given Lie
algebra. We shall not prove the Campbell-Hausdorff formula here but shall give another
proof that the group operation is uniquely determined near e by the Lie algebra.
Introducing in a neighborhood of the identity in G the canonical coordinates in g
provided by the exponential map has the advantage that the one parameter groups
become rays through the origin. Identifying g with Rn , we denote by X1 , . . . , Xn the
left invariant vector fields which are equal to the unit vectors in Rn at the origin. Then
we know that for a ∈ Rn , s ∈ R and |a||s| small enough the map s 7→ sa is a one
parameter subgroup. P The derivative is a, and the left
P invariant vector field with this
j j
value at the origin is a Xj , so it follows that a = a Xj (sa), that is,
X
(5.3.6) x= xj Xj (x), |x| < δ.
P j
In the left-hand side x denotes of course the radial vector field ̺ = x ∂/∂xj . Taking
the scalar product with the left invariant forms ω k gives h̺, ω k i = xk . We shall derive
equations for the determination of ω i by writing down (5.3.1) explicitly.
differential P
With ω i = ωνi dxν the equation means that
n
X
(5.3.1)′′ ∂ωµi /∂xν − ∂ωνi /∂xµ = cijk ωνj ωµk , i, ν, µ = 1, . . . , n,
j,k=1
X
(̺ + 1)ωµi = δµi + cijk xj ωµk .
if Xe is the left invariant vector field equal to X at the identity. This has a unique
solution which is analytic in a and tX, when a is in a neighborhood of e in G and tX
is in a neighborhood of 0 in g. The proof that the structure constants determine the
multiplication uniquely in a neighborhood of the identity is now complete. Note that it
would have been enough to assume that G is a C 3 manifold with C 3 group operations
96 V. CONNECTIONS, CURVATURES AND CHERN CLASSES
(5.3.7) ω(L′a X) = X, X ∈ g = Te G.
(5.3.1)′′′ dω = − 12 [ω ∧ ω],
To justify the notation we note that in general, if ω1 , ω2 are differential forms with
values in finite dimensional vector spaces W1 and W2 , then we can define ω1 ∧ ω2
uniquely as a form with values in the tensor product W1 ⊗ W2 so that it is equal to
(w1 ⊗ w2 )σ1 ∧ σ2 if ωj = wj σj , j = 1, 2, with wj ∈ Wj and scalar differential forms σj .
In fact, this expression is bilinear in w1 , σ1 and in w2 , σ2 , and the space of forms with
values in Wj is the tensor product of Wj and the space of scalar forms. If we have
a bilinear map β : W1 × W2 → W3 , a third vector space, then composition with the
corresponding map β̃ : W1 ⊗ W2 → W3 gives a form β̃ω1 ∧ ω2 with values in W3 . In
particular, if W1 = W2 = g and β is the Lie bracket we get precisely the definition of
[ω ∧ ω] in (5.3.8). Of course the equation (5.3.1)′′′ has the same contents as (5.3.1) or
(5.3.1)′′ but it is much more compact thanks to the coordinate free notation.
In g there is a natural symmetric bilinear form
gij = ϕi ϕ−1
j
The bundle E can be reconstructed using the system of transition functions gij . In
e be the set of all (i, x, w) ∈ I × M × CN with x ∈ Ui and the equivalence
fact, let E
relation
The reflexivity and symmetry of this relation follow from the first part of (5.4.1), and
the second part of (5.4.1) gives the transitivity. For any given family of transition func-
tions with values in GL(N, C) satisfying (5.4.1), it is clear that the space of equivalence
classes of E,
e with the projection induced by the map E e ∋ (i, x, w) → x ∈ M is a C ∞
complex vector bundle of fiber dimension N . (The linear structure is inherited from
CN of course.) If the transition matrices are obtained from a given vector bundle E
as above, it is also clear that we get back an isomorphic bundle.
For a real vector bundle there is no real change in the argument except that the
transition functions gij will take their values in the smaller group GL(N, R). The
group can be reduced further if we recall that by Proposition 5.1.6 it is always possible
to introduce a Hermitian (Euclidean) structure in a complex (real) vector bundle. The
local trivializations can then be chosen so that they respect the structure, for the local
frames in an arbitrary local trivialization can be orthonormalized using the Gram-
Schmidt procedure. If only such trivializations are used, then the transition functions
take their values in U(N ) (resp. O(N )). If we have a real vector bundle which is
oriented, that is, each fiber is oriented and there exist local trivializations respecting
the orientation, then gij takes its values in SO(N ).
To avoid looking at a large number of cases we now consider an arbitrary Lie group
G. Assume that we have an open covering {Ui }i∈I of the C ∞ manifold M and C ∞
transition functions gij : Ui ∩ Uj → G satisfying (5.4.1). Let ̺ be any representation
of G in a finite dimensional vector space F , that is, ̺(g) is for every g ∈ G a linear
transformation in F such that
Following the reconstruction of E above we now form the set of all (i, x, w) ∈ I ×M ×F
with x ∈ Ui . Then
w′′ = ̺(gkj )w′ = ̺(gkj )(̺(gji )w) = ̺(gkj gji )w = ̺(gki )w,
so (i, x, w) ∼ (k, x, w′′ ). The set of equivalence classes is a vector bundle F̺ with fiber
dimension dim F and the projection induced by the map (i, x, w) 7→ x ∈ M . The
restriction π −1 (Ui ) is identified with Ui × F by the map
ϕ(pg) = ϕ(p)g, p ∈ π −1 (U ), g ∈ G,
For the principal bundle constructed above using the transition functions gij we
have obvious trivializations of π −1 (Ui ) which give back the transition functions. Now
we can pass directly from the principal bundle to any one of the bundles F̺ ; they are
100 V. CONNECTIONS, CURVATURES AND CHERN CLASSES
said to be associated to the principal bundle. To do so we note that G acts to the right
(see Definition 5.4.1, (i)) on P × F by
In fact,
ψab (p, w) = (p(ab), ̺((ab)−1w) = ((pa)b, ̺(b−1)(̺(a−1 w)) = ψb (ψa (p, w)),
which is just the definition of action to the right. We define two elements in P × F to
be equivalent if they are carried into each other by some element in G, thus (pa, w) ∼
(p, ̺(a)w), and form the quotient (P × F )/G by this equivalence relation. The map
is a bijection, and each equivalence class under the right action of G contains precisely
one element with g = e; (x, i, g, w) is equivalent to (x, i, e, ̺(g)w) under the G action.
If x ∈ Uj also, then (x, i, e, w) and (x, j, gji, w) define the same element in P × F , and
under the G action the latter is equivalent to (x, j, e, ̺(gji)w). Thus our trivializations
of (P ×F )/G over Uj and Ui are related by the transition function ̺(gji ), which means
that (P × F )/G is isomorphic to F̺ . We have now proved:
Proposition 5.4.2. If P is a principal G bundle over M , and ̺ is a representation
of G on a vector space F , then the quotient of P × F by the right G action defined by
is a vector bundle with fiber dimension dim F , said to be associated with P and ̺,
denoted by P ×̺ F .
Let us now as an example consider the case of a real vector bundle E of fiber
dimension N over M . Let P be the frame bundle over M with fiber Px consisting
of all bases (t1 , . . . , tN ) for Ex . The full linear group GL(N, R) acts to the right
on PP , mapping (t1 , . . . , tN ) ∈ Px and A = (ajk ) ∈ GL(N, R) to (t′1 , . . . , t′N ), where
′
tj = akj tk , and it is clear that P is a principal GL(N, R) bundle. The vector bundle
associated to P and the natural action of GL(N, R) in RN is isomorphic to E. In
fact, the map
N
X
N 1 N
Px × R ∋ (t1 , . . . , tN , x , . . . , x ) 7→ x j tj ∈ E x
1
N
X N
X N
X
xj t′j = xj akj tk = (Ax)k tk .
j=1 j,k=1 k=1
PRINCIPAL AND ASSOCIATED BUNDLES 101
PN
as in (5.1.8). For a general section u = 1 ui ei we have
N
X N
X N
X N
X
∇u = dui ei + ui ωji ej = (dui + uj ωij )ei .
i=1 i,j=1 i=1 j=1
The composition ωp = γ(p)−1 vp of the projection (5.4.6) with the inverse γ(p)−1 is
thus a linear map from Tp P to g, that is, a one form on P with values in g, which
vanishes precisely in the horizontal spaces Hp . To define it we have used part (i) of
Definition 5.4.3, but we must also express condition (ii) there in terms of ω. To do so
we note that the condition means that
vpa = Ra′ vp (Ra′ )−1 , hence ωpa = γ(pa)−1 Ra′ vp (Ra′ )−1 = γ(pa)−1 Ra′ γ(p)ωp (Ra′ )−1 .
If X, Y ∈ g and
Ra′ γ(p)X = γ(pa)Y
then t 7→ p exp(tX)a and t 7→ pa exp(tY ) have the same derivative for t = 0, hence
t 7→ p exp(tX) and t 7→ pa exp(tY )a−1 = p exp(Ad a (tY )) have the same derivatives for
t = 0. By the injectivity of γ(p) we conclude that X = Ad a Y , that is, Y = Ad(a−1 )X,
so
ωpa Ra′ = Ad(a−1 )ωp .
The left-hand side is the pullback Ra∗ ω in Tp . We have now proved the first half of the
following:
PRINCIPAL AND ASSOCIATED BUNDLES 103
(ht, dwi is well defined since w is a function with values in a vector space.) If p′ (x)t
is not horizontal, we replace p and w by p̃ = p exp(−ϕ) and w̃ = ̺(exp ϕ)w where
ϕ : U → g, ϕ(x) = 0. If ht, ϕ′ (x)i = X ∈ g, then
where ̺′ = ̺′e is a linear map from g to linear maps in F . Thus p̃′ (x)t ∈ Hp(x) if and
only if vp p′ (x)t = γ(p)X, that is, X = hp′ (x)t, ωp(x) i = ht, p∗ ωi, so (5.4.10) leads to
∼
(5.4.11) ∇t s(x) = p(x), ht, dw(x)i + ̺′ (ht, p∗ ωi)w(x) .
The preceding arguments show that (5.4.11) is the only definition satisfying (5.4.10),
but it remains to show that it does not depend on p(x). To do so we must prove that
(5.4.11) does not change if p is replaced by pa = Ra p and w is replaced by ̺(a−1 )w
with a constant a ∈ G. If we do that then ht, p∗ ωi is replaced by
which is precisely ̺(a−1 ) times the value in (5.4.11). The first component is replaced
by p(x)a, so the equivalence class under the G action is unchanged. Hence (5.4.11)
gives a unique definition of ∇t s(x), t ∈ Tx M , when s is a section of F̺ . It is obvious
that ∇t s(x) is linear in t and linear in s, and we have for ϕ ∈ C ∞
so ∇ is a connection in F̺ .
To motivate the definition of the curvature form of a principal bundle with connec-
tion we shall write the preceding formulas explicitly in the case of the frame bundle P
of a vector bundle, which is a principal G = GL(N, R) bundle. As above we choose a
basis e1 , . . . , eN for the sections of E in a coordinate patch U ⊂ M and define forms
ωij by (5.1.8). As coordinates in P |U we use the local coordinates x1 , . . . , xn in U and
the components A = (aij ) ∈ G of a general frame
X
Ej = aij ei , j = 1, . . . , N.
The Lie algebra g is identified with the space M(N ) of N × N matrices, which is
mapped bijectively to the tangent to the fiber of P at E1 , . . . , EN by
d
M(N ) ∋ X 7→ (AetX )|t=0 = AX.
dt
The horizontal space is defined by the equations (5.4.5), hence
N
X
(Aω)ij = daij + ωik akj
k=1
if ω denotes the connection form in Definition 5.1.3, with values in M(N ), for both
sides vanish in the horizontal space and are equal in the fiber direction. If ω e denotes
the matrix (ωij ), this means that the connection form ω satisfies
Aω = dA + ω e A,
or after insertion of dA = Aω − ω e A
The matrix of two forms dω e + ω e ∧ ω e in the right-hand side of (5.4.12) is the matrix
Ω in (5.1.10).
It is now clear that we should define the curvature form Ω for a principal G bundle
with g valued connection form ω by
(5.4.13) Ω = dω + 12 [ω ∧ ω].
Here the right-hand side is defined as explained in Section 5.3, using the bracket in
g. When G = GL(N, R), g = M(N ), this is equal to ω ∧ ω. We shall now prove a
general analogue of (5.4.12) where transformation by the matrix A is replaced by the
adjoint action.
Theorem 5.4.5. Let P be a principal G bundle with connection form ω. The curva-
ture form Ω defined by (5.4.13) is then for every p ∈ P an antisymmetric bilinear form
in Tp P/Tp0 P with values in g, and
Ω(t1 , t2 ) = 0, if t ∈ Hp P, t2 ∈ Tp0 P.
−h[h, X],
e ωi + hhX,
e ωi − Xhh,
e ωi = 0.
106 V. CONNECTIONS, CURVATURES AND CHERN CLASSES
Theorem 5.4.5 means that at every p ∈ P we can regard the curvature form Ω as the
pullback of a two form with values in g at the base point π(p), and this form transforms
by Ad(a−1 ) under right translation by a. Thus we have obtained an extension of
(5.4.12) with considerable precision added when we can reduce the group G to a
subgroup.
Exercise 5.4.1. Show using (5.4.14) that if ̺ : G → F is a representation and ̺′ = ̺′e ,
then ̺′ (p∗ Ω) is a two form which at x ∈ M only depends on p(x), and that
which means that ̺′ (p∗ Ω) defines a two form on M with values in End(P ×̺ F ).
For a Riemannian manifold of dimension n the tangent bundle is associated to the
principal O(n) bundle of orthonormal frames. So are all tensor bundles and more
generally the subbundles associated to O(n) invariant subspaces of tensor products
of a number of factors Rn . All the covariant differentiations discussed in Section
3.1.5, including that in Exercise 3.1.5, are therefore induced by the same Riemannian
connection in the orthonormal frame bundle, and its curvature form is essentially
equivalent to the Riemann curvature tensor by (4.2.5).
CHAPTER VI
6.1. Metric elliptic operators. Already in (3.4.4) and Proposition 3.4.6 we de-
fined the Laplace operator acting on a scalar function P
on a Riemannian manifold. It
is a second order operator with principal symbol ξ 7→ g jk ξj ξk . (See (5.1.3) for the
definition of the principal symbol.) However, in geometrical questions the most com-
mon objects are sections of vector bundles, such as tensor bundles, rather than scalar
functions. We shall therefore consider a more general class of operators here.
Definition 6.1.1. If E is a C ∞ vector bundle on a C ∞ manifold M , then a second order
differential operator P : C ∞ (M, E) → C ∞ (M, E) is said to be metric if the principal
symbol p(x, ξ) : Ex → Ex is a positive multiple p0 (x, ξ)IEx of the identity in Ex for
every ξ ∈ Tx M .
The positive definite quadratic form p0 defines a symmetric tensor of type 2, 0,
and its dual tensor, of type 0, 2, is a Riemannian metric. With local coordinates
x1 , . . . , xn P
in M and corresponding coordinates x1 , . . . , xn , ξ1 , .P
. . , ξn in T ∗ M , we have
p0 (x, ξ) = g jk (x)ξj ξk , and the Riemannian metric is ds2 = gjk (x)dxj dxk , where
jk −1
(gjk ) = (g ) . Thus a metric operator is associated with a unique Riemannian
metric, hence the term “metric operator” (which is used by some authors but not
universally).
Typeset by AMS-TEX
107
108 LINEAR DIFFERENTIAL OPERATORS IN RIEMANNIAN GEOMETRY
e E ∇E (ϕu) = ∇
∇ e E (ϕ∇E u + (dϕ)u) = ϕ∇
e E ∇E u + 2((dϕ)∇E u)s + (∇dϕ)u,
hence
Now suppose that P is any metric operator in E, introduce the associated Riemannian
metric, and define for t ∈ Tx M
n
X n
X
Pu = g ij ∂i ∂j u + cj ∂j u + cu,
i,j=1 j=1
n n n
− 21 1
X X X
ij i
g ∂i ϕ∂j u + 1
2 (c − g ∂j (g 2 g ij ))∂i ϕ u
i,j=1 i=1 j=1
which is a linear function of ϕ′ , hence of t = ϕ′♯ , with principal part equal to the
identity. If we write
X
P = g ij (∇
e E ∇E u)ij + Q,
it follows from (6.1.2) and the way we have defined ∇E that Q(ϕu) = ϕQu, ϕ ∈ C ∞ ,
so Q is a differential operator of order 0. We have proved:
METRIC ELLIPTIC OPERATORS 109
n n
− 12 1
X X
jk
(6.1.5) Pu = g ∂j (g g ∂k u) +
2 cj ∂j u + cu,
j,k=1 j=1
n
X
E
ht, ∇ ui = 1
2
(P (ϕu) − ϕP u − (∆ϕ)u) = (dϕ, du) + 1
2
cj (∂j ϕ)u,
j=1
hence
n
X
E
∇ u = du + 1
2 cj gjk dxk u.
j,k=1
Pn
If ̺ = 1 xj ∂j is the radial vector field, we obtain
n
X n
X
(6.1.6) ∇E
̺ u = ̺u +
1
2 ck xk u, ck = cj gjk .
k=1 j=1
110 LINEAR DIFFERENTIAL OPERATORS IN RIEMANNIAN GEOMETRY
Lemma 6.1.3. In a neighborhood of the origin in the local coordinates one can choose
the local frame e1 , . . . , eN for E so that ∇E
̺ eν = 0, ν = 1, . . . , N , where ̺ is the
radial vector field. Such a frame is called synchronous; it is uniquely determined by
e1 (0), . . . , eN (0).
Proof. By (6.1.6) we want all eν to be solutions of the system of differential equations
n
X n
X
xj ∂ j u + 1
2 ck xk u = 0.
j=1 k=1
We shall prove that there is a unique solution with given value at the origin. At the
origin the equations imply that ∂j u(0) = 21 cj (0)u(0), so we set
n
X
v= exp( 21 cj (0)xj )u.
j=1
Then
X n
X n
X n
X
j j j
exp( 21 1
cj (0)xj u + O(|x|2 )u ,
x ∂j v = cj (0)x ) x ∂j u + 2
j=1 j=1 j=1
where R vanishes of second order at the origin. But this equation is of the form
discussed in the proof of Lemma 3.3.4, which completes the proof.
Remark. We have assumed aboved that the vector bundle is real, but there is no
difference in statements or proofs if it is complex.
In the following construction we shall use geodesic coordinates and a frame for E
satisfying the conclusions in Lemma 6.1.3, that is,
n
X n
X n
X
j k j k
(6.1.7) gjk (x)c (x)x = gjk (0)c (x)x = ck (x)xk = 0.
j,k=1 j,k=1 k=1
Lemma 6.1.4. If z ∈ C\R+ then there is for every integer ν ≥ 0 a unique distribution
Fν ∈ S ′ (Rn ) such that
(6.1.8) (∆ + z)ν+1 Fν = δ0 .
Proof. The equation (6.1.8) means that (z − |ξ|2 )ν+1 Fbν = 1 if Fbν is the Fourier trans-
form of Fν . Hence
Z
Fν (x) = (2π) −n
eihx,ξi (z − |ξ|2 )−ν−1 dξ
with the integral taken in the sense of distribution theory. Thus xβ ∂ α Fν is the inverse
Fourier transform of (i∂ξ )β (iξ)α (z − |ξ|2 )−ν−1 which is integrable if |α| − |β| − 2(ν +
1) < −n, so xβ ∂ α Fν is then a bounded continuous function. (6.1.9) is an obvious
consequence of the fact that Fν is the unique solution of (6.1.8). The Fourier transform
of ∂Fν /∂xj is
iξj (z − |ξ|2 )−ν−1 = i∂ξj (z − |ξ|2 )−ν /2ν,
which proves (6.1.10) and ends the proof of the lemma.
1
The uniqueness implies that Fν is a function of |x| = (x21 + · · · + x2n ) 2 , and it would
be easy to express Fν in terms of Bessel functions. We shall not do so but note that
Fν (x) = O(|x|2(ν+1)−n log |x|) as x → 0, where the logarithm can be dropped if n is
odd or 2(ν + 1) < n. This follows from the corresponding well known facts for the
Laplacian. In particular, Fν ∈ L1 .
If f ∈ D ′ (R), we have for x 6= 0
n
X n
X
jk 2 2
′
g (x)∂k f (|x| ) = 2f (|x| ) g jk (x)xk
k=1 k=1
= 2f ′ (|x|2 )xj = ∂j f (|x|2 ), j = 1, . . . , n,
Pn
for k=1 gjk (x)xk = xj since the coordinates are geodesic. When f (|x|2 ) is replaced
by Fν the first and last expressions are equal as distributions in Rn since both are
locally integrable functions. Hence it follows from (6.1.9), (6.1.10) that
n
X
g jk (x)∂k Fν = (2ν)−1 xj Fν−1 , if ν > 0;
k=1
n
X
∂j (g jk (x)∂k Fν ) + zFν = Fν−1 , if ν ≥ 0,
j,k=1
112 LINEAR DIFFERENTIAL OPERATORS IN RIEMANNIAN GEOMETRY
where F0 = f (|x|2 ). Here we have used (6.1.7) and introduced the notation
1 1
X
(6.1.12) h = 12 xj g − 2 ∂ j g 2 .
here the undefined term P u−1 should be omitted when ν = 0. These equations are
chosen so that the terms involving F0 , . . . , Fµ−1 drop out. Note that the error term
(P uµ )Fµ in the right-hand side of (6.1.13) is as smooth as desired if µ is chosen large.
The first equation (6.1.14) can be integrated explicitly, for
n n
1 1 1 1
X X
(6.1.12)′ h= 1
2
xj g − 2 ∂ j g 2 = g − 4 xj ∂j (g 4 ).
j=1 j=1
and that the other equations (6.1.13) can be written in the form
n
1 1
X
(6.1.14) ′
(ν + xj ∂j )(g 4 uν ) + νg 4 P uν−1 = 0, ν = 1, . . . , µ.
j=1
Thus the equations (6.1.14) have a unique solution when u0 (0) is prescribed.
The functions uν obtained depend linearly on u0 (0) = e(0), so we can write them
in the form ũν (x)e(0), where ũ(x) is a N × N matrix. The corresponding linear trans-
formation from E0 to Ex , both of which have been identified with CN , is independent
of the choice of basis made. In fact, by the uniqueness in Lemma 6.1.3 the basis is
uniquely determined up to a linear transformation independent of x. From now on we
drop tilde from the notation, so uν (x) ∈ L(E0 , Ex ) ∼
= Ex ⊗ E0∗ and u0 (0) is the identity
in E0 . Note that uν is independent of z but the distributions Fν depend on z.
For every y ∈ M the exponential map Ty M ∋ X 7→ expy (X) at y gives geodesic
coordinates centered at y when we introduce an orthonormal basis in Ty M . Using
Lemma 6.1.3 we then construct a synchronous local frame for E, so that P can be
written in the form (6.1.5) with (6.1.7) fulfilled, in a neighborhood of the origin. The
functions uν obtained depend of course in a C ∞ fashion on the parameters y. Now
the distributions Fν (X) pull back to distributions Fν (x, y) defined in a neighborhood
of the diagonal in M × M , which only depend on the geodesic distance between x and
y, and are locally integrable in y for fixed x, hence also in x for fixed y. We multiply
the functions uν by a C ∞ cutoff function which is 1 near the center of the geodesic
coordinates and has support in a small neighborhood and pull them back to M × M .
Since Fν (X) ∈ C ∞ when X ∈ Rn \ {0}, this does not introduce any new singularities
in (6.1.13). We have then obtained C ∞ functions uν ∈ C ∞ (M × M, E ⊠ E ∗ ) where
E ⊠ E ∗ is the vector bundle with fiber Ex ⊗ Ey∗ at (x, y) ∈ M × M , such that Fν (x, y)
is defined in a neighborhood of supp uν and for every µ
µ
X
(Px + z) uν Fν − δdiag ∈ C 2µ+1−n (M × M, E ⊠ E ∗ ).
ν=0
Here δdiag is the kernel of the identity map in C ∞ (M, E), that is, the distribution in
D ′ (M × M, E ⊠ E ∗ ) such that
Z
δdiag (x, y)ϕ(y) dvol(y) = ϕ(x), ϕ ∈ C0∞ (M, E),
p
which means that in terms of local coordinates δdiag (x, y) is equal to δ(x − y)/ g(y)
times the identity matrix. We can choose F ∈ D ′ (M × M, E ⊠ E ∗ ) so that
µ−1
X
F− uν Fν ∈ C 2µ+1−n (M × M, E ⊠ E ∗ ), ∀µ.
ν=0
Kν contains an arbitrary
P∞ compact set in M × M when ν is large enough, this implies
that the series F = 0 (uν Fν − ψν ) converges in D ′ (M × M, E ⊠ E ∗ ), and F has the
required properties. Thus
If M is not compact it is useful to take the cutoffs in the definition so close to the
diagonal that the maps
F ϕ − Gϕ = (G(P + z) − S)F ϕ − G((P + z)F − R)ϕ = (GR − SF )ϕ, ϕ ∈ C0∞ (M, E),
PN ∗
If (λ 1 , . . . , λN ) is a normal of the linear space V , then 1 λj hj ∈ Ker P , so
λj hf, hj i = 0 if hf, Ker P ∗ i = 0. Hence (hf, h1 i, . . . , hf, hN i) ∈ V , which completes
P
the proof.
It is easy to give other spaces than C ∞ and D ′ , such as Sobolev spaces and Hölder
spaces, for which (ii) and (iii) in Theorem 6.1.5 are true, but we shall postpone intro-
ducing these spaces until we need them.
Theorem 6.1.5 has an important corollary for first order operators:
116 LINEAR DIFFERENTIAL OPERATORS IN RIEMANNIAN GEOMETRY
Conversely, if P is given and such a linear symbol q exists, then we can choose Q so
that the hypotheses in the theorem are fulfilled. When E and F have the same fiber
dimension, then (6.1.19) implies that p(ξ)q(ξ) = r(ξ)IFx , for p(ξ)q(ξ)p(ξ) = r(ξ)p(ξ)
and p(ξ) is invertible. Hence the roles of P and Q may be interchanged, so the equation
P u = f ∈ D ′ (M, F ) has a solution u ∈ D ′ (M, E) if and only if u is orthogonal to
Ker P ∗ , which is a finite dimensional subspace of C ∞ (M, F ∗ ); we have statements
analogous to (i) and (ii) for Q.
6.2. The exterior algebra of a Euclidean vector space. As a preparation for
the Hodge theory in Section 6.3 we shall discuss here some elementary algebraic aspects
of exterior differential forms. Let V be a vector space of dimension n < ∞ over R,
and denote its dual space by V ∗ . Recall that the space ∧p V ∗ of alternating p linear
forms on V is spanned by the forms θ1 ∧ · · · ∧ θp , θj ∈ V ∗ , defined by
In fact, if L is such a form and e1 , . . . , en is a basis for V , θ1 , . . . , θn the dual basis for
V ∗ , then
X X
L(v1 , . . . , vp ) = L( hv1 , θj1 iej1 , . . . , hvp , θjp iejp )
dethvk , θjl ipk,l=1 L(ej1 , . . . , ejp )/p!.
X X
= hv1 , θj1 i . . . hvp , θjp iL(ej1 , . . . , ejp ) =
THE EXTERIOR ALGEBRA OF A EUCLIDEAN VECTOR SPACE 117
Thus the forms (6.2.1) give a basis for ∧p V ∗ even if θ1 , . . . , θp are restricted to elements
in a fixed basis for V ∗ .
The forms
span ∧p V for reasons of symmetry. We have used the duality between ∧p V and ∧p V ∗
such that
(Many authors use another definition where the right-hand side is divided by p!, for
that is the duality inherited from the natural duality of the tensor products. See
Sternberg [1, p. 19] for a discussion of this point. Kobayashi and Nomizu [1] use the
division by p!, which should be kept in mind when comparing identities.) The existence
and uniqueness of a bilinear form on ∧p V × ∧p V ∗ such that (6.2.2) holds is obvious,
for there is a unique bilinear form such that (6.2.2) holds when vk are chosen in a fixed
basis for V and θj in a fixed basis for V ∗ , and in view of the multilinearity and skew
symmetry of the two sides they must then be equal for arbitrary v1 , . . . , vp , θ1 , . . . , θp .
Assume now that V is a Euclidean vector space with scalar product denoted by (·, ·).
The linear form v ♭ : V ∋ v ′ 7→ (v, v ′ ) is an element 6= 0 in V ∗ , so we have a bijection
V ∋ v 7→ v ♭ ∈ V ∗ with inverse denoted by ♯; these are the musical isomorphisms we
used already in Chapter III. In particular, this gives a norm in V ∗ , which is of course
the dual norm
kθk = sup |hv, θi|/kvk,
06=v∈V
where we have used the convention that (·, ·) denotes scalar product in a Euclidean
space while h·, ·i denotes the bilinear form in a space and its dual. The map θ 7→ θ ♯
extends to a map ∧p V ∗ ∋ ϕ 7→ ϕ♯ ∈ ∧p V , and we define
I are different and J is a permutation of them; in that case, we have the sign of the
permutation. This means that for every I we get p! equal non-zero terms ϕI ψJ hε♯I , εJ i
which proves (6.2.4). By (6.2.4) it is clear that (6.2.3) gives a Euclidean structure in
∧p V ∗ . We can suppress the factorials in (6.2.4) if we sum only for strictly increasing
P′
multiindices I, that is, i1 < · · · < ip ; when we do that we shall use the notation .
∗
Note that our definition (6.2.3) did not use a basis for V or V . If O is an orthogonal
transformation in V and we define
∗(εp+1 ∧ · · · ∧ εn ) = (−1)p(n−p) ε1 ∧ · · · ∧ εp ,
(6.2.8) ∗ ∗ ϕ = (−1)p(n−p) , ϕ ∈ ∧p .
THE EXTERIOR ALGEBRA OF A EUCLIDEAN VECTOR SPACE 119
This is independent of the choice of orientation, for changing the orientation means
replacing ∗ by −∗, which preserves any formula where there is an even number of
factors ∗.
It is often convenient to write (6.2.6) in a different way. If we replace ϕ by ∗ϕ, we
obtain
ε(ϕ, ψ) = (−1)p(n−p) (∗ϕ) ∧ ψ = ψ ∧ ∗ϕ,
hence
(6.2.6)′ ψ ∧ ∗ϕ = ε(ϕ, ψ), ϕ, ψ ∈ ∧n−p V ∗ .
Using this formula it is easy to write down the ∗ operator explicitly for an arbitrary
basis ε1 , . . . , εn in V ∗ . Set g jk = (εj , εk ), and let I, J be multiindices of length p, n − p,
such that I, J is a positive permutation of 1, . . . , n. If the basis ε1 , . . . , εn is positively
oriented, that is, ε = ε1 ∧ · · · ∧ εn /kε1 ∧ · · · ∧ εn k, then taking ψ = εI in (6.2.6)′ gives
(6.2.9)
′
X p 1
(∗ϕ)J = (ϕ, εI )/kε1 ∧ · · · ∧ εn k = |K| = pϕK det g kν Iµ ν,µ=1 / det(g νµ )nν,µ=1 2 .
2
If n is even, n = 2l, then ∗ maps ∧l V ∗ into itself, and the square is (−1)l . For a
positively oriented orthonormal basis θ1 , . . . , θ2l in V ∗ we have
2
∗(θ1 ∧ · · · ∧ θl + cθl+1 ∧ · · · ∧ θ2l ) = θl+1 ∧ · · · ∧ θ2l + (−1)l cθ1 ∧ · · · ∧ θl .
If l is even, then taking c = ±1 we see using such elements in ∧l V ∗ that ∧l V ∗ = ∧+ ⊕∧−
where ∧+ and ∧− are subspaces of the same dimension 21 2ll and ∗λ = ±λ when
λ ∈ ∧± . This is the general form of the decomposition we noticed in Lemma 2.3.3
when n = 4. If l is odd, we must take c = ±i, so it is the complexification ∧l VC∗ which
is the direct sum of the eigenspaces of ∗ with eigenvectors ±i.
To generalize this to ∧∗ V ∗ = ⊕np=0 ∧p V ∗ , still with n = 2l, we define
τ 2 ϕ = ip(p−1)+l+(2l−p)(2l−p−1)+l (−1)p(2l−p) ϕ.
The exponent of −1 here is congruent to p2 mod 2 and the exponent of i is 4l2 −
4lp + 2p2 ≡ 2p2 mod 4, so τ 2 ϕ = ϕ. If p 6= l it follows that ∧p VC∗ ⊕ ∧2l−p VC∗ is the
direct sum of the eigenspaces of τ corresponding to the eigenvalues ±1, and these must
have the same dimension 2l p ∗
p since none intersects ∧ VC ⊕ {0}. Summing up, we have
proved:
Proposition 6.2.1. If the oriented Euclidean vector space V is of even dimension 2l,
then the complexification ∧∗ VC∗ of ∧∗ V ∗ is the direct sum of the eigenspaces ∧± of the
operator τ defined by (6.2.10) corresponding to the eigenvalues ±1. ∧+ and ∧− have
the same dimension, and ∧± is the direct sum of its intersections with ∧p VC∗ +∧2l−p VC∗
when 0 ≤ p ≤ l.
120 LINEAR DIFFERENTIAL OPERATORS IN RIEMANNIAN GEOMETRY
∆ = dd∗ + d∗ d : λp → λp .
so d∗ ψ = dψ = 0.
Definition 6.3.2. The differential forms satisfying the equation ∆ψ = 0 are called
harmonic forms.
To justify the notation ∆ and the term “harmonic” we shall calculate the Hodge
Laplacian, starting with the case where ϕ ∈ λp (Rn ) and the metric in Rn is the
standard Euclidean. Write
X′
ϕ= |I| = pϕI dxI
P′
where denotes summation over increasing sequences I = (i1 , . . . , ip ) and dxI =
dx ∧ · · · ∧ dxip . Then
i1
′
X ′
X
i I
dϕ = I, i∂i ϕI dx ∧ dx , ∗ϕ = I, JϕI dxJ sgn(I, J)
If I 6= I ′ then the contribution is 0 unless the difference is in just one index. Assume
for example that I = (2, . . . , p + 1), I ′ = (1, . . . , p). Then we must have i = j ′ = 1,
i′ = j = p + 1, and we obtain
i I 1 2 ... p+1
sgn = sgn = (−1)p ,
i′ I′ p+1 1 ... p
j J p+1 1 p+2 ...
sgn = sgn = −1
j′ J′ 1 p+1 p+2 ...
I J 2 ... p +1 1 p+2 ...
sgn = sgn = (−1)p .
I′ J′ 1 ... p p+1 p+2 ...
122 LINEAR DIFFERENTIAL OPERATORS IN RIEMANNIAN GEOMETRY
If we integrate by parts moving derivatives to the first factor the two terms will therefore
give contributions which cancel each other, and the only remaining terms are those with
I = I ′ and i = i′ ∈ J or j = j ′ ∈ I, hence
Z X ′ Xn
2 2
∗
−∂i2 ϕI ϕI dx.
(∆ϕ, ϕ) = kdϕk + kd ϕk = I
i=1
Thus the Hodge Laplacian operates componentwise as minus the classical Laplacian;
the minus sign is natural since the Hodge Laplacian is a positive operator by its defi-
nition. We shall write ∆H whenever confusion seems possible.
To compute ∆ generally we can use the formula
∆ = (−1)n(p−1)+1 d ∗ d ∗ +(−1)np+1 ∗ d ∗ d on λp ,
which follows from (6.3.2) and can be used locally even if M is not orientable. Choose
geodesic coordinates x1 , . . . , xn centered at the point where we want to calculate ∆ϕ.
Then gjk (x) = δjk + O(|x|2 ), by (3.1.15) the second derivatives of gjk are linear func-
tions of the Riemann curvature tensor, and gjk + g jk − 2δjk = O(|x|3 ). We can use
(6.2.9) to express the ∗ operator. When we compute ∆H ϕ(x) the result will be the
same as for the Euclidean metric unless two derivatives fall on a coefficient involving
g jk . Hence (6.3.3) only needs modification by a term of order 0.
Since ϕ can also be regarded as a P skew symmetric tensor field, we could also form
another Laplacian by the contraction g jk ϕ,jk . At the center of a geodesic coordinate
system this will also apart from lower order terms be the classical Laplacian acting on
each component of ϕ, so
X
∆H ϕ + g jk ϕ,jk is of order 0.
When p = 0 there P isjkno term d ∗ d∗ so no second order derivatives of gjk can occur,
hence −∆H ϕ = g ϕ,jk is the standard Laplace-Beltrami operator. We shall return
to the exact calculation for p 6= 0 later on in this section. For the moment it suffices
for us just to observe that the calculation of the leading term just made shows that
the principal symbol of the Hodge Laplacian is −|ξ|2 . Hence ∆H (or rather −∆H )
satisfies the hypotheses of Theorem 6.1.5. Thus the space Hp of harmonic p forms is
finite dimensional, and every ψ ∈ λp can be written in the form
ψ = h + ∆ϕ
where h is the orthogonal projection of ψ on Hp and ϕ ∈ λp is uniquely determined
mod Hp . Here we have of course used that ∆H is formally self-adjoint. With ϕ+ = dϕ
and ϕ− = d∗ ϕ, it follows that
(6.3.4) ψ = h + dϕ− + d∗ ϕ+ , h ∈ Hp , ϕ− ∈ λp−1 , ϕ+ ∈ λp+1 .
THE HODGE DECOMPOSITION 123
Hence the decomposition (6.3.4) is unique (but ϕ− and ϕ+ are not). If dψ = 0 then
dd∗ ϕ+ = 0, which implies
(dd∗ ϕ+ , ϕ+ ) = (d∗ ϕ+ , d∗ ϕ+ ) = 0,
which implies
From (6.3.5) it follows that the kernel consists of the harmonic forms, and D is formally
selfadjoint by its definition. Hence the range is the orthogonal space of the harmonic
forms, and the index is equal to 0. However, one can obtain operators with non-trivial
index by restricting D to suitable subbundles of ∧∗ T ∗ M .
First note that D maps forms of even (odd) degree to forms of odd (even) degree.
If we write
λeven = ⊕λ2p , λodd = ⊕λ2p+1 ,
it follows that the restriction Deo to λeven is a differential operator from λeven to λodd
with adjoint equal to the restriction of D to λodd , with values in λeven . Hence the
124 LINEAR DIFFERENTIAL OPERATORS IN RIEMANNIAN GEOMETRY
kernel of Deo is ⊕H2p , and the range is the orthogonal space of ⊕H2p+1 in λodd . This
means that
n
X n
X
eo p p
(6.3.6) ind D = (−1) dim H = (−1)p dim H p (M ),
p=0 p=0
which is the Euler characteristic of M . If we accept the fact that the Euler characteristic
is equal to the number of critical points of functions on M , counted with signs, then
Theorem 4.3.2 means that the index of Deo can be expressed in terms of the curvature
forms. This is the model for the index theorems which will be a major subject of this
chapter.
Assume now that M is a compact oriented C ∞ Riemannian manifold of even dimen-
sion n = 2l. By Proposition 6.2.1 the map τ defined in ∧∗ Tx∗ for every x ∈ M gives a
decomposition
(∧∗ T ∗ M )C = ∧+ ⊕ ∧− ,
where ∧± are complex vector bundles of fiber dimension 2n−1 . We claim that
(6.3.7) Dτ = −τ D,
which will prove that D maps sections of ∧+ (∧− ) to sections of ∧− (∧+ ). To prove
(6.3.7) we let ϕ be a p form. Since the dimension n is even, we have d∗ = − ∗ d∗, hence
2
Dτ ϕ = ip(p−1)+l (d − ∗d∗) ∗ ϕ = ip(p−1)+l (d ∗ ϕ − (−1)p ∗ dϕ)
τ Dϕ = τ (dϕ − ∗d ∗ ϕ) = i(p+1)p+l ∗ dϕ − i(p−1)(p−2)+l ∗ ∗d ∗ ϕ
= ip(p−1)+l (−1)p ∗ dϕ − (−1)p−1 (−1)2l−p+1 d ∗ ϕ
= ip(p−1)+l (−1)p ∗ dϕ − d ∗ ϕ .
2
(If l is odd then ϕ ∧ ϕ = 0 since ϕ ∧ ϕ = (−1)l ϕ ∧ ϕ, so (6.3.9) vanishes then.)
We can give (6.3.9) an interpretation which does not involve harmonic forms. To
do so we note that the bilinear form
Z
(6.3.10) (ϕ, ψ) 7→ ϕ∧ψ
M
The expression of ind D+ in terms of Pontrjagin classes will be discussed later on.
We shall close this section by making a few
P additional calculations involving ∗, d∗
and ∆. Let ϕ be a one form and write ϕ = ϕj dxj in local coordinates. By (6.2.9)
n
n√
X
∗ϕ= ϕk g ki (−1)i−1 dx1 ∧ . . . |∧dx i
{z } ∧ · · · ∧ dx g
k,i=1 omit
n
n√
X
= Φi (−1)i−1 dx1 ∧ . . . ∧dx i
| {z } ∧ · · · ∧ dx g,
i=1 omit
i√
where Φ = ϕ♯ . Hence d ∗ ϕ = g)dx1 ∧ · · · ∧ dxn , so
P
i (∂i (Φ
n
− 12 1
X
(6.3.12) ∗
d ϕ = − ∗ d ∗ ϕ = −g ∂i (Φi g 2 ) = − div Φ,
i=1
126 LINEAR DIFFERENTIAL OPERATORS IN RIEMANNIAN GEOMETRY
where the last equality is a definition. The invariance of div Φ when Φ is a vector field
is also clear since for χ ∈ C0∞ (M ) we have
Z Z
(6.3.13) − χ div Φ dvol = hΦ, dχi dvol.
M M
This is also true if χ = 1 and Φ has compact support, that is, the integral of the
divergence of a vector field with compact support is equal to 0. We can also write
X
(6.3.14) div Φ = Φi ,i ,
for (6.3.14) is true in a geodesic coordinate system. (See also the proof of (3.4.15) and
Exercise 3.4.2.) Thus
Z X n
(6.3.14) ′
Φi ,i dvol = 0,
M i=1
since ϕj,i = ∂i ϕj + k Γij k ϕk and Γij k = Γji k . However, one must remember when
P
using this expression that ϕj,i is not antisymmetric in general. We shall now compute
d∗ dϕ by taking another one form ψ with compact support and calculating (dϕ, dψ).
At a point where gij = δij the pointwise scalar product is
n
X n
X n
X
1
2 (ϕj,i − ϕi,j )(ψj,i − ψi,j ) = (ϕj,i ψj,i − ϕi,j ψj,i ) = (∇ϕ, ∇ψ) − Φi ,j Ψj ,i ,
i,j=1 i,j=1 i,j=1
for the natural scalar product in Tx∗ M ⊗ Tx∗ M . The last expression is invariant so it is
always equal to (dϕ, dψ)(x), which proves that
Z Z X n
(dϕ, dψ) = (∇ϕ, ∇ψ) dvol − Φi ,j Ψj ,i dvol.
M M i,j=1
′
In
P thei lastj integral we can integrate by parts using (6.3.14) applied to the vector field
j Φ ,j Ψ , which gives
Xn
∗ ∗
d dϕ = ∇ ∇ϕ + Φi ,ji dxj .
i,j
HEAT EQUATIONS 127
Here ∇∗ is the adjoint of the operator ∇ from one forms, that is, sections of T ∗ M , to
sections of T ∗ M ⊗ T ∗ M , with the scalar product above. Now we add
n
X n
X
dd∗ ϕ = −d Φi ,i = − Φi ,ij dxj
i=1 i,j=1
This makes the remarks following (6.3.3) explicit. The identity (6.3.15) is of the form
(6.1.4), corresponding to the Levi-Civita connection in T ∗ M , and c has been identified
as the Ricci curvature. There are similar formulas for forms of degree p > 1, due to
Weitzenböck [1]; see also de Rham [1, p. 131]. Formulas of the same structure as
(6.3.15) are called Weitzenböck decompositions in general.
To show the importance of (6.3.15) we now assume that M is compact and con-
nected, and take the scalar product of (6.3.15) with ϕ. This gives
Z X
2
(6.3.16) (∆H ϕ, ϕ) = k∇ϕk + Rij Φi Φj dvol, ϕ ∈ λ1 , Φ = ϕ♯ .
M
If ϕ is a harmonic one form, that is, ∆ϕ = 0, and the Ricci tensor is non-negative, it
follows from (6.3.16) that both terms in the right-hand side must vanish. Hence the
covariant differential of ϕ is equal to 0, and ϕ♯ is at every point contained in the kernel
of the Ricci tensor. If the Ricci tensor is strictly positive definite at some point, then ϕ
must vanish in a neighborhood. But kϕ(x)k is a constant since ∇ϕ = 0, so we obtain
the following theorem of Bochner:
Theorem 6.3.5. If h is a harmonic one form on a compact, connected Riemannian
manifold M with non-negative Ricci tensor, then ∇h = 0 and h♯ is in the kernel of R
at every point. Hence dim H 1 (M ) ≤ n. If in addition R is strictly positive definite at
some point, then h = 0, hence H 1 (M ) = 0 then.
Theorem 6.3.5 has been the starting point of much work on the connection between
the topology of a manifold and its curvature.
6.4. Heat equations. The first step to a calculation of the index of the operators
Deo and D+ introduced in Section 6.3 is to study the heat equations associated with the
corresponding metric elliptic operators. Let P be a metric differential operator acting
on sections of the C ∞ vector bundle E over the compact C ∞ Riemannian manifold
128 LINEAR DIFFERENTIAL OPERATORS IN RIEMANNIAN GEOMETRY
where v is given in C ∞ (M, E). Formally this means that u = etP v, t ≥ 0, where
e−tz
Z
tP −1
e = (2πi) dz
Γ z+P
with an integration contour Γ such as a sector < π of the unit circle with the two
tangents in the direction of the bisector of the first and fourth quadrants. (We have
defined the principal symbol without the customary factors i, so our conventions make
P bounded above.) The distributions Fν (x) in (6.1.8), which depend on z ∈ C \ R+ ,
are inverse Fourier transforms of ξ 7→ (z − |ξ|2 )−ν−1 , and
Z
2
(2πi) −1
(z − |ξ|2 )−ν−1 e−tz dz = e−t|ξ| (−t)ν /ν!.
Γ
If we multiply Fν by e−tz /(2πi) and integrate over Γ, we thus obtain at least formally
2
the inverse Fourier transform of (−t)ν e−t|ξ| /ν!, that is,
2
(6.4.2) Hν (t, x) = (−t)ν H0 (t, x)/ν!, where H0 (t, x) = (4πt)−n/2 e−|x| /4t
, t > 0,
is the fundamental solution of the scalar heat equation. The identities (6.1.9), (6.1.10)
are transformed to the obvious identities
Recall that in (6.1.13) the coefficients uν were independent of z. Repeating the proof
of (6.1.13) we now obtain with the same coefficients uν
µ
X
(6.4.3) (P − ∂t ) uν Hν = (P uµ )Hµ .
0
Since ze−tz = −∂t e−tz , this can easily be justified also by an integration
Pµ of (6.1.13)
−tz
multiplied by e . Since H0 (t, ·) → δ0 as t → +0, it is clear that 0 uν Hν → δ0 as
t → +0. If we define Hν (t, x) = 0 when t ≤ 0, then Hν ∈ C k (R×U ), if νP > k +n/2 and
∞ µ
U is a geodesic coordinate patch in M . Every Hν is C outside (0, 0), so 0 uν Hν (t, x)
is a good parametrix for (∂t − P ) when µ is large.
Using a parametrix it is easy to solve the Cauchy problem (6.4.1). At first we put the
inhomogeneity in the equation instead of the data and consider the Cauchy problem
(∂t − Px )H(t,
e x, y) = R(t, x, y) when t > 0; H(t,
e ·, ·) → δdiag as t → +0,
is an exact fundamental solution for the Cauchy problem. (The product here is the
duality between Ez∗ and Ez .) The correction term is infinitely differentiable and van-
ishes for t < 0, so the asymptotic behavior at the diagonal is still given by the sum in
(6.4.3).
Now the Cauchy problem (6.4.1) has the exact solution
Z
u(t, x) = H(t, x, y)v(y) dvol(y).
There is no other solution. In fact, for the scalar product with a solution U of the
adjoint equation (∂t + P ∗ )U = 0 we have
We can choose U as a solution for t < T which is equal to a given section of E ∗ when
t = T by the preceding existence proof, for the change of sign in front of ∂t changes
the direction of “time”. Since
it follows that u(T, ·) is uniquely determined by the initial data v. Hence H is also
uniquelyP determined. (It would therefore have been enough to start from an approx-
µ
imation 0 uν Hν with a fixed large µ; the resulting H is then independent of µ.) In
particular it follows that
∞
X
(6.4.4) H(t, x, x) ∼ tj−n/2 hj (x), t → +0,
0
u = F D+∗ D+ u + Ru
HIRZEBRUCH’S INDEX FORMULA AND GILKEY’S THEOREM 131
If ϕ ∈ Γ+
λ then
When N > n/4 the kernel of F N is square integrable in each variable, so we obtain
N
(6.5.2) sup |ϕ(x)| ≤ C(1 + λ)1+ 4 kϕkL2 , ϕ ∈ Γλ .
and if we sum over all α in an orthonormal basis in ∧+x we get for 0 < t < 1
X X
(6.5.4) e−2tλj |ϕj (x)|2 ≤ C ′ t−n , hence e−2tλ dim Γ+λ ≤C t
′′ −n
.
The last bound follows by integrating the first over M , and it proves the convergence
of the sum in (6.5.3) when χ(λ) = e−tλ , t > 0.
132 LINEAR DIFFERENTIAL OPERATORS IN RIEMANNIAN GEOMETRY
Since such sections of ∧+ are dense in the continuous sections it follows in view of
(6.5.2) and (6.5.4) that for every f ∈ ∧+x
X
H + (t, x, x)f = e−tλj ϕj (x)(f, ϕj (x)).
∞
n
X
+ −
Tr H (t, x, x) − Tr H (t, x, x) ∼ tj− 2 hj (x),
0
where hj is a polynomial in the components of the Riemann curvature tensor and its
covariant derivatives. The product hj (x) dvol(x) is a density on the oriented manifold
M . If we reverse the orientation, then τ is replaced by −τ , which means that the
spaces ∧+ and ∧− and therefore the kernels H + and H − are interchanged. Thus hj is
replaced by −hj , which means that we can regard hj (x) dvol(x) as a n form ωj . From
(6.5.1) it follows then that
Z
(6.5.5) ωj = 0, j < n/2,
M
Z
(6.5.6) ωn/2 = ind D+ .
M
Our aim now is to prove that ωj = 0 when j < n/2, which is a very much stronger
statement than the integral condition (6.5.5), and that ωn/2 is a polynomial in the
Pontrjagin forms. The following simple observation is crucial:
HIRZEBRUCH’S INDEX FORMULA AND GILKEY’S THEOREM 133
Lemma 6.5.2. If the metric ds2 is replaced by λ2 ds2 where λ is a positive constant,
then ∆ is multiplied by λ−2 and ωj is multiplied by λn−2j .
Proof. In a local coordinate system gjk is replaced by λ2 gjk and g jk by λ−2 g jk , which
means that ∆ is replaced by λ−2 ∆. Now
if τ = λ−2 t. Hence the initial value problem for λ−2 D+∗ D+ + ∂t has the solution
Z
H + (t/λ2 , x, y)f (y) dvol(y)
where H + and dvol(y) belong to the metric ds2 . For the metric λ2 ds2 the vol-
ume element is λn dvol(y), so the new fundamental solution is H + (t/λ2 , x, y)λ−n .
The product of H ± (t/λ2 , x, x)λ−n by the new volume element λn dvol(x) is equal to
H + (t/λ2 , x, x) dvol(x), so
n n
X X
tj− 2 ωj is replaced by λn−2j tj− 2 ωj ,
where ΦI is a polynomial in the curvature tensor and its covariant derivatives of order
≤ k at p, which is independent of M and p. The invariant is said to have weight k if
ω is multiplied by λk when the first fundamental form is multiplied by λ2 .
Let V be the Euclidean vector space Rn . We can regard the curvature tensor R as
N4 N4+j
an element of V and R(j) as an element of V , so (6.5.7) is a polynomial map
k
M 4+j
O
q
Φ : W → ∧ V, W = Wj , Wj = V.
j=0
134 LINEAR DIFFERENTIAL OPERATORS IN RIEMANNIAN GEOMETRY
The orthogonal group operates both in W and in ∧q V , and we denote the operation of
O by O ∗ in both cases. Since an orthogonal transformation of x1 , . . . , xn gives a new
geodesic system of coordinates centered at p, we must have
Φ(O ∗ w) = O ∗ Φ(w)
if w = (R, . . . , R(k) ) for some choice of a Riemannian metric. (Note that already the
L4
Bianchi identities show that R is not an arbitrary element in V .) Hence we have
Φ(w) = (O −1 )∗ Φ(O ∗ w)
for all such w. Taking the average of the right-hand side over the orthogonal group,
we obtain a polynomial Φ e which can still be used in (6.5.7) and has the advantage of
being equivariant, that is,
e ∗ w) = O ∗ Φ(w),
Φ(O e ∀w ∈ W.
To simplify notation we assume in what follows that Φ already has this property. It
is clear that if we split Φ into a sum of polynomials which are homogeneous in each of
the variables wj ∈ Wj , then all the terms will be equivariant.
A polynomial map Ψ(w) of degree ν in w ∈ Wj can be written in one and only one
way in the form Ψp (w, . . . , w) where Ψp (w1 , . . . , wν ) is a symmetric multilinear map
in w1 , . . . , wν ∈ Wj ; the polarization Ψp is given by
ν
Y
Ψp (w1 , . . . , wν ) = hwj , ∂/∂wiΨ(w)/ν!.
j=1
Np
Ψp defines a linear map on Wj which is equivariant if Ψ is. If we use polarization
for each Wj we may conclude that Φ is a sum of polynomials each of which is induced
by a linear equivariant map
ON
ϕ: V → ∧q V.
The fundamental theorem on O(n) invariants, Theorem D.1, gives a complete descrip-
NN
tion of such maps when q = 0. One calls a linear form ϕ on V elementary if
N = 2k and
ϕ(v1 ⊗ · · · ⊗ v2k ) = (v1 , v2 ) . . . (v2k−1 , v2k ),
that is,
X N
O
ϕ(ξ) = ξα1 α1 α2 α2 ...αk αk , ξ∈ V,
or if ϕ differs from this linear form just by a permutation of the indices.
NN
Theorem 6.5.4. Every invariant linear form V → R is a linear combination of
elementary forms; in particular there are no such forms 6= 0 unless N is even.
This is just a reformulation of Theorem D.1, so we pass to:
HIRZEBRUCH’S INDEX FORMULA AND GILKEY’S THEOREM 135
NN
Corollary 6.5.5. If ϕ : V → ∧q V is equivariant and not identically 0, then N −q
is an even integer 2r ≥ 0, and ϕ is a linear combination of elementary maps, that is,
maps which apart from a permutation of the indices are of the form
X
ϕ(ξ) = ξα1 α1 α2 α2 ...αr αr [β1 ...βq ] dxβ1 ∧ · · · ∧ dxβq ,
α,β
where [. . . ] denotes alternation over the enclosed indices, that is, summation over all
permutations after multiplication by the sign of the permutation.
NN+q
Proof. We define an invariant form ϕ̃ : V → R by
We shall now apply the corollary to the q form invariant ω in Definition 6.5.3. When
α = (α1 , . . . , αν ) is a sequence of |α| = ν indices between 1 and n, we shall denote by
Rα the corresponding components of the ν − 4th covariant derivative of the Riemann
curvature tensor Rijkl . By an elementary monomial of degree r in R we shall mean an
expression of the form
X∗
(6.5.8) m(R) = R α1 R α2 . . . R αr ,
q
where the summation indicates alternation of precisely q indices and pairwise contrac-
tion of the others in the total index sequence α1 α2 . . . αr , where the number of indices
shall exceed q by an even number (or 0). What we have proved so far is that with
coefficients am ∈ R we have
X
(6.5.7)′ ω(p) = am m(R).
Lemma 6.5.6. The weight of m(R) is 2r + q − |α| where |α| = |α1 | + . . . |αr |.
Proof. With a fixed system of local coordinates the Christoffel symbols Γikj are mul-
tiplied by λ2 if the metric is multiplied by λ2 , but the Christoffel symbols Γik j remain
unchanged. By (2.1.13) it follows that Rijkl is multiplied by λ2 , and so are the covari-
ant derivatives. If t1 , . . . , tq are tangent vectors, then m(R) evaluated with respect to
these in a fixed coordinate system is
X∗
(6.5.8)′ Rα1 . . . Rαr g i1 i2 . . . tj11 . . . tjqq ,
q
Such identities are preserved by covariant differentiation since it commutes with the
permutation group acting on the tensor product. Thus differentiation of the first
Bianchi identity gives
Rijkl,m + Riljk,m + Riklj,m = 0
if we write out all terms explicitly. The alternation of Rijkl over three arbitrary indices
is zero by the first Bianchi identity and the symmetries (6.5.9), and this is also true
for Rijkl,m . When one alternates over mij or mkl this is the second Bianchi identity.
If one alternates over jkm and notes that
it follows that two times the alternation is 0, for exchanging k and j changes the sign
of the permutation. This proves the claim. By covariant differentiation we conclude
that the alternation of Rα over three of the first four or five indices is always equal to
0.
HIRZEBRUCH’S INDEX FORMULA AND GILKEY’S THEOREM 137
If the term m(R) is not zero it follows that the alternation in (6.5.8) involves at
most two of the first four indices in each factor. Altogether we can then alternate in
at most 2r + ε indices, hence q ≤ 2r + ε, which means that the weight of m(R) is
≤ 0. This proves the first part of Gilkey’s theorem. If the weight is 0 we have also
found that q = 2r + ε and that we must alternate over all indices corresponding to
covariant differentiation and in addition two of the first four indices in each factor. If
a covariant differentiation occurs in some factor, it follows that we alternate over three
of the first five indices in it, which implies that m(R) = 0. Hence ε = 0, that is, no
covariant derivatives occur. Furthermore we have alternation with respect to precisely
two indices in each factor Rijkl . If they are the first two, we can use the symmetry
Rijkl = Rklij to replace them by the last two. If they are the middle two, we can use
that
Rijkl − Rikjl = Rilkj = (Rilkj − Riljk )/2
to replace them by the last two indices, at the expense of a factor 12 . Hence it follows
that in a geodesic coordinate system ω is a sum of products of expressions of the form
X
Ri1 i2 j1 j2 Ri2 i3 j3 j4 . . . Rik i1 j2k−1 j2k dxj1 ∧ · · · ∧ dxj2k .
i,j
Let us now note that for n × n matrices Aij with complex coefficients we have
n
Y n
X
det(λδij − Aij ) = (λ − λj ) = λn−j (−1)j cj ,
1 0
where λj are the eigenvalues and cj the elementary symmetric functions of them. We
have for every positive integer k
n
X X
sk = λkj = Ai1 i2 Ai2 i3 . . . Aik i1 ,
j=1
for the sum is invariant under conjugation of the matrix A, and the formula is valid
for diagonal matrices. Since c1 = s1 , and by Newton’s formulas
for every oriented manifold M of dimension 4k. One can now determine the coefficients
of L by specializing M to manifolds for which the signature and the Pontrjagin classes
are easy to describe, such as products of complex projective spaces. We refer to Atiyah,
Bott and Patodi [1] for the calculation of Lk and only give the result. Set
X Y Y X
(6.5.12) Lk = xj / tanh xj , where (1 + x2j ) = pk .
We have followed that paper here apart from substituting the Hadamard construction
of a parametrix for pseudo-differential operator theory. The Hadamard construction
is not only more elementary, it fits precisely with the differential geometric context.
For the operator Deo having the Euler characteristic as index, it was first proved by
Patodi [1] that a corresponding phenomenon occurs, leading to the Gauss-Bonnet-
Chern formula. A variant of his proof using ideas of supersymmetry from physics is
given in Cycon, Froese, Kirsch and Simon [1, Chapter 12]. (See page 258 for a criticism
of the methods used here which do not go all the way to a direct computation of the
coefficients in the index formula.) More recently, a direct computational proof has
been obtained for twisted Dirac operators by Bismut [1] and Getzler [1,2]. Our next
aim is to give an exposition of the methods of Getzler.
6.6. Operators of Dirac type. We started this chapter with a study of second
order metric elliptic operators, but all the geometric applications in Sections 6.3 and
6.5 concerned first order operators D such that −D∗ D is metric. (The minus sign
comes from the convention for defining the principal symbol introduced in Section 5.1,
which does not contain the factor i which is customary in pseudo-differential operator
theory.) The rest of this chapter will be devoted to a more systematic search for such
operators.
Definition 6.6.1. If M is a C ∞ manifold and E0 , E1 two C ∞ Hermitian vector bundles
on M with the same fiber dimension, then a first order differential operator
D : C ∞ (M, E0 ) → C ∞ (M, E1 )
where σ(x, ξ)∗ : E1x → E0x is the adjoint with respect to the Hermitian metrics and
p(x, ξ) is a positive definite quadratic form in Tx∗ M . Since E0x and E1x have the same
dimension, it follows from (6.6.1) that
originally introduced what is now known as a Dirac operator. He wanted to write the
Klein-Gordon equation
The first step in the study of (symmetric) Dirac operators is to determine matrices
σ(x, ξ) depending linearly in ξ, which satisfy (6.6.3) for a fixed x. This leads to the
definition of Clifford algebras and spinors, which will be studied in Section 6.7. We can
then define (twisted) Dirac operators in Section 6.8, where we also prove an analogue
of the Weitzenböck formula due to Lichnerowicz. Section 6.9 is an analytical interlude
devoted to the classical Mehler formula giving the heat kernel for the harmonic oscil-
lator explicitly, and to some extensions needed here. We are then prepared to prove
the local index theorem for Dirac operators in Section 6.10.
6.7. Clifford and spinor algebra. Let V be a real vector space, q a quadratic
form in V , E a complex vector space, and σ : V → End(E) = L(E, E) a linear map
such that
that is, σ̃ vanishes on the two sided ideal I generated in the tensor algebra by all
elements of the form v ⊗v −q(v) · 1 with v ∈ V , so σ̃ induces an algebra homomorphism
from the quotient algebra to End(E). Note that with the notation q also for the
polarized form
(6.7.3) εj ⊗ εk + εk ⊗ εj , j 6= k; εj ⊗ εj − qj · 1; j, k = 1, . . . , n.
Here ai1 ,...,ij is a linear form on the tensor algebra vanishing in I, defined by
n
qiµi ,
Y
ι
ai1 ,...,ij (εk1 ⊗ · · · ⊗ εkν ) = (−1)
i=1
are represented by sums (6.7.4) with j even (odd), so the dimensions are 2n−1 . We
have
Cli (V, q) · Clj (V, q) ⊂ Cli+j mod 2 (V, q),
which gives a Z2 grading of the algebra Cl(V, q). The set Cl[j] (V, q) of elements of the
form (6.7.4) with j fixed ∈ {0, . . . , n} is a linear subspace of Cl(V, q), equal to the linear
span of all images of elements of the form
X
(6.7.5) sgn πvπ(1) ⊗ · · · ⊗ vπ(j) , v1 , . . . , vj ∈ V,
π
and since the sum over π is skew symmetric in k1 , . . . , kj , we have an element of the
form (6.7.4) with j fixed.
If x is the class of the form (6.7.4) with P a basis ε1 , . . . , εn satisfying (6.7.2), then
t t
x · x and x · x both have the constant term a2i1 ...ij qi1 . . . qij , for εi1 · · · εij · εk1 · · · εkl
can only be a constant if there is some ν ∈ {i1 , . . . , ij } ∩ {k1 , . . . , kl } with qν = 0 or
{i1 , . . . , ij } = {k1 , . . . , kl }. The definition of Q is obviously invariant since it makes
CLIFFORD AND SPINOR ALGEBRA 143
no reference to the choice of a basis, and the other statements are obvious since for
example t (v · x) · (v · x) = t x · v · v · x.
Note that Cl[j] · Cl[k] is not contained in Cl[j+k] but may also have components in
Cl[i] with i < j + k.
Exercise 6.7.1. Show that when v1 , . . . , v4 ∈ V we have with all products taken in
Cl(V, q)
[v1 · v2 − v2 · v1 , v3 · v4 − v4 · v3 ] = 4 q(v1 , v4 )(v2 · v3 − v3 · v2 ) − q(v1 , v3 )(v2 · v4 − v4 · v2 )
+ q(v2 , v3 )(v1 · v4 − v4 · v1 ) − q(v2 , v4 )(v1 · v3 − v3 · v1 ) .
Thus Cl[2] (V, q) is a Lie algebra. Deduce that if q is positive definite and ε1 , . . . , εn is
an orthonormal basis, then
X X X
[ aij εi · εj , bk,l εk · εl ] = 4 [a, b]ij εi · εj
i,j k,l i,j
if a and b are skew symmetric matrices. (Compare with the Lie algebra of SO(n).)
When q = 0, the Clifford algebra becomes the exterior algebra ∧∗ V on V . However,
the case of interest to us here is primarily the case where q is positive definite. The
negative definite case is often preferred but the difference is not important in the
arguments below where we introduce the complexification of Cl, for all non-degenerate
quadratic forms in a complex vector space are equivalent. However, before doing so
we introduce an element in the Clifford algebra which will be very important later on.
Theorem 6.7.3. If V is oriented and q is non-degenerate, then
n q
Y
(6.7.6) ν = ε1 · · · εn / |qj |
1
is independent of the choice of positively oriented diagonalizing basis ε1 , . . . , εn for V .
We have
n
1
Y
(6.7.7) ν 2 = (−1) 2 n(n−1) sgn qj , νu = (−1)j(n−1) uν, if u ∈ Cl[j] (V, q).
1
n q
Y
|qj |νεj = (−1)n−j ε1 · · · εj−1 qj εj+1 · · · εn ,
1
n q
Y
|qj |εj ν = (−1)j−1 ε1 · · · εj−1 qj εj+1 · · · εn .
1
Hence νεj = (−1)n−1 εj ν, which immediately gives the second part of (6.7.7).
A complete description of the real Clifford algebras is fairly complicated (see Atiyah,
Bott and Shapiro [1]) so we content ourselves with discussing the complexification
ClC (V, q) obtained by allowing the coefficients in (6.7.4) to be complex, or equivalently,
replacing V by the complexification VC in the definition by the tensor algebra. All non-
degenerate quadratic forms in a complex vector space are equivalent under complex
linear coordinate transformations, and we shall only discuss ClC (n) = ClC (Rn , e)
where e is the Euclidean quadratic form in Rn .
Theorem 6.7.4. There are complex algebra isomorphisms
k k k
(6.7.8) ClC (2k) ∼
= End(C2 ), ClC (2k + 1) ∼
= End(C2 ) ⊕ End(C2 ).
Proof. We start with dimensions 1 and 2. The algebra ClC (1) consists of all (α, β) ∈ C2
with componentwise addition and
Since
αα′ + ββ ′ ± (αβ ′ + βα′ ) = (α ± β)(α′ ± β ′ ),
CLIFFORD AND SPINOR ALGEBRA 145
the map (α, β) → (α+β, α−β) changes the operations in ClC (1) to the coordinatewise
operations in C ⊕ C.
To determine ClC (2) we define a linear map σ : R2 → End(C2 ) by
1 0 0 i
σ(ε1 ) = , σ(ε2 ) = ,
0 −1 −i 0
(6.7.9) ClC (n + 2) ∼
= ClC (n) ⊗ ClC (2), n ≥ 1.
To prove (6.7.9) we define a linear map σ : Rn+2 → ClC (n) ⊗ ClC (2) by
Since σ(εj )2 = 1 ⊗ 1 for j ≤ n + 2 and σ(εj )σ(εk ) + σ(εk )σ(εj ) = 0 if j 6= k, the map
σ induces a homomorphism σ̃ : ClC (n + 2) → ClC (n) ⊗ ClC (2). The range of σ̃ is a
subalgebra containing 1 ⊗ ClC (2) and ClC (n) ⊗ 1, so σ̃ is surjective, henced bijective
because the dimensions agree. Now the isomorphisms (6.7.8) follow inductively since
for arbitrary (complex) vector spaces E and F we have
(End E) ⊗ (End F ) ∼
= End(E ⊗ F ).
where ∧∗ Ck denotes the full complex exterior algebra over Ck . (Since ∧∗ Ck has
complex dimension 2k , the two sides of (6.7.10) are isomorphic by Theorem 6.7.4.)
To define the isomorphism we must recall a construction made in Section 6.2. For
w ∈ W = Ck we denote by Tw the exterior multiplication Tw u = w ∧ u, u ∈ ∧∗ W , and
by Tw∗ we denote the adjoint with respect to the natural hermitian scalar product in
∧∗ W . If w is the basis vector e1 ∈ W , then
Tw u = e1 ∧ u0 , if u = u0 + e1 ∧ u1 ,
146 LINEAR DIFFERENTIAL OPERATORS IN RIEMANNIAN GEOMETRY
where I is the identity in ∧∗ W , and |w|2 is the square of the norm of w as an element
in Ck , or equivalently as an element in R2k . Since Tw2 = 0 and Tw∗ 2 = 0, it follows that
(6.7.12) V ∋ x 7→ u · x · u∗ , x ∈ V,
such that Spin(V, q) is the inverse image of SO(V, q). The maps Pin(V, q) → O(V, q)
and Spin(V, q) → SO(V, q) are surjective by the following lemma, which is very close
to Exercise 1.4.3:
Lemma 6.7.8. Every element in O(n) is a product of at most n orthogonal reflections
in Rn .
Proof. We can extend an orthogonal transformation O in Rn to a unitary transforma-
tion in Cn . The projection in Rn of an eigenvector in Cn is an invariant subspace V1
for O of dimension 1 or 2, and the orthogonal space V1⊥ is also invariant since
(OV1⊥ , V1 ) = (V1⊥ , O −1 V1 ) = 0.
Hence it suffices to prove the lemma in dimension 1 and dimension 2. In the first case
an orthogonal transformation is a reflection or the identity, in the second case it is a
reflection or a rotation by an angle θ and therefore the product of reflections in two
lines with an angle θ/2 between them.
Theorem 6.7.9. If q is a positive definite quadratic form in V , then the homomor-
phisms τ : Pin(V, q) → O(V, q) and τ : Spin(V, q) → SO(V, q) are surjective with
kernel {±1}. Both Pin(V, q) and Spin(V, q) are compact subsets of Cl(V, q), Spin(V, q)
is arcwise connected and so is Pin(V, q) \ Spin(V, q). If dim V ≥ 3 then Spin(V, q) is
simply connected.
Proof. We have already proved surjectivity, and it is clear that ±1 is in the kernel.
Suppose that u ∈ Pin(V, q) and that u · x · u∗ = x, x ∈ V . Then u ∈ Spin(V, q) since
τ (u) preserves the orientation, so u∗ · u = 1, hence u · x = x · u and
x · u · x = q(x)u.
148 LINEAR DIFFERENTIAL OPERATORS IN RIEMANNIAN GEOMETRY
If we take x = εl and note that εl · εi1 · · · εij · εl = ±εi1 · · · εij with the minus sign when
l ∈ {i1 , . . . , ij }, we conclude that only a term with j = 0 can occur. This means that
u is a real number, and since uxu = x, x ∈ V , we have u = ±1.
The compactness follows from the bound for the number of reflections in Lemma
6.7.8. To prove that Spin(V, g) is connected it suffices to note that v1 · · · v2k , where
vj ∈ V and q(vj ) = 1, is connected to v1 · v1 · · · v1 = 1 if each vj is connected to v1 in
the connected unit sphere in V .
What remains is to prove that Spin(V, q) is simply connected if dim V ≥ 3. In
the proof we shall need that τ is a local homeomorphism. It suffices to prove that at
±1. Take disjoint compact neighborhoods U± of ±1 in Spin(V, q) with U− = −1 · U+
and, using the compactness, a compact neighborhood U of the identity in SO(V, q)
such that τ −1 U ⊂ U+ ∪ U− . By the results already proved it follows that τ maps the
′
neighborhoods U± = U± ∩ τ −1 U of ±1 bijectively, hence homeomorphically, on U . For
the proof we shall also need the following elementary lemma:
Lemma 6.7.10. Every closed loop in SO(n), n ≥ 2, is homotopic to a loop in SO(2),
embedded in SO(n) as SO(2) × In−2 .
Proof. Let R/Z ∋ t 7→ O(t) ∈ SO(n) be the given loop, n ≥ 3. If εn = (0, . . . , 0, 1) is
left fixed by O(t) for every t, then O(t) ∈ SO(n − 1) × 1, and the lemma follows by
induction from lower dimensions. In the general case we first regularize O(t) to a C 1
map. Then the curve {O(t)εn ; t ∈ R/Z} is of measure 0 in S n−1 , so we can choose
ξ ∈ S n−1 such that ±ξ ∈ / {O(t); t ∈ R/Z}. Let O1 ∈ SO(n) map ξ to εn . Our loop is
homotopic to the loop t 7→ O1 O(t) = O(t),
e and O(t)ε
e n 6= ±εn for every t. Hence
O(t)ε
e n = εn cos θ(t) + ξ(t) sin θ(t)
with uniquely determined continuous θ(t) ∈ (0, π) and ξ(t) ∈ S n−1 orthogonal to
εn .Denote by O2 (t, s) the rotation by the angle −sθ(t) in the εn ξ(t) plane. Then
(t, s) 7→ (1, 0, . . . , 0) · (cos( 21 πs) cos(πkt), − cos( 12 πs) sin(πkt), sin( 21 πs), 0, . . . , 0)
defined by the Taylor expansion which converges in Cl(V, q), is an element in Spin(V, q).
The image in SO(V, q) has the matrix exp(4A) in the basis ε1 , . . . , εn . Thus (6.7.13)
gives a bijection of a neighborhood of 0 in the space of skew symmetric matrices on a
neighborhood of the identity in Spin(V, q), which is therefore an analytic manifold with
tangent space Cl[2] (V, q) at the identity, and
n
X
so(n) ∋ 4(ajk )nj,k=1 7→ ajk ej ek ∈ Cl[2] (V, q)
j,k=1
Proof. Choose
P∞ a norm k·k in Cl(V, q) such that kx·yk ≤ kxkkyk. The formal expansion
exp x = 0 xν /ν! converges for every x ∈ Cl(V, q), for the norm of the term is bounded
by kxkν /ν!. The sum is an analytic function of x, and we obtain exp x · exp(−x) =
exp 0 = 1 by rearranging the terms in the product. For small t ∈ R let
n
Y q
P (t) = ·Pij (t), Pij (t) = εi · ( 1 − a2ij t2 εi + aij tεj ),
i,j=1
with the product taken say in the lexicographical order. We have Pjj = 1, Pij (t) ∈
Spin(V, q) for any i, j, and
n
X
Pij (t) = 1 + aij tεi · εj + O(t2 ), hence P (t) = 1 + tx + O(t2 ), x = aij εi · εj ,
i,j=1
ν
This implies that τ (P (1/ν)P ) = (τ (P (1/ν)))ν → e4A where A is the skew symmetric
operator in V with Aεk = i aik εi , that is, with matrix (ajk ) in the basis ε1 , . . . , εn .
The proof is complete.
From now on we shall use the abbreviated notation Pin(n) and Spin(n) for
Pin(Rn , e) and Spin(Rn , e). The composition
µ
(6.7.14) ̺ : Pin(2k) → Cl(2k) → ClC (2k) → End(∧∗ Ck ))
(6.7.15) D±
1 : Spin(2k) → Aut(S± (2k)), S+ (2k) = ∧even (Ck ), S− (2k) = ∧odd (Ck ),
2
called the half-spin representations. One calls (6.7.14) the spin representation.
Proposition 6.7.13. The half-spin representations are unitary and irreducible.
Proof. If W ⊂ S+ (2k) is an invariant subspace for D+ ∗ k
1 , then W ⊕ (̺(ε1 )W ) ⊂ ∧ (C )
2
is an invariant subspace for ̺, so W = {0} or W = S+ (2k) by Proposition 6.7.11,
which also shows that the representations are unitary.
In the odd dimensional case we can use the isomorphism ClC (2k − 1) → Cl0C (2k)
mapping v ∈ R2k−1 to iv · ε2k , where ε2k is the last basis vector in R2k . This means
[j] 2
that the elements in CC (2k − 1) are multiplied by ij , followed by right multiplication
by ε2k if j is odd. Hence we get an inclusion Spin(2k − 1) → Spin(2k), so (6.7.15) gives
a representation
+
D
1
2
Spin(2k − 1) → Spin(2k) −→ Aut(S+ (2k)),
which is also denoted by D+ −
1 . Similarly we get a representation D 1 in S− (2k), but
2 2
since
+
D−1 µ(ε2k ) = µ(ε2k )D 1 on S+ (2k),
2 2
it is an equivalent representation.
For reasons analogous to Proposition 6.5.1 we shall have to calculate the difference
between the traces of an endomorphism K+ of S+ (2k) and one K− of S− (2k). We
combine them to
K+ 0
K= ∈ End(S(2k)) = End(∧∗ (Ck )) ∼= ClC (2k),
0 K−
and we write
(6.7.16) Str K = Tr K+ − Tr K− .
One calls Str K the supertrace of K; it is defined for any endomorphism in a space S
with a given decomposition S = S+ ⊕ S− .
152 LINEAR DIFFERENTIAL OPERATORS IN RIEMANNIAN GEOMETRY
for Te∗j−1 annihilates the right-hand side. Starting from the case j = 1 we obtain by
induction for increasing j
for Tej+1 annihilates the right hand side while Te∗j+1 removes the factor ej+1 . When
j = k we obtain ̺(ν)1 = ik , for
since ε1 · ε2 , ε3 · ε4 , . . . commute with the other factors which allows us to to move them
out to the left starting from the middle. The lemma is proved.
Proof of Theorem 6.7.14. From Lemma 6.7.15 it follows that
[2k]
If we set w = νz, then z = (−1)k νw by (6.7.7), so the component of z in ClC is
(−1)k w0 ν where w0 is the constant term in w. We shall prove that
(6.7.18) Tr µ(w) = 2k w0 .
In view of (6.7.17) this will show that (−1)k w0 = (−2)−k Tr µ(w) = (2i)−k Str µ(z),
which is the Berezin-Patodi formula. To prove (6.7.18) we note first that µ(1) is the
identity in a space of dimension 2k , so the trace is 2k . We also have to show that for
j = 1, . . . , 2k we have Tr µ(w) = 0 if
w = εi1 · · · εij , 1 ≤ i1 < · · · < ij ≤ 2k.
If j is even we write w = εi1 · w1 = −w1 · εi1 and obtain
Tr µ(w) = Tr(µ(εi1 )µ(w1 )) = Tr(µ(w1 )µ(εi1 )) = − Tr µ, hence Tr µ(w) = 0,
for Tr(AB) = Tr(BA). If j is odd we choose i ∈ {1, . . . , 2k} \ {i1 , . . . , ij }. Then we
have w = εi · εi · w = −εi · w · εi , and it follows that
Tr µ(w) = Tr(µ(εi )2 µ(w)) = Tr(µ(εi )µ(w)µ(εi )) = − Tr µ(w),
hence Tr µ(w) = 0. The proof is complete.
The important feature of the Berezin-Patodi formula is that the information about
the supertrace is contained in the highest part of the Clifford algebra. As we shall
see in Section 6.10, the parametrix construction will show that the jth term has no
[2j]
component above the level ClC , and that the term there can be calculated for j ≤ k;
we shall precisely need the kth term to compute an index.
6.8. Clifford and spinor analysis. Let M be a C ∞ Riemannian manifold of
dimension n. For every x ∈ M we have defined in Section 6.7 a Clifford algebra
Clx (M ) = Cl(Tx∗ , gx∗ ), where gx∗ is the quadratic form in Tx∗ M dual to the metric form
gx in Tx M . It is clear that Clx (M ) is the fiber at x of a vector bundle Cl(M ), which
is a quotient of ⊕j≤n ⊗j T ∗ M . We can also view Cl(M ) as a bundle associated to the
orthonormal frame bundle and the representation of the orthogonal group on Cl(Rn , e)
induced by the natural action of O(n) on Rn . In either way we see that the Levi-Civita
connection is defined in Cl(M ).
As motivated in Section 6.6, Dirac operators involve bundles with a special structure:
Definition 6.8.1. A (complex) vector bundle E on M is called a Clifford bundle if for
every x ∈ M we have a map
Clx (M ) × Ex ∋ (v, ϕ) 7→ v · ϕ ∈ Ex
which makes Ex a Clx (M ) module, and the map Cl(M ) ⊕ E → E is in C ∞ . We shall
say that E is a Hermitian Clifford bundle if E is provided with a Hermitian metric
such that the map σ(v) : ϕ 7→ v · ϕ in End Ex is self-adjoint for every v ∈ Tx∗ M .
Note that the last assumption implies that (σ(v)ϕ, σ(v)ϕ) = (σ(v)2 ϕ, ϕ) =
|v|2x (ϕ, ϕ), where |v|x is the norm in Tx∗ M , so σ(v) is an isometry if |v|x = 1.
This implies that the whole group Pin(Tx∗ , gx∗ ) acts as a group of isometries on Ex .
154 LINEAR DIFFERENTIAL OPERATORS IN RIEMANNIAN GEOMETRY
Proposition 6.8.2. In a Hermitian Clifford bundle one can always define a connec-
tion ∇E compatible with the metric in the sense of (5.1.12) such that for every vector
field X, one form v and section ϕ of E we have
(6.8.1) ∇E E
X (v · ϕ) = (∇X v) · ϕ + v · ∇X ϕ.
∇E E
X ϕ(x) = v · ∇X (V · ϕ)(x),
(6.8.2) ∇E E ∗
X ϕ(x) = u · ∇X (u · ϕ)(x), u ∈ Pin(Tx , gx∗ ),
where du is the invariant measure on the compact group Pin. Since the metric is
invariant under Pin by assumption, this is still a connection compatible with the metric,
e E , which means that (6.8.1) holds at x when ∇v = 0 at
and (6.8.2) is now valid for ∇ Pn
x. In general we can for any given point x write v = 1 ψj vj in a neighborhood of x,
where ∇vj = 0 at x and ψj ∈ C ∞ . Then we obtain
X X
e E (v · ϕ) =
∇ (Xψj )vj · ϕ + e E ϕ = (∇X v) · ϕ + v · ∇
ψj vj · ∇ e E ϕ,
X X X
∇E : C ∞ (M, E) → C ∞ (M, T ∗ ⊗ E)
m : C ∞ (M, T ∗ ⊗ E) → C ∞ (M, E)
CLIFFORD AND SPINOR ANALYSIS 155
so the principal symbol as defined in Section 5.1 is at (x, ξ) ∈ Tx∗ the map
Ex ∋ w 7→ ξ · w.
In other words,
The square of the symbol is equal to |ξ|2 times the identity in Ex , and if E is a
Hermitian Clifford bundle it follows that D is of Dirac type. We can say more if the
connection is well chosen:
Proposition 6.8.3. If ∇E is a Clifford connection in the Hermitian Clifford bundle
E, then D is skew symmetric.
Proof. We must show that if ϕ and ψ are in C0∞ (M, E), then
Z
(6.8.4) (Dϕ, ψ) + (ϕ, Dψ) dvol = 0.
M
At any point we can find an orthonormal basis e1 , . . . , en for the vector fields, hence
a dual orthonormal frame ε1 , . . . , εn for the one forms, with ∇εj = 0 for all j at the
chosen point. Since
Xn
E
∇ ϕ= εj ⊗ ∇Eej ϕ,
1
we have
n
X
Dϕ = εj · ∇E
ej ϕ.
1
where v is a one form. Taking v = εj we obtain at a point where ∇εj = 0 for every j,
using (6.8.1) and (5.1.12),
n
X n
X
(∇ej X)(εj ) = (εj · ∇E E
ej ϕ, ψ) + (εj · ϕ, ∇ej ψ) = (Dϕ, ψ) + (ϕ, Dψ).
1 1
Since X has compact support and the divergence in the left-hand side is independent
of the choice of frames (cf. (6.3.14)), we conclude using (6.3.14)′ that (6.8.4) holds.
156 LINEAR DIFFERENTIAL OPERATORS IN RIEMANNIAN GEOMETRY
Remark. Usually one includes a factor i in the definition to make D symmetric. How-
ever, we have stayed here with the definition of principal symbol given in Section 5.1
and hope to be consistent.
We shall say that a Clifford module E is graded if a direct sum decomposition
E = E 0 ⊕E 1 is given which is compatible with the grading in Cl(M ), that is, Clix ·Exj ⊂
Exi+j mod 2 . The Clifford connection can clearly be chosen so that it respects this
decomposition, and then the Dirac operator takes the form (6.6.2).
Exercise 6.8.1. Prove that if ∇E is a Clifford connection and v is a one form, then
where D is the Dirac type operator corresponding to the Clifford module Cl(M ), that
is, with the notation in the proof of Proposition 6.8.3,
X
Dv = εj · ∇ej v.
We shall now prove Weitzenböck type formulas, similar to (6.3.15), for a Dirac type
operator D corresponding to a Hermitian Clifford bundle E with a Clifford connection.
Then D is skew adjoint and the principal symbol of D2 is |ξ|2 . For the connection
∇ : C ∞ (M, E) → C ∞ (M, T ∗ ⊗ E) we can form the adjoint with respect to the metric
in E and the metric in T ∗ ⊗ E such that
n
X n
X
2
k εj ⊗ wj k = kwj k2 , wj ∈ Ex ,
1 1
(∇∗ ∇(χϕ), ψ) = (∇(χϕ), ∇ψ) = (χ∇ϕ, ∇ψ) + ((dχ) ⊗ ϕ, ∇ψ) = (∇ϕ, ∇(χψ))
− (∇ϕ, (dχ) ⊗ ψ) + ((dχ) ⊗ ϕ, ∇ψ) = (χ∇∗ ∇ϕ, ψ) − (∇X ϕ, ψ) + (ϕ, ∇X ψ).
CLIFFORD AND SPINOR ANALYSIS 157
Since the connection is compatible with the Hermitian metric we have with scalar
products in the fiber at x
that is,
∇∗ ∇(χϕ) = χ∇∗ ∇ϕ − 2∇X ϕ − (div X)ϕ.
Combining this result with (6.8.7) we conclude that
where ε1 , . . . , εn and e1 , . . . , en are local dual orthonormal frames for the one forms
and the vector fields, and R∇ is the curvature of E with the connection ∇, defined by
(5.1.5).
Proof. Writing ∇E for the connection in E to distinguish it from the Levi-Civita
connection in T ∗ M , we have
n
X
Dϕ = εj · ∇E
ej ϕ, hence
1
n
X Xn n
X
2
D ϕ= εk · ∇E
ek (εj · ∇E
ej ϕ) = εk · (∇ek εj ) · ∇E
ej ϕ + εk · εj · ∇E E
ek ∇ej ϕ,
j,k=1 j,k=1 j,k=1
εj ⊗ ∇E ∞
P
by (5.1.5). Since ∇ϕ = ej ϕ we obtain if ψ ∈ C0 (M, E) has support in the
coordinate patch U where the local frames are defined
X
(∇∗ ∇ϕ, ψ) = (∇ϕ, ∇ψ) = (∇E E
ej ϕ, ∇ej ψ).
158 LINEAR DIFFERENTIAL OPERATORS IN RIEMANNIAN GEOMETRY
(∇E E E E E E
ej ϕ, ∇ej ψ)(x) = ∇ej (∇ej ϕ, ψ)(x) − (∇ej ∇ej ϕ, ψ)(x).
If χ ∈ C0∞ (U ) then
Z Z Z
∇ej χ dvol = hej , dχi dvol = − χ div ej dvol,
U U
For any point x ∈ U we can choose ϕ with given value and ∇ϕ = 0 at x, which makes
the right-hand side equal to 0 at x. But we have already proved that the left-hand side
only depends on ϕ(x), which completes the proof of (6.8.8).
As an example we shall derive (6.3.16) again by applying Proposition 6.8.4 to the
Dirac type operator d + d∗ in C ∞ (M, ∧∗C M ) (see (6.3.5)). Recall that by Exercise
3.1.2 and (5.1.5) the matrix of R∇ (X, Y ) for the cotangent bundle with the Levi-
Civita connection is − k,l R ikl X k Y l , i, j = 1, . . . , n, in the coordinate frame. With
j
P
n
X
(6.8.9) − 12 εk · εj · εl Ri lkj ϕi .
i,j,k,l=1
where R denotes the Ricci tensor in the last formula. This gives (6.3.16). The
Weitzenböck formulas for forms of higher degree could be obtained in the same way,
but then we would have to work out the curvature form in the higher exterior powers
using Exercise 5.4.1.
We shall now pass to the main examples of Dirac operators, based on the half spin
representations D±1 in (6.7.15). These are not representations of SO(2k) but of the
2
covering group Spin(2k), so we cannot define associated vector bundles for an arbitrary
Riemannian manifold.
Definition 6.8.5. An oriented Riemannian manifold M of dimension n has a spin struc-
ture if there exists a principal Spin(n) bundle Pe on M , which is a double cover of the
oriented orthonormal frame bundle P , such that with the map Spin(n) → SO(n) de-
fined by (6.7.12) we have a commutative diagram, with horizontal arrows defined by
the right group action,
Pe × Spin(n) −−−−→ Pe
y y
P × SO(n) −−−−→ P.
and the full spin bundle S(Pe) = S+ (Pe) ⊕ S− (Pe) are defined. The latter is associated
with the direct sum ̺ of the representations D± 1 , which is the restriction of (6.7.14)
2
to Spin(2k). The Clifford bundle can be regarded as the bundle associated with the
representation of Spin(2k) on Cl(2k) by the representation
for this is a representation since a∗ is the inverse of a ∈ Spin(2k), and it gives the
orthogonal transformation τ (a) when applied to an element in R2k .
160 LINEAR DIFFERENTIAL OPERATORS IN RIEMANNIAN GEOMETRY
Proposition 6.8.6. The spin bundle S(Pe) is a Hermitian Clifford bundle, and the
Levi-Civita connection is a Clifford connection in S(P̃ ). For the corresponding Dirac
operator we have the Lichnerowicz formula
(6.8.11) D2 + ∇∗ ∇ = −S/4,
and (pa−1 , x′ , w′ ) with a ∈ Spin(2k) define the same element in Clx (M ) × Sx (Pe), then
x′ = a · x · a∗ and w′ = a · w, which implies that x′ · w′ = a · (x · w), hence that (p, x · w)
and (pa−1 , x′ · w′ ) = (pa−1 , a · (x · w)) define the same element in S(P̃ ). It is clear
that the multiplication so defined depends smoothly on x. Multiplication by a vector
v ∈ Tx∗ M is self-adjoint, since by our original definition it is the sum of an operator
and its adjoint, so S(Pe ) is a Clifford module. At the center of a geodesic coordinate
system the Levi-Civita connection cannot be distinguished from the flat connection in
any of the bundles involved, so it is obvious that we have a Clifford connection.
With Ri jkl denoting the components of the Riemannian curvature tensor in the
oriented orthonormal frame ε1 , . . . , εn of Tx∗ , with dual frame e1 , . . . , en in Tx , we
have seen above that the curvature R∇ (ek , ej ) of the cotangent bundle has the skew
symmetric matrix −(Rl ikj )l=1,...,n l=1,...,n
i=1,...,n = −(Rlikj )i=1,...,n . By Theorem 6.7.11 this skew
symmetric matrix as an element in so(n) corresponds in the Lie algebra of Spin(n) to
1
P
− 4 i,l Rlikj εi · εl . Hence the right-hand side of (6.8.8) becomes
X X
− 81 Rlikj εk · εj · εi · εl = − 14 εj Rlj · εl = − 14 S,
i,j,k,l
k∇ϕk2 + 41 (Sϕ, ϕ) = 0.
where I is the identity in the spin bundle. We shall need (6.8.12) in Section 6.10.
6.9. Hermite polynomials and Mehler’s formula. The proof of the local index
formula for twisted Dirac operators will culminate in an identification of the difference
of the traces of the heat kernels involved with the heat kernel belonging to an operator
closely related to the harmonic oscillator. As a preparation we shall now give the
classical background.
The Hermite polynomial Hn (x), x ∈ R, of order n is defined by
2 2
(6.9.1) Hn (x) = ex (−d/dx)n (e−x ) = (2x)n + . . . ,
2
which by Taylor’s formula for e−(x−z) implies that we have the generating function
∞
X Hn (x) n 2
(6.9.2) z = e2xz−z .
n=0
n!
2
They are orthogonal with respect to the weight function e−x , for if n < m we have
Z Z
−x2 2
Hn (x)Hm (x)e dx = Hn (x)(−d/dx)m e−x dx = 0
R R
Summing up,
√
Z
2
(6.9.3) Hn (x)Hm (x)e−x dx = δnm 2n n! π
R
1 2
p √
Hence the functions e− 2 x Hn (x)/ 2n n! π form an orthonormal system in L2 (R). It
1 2
is complete, for if u ∈ L2 (R) is orthogonal to all of them, then U (x) = u(x)e− 2 x is
orthogonal to all polynomials; the Fourier transform is then an entire analytic function
with all derivatives equal to 0 at 0, hence identically 0, which proves that u = 0.
If in the identity
∞
1 2 1 2 2
X
Hn (x)e− 2 x z n /n! = e− 2 x +2xz−z
0
P n n
The right-hand side is obtained from the power series h z /n! = ehz when hn is
replaced by hn . Such a substitution is legitimate in any equality between two power
series in h provided that no convergence difficulties occur, for the coefficients of corre-
sponding powers of h are identical. If we apply this to the identity
∞
X
2xz hz 2xz+hz
e e =e = (2x + h)n z n /n!
0
Now we wish to determine the kernel of e−tL when t > 0, which is equal to
X 1 2
+y 2 )
√
e−2tn Hn (x)Hn (y)e− 2 (x /(2n n! π).
The expansion of the right-hand side is not obvious unless the coefficient of h vanishes,
that is, y = xs, but then we can use that
∞ ∞
−h2 s2 /4
X
n 2 n
X (2n)!
Ee = (−1) h2n (s /4) /n! = (s2 /4)n .
0 0
n!2
13 − 12
Since (2n)!/(4n n!2 ) = . . . ( 21 + n − 1)/n! = (−1)n
22 n
, we conclude that
2 2
p
s /4
(6.9.10) Ee−h = 1/ 1 − s2 .
In view of the symmetry we may exchange x and y in our earlier calculations which
gives
2 2 2 2 2 2 2
p
Ee−h s /4+h(xs−ys )+2xys−y s = e2xys−x s / 1 − s2 , if y = xs.
With the notation v = xs − ys2 = xs(1 − s2 ) we have (y 2 − x2 )s2 = x2 (s2 − 1)s2 =
−v 2 /(1 − s2 ) and conclude that for any v
2 2 2
/(1−s2 )
p
s /4+hv
(6.9.11) Ee−h = e−v / 1 − s2 .
This simplifies to
∞
2 2
s +y 2 s2 −2xys)/(1−s2 )
X p
(6.9.12) Hn (x)Hn (y)(s/2)n/n! = e−(x / 1 − s2 ,
0
which is known as Mehler’s formula. I owe the trick in the proof to Marcel Riesz who
showed it to me in a private conversation with him in 1952. My notes state that the
proof goes back to the 1920’s. I have inserted the operator E to justify the formal
argument.
From (6.9.12) it follows that
∞
2
− 12 1
+y 2 )
X
π Hn (x)Hn (y)(s/2)ne− 2 (x /n!
0
2
1
+y 2 )(1+s2 )−4xys)/(1−s2 )
p
= e− 2 ((x / π(1 − s2 ),
and with s = e−2t we obtain the kernel of e−tL as
√
exp − ( 21 (x2 + y 2 ) cosh 2t − xy)/ sinh 2t et / 2π sinh 2t.
The factor et is caused by the term −1 in L = −d2 /dx2 + x2 − 1. If we define
L1 = −d2 /dx2 + x2 , it follows that the kernel of e−tL1 is
√
K(t, x, y) = exp − 12 ((x2 + y 2 ) cosh 2t − 2xy)/ sinh 2t) / 2π sinh 2t.
√
To find the kernel of the operator La = −d2 /dx2 + a2 x2 we take x̃ = x a as new
variable so that La = aL e 1 . Then
√
Z
−tLa e1 ˜
−taL
(e f )(x) = (e f )(x̃) = K(ta, x̃, ỹ)f (ỹ/ a)dỹ
√ √ √
Z
= K(ta, x a, y a)f (y) a dy,
√ √ √
so the kernel of e−tLa is aK(ta, x a, y a), that is,
r
1 2at 1 2at 2 2
(6.9.13) √ exp − (x + y ) cosh 2at − 2xy .
4πt sinh 2at 4t sinh 2at
This formula gives immediately the kernel of a harmonic oscillator
−∆ + hAx, xi
when A is a positive diagonal matrix, the only difference is that we get a product of
such factors. By the orthogonal invariance the kernel is therefore in general, when the
dimension is n,
s √
1 2 At
(6.9.14) n det √
(4πt) 2 sinh 2 At
√ √ √
1 2 At 2 At 2 At
× exp − √ x, x + √ y, y − 2 √ x, y .
4t tanh 2 At tanh 2 At sinh 2 At
HERMITE POLYNOMIALS AND MEHLER’S FORMULA 165
where Ω is a real skew symmetric matrix and i is the imaginary unit. At least formally
we can write L = L0 + L1 , where
n
X X
L0 = − ∂j2 + hΩx, Ωxi, L1 = −2i Ωjk xk ∂j .
1
Formally this means that e−tL = e−tL0 e−tL1 , but this is rather delicate since L1 is a
formally self-adjoint operator not bounded from below. To proceed, at first formally,
we shall now assume that n = 2, which is no essential restriction since every skew
symmetric Ω is the direct sum of such operators and trivial ones.
With n = 2 we now assume that
2
0 a 2 a 0
Ω= , hence − Ω = ,
−a 0 0 a2
L0 = −∆ + a2 |x|2 , L1 = −2ihΩx, ∂i.
The vector field hΩx, ∂i = ax2 ∂/∂x1 − ax1 ∂/∂x2 is infinitesimal generator of rotation
with the speed a, so for real s we conclude that eshΩx,∂i is rotation by the angle sa,
that is, composition with the map x 7→ (x1 cos(sa) −x2 sin(sa), x1 sin(sa) +x2 cos(sa)),
so we can calculate eshOx,∂i e−tL0 by just composing in the x variables with this map.
By analytic continuation to s = 2it this gives for the inner parenthesis in the exponent
By analytic continuation
√ it follows at once that ∂/∂t + Lx annihilates the kernel, and
taking x = y + z t we find at once that it converges to δ as t → 0, which easily
justifies that we√have indeed found the fundamental solution. With the suggestive
notation |Ω| = −Ω2 it follows that the heat kernel for the operator (6.9.15) with a
general real skew symmetric Ω is
s
1 2|Ω|t 1 2|Ω|t
(6.9.16) n det exp − (x − y), x − y + 4ithΩx, yi .
(4πt) 2 sinh 2|Ω|t 4t tanh 2|Ω|t
Note that z/ sinh z and z/ tanh z are analytic functions of z 2 in a neighborhood of the
real axis. For small t the right-hand side of (6.9.16) is therefore well defined even if Ωjk
are not real valued but take their values in some commutative finite dimensional alge-
bra, such as ∧∗ Ck , and it will still be a fundamental solution. This will be important
in Section 6.10.
6.10. The local index formula for twisted Dirac operators. Let M be a
Riemannian manifold of even dimension n with a spin structure, and let F be a Her-
mitian vector bundle on M with a connection compatible with the metric. At the end
of Section 6.8 we defined the skew adjoint twisted Dirac operator DF in C ∞ (M, E)
where E = E+ ⊕ E− and E± = S± (Pe) ⊗ F . It maps sections of E± to sections of E∓ .
By Proposition 6.5.1 we know that the index of DF+ : C ∞ (M, E+ ) → C ∞ (M, E− ) for
2
t > 0 is equal to the difference between the trace of etDF on C ∞ (M, E+) and the trace
2
on C ∞ (M, E− ). Thus it is the supertrace of etDF on C ∞ (M, E) with the grading by
2
E+ and E− . Let K(t, x, y) ∈ Hom(Ey , Ex ) be the kernel of etDF . We want to show
in analogy to (6.5.5)′ and (6.5.6)′ that the supertrace of K(t, x, x) for fixed x, without
integration over M , has an asymptotic expansion in non-negative integer powers of
t and calculate the constant term. The crucial point is the Berezin-Patodi formula
(Theorem 6.7.14), which has an obvious extension to Clx (M )C ⊗ Fx . Let
be the natural filtration of the even part of End E; recall that End S(P̃ ) ∼
= Cl(M )C .
(j)
When j ≥ n/2 then Lx is the whole even part.
We have seen in Section 6.4 that using geodesic coordinates centered at y and
synchronous frames for E and T ∗ M , hence for S(Pe), we have for fixed y an asymptotic
expansion
∞
X
K(t, x, y) ∼ H0 (t, x) (−t)ν uν (x)/ν!
0
1
where H0 is the fundamental solution of the scalar heat equation, u0 = g(x)− 4 and
u1 , u2 , . . . are determined successively by integrating the equations (6.1.14) (or equiv-
alently (6.1.14)′ ). What makes the supertrace on the diagonal accessible to explicit
calculation is the following fact:
THE LOCAL INDEX FORMULA FOR TWISTED DIRAC OPERATORS 167
(ν)
Lemma 6.10.1. The coefficients uν (0) are in Ly for ν = 0, 1, . . .
This is clear when ν = 0 and will be proved by induction for increasing ν. The
statement is void when ν ≥ n/2, but in the course of the proof we shall isolate the
terms which are not better than the lemma states, and when ν = n/2 this will give uν
mod Lk−1 y , which is all that we need. Before the proof we need some preliminaries. The
first point is to express DF2 in terms of geodesic coordinates and synchronous frames
ε1 , . . . , εn for T ∗ M , f1 , . . . , fN for F . With the dual frame e1 , . . . , en in T M we write
for l = 1, . . . , n
n
X N
X
(6.10.1) ∇el εj = e lj k εk , j = 1, . . . , n,
Γ ∇e l f j = ΓF k
lj fk , j = 1, . . . , N.
k=1 k=1
e lj k = −Γ
Here Γ e lk j since the frame is orthonormal, and by Theorem 3.3.6 we have
n
X
(6.10.2) eklj (x) =
Γ 1
Rkjil (0)xi + O(|x|2 ).
2
i=1
(6.10.3) ΓF k
lj (0) = 0, for l = 1, . . . , n; j, k = 1, . . . , N.
X n
X
∇e l ( ϕj εj ) = hel , ∂iϕj + elj k ϕj εk .
Γ
j,k=1
The matrix (Γe lj k )j=1,...,n is in so(n) for fixed l, and as in the proof of Proposition 6.8.6
k=1,...,n
it corresponds to multiplication by − 1
P e k
4
Γlj εj · εk in the Clifford algebra, so
j,k
n
X n
X N
X
1 e lj k εj · εk + ΓF k
(6.10.4) ∇ϕ = εl ⊗ hel , ∂i − 4 Γ lj · ϕ.
l=1 j,k=1 j,k=1
P
To find the adjoint acting on ψ = εl ⊗ ψl we form
n
X n
X N
X
1 elj k εj · εk + ΓF k
(ψ, ∇ϕ) = (ψl , hel , ∂i − 4 Γ lj · ϕ).
l=1 j,k=1 j,k=1
168 LINEAR DIFFERENTIAL OPERATORS IN RIEMANNIAN GEOMETRY
Here the scalar product is of course taken using the Riemannian volume form, which
is 1 + O(|x|2 ) times the Lebesgue measure. Hence we obtain by partial integration
X n
X N
X
∗ 1 k eF k
∇ ψ=− hel , ∂i − Γlj εj · εk + Γ ψl ,
e
e
4 lj
j,k=1 j,k=1
where Γe lj k and Γ
e eF k also satisfy (6.10.2), (6.10.3) but include the terms arising when
lj
derivatives fall on el or on the Riemannian volume density. Summing up, we obtain in
view of (6.8.12)
n
X F
(6.10.5) DF2 = − 14 S − 1
2 εi · εj R∇ (ei , ej )
i,j=1
n
X n
X N
X n
X N
X
k Fk k
1 1
ΓF k
+ hel , ∂i − Γlj εj · εk + Γlj hel , ∂i − 4 Γlj εj · εk + .
e
e e e
4 lj
l=1 j,k=1 j,k=1 j,k=1 j,k=1
[2]
The coefficients can be viewed as elements in ClC ⊗ End Fy or just End Fy .
Proof of Lemma 6.10.1. We are now ready to prove that for the coefficients uν con-
structed for DF2 as in Section 6.1 we have
X
(6.10.6) uν = uνµ , uνµ ∈ L(µ)
y , uνµ = O(|x|2µ−2ν ), if µ > ν.
µ≤2ν
This is obvious when ν = 0. Assuming that (6.10.6) holds for a certain ν we shall
prove that
X
(6.10.6)′ DF2 uν = vνµ , vνµ ∈ L(µ)
y , vνµ = O(|x|2µ−2ν−2 ), if µ > ν + 1.
µ≤2(ν+1)
This is obvious for the terms coming from the terms in (6.10.5) after the first sum. To
handle the others we observe that multiplication by ΓF k
lj does not raise the Clifford
degree but increases the order of the zeros. Multiplication by Γ̃lj k εj · εk may raise the
Clifford degree by at most 2, but at the same time one gets a factor vanishing at 0.
Differentiation reduces the order of the zero but does not affect the Clifford degree.
Altogether, in the terms coming from the first sum in (6.10.5) where the Clifford degree
is raised 4 units we also get a compensating factor O(|x|2 ), and O(|x|2µ−2ν+2 ) =
O(|x|2(µ+2)−2(ν+1) ). In those where the Clifford degree is raised 2 units we do not
lose any zero at 0, and O(|x|2µ−2ν ) = O(|x|2(µ+1)−2(ν+1) ). In the terms where the
Clifford degree is not raised we cannot lose more than two zeros, and O(|x|2µ−2ν−2 ) =
O(|x|2µ−2(ν+1) ), which completes the proof of (6.10.6)′ .
The integration of the transport equations (6.1.14) for uν+1 , using (6.1.14)′ and
(6.1.17), does not affect the order of the zeros at 0, so (6.10.6) follows by induction,
which proves the lemma.
THE LOCAL INDEX FORMULA FOR TWISTED DIRAC OPERATORS 169
In the proof the estimate for the order of the zeros was always too low except for
terms coming from the simplified operator
n n n
X X X F
1 i 2 1
(6.10.7) (∂l − 8 Rkjil (y)x εj · εk ) − 2 εi · εj R∇ (ei , ej )(y),
l=1 i,j,k=1 i,j=1
where we have put in the argument y instead of 0 since 0 was the geodesic coordinate
of y. Hence limt→0 Str K(t, y, y) exists and depends only on the Riemann curvature
tensor and the curvature forms of F at y.
An explicit computation can be obtained from the results related to Mehler’s for-
mula given in Section 6.9. As a first step we observe that in the computation of the
component of un/2 in Cl[n]y C ⊗ End Fy there will be no contributions where two equal
factors εi in the Clifford algebra have been multiplied, and different εi anticommute.
The component is therefore equal to the term of degree n obtained if one replaces
Clifford multiplication by exterior multiplication throughout. With the notation
n
X F
(6.10.8) Ωli = − 18 Rkjil (y)dxj ∧ dxk , ΩF = − 12 R∇ (ei , ej )(y)dxi ∧ dxj ,
j,k=1
acts on functions in Rn with values in (∧∗ Cn ) ⊗ Fy . The second term acts only in the
F
factor Fy and commutes with the first, so it only contributes to etL a factor etΩ , and
for the other part we get the fundamental solution from (6.9.16) with Ω replaced by
Ω/i. This means that |Ω|2 is replaced by Ω2 , so the kernel with the pole at 0 is
r
1 2Ωt 1 D 2Ωt E F
(6.10.10) n det exp − x, x etΩ ,
(4πt) 2 sinh 2Ωt 4t tanh 2Ωt
which simplifies to
r
−n 2Ωt F
(6.10.10) ′
(4πt) 2 det etΩ
sinh 2Ωt
it follows that
n
X
xj = cjk x′k
1
where X
εk ( xj ej ) = xk , that is, εk (ej ) = δjk .
Thus ε1 , . . . , εn form a dual basis in V ′ which is also a vector space of dimension n
with the natural definition of addition of linear forms and multiplication of them by
scalars. When ξ ∈ V ′ and x ∈ V we usually write hx, ξi for the linear form ξ(x) to
emphasize that it is bilinear, that is, linear in x for fixed ξ (because ξ is a linear form)
and linear in ξ for fixed x (by definition of the vector operations in V ′ ). The bilinear
form is non-degenerate, that is
hx, ξi = 0 ∀x ∈ V =⇒ ξ = 0; hx, ξi = 0 ∀ξ ∈ V ′ =⇒ x = 0.
171
172 APPENDIX A. PREREQUISITES FROM MULTILINEAR ALGEBRA
for the left-hand side is for fixed η a linear form on V1 , hence an element T ′ η ∈ V1′ ,
and it depends linearly on η. T1 and T1′ have the same rank.
If V and W are two vector spaces and B : V × W ∋ x, y 7→ B(x, y) ∈ K is a bilinear
form, then we get in the same way a map W → V ′ and an adjoint map V → W ′ .
The form is called non-degenerate if one (and therefore both) are bijective. Thus the
bilinear forms on V × W can be identified either with the linear maps L(W, V ′ ) or the
linear maps L(V, W ′ ).
Every bilinear form B on V ×V can in one and only one way be written B = B0 +B1
where
B0 (x, y) = B0 (y, x), x, y ∈ V ; B1 (x, y) = −B1 (y, x), x, y ∈ V ;
one calls B0 symmetric and B1 skew symmetric, and we have
From the matrix (bjk ) for B one obtains the matrices forPB0 and B1 as (bjk ± bkj )/2.
Only B0 is determined by the quadratic form B(x, x) = bjk xj xk ; we have
Thus we have a one to one correspondence between quadratic forms and symmetric
bilinear forms. A quadratic form Q is called non-degenerate if the corresponding
symmetric bilinear form B is non-degenerate. If the bilinear form is degenerate we
can find z 6= 0 so that B(V, z) = 0, B(z, V ) = 0, which implies that Q(x + tz) =
B(x + tz, x + tz) = B(x, x) = Q(x) for any t ∈ K, so Q is defined in the quotient space
V /Kz. In what follows we shall use the same notation for a quadratic form and the
corresponding symmetric bilinear form.
If G is a non-degenerate quadratic form in V , then the polarized form defines a
bijection g : V → V ′ and we can define a dual quadratic form G′ in V ′ by
G′ (gx) = G(x), x ∈ V.
If we introduce dual bases in V and V ′ and (gjk ) is the symmetric matrix for G, then
the matrix for G′ is (g jk ) if (g jk ) is the inverse of the matrix gjk .
APPENDIX A. PREREQUISITES FROM MULTILINEAR ALGEBRA 173
which proves the existence of the form Tr with the required properites, independent
of the choice
Pn of basis. Note that if Tjk is the matrix of T in terms of a basis then
Tr T = 1 Tjj .
Let V and W be two finite dimensional vector spaces over K. The bilinear forms on
V × W ′ form another vector space called the tensor product of V and W and denoted
′
This proves that V ⊗ W is a vector space of dimension dim V dim W with basis ej ⊗ fk ,
j = 1, . . . , n, k = 1, . . . , N . If x1 , . . . , xn are the coordinates of an element in V in
the basis e1 , . . . , en and y1 , . . . , yN are the coordinates of an element in W in the basis
f1 , . . . , fN , then xj yk are the coordinates of x ⊗ y.
If T is a linear map from V ⊗ W to a third vector space Z, then
S : V × W ∋ x, y 7→ T (x ⊗ y) ∈ Z
L(V, W ) ∼
= W ⊗ V ′; in particular, L(V, V ) ∼
= V ⊗ V ′.
We can now look at the trace in a new way. The bilinear form
V × V ′ ∋ x, ξ 7→ hx, ξi
W × V × V ′ ∋ w, x, ξ 7→ whx, ξi ∈ W
Pn
coordinates of the contraction are j=1 Tijj . This contraction operation can of course
be applied to any tensor product
W1 ⊗ · · · ⊗ V ⊗ · · · ⊗ V ′ ⊗ Wk
containing two dual vector spaces; contraction gives an element in the tensor product
where they are both removed.
In Riemannian geometry where V is the tangent space and V ′ the cotangent space, it
is customary to put coordinate indices corresponding to factors V in a tensor product as
superscripts and to put indices corresponding to factors V ′ as subscripts. Contraction
then means putting a superscript equal to a subscript and summing over it.
If B is a bilinear form in V ⊗V and a non-degenerate quadratic form G is given in V ,
then we can define the trace of B with respect to G as follows: We have B ∈ V ′ ⊗ V ′ ,
and G gives an identification V ∼ = V ′ , so B can be identified with an element in V ⊗ V ′
for which the trace is defined. In terms
P of kj the matrices (bjk ) and (gjk ) for B and G
with respect to a basis, the trace is bjk g where (g jk ) is the inverse of the matrix
(gjk ).
We have defined the tensor product V ⊗V as the space of bilinear forms on V ′ . Now
every bilinear form on V ′ can be written as the sum of one symmetric and one skew
symmetric bilinear form in one and only one way. We shall denote by S 2 (V ) ⊂ V ⊗ V
the space of symmetric bilinear forms on V ′ and by ∧2 V ⊂ V ⊗ V the space of skew
symmetric bilinear forms on V ′ . Thus the tensor product V ⊗ V is the direct sum of
the symmetric tensor product S 2 (V ) and the exterior product ∧2 V ,
V ⊗ V = S 2 (V ) ⊕ ∧2 V.
If L ∈ S k (V ), then
e : V ′ ∋ ξ 7→ L(ξ, . . . , ξ)
L
is a homogeneous polynomial of degree k in the sense that
(t1 , . . . , tj ) 7→ L(t
e 1 ξ 1 + · · · + tj ξ j )
e 1 ξ + · · · + tk ξ) = (t1 + · · · + tk )k L(ξ),
L(t e
176 APPENDIX A. PREREQUISITES FROM MULTILINEAR ALGEBRA
V k ∋ (x1 , . . . , xk ) 7→ x1 ∧ · · · ∧ xk
is for fixed xk+1 , . . . , xk+κ an alternating multilinear map from V k , so it defines a linear
map
T (xk+1 , . . . , xk+κ ) : ∧k V → ∧k+κ V.
For every w ∈ ∧k V the map
such that
(A.5) w ∧ w′ = (−1)kκ w′ ∧ w, if w ∈ ∧k V, w′ ∈ ∧κ V.
178 APPENDIX A. PREREQUISITES FROM MULTILINEAR ALGEBRA
Note that w and w′ commute unless the degrees k and κ of w and w′ are both odd;
then they anticommute.
If V1 and V2 are finite dimensional vector spaces, then every linear map A : V1 → V2
induces a linear map ∧k A : ∧k V1 → ∧k V2 . In fact, the map
We can define the exterior powers ∧k V ′ of the dual V ′ of V in the same way. The
multilinear map
is alternating in the first k and in the last k variables. Hence it induces a bilinear form
on ∧k V × ∧k V ′ . It defines a duality for if e1 , . . . , en and ε1 , . . . , εn are dual bases for V
and for V ′ , then the bases ej1 ∧ · · · ∧ ejk and εi1 ∧ · · · ∧ εik with 1 ≤ j1 < · · · < jk ≤ n
and 1 ≤ i1 < · · · < ik ≤ n are dual with respect to this form. Thus we have a natural
duality between ∧k V and ∧k V ′ such that
(Many authors use another definition where the right-hand side is divided by k!, for
that is the duality inherited from the natural duality of the tensor products. See
Sternberg [1, p. 19] for a discussion of this point. Kobayashi and Nomizu [1] use the
division by k!, which should be kept in mind when comparing identities.)
APPENDIX B. THE CALCULUS OF DIFFERENTIAL FORMS
Tx∗ M ∼= Rn , and ∧k Tx∗ M has the basis dxj1 ∧ · · · ∧ dxjk , where 1 ≤ j1 < · · · < jk ≤ n;
this is the alternating multilinear form
For any C 1 map f : N → M , where N is another smooth manifold, the adjoint of the
differential
f ′ (y) : Ty N → Tf (y) M
is a map Tf∗(y) M → Ty∗ N inducing a map
f ∗ : λk (M ) → λk (N ).
where df j (y) =
P j
∂f (y)/∂y idy i . In fact, if ϕ is a function on M (such as a local
coordinate), then
by the chain rule, which means that f ∗ dϕ = d(f ∗ ϕ). We have natural rules of com-
putation such as (f g)∗ = g ∗ f ∗ ; the notation with the upper star is meant to be a
reminder of this reversal of factors.
We shall now define the exterior differential of a k form. We do this first in a
fixed local coordinate system and verify the independence of the chosen coordinates
afterwards. For the form u in (B.2) we thus define
X
(B.4) du = daj1 ...jk (x) ∧ dxj1 ∧ · · · ∧ dxjk .
1≤j1 <···<jk ≤n
in view of the linearity it suffices to verify (B.5) when u = adxj1 ∧ · · · ∧ dxjk and
v = bdxjk+1 ∧ · · · ∧ dxjk+κ , and then it follows from the fact that d(ab) = adb + bda;
commutation of db through u gives the sign (−1)k . If u is a smooth function we have
(B.6) d2 u = 0.
∂j udxj , so
P
In fact, du =
X X
d2 u = d(∂j u) ∧ dxj = ∂k ∂j udxk ∧ dxj = 0
which proves (B.7). In particular, it follows from (B.7) that our definition (B.4) of the
exterior differential is independent of the choice of local coordinates.
Locally there is a converse of (B.6):
APPENDIX B. THE CALCULUS OF DIFFERENTIAL FORMS 181
Theorem B.1 (Poincaré’s lemma). Let v ∈ λk+1 (X) where X is a convex open
set in Rn , assume that the coefficients of v are in C µ and that dv = 0. Then there is
a form u ∈ λk (X) with C µ coefficients such that du = v.
Proof. We may assume that 0 ∈ X. Then X b = {(x, t) ∈ Rn × R; tx ∈ X} is an
open set containing X × [0, 1], and f (x, t) = tx is a C ∞ map from X b to X, so we can
form
f ∗ v = ft∗ v + dt ∧ wt
where ft∗ is the pullback of v when t is regarded as a parameter, so it is a differential
form which only involves the differentials of the coordinates in X. Since df ∗ v = f ∗ dv =
0 it follows that
0 = dt ∧ (∂(ft∗ v)/∂t − dx wt ) + . . .
where the dots indicate a form which has no factor dt and dx wt is the differential of
wt for fixed t. Hence
∂(ft∗ v)/∂t = dx wt ,
and integration from 0 to 1 gives
Z 1
f1∗ v − f0∗ v = du, u= wt dt.
0
But f1∗ v = v and f0∗ v = 0 since f1 is the identity and f0 maps X to {0}, which proves
the theorem.
Poincaré’s lemma reflects the fact that the topology of a convex set in Rn is very
simple; in general there is a topological obstruction:
Definition B.2. If X is a smooth manifold then the quotient H k (X) of the closed
forms {u ∈ λk (X); du = 0} by the linear subspace {dv; v ∈ λk−1 (X)} of exact forms,
k > 0, is called the de Rham cohomology of degree k. The residue class in H k (X) of
a form u ∈ λk (X) with du = 0 is called the cohomology class of u.
Exercise B.1. Prove that if M and N are smooth manifolds, f : M × [0, 1] → N
is a smooth map, and u ∈ λk (N ) is a closed form, then the forms f (·, t)∗u ∈ λk (M )
are in the same cohomology class for all t ∈ [0, 1].
Poincaré’s lemma is closely related to Stokes’ formula, which we shall now discuss.
First we define the integral of a form u ∈ λn (M ) with compact support over M when
M is an oriented manifold of dimension n. To do so we first assume that the support
is contained in a local coordinate patch with positively oriented local coordinates x ∈
ω ⊂ Rn . In terms of these coordinates we can then write
u = a(x)dx1 ∧ · · · ∧ dxn
and we define Z Z
u= a(x)dx,
182 APPENDIX B. THE CALCULUS OF DIFFERENTIAL FORMS
The positive orientation means precisely that the Jacobian here is positive, so Rthe
definitions in terms of the two coordinate systems
P agree. Now we can define u
in general by using a partition of unity 1 =
P ϕj to split u into a finite sum u =
ϕj u where each term has support in a local coordinate patch and using these local
coordinates set
XZ
u= ϕj u.
Remark. We have not given a precise definition of orientation above. One way
to do so is to say that an orientation on a manifold of dimension n is a n form o,
the orientation form, which is everywhere different from 0; orientation forms differing
by multiplication with a positive function are considered equivalent. A system of
local coordinates x1 , . . . , xn is then positively oriented if with these coordinates o =
g(x)dx1 ∧ · · · ∧ dxn where g(x) > 0. The manifold is said to be oriented by o > 0.
n
X
u= (−1)j−1 uj (x)dx1 ∧ · · · ∧ dxj−1 ∧ dxj+1 ∧ · · · ∧ dxn ,
1
n
X
du = ( ∂j uj )dx1 ∧ · · · ∧ dxn ,
1
which is obvious.
Now we can orient ∂X uniquely by using at points in ∂X positively oriented local
coordinates in M such that X is located as just described where x1 < 0; the coordinates
x2 , . . . , xn in ∂X are then by definition positively oriented. From the local formula
(B.8) it follows by using a partition of unity that
Z Z
(B.9) u= du, Stokes’ formula,
∂X X
To study flat structures we shall need the Frobenius existence theorem for first order
systems of differential equations. One of the standard forms of this result is as follows:
Theorem C.1. Let v1 , . . . , vr be C ∞ vector fields in a neighborhood of 0 in Rn
such that
where [v, w] denotes the commutator of v and w, and cijk ∈ C ∞ . Then there exist new
local coordinates y1 , . . . , yn in a neighborhood of 0 such that
r
X
∂/∂yi = bij vj , i = 1, . . . , r.
j=1
n
X
vj = vjl ∂/∂xl , j = 2, . . . , r.
l=2
By the inductive hypothesis we may also assume that vjl = 0 for j = 2, . . . , r if l > r
and x1 = 0. By (C.1)
r
X r
X
∂vjl /∂x1 = [v1 , vj ]xl = c1jk vk xl = c1jk vkl , j = 2, . . . , r.
k=2 k=2
(C.3) ωj = 0, j = r + 1, . . . , n,
where ωr+1 , . . . , ωn are linearly independent differential one forms at the origin. We
want to rewrite the Frobenius condition (C.2) in terms of these forms. To do so we
need a lemma:
Lemma C.2. If X and Y are C 1 vector fields and ω is a C 1 one form, then
Proof. If ω = du for some u then (C.4) is the definition of the commutator vector
field. If ω is replaced by ϕω, then both sides are multiplied by ϕ and in addition we
get on the left-hand side a term hX ∧ Y, dϕ ∧ ωi which is equal to the additional term
hvj , ωk i = δjk , j, k = 1, . . . , n.
Then
n
X n
X
(C.5) [vi , vj ] = cijk vk , i, j = 1, . . . , n =⇒ dωk = − 12 cijk ωi ∧ ωj ,
k=1 i,j=1
where Fµν are smooth functions defined in a neighborhood of (x0 , y0 ) ∈ Rn+m . The
system is called completely integrable if there is a smooth solution in a neighborhood
of x0 with y(x0 ) given in a neighborhood of y0 . Since the equations give
m
X
∂ 2 yµ /∂xν ∂xκ = ∂Fµν /∂xκ + ∂Fµν /∂yσ Fσκ
σ=1
This condition is also sufficient. In fact, the equations (C.6) mean precisely that the
graph of y(x) shall be defined by the equations
n
X
ωµ = dyµ − Fµν dxν = 0.
ν=1
Since
X X X
dωµ = − ∂Fµν /∂xκ dxκ ∧ dxν − ∂Fµν /∂yσ (ωσ + Fσκ dxκ ) ∧ dxν
κ,ν
X
=− ∂Fµν /∂yσ ωσ
Sometimes f is then called an even invariant, and one calls f an odd invariant if
where SO(n) = {O ∈ O(n); det O = 1}. On the other hand, it follows from (D.3) that
The diagonal elements are here Dk,k + (k − 1), the off diagonal elements are Dj,k , and
the expansion of the determinant shall be made so that the elements are multiplied in
the order of their columns. To prove (D.4) we first consider the “minors” formed from
the last two columns. We have
Dj,2 Dk,1 − Dk,2 Dj,1 = Dj,2 #Dk,1 − Dk,2 #Dj,1 + δk,2 Dj,1 − δj,2 Dk,1 ,
that is,
(Dj,2 + δj,2 )Dk,1 − (Dk,2 + δk,2 )Dj,1 = Dj,2 #Dk,1 − Dk,2 #Dj,1 .
Thus the minors of (D.4) in the left-hand side are the same as in a determinant with
diagonal elements Dk,k expanded using the formal product. We shall prove by induc-
tion for increasing k that this is true for the minors taken from the last k ≥ 2 columns.
When k = N this will prove (D.4), for
0, if N > n
dethxj , ξk iN
j,k=1 =
det(x1 , . . . , xn ) det(ξ1 , . . . , ξn ), if N = n.
We shall let i1 , . . . , ik run through all permutations of k fixed indices, multiply by the
sign of the permutation and sum. Now the permutations
have opposite signs since they differ by one inversion. Hence the terms in the last sum
will cancel the preceding one, which completes the proof of the inductive statement
and hence of the Capelli identity.
End of Proof of Theorem D.1. With the Capelli identity available we can now
finish the proof. By the first part of the proof we may assume that N ≥ n and that
the theorem has already been proved when N is replaced by N − 1. Furthermore we
may assume that f is homogeneous in each of the vectors x1 , . . . , xN , for if we split f
into a sum of polynomials of such separate homogeneities, then each term must be an
PN
invariant. Let rj be the degree of homogeneity with respect to xj , and let |r| = 1 rj
be the total degree of homogeneity. We order all multiindices r = (r1 , . . . , rN ) first
for increasing |r|, and then lexocographically for fixed |r|, that is, first according to
increasing r1 , then for fixed r1 according to increasing r2 , and so on. If r1 = 0 then
f depends on N − 1 vectors so the theorem is true then by inductive hypothesis. By
still another inductive argument we may therefore assume that r1 > 0 and that the
theorem has already been proved for lower values of r.
Now consider (D.4). If N = n, then g = det(∂/∂x1 , . . . , ∂/∂xn)f is an odd (even)
invariant if f is an even (odd) invariant, and the degree of g is lower than the degree
of f , so we know that g has the form stated in the theorem. Since
it follows that the right-hand side has the stated form for any N ≥ n. In the left-hand
side of (D.4) the diagonal term is cf where c = r1 (r2 + 1) . . . (rN + N − 1) 6= 0 since
r1 6= 0. All other terms contain some factor Dj,k with j 6= k. For a given term we
choose the factor Dj,k with j 6= k which is furthest to the right. To the right of it we
then have only the factors Dk−1,k−1 . . . D1,1 which multiply f by r1 . . . (rk−1 + k − 2).
We must have j > k since two factors must not be chosen in the same row. The full
expression of the term is of the form ajk Dj,k f where ajk is a product of operators
Dp,q . But Dj,k f is also an invariant and its total degree is equal to that of f while
its lexicographic order is lower since Dj,k lowers the degree in xk at the expense of
a raise of degree in xj for some j > k. Thus Dj,k has the desired form by inductive
hypothesis. So has ajk Dj,k f , for
Chapter 1. Starting with the notion of curvature for a curve, which is as old as
calculus, we study the tangent and normals associated to a curve which leads up to a
first encounter with the idea of moving orthogonal frames. It goes back to Euler and
is basic for the methods of E. Cartan presented in Chapter IV. We introduce it in the
elementary context of curves in a Euclidean space and take the opportunity to give at
the same time a first introduction to Lie groups, restricted to subgroups of the linear
group. A more extensive discussion is given in Section 5.3, and the reader who wants
more information is referred to Chevalley [1], Helgason [1], Sternberg [1], Warner [1]
and Weyl [1].
Chapters 2, 3. Principal curvatures of surfaces in R3 were also known to Euler,
but it was Gauss [1] who discovered that the total curvature is an inner property of the
surface. Reading Gauss [1] is still very pleasant, and one will recognize many of the
ideas presented in Chapters 2 and 3 there. The extension to higher dimensions outlined
in Riemann [1], with emphasis on ideas rather than formulas, is equally enjoyable to
read. The later formal development of his ideas by Bianchi, Christoffel, Ricci and
others is the main theme of Chapters 2 and 3. Our presentation is in the same spirit as
Klingenberg [1,2] and to some extent the classic Eisenhart [1,2]. One can find a good
presentation in Berger, Gauduchon and Mazet [1], and excellent summaries are given in
Aubin [1], Besse [1]. Chapter 2 ends with some detailed analysis of the curvature tensor
at a point taken from Atiyah, Hitchin and Singer [1], where much more information is
available.
To give a bridge between the study of submanifolds of Rn in Chapter 2 and abstract
manifolds in Chapter 3 we have devoted much space in Chapter 3 to the problem of
embedding an abstract manifold in a Euclidean space. After the classical results of
Cartan and Janet for the analytic case, the big advance on the problem was made by
Nash [1,2]; recently a technically simpler variation has been found by Günther [1,2].
For additional information on global embedding theorems we refer to Gromov and
Rohlin [1], Griffiths [1], Griffiths and Jensen [1], Jacobowitz [1,2] and references in
these papers.
The discussion of spaces of constant curvature in Section 3.3 has been taken to
a large extent from Kobayashi and Nomizu [1]. This is a very careful and useful
presentation of basic differential geometry, though provided with very little motivation
for the reader.
Chapter 4. In this chapter we leave the Ricci tensor calculus which dominated
Chapter 3 and introduce differential forms in the spirit of E. Cartan. The reader can
193
194 NOTES AND REFERENCES
consult Chern [1,2], Kobayashi and Nomizu [1] and Sternberg [1] for a more thorough
discussion of these topics. Section 4.3 is almost entirely taken from Chern [1].
Chapter 5. We have started the chapter with a discussion of connections in a vector
bundle, as they are encountered naturally by an analyst. However, at the end we also
give the more geometrical approach using connection forms on a principal bundle. The
purpose of this material is to give the language required to state index theorems for
elliptic differential operators and to give the basic concepts required in gauge theory.
Also in this chapter one can consult Kobayashi and Nomizu [1] or Sternberg [1] for
most of the topics covered.
Chapter 6. The term “metric operator” is taken from Günther [3], and Proposition
6.1.2 can also be found there. The Hadamard construction of parametrices (Hadamard
[1], see also e.g. Hörmander [1, Chapter 17]) is made more transparent by using the
connection provided by the operator itself, and this is useful in the proof of the local
index theorem in Section 6.10. Section 6.2 is devoted to the algebra of differential
forms, and Hodge theory is presented in Section 6.3. For a more detailed discussion
of differential forms on a Riemannian manifold including Hodge theory, the reader can
also consult de Rham [1] or Warner [1]. In Section 6.4 the Hadamard construction of
parametrices is adapted to heat equations associated to metric operators. This gives
the analytical tools required for the discussion of Hirzebruch’s signature theorem in
Section 6.5. (For additional background to this result one should consult Hirzebruch
[1].) The presentation here follows Atiyah, Bott and Patodi [1] in principle, but we
have substituted the parametrix construction of Hadamard for the application of pseu-
dodifferential operators. Sections 6.6–6.10 are devoted to a direct proof of the local
index formula for Dirac operators, following Bismut [1] and particularly Getzler [1, 2].
The presentation owes much to Roe [1], and we have also benefited from some lecture
notes of M. Taylor.
Unwritten chapters. One of the possible directions for a continuation is to cover
the solution of the Yamabe problem, following the excellent exposition in Lee and
Parker [1]. After this introduction to non-linear problems in geometry one might
study some gauge theory. Another natural direction is to study pseudo-Riemannian
manifolds of Lorentz signature. There are a number of papers exploiting conformal
invariance to prove global existence theorems for non-linear hyperbolic systems, most
recently Christodoulou and Klainerman [1] where the stability of Minkowski space un-
der small perturbations is proved. This requires of course some preparations concerning
general relativity theory. The classical survey by Pauli [1] is still very readable. More
recent information can be found in Bergmann [1], but this reference is directed towards
mathematical physicists rather than mathematicians interested in physics. With some
background from these sources, including the Schwarzschild solution, the way is also
open to discuss the interesting proof of the positive mass conjecture in Schoen and Yau
[1,2], Witten [1] and Parker and Taubes [1].
Bibliography
M. F. Atiyah [1], Classical groups and classical differential operators on manifolds, Differential op-
erators on manifolds, Edition Cremonese, Roma, 1975, CIME conference in Varenna September
1975..
M. F. Atiyah, R. Bott and V. K. Patodi [1], On the heat equation and the index theorem, Inv. math.
19 (1973), 279–330, Correction 28 (1975), 277–280..
M. F. Atiyah, R. Bott and A. Shapiro, Clifford modules, Topology 3 (1964), 3–38, Suppl. 1..
M. F. Atiyah, N. J. Hitchin and I. M. Singer [1], Self-duality in four-dimensional Riemannian geom-
etry, Proc. R. Soc. Lond. A. 362 (1978), 425–461.
T. Aubin [1], Nonlinear analysis on manifolds. Monge-Ampère equations, Springer-Verlag, New York,
Heidelberg, Berlin, 1982.
E. Berger, R. Bryant and Ph. Griffiths [1], The Gauss equations and rigidity of isometric embeddings,
Duke Math. J. 50 (1983), 803–892.
M. Berger, P. Gauduchon and E. Mazet [1], Le spectre d’une variété Riemannienne, Springer-Verlag,
1971, SLN 194..
P. G. Bergmann [1], The general theory of relativity, Handbuch der Physik IV, Springer-Verlag, 1962.
A. L. Besse [1], Einstein Manifolds, Springer-Verlag, 1987, Erg. d. Math. Band 10..
J.M. Bismut [1], The Atiyah-Singer theorem for classical elliptic operators: a probabilistic approach.
I. The index theorem., J. Func. Anal. 57 (1984), 56–99.
Yu. D. Burago and V. A. Zalgaller [1], Geometric inequalities, Springer-Verlag, 1988, Grundl.d.Math.
Wiss. 285..
S. S. Chern [1], On the curvature integral in a Riemannian manifold, Ann. of Math. 46 (1945),
674–684.
S.S. Chern [2], Topics in differential geometry, Inst. for Adv. Study, Princeton, N.J., 1951.
C. Chevalley [1], Theory of Lie groups I, Princeton University Press, Princeton, N.J., 1946.
D. Christodoulou and S. Klainerman [1], The global nonlinear stability of the Minkowski space, Ann.
of Math., Manuscript of 565 pages, to appear..
H. L. Cycon, R. G. Froese, W. Kirsch and B. Simon [1], Schrödinger operators with applications
to quantum mechanics and global geometry, Springer-Verlag, 1987, Texts and Monographs in
Physics..
N. V. Efimov [1], Generation of singularities on surfaces of negative curvature, Mat. Sb. 64 (1964),
286–320, (Russian; translation in AMS translations.).
L. Eisenhart [1], Introduction to differential geometry, Princeton University Press, Princeton, N. J.,
1940.
L. Eisenhart [2], Riemannian geometry, Princeton University Press, Princeton, N. J., 1926.
C. F. Gauss [1], Disquisitiones generales circa superficies curvas, Comm. soc. reg. scient. Gött. 6
(1823–27), German translation in Ostwald’s Klassiker der exakten Wissenschaften 5(1889), Eng-
lish translation in “General investigations of curved surfaces”, Raven Press, Hewlett, N.Y. (1965),
which also contains an earlier version..
J. C. H. Gerretsen [1], Lectures on tensor calculus and differential geometry, P. Noordhoff N. V.,
Groningen, 1962.
E. Getzler [1], Pseudodifferential operators on supermanifolds and the Atiyah-Singer index theorem,
Comm. Math. Phys. 92 (1983), 163–178.
E. Getzler [2], A short proof of the local Atiyah-Singer index theorem, Topology 25 (1986), 111–117.
P. Gilkey [1], Curvature and the eigenvalues of the Laplacian for elliptic complexes, Adv. in Math.
10 (1973), 344–382.
Ph. A. Griffiths [1], Exterior differential systems and the calculus of variations, Birkhäuser, Boston,
Basel, Stuttgart, 1983, Progress in Math. 25..
Ph. A. Griffiths and G. R. Jensen [1], Differential systems and isometric embeddings, Princeton
University Press, Princeton, N. J., 1987, Annals of Math. Studies 114..
M. L. Gromov and V. A. Rokhlin [1], Embeddings and immersions in Riemannian geometry., Russian
Math. Surveys 25:5 (1970), 1–57, Russian original in Usp. Mat. Nauk 25:5(155), 3–62 (1970)..
195
196 BIBLIOGRAPHY
Index