Distributions: Rodica D. Costin
Distributions: Rodica D. Costin
DISTRIBUTIONS
RODICA D. COSTIN
Contents
1. Setting the stage 2
1.1. Test functions 2
1.2. Distributions 3
1.3. Examples of distributions 3
1.4. Dirac’s delta function 3
2. Operations with distributions 4
2.1. Linear combinations 4
2.2. Multiplication by C ∞ functions 4
2.3. Any distribution is differentiable 4
2.4. Convolution of distributions with test functions 5
2.5. The Fourier transform of distributions 6
3. Calculus with distributions 7
3.1. Limits of distributions 7
3.2. Fundamental sequences for δ 7
3.3. A sequence of step functions 7
3.4. Other sequences 7
4. Green’s function 8
4.1. Non-homogeneous conditions can be reduced to homogeneous
conditions, but for a non-homogeneous equation 8
4.2. Superposition 9
4.3. First examples 10
4.4. Autonomous equations 12
5. Green’s function of self-adjoint Sturm-Liouville problems 13
5.1. Liouville transformation 13
5.2. Existence and uniqueness of the Greens’s function for
non-homogeneous problems are not guaranteed 16
6. Conditions for existence and uniqueness of the solution to
real-valued, non-homogeneous Sturm-Liouville problems 17
6.1. Existence of eigenvalues for real-valued Sturm-Liouville
problems 17
7. Green’s function for self-adjoint Sturm-Liouville problems 18
7.1. When the Green’s function of Sturm-Liouville problems exists
and is unique 18
7.2. When the Green’s function for self-adjoint Sturm-Liouville
problems exists and it is not unique 22
7.3. The resolvent 22
1
2 RODICA D. COSTIN
8. Continuous spectrum 24
8.1. Example: Bessel operator 26
d
8.2. A cautionary tale: −i dx on L2 ([0, ∞) Formally self-adjoint is
not the same as self-adjoint! 27
1.1. Test functions. These are defined on a domain of interest for a partic-
ular problem: on R, or R2 , Rn or a subset (a, b) ⊂ R, Ω ⊂ Rn ) and are very,
very, nice functions, having all the properties one may wish for in tackling
a problem at hand. Namely:
- they are infinitely many times differentiable (we say for short that they
are class C ∞ ), and
- they vanish, together with all their derivatives, at the boundary of the
domain (so that when we integrate by parts the boundary terms vanish).
Here are the most popular spaces of test functions.
S(R), the rapidly decaying functions (the Schwartz space) contains func-
tions which decay, towards ±∞, more rapidly than any power, and so do all
2
their derivatives. For example, φ(x) = e−x ∈ S(R).
It can be shown that the Fourier transform F : S → S is one-to-one and
onto. This requires a proof, but here is the intuitive reason: f has decay
↔ Ff is smooth, and f is smooth ↔ Ff has decay, and functions in S are
both smooth and have decay.
C0∞ (Ω), the functions with compact support on Ω ⊂ Rn are functions
which are non-zero only on a bounded set, and they are identically zero
before the boundary of Ω. For example,
(
− 12
φ(x) = e 1−x if |x| < 1 ∈ C0∞ (R)
0 otherwise
and in fact φ(x) ∈ C0∞ (−a, a) for any a > 1.
Note: the spaces of test functions are linear subspaces of L2 (and they
are dense in L2 ).
DISTRIBUTIONS 3
1.3.1. Function type distributions. Any function which is not too wild (i.e. it
is integrable on compact sets) can be regarded as a distribution: say, f (x) =
1, or f (x) = sin x, or f (x) = eix , or f (x) = H(x), defines a distribution by
the formula
Z ∞
(1) (φ, f ) = f (x)φ(x) dx
−∞
The integral converges since any test function φ(x) vanishes outside some
interval [−M, M ].1 The functional defined by (1) is clearly linear; a bit of
mathematical argumentation would also show continuity (which we do not
pursue here).
Note the underlying idea: (1) is, essentially, the inner product of L2 . We
want to use it for more general functions f , like the examples above, which
are not in L2 ; we do that, but the price we pay is that we can only do pairing
with very rapidly decaying functions φ.
1Of course, if the test functions are S, then we should require f not to increase faster
than some power.
4 RODICA D. COSTIN
Note that (2),(3) are commonly written, in the spirit of (1), using an
integral sign:
Z ∞
(φ, δ) = φ(x)δ(x) dx = φ(0)
−∞
and
Z ∞ Z ∞
(φ, δ(x − a)) = φ(x)δ(x − a) dx = φ(t + a)δ(t) dt = φ(a)
−∞ −∞
These are not bonafide integrals, but they work like integrals.
This justifies
Definition 1. The derivative u0 of a distribution u is defined by the rule
(φ, u0 ) = −(φ0 , u) for any test function φ
Theorem 2. If u is a function-type distribution, and u is differentiable,
then its derivative in the sense of distributions coincides with its derivative
as a function.
DISTRIBUTIONS 5
2.5. The Fourier transform of distributions. Let us see how the Fourier
transform Ff of a super-nice function (from the point of view of the Fourier
transform), say, f ∈ S acts on test functions:
Z ∞ Z ∞ Z ∞
1
(φ, Ff ) = (Ff )(ξ)φ(ξ)dξ = dξ √ dx e−ixξ f (x)φ(ξ)
−∞ −∞ 2π −∞
Z ∞ Z ∞ Z ∞
1
= dx f (x) √ dξ e−ixξ φ(ξ) = dx f (x) (Fφ)(x) = (Fφ, f )
−∞ 2π −∞ −∞
It is natural then to define:
Definition 3. For any distribution u on S (i.e. tempered distribution)
define its Fourier transform Fu by the rule
(φ, Fu) = (Fφ, u) for any test function φ
Examples.
1. The Fourier transform of δ is 1 (well, up to a normalization constant...
you need to check the normalization you are using). For the one here
1
Fδ = √
2π
(check!) Also
1
Fδ(x − a) = √ e−iaξ
2π
Exercise.
Show that δ(−x) = δ(x). Find the Fourier transforms of 1, and of eibx , sin x, cos x.
DISTRIBUTIONS 7
(by the mean value theorem) which converges to φ(0) = (φ, δ).
3.4. Other sequences. We could smooth out the corners of our step func-
tions to make them even C ∞ .
In fact, one can find many fundamental sequences for δ by rescaling, as
follows. Pick a function f which is C ∞ on R Rwhich is zero outside some
∞
interval [−M, M ] and has total integral one: −∞ f (x)dx = 1. For each
> 0 use rescaling to shrink [−M, M ] to [−M, M ] while dialting f so that
the total area remains one: define
1 x
f (x) = f
R∞
Then f (x) = 0 outside [−M, M ] and −∞ f (x)dx = 1 (check!). We have
lim→0 f = δ since
Z ∞ Z ∞ Z M
(φ, f ) = f (x)φ(x)dx = f (y)φ(y)dy = f (y)φ(y)dy
−∞ −∞ −M
Z M
= {f (y) [φ(y) − φ(0)] + f (y)φ(0)} dy
−M
8 RODICA D. COSTIN
M M
φ(y) − φ(0)
Z Z
= φ(0) f (y) dy + yf (y) dy
−M −M y
and using the mean value theorem for the function φ(y)
Z M
= φ(0) + yf (y)φ0 (c) for some c ∈ (0, M )
−M
hence the last term in the sum goes to 0 as → 0 and the limit is φ(0).
4. Green’s function
4.1. Non-homogeneous conditions can be reduced to homogeneous
conditions, but for a non-homogeneous equation. If the initial and
boundary conditions are not zero, the problem is first turned into one with
homogeneous conditions (but with a non-homogeneous term) by subtract-
ing from the unknown function a function witch has the prescribed initial
and/or boundary conditions. This is easiest to see when illustrated on a few
examples.
∂ ∂2
(9) L= ∂t − ∂x2
for equation ut − uxx = f
and so on.
Then: if u1 solves Lu1 = f1 (x), and u2 solves Lu2 = f2 (x), then a
linear combination u = c1 u1 + c2 u2 solves Lu = c1 f1 + c2 f2 (because L is
linear). Similarly, we can superimpose P any number of solutions:
P if uk solves
Luk = fk (x) for k = 1, . . . , n then u = k ck uk solves Lu = k ck fk .
The same is true if we also have some homogeneous conditions, initial
(IC) or boundary (BC). A condition is called homogeneous when a linear
functional of u is required to be zero. For example, we could consider the
problem Lu = f with
-L as in (7) and the (IC) u(x0 ) = 0
-L as in (8) and the (IC) u(x0 ) = 0, u0 (x0 ) = 0
-L as in (8) and the (BC) u(x0 ) = 0, u(x1 ) − 3u0 (x1 ) = 0
-L as in (9) and the (IC) u(t0 , x) = 0 and (BC) u(x0 ) = 0, u(x1 ) = 0
and so on.
For this non-homogeneous problems with homogeneous conditions, again,
a superposition of forcing terms is solved by a superposition of solutions.
The tasks ahead are then: to find ways of calculating the Green’s function
of given problems, to show that the superposition (10) is indeed a solution,
10 RODICA D. COSTIN
and to see what information about solutions can be read directly in the
Green’s function.
4.3. First examples.
4.3.1. The simplest example, C ∞ case. Consider L as in (7) and the homo-
geneous initial value problem
du
(11) dx + a(x)u = f (x), u(x0 ) = 0, for x ≥ x0
Let us find its Green’s function: solve
dG
(12) + a(x)G = δ(x − t), G(x0 , t) = 0 for all t
dx
R
The integrating factor is exp( a(x) dx).
Assume a ∈ C ∞ . Then so is the integrating factor, and then we can
multiply the equation by it (since δ can be multiplied by C ∞ functions) and
we get
d Rxx a(s) ds Rx
a(s) ds
e 0 G = e x0 δ(x − t)
dx
hence
d Rxx a(s) ds Rt
a(s) ds
(13) e 0 G = e x0 δ(x − t)
dx
Recalling that
dH d
= δ(x), and similarly H(x − t) = δ(x − t)
dx dx
integrating (13)2
Rx Rt
a(s) ds a(s) ds
e x0 G(x, t) = e x0 H(x − t) + C(t)
therefore
Rx Rx
− a(s) ds
G(x, t) = e− t a(s) ds
H(x − t) + e x0 C(t)
| {z } | {z }
particular sol. general sol. homog. eq.
R x0
Imposing the initial condition we obtain C(t) = −e− t a(s) ds H(x 0 − t)
hence
Rx
(14) G(x, t) = e− t a(s) ds
χ[x0 ,x] (t)
where
1 if t ∈ [x0 , x]
χ[x0 ,x] (t) = H(x − t) − H(x0 − t) =
0 otherwise
is a step function.
The solution of (11) is
Z ∞ Z x Rx
u(x) = f (t)G(x, t) dt = f (t)e− t a
dt
x0 x0
2We need to use here the fact that the only distributions whose derivatives are 0 are
the constants, which requires a proof (not included here).
DISTRIBUTIONS 11
Fact 1. We can rely on the intuition that δ(x) = 0 for x < 0 and also
δ(x) = 0 for x > 0. (This can be rigorously stated, but we will not do that.)
Therefore δ(x − t) = 0 for x < t and also for x > t.
d
4.4. Autonomous equations. If the coefficients of L = L( dx ) do not de-
pend on x, the equation Lu = f (x) is called autonomous. For example, (6),
(9) are autonomous equations.
Autonomous equations are translation-invariant, and we want to take
advantage of that. More precisely, if u(x) solves Lu = 0 then u(x − t) solves
the same (homogeneous) equation for any t. And is u(x) solves Lu = f (x)
then also u(x − t) solves L[u(x − t)] = f (x − t) for any t. So instead of
searching for a Green’s function G(x, t) by solving LG = δ(x − t) with
boundary conditions, we could first solve for LF = δ, then add to F (x − t)
the general solution of the homogeneous equation, and then impose the
boundary conditions, the result being the Green’s function. The procedure
is illustrated below. But first, a definition.
d
Definition 5. Let L = L( dx ) be an autonomous differential operator.
A solution F (x) of LF = δ is called a fundamental solution of the differ-
ential operator L.
R
Then upart (x) = (f ∗ F )(x) = f (t)F (x − t)dt is a particular solution of
Lu = f . The general solution of Lu = f is upart + uhomog where uhomog is
the general solution of Lu = 0.
4.4.1. A simple example. Consider the autonomous problem (6) (which, of
course, is a particular case of (7), for a(x) ≡ α). Its fundamental solution
is found as in §4.3.1 (only calculating for t = 0): we find a solution of
dF
(15) dx + αF = δ(x)
(no conditions imposed). Multiplying be the integrating factor eαx and
integrating we obtain
F (x) = e−αx [H(x) + C]
Any particular value for C gives a fundamental solution of L. For exam-
ple, for C = − 12 we obtain the symmetric fundamental solution F (x) =
1 −αx .
2 sign(x) e
To find the Green’s function of L with condition u(x0 ) = 0, for x ≥ x0
we look for it in the form F (x − t) + C(t)e−αx (note that the free con-
stant in the solution of the homogeneous equation is allowed to depend on
t). We determine C(t) by imposing the boundary condition: G(x0 , t) =
e−α(x0 −t) H(x − t) − C(t)e−α(x0 −t) = 0 and solving for C(t) we get
G(x, t) = e−α(x−t) (H(x − t) − H(x0 − t))
R∞
The solution of the initial value problem is then given by u(x) = x0 G(x, t)f (t) dt
Rx
(which for x > x0 equals the familiar u(x) = e−αx x0 eαt f (t) dt).
Alternatively, the solution can be found as follows. A particular solution
of du
dx + αu = f is then (for, say, C = 0)
Z ∞
upart (x) = (f ∗ F )(x) = f (t)e−α(x−t) H(x − t) dt
−∞
DISTRIBUTIONS 13
we saw that if
Z x
Q 1 R
p(x) = exp , w= p, q = − p
x0 P P P
then substituting
p p0 q
P = , Q= , R=−
w w w
14 RODICA D. COSTIN
d2
1 d
L= −p 2 − p0 +q u
w dx dx
becomes
d2
1 d
L= −pz 02 2 − (pz 00 + p0 z 0 ) + q u
w dz dz
pz 02 d2 pz 00 + p0 z 0 d
q
= − 2− +
w dz pz 02 dz pz 02
R x q w(t)
We want pz 02 /w = 1, therefore z(x) = a p(t) dt where a can be chosen
at will. With this choice L is
00
d2 p0
z d q
L=− 2 − + +
dz z 02 pz 0 dz w
II. The first derivative term can be eliminated (for any linear second order
equation) by a substitution of the dependent variable of the form u = φy
where the function φ is suitably chosen. In our case, this substitution in the
eigenvalue equation Lu = λu gives
00
p0
00
d2 y d2 φ p0
dφ dy z dy z dφ q
−φ 2 −2 −y 2 − 02 + 0 φ − 02 + 0 y + φy = λφy
dz dz dz dz z pz dz z pz dz w
hence, and dividing by φ
(17)
z 00 p0 dy
00
d2 y 1 d2 φ p0
1 dφ z 1 dφ q
− 2 + + + − − + + y = λy
dz 2 φ dz z 02 pz 0 dz φ dz 2 z 02 pz 0 φ dz w
dy
The condition that the coefficient of dz vanishes is that
1 dφ z 00 p0
(18) 2 + 02 + 0 = 0
φ dz z pz
or
φ0 z 00 p0
2 + 0 + = 0 therefore φ2 pz 0 = C and choose φ = (pw)−1/4
φ z p
Furthermore, this change of coordinates is unitary:
DISTRIBUTIONS 15
Proof of Theorem 6. The formula for z(x) and φ have been found before.
To deduce the formula for g(z) we only need to simplify the coefficient of y
in (17); using (18) this coefficient becomes
1 d2 φ 1 dφ 2
q d 1 dφ q
− 2
+2 + =φ 2
+
φ dz φ dz w dz φ dz w
giving the first equality in (21); the second equality follows immediately
using the formulas for z 0 and for φ.
To show that the Liouville transformation preserves the norm we simply
change the variable of integration:
Z b Z c Z c √
2 2 2 1 1 2 p
|u(x)| w(x)dx = φ |y| w 0 dz = 1/2
|y| w √ dz
a 0 z 0 (pw) w
2
5. Furthermore,
∞
X
(26) w(t) un (t) un (x) = δ(x − t)
n=1
Now,
Z b
(−pu0n )0 + qun G(x, t) dx = (un , wLG(·, t))
hLun , G(·, t)iw =
a
(we define, in the sense of distributions, as usual, (u, Lφ) = (Lu, φ)) hence
(un , wLG(·, t)) = (un , δ(x − t) w(t)) = un (t)w(t)
Question: Which steps above are not quite correctly justified?
Therefore
un (t)w(t)
cn (t) =
λn
which used in (29) gives (23).
4. Formula (24) for the solution of a second order linear equation follows
from the classical theory of linear differential equations (or can be checked
directly). Using (23) in (24) gives (25).
5. Besides the expansion (25), we can expand any u ∈ H as
∞
X X∞ Z b
u(x) = hun , uiw un (x) = un (t)u(t)w(t)dt un (x)
n=1 n=1 a
∞
Z b "X #
= [exchange order in distribution sense] un (t)un (x) u(t)w(t) dt
a n=1
therefore
∞
X
w(t) un (t)un (x) = δ(x − t)
n=1
2
22 RODICA D. COSTIN
Proof.
We only need to retrace the proof of Theorem 8 and see what happens if
0 is an eigenvalue.
We first see that we cannot find a Green’s function. Then a direct verifi-
cation show that (31) satisfies Lu = f . Are there other solutions? Suppose
that u2 also satisfies Lu2 = f . Then u − u2 satisfies L(u − u2 ) = 0, hence
u − u2 is a solution of the homogeneous equation, hence u2 = u+ [a solution
in D(L) of the homogeneous equation].
If u2 is linearly dependent of u0 , then u2 has the same formula as u, only
with a different constant C.
If u2 and u0 are linearly independent, then there are two independent
solutions of the homogeneous equation satisfying the boundary conditions,
hence any solutions satisfies the boundary conditions, but this is not possible
unless α = α0 = β = β 0 = 0, contradiction! 2
7.3. The resolvent. We can use a line of reasoning similar to that of §7.1
and §7.2 to find the formula for the resolvent of L. Let z ∈ C, not an
eigenvalue of L. This means that 0 is not an eigenvalue of L − zI. Using
Theorems 8 and 9 for the operator L − z instead of L we find
Remark: formula (32) shows that the Green’s function of the resolvent
(L−z)−1 of (L, D(L)) is analytic in the parameter z, except for z equal to the
eigenvalues of (L, D(L)), where G has poles of order one. This is intuitively
expected, by functional calculus. In infinite dimensions this happens only
for special operators, though.
24 RODICA D. COSTIN
8. Continuous spectrum
Eigenvalue problems on infinite intervals, or when coefficients are not
smooth enough, may have a continuum of ”eigenvalues”.
For example, consider the problem
(34) u00 + λu = 0, for x ∈ (−∞, ∞)
with the boundary condition
(35) u is bounded at ± ∞
Then√ any λ ≥ √0 is an ”eigenvalue”, corresponding to the ”eigenfunctions”
sin( λx), cos( λx). I used quotation makes because we can only use these
words in a specified a Hilbert space, and we specified none. However, the
problem (34), (35) has a valid physical significance, and we need a solid
mathematical foundation for its treatment.
To have an intuitive picture on what is going on with the problem (34),
(35), we should consider the even simpler operator
d
(36) L = −i dx
with the boundary conditions (35), that is, on
(37) D(L) = {u ∈ L2 (R) | u0 ∈ L2 (R), u is bounded at ± ∞}
We have
d iξx
−i dx e = ξ eiξx
so eiξx behaves like an eigenfunction corresponding to the eigenvalue ξ, for
any ξ ∈ R. Since eiξx does not belong to the Hilbert space L2 (R), we will call
it a generalized eigenfunction, and ξ will be called a generalized eigenvalue.
Moreover, these generalized eigenfunctions are ”complete” in L2 (R) in
the sense that any f in the Hilbert space can be expanded in terms of the
generalized eigenfunctions—only not as a series, but as an integral: any
f ∈ L2 (R) can be written as
Z ∞
f (x) = eixξ g(ξ) dξ
−∞
where, of course, g(ξ) = √12π fˆ(ξ) ∈ L2 (R) is the continuous analogue of the
coefficients of the expansion of f .
The Fourier transform is the unitary transformation of L2 (R) which di-
d
agonalizes the operator dx . Indeed, since
Z ∞
d
−i dx d
f (x) = −i dx F −1 fˆ(ξ) = −i dx
d
eixξ √12π fˆ(ξ) dξ
−∞
Z ∞
= eixξ √12π ξ fˆ(ξ) dξ = F −1 [ξ fˆ(ξ)]
−∞
we see that −1
d
F −i dx F =ξ
DISTRIBUTIONS 25
fˆ(ξ)
Z ∞ Z ∞ ixξ Z ∞
e
(40) u(x) = √12π eixξ 1
dξ = 2π f (y)e−iξy dydξ
−∞ ξ − z −∞ ξ − z −∞
Z ∞ Z ∞ i(x−y)ξ
1 e
= f (y) dξ
2π −∞ −∞ ξ − z
26 RODICA D. COSTIN
hence
∞
eiξ(x−y)
Z
1
(41) G(x, y, z) = dξ
2π −∞ ξ−z
We can further simplify (41): for =z > 0 we have G(x, t) = ieizt H(t), which
is seen by deformation of the path of integration upwards in the complex ξ
plane, towards ∞ and collecting the residue at ξ = z if t > 0, and downwards
if t < 0, yielding zero. Similarly, for =z < 0, G(x, t) = ieizt H(−t).
We see that u(x) in (38) is not defined if z ∈ R. This alone does not imply
that the equation (L − z)u = f has no solution, for now it only implies that
a particular formula for a possible solution does not apply.
Let’s however try to solve
2
8.0.2. A second order example: L = − xd2 on R. Let
d 2
(43) L = − dx 2
Since (43) is the square of (36) we expect these two operators to have a
d2 ±ixξ
common set of generalized eigenvectors, and indeed, − dx 2e = ξ 2 e±ixξ so
that the spectrum of (43) , (44) is [0, +∞).
8.1.2. An eigenvalue problem for the Bessel operator. Consider the eigen-
value problem
1 2
(xu0 )0 + λ − nx2 u = 0, for x ∈ [0, ∞)
x
with boundary conditions
u is finite for x → 0 and for x → ∞
√
For any λ > 0 the Bessel functions Jn ( λx) are generalized eigenfunctions
and it can be shown that any f ∈ L2 ([0, ∞), xdx) can be expanded in terms
of the eigenfunctions: there exists a function F so that
Z ∞ √ Z ∞
f (x) = Jn ( λx)F (λ) dλ = Jn (tx)F (t2 ) 2t dt
0 0
A general representation theorem of this nature exists for self-adjoint
operators.
d
8.2. A cautionary tale: −i dx on L2 ([0, ∞) Formally self-adjoint is not
d
the same as self-adjoint! Consider the operator L = −i dx on D(L) =
2 0 2
{f ∈ L ([0, ∞), f ∈ L ([0, ∞), f (0) = 0}. Check first that the operator is
symmetric.
Now let’s look at (L + i) and see whether it is invertible.
For that, we have to solve the equation
−iu0 + iu = f ; f ∈ L2 ([0, ∞)
Equivalently,
u0 − u = if ; f ∈ L2 ([0, ∞)
By variation of parameters, and imposing the condition u(0) = 0 we get
Z x
u(x) = ex e−s f (s)ds
0
28 RODICA D. COSTIN
R∞
Take any f > 0 in L2 . Check that e−s f ∈ L2 and 0 e−s f (s)ds = Cf > 0.
It then follows that
u(x)e−x → Cf as x → ∞
and thus u 6∈ L2 .
In fact you can similarly check that for any z = x + iy with y < 0 (L − z)
is not invertible. Thus the spectrum of L contains the whole lower half plane
in z, far from being purely real! That’s not what happens with self-adjoint
matrices. In fact : −i dx d
on L2 ([0, ∞) on D(L) = {f ∈ L2 ([0, ∞), f 0 ∈
L2 ([0, ∞), f (0) = 0} is not self-adjoint!