Mat 306 Study Guide
Mat 306 Study Guide
Prof M Grobbelaar
Revised by
Dr SA van Aardt
Contents
PREFACE iv
Preface
The module MAT306—X is a continuation of the module MAT216—W in which one—dimensional differential
equations are treated. Obviously you should have a good knowledge of the contents of this module before
starting to study MAT306—X.
Although the study guide is to a great extent complete, it is essential that you use the prescribed book
together with the study guide (this does not apply in Chapter 1), since exercises have been taken from the
prescribed book, while in may cases we have referred to the prescribed book for the proofs of theorems.
As you will notice, Chapter 8 has been adopted from the book Differential Equations with Boundary—value
Problems by Dennis G. Zill and Michael R. Cullen, PWS—Kent Publishing Company, Boston, 1993, which
is the prescribed book for MAT216—W. This procedure is followed in view of the fact that in this book the
subject which is treated in this chapter, viz qualitative theory of differential equations, is presented in a
avery lucid and “user—friendly” manner. Students who are not in possession of this book for any reason
whatsoever, will not be at any disadvantage, since the relevant chapter is sufficiently self—contained with
very few references to other sections of the book. Should references to other sections of the book occur,
explanatory notes will be provided in a tutorial letter whenever deemed necessary.
You will notice that the study guide contains a large number of exercises. Do as many as you find necessary
to master the tutorial matter.
In compiling Chapters 1 to 7, we have, apart from the prescribed book, made use of the following books:
FINIZIO, N & LADAS, G, (1989), Ordinary Differential Equations with Modern Applications, Wadsworth
Publishing Company, Belmont, California.
RICE BERNARD J& STRANGE JERRY D, (1989), Differential Equations with Applications, The Ben-
jamin/Cummings Publishing Company, Inc, Redwood City, California.
RAINVILLE, ED, (1981), Elementary Differential Equations, Collier Macmillan Publishers, London.
TENENBAUM, M & POLLARD, H, (1964), Ordinary Differential Equations, Harper & Row, New York.
Chapter 1
• degeneracy of systems;
• apply the solution techniques for systems of DE’s in solving relevant physical problems.
1
2
1.1 INTRODUCTION
Mathematical models of real—life situations often lead to linear systems of differential equations. Linear
systems of differential equations arise for instance in the theory of electric circuits, in Economics and in
Ecology, the branch of science in which the interaction between different species is investigated. Systems of
differential equations have been used by ao L.F. Richardson1 in as early as 1939 in devising a mathematical
model of arms races, and by E. Ackerman1 and colleagues in the study of a mathematical model for the
detection of diabetes. As we shall see later on (see Chapter 5) systems of differential equations also appear
when an n—th order linear differential equation is reduced to a linear system of differential equations.
Definition 1.1 By a system of n first—order linear differential equations we mean a set of n simulta—
neous equations of the form
where the functions aij , j = 1, 2, . . . , n and the functions Qk , k = 1, 2, . . . , n, are all given functions of t
on some interval J. We assume that all the functions aij and all the functions Qk are continuous on J.
The functions aij (t) and Qk (t) are known functions, while x1 (t) , x2 (t) , . . . , xn (t) are the unknowns.
The functions aij (t) are called the coefficients of the linear system (1.1). We say that the system (1.1)
has constant coefficients when each aij is a constant function. The terms Qk (t) are called forcing terms.
The system (1.1) is called homogeneous when Qk (t) is zero for all k = 1, 2, . . . , n and non—homogeneous
(inhomogeneous) if at least one Qk (t) is not zero.
Remark:
The homogeneous system obtained by putting Qk (t) = 0 for all k = 1, 2, . . . , n,
It is clear that a solution of this system (see Definition 1.3) should be an n—dimensional vector. Use your
knowledge of matrix multiplication to show that the product of the matrix A (t) and the column vector
X actually produces the vector
a11 (t) x1 + a12 (t) x2 + . . . + a1n (t) xn
a21 (t) x1 + a22 (t) x2 + . . . + a2n (t) xn
.. .
.
an1 (t) x1 + an2 (t) x2 + . . . + ann (t) xn
Definition 1.2 A higher—order system with constant coefficients is a system of the form
where all the Pij , i, j = 1, 2, . . . , n are polynomial differential operators and all the functions hi (t) ,
i = 1, 2, . . . , n are defined on an interval J. The determinant
P11 (D) · · · P1n (D)
P21 (D) · · · P2n (D)
det
..
.
Pn1 (D) · · · Pnn (D)
is called the determinant of the system (1.4). The system (1.4) is non—degenerate if its determinant is
non—zero, otherwise it is degenerate.
The determinant of a system of differential equations is of great importance, since it provides a method of
determining the correct number of arbitrary constants in the solution of the system (see Theorem 1.5).
Definition 1.3 A solution of the system (1.1) of differential equations is an n—dimensional vector
function
[x1 (t) , x2 (t) , . . . , xn (t)]T ,
of which each component is defined and differentiable on the interval J and which, when substituted
into the equations in (1.1), satisfies the equations for all t in J.
Theorem 1.1 (The Superposition Principle) Any linear combination of solutions of a system of
differential equations is also a solution of the system.
4
T T
For example, each of the two vectors −2e2t , e2t and e3t , −e3t is a solution of the system
ẋ1 = x1 − 2x2
ẋ2 = x1 + 4x2 .
T
By the superposition principle, the linear combination −2c1 e2t + c2 e3t , c1 e2t − c2 e3t , with c1 , c2 arbitrary
constants, is also a solution of the system.
Before defining the concept of a general solution of a system of differential equations, we need the con-
cept of linear independence of vector functions, which is obviously an extension of the concept of linearly
independent functions.
are solutions, defined on an interval J, of a system of n first order linear differential equations in n
unknowns. We say that X1 (t) , X2 (t) , . . . , Xn (t) are linearly independent if it follows from
c1 X1 + c2 X2 + . . . + cn Xn = 0 for all t in J
that
c1 = c2 = . . . = cn = 0.
The following theorem provides a criterion which can be used to check the linear dependence of solutions
of a system of differential equations:
of the system (1.1) are linearly independent on the interval J if and only if the Wronskian determinant
x11 (t) . . . x1n (t)
x21 (t) . . . x2n (t)
det
..
.
xn1 (t) . . . xnn (t)
We state without proof the following basic existence theorems for linear systems of differential equations:
are linearly independent solutions of the linear system (1.2), a general solution is given by
x1 (t) x11 (t) x12 (t) x1n (t)
x2 (t) x21 (t) x22 (t) x2n (t)
. = c1 .. + c2 .. + . . . + cn .. ,
.
. . . .
xn (t) xn1 (t) xn2 (t) xnn (t)
i.e.
x1 (t) = c1 x11 (t) + c2 x12 (t) + . . . + cn x1n (t)
x2 (t) = c1 x21 (t) + c2 x22 (t) + . . . + cn x2n (t)
..
.
xn (t) = c1 xn1 (t) + c2 xn2 (t) + . . . + cn xnn (t) ,
where c1 , c2 , . . . , cn are arbitrary constants.
of the system (1.2) are linearly independent and if [x1p (t) , x2p (t) , . . . , xnp (t)]T is a particular solution of
the inhomogeneous system (1.1), then a general solution of (1.1) is given by
x1 (t) x11 (t) x12 (t) x1n (t) x1p (t)
x2 (t) x21 (t) x22 (t) x2n (t) x2p (t)
.. = c1 .. + c2 .. + . . . + cn .. + ..
. . . . .
xn (t) xn1 (t) xn2 (t) xnn (t) xnp (t)
i.e.
x1 (t) = c1 x11 (t) + c2 x12 (t) + . . . + cn x1n (t) + x1p (t)
x2 (t) = c1 x21 (t) + c2 x22 (t) + . . . + cn x2n (t) + x2p (t)
..
.
xn (t) = c1 xn1 (t) + c2 xn2 (t) + . . . + cn xnn (t) + xnp (t)
where c1 , c2 , . . . , cn are arbitrary constants.
6
Remark 1.1
In Theorem 1.3 we learn that a general solution of a system of n first—order linear differential equations
in n unknown functions contains n arbitrary constants. As we shall see in the next section the method
of elimination sometimes introduces redundant constants. The redundant constants can be eliminated by
substituting the solutions into the system and equating coefficients of similar terms.
In general the following theorem determines the correct number of arbitrary constants in a general solution
of a system of differential equations (which may be a higher—order system):
Theorem 1.5 The correct number of arbitrary constants in a general solution of a system of differential
equations is equal to the order of the determinant of the system, provided this determinant has non—zero
value.
In particular: The number of arbitrary constants in a general solution of the higher—order system
We will return to the proof of the latter theorem in Section 1.4, where we will also discuss the degenerate
case when = 0 in order to be able to decide when a degenerate system has an infinite number of solutions
or no solution. We end this section with the Existence and Uniqueness Theorem for a linear system with
initial conditions.
Theorem 1.6 (Existence and Uniqueness Theorem for Linear Initial Value Problems)
Assume that the coefficients aij , i, j = 1, 2, . . . , n and the functions Qk (t) , k = 1, 2, . . . , n of the system
(1.1) are all continuous on the interval J. Let t0 be a point in J and let x10 , x20 , . . . , xn0 , be n given
constants. Then the interval problem (IVP) consisting of the system (1.1) and the initial conditions
has a unique solution [x1 (t) , x2 (t) , . . . , xn (t)]T . Furthermore this unique solution is valid throughout the
interval J.
Before carrying on with this section, you should revise the section of MAT216—W in which the properties
of differential operators are discussed.
1.3.1 POLYNOMIAL OPERATORS
First we define:
Definition 1.5 An nth—order polynomial operator P (D) is a linear combination of differential opera—
tors of the form
an Dn + an−1 Dn−1 + . . . + a1 D + a0
The symbol P (D) indicates that P (D) is applied to an n—times differentiable function y. To emphasize
the fact that P (D) is an operator, it is actually preferable (but not compulsory) to write P (D) [y] . This
notation is used in ao the book Fundamentals of Differential Equations, by R. Kent Nagle and B. Saff (see
bibliography at the end of this chapter for complete reference).
Definition 1.6 (Linearity property) The polynomial differential operator P (D) is a linear opera—
tor in view of the fact that the following linearity properties are satisfied:
Definition 1.7 (Sum and Product of polynomial differential operators) The sum
P1 (D) + P2 (D) of two operators, P1 (D) and P2 (D) , is obtained by writing P1 and P2 as linear combi—
nations of the D operator and adding coefficients of equal powers of D.
The product P1 (D) P2 (D) of two operators P1 (D) and P2 (D) is obtained by applying the operator
P2 (D) and then applying the operator P1 (D) .
Remark:
Polynomial operators can be shown to satisfy all the laws of elementary Algebra with regards to the basic
operations like addition and multiplication. They may therefore be handled in the same way as algebraic
polynomials. We can, for instance, factorize polynomial differential operators by using methods similar to
those used in factorizing algebraic polynomials.
Example 1.1
(c) The operator D3 − 2D2 − 15D may be factorized into the factors D (D − 5) (D + 3) .
(d) Any linear system of differential equations may be written in the form (1.5). For example, the system
d2 x dy
− 4x + =0
dt2 dt
(1.7)
dx d2 y
−4 + 2 + 2y = 0
dt dt
Example 1.2
Solve the system (1.7) by using the elimination method (operator method).
Solution:
Apply the operator 4D to the first equation and the operator D2 − 4 to the second equation of the system.
For brevity’s sake, we denote these operations by 4 (D) [(1)] and D2 − 4 [(2)] . We thus have
we find
m2 = −4 or m2 = 2
9 MAT306—X/1
so that
√
m = ±2i or m = ± 2.
whence
√ √ 1 1
2t
D [x] = c1 e + c2 e− 2t
− c3 cos 2t − c4 sin 2t.
2 2
Integration yields √ √
2 √2t 2 −√2t 1 1
x (t) = c1 e − c2 e − c3 sin 2t + c4 cos 2t + c5 ,
2 2 4 4
with ci , i = 1, 2, . . . , 5 arbitrary constants.
The solution, therefore, contains five arbitrary constants whereas the determinant
D2 − 4 D
det 2
= D2 − 4 D2 + 2 + 4D2
−4D D + 2
is of the fourth order. Substitution of the solution (x (t) , y (t)) into (1) yields c5 = 0. Consequently, a
general solution of the system is given by
√ √
2 √2t 2 −√2t 1
x (t) = c1 e − c2 e − c3 sin 2t + c4 cos 2t
2 2 4
√ √
2t
y (t) = c2 e + c2 e− 2t
+ c3 cos 2t + c4 sin 2t.
Example 1.3
Solve the system
Solution:
Eliminating y first we find:
with c1 and c2 arbitrary constants. We find the particular integral xP.I. (t) by using the method of unde-
termined coefficients:
Let
xP.I. (t) = Aet + Bt
so that
6Aet + 3B = 3et + 5.
(D + 1) [y] = (D + 1) [x] − et
5 5
= c1 + t + . (6)
3 3
From the auxilliary equation m + 1 = 0 it clearly follows that the complementary function yC.F. (t) is given
by
yC.F. (t) = c3 e−t .
Using the method of undetermined coefficients we solve the particular integral yP.I. (t) as follows:
Let
yP.I. (t) = At + B.
Then
ẏP.I. (t) = A.
11 MAT306—X/1
5 5
At + (A + B) = c1 + t + .
3 3
Comparing coefficients we find that
5
A= and B = c1
3
so that
5
yP.I. (t) = t + c1 .
3
Therefore
D+1 −D − 1
det
D−1 2D + 1
is of the second order. Substitution of the solution (x (t) , y (t)) into (2) yields c3 = −2c2 . Consequently a
general solution of the system is given by
1 5
x (t) = c1 + c2 e−t + et + t
2 3
5
y (t) = c1 − 2c2 e−t + t.
3
Exercise 1.1
(a) ẋ = 5x − 6y + 1
ẏ = 6x − 7y + 1
(b) ẋ = 3x − 2y + 2t2
ẏ = 5x + y − 1
(c) ẍ − ẏ = t
ẋ + 3x + ẏ + 3y = 2.
ÿ = −y + ẏ + z
ż = y − z − 1
y (0) = 1, ẏ (0) = 0, z (0) = 1.
12
In Remark 1.1 it was stated that the method of elimination sometimes introduces redundant arbitrary
constants into a general solution of a system of differential equations, the reason being, as we have in
fact seen in the above example, that a polynomial operator in D operates on each equation. Since the
elimination of such constants is generally a tedious and time consuming process, we shall develop a method
which immediately yields the correct number of constants in a general solution of a (non—degenerate) system
of differential equations. The method amounts to a reduction of a system of differential equations to an
equivalent triangular system.
is reducible by a method analogous to that of row operations in Matrix Algebra to the system
A linear system of differential equations is thus in triangular form if each succeeding equation in the system
has at least one unknown function more (or less) than the previous equation.
From the definition and from the properties of determinants, the following conclusion can immediately be
drawn:
If a system of differential equations is reduced to an equivalent triangular system, the determinant and the
order of the determinant of the system remain unchanged.
In Theorem 1.7 we will show that a solution of an equivalent triangular system is also a solution of the
original system of differential equations. The proof contains the proof of Theorem 1.5 for n = 2.
In the next example we show how a linear system of differential equations can be reduced to an equivalent
triangular system.
13 MAT306—X/1
Example 1.4
Reduce the following system of equations to an equivalent triangular system:
Solution:
Consider the polynomials in D, i.e. the polynomials P (D); retain the equation in which the order of
P (D) is the lowest – in this case (2). The notation (1) → (1) + k (D) [(2)] denotes that (1) is replaced
by (1) + k (D) [(2)] with k (D) a polynomial in D. By executing the operations as indicated, we obtain the
following systems, of which the last is the required triangular system.
to eliminate x.)
Instead, we did not eliminate x, but only got rid of the coefficient D3 of x. The next step would be to
reduce the order of coefficient polynomial P (D) of x in (3) (i.e. 5D2 − D + 1) by one which means that we
have to get rid of the coefficient 5D2 :
(3) → (3) − 5D [(2)] : (14D + 1) [x] + −D5 − 4D4 − 3D2 − 5D + 2 [y] = 0 (4)
(D − 3) [x] + D3 + 1 [y] = 0. (2)
The order of the coefficient polynomial P (D) of x in the two equations is now the same. Retain either of
the two equations.
Remark 1.2
An equivalent triangular system is not unique. The system
D2 + D − 2 [y1 ] = 2 + 20e2x
D [y1 ] − 2y2 = −6 + 12e2x
and
D2 + D − 2 [y2 ] = −6 − 4e2x
y1 − (D + 1) [y2 ] = −4 + 8e2x .
Exercise 1.2
Derive the two triangular systems mentioned in Remark 1.2. Use the method of elimination to solve these
two systems as well as the original system of differential equations. What conclusion can you draw with
regard to the solutions?
We now proceed to the proof of Theorem 1.7. We formulate and prove the theorem only for a linear system
consisting of two equations, since the proof is completely analogous for systems consisting of more equations.
Proof:
The proof is in the form of a discussion:
Suppose we solve equation (3) for x. The correct number of arbitrary constants in the solution is equal to
the order of F1 (D) . Substitute this solution into equation (4), which is then reduced to an equation in y of
which the order equals that of G2 (D) . The number of arbitrary constants in the solution of (4) equals the
order of G2 (D) . Consequently the total number of constants in the solution (x, y) of (1.11) is equal to the
sum of the orders of F1 (D) and G2 (D) . This is, however, precisely the order of the determinant of (1.11),
i.e. the order of F1 (D) G2 (D) . We have thus shown that, if we solve the system as described, the solution
contains the correct number of constants, (since the pair satisfies each equation identically) and that this
number equals the order of the determinant of (1.11). But the order of the determinant of the reduced
system (1.11) is the same as the order of the determinant of the original system (1.10). Thus we have not
only proved that a solution of (1.11) will also be a solution of (1.10) and contain the correct number of
arbitrary constants, but have at the same time proved Theorem 1.5.
15 MAT306—X/1
In all the examples that we have studied, the systems of differential equations were non—degenerate. We
now study:
We recall that a system of differential equations is degenerate when its determinant is zero. As in the case
of systems of algebraic equations, a degenerate system of differential equations has either no solution or
otherwise an infinite number of solutions.
By systematizing the elimination procedure and letting it assume Cramer’s rule, we will be able to decide
when a degenerate system of differential equations has no solution or infinitely many. For simplicity’s sake,
we consider the system
P11 (D) [y1 ] + P12 (D) [y2 ] = f1 (t)
(1.12)
P21 (D) [y1 ] + P22 (D) [y2 ] = f2 (t) .
By applying P22 (D) to the first equation of (1.12) and P12 (D) to the second equation of (1.12), and by
subtracting the resulting equations, we obtain
(P11 (D) P22 (D) − P12 (D) P21 (D)) [y1 ] = P22 (D) [f1 (x)] − P12 (D) [f2 (x)] .
where the determinant on the right hand side is interpreted to mean P22 [f1 ] − P12 [f2 ] .
Similarly we have
P11 P12 P11 f1
det [y2 ] = det . (1.14)
P21 P22 P21 f2
By using equations (1.13) and (1.14), we can make the last part of Theorem 1.5 precise for the case n = 2, by
stating:
If = P11 P22 − P12 P21 = 0, then the system (1.12) has infinitely many solutions if the determinants on
the right hand sides of equations (1.13) and (1.14) vanish; otherwise there is no solution.
The theorem can easily be generalized to hold for the case n > 3.
Example 1.5
Find a general solution, if it exists, of the system
D [x] − D [y] = et
3D [x] − 3D [y] = 3et .
16
Solution:
The system is clearly degerate since
D −D
= = D (−3D) − (−D) (3D) = 0.
3D −3D
Furthermore,
D et
= D 3et − 3D et = 3et − 3et = 0
3D 3et
and similarly
et D
= 3D et − D 3et = 0
3et 3D
so that an infinite number of solutions exist (note that the system is equivalent to a single equation in two
unknowns). We could for instance choose
Example 1.6
Find a general solution, if it exists, of the system
D [x] + 2D [y] = et
D [x] + 2D [y] = t.
Solution:
The system is degenerate since
D 2D
= = D (2D) − 2D (D) = 0.
D 2D
However, since
D et
= D (t) − D et = 1 − et = 0,
D t
the right hand side of the system does not become zero when x is eliminated. (In this case, the same is true
when y is eliminated.) Hence the system has no solution. (This could also quite easily be seen by noting
that, since the left hand sides of the two equations are identical, we must have that the right hand sides
should also be equal, i.e. et = t for all values of t!)
Exercise 1.3
In the following exercises determine if the given systems have no solutions or infinitely many solutions. If
a system has infinitely many solutions, indicate how the set of solutions may be obtained.
1.5 APPLICATIONS2
As stated in Section 1.1, linear systems of differential equations are encountered in a large variety of physical
problems. We will treat a problem which occurs in the study of electrical circuits, as well as a mixture
problem, and finally an example in which the love and hate between two lovers are modelled (believe it if
you can!) Many other applications of linear differential equations may be found in books on differential
equations, a few of which will be mentioned at the end of this section.
Section 1.5.1 as well as Section 1.5.2 are adopted from Section 9—8 of Bernard J. Rice and Jerry D. Strange’s
book Ordinary Differential Equations with Applications, Second Edition, Brooks/Cole Publishing Company,
Pacific Grove, California, 1989, since these sections effectively illustrate the use of systems of differential
equations in modelling physical processes. Only numbers of equations and units (we use metric units) have
been changed.
The current in the electrical circuit shown in Figure 1.1 can be found by applying the elements of circuit
analysis (which you are, for the purpose of this module, not expected to be familiar with). Essential to the
analysis is Kirchoff ’s law, which states that the sum of the voltage drops around any closed loop is zero. In
applying Kirchoff’s law, we use the fact that the voltage across an inductance is
dI
vL = L
dt
and the voltage across a resistance is vR = IR.
Figure 1.1
Let the current in the left loop of the circuit be I1 , and the current in the right loop be I2 . From the figure
we conclude that the current in the resistor R1 is I1 − I2 relative to the left loop, and I2 − I1 relative to the
right loop. Applying Kirchoff’s law to the left loop, we get
dI1
where I˙1 = . Similarly, the sum of the voltage drops around the right loop yields
dt
If the components of the circuit are given, the values of I1 and I2 can be found by solving this system of
differential equations.
Example 1.7
Consider the circuit shown in Figure 1.1. Determine I1 and I2 when the switch is closed if
L1 = L2 = 2 henrys,
R1 = 3 ohms,
R2 = 8 ohms, and
v (t) = 6 volts.
Solution:
As previously noted, the circuit is described by the system
Multiplying the first equation by 3 and applying the operator 2D + 3 to the second and then adding the
two equations, we obtain
4D2 + 28D + 24 [I2 ] = 18,
or dividing by 4,
9
D2 + 7D + 6 [I2 ] = .
2
The solution of this second—order non—homogeneous linear differential equation consists of the sum of a
general solution of the corresponding homogeneous equation and a particular solution of the given non—
homogeneous equation. The solution of D2 + 7D + 6 [I2 ] = 0, is c1 e−t + c2 e−6t .
The method of undetermined coefficients can now be used to show that
3
Ip2 =
4
is a particular solution of the non—homogeneous equation. Thus the expression for I2 is
3
I2 = c1 e−t + c2 e−6t + .
4
19 MAT306—X/1
To find an expression for I1 , we substitute I2 into the second equation of the system:
3
2 −c1 e−t − 6c2 e−6t + 11 c1 e−t + c2 e−6t + − 3I1 = 0.
4
Solving this equation for I1 yields
1 11
I1 = 3c1 e−t − c2 e−6t + .
3 4
Finally, using I1 (0) = 0 in the expressions for I1 and I2 , we get
1 11
0 = 3c1 − c2 + , (I1 )
3 4
3
0 = c1 + c2 + . (I2 )
4
9 3
Solving this system, we get c1 = − and c2 = , so the desired currents are
10 20
27 −t 1 11
I1 = − e − e−6t + ,
10 20 4
9 −t 3 3
I2 = − e + e−6t + .
10 20 4
Exercise 1.4
This exercise is Exercise 3 from Section 9—8 of the quoted book by Rice and Strange. Referring to Figure
1.1, we assume that
L1 = L2 = 1, R1 = 6, R = 1, and R2 = 3.
Prior to t = 0, the generator has an output of 6v which at t = 0 is increased to 12v. Determine the initial
current and the current for t > 0.
1.5.2 MIXTURE PROBLEMS
Consider the two tanks shown in Figure 1.2, in which a salt solution of concentration ci kg/litre is pumped
into tank 1 at a rate gi kg/min. A feedback loop interconnects both tanks so that solution is pumped from
tank 1 to tank 2 at a rate of g1 litre/min and from tank 2 to tank 1 at a rate of g2 litre/min. Simultaneously,
the solution in tank 2 is draining out at a rate of g0 litre/min. Find the amount of salt in each tank.
Figure 1.2
20
Let y1 and y2 represent the amount of salt (in kilograms) in tanks 1 and 2 at any time and G1 and
G2 represent the amount of liquid in the tank at any time t. Then the concentration in each tank is given
by
y1
c1 = = concentration of salt in tank 1
G1
y2
c2 = = concentration of salt in tank 2.
G2
The rate at which the salt is changing in tank 1 is
dy1
= ci gi − c1 g1 + c2 g2
dt
y1 y2
= ci gi − g1 + g2 (1.17)
G1 G2
Also, in tank 2, we have
dy2 y1 y2 y2
= g1 − g2 − g0 . (1.18)
dt G1 G2 G2
Equations (1.17) and (1.18) form a system in y1 and y2 . If G10 and G20 are the initial volumes of the two
tanks, then
G1 = G10 + gi t − g1 t + g2 t
G2 = G20 + g0 t − g2 t + g1 t.
Comment:
In many mixture problems the rates are assumed to be balanced; that is
gi = g0 , gi + g2 = g1 ; and g0 + g2 = g1 .
In this problem we assume the volumes are constant, that is G1 = G10 and G2 = G20 .
Example 1.8
Assume both tanks in Figure 1.2 are filled with 100 litres of salt solutions of concentrations c1 and c2 ,
respectively. Pure water is pumped into tank 1 at the rate of 5 litres/min. The solution is thoroughly
mixed and pumped into and out of tank 2 at the rate of 5 litres/min. Assume no solution is pumped from
tank 2 to tank 1. The system of equations describing this process is obtained by using equations (1.17) and
(1.18) with
ci = 0, gi = g1 = g0 = 5, and g2 = 0.
Thus
dy1 y1
= −5
dt 100
dy2 y1 y2
= 5 −5 .
dt 100 100
21 MAT306—X/1
Exercise 1.5
The mixture problems below are Exercises 11—13 from Section 9—8, again from the cited book by Rice and
Strange.
(1) Solve for the amount of salt at any time in two 200—litre tanks if the input is pure water and
gi = g0 = 5, g1 = 7, and g2 = 2.
(2) Solve for the amount of salt at any time in two 200—litre tanks if the input is a salt solution with
a concentration of 0.5 kilogram/litre and gi = g1 = g0 = 12 litres/minute. Assume g2 = 0 and
y1 (0) = y2 (0) = 0.
(3) Solve for the amount of salt at any time in two 400—litre tanks if the input is a salt solution with a
concentration of 0.5 kilogram/litre,
gi = g0 = 5, g1 = 7, and g2 = 2.
Our final application on linear systems of differential equations is a rather humoristic one, just in case none
of the above applications appeal to you.
This application is due to Steven H. Strogaz, “Love Affairs and Differential Equations”, Math Magazine 61
(1988): 35.
Two lovers, Romeo and Juliet, are such that the more Juliet loves Romeo, the more he begins to dislike her.
On the other hand, when Juliet’s love for Romeo begins to taper off, the more his affection for her begins
to grow. For Juliet’s part, her love for Romeo grows when he loves her and dissipates when he dislikes her.
Introduce r (t) to represent Romeo’s love/hate for Juliet at time t and j (t) to represent Juliet’s love/hate
for Romeo at time t. For either function a positive value represents love and a negative value represents
hate. If a and b are positive constants, then this love affair is modelled by the following system:
dr
= −aj,
dt
dj
= br.
dt
Solve this system. Show that their love affair is a never—ending cycle (ellipse) of love and hate. What
percentage of the time do they achieve simultaneous love?
22
We end this section and chapter by giving the titles of a few books which you can use complementarily to
the study guide and in which more interesting applications on linear systems of differential equations may
be found (there are, of course, numerous books which are worthwhile to page through). In alphabetical
order the books are:
1. N. FINIZIO & G. LADAS. Ordinary Differential Equations with Modern Applications, Wadsworth
Publishing Company, Belmont, California, 1989.
2. R. KENT NAGLE & EDWARD B. SAFF. Fundamentals of Differential Equations, The Benjamin/Cum—
mings Publishing Company, Inc., Redwood City, California, 1986.
3. BERNARD J. RICE & JERRY D. STRANGE. Differential Equations with Applications, Brooks—Cole
Publishing Company, Pacific Grove, California, 1989.
4. DENNIS G. ZILL. A First Course in Differential Equations with Applications, Prindle Webber &
Schmidt, Boston, 1986.
Chapter 2
1
Before studying this chapter, the student should read Chapter 1 and the first two sections of Chapter 2 of the prescribed
book.
23
24
• find real solutions for a sytem of DE’s which has complex eigenvalues and eigenvectors;
• find two linear independent eigenvectors which correspond to a double (repeated) root of C (λ) = 0;
2.1 INTRODUCTION
We have already pointed out that a system of n linear equations in functions x1 (t) , x2 (t) , . . . , xn (t) and
with constant coefficients aij , may be expressed in the form Ẋ = AX, with A the n × n coefficient matrix
and X the vector [x1 , . . . , xn ]T . Thus the system
ẋ − y = x
ẏ + y = 0
ż − 2z = 0
In this chapter it will be shown that the solutions of Ẋ = AX may be found by studying the eigenvalues
and eigenvectors of A. This approach is known as the eigenvalue—eigenvector method. If this method yields
n linearly independent solutions of a system of differential equations, then a general solution is a linear
combination of these solutions – the same as that obtained by applying the method of elimination which
was described in the previous chapter.
In what follows, A is an n × n matrix with real entries. For the sake of completeness we give the following
definitions and theorem with which you should be familiar:
Definition 2.2
a11 − λ a12 ... a1n
a21 a22 − λ ... a2n
C (λ) = |A − λI| = det
.. .. .. ..
. . . .
an1 an2 . . . ann − λ
Exercise 2.1
Study examples 2.3.1 to 2.3.5, pp. 58—60 of the prescribed book.
As far as solutions of Ẋ = AX are concerned, the analogy between ẋ = ax (a a constant) and Ẋ = AX,
gives rise to the conjecture that exponentials should satisfy the equation Ẋ = AX. Let us assume that
X (t) = eλt U is a solution. Since Ẋ =λeλt U, U and λ must satisfy (from Ẋ = AX)
i.e.
eλt (A − λI) U = 0,
i.e.
(A − λI) U = 0,
since eλt = 0. We note that this equation expresses precisely the connection between an eigenvalue λ of
A and the associated eigenvector U. If λ is real, then U = 0 is the solution of a homogeneous system of
algebraic equations with real coefficients and may, therefore, be chosen as a real vector.
The following theorem can now be formulated:
Theorem 2.2 If U1 is a real eigenvector of A corresponding to the real eigenvalue λ1 , then X1 (t) =
eλ1 t U1 is a solution of Ẋ = AX for all t.
Exercise 2.2
Answer:
U1 = k1 [0, 1, −1]T , U2 = e−2t k2 [−2, 1, 0]T , U3 = e−3t k3 [1, 0, −1]T .
(In general we will not provide answers, as solutions can be checked by substituting into the original
equation. Do this as an exercise.)
As in Chapter 1 we have
Theorem 2.3 (The Superposition Principle) Assume that X1 (t) , . . . , Xr (t) are solutions of
Ẋ = AX. Then
X (t) = k1 X1 (t) + . . . + kr Xr (t)
with ki , i = 1, 2, . . . , r arbitrary constants, is also a solution.
Exercise 2.3
Prove Theorem 2.3.
Remark 2.1
(1) In the proof use is made of the fact that for any n × n matrix A
i.e.
r r
A ki Xi (t) = ki AXi (t) .
i=1 i=1
(2) It follows from Theorem 2.3 that the set of solutions of Ẋ = AX forms a vector space (choose the
trivial solution X ≡ 0 as the zero element). We may, therefore, now use the expression “the space of
solutions” or “the solution space”.
It will be shown later on that the dimension of the solution space corresponds with that of A, i.e. if
A is an n × n matrix, the space of solutions Ẋ = AX will be n—dimensional.
to be linearly independent.
27 MAT306—X/1
Theorem 2.4 Suppose the n × n matrix A has different eigenvalues λ1 , . . . , λn corresponding to eigen-
vectors U1 , . . . Un . Then U1 , . . . , Un are linearly independent.
Exercise 2.4
(2) Find an n × n matrix A such that although A does not have n different eigenvalues, n linearly
independent corresponding eigenvectors still exist.
A natural question to ask is: which n × n matrices have n linearly independent eigenvectors? An important
class of matrices with this property, is the class of symmetric matrices. These matrices appear frequently
in applications. We will return to this class of matrices in Chapter 3.
Exercise 2.5
(1) Study Section 2.5, pp. 71—73 of the prescribed book, which deals with symmetric matrices. Prove
that a symmetric matrix has only real eigenvalues.
of a system of linear equations are linearly independent, is to find the Wronskian of the solutions. We recall
that the Wronskian W (t) of the solutions X1 (t) , . . . , Xn (t) is the determinant
x11 (t) . . . x1n (t)
.. .. ..
W (t) = |X1 (t) , X2 (t) , . . . , Xn (t)| = det
. . . .
xn1 (t) . . . xnn(t)
By recalling that the value of a determinant in which one of the rows or columns is a linear combination
of another row or column, is zero, it follows immediately that W = 0 if and only if X1 , . . . , Xn are linearly
dependent. Therefore, if W = 0, the solutions X1 , . . . , Xn are linearly independent.
Since every Xi , i = 1, . . . , n is a function of t, we must investigate the questions whether it is possible that
on an interval I ⊂ R, W could be zero in some points and in others not. The following theorem, to which
we shall return in Chapter 4, settles the question.
Theorem 2.5 If W, the Wronskian of the solutions of a system of differential equations, is zero at a point
t0 in an interval I, it is zero for all t ∈ I.
28
W (t) = 0 ∀t ∈ I
or
W (t) = 0 ∀t ∈ I.
Example 2.1
Solve the initial value problem
6 8 1
Ẋ = X ≡ AX, with X (0) = .
−3 −4 2
Solution:
We solve the characteristic equation
6−λ 8
C (λ) = =0
−3 −4 − λ
C (λ) = (6 − λ) (−4 − λ) + 24 = λ (λ − 2) = 0
so that λ1 = 0 and λ2 = 2.
To find the eigenvector U1 corresponding to λ1 = 0 we solve (using the equation (A−λI) U = 0 with
λ = λ1 = 0)
6 8 u1 0
=
−3 −4 u2 0
which is equivalent to the equation
3u1 + 4u2 = 0.
is a solution of Ẋ = AX corresponding to λ = 0.
4 8 u1 0
=
−3 −6 u2 0
29 MAT306—X/1
to find
u1 −2
U2 = = c2
u2 1
so that
−2
X2 (t) = c2 e2t
1
is a solution of Ẋ = AX corresponding to λ = 2.
4 −2
X (t) = c1 + c2 e2t .
−3 1
1
From the initial value X (0) = , we now solve for c1 and c2 :
2
1 = 4c1 − 2c2
2 = −3c1 + c2
5 11
which yields c1 = − and c2 = − .
2 2
The solution of the initial value problem is therefore given by
x1 (t) 5 4 11 −2
X = =− − e2t
x2 (t) 2 −3 2 1
−10 + 11e2t
=
15 11
.
− e 2t
2 2
Exercise 2.6
4 2
(a) Ẋ = X
3 3
1 −1 4
(b) Ẋ = 3 2 −1 X
2 1 −1
1 −1 0
(c) Ẋ = −1 2 −1 X
0 −1 1
30
−3 2
(d) Ẋ = X
0 −3
5 4 2
(e) Ẋ = 4 5 2 X.
2 2 2
By using techniques to be developed in the next section, n linearly independent solutions X1 (t) , X2 (t) , . . . ,
Xn (t) of the equation Ẋ = AX, can always be found, even if some of, or all, the roots of the characteristic
equation are coincident. This implies that the space of solutions of Ẋ = AX is n—dimensional: the solution
space is namely a subspace of an n—dimensional space, so that its dimension cannot exceed n. That its
dimension is exactly n, now follows from the existence of n linearly independent solutions.
From the above it follows that any solution of Ẋ = AX is of the form X (t) = k1 X1 (t) + . . . + kn Xn (t) with
Xi (t) , i = 1, . . . , n, the above solutions.
Since any other set of n linearly independent solutions (known as a fundamental set of solutions) also forms
a base for the solution space, it is sufficient, in order to determine a general solution of Ẋ = AX, to find in
any manner any set of n linearly independent solutions.
If one verifies, by differentiation and by evaluating the Wronskian, that they are actually linearly indepen-
dent solutions, no theoretical justification for one’s method is needed.
From the above, it is clear that if the equation C (λ) = 0 has m—fold roots (m < n) , a method will have to
be found for determining the (n − m) solutions of Ẋ = AX that may still be lacking and which, together
with the solutions already found, constitute a set of n linearly independent solutions. The technique for
finding these solutions will be described in Sections 2.5 as well as in Chapter 3.
Even if all the entries of A are real, it often happens that the eigenvalues of A are complex. In order to
extend Theorem 2.2 to the case where λ is complex, we must define eλt for complex λ. Let λ = a + ib with
a and b real numbers. In view of Euler’s equation we have
d
eλt U = λeλt λU = eλt AU = Aeλt U.
dt
Exercise 2.7
Prove that
d
e(a+ib)t = (a + ib) e(a+ib)t ,
dt
i.e.
d
eλt = λeλt
dt
for λ complex.
Consider again the complex vector eλt U. If U = V+iW with V, W ∈ Rn , λ = a + ib, then
Consequently
Re eλt U = eat (V cos bt − W sin bt)
and
Im eλt U = eat (V sin bt + W cos bt) .
The following theorem shows that complex eigenvalues of A may lead to real solutions of Ẋ = AX.
Theorem 2.6 If X (t) is a solution of Ẋ = AX, A real, then Re X (t) and Im X (t) are solutions.
Exercise 2.8
Do Exercise 2.7, Part 1:1, p. 79 of the prescribed book.
Example 2.2
Find real—valued fundamental solutions of
2 3
Ẋ = X.
−3 2
Solution:
The characteristic equation is λ2 − 4λ + 13 = 0. The roots are λ1 = 2 + 3i and λ2 = 2 − 3i. Choose
λ = 2 + 3i. The eigenvector equation is now
A solution of this linear system is v2 = iv1 , where v1 is any nonzero constant (note that the two equations
are equivalent). We take v1 = 1 so that V = (1, i)T is an eigenvector. A complex—valued solution is
1
X1 (t) = e(2+3i)t
i
1 0
= e2t (cos 3t + i sin 3t) +i
0 1
cos 3t sin 3t
= e2t +i .
− sin 3t cos 3t
cos 3t sin 3t
e2t and e2t .
− sin 3t cos 3t
As for solutions corresponding to λ = 2 − 3i : This eigenvalue yields identical solutions up to a constant, i.e.
solutions which can be obtained by multiplying the solutions already found by a constant. This happens
whenever eigenvalues occur in complex conjugate pairs — cf again the result of Exercise 2.82 .
cos 3t sin 3t
c1 e2t + c2 e2t .
− sin 3t cos 3t
2
The implication of this is that you need not construct solutions corresponding to the second complex eigenvalue as this
yields nothing new.
33 MAT306—X/1
Exercise 2.9
Do Exercise 2.7, Part 1:3, 5, p. 79 of the prescribed book.
Assume that the equation C (λ) = 0 has multiple roots. In order to find the general solution of Ẋ = AX,
we seek a method for determining the eigenvectors which may be lacking.
1 1
so that λ = 1 is a double root of C (λ) = 0. An associated eigenvector is . Thus X1 (t) = et is
0 0
a solution.
x1
Suppose that X̂ (t) = is a second solution. Then
x2
d x1 1 1 x1 x1 + x2
= = .
dt x2 0 1 x2 x2
ẋ1 = x1 + x2 (1)
ẋ2 = x2 (2)
x2 (t) = et . (3)
so that x1 (t) = et + tet together with x2 (t) = et is a solution of the system. Consequently
et (t + 1) 1 1
X̂ (t) = = tet + et .
et 0 1
34
By differentiation we see that X̂ (t) is indeed a solution. We check for linear independence:
et et (t + 1) 2
det X1 (t) X̂ (t) = det = et = 0.
0 et
with k1 , k2 arbitrary constants. If we compare X1 (t) and X̂ (t) , we see that X̂ (t) is of the form
1
et (B + tC) with C =
0
are also solutions. We return to this method for finding new solutions in Chapter 3.
Example 2.3
Find a general solution of
0 1 0
Ẋ = 0 0 1 X ≡ AX.
−2 −5 −4
Solution:
Put C (λ) = 0, i.e.
−λ 1 0
det 0 −λ 1 = 0.
−2 −5 −4 − λ
Then −4λ2 − λ3 − 5λ − 2 = 0 from which we obtain λ = −1, −1, or −2. We first determine an eigenvector
u1
U = u2
u3
−C = AC (2.3)
−B + C = AB. (2.4)
This yields
b1 1 0 1 0 b1 b1
− b2 + −1 = 0 0 1 b2 where b2 = B,
b3 1 −2 −5 −4 b3 b3
i.e.
−b1 + 1 = b2
−b2 − 1 = b3
−b3 + 1 = −2b1 − 5b2 − 4b3 .
with c1 , c2 , c3 arbitrary constants. That X1 (t) , X2 (t) and X3 (t) are linearly independent, can be checked
by calculating their Wronskian.
1
Note that in this example 0 = B is not an eigenvector associated with the eigenvalue λ = −1 of
−1
0 1 0
A = 0 0 1 ,
−2 −5 −4
since
1 1 0 1
(A − λI) B = 0 1 1 0 = C = 0.
−2 −5 −3 −1
B is known as a generalised eigenvector or root vector of A. This subject is dealt with in Chapter 3.
Exercise 2.10
−3 4
(a) A = ,
−1 1
2 0 0
(b) A = 1 2 0 ,
−1 0 2
2 1 0
(c) A = 0 2 1 .
0 0 2
Definition 2.4 The system Ẋ = AX together with an (initial) condition X (t0 ) = X0 is known as an
initial value problem. A solution to the problem is a differentiable function X (t) that is a solution of
Ẋ = AX and such that X (t0 ) = X0 .
Exercise 2.11
Study Examples 2.4.2 and 2.4.3, p. 65 of the prescribed book. These examples illustrate that application
of the eigenvalue—eigenvector method does not always lead to a solution of Ẋ = AX which satisfies the
initial condition. This problem, as can be expected, is experienced when the characteristic equation has
multiple roots. The following theorems contain sufficient conditions under which a solution for the initial
value problem Ẋ = AX, X (t0 ) = X0 can be found by means of the eigenvalue—eigenvector method.
Theorem 2.8 If n linearly independent eigenvectors exist with corresponding real (not necessarily differ-
ent) eigenvalues of the n×n matrix A, then the initial value problem Ẋ = AX, X (t0 ) = X0 has a solution
for every X0 ∈ Rn .
Exercise 2.12
Prove Theorem 2.8.
Theorem 2.9 If A has n real different eigenvalues, then Ẋ = AX, X (t0 ) = X0 , has a solution for every
X0 ∈ Rn .
Exercise 2.13
Study Examples 2.4.4 and 2.4.5, p. 67 of the prescribed book and Summary 2.8, p. 79 of the prescribed
book. Why is there a unique solution to the problem
0 1 0 1
Ẋ = 0 0 1 X, X (0) = 0 ?
2 1 −2 1
The following example illustrates the third point of the summary.
Example 2.4
Find a solution of
1 1 0 x1 2
Ẋ = 0 1 0 x2 ≡ AX, X0 = 0 .
0 0 2 x3 −2
Solution:
The characteristic equation has roots λ = 1, 1, 2. If we put λ = 1 into (A − λI) U = 0, we have
0 1 0 u1 0
0 0 0 u2 0 .
=
0 0 1 u3 0
Now
u1 1
U = u2 = k1 0
u3 0
satisfies this equation and
1
X1 (t) = et k1 0
0
38
is a solution of Ẋ = AX which, however, does not satisfy the initial condition, since
1
t
X1 (t) = e k1 0
0
2
can, for no values of t and k1 , yield the vector 0 . Next, put λ = 2 into (A − λI) V = 0. We then
−2
have
−1 1 0 v1 0
0 −1 0 v2 = 0
0 0 0 v3 0
from which follows
−v1 + v2 = 0
−v2 = 0.
Therefore, v1 = 0, v2 = 0, while, since v3 does not appear in the equations, v3 may be chosen arbitrarily.
Put v3 = k2 . Then
0
2t
V = k2 e 0
1
is a solution of Ẋ = AX which, however, does not satisfy the initial condition. Consequently we must check
whether X0 belongs to the span of the eigenvectors U and V.
whence
m = 2,
n = −2.
Consequently
1 0
t 2t
X (t) = 2e 0 − e 0
0 1
Exercise 2.14
Exercise 2.4, Part I, p. 69 and Exercise 2.8, p. 81. Study Example 3.1.1, p. 83. (All page numbers refer to
the prescribed book.)
Chapter 3
• solve initial value problems Ẋ = AX, X (t0 ) = X0 where the matrix A has multiple roots;
Suppose A is a square matrix and λ a multiple root of the characteristic equation with corresponding
eigenvector C = 0, i.e. eλt C is a solution of Ẋ = AX. In Section 5 of Chapter 2 we derived a second
solution, (B + Ct) eλt where B is the solution of the equation
(A−λI) B = C = 0. (3.1)
39
40
As has been previously observed, B is not an eigenvector corresponding to λ. This is clear from (3.1),
the defining equation for B. Equations of the form (3.1) in which B has to be determined and C is an
eigenvector leads to the concept of a generalised eigenvector or root vector.
Before formally introducing the concept of a generalised eigenvector or root vector, we observe that equation
(3.1) may be expressed in a form in which C does not appear at all and in which the right hand side equals
zero, viz
(A − λI)2 B = (A − λI) C = 0 (3.2)
by applying A−λI to (3.1). If λ is at least a triple root of the characteristic equation, then
1 2
C t + Bt + D eλt
2
(A−λI) C = 0, C = 0
(A−λI) B = C
(A−λI) D = B
are satisfied. The last two equations may, by applying (A − λI) and (A − λI)2 , be written as
(A − λI)2 B = 0 (3.3)
(A−λI)3 D = 0, (3.4)
because
Note that on the other hand, we have (A − λI) B = 0 and (A − λI)2 D = 0, because B and D are not
eigenvectors corresponding to λ.
Generally, we now consider equations of the form (A − λ∗ I)k U = 0, U = 0, with k the smallest number
for which this relation holds, in other words, for i = 1, . . . , k, the relationship
(A − λ∗ I)k−i U = 0
holds.
Definition 3.1 The non—zero vector V is said to be a root vector of order k, k ≥ 1, of the matrix A,
if a number λ∗ exists such that
(A − λ∗ I)k V = 0 (3.5)
(A − λ∗ I)k−1 V = 0. (3.6)
441 MAT306—X/1
If we bear in mind that V is said to be annihilated by (A − λ∗ I)k if (A − λ∗ I)k V = 0, Definition 3.1 can
be reformulated as:
Definition 3.2 The vector V is a root vector of order k, k ≥ 1, of A, if a number λ∗ exists such that
V is annihilated by (A − λ∗ I)k , but not by (A − λ∗ I)k−1 .
Note that unless (A − λ∗ I) is singular (non—invertible), (A − λ∗ I)k would be a nonzero vector for all k.
Hence the existence of k, V = 0 and λ∗ satisfying (3.5) — (3.6) implies A−λ∗ I is singular, which means that
λ∗ must be an eigenvalue of A. It is now clear, by putting k = 1, that an eigenvector is a root vector of
order 1 while the vectors B and D above are root vectors of order 2 and 3 respectively.
Exercise 3.1
Study p. 89 and Example 3.2.1 of the prescribed book.
The following theorem, which is a summary of the contents of p. 89 of the prescribed book, provides a
condition under which a vector V is a root vector of order k, while it also supplies a method for constructing
a sequence of root vectors such that the order of each root vector is one less than that of its predecessor.
This theorem will, after a few more formalities, supply us with an easy method for finding the general
solution of Ẋ = AX in the case where some or all of the eigenvalues of A are coincident.
Then Vk , Vk−1 , . . . , V1 are root vectors of order k, k − 1, . . . , 1 respectively if and only if V1 is an eigen-
vector of A corresponding to λ∗ .
If the root vectors correspond to different eigenvalues, one must distinguish between the different sequences
of root vectors.
Definition 3.3 If Ui,k is a root vector of order k, corresponding to the eigenvalue λi of A, the sequence
of root vectors {Ui,k−j } , j = 0, . . . , k − 1, defined by
(A − λi I) Ui,k−j = Ui,k−(j+1) , j = 0, . . . , k − 2
Exercise 3.2
Theorem 3.2 Suppose Uk is a root vector of order k, corresponding to the eigenvalue λ of A. Then
tk−1
X (t) = eλt Uk + tUk−1 + . . . + U1
(k − 1)!
is a solution of Ẋ = AX.
tk−2
(A − λI) X (t) = eλt Uk−1 + tUk−2 + . . . + U1 .
(k − 2)!
(Note that (A − λI) U1 = 0.)
Hence
tk−1
Ẋ (t) = λeλt Uk + tUk−1 + . . . + U1
(k − 1)!
tk−2
+eλt Uk−1 + . . . + U1
(k − 2)!
= λX (t) + (A − λI) X (t)
= AX (t) .
Exercise 3.3
Prove that if Uk are root vectors of order k, corresponding to the eigenvalue λ of A, then
X1 (t) = eλt U1
X2 (t) = eλt (U2 + tU1 )
..
.
tk−2
Xk−1 (t) = eλt Uk−1 + tUk−2 + . . . + U1
(k − 2)!
tk−2
Xk (t) = eλt Uk + tUk−1 + . . . + U1
(k − 2)!
is a solution of Ẋ = AX), we have that k solutions can be found from k root vectors of orders 1, . . . , k.
443 MAT306—X/1
If we can prove that these solutions are linearly independent, we can find the general solution of Ẋ = AX when
the zeros of the characteristic polynomial are coincident, i.e. we now have at our disposal sufficient methods
for finding the solution of Ẋ = AX if the roots of the characteristic equation are different and if the roots
of C (λ) = 0 are coincident.
We first prove
Lemma 3.1 If U1 , . . . , Uk are root vectors of order 1, . . . , k corresponding to the eigenvalue λ of A, then
U1 , . . . , Uk are linearly independent.
From Definition 3.1 we have that (A − λI)k−i Uk = 0, i ≥ 1. Apply (A − λI)k−1 to (3.8). Then
0 = ck (A − λI)k−1 Uk (3.9)
since all the other terms vanish in view of the index of A − λI being equal to or greater than k. Since
(A − λI)k−1 Uk = 0, (3.9) can only hold if ck = 0.
Proof: Put
c1 X1 (t) + . . . + ck Xk (t) = 0,
i.e.
t2
eλt c1 U1 + c2 (U2 + tU1 ) + c3 U3 + tU2 + 2! U1 + ...
tk−2
. . . + ck−1 Uk−1 + tUk−2 + . . . + (k−2)! U1 + ck (Uk + tUk−1 + . . .
tk−1
... + (k−1)! U1 ) = 0.
44
t2
. . . + Uk−2 ck−2 + ck−1 t + ck + Uk−1 (ck−1 + ck t) + Uk ck = 0.
2!
This yields
tk−1
c1 + c2 t + . . . + ck = 0 ∀t,
(k − 1)!
tk−2
c2 + c3 t + . . . + ck = 0 ∀t,
(k − 2)!
..
.
ck−1 + ck t = 0 ∀t,
ck = 0,
since eλt = 0 and U1 , . . . , Uk are linearly independent. Consequently ci = 0, i = 1, . . . , k.
Example 3.1
Find a general solution of
0 1 0
Ẋ = 0 0 1 X ≡ AX
8 −12 6
by using root vectors.
Solution:
Put
−λ 1 0
det 0 −λ 1 = 0,
8 −12 6 − λ
i.e.
−λ {(6 − λ) (−λ) + 12} + 8 = 0.
Then −λ3 + 6λ2 − 12λ + 8 = 0 from which we obtain λ = 2, 2, 2. Next we determine the root vectors
U, V, W of order 1, 2, 3 respectively.
is a solution of Ẋ = AX.
from which
w1 3/4
W = w2 = 1/2 .
w3 0
It is easy to verify that we have three linearly independent solutions, by determining their Wronskian.
Exercise 3.4
Exercise 3.3:7—17, p. 101 of the prescribed book.
Remark:
In the above example we have constructed the solution of an initial value problem of the form Ẋ = AX if
A has constant entries, i.e. the entries are (real) numbers.
In Section 3.5 of the prescribed book an alternative method is used to construct the general solution of
Ẋ = AX by treating the initial value problem Ẋ = AX, X (0) = X0 as the fundamental problem and then
deducing a general solution of Ẋ = AX from the solution of the initial value problem. The method in
which use is made of annihilating polynomials is interesting, but cumbersome and therefore in general not
recommended.
The aim of this section is to present a theorem on the solvability of the initial value problem Ẋ = AX,
X (0) = X0 and to “revisit” symmetric matrices.
The result of the next theorem is stronger than the results in Theorems 2.8 and 2.9 in the sense that the
conditions “If ....” are now no longer necessary.
Theorem 3.3 (Existence Theorem for Linear Systems with Constant Coefficients) For
any choice of X0 ∈ Rn and any n × n matrix A with constant entries, a vector function X (t) exist such
that Ẋ = AX (t) , X (0) = X0 .
Remark:
(i) Note that the only condition is that the entries of A must be real – by virtue of the solution techniques
developed in the preceding sections, the conditions imposed in Theorems 2.8 and 2.9 can be removed.
(ii) The solution whose existence is guaranteed by this theorem, is unique. This will be proved in the
next chapter. In Chapter 7 we will prove a more general result in the sense that the matrix A can be
dependent on t, i.e. the entries of A may be functions of t. Students should take care to distinguish
between the two cases.
In case the reader doubts the value of existence and uniqueness theorems, he should read an article by AD
Snider on this subject: “Motivating Existence—Uniqueness Theory for Applications Oriented Student”2 .
1
The proof which makes use of annihilating polynomials, can be found in Section 3.6 of the prescribed book.
2
American Mathematical Montly, 83(1967), 805—807.
447 MAT306—X/1
In this article the writer applies the fact that the one—dimensional wave problem
∂2u 2
2∂ u
= c , x > 0, t > 0,
∂t2 ∂x2
u (x, 0) = f (x)
∂u
(x, 0) = g (x)
∂t
u (0, t) = 0
has one and only one solution, to answer the following question: if a moving wave meets the boundary, does
the reflected wave travel below or above the position of equilibrium? The uniqueness of the solution of the
wave equation enables one to deduce that the wave travels below the position of equilibrium.
The last result of this chapter pertains to symmetric matrices. We already know that symmetric matrices
have only real eigenvalues. In view of our knowledge of root vectors, we can prove the following result:
Theorem 3.4 If A is a real symmetric matrix, then A has no root vectors of order n ≥ 2.
Study the proof, which can be found on p. 116 of the prescribed book. Give yourself account for every
step. Please note that the second line in the proof should be (A−λI) a = 0 (since a is an eigenvector).
Exercise 3.5
Exercise 3.2, Part 2, no. 6, p. 94 of the prescribed book.
48
Chapter 4
Fundamental Matrices
Non—homogeneous Systems
The Inequality of Gronwall
Objectives for this Chapter
The main objective of this chapter is to gain an understanding of the following concepts regarding funda-
mental matrices and non—homogeneous systems:
• a fundamental matrix of Ẋ = AX at t0 ;
• use the method of variation of parameters to solve the non—homogeneous problem Ẋ = AX + F (with
or without the initial condition X (t0 ) = X0 );
The concept of the fundamental matrix of a system of differential equations arises when the solutions (which
are vectors) of the system are grouped together as the columns of a matrix. Fundamental matrices provide
a convenient method of formulating the behaviour of the system Ẋ = AX in an elegant way.
Important results in this chapter are the uniqueness theorems for the homogeneous initial value problem
Ẋ = AX, X (t0 ) = X0 and the non—homogeneous initial value problem Ẋ = AX + F (t) , X (t0 ) = X0 .
Definition 4.1 An n × n matrix Φ with the property that its columns are solutions of Ẋ = AX, and
linearly independent at t0 , is said to be a fundamental matrix of Ẋ = AX at t0 .
Exercise 4.1
Study Example 4.1.1, p. 123 of the prescribed book. It illustrates the fact that a system may have two
fundamental matrices at a point t0 , i.e. the fundamental matrix of a system at a point t0 is not unique.
In order to construct fundamental matrices, it is necessary to find n solutions of Ẋ = AX which are linearly
independent at t0 . From the contents of the two previous chapters, we know that this is always possible.
Therefore the following theorem holds.
For matrices the equation χ̇ = Aχ, with χ a matrix function of t, is analogous to the equation Ẋ = AX with
X a vector function of t. (Note that we use a notation different from that of the prescribed book as by X
we mean a vector function.) A fundamental matrix of Ẋ = AX at t0 may be formulated in terms of this
equation. To begin with, we define:
d
Definition 4.2 If Φ (t) is the matrix [aij (t)] , then Φ̇ (t) = [Φ (t)] is the matrix [ȧij (t)], i.e. the
dt
matrix of which the entries are the derivatives of the corresponding entries of Φ (t) .
On pp. 45—46 of the prescribed book the reader will find more on the differentiation of matrices.
Remark:
Condition (b) is equivalent to the condition: (b ) Φ−1 (t0 ) exists.
50
Exercise 4.2
On the strength of Theorem 4.2 it is possible to give a representation of the solutions of Ẋ = AX in terms
of fundamental matrices.
Corollary 4.1 If Φ is a fundamental matrix of Ẋ = AX at t0 , then X (t) = Φ (t) K with K = Φ−1 (t0 ) X0 ,
is a solution of the initial value problem Ẋ = AX, X (t0 ) = X0 .
In the expression Φ (t) Φ−1 (t0 ) X0 , we are dealing with the product of two matrices of which the second is
constant. From Theorem 4.2 we deduce
If t = t0 , then Φ (t) Φ−1 (t0 ) is the identity matrix, so that at t = t0 we have Φ (t) Φ−1 (t0 ) X0 = X0 . This
leads to the concept of normalised fundamental matrix at t0 .
Definition 4.3 A fundamental matrix Φ (t) of Ẋ = AX at t0 with the property that Φ (t0 ) = I, the
identity matrix, is said to be a normalised fundamental matrix at t0 .
Remark:
A fundamental matrix Φ (t) of Ẋ = AX at t0 can therefore be normalized by multiplying by Φ−1 (t0 ) from
the right as follows:
Ψ (t) = Φ (t) Φ−1 (t0 ) (4.1)
where Ψ (t) is then the normalized fundamental matrix of Ẋ = AX at t0 . (Clearly, from Corollary 4.2, the
matrix Ψ (t) is a fundamental matrix and furthermore, at t = t0 , we have
The initial value problem Ẋ = AX, X (t0 ) = X0 , can therefore be solved simply by finding a normalised
fundamental matrix at t0 .
51 MAT306—X/1
Exercise 4.3
Prove Theorem 4.3 and Corollaries 4.1, 4.2.
Example 4.1
Find the solution of the initial value problem
0 1 2
Ẋ = X, X (0) = .
−3 4 4
Solution:
The vector functions
1 1
X1 (t) = et and X2 (t) = e3t
1 3
are two linearly independent solutions of the given system. Therefore
et e3t
Φ (t) =
et 3e3t
Exercise 4.4
Exercise 4.1, p. 128 of the prescribed book.
In the next theorem it is shown that if we take B (t) as the transpose of a normalised fundamental matrix
of
T
Ẋ = −A X (4.3)
T
Theorem 4.4 Assume that Φ and Ψ are normalised fundamental matrices of Ẋ = AX and Ẋ = −A X at
t0 respectively. Then
ΨT (t) Φ (t) = I ∀t,
i.e.
(Φ (t))−1 = ΨT (t) .
In words the theorem reads: The inverse of a normalised fundamental matrix Φ (t) of Ẋ = AX at t0 , is
T
equal to the transpose of the normalised fundamental matrix of the conjugate equation Ẋ = −A X at t0 .
Proof: Since Ψ is differentiable, ΨT and ΨT Φ are likewise differentiable. From the rules for differentiating
matrix functions (see again p. 46 of the prescribed book) we have
d d dΦ
ΨT Φ = ΨT Φ + ΨT
dt dt dt
T
dΨ
= Φ + ΨT AΦ
dt
T
= −AT Ψ Φ + ΨT AΦ
= −ΨT AΦ + ΨT AΦ
= 0
T
where in the second last step we have made use of (AB)T = BT AT and AT = A for any matrices A
and B.
Consequently ΨT Φ = C, a constant matrix, i.e. ΨT Φ (t) = C for all t. In particular, ΨT (t0 ) Φ (t0 ) = C
holds. Since Ψ and Φ are normalised fundamental matrices, we have I = ΨT (t0 ) Φ (t0 ) = C, so that
C = I holds. This completes the proof.
Proof: Suppose Φ is a fundamental matrix at t0 as well as non—singular in t0 , i.e. Φ−1 (t0 ) exists. From
the definition Φ (t) Φ−1 (t0 ) is then a normalised fundamental matrix at t0 , so that Φ (t) Φ−1 (t0 ) is non—
singular for every t, i.e. a matrix B exists so that Φ (t) Φ−1 (t0 ) B (t) = I for every t. This implies that
Φ (t) itself is non—singular for every t.
Corollary 4.5 Solutions of Ẋ = AX which are linearly independent at t = t0 , are linearly independent
for all t.
Proof: The proof follows immediately from the preceding corollary by recalling that solutions of Ẋ = AX are
linearly independent at t0 if det Φ−1 (t0 ) = 0, where Φ is a fundamental matrix of Ẋ = AX at t0 . This
requirement is equivalent to that of the non—singularity of Φ at t0 .
Proof: From Definition 4.1, the result is merely a formulation of the preceding result.
53 MAT306—X/1
From the above discussion, it follows that vector functions satisfying Ẋ = AX, cannot be linearly indepen-
dent at one point and linearly dependent at another.
Exercise 4.5
Show by means of an example that arbitrary chosen vectors, linearly independent at a point t0 , are not
necessarily linearly independent for all t. The property of Corollary 4.5 is, therefore, a specific property
of solutions of Ẋ = AX. The converse, however, is not true, i.e. a system of vector functions, linearly
independent for all t, does not necessarily constitute a solution of Ẋ = AX with A some or other matrix.
See Exercise 4.2.II:2, p. 134 of the prescribed book for an example.
Theorem 4.5 Any solution of Ẋ = AX is of the form X (t) = Φ (t) X (t0 ) with Φ a normalised funda-
mental matrix of Ẋ = AX at t0 .
Proof: We show that Φ−1 (t) X (t) = Z (t) is a constant vector by showing that Ż (t) = 0. From Theorem
4.4 we have Z (t) = ΨT (t) X (t) with Ψ (t) a solution of χ̇ = −AT χ. Hence
T
Ż = Ψ̇ X+ ΨT Ẋ
T
= −AT Ψ X + ΨT AX
= −ΨT AX + ΨT AX
= 0.
Consequently, Z (t) = Φ−1 (t) X (t) = C, with C a constant vector. If we put t = t0 , it follows that
C = X (t0 ) , so that Φ−1 (t) X (t) = X (t0 ) , i.e. X (t) = Φ (t) X (t0 ) .
Definition 4.4 The vector function Φ (t) K is said to be a general solution of Ẋ = AX if Φ is any
fundamental matrix of the system Ẋ = AX.
Corollary 4.7 Any solution of Ẋ = AX that vanishes at t0 , is identically zero. In particular, the problem
Ẋ = AX, X (0) = 0, has the unique solution X (t) ≡ 0.
We now have all the tools for proving the uniqueness theorem for linear systems with constant coefficients.
Theorem 4.61 (Uniqueness Theorem for Linear Systems with Constant Coefficients)
The initial value problem Ẋ = AX, X (t0 ) = X0 in which A is a constant matrix, has a unique solution.
1
This theorem must not be confused with the Existence and Uniqueness Theorem for linear systems of the form
Ẋ = A (t) X, i.e. the case where the entries of A vary with time t (Chapter 7).
54
Proof: We have to show that if X (t) and Y (t) are both solutions, then X (t) ≡ Y (t) . Put Z (t) =
X (t) − Y (t) . Then
Therefore Z (t) is a solution of Ẋ = AX which vanishes at t0 . By Corollary 4.7 we conclude that Z (t) = 0 ∀t,
i.e. X (t) ≡ Y (t) .
Corollary 4.8 Let Φ be a normalised fundamental matrix for the system Ẋ = AX at t0 . Then χ (t) =
Φ (t) χ0 is the unique solution of the matrix initial value problem χ̇ = Aχ, χ (t0 ) = χ0 .
The uniqueness theorem for the initial value problem Ẋ = AX, X (t0 ) = X0 , may be reformulated in terms
of fundamental matrices.
Theorem 4.7 There exists one and only one normalised fundamental matrix of Ẋ = AX at t0 .
Exercise 4.6
Prove Corollary 4.8 and Theorem 4.7.
The following theorems, which deal with the properties of fundamental matrices, which may be normalised,
are proved by using the Uniqueness Theorem 4.6.
Proof: Let s be fixed, say s = s0 . We show that both sides of (4.4) are solutions of the matrix initial value
problem
χ̇ = Aχ, χ (0) = Φ (s0 ) . (4.5)
From Corollary 4.8, we have that χ (t) = Φ (t) Φ (s0 ) satisfies (4.5), i.e.
d
[Φ (t) Φ (s0 )] = A [Φ (t) Φ (s0 )] and χ (0) = Φ (s0 ) .
dt
d
[Φ (t)] = AΦ (t) ∀t,
dt
so that in particular
d
[Φ (s0 + t)] = AΦ (s0 + t) ∀t.
dt
55 MAT306—X/1
Moreover Φ (s0 + t) = Φ (s0 ) at t = 0. Consequently χ (t) = Φ (s0 + t) is a solution of (4.5). Since the
solution of (4.5) is unique, Φ (s0 + t) = Φ (t) Φ (s0 ) holds for arbitrary s0 and all real numbers t. Therefore
Exercise 4.7
1. Show by means of an example that the hypothesis that Φ is a normalised fundamental matrix at
t0 = 0, cannot be scrapped.
Note that Φ (0) = I follows from the definition, so that Φ indeed displays the properties of the
exponential function.
Corollary 4.9 If Φ is a normalised fundamental matrix at t0 = 0, then Φ−1 (t) = Φ (−t) for all t.
Remark:
This formula provides a method for determining the inverse of a fundamental matrix of a system of differ-
ential equations.
Exercise 4.8
2. Study the examples of Section 4.3 of the prescribed book (pp. 136—137).
We merely mention the following results – more detail can be found in the prescribed book.
Theorem 4.9 (Transition Property) Suppose the system Ẋ = AX has the fundamental matrices
Φ0 and Φ1 . Suppose Φ0 and Φ1 are respectively normalised at t0 and t1 . Then
Corollary 4.11 With the notation the same as in the previous theorem we have
Φ1 (t0 ) Φ0 (t1 ) = I.
AΦ (t) = Φ (t) A.
56
Exercise 4.9
1. Show by means of an example that the hypothesis that Φ is a normalised fundamental matrix, is
indispensable.
In the case of the n—dimensional problem Ẋ = AX + F (t) , the part played by the exponential function in
the one—dimensional problem, is played by a fundamental matrix of the complementary system Ẋ = AX. We
assume the existence of a vector U such that
is a solution of
Ẋ = AX + F (t) .
Then
Consequently
Φ (t) U̇ (t) = F (t) ,
so that
U̇ (t) = Φ−1 (t) F (t)
since Φ−1 exists. We therefore have
t
Xp (t) = Φ (t) U (t) = Φ (t) Φ−1 (s) F (s) ds .
t0
Proof: We have to show that Ẋp (t) = AXp (t) + F (t) . Since Φ is a fundamental matrix of Ẋ = AX, we
have from Theorem 4.2 that Φ̇ (t) = AΦ (t) . We therefore have
t t
d
Ẋp (t) = AΦ Φ−1 (s) F (s) ds + Φ (t) Φ−1 (s) F (s) ds
t0 dt t0
t
= AΦ (t) Φ−1 (s) F (s) ds + Φ (t) Φ−1 (t) F (t)
t0
t
= AΦ (t) Φ−1 (s) F (s) ds + F (t)
t0
Example 4.2
Determine a solution of
1 0 et
Ẋ = X+ ≡ AX + F.
−1 3 1
Solution:
From (4.6), Xp (t) = Φ (t) U (t) is a solution of Ẋ = AX + F, where Φ (t) U̇ (t) = F (t) with Φ a funda-
mental matrix of Ẋ = AX. A fundamental matrix is
2et 0
et e3t
so that we must solve for U from
2et 0 et
U̇ = .
et e3t 1
If
u̇1
U̇ = ,
u̇2
we have
Consequently
t
U= 2
e−3t e−2t
3 + 4
and
t
2et 0 2 tet
Φ (t) U (t) = e−3t e−2t
= t
et e3t 3 + 4 2 + 14 et − 1
3
is a solution of the given problem.
Example 4.3
Use Theorem 4.11 to find the general solution of the inhomogeneous system
−1 1 1
Ẋ = X+
−2 1 cot t
π
where t0 = .
2
Solution:
The characteristic equation C (λ) = 0 has roots λ = ±i. An eigenvector corresponding to λ = i is
1
U= .
1+i
From this we obtain, with the aid of Theorem 2.7, the two real solutions
cos t sin t
and .
cos t − sin t cos t + sin t
A fundamental matrix is therefore
cos t sin t
Φ (t) = .
cos t − sin t cos t + sin t
Now
cos t + sin t − sin t
Φ−1 (t) = .
sin t − cos t cos t
π
Therefore, by choosing t0 = , we have from Theorem 4.11
2
t
cos s + sin s − sin s 1
Xp (t) = Φ (t) ds
π
2
sin s − cos s cos s cot s
= Φ (t) J
where
t
sin s
J = ds
π
2
− cos s + cosec s
cos2 s 1 − sin2 s
(recall = = cosec s − sin s)
sin s sin s
t
− cos s
=
− sin s − ln |cosec s + cot s| π
2
− cos t
= .
− sin t − ln |cosec t + cot t| + 1
59 MAT306—X/1
This yields
Xp (t) = Φ (t) J
Remark:
The inverse Φ−1 (t) of the matrix
cos t sin t
Φ (t) =
cos t − sin t cos t + sin t
could quite easily be determined by recalling that the inverse of the matrix
A B
S=
C D
is given by
1 D −B
S−1 = , provided that AD − BC = 0.
AD − BC −C A
is a solution of Ẋ = AX + F.
Remark:
Note that the solution Xp (t) of Ẋ = AX + F, given by (4.7) or (4.9) vanishes at t0 , since the integral
becomes zero at t0 = 0.
Example 4.4
Use Corollary 4.12 to solve the inhomogeneous system
4 5 4et cos t
Ẋ = X+
−2 −2 0
by taking t0 = 0.
Solution:
The corresponding homogeneous system has the characteristic equation
C (λ) = λ2 − 2λ + 2 = 0
1 + i : Solve
3−i 5 v1 0
=
−2 −3 − i v2 0
or equivalently
(3 − i) v1 + 5v2 = 0 (1)
−2v1 + (3 + i) v2 = 0. (2)
(Note that the second equation is identical to the first one, i.e. − (3 − i) /2×(2) = (1) .) Choose, for example,
v1 = 5, then v2 = −3 + i, so that an eigenvector corresponding to 1 + i is
5 5 0
V= = +i .
−3 + i −3 1
Thus we get the two real solutions of the corresponding homogeneous problem
5 0
X1 (t) = et cos t − sin t
−3 1
5 cos t
= et
−3 cos t − sin t
and
5 0
X2 (t) = et sin t + cos t
−3 1
5 sin t
= et .
−3 sin t + cos t
5 cos t 5 sin t
Φ (t) = et .
−3 cos t − sin t −3 sin t + cos t
Hence
5 0
Φ (0) =
−3 1
and thus
1 1 0
Φ−1 (0) =
5 3 5
t
cos (t − s) + 3 sin (t − s) 5 sin (t − s) 4es cos s
= et−s ds
0 −2 sin (t − s) −3 sin (t − s) + cos (t − s) 0
t
cos (t − s) cos s + 3 sin (t − s) cos s
= 4et ds
0 −2 sin (t − s) cos
t
cos t + cos (t − 2s) + 3 sin t + 3 sin (t − 2s)
= 2et ds
0 −2 sin t − 2 sin (t − 2s)
t
t s cos t + 3s sin t − 12 sin (t − 2s) + 32 cos (t − 2s)
= 2e
−2s sin t − cos (t − 2s)
0
Exercise 4.10
1. Find a particular solution of the problem in Example 4.2 by using formula (4.7). Show that the answer
3t T
obtained and the solution already found, differ by 0, e12 – a solution of the complementary system
Ẋ = AX.
2. Prove that the difference between any two particular solutions of a non—homogeneous system, is a
solution of the complementary homogeneous system.
From this it follows that the solutions of the non—homogeneous system Ẋ = AX + F may be expressed
as X (t) = Xc (t) + Xp (t) with Xp (t) a particular solution of Ẋ = AX + F and Xc (t) a general solution
of Ẋ = AX. As Xc (t) assumes all solutions of Ẋ = AX, it follows that X (t) assumes all solutions of
Ẋ = AX + F. As a result of Definition 4.4, we define a general solution of Ẋ = AX + F as follows:
Definition 4.5 Suppose that Xp (t) is any particular solution of the non—homogeneous system
Ẋ = AX + F and Φ a fundamental matrix of the complementary system Ẋ = AX. Then the vector
function
X (t) = Xp (t) + Φ (t) K (4.10)
The fact proved in Exercise 4.9, no. 2, together with the uniqueness theorem for homogeneous initial value
problems, enables us to prove a uniqueness theorem for the corresponding non—homogeneous initial value
problem.
62
Theorem 4.12 (Uniqueness Theorem for the Non—Homogeneous Initial Value Problem
Ẋ = AX + F, X (t0 ) = X0 ) The problem
Ẋ = AX + F, X (t0 ) = X0 (4.11)
Proof: By differentiation it follows that the function given by equation (4.12), is a solution of Ẋ = AX + F,
which also satisfies the initial condition. Suppose Y (t) is any solution of (4.11). We then have to prove
that X (t) ≡ Y (t) . Define Z (t) = X (t) − Y (t) . Then
and
Z (t0 ) = X (t0 ) − Y (t0 ) = 0.
The vector function Z (t) is, therefore, a solution of a homogeneous system which vanishes at t0 . From
Corollary 4.7, Z (t) is identically zero, so that X (t) ≡ Y (t) . This completes the proof.
Example 4.5
Find a general solution of Ẋ = AX + F, where
1 0 et
A= , F (t) = .
−1 3 1
Solution:
Take t0 = 0.
A general solution of Ẋ = AX is
2 0
X (t) = c1 et + c2 e3t ,
1 1
2et 0
Φ (t) = .
et e3t
663 MAT306—X/1
We find a particular solution by using Corollary 4.12. A normalized fundamental matrix at t0 = 0 for the
homogeneous system is given by
2et 0 1 1 0
=
et e3t 2 −1 2
et 0
= et − e3t .
e3t
2
From equation (4.9), a particular solution is
t
Xp (t) = Ψ (t − s) F (s) ds
0
t et−s 0 es
= et−s −e3(t−s)
ds
0 2 e3(t−s) 1
t
et
= et −e3t−2s
ds
0 2 + e3(t−s)
t
set
= set e3t−2s e3(t−s)
2 − −4 + −3 0
by using the fact that the integral of a matrix function A (t) = [aij (t)] is the matrix function with entries
aij (t) dt (see p. 46 of the prescribed book).
tet tet
Xp (t) = tet et e3t e3t
= t 1 e3t 1 .
2 + 4 − 13 − 4 + 3 2 + 4 et + 12 − 3
tet 2 0
X (t) = t 1 e3t 1 + c1 et + c2 e3t ,
2 + 4 et + 12 − 3 1 1
with c1 , c2 arbitrary constants. Special additional methods still exist for determining a general solution of
the non—homogeneous system Ẋ = AX + F (see Section 4.5 of the prescribed book.)
Exercise 4.11
Exercise 4.4, p. 144—145 of the prescribed book.
64
In applications of differential equations, it is not always possible to obtain an exact solution of the differential
equation. In such cases it is sometimes necessary to find, instead, an estimate for, or an upperbound of,
the so—called norm of the solution. The latter quantity, as we will see below, measures the “size” of the
solution.
In this section we treat the inequality of Gronwall, which is a useful tool in obtaining estimates for the norm
of a solution of a differential equation of the form Ẋ = AX.
then
1
n 2
= x2i (t) .
i=1
Next we recall that for an n × n matrix A = [aij ] the number ||A|| is defined 2 by
1
n 2
||A|| = a2ij .
i,j=1
We show that
1
n 2
||A|| = a2ij
i,j=1
is indeed a norm. The system of all n × n matrices is then a normed linear space. If A denotes [aij ] and
B = [bij ] , we have
1
n 2
(ii) ||λA|| = 2
(λaij )
i,j=1
2
The norm of A may also be defined in another way, as long as the properties of a norm are satisfied – see (i)—(iii) further
on.
665 MAT306—X/1
1
n 2
i,j=1
1 1
n 2 n 2
≤ a2i,j + b2ij
i,j=1 i,j=1
= ||A|| + ||B||
where in the second last step we used the inequality of Minkowski.
Before formulating the inequality of Gronwall, we prove the following auxiliary result, which will be needed
in proving Gronwall’s inequality.
Lemma 4.1 Suppose that f and g are continuous, real—valued functions in a ≤ t ≤ b and that f˙ exists in
this interval. If, in addition, f˙ (t) ≤ f (t) g (t) in a ≤ t ≤ b, then
t
f (t) ≤ f (a) exp g (s) ds ∀t ∈ [a, b] . (4.13)
a
Exercise 4.12
Prove Lemma 4.1.
Hint: See p. 151 of the prescribed book. The only difficulty in the proof is to determine ṗ (t) . Once more
you have to use the differentiation formula (4.8), from which it follows that if
t
p (t) = exp − g (s) ds ,
a
then
t t
d
ṗ (t) = exp − g (s) ds · − g (s) ds
a dt a
t
= exp − g (s) ds [−g (t)]
a
= −p (t) g (t) .
We now proceed to
then
t
f (t) ≤ K exp g (s) ds , a ≤ t ≤ b. (4.15)
a
66
Proof: Put
t
G (t) = K + f (s) g (s) ds.
a
Then f (t) ≤ G (t) in [a, b] . G is now differentiable in [a, b] . Indeed G (t) = f (t) g (t) . Since g ≥ 0 and
f ≤ G, we have f g ≤ Gg; therefore G ≤ Gg. All the requirements of Lemma 4.1 are satisfied with the role
of f assumed by G. Consequently we have, according to inequality (4.13),
t
G (t) ≤ G (a) exp g (s) ds , a ≤ t ≤ b.
a
Now
a
G (a) = K + f (s) g (s) ds = K.
a
Therefore
t
G (t) ≤ K exp g (s) ds , a ≤ t ≤ b.
a
Exercise 4.13
Study Examples 4.6.1 and 4.6.2, p. 152 of the prescribed book, in which the inequality of Gronwall is
applied.
The technique used in these examples is applied to systems of differential equations in the following section.
As in the case of the one—dimensional equation, the procedure is to replace the differentiable equation by
an equation containing an integral, and then to take the norm.
The Gronwall Inequality may be used to obtain an estimate of the norm of solutions of the system Ẋ = AX,
where A is a constant matrix.
Theorem 4.14 If X (t) is any solution of Ẋ = AX, then the following inequality holds for all t and t0 :
(See Problem 2.4.II:3, p. 71 of the prescribed book for an interpretation of exp (tA) .)
Proof: If t = t0 , the assertion holds. If t > t0 , we have |t − t0 | = t − t0 , so that the inequality (4.16)
reduces to
||X (t)|| ≤ ||X (t0 )|| exp [||A|| (t − t0 )] .
From the properties of norms (see also Theorem 1.10.2 and Problem 1.9I:5 of the prescribed book), we have
t
||X (t)|| ≤ ||X (t0 )|| + ||A|| ||X (s)|| ds. (4.17)
t0
By putting f (t) = ||X (t)|| , g (t) = ||A|| and K = ||X (t0 )|| , it follows from Theorem 4.13 that
t
||X (t)|| ≤ ||X (t0 )|| exp ||A|| ds = ||X (t0 )|| exp [||A|| (t − t0 )] .
t0
We still have to prove the theorem for t < t0 . In this case, put Y (t) = X (−t) . If we assume that X (t) is a
solution of Ẋ = AX, it follows that
Therefore Y (t) is a solution of Ẏ = −AY. Moreover −t > −t0 . Therefore, by applying (4.17) to the system
Ẏ = −AY, we obtain
Remark:
If t − t0 > 0, the inequality (4.16) shows that ||X (t)|| does not increase more rapidly than a certain
exponential function. Such functions X (t) , are known as functions of exponential type.
Example 4.6
The norms of a vector
X (t) = (x1 (t) , x2 (t) , . . . , xn (t))T
and an n × n matrix A = [ai,j ] were previously defined by
n n
||X|| = x2i and ||A|| = a2ij .
i=1 i,j
The latter definition means that the norm of a matrix A is defined as the maximum of the sums obtained
by addition of the absolute values of the entries in each column of A.
ẋ1 = − (sin t) x2 + 4
ẋ2 = −x1 + 2tx2 − x3 + et
1
ẋ3 = 3 (cos t) x1 + x2 + x3 − 5t2 .
t
68
Assume that there exists solutions X and Y on the interval (1, 3) with
Solution:
For the given system we have
0 − sin t 0
A (t) =
−1 2t −1
1
3 cos t 1
t
so that
1
||A (t)|| = max 1 + 3 |cos t| , |sin t| + 2t + 1, 1 +
t
i.e.
Ut
||X (t) − Y (t)|| ≤ ||X (2) − Y (2)|| e| 2 (2+2s)ds |
The term e7 is obtained by computing the integral and bearing in mind that t < 3. (The proof of ∗ is similar
to the proof of (4.14).)
Exercise 4.14
2
1. Show that no solution of Ẋ = AX can satisfy ||X (t)|| = et .
2. Study Examples 4.7.1 and 4.7.2, p. 54 of the prescribed book. These illustrate that inequality (4.16)
could yield an over—estimate, but also a precise estimation.
Chapter 5
• the link between higher order one—dimensional equations and linear systems of first order differential
equations;
• companion matrix;
• companion system for the n—th order equation with constant coefficients as given by (5.1).
• rewrite the n—th order equation (5.1) as a linear system of first order differential equations;
• solve the corresponding linear system of first order differential equations and interpret the solution in
terms of the original higher order one dimensional equation.
5.1 INTRODUCTION
In Chapter 5 of the prescribed book the link that exists between higher order one—dimensional linear
differential equations and linear systems of first order differential equations is explored. The study of
Sections 5.1 through 5.5 from this chapter in the prescribed book is set as an assignment for the student,
as it is very important for students to get practise in studying from text books.
Being able to “read” a variety of books on the particular topic that you are studying, should be one of one’s
ultimate goals when studying at a university.
We shall therefore only discuss the concept of companion systems and illustrate by means of an example.
69
70
The rest is left to the student. Mark theorems which look interesting, eg. Theorem 5.3.3 and Theorem 5.5.3,
which deal with the relationship between the auxiliary equation of the general n—th order one—dimensional
equation and the characteristic equation of the companion system. In this way your understanding of n—th
order one—dimensional linear differential equations will be deepened. Work through the proofs of the
theorems (these are not required for examination purposes) and select problems from Exercise 5.2 through
5.5.
has been treated in MAT216—W. This equation can be reduced to a system of n first order equations. We
illustrate by means of an example:
If y (t) represents the distance of a body from a fixed point on a line, this equation describes the movement
of that body along the line, subject to different forces. Since y measures distance, y (1) measures velocity
and y(2) acceleration. By putting y (1) = v and y (2) = a, equation (5.2) may be expressed as
y(1) = v
y(2) = a (5.3)
y(3) = −5v + 6a + sin t.
Equation (5.5) is now merely a special case of Ẋ = AX + F (t) with the initial condition
y (0) y (0)
X (0) = v (0) = y(1) (0) = 0.
a (0) y(2) (0)
Let
y
X = y (1) .
y (2)
Then
y(1)
Ẋ = y(2)
y(3)
so that
y (1) y(1) 0
Ẋ = y (2) = y(2) + 0
y (3) −a0 y − a1 y (1) − a2 y (2) f (t)
0 1 0 0
= 0 0 1 X + f (t) 0 (5.7)
−a0 −a1 −a2 1
= AX + F (t)
where
0 1 0
A = 0 0 1
−a0 −a2 −a3
and
0
F (t) = f (t) 0 .
1
The system (5.7) is known as the companion system for the third order equation (5.6). The companion
system for the n—th order equation (5.1) may be defined in a similar way.
is known as the companion system for the n—th order equation (5.1).
is a solution of the system (5.8), then x1 , the first component of X, is a solution (5.1) and
x1
(1)
x1
X = .. ,
.
(n−1)
x1
i.e.
(1) (2) (n−1)
x2 = x1 , x3 = x1 , . . . , xn = x1 .
Exercise 5.1
We summarize: Equations of order n in one dimension may be solved by determining a general solution
x1 (t)
.
X (t) = ..
xn (t)
of the companion system. The function x1 (t) is then a solution of the given n—th order equation.
As the companion system is but a special case of the non—homogeneous linear system with constant coeffi-
cients, one needs no new techniques.
Example 5.1
Determine the general solution of
d2 y
+ 4y = sin 3t (5.9)
dt2
by using the companion system.
Solution:
The companion system is found as follows:
Rewrite (5.9) as
y (2) = −4y + sin 3t.
Let
y
X= ,
y (1)
then
y(1)
Ẋ =
y(2)
so that
y(1) 0 1 y 0
Ẋ = = +
y(2) −4 0 y (1) sin 3t
0 1 y 0
= + sin 3t
−4 0 y (1) 1
= AX + F (t) . (5.10)
−λ 1
det = 0 ⇒ λ = ±2i.
−4 −λ
Choose λ = 2i. (Remember λ = −2i yields identical solutions up to a constant). Now put
−2i 1 u1 0
= .
−4 −2i u2 0
74
Then
−2iu1 + u2 = 0
−4u1 − 2iu2 = 0.
These two equations are identical. Choose u1 = 1, then u2 = 2i. The vector
u1 1
U≡ =
u2 2i
is then an eigenvector corresponding to the eigenvalue λ = 2i. Consequently
1 0 cos 2t
X1 (t) = cos 2t − sin 2t =
0 2 −2 sin 2t
and
1 0 sin 2t
X2 (t) = sin 2t + cos 2t =
0 2 2 cos 2t
are solutions of Ẋ = AX and a general solution is
1 2 cos 2t sin 2t
Ψ (t) = Φ (t) Φ−1 (0) = .
2 −4 sin 2t 2 cos 2t
Therefore, a particular solution of the system (5.10) is
t
1 2 cos 2 (t − s) sin 2 (t − s) 0
Xp (t) = ds
2 0 −4 sin 2 (t − s) 2 cos 2 (t − s) sin 3s
t
1 sin 2 (t − s) sin 3s
= ds
2 0 2 cos 2 (t − s) sin 3s
1
1 t
2 cos (2t − 5s) − 12 cos (2t + s)
= ds
2 0 sin (2t + s) + sin (5s − 2t)
− sin(2t−5s) sin(2t+s)
t
10 − 2
1
=
2 cos(5s−2t)
− cos (2t + s) − 5 0
− 2 sin
5
3t
+ 3 sin 2t
5
1
= .
2
− 6 cos
5
3t
+ 6 cos 2t
5
75 MAT306—X/1
According to Theorem 5.1, a general solution of (5.9) is given by the first component of X (t) . Therefore
sin 3t 3 sin 2t
y (t) = c1 cos 2t + c2 sin 2t − + . (5.11)
5 10
Note that the second component of X (t) is the derivative of y (t) .
Remark:
This solution can be obtained more readily by applying the standard techniques for solving higher order one—
dimensional equations: For the equation D2 + 4 y = sin 3t, we have the auxiliary equation m2 +4 = 0 with
roots ±2i. Hence
yC.F. (t) = d1 cos 2t + d2 sin 2t.
Furthermore
sin 3t sin 3t
yP.I. (t) = 2
=− .
−3 + 4 5
Hence
sin 3t
y (t) = d1 cos 2t + d2 sin 2t − .
5
This is equivalent to the solution (5.11).
Exercise 5.2
Exercise 5.2, p. 162 of the prescribed book.
Chapter 6
• analytic matrices;
6.1 INTRODUCTION
The system Ẋ = AX, with A an n × n matrix with constant entries, is a special case of the more general
system
Ẋ = A (t) X, a < t < b, (6.1)
with A (t) = [aij (t)] a matrix whose entries are functions of time and which we will call continuous and
real—valued if all the functions aij are continuous and real—valued. Everything that holds for the system
Ẋ = A (t) X, is therefore applicable to the case where A is constant. The converse, is, however, not easy.
76
77 MAT306—X/1
Whereas we have to our disposal standard techniques for finding n linearly independent solutions of
Ẋ = AX, each of which is usually a linear combination of known functions, while we also know that exactly
n linearly independent solutions exist, the construction of only a single solution of Ẋ = A (t) X, usually
entails an enormous amount of work. Moreover, it requires advanced mathematics to prove theoretically
that n linearly independent solutions always exist.
A discussion of this theory, which is outside the scope of this course, is found, amongst others, in “Lectures
on Ordinary Differential Equations” by W. Hurewicz1 . It is sufficient to say that, as in the case of the
equation Ẋ = AX, fundamental and normalised fundamental matrices Φ (t) of Ẋ = A (t) X, are used in
the development of this theory.
Whereas the solution of Ẋ = AX are usually linear combinations of known functions such as sin t and cos t,
such linear combinations are rarely solutions of Ẋ = A (t) X. Indeed, systems Ẋ = A (t) X are often used
to define and to study new functions. For instance, the functions of Bessel, Legendre and Laguerre and the
Hermite Functions (some of which you have encountered in MAT216—W), are examples of functions defined
as solutions of certain differential equations.
The main result of this chapter is that, under certain conditions on the matrix A (t) , viz that it is possible
to write the entries of A (t) as power series, a solution of Ẋ = A (t) X can be expressed as a power series
X (t) = U0 + U1 (t − t0 ) + . . . + Uk (t − t0 )k + . . . (6.2)
about t0 . If we wish to find a power series solution of Ẋ = A (t) X, the vectors U0 , U1 , . . . have to be deter-
mined, while we also have to show that the infinite sum (6.2) converges to a solution of Ẋ = A (t) X. The
property which A (t) must satisfy, is that of analyticity, a powerful property, since continuity and even the
existence of derivatives of all orders, are not sufficient to ensure this property. However, in most cases,
the coefficient matrices of differential equations, as well as of systems that are important for all practical
purposes, do have the property of analyticity.
Although the property of analyticity is usually associated with functions of a complex variable, we can
define the property of analyticity for real functions, in view of the fact that a function of a real variable is
only a special case of a function of a complex variable.
Definition 6.1 A function F of a real variable is analytic at a real number t0 if a positive number r
and a sequence of numbers a0 , a1 , a2 , . . . exist such that
N ∞
F (t) = lim ak (t − t0 )k = ak (t − t0 )k for |t − t0 | < r. (6.3)
N→∞
k=0 k=0
The following theorem deals with the differentiation of power series. This is no trivial matter, since infinite
sums are involved.
1
W. Hurewicz, Lectures on Ordinary Differential Equations, The Technology Press, Cambridge, Mass., 1958.
78
then
∞
(k + j)!
F (j) (t) = ak+j (t − t0 )k for |t − t0 | < r,
k!
k=0
for every positive integer j.
From the above it is clear that the j—th derivative F (j) is also analytic at t0 and F (j) (t0 ) = j!aj . The
constants
1
ak = F (k) (t0 )
k!
are known as the Taylor coefficients of F at t0 , and the expansion
∞ ∞
1 (k)
ak (t − t0 )k = F (t0 ) (t − t0 )k
k!
k=0 k=0
Remark:
(1) It is worthwhile selecting analytic functions to form a class of functions for the following reasons:
(a) It is possible that a function may not possess derivatives of all orders. (Can you think of a good
example?)
(b) Functions F exist which have derivatives of all orders, but of which the Taylor series
∞
1 (k)
F (t0 ) (t − t0 )k
k!
k=0
converges only at t = t0 . It may happen that a function F has a Taylor series expansion
∞
1 (k)
F (t0 ) (t − t0 )k
k!
k=0
converging for |t − t0 | < r for some r > 0, but not to F (t) , i.e.
∞
1 (k)
F (t) = F (t0 ) (t − t0 )k ,
k!
k=0
unless t = t0 .
(2) If F is a function of the complex variable z and F is analytic at z = z0 (with the concept of analyticity
defined as in Complex Analysis), a power series expansion of the form
∞
ak (z − z0 )k
k=0
79 MAT306—X/1
equal to zero.
In this case the coefficients ak (k = 0, 1, 2, . . .) are given in terms of contour integrals by means of the
formula
1 F (z)
ak = dz
2πi C (z − z0 )k+1
with C a circle with centre z0 . This is, of course, a result from Complex Analysis, quoted here only
for the sake of interest (as this result is dealt with in the module on Complex Analysis).
A list of Taylor series for some elementary functions is supplied on p. 196 of the prescribed book.
The following theorem deals with addition, subtraction and multiplication of power series:
Theorem 6.2 If
∞
F (t) = Fk (t − t0 )k , −r < t − t0 < r,
k=0
∞
G (t) = Gk (t − t0 )k , −r < t − t0 < r
k=0
and
∞ k
F (t) G (t) = Fi Gk−i (t − t0 )k for |t − t0 | < r.
k=0 i=0
Theorem 6.3 If
∞
F (t) = Fk (t − t0 )k ≡ 0, −r < t − t0 < r,
k=0
Remark:
F (0) = 0 does not imply F (0) = 0! For example
Definition 6.2 The n × m matrix A (t) = [aij (t)] is said to be analytic at t = t0 if every entry
aij (t) , i = 1, . . . , n, j = 1, . . . , m
is analytic at t = t0 .
In this definition it may happen that the intervals of convergence of the expansions for the separate entries
differ from one another. The interval of convergence of A (t) is then taken as that interval in which all the
entries converge.
The preceding theorems may now be reformulated for analytic A (t) by simply applying the relevant con-
ditions to each entry of A (t) and then making the conclusion in a similar way.
The notation ∞
A (t) = Ak (t − t0 )k ,
k=0
with
1 (k)
Ak = A (t0 ) , |t − t0 | < r,
k!
is, therefore, just an abbreviated notation for n × m expansions – one for each entry of A (t) .
Example 6.1
Determine Ak such that
cos t arctan t ∞
1 = Ak tk ≡ A (t) .
et k=0
1−t
Solution:
Note that t0 = 0. Since cos t is an even function and arctan t is an odd function, we rewrite the series
expansion
∞
Ak tk
k=0
as the sum of its even and odd expansions as follows:
∞ ∞ ∞
Ak tk = A2i t2i + A2i+1 t2i+1 .
k=0 i=0 i=0
81 MAT306—X/1
Then
(−1)2i
0
(2i)!
A2i =
1
1
(2i)!
and
(−1)2i
0
A2i+1 = (2i + 1)!
1
1
(2i + 1)!
for i = 1, 2, . . . by using the expansions for cos t, et and arctan t given on page 96 of the textbook. The
common interval of convergence is −1 < t < 1, so that A (t) is analytic at t0 = 0.
Exercise 6.1
Ẋ = A (t) X,
X (t0 ) = X0 . (6.4)
If we want to find a solution for (6.4) in the form of the expansion (6.5), we have to determine formulas
for Xk . Since A (t) and X (t0 ) = X0 are known, we therefore determine the relation between the constant
vectors Xk , and A (t) and X0 .
Since
X(k) (t0 )
Xk = , k = 0, 1, 2, . . . (6.6)
k!
we must differentiate (6.5) (k − 1) times, and determine the value of the derivative at t = t0 in each case.
In order to do this, we use Leibniz’s Rule, according to which
k
dk k
[A (t) X (t)] = A(i) (t) X(k−i) (t) (6.7)
dtk i
i=0
with
k k!
= .
i i! (k − i)!
82
X(k+1) (t0 ) 1 dk
Xk+1 = = Ẋ (t)
(k + 1)! (k + 1)! dtk t=t0
1 dk
= (A (t) X (t))
(k + 1)! dtk t=t0
k
1 k
= A(i) (t0 ) X(k−i) (t0 )
(k + 1)! i
i=0
from (6.7).
Since
∞
A(k) (t0 )
A (t) = (t − t0 )k
k!
k=0
A(k) (t0 )
Ak = ,
k!
i.e.
k
1
Xk+1 = Ai Xk−i (6.8)
k+1
i=0
Since a unique solution of (6.4), analytic at t0 , exists, as is proved in Theorem 7.1, we have the following
theorem:
Theorem 6.4 Suppose that A (t) is analytic at t0 with a power series expansion
∞
A (t) = Ak (t − t0 )k , |t − t0 | < r.
k=0
k
1
Xk+1 = Ai Xk−i ,
k+1
i=0
83 MAT306—X/1
i.e.
X1 = A0 X0
1
X2 = {A0 X1 + A1 X0 }
2 (6.9)
..
.
1
Xk+1 = {A0 Xk + A1 Xk−1 + . . . + Ak X0 }
k+1
Example 6.2
Determine a series solution of
0 1
Ẋ = X ≡ AX.
0 2
Compare your answer with the solution obtained by the eigenvalue—eigenvector method.
Solution:
Since all entries are constants, the matrix A is analytic at t = 0. The sum
∞
Ak (t − t0 )k
k=0
1 0 1
Xn = Xn−1
n 0 2
2
1 0 1
= Xn−2
n (n − 1) 0 2
= ...
n
1 0 1
= X0 .
n! 0 2
we have
1 0 2n−1 a
Xn =
n! 0 2n b
1 2n−1 b
=
n! 2n b
b
n
2 2
= , n ≥ 1.
n!
b
Therefore
∞b
a 2 2n tn
X (t) = +
b n!
n=1 b
b ∞
a (2t)n
= + 2
b n!
b n=1
b
a
= + 2 e2t − 1
b
b
b b
a −
= 2
+ 2 e2t .
0 b
∞ (2t)n
(Recall that e2t − 1 = .)
n=1 n!
b b
By putting a − = c1 and = c2 , we obtain
2 2
1 1
X (t) = c1 + c2 e2t .
0 2
This solution is the same as that obtained by the method of eigenvalues and eigenvectors.
Example 6.3
Find a series solution for the second order initial value problem
ẍ + x = 0, x (0) = 1, ẋ (0) = 0.
Solution:
Let ẋ = y so that the second order initial value problem becomes a system of two first order equations
ẋ = y
ẏ = −x
x
By putting X = the problem can be rewritten as
y
0 1 1
Ẋ = X, X (0) = .
−1 0 0
Since all entries are constants, the matrix A is analytic at t = 0. Thus the sum
∞
Ak (t − t0 )k
k=0
1 0 1
Xn = Xn−1
n −1 0
n
1 0 1
= X0 . (1)
n! −1 0
Now
∞
X (t) = Xk tk
k=0
so that, from (1),
∞ k
1 0 1
X (t) = X0 tk
k! −1 0
k=0
∞ 2k ∞ 2k+1
1 0 1 2k 1 0 1
= X0 t + X0 t2k+1 .
(2k)! −1 0 (2k + 1)! −1 0
k=0 k=0
0 1
But for A = , we have (check this for a few values of k)
−1 0
2k k (−1)k 0
A = (−1) I =
0 (−1)k
0 (−1)k
A2k+1 = (−1)k A =
(−1)k+1 0
with I the identity matrix. This gives
∞ ∞
1 (−1)k 0 1 1 0 (−1)k 1
X (t) = t2k + t2k+1
k=0
(2k)! 0 (−1)k 0 k=0
(2k + 1)! (−1)k+1 0 0
∞ ∞
(−1)k 1 (−1)k+1 0
= t2k + t2k+1 . (2)
(2k)! 0 (2k + 1)! 1
k=0 k=0
86
Remark:
If we used the eigenvalue—eigenvector method to solve this problem, we would have found the exact solution
This result can easily be obtained from equation (2) by making use of standard series expansions of sin t
and cos t (see page 196 of the textbook); we find
cos t 0 cos t
X (t) = + = .
0 sin t − sin t
Exercise 6.2
1. Study Exercise 6.3.2 and 6.3.3, pp. 204—206 of the prescribed book.
3. Show that
∞
(t − t0 )k k
X (t) = A X0
k!
k=0
is the solution of
Ẋ = AX, X (t0 ) = X0 .
Remark:
We have restricted ourselves to series solutions about “ordinary” points, i.e. points where A (t) is analytic.
Series solutions about certain types of singular points are also possible, but for the purpose of this course,
we shall not deal with them. This subject is treated in Section 6.6 of the prescribed book, whilst the
equation of Legendre, to which we have referred earlier in this chapter, is treated in Section 6.4.
Our object is this section is to generalize the exponential function eta , where a is a constant. We wish to at-
tach meaning to the symbol etA when A is a constant matrix, say, an n×n matrix. That this has significance
in the process of a study of the system Ẋ = AX, and the initial value problem Ẋ = AX, X (t0 ) = X0 , can
be “suspected” from the fact that eta y is the (unique) solution of ẋ = ax, x (0) = y.
87 MAT306—X/1
In view of our previous work, it is possible to define in either of the following two ways:
∞
tk k
X (t) = A X0
k!
k=0
∞
tk k
etA U = A U (6.10)
k!
k=0
∞
tk k
etA X0 = A X0 = X (t)
k!
k=0
(2) By recalling that the solution X (t) of Ẋ = AX, X (t0 ) = X0 , may also be expressed in terms of the
normalised fundamental matrix Φ (t) of A at t = 0, viz as X (t) = Φ (t) X0 , we can also define
By using the latter definition, and the properties of a normalised fundamental matrices, it is easy to show
that etA has all the properties of the exponential function.
Exercise 6.3
Show that
etA = I for t = 0.
88
Remark:
The generalization of the exponential function has been extended to the case where A, which may be
dependent on time, is a bounded or unbounded operator with domain and range in a Banach space. A
system {Tt : t > 0} = {etA : t > 0} of operators is then known as a semigroup of bounded operators with
infinitesimal generator A. The concept of semigroups of bounded operators2 is of cardinal importance in
the study of abstract differential equations and has wide application to the field of the qualitative theory
of partial differential equations. A good knowledge of ordinary differential equations, as well as a course in
Functional Analysis and Distribution Theory will make this exciting field accessible to you.
2
The concept of semigroup was named in 1904 and after 1930 the theory of semigroups developed rapidly. One of the most
authentic books on this subject is Functional Analysis and Semigroups, Amer. Math. Soc. Coll. Publ. Vol. 31, Providence
R.I., 1957, by E. Hille and R.S. Phillips. A recent addition to the literature is A Concise Guide to Semigroups and Evolution
Equations by Aldo Belleni—Morante, World Scientific, Singapore, 1994.
Chapter 7
Nonlinear Systems
Existence and Uniqueness Theorem for
Linear Systems
• determine under which conditions the initial value problem Ẋ = A (t) X, X (t0 ) = X0 has a unique
solution.
89
90
If for some or other i, xi cannot be expressed as a linear combination of the components of X (the coefficients
of these components may be either functions of t or else constants) we have a nonlinear equation
Systems containing nonlinear equations, i.e. nonlinear systems, are written in vector form as
Ẋ = F (t, X) . (7.2)
is linear iff F (t, X) = A (t) X + G (t) , for some other matrix A (t) and a vector function G (t) .
An autonomous system is, therefore, of the form Ẋ = F (X) . Autonomous systems have certain properties
which are not generally valid.
Example 7.1
ẋ1 = x1 x2
ẋ2 = tx1 + x2
is a non—autonomous system.
Example 7.2
Consider the non—autonomous equation
ẋ = t2 x − et .
91 MAT306—X/1
Put x1 = t and x2 = x. Then ẋ1 = 1 and ẋ2 = ẋ = t2 x − et , i.e. ẋ2 = x21 x2 − ex1 .
This yields the system
ẋ1 1
Ẋ = = ,
ẋ2 x1 x2 − ex1
2
Put
t 1
x1 f1
Y =
.. , G =
.. .
. .
xn fn
Then
1 1
ẋ1 f1
Ẏ =
.. =
.. = G (Y)
. .
ẋn fn
Exercise 7.1
It is seldom possible to solve nonlinear equations explicitly1 . Certain principles, applicable to linear systems
are also not even valid in the case of nonlinear equations, as for instance the principle of uniqueness of
solutions and the principle of superposition of solutions. The following counterexamples illustrate the truth
of these statements.
Example 7.3
3 1
The initial value problem ẏ = y 3 , y (0) = 0 has an infinite number of solutions. Check it yourself. (See
2
p. 231 of the prescribed book.)
1
One can, however, obtain important information on certain characteristics of the solution of a nonlinear differential equation
without actually solving the equation. One method is the method of phase portraits in the phase space. Nonlinear first—order
autonomous systems of differential equations also are studied by investigating the points of equilibrium of the so—called linearized
system. This subject is dealt with in Chapter 8.
92
Example 7.4
t
The principle of superposition of solutions is not valid for the equation ẏ = y −1 .
2
1
t2 2
Differentiation confirms that y (t) = +c satisfies, for all c, the equation 2yy = t, which is equivalent
2
t
to y = . However,
2y
1 1
t2 2 t2 2
+ c0 + + c1
2 2
is in no case a solution, in view of
t2 t2 t t
2 + c0 + + c1 +
2 2 t2 t2
2 2 + c0 2 2 + c1
t2
+ c0 t2
+ c1
2 2
= t+t+t t2
+ t2
+ c1 + c0
2 2
= t for all c0 , c1 .
It is, indeed, true that the validity of the principle of superposition would imply linearity – thus leading
to a contradiction. We prove this:
Suppose that F (t, X) is a function such that every initial value problem Ẋ = F (t, X) has a solution.
Suppose further that the principle of superposition is valid, i.e. if Xi (t) is a solution of Ẋ = F (t, X) for
i = 1, 2, then Z = aX1 + bX2 is a solution of
and
d
F (t, aX1 + bX2 ) = (aX1 (t) + bX2 (t))
dt
= aF (t, X1 ) + bF (t, X2 ) .
93 MAT306—X/1
Exercise 7.2
Study Example 7.3.3, p. 233 of the prescribed book, from which it appears that the term “general solution”
has to be used very carefully in the case of nonlinear equations.
As far as the solution of nonlinear equations is concerned, we note that the Method of Separation of
Variables can, in the case of simple one—dimensional equations, often be applied successfully.
Example 7.5
Determine the solution of √
1 − t2
ẏ = − √ − 1 ≤ t ≤ 1, y > −5. (7.3)
5+y
Solution:
By separation of variables, equation (7.3) is equivalent to
1 − t2 dt + 5 + y dy = 0.
If a differential equation cannot be solved by means of a standard technique, a numerical solution can still
be obtained by means of techniques of approximation. By a numerical solution of a differential equation
in x, a function of t, is meant a table of values such that for every value of t, a corresponding value of
x (t) is given. Such a table always has a column giving the magnitude of the error involved. The theory
of numerical solutions covers a wide field and is outside the scope of this course. Should the reader be
interested, an easily readable chapter on numerical solutions is found in Ordinary Differential Equations
by M. Tenenbaum and H. Pollard (see references) as well as Differential Equations with Boundary Value
Problems by Dennis G. Zill and M.R. Cullen.
It is interesting to note that a computer can yield a “solution” for a differential equation even when
theoretically no solution exists. An example is the boundary value problem
For this reason it is necessary to pay particular attention to those conditions under which a unique solution
of a system of differential equations exists. We shall confine ourselves to the case of linear systems of
differential equations.
94
Ẋ = A (t) X,
X (t0 ) = X0 ,
Ẋ = F (t, X) ,
X (t0 ) = X0 .
The theorem will, therefore, be proved by applying the Existence and Uniqueness Theorem for the differ-
ential equation
ẋ = f (t, x) , x (t0 ) = x0 .
Ẋ = f (t, X)
X (t0 ) = X0
Remark:
(1) In the above theorem the constant δ is defined as δ = min (T, R/M ) , with the constant M the
upperbound of ||F|| . (See note later on.)
(2) If the range of F is the Banach Space E n , i.e. the n—dimensional Euclidian space with the Euclidian
metric (recall that a Banach space is a normed space which is complete), then ||F (t, X)|| is precisely
|F (t, X)| .
(3) If A (t) is an n × n matrix with real entries, and X an n—tuple of real numbers, then A (t) X is an
n—dimensional vector with components real numbers, so that A (t) X is an element of the Banach
space Rn . As in Chapter 4, Section 5, we define
1
n 2
Since the role of F (t, X) is played by A (t) X, we must have A (t) X continuous for all t in |t − t0 | < T and
every X in ||X − X0 || ≤ R, with R an arbitrary constant.
It is sufficient to require that A (t) should be continuous in |t − t0 | ≤ T, i.e. that fij (t) should be continuous
for i = 1, . . . , n, j = 1, . . . , n. The vector X is, indeed, continuous as a function of X for every “interval”
||X − X0 || ≤ R. Since the product of continuous functions is continuous, it follows that A (t) X is continuous
in |t − t0 | ≤ T and ||X − X0 || ≤ R.
For the proof of Theorem 7.1 we need
Lemma 7.1 If
x1
.
A (t) = [fij (t)] , X = .
. ,
xn
then we have the inequality
||A (t) X|| ≤ ||A (t)|| ||X|| . (7.7)
Proof: By putting
f11 . . . f1n
f1
f21 . . . f2n ..
A (t) =
.. =
.
,
.
fn
fn1 . . . fnn
i.e. fi = (fi1 , fi2 , . . . , fin ) , i = 1, . . . , n, we have
f1 · X
f2 · X
AX =
..
.
fn · X
Since
n
||fi ||2 = fi1
2 2
+ fi2 2
+ . . . + fin = (fij )2 ,
j=1
96
we have
n n n
2 2 2 2
||f1 || + . . . + ||fn || = (f1j (t)) + (f2j (t)) + . . . + (fnj (t))2
j=1 j=1 j=1
n
= (fij (t))2
t,j=1
= ||A (t)||2 .
It follows that
||A (t) X|| ≤ (||A (t)|| ||X||) .
In proving Theorem 7.1, we finally make use of the fact that if a function f is continuous in |t − t0 | ≤ T,
||X − X0 || ≤ R, then a constant M, independent of t and X, exists, such that ||f || ≤ M. Indeed, continuity
and boundedness are equivalent concepts in normed spaces. For the proof, the reader may refer to, amongst
others, “Mathematical Analysis” by Apostol, p. 832 .
Ẋ = A (t) X
X (t0 ) = X0
Proof: From the assumption that A (t) is continuous in |t − t0 | ≤ T we have, according to our previous
remarks, that A (t) X is continuous in |t − t0 | ≤ T and ||X − X0 || ≤ R for every R. We also have, from a
previous remark, that ||A (t)| ≤ M | in |t − t0 | ≤ T.
Therefore
Consequently A (t) X = F (t, X) satisfies the Lipschitz condition (7.5) with K = M. All the conditions of
Theorem 7.1 are now satisfied. We can conclude that the problem
Ẋ = A (t) X
X (t0 ) = X0
has a unique solution in |t − t0 | < δ. This completes the proof of the theorem.
2
Mathematical Analysis, T.M. Apostol, Addison—Wesley Publishing Co., London.
Chapter 8
• autonomous systems;
• critical point;
• periodic solutions;
• classification of a critical point: stable/unstable node, saddle, center, stable/unstable spiral point,
degenerate node;
• Jacobian matrix;
• phase—plane method.
• apply the stability criteria to determine whether a critical point is locally stable or unstable;
97
98
The final chapter of this module is devoted to the qualitative theory of differential equations. In this
branch of theory of differential equations, techniques are developed which will enable us to obtain important
information about the solutions of differential equations without actually solving them. We will, for instance,
be able to decide whether a solution is stable, or what the long time behaviour of the solution is, without
even knowing the form of the solution. This is extremely useful in view of the fact that it is often very
difficult, or even impossible, to determine an exact solution of a differential equation.
As already mentioned in the preface, this chapter is taken from the book Differential Equations with
Boundary—value Problems by Dennis G. Zill and Michael R. Cullen, PWS—Kent Publishing Company,
Boston (1993), where it appears as Chapter 10.