Lecture Notes For ECON660 and ECON460-2022-08
Lecture Notes For ECON660 and ECON460-2022-08
Mathematical Ecnomics
Guoqiang TIAN
Department of Economics
Texas A&M University
College Station, Texas 77843
(gtian@tamu.edu)
2 Economic Models 5
2.1 Ingredients of a Mathematical Model . . . . . . . . . . . . . . 5
2.2 The Real-Number System . . . . . . . . . . . . . . . . . . . . 5
2.3 The Concept of Sets . . . . . . . . . . . . . . . . . . . . . . . . 6
2.4 Relations and Functions . . . . . . . . . . . . . . . . . . . . . 9
2.5 Types of Function . . . . . . . . . . . . . . . . . . . . . . . . . 11
2.6 Functions of Two or More Independent Variables . . . . . . . 12
2.7 Levels of Generality . . . . . . . . . . . . . . . . . . . . . . . . 13
i
ii CONTENTS
1
2 CHAPTER 1. THE NATURE OF MATHEMATICAL ECONOMICS
(3) Mathematics can help obtain the results that cannot be eas-
ily attained through intuition.
issues. The study of economics not only calls for the understanding of var-
ious economic terms, concepts and results from the perspective of mathe-
matics (including geometry), but more importantly, even when those are
given by mathematical language or geometric figure, we need to get to
their economic meaning and the underlying profound economic thoughts
and ideals. Thus we should avoid being confused by the mathematical
formulas or symbols in the study of economics.
Economic Models
Whole numbers such as 1, 2, · · · are called positive numbers; these are the
numbers most frequently used in counting. Their negative counterparts
−1, −2, −3, · · · are called negative integers. The number 0 (zero), on the
other hand, is neither positive nor negative, and it is in that sense unique.
Let us lump all the positive and negative integers and the number zero in-
5
6 CHAPTER 2. ECONOMIC MODELS
A set with finite number of elements is called a finite set. Set I with an
infinite number of elements is an example of an infinite set. Finite sets are
always denumerable (or countable), i.e., their elements can be counted
one by one in the sequence 1, 2, 3, · · · . Infinite sets may, however, be either
denumerable (set I above) or nondenumerable (for example, J = {x|2 <
x < 5}).
then S1 and S2 are said to be equal (S1 = S2 ). Note that the order of
appearance of the elements in a set is immaterial.
this occurs, however, we can be sure that these two sets are equal.
If a set have n elements, a total of 2n subsets can be formed from those
elements. For example, the subsets of {1, 2} are: ∅, {1}, {2} and {1, 2}.
If two sets have no elements in common at all, the two sets are said to
be disjoint.
The union of two sets A and B is a new set containing elements belong
to A, or to B, or to both A and B. The union set is symbolized by A ∪ B
(read: “A union B").
A ∪ (B ∪ C) = (A ∪ B) ∪ C
A ∩ (B ∩ C) = (A ∩ B) ∩ C
A ∪ (B ∩ C) = (A ∪ B) ∩ (A ∪ C)
A ∩ (B ∪ C) = (A ∩ B) ∪ (A ∩ C)
2.4. RELATIONS AND FUNCTIONS 9
Example 2.4.1 To show the age and weight of each student in a class, we
can form ordered pairs (a, w), in which the first element indicates the age
(in years) and the second element indicates the weight (in pounds). Then
(19, 128) and (128, 19) would obviously mean different things.
Suppose, from two given sets, x = {1, 2} and y = {3, 4}, we wish to
form all the possible ordered pairs with the first element taken from set x
and the second element taken from set y. The result will be the set of four
ordered pairs (1,2), (1,4), (2,3), and (2,4). This set is called the Cartesian
product, or direct product, of the sets x and y and is denoted by x × y
(read “x cross y").
Extending this idea, we may also define the Cartesian product of three
sets x, y, and z as follows:
x × y × z = {(a, b, c)|a ∈ x, b ∈ y, c ∈ z}
Example 2.4.2 If the sets x, y, and z each consist of all the real numbers,
the Cartesian product will correspond to the set of all points in a three-
dimension space. This may be denoted by R × R × R, or more simply,
R3 .
10 CHAPTER 2. ECONOMIC MODELS
Example 2.4.3 The set {(x, y)|y = 2x} is a set of ordered pairs including,
for example, (1,2), (0,0), and (-1,-2). It constitutes a relation, and its graph-
ical counterpart is the set of points lying on the straight line y = 2x.
Example 2.4.4 The set {(x, y)|y ≤ x} is a set of ordered pairs including,
for example, (1,0), (0,0), (1,1), and (1,-4). The set corresponds the set of all
points lying on below the straight line y = x.
As a special case, however, a relation may be such that for each x value
there exists only one corresponding y value. The relation in example 2.4.3
is a case in point. In that case, y is said to be a function of x, and this
is denoted by y = f (x), which is read: “y equals f of x." A function is
therefore a set of ordered pairs with the property that any x value uniquely
determines a y value. It should be clear that a function must be a relation,
but a relation may not be a function.
A function is also called a mapping, or transformation; both words
denote the action of associating one thing with another. In the statement
y = f (x), the functional notation f may thus be interpreted to mean a rule
by which the set x is “mapped" (“transformed") into the set y. Thus we
may write
f :x→y
where the arrow indicates mapping, and the letter f symbolically specifies
a rule of mapping.
In the function y = f (x), x is referred to as the argument of the func-
tion, and y is called the value of the function. We shall also alternatively
refer to x as the independent variable and y as the dependent variable.
The set of all permissible values that x can take in a given context is known
as the domain of the function, which may be a subset of the set of all real
numbers. The y value into which an x value is mapped is called the image
of that x value. The set of all images is called the range of the function,
2.5. TYPES OF FUNCTION 11
which is the set of all values that the y variable will take. Thus the domain
pertains to the independent variable x, and the range has to do with the
dependent variable y.
y = a0 + a1 x + a2 x2 + · · · + an xn
A function such as
x−1
y=
x2 + 2x + 4
in which y is expressed as a ratio of two polynomials in the variable x,
is known as a rational function (again, meaning ratio-nal). According to
12 CHAPTER 2. ECONOMIC MODELS
Rules of Exponents:
Rule 1: xm × xn = xm+n
xm
Rule 2: = xm−n (x ̸= 0)
xn
1
Rule 3: x−n = n
x
Rule 4: x0 = 1 (x ̸= 0)
1 √
n
Rule 5: x n = x
Rule 7: xm × y m = (xy)m
z = g(x, y)
a given pair of x and y values will uniquely determine a value of the de-
pendent variable z. Such a function is exemplified by
z = ax + by or z = a0 + a1 x + a2 x2 + b1 y + b2 y 2
y = a1 x1 + a2 x2 + · · · + an xn
y = 7, y = 6x + 4, y = x2 − 3x + 1 (etc.)
14 CHAPTER 2. ECONOMIC MODELS
Not only are these expressed in terms of numerical coefficients, but they al-
so indicate specifically whether each function is constant, linear, or quadrat-
ic. In terms of graphs, each such function will give rise to a well-defined
unique curve. In view of the numerical nature of these functions, the so-
lutions of the model based on them will emerge as numerical values also.
The drawback is that, if we wish to know how our analytical conclusion
will change when a different set of numerical coefficients comes into effec-
t, we must go through the reasoning process afresh each time. Thus, the
result obtained from specific functions have very little generality.
On a more general level of discussion and analysis, there are functions
in the form
y = a, y = bx + a, y = cx2 + bx + a (etc.)
Since parameters are used, each function represents not a single curve but
a whole family of curves. With parametric functions, the outcome of math-
ematical operations will also be in terms of parameters. These results are
more general.
In order to attain an even higher level of generality, we may resort to
the general function statement y = f (x), or z = g(x, y). When expressed
in this form, the functions is not restricted to being either linear, quadrat-
ic, exponential, or trigonometric – all of which are subsumed under the
notation. The analytical result based on such a general formulation will
therefore have the most general applicability.
Chapter 3
Equilibrium Analysis in
Economics
Like any economic term, equilibrium can be defined in various ways. One
definition here is that an equilibrium for a specific model is a situation
where there is no tendency to change. More generally, it means that from
an available set of choices (options), choose the "best” one according to
a certain criterion. It is for this reason that the analysis of equilibrium is
referred to as statics. The fact that an equilibrium implies no tendency to
change may tempt one to conclude that an equilibrium necessarily consti-
tutes a desirable or ideal state of affairs.
This chapter provides two typical examples of equilibrium analysis.
One that is from microeconomics is the equilibrium attained by a mar-
ket under given demand and supply conditions. The other that is from
macroeconomics is the equilibrium of Keynesian national income model
under given conditions of consumption and investment patterns. We will
use these two models as running examples throughout the course.
15
16 CHAPTER 3. EQUILIBRIUM ANALYSIS IN ECONOMICS
Qd = Qs ,
Qd = a − bP (a, b > 0),
Qs = −c + dP (c, d > 0),
Qd , Q s
a
Qd = a - b P Qs = - c + d P
(demand) (supply)
Q= Qd= Qs (P, Q )
O P
P1 P
-c
From Qs = Qd , we have
a − bP = −c + dP
and thus
(b + d)P = a + c.
a+c
P̄ = .
b+d
The partial market model can be nonlinear. For example, suppose the
model is given by
Qd = Qs (equilibrium condition);
Qd = 4 − P 2 ;
Qs = 4P − 1.
4 − P 2 = 4P − 1,
or
P 2 + 4P − 5 = 0,
where the “+" part of the “±" sign yields x̄1 and “−" part yields x̄2 . Thus,
by applying the quadratic formulas to P 2 +4P −5 = 0, we have P̄1 = 1 and
P̄2 = −5, but only the first is economically admissible, as negative prices
are ruled out.
3.4. GENERAL MARKET EQUILIBRIUM 19
Qd , Qs
4
Qs = 4P - 1
3 ( 1, 3 )
Qd = 4 - P!
1
P
-2 -1 0 1 2
-1
where Qdi = Qdi (P1 , P2 , · · · , Pn ) and Qsi = Qsi (P1 , P2 , · · · , Pn ) are the de-
mand and supply functions of commodity i, and (P1 , P2 , · · · , Pn ) are prices
of commodities.
Thus, solving n equations for P = (P1 , P2 , · · · , Pn ):
Ei (P1 , P2 , · · · , Pn ) = 0,
we obtain the n equilibrium prices P̄i – if a solution does indeed exist. And
then the Q̄i may be derived from the demand or supply functions.
By substituting the second and third equations into the first and the
fifth and sixth equations into the fourth, the model is reduced to two e-
quations in two variable:
If we let
ci = ai − bi (i = 0, 1, 2),
γi = αi − βi (i = 0, 1, 2),
c1 P1 + c2 P2 = −c0 ;
γ1 P1 + γ2 P2 = −γ0 ,
For these two values to make sense, certain restrictions should be im-
posed on the model. Firstly, we require the common denominator c1 γ2 −
c2 γ1 ̸= 0. Secondly, to assure positivity, the numerator must have the same
22 CHAPTER 3. EQUILIBRIUM ANALYSIS IN ECONOMICS
Numerical Example
Suppose that the demand and supply functions are numerically as follows:
Qd1 = 10 − 2P1 + P2 ;
Qs1 = −2 + 3P1 ;
Qd2 = 15 + P1 − P2 ;
Qs2 = −1 + 2P2 .
By substitution, we have
5P1 − P2 = 12;
which are two linear equations. The solutions for the equilibrium prices
and quantities are P̄1 = 52/14, P̄2 = 92/14, Q̄1 = 64/7, Q̄2 = 85/7.
Similarly, for the n−commodities market model, when demand and
supply functions are linear in prices, we can have n linear equations. In the
above, we assume that an equal number of equations and unknowns has
a unique solution. However, some very simple examples should convince
us that an equal number of equations and unknowns does not necessarily
guarantee the existence of a unique solution.
For the two linear equations,
x + y = 8,
,
x+y =9
3.5. EQUILIBRIUM IN NATIONAL-INCOME ANALYSIS 23
These two equations are functionally dependent, which means that one
can be derived from the other. Consequently, one equation is redundant
and may be dropped from the system. Any pair (x̄, ȳ) is the solution as
long as (x̄, ȳ) satisfies y = 12 − x.
Now consider the case of more equations than unknowns. In gener-
al, there is no solution. But, when the number of unknowns equals the
number of functionally independent equations, the solution exists and is
unique. The following example shows this fact.
2x + 3y = 58;
y = 18;
x + y = 20.
model,
Y = C + I0 + G0 (equilibrium condition);
C = a + bY (consumption function),
where Y and C stand for the endogenous variables national income and
consumption expenditure, respectively, and I0 and G0 represent the ex-
ogenously determined investment and government expenditures, respec-
tively.
Solving these two linear equations, we obtain the equilibrium national
income and consumption expenditure:
a + I0 + G0
Ȳ = ,
1−b
a + b(I0 + G0 )
C̄ = .
1−b
Chapter 4
From the last chapter we have seen that for the one-commodity partial
market equilibrium model, the solutions for P̄ and Q̄ are relatively sim-
ple, even though a number of parameters are involved. As more and
more commodities are incorporated into the model, such solutions for-
mulas quickly become cumbersome and unwieldy. We need to have new
methods suitable for handing a large system of simultaneous equations.
Such a method is provided in matrix algebra.
25
26 CHAPTER 4. LINEAR MODELS AND MATRIX ALGEBRA
c1 P1 + c2 P2 = −c0 ,
γ1 P1 + γ2 P2 = −γ0 .
Matrix as Arrays
6x1 + 3x2 + x3 = 22
x1 + 4x2 − 2x3 = 12 ,
4x1 − x2 + 5x3 = 10
we can write
6 3 1
A=
1 4 −2
,
4 −1 5
x1
x=
x2 ,
x3
22
d=
12 .
10
more simple as
x′ = [x1 , x2 , · · · , xn ].
With the matrices defined in (3.2), we can express the equations system
(3.1) simply as
Ax = d.
Remark 4.2.1 Two matrices can be added (equal) if and only if they have
the same dimension.
Example 4.2.1
4 9 2 0 6 9
+ = .
2 1 0 7 2 8
Example 4.2.2
a11 a12 a13 b11 b12 b13 a11 + b11 a12 + b12 a13 + b13
+ = .
a21 a22 a23 b21 b22 b23 a21 + b21 a22 + b22 a23 + b23
A − B is defined by
[aij ] − [bij ] = [aij − bij ].
Example 4.2.3
19 3 6 8 13 −5
− = .
2 0 1 3 1 −3
30 CHAPTER 4. LINEAR MODELS AND MATRIX ALGEBRA
Scalar Multiplication:
λA = λ[aij ] = [λaij ],
Example 4.2.4
3 −1 21 −7
7 = .
0 5 0 35
Example 4.2.5
a11 a12 a13 −a11 −a12 −a13
−1 = .
a21 a22 a23 −a21 −a22 −a23
Multiplication of Matrices:
Given two matrices Am×n and Bp×q , the conformability condition for mul-
tiplication AB is that the column dimension of A must be equal to the row
dimension of B, i.e., the matrix product AB will be defined if and only if
n = p. If defined, the product AB will have the dimension m × q.
The product AB is defined by
AB = C
∑n
with cij = ai1 b1j + ai2 b2j + · · · + ain bnj = l=1 ail blj .
Example 4.2.6
a11 a12 b11 b12 a11 b11 + a12 b21 a11 b12 + a12 b22
= .
a21 a22 b21 b22 a21 b11 + a22 b21 a21 b12 + a22 b22
4.2. MATRIX OPERATIONS 31
Example 4.2.7
3 5 −1 0 −3 + 20 35 17 35
= = .
4 6 4 7 −4 + 24 42 20 42
Example 4.2.8
∑
n
u′ v = u1 v1 + u2 v2 + · · · + un vn = ui vi .
i=1
This can be described by using the concept of the inner product of two
vectors u and v.
v · v = u1 v1 + u2 v2 + · · · + un vn = u′ v.
Example 4.2.9 For the linear-equation system (4.1.1), the coefficient matrix
and the variable vector are:
a
11
a12 · · · a1n x
1
a21 a22 · · · a2n x2
A=
and x = ,
··· ··· ··· ··· · · ·
am1 am2 · · · amn xn
Ax = d.
3
Example 4.2.10 Given u = and v ′ = [1, 4, 5], we have
2
3 × 1 3 × 4 3 × 5 3 12 15
uv ′ = = .
2×1 2×4 2×5 2 8 10
If there are only two vectors, linear dependence means one is propor-
tional to the other.
or
3v1 − 2v2 − v3 = 0,
0
where 0 = represents a zero vector.
0
1 1
v1 = v2 + v3 .
2 2
∑
n
ki vi = 0. (4.3.2)
i=1
V k = 0, (4.3.3)
where
v′
1
′
v
2
V =
···
′
vm ,
pendent.
In general, it is hard to check if a set of m−vectors v1 , v2 , · · · , vn is lin-
early dependent by the definition when m is bigger than 2. However, we
can reduce the test by checking if the homogeneous linear-equation sys-
tem (4.3.3) has a zero solution. We will provide a relative simple way to
check this.
Proof: A + B = [aij ] +[bij ] = [aij +bij ] = [bij +aij ] = [bij ]+[aij ] = B +A.
Associative Law:
(A + B) + C = A + (B + C).
Matrix Multiplication
AB ̸= BA.
4.5. IDENTITY MATRICES AND NULL MATRICES 35
Even when AB is defined, BA may not be; but even if both products
are defined, AB = BA may not hold.
1 2 0 −1 12 13
Example 4.4.1 Let A = , B = . Then AB = , but
3 4 6 7 24 25
−3 −4
BA = .
27 40
kA = Ak
if k is a scalar.
Associative Law:
(AB)C = A(BC)
provided A is m × n, B is n × p, and C is p × q.
Distributive Law
1 0 ··· 0
0 1 ··· 0
In = .
··· ··· ··· ···
0 0 ··· 1
Im A = AIn = A.
Fact 2:
Am×n In Bn×p = (AI)B = AB.
Fact 3:
(In )k = In .
we have
5 8
CD = CE = ,
15 24
even though D ̸= E.
Then a question is: Under what condition, does CD = CE imply
D = E? We will show that it is so if C has the inverse that we will discuss
shortly.
(b) Even if A and B ̸= 0, we can still have AB = 0. Again, we will see
this is not true if A or B has the inverse.
2 4 −2 4
Example 4.5.1 A = , B = .
1 2 1 −2
We have AB = 0.
3 1
A′ =
8 0
.
−9 4
38 CHAPTER 4. LINEAR MODELS AND MATRIX ALGEBRA
Properties of Transposes:
a) (A′ )′ = A;
b) (A + B)′ = A′ + B ′ ;
d) (AB)′ = B ′ A′ .
The property d) states that the transpose of a product is the product of the
transposes in reverse order.
For a given square matrix A, while its transpose A′ is always derivable, its
inverse matrix may or may not exist.
B = BI = BAC = IC = C.
So B is the inverse of A.
Ax = d.
41
42CHAPTER 5. LINEAR MODELS AND MATRIX ALGEBRA (CONTINUED)
Ak = 0
Rank of a Matrix
a11 a12
|A| = = a11 a22 − a12 a21 .
a21 a22
10 4 3 5
Example 5.2.1 Given A = and B = , then
8 5 0 −1
10 4
|A| = = 50 − 32 = 18;
8 5
3 5
|B| = = −3 − 5 × 0 = −3.
0 −1
2
6
Example 5.2.2 A = . Then its determinant is
8 24
2 6
|A| = = 2 × 24 − 6 × 8 = 48 − 48 = 0.
8 24
This example shows that the determinant is equal to zero if and only if
its rows are linearly dependent. As will be seen, the value of a determinant
|A| can serve as a criterion for testing the linear independence of the rows
(hence nonsingularity) of matrix A but also as an input in the calculation
of the inverse A−1 , if it exists.
5.2. TEST OF NONSINGULARITY BY USE OF DETERMINANT 45
= a11 a22 a33 − a11 a23 a32 − a12 a21 a33 + a12 a23 a31 + a13 a21 a32 − a13 a22 a31 .
Figure 5.1: The graphic illustration for calculating the third-order determi-
nant.
46CHAPTER 5. LINEAR MODELS AND MATRIX ALGEBRA (CONTINUED)
Example 5.2.3
2 1 3
4 5 6 =2×5×9+1×6×7+4×8×3−3×5×7−1×4×9−6×8×2
7 8 9
= 90 + 42 + 96 − 105 − 36 − 96 = −9.
Example 5.2.4
0 1 2
3 4 5 =0×4×8+1×5×6+3×7×2−2×4×6−1×3×8−5×7×0
6 7 8
= 0 + 30 + 42 − 48 − 24 − 0 = 0.
Example 5.2.5
−1 2 1
0 3 2 = −1 × 3 × 2 + 2 × 2 × 1 + 0 × 0 × 1 − 1 × 3 × 1 − 2 × 0 × 2 − 2 × 0 × (−1)
1 0 2
= −6 + 4 + 0 − 3 − 0 − 0 = −5.
For instance, for a third determinant, the minors of a11 , a12 and a13 are
Thus, if the sum of the two subscripts i and j in Mij is even, then
|Cij | = |Mij |. If it is odd, then |Cij | = −|Mij |.
∑
n
|A| = aij |Cij | [expansion by the ith row]
j=1
∑n
= aij |Cij | [expansion by the jth column].
i=1
Even though one can expand |A| by any row or any column, as the
numerical calculation is concerned, a row or column with largest number
of 0’s or 1’s is always preferable for this purpose, because a 0 times its
cofactor is simply 0.
5 6 1
Example 5.2.6 For the |A| = 2 3 0 , the easiest way to expand the
7 −3 0
determinant is by the third column, which consists of the elements 1, 0,
and 0. Thus,
2 3
|A| = 1 × (−1)1+3 = −6 − 21 = −27.
7 −3
Example 5.2.7
0 0 0 1
0 0 2
0 0 2 0
= 1 × (−1)1+4 0 3 0 = −1 × (−24) = 24.
0 3 0 0
4 0 0
4 0 0 0
we have
a c
|A′ | = = ad − bc = |A|.
b d
Property II. The interchange of any two rows (or any two columns)
will alter the sign, but not the numerical value of the determinant.
50CHAPTER 5. LINEAR MODELS AND MATRIX ALGEBRA (CONTINUED)
a b
Example 5.3.2 = ad − bc, but the interchange of the two rows yields
c d
c d
= bc − ad = −(ad − bc).
a b
Property III. The multiplication of any one row (or one column) by a
scalar k will change the value of the determinant k-fold, i.e., for |A|,
Example 5.3.3
a b a b
= a(d + kb) − b(c + ka) = ad − bc = .
c + ka d + kb c d
Example 5.3.4
a b b b a + 3b b b b a + 3b b b b
b a b b a + 3b a b b 0 a−b 0 0
= = = (a+3b)(a−b)3 .
b b a b a + 3b b a b 0 0 a−b 0
b b b a a + 3b b b a 0 0 0 a−b
1 2 3 4 10 2 3 4 10 2 3 4
2 3 4 1 10 3 4 1 0 1 1 −3
= =
3 4 1 2 10 4 1 2 0 2 −2 −2
4 1 2 3 10 1 2 3 0 −1 −1 −1
1 1 −3 1 1 −3
1+1
= (−1) 10 2 −2 −2 = 10 0 −4 4 = 160.
−1 −1 −1 0 0 −4
52CHAPTER 5. LINEAR MODELS AND MATRIX ALGEBRA (CONTINUED)
Example 5.3.6
−2 5 −1 3 0 −13 25 17
−13 25 17
1 −9 13 7 1 −9 13 7
= = (−1)1+2 26 −34 −26
3 −1 5 −5 0 26 −34 −26
26 −33 −24
2 8 −7 −10 0 26 −33 −24
1 25 17 1 25 17
= 13 −2 −34 −26 = 13 0 16 8
−2 −33 −24 0 17 10
16 8
= (−1)1+1 13 = 312.
17 10
Example 5.3.7
ka kb
= kab − kab = 0.
a b
Property VI. If A and B are both square matrices, then |AB| = |A||B|.
5.3. BASIC PROPERTIES OF DETERMINANTS 53
The above basic properties are useful in several ways. For one thing,
they can be of great help in simplifying the task of evaluating determi-
nants. By adding (or subtracting) multipliers of one row (or column) from
another, the elements of the determinant may be reduced to much simpler
numbers. If we can indeed apply these properties to transform some row
or column into a form containing mostly 0’s or 1’s, Laplace expansion of
the determinant will become a much more manageable task.
Property VII. |A−1 | = 1
|A|
. As a consequence, if A−1 exists, we must
have |A| =
̸ 0. The converse is also true.
Recipe - How to Calculate the Determinant:
4. If two rows (or columns) are proportional, i.e., they are lin-
early dependent, then the determinant will vanish.
Using these rules, we can simplify the matrix (e.g. obtain as many zero
elements as possible) and then apply Laplace expansion.
54CHAPTER 5. LINEAR MODELS AND MATRIX ALGEBRA (CONTINUED)
Example 5.3.8
1 3 2
γ
2 6 4
=2
−5 7 1
1 3 2
1 3
since 2 6 4 = 0 and ̸= 0.
2 6
−5 7 1
One can also see this because the first two rows are linearly dependent,
but the last two are independent, therefore the maximum number of lin-
early independent rows is equal to 2.
1) The column rank and the row rank of a matrix are equal.
∑
n
|A| = aij |Cij | [expansion by the ith row]
j=1
∑n
= aij |Cij | [expansion by the jth column]
i=1
Now what happens if we replace one row (or column) by another row
(or column), i.e., aij by ai′ j for i ̸= i′ or by aij ′ for j ̸= j ′ . Then we have the
following important property of determinants.
Property VIII. The expansion of a determinant by alien cofactors (the
cofactors of a “wrong" row or column) always yields a value of zero. That
is, we have
∑
n
ai′ j |Cij | = 0 (i ̸= i′ ) [expansion by the i′ th row and use of cofactors of ith row]
j=1
∑
n
aij ′ |Cij | = 0 (j ̸= j ′ ) [expansion by the j ′ th column and use of cofactors of jth column]
j=1
The reason for this outcome lies in the fact that the above formula can
be considered as the result of the regular expansion of a matrix that has
two identical rows or columns.
∑
3
0 = |A∗ | = a11 |C21 | + a12 |C22 | + a13 |C23 | = a1j |C2j |.
j=1
Matrix Inversion
since each element of A has a cofactor |Cij |, we can form a matrix of cofac-
tors by replacing each element aij with its cofactor |Cij |. Such a cofactor
matrix C = [|Cij |] is also n × n. For our present purpose, however, the
transpose of C is of more interest. This transpose C ′ is commonly referred
to as the adjoint of A and is denoted by adj A. That is,
|C | |C21 | · · · |Cn1 |
11
|C12 | |C22 | · · · |Cn2 |
C ′ ≡ adj A ≡ .
··· ··· ··· ···
|C1n | |C2n | · · · |Cnn |
By utilizing the formula for the Laplace expansion and Property VI, we
58CHAPTER 5. LINEAR MODELS AND MATRIX ALGEBRA (CONTINUED)
have
a
11
a12 · · · a1n |C11 | |C21 | · · · |Cn1 |
a21 a22 · · · a2n
′ |C12 | |C22 | · · · |Cn2 |
AC =
· · · ··· ···
··· ··· ··· ···
an1 an2 · · · ann |C1n | |C2n | · · · |Cnn |
∑ ∑n ∑n
n
j=1 a1j |C1j | j=1 a1j |C2j | · · · j=1 a1j |Cnj |
∑n ∑n ∑n
a2j |Cnj |
j=1 a2j |C1j | j=1 a2j |C2j | · · · j=1
=
··· ··· ··· ···
∑n ∑n ∑n
j=1 anj |C1j | j=1 anj |C2j | · · · j=1 anj |Cnj |
|A| 0 ··· 0
0 |A| · · · 0
=
··· ··· ··· ···
0 0 · · · |A|
= |A|In .
−1 C′ adj A
A = = .
|A| |A|
Remark 5.4.1 In summary, the general procedures for finding the inverse
of a square A are:
(2) find the cofactors of all elements of A and form C = [|Cij |];
|C11 | |C12 | d −c
C= = .
|C21 | |C22 | −b a
adj A
A−1 =
|A|
1 d −b
= ,
ad − cb −c a
3 2
Example 5.4.2 A = .
1 0
4 1 −1
Example 5.4.3 Find the inverse of B =
0 3 2
.
3 0 7
60CHAPTER 5. LINEAR MODELS AND MATRIX ALGEBRA (CONTINUED)
Then
21 −7 5
adj B = C ′ =
6 31 −8
.
−9 3 12
Therefore, we have
21 −7 5
−1 1
6
B = 31 −8
.
99
−9 3 12
5.5. CRAMER’S RULE 61
2 4 5
Example 5.4.4 A =
0 3 0.
1 0 1
We have |A| = −9 and
3 −4 −15
−1 1
A =−
0 −3 0
.
9
−3 4 6
1
x̄ = A−1 d = (adj A)d
|A|
1 ∑ n
x̄j = di |Cij |
|A| i=1
11
a a12 · · · d1 · · · a1n
1 a21 a22 · · · d2 · · · a2n
=
|A|
· · · ··· ··· ···
an1 an2 · · · dn · · · ann
1
= |Aj |,
|A|
where |Aj | is obtained by replacing the jth column of |A| with constant
terms d1 , · · · , dn . This result is the statement of Cramer’s rule.
we have
−42 −28
x1 = = 3, x2 = = 2.
−14 −14
Example 5.5.2
We then have
5 3
|A| = = −28;
6 −2
30 3
|A1 | = = −84;
8 −2
5 30
|A2 | = = −140.
6 8
Example 5.5.3
x1 + x2 + x3 = 0
12x1 + 2x2 − 3x3 = 5
3x1 + 4x2 + x3 = −4.
We have
|A| = 35, |A3 | = 35, and thus x3 = 1.
64CHAPTER 5. LINEAR MODELS AND MATRIX ALGEBRA (CONTINUED)
Example 5.5.4
7x1 − x2 − x3 = 0
10x1 − 2x2 + x3 = 8
6x1 + 3x2 − 2x3 = 7.
We have
Thus
|A1 |
x̄1 = = 1,
|A|
|A2 |
x̄2 = = 3,
|A|
|A3 |
x̄3 = = 4.
|A|
Example 5.5.5
a11 x1 + a12 x2 = 0,
a21 x2 + a22 x2 = 0.
If |A| = 0, then its rows are linearly dependent. As a result, one of two
equations is redundant. By deleting, say, the second equation, we end up
with one equation with two variables. The solutions are
a12
x̄1 = − x2 if a11 ̸= 0
a11
Proposition 5.5.1 A necessary and sufficient condition for the existence of solu-
tion for a linear-equation system Am×n x = d with n variables and m equations
is that the rank of A and the rank of the added matrix [A; d] are the same, i.e.,
d
d ̸= 0 d=0
|A|
The solution is unique The solution is unique
|A| ̸= 0
and x̄ ̸= 0 and x = 0
Equations An infinite number of There is an infinite
|A| = 0 dependent solutions and x̄ ̸= 0 number of solutions
Equations
No solution exists [Not Applicable]
inconsistent
Market Model:
c1 P1 + c2 P2 = −c0 ,
γ1 P1 + γ2 P2 = −γ0 .
Thus
c1 c2
|A| = = c1 γ2 − c2 γ1 ,
γ1 γ2
−c0 c2
|A1 | = = c2 γ0 − c0 γ2 ,
−γ0 γ2
c1 −c0
|A2 | = = c0 γ1 − c1 γ0 .
γ1 −γ0
5.6. APPLICATION TO MARKET AND NATIONAL-INCOME MODELS67
|A1 | c2 γ0 − c0 γ2
P̄1 = =
|A| c1 γ2 − c2 γ1
and
|A2 | c0 γ1 − c1 γ0
P̄2 = = .
|A| c1 γ2 − c2 γ1
Consider a market for three goods. The demand and supply for each
good are given by:
D1 = 5 − 2P1 + P2 + P3 ,
S1 = −4 + 3P1 + 2P2 .
D2 = 6 + 2P1 − 3P2 + P3 ,
S2 = 3 + 2P2 .
D3 = 20 + P1 + 2P2 − 4P3 ,
S3 = 3 + P2 + 3P3 ,
|A1 | 356
P̄1 = = = 2,
|A| 178
|A2 | 356
P̄2 = = = 2,
|A| 178
|A3 | 534
P̄3 = = = 3.
|A| 178
National-Income Model
Y = C + I0 + G0 ,
Y − C = I0 + G0 ,
−bY + C = a.
While we can solve Ȳ and C̄ by Cramer’s rule, here we solve this model
by inverting the
coefficient
matrix.
1 −1 1
1 1
Since A = , then A−1 = 1−b .
−b 1 b 1
Hence
[ ]
1 1 1 I0 + G0
Ȳ C̄ =
1−b b 1 a
1 I0 + G0 + a
= .
1 − b b(I0 + G0 ) + a
5.7. QUADRATIC FORMS 69
This is, it is a polynomial having only second-order terms (either the square
of a variable or the product of two variables).
= u′ Du,
where
d
11
d12 · · · d1n
d21 d22 · · · d2n
D=
,
· · · · · · · · · · · ·
dn1 dn2 · · · dnn
Example 5.7.2
q = u21 + u22
q = u21 − u22
5.7. QUADRATIC FORMS 71
is indefinite.
We state without proof that for the quadratic form q(u) = u′ Du, the
necessary and sufficient condition for positive definiteness is the principal
minors of |D|, namely,
|D1 | = d11 > 0,
d11 d12
|D2 | = > 0,
d21 d22
···
5 1.5
|D2 | = = 10 − 2.25 = 7.75 > 0,
1.5 2
72CHAPTER 5. LINEAR MODELS AND MATRIX ALGEBRA (CONTINUED)
so q is positive definite.
|D1 | = 1 > 0,
1 −1
|D2 | = = 6 − 1 = 5 > 0,
−1 6
and
1 −1 0
|D3 | = −1 6 −2 = 11 > 0.
0 −2 3
form is
−3 −1 0
D=
−1 −3 0
.
0 0 −5
|D1 | = −3 < 0,
|D2 | = 8 > 0,
Dx = λx.
(D − λI)x = 0.
3−λ −1 0
|D − λI| = −1 3−λ 0
0 0 5−λ
= (3 − λ)(3 − λ)(5 − λ) − (5 − λ)
= (5 − λ)(λ − 2)(λ − 4) = 0,
Properties of Eigenvalues:
P −1 AP = D.
Theorem 5.8.1 (The Spectral Theorem for Symmetric Matrices) Suppose that
A is a symmetric matrix of order n and λ1 , · · · , λn are its eigenvalues. Then there
exists an orthogonal matrix U such that
λ1 0
..
U −1 AU =
. .
0 λn
1−λ 2
= λ(λ − 5) = 0,
2 4−λ
i.e., λ1 = 0 and λ2 = 5.
For λ1 = 0, we solve
1 − 0 2 x1 0
= .
2 4 − 0 x2 0
Let us normalize the eigenvectors, i.e. let us pick constants Ci such that
vi′ vi = 1. We get
( ) ( )
2 −1 1 2
v1 = √ , √ , v2 = √ ,√ .
5 5 5 5
1) tr(A) = λ1 + · · · + λn ;
5) tr(A′ ) = tr(A);
∑n ∑n
6) tr(A′ A) = i=1 j=1 a2ij .
1. (u + v) + w = u + (v + w);
2. u + v = v + u;
3. 0 + u = u;
4. u + (−u) = 0;
5. a(u + v) = au + av;
6. (a + b)u = au + bu;
7. a(bu) = (ab)u;
8. 1u = u.
The objects of a vector space V are called the vectors, the operations +
and · are called the vector addition and scalar multiplication, respectively.
The element 0 ∈ V is the zero vector and −v is the additive inverse of V .
78CHAPTER 5. LINEAR MODELS AND MATRIX ALGEBRA (CONTINUED)
u + v = (u1 + v1 , · · · , un + vn )′ ,
au = (au1 , · · · , aun )′ .
1. u · v = v · u,
2. u · (v + w) = u · v + u · w,
4. u · u ≥ 0 and u · u = 0 iff u = 0.
1. ||au|| = |a|||u||;
3. |u · v| ≤ ||u|| × ||v||.
5.9. VECTOR SPACES 79
The nonzero vectors u and v are parallel if there exists a ∈ R such that
u = av.
The vectors u and v are orthogonal or perpendicular if their scalar
product is zero, that is, if u · v = 0.
uv
The angle between vectors u and v is arccos( ||u||||v|| )
A nonempty subset S of a vector space V is a subspace of V if for any
u, v ∈ S and a ∈ R
u + v ∈ S and au ∈ S.
S = {a1 u1 + a2 u2 + · · · , ak uk |a1 , · · · , ak ∈ R}
Example 5.9.5 Let u1 = (2, −1, 1)′ , u2 = (3, 4, 0)′ . Then the subspace of R3
generated by u1 and u2 is
Example 5.9.7 Consider the vector space Rn with the standard basis En .
For any u = (u1 , · · · , un )′ , we can represent u as u = u1 e1 + · · · + un en ;
therefore, (u1 , · · · , un )′ is the vector of coordinates of u with respect to En .
Example 5.9.8 Consider the vector space R2 . Let us find the coordinate
vector of (−1, 2)′ with respect to the basis B = (1, 1)′ , (2, −3)′ (i.e., find
(−1, 2)′B ). We have to solve for a, b such that (−1, 2)′ = a(1, 1)′ + b(2, −3)′ .
Solving the system a + 2b = −1 and a − 3b = 2, we find a = 1
5
and b = − 35 .
Thus, (−1, 2)′B = ( 15 , − 53 )′ .
5.9. VECTOR SPACES 81
Example 5.9.9 The dimension of the vector space Rn with the standard
basis En is n.
Properties:
y = AT x for any u ∈ U.
T ((1, 1, 1)′ ) = (−1, 2), T ((1, 1, 0)′ ) = (−1, 1), T ((1, 0, 0)′ ) = (1, 1).
Example 5.9.13 Given the following bases for R2 : B = {(1, 1)′ , (1, 0)′ } and
C = {(0, 1)′ , (1, 1)′ }, find the change-of-basis matrix D relative to B, C.
The columns of D are the coordinate vectors of (1, 1)′ , (1, 0)′ relative to C.
5.9. VECTOR SPACES 83
85
86CHAPTER 6. COMPARATIVE STATICS AND THE CONCEPT OF DERIVATIVE
We use the symbol ∆ to denote the change from one point, say x0 , to
another point, say x1 . Thus ∆x = x1 − x0 . When x changes from an initial
value x0 to a new value x0 +∆x, the value of the function y = f (x) changes
from f (x0 ) to f (x0 + ∆x). The change in y per unit of change in x can be
represented by the difference quotient.
and thus,
The Derivative
dy ∆y
≡ y ′ ≡ f ′ (x) ≡ lim .
dx ∆x→0 ∆x
Remark 6.2.1 Several points should be noted about the derivative: (1) a
derivative is a function. Whereas the difference quotient is a function of x0
and ∆x, the derivative is a function of x0 only; and (2) since the derivative
is merely a limit of the difference quotient, it must also be of necessity a
measure of some rate of change. Since ∆x → 0, the rate measured by the
derivative is in the nature of an instantaneous rate of change.
∆y
= 6x + 3∆x,
∆x
dy
we have dx
= 6x.
88CHAPTER 6. COMPARATIVE STATICS AND THE CONCEPT OF DERIVATIVE
Figure 6.1: Graphical illustrations of the slope of the total cost curve and
the marginal cost.
6.4. THE CONCEPT OF LIMIT 89
Graphical Illustrations
Evaluation of a limit
q = (1 − v 2 )/(1 − v).
1 − v2 (1 − v)(1 + v)
q= = = 1 + v (v ̸= 1)
1−v 1−v
2v+5
Example 6.4.3 Find limv→∞ v+1
.
2v+5 2(v+1)+3 3 3 2v+5
Since v+1
= v+1
= 2 + v+1 and limv→∞ v+1
= 0, so limv→∞ v+1
= 2.
Transitivity:
a > b and b > c implies a > c;
a ≥ b and b ≥ c implies a ≥ c.
Addition and Subtraction:
a > b =⇒ a ± k > b ± k;
a ≥ b =⇒ a ± k ≥ b ± k.
Multiplication and Division:
a > b =⇒ ka > kb (k > 0);
a > b =⇒ ka < kb (k < 0).
6.5. INEQUALITY AND ABSOLUTE VALUES 93
Squaring:
a > b with b ≥ 0 =⇒ a2 > b2 .
For any real number n, the absolute value of n is defined and denoted
by
n if n > 0,
|n| = −n if n < 0,
0 if n = 0.
Thus we can write |x| < n as an equivalent way −n < x < n (n > 0).
Also |x| ≤ n if and only if −n ≤ x ≤ n (n > 0).
The following properties characterize absolute values:
1) |m| + |n| ≥ |m + n|;
2) |m| · |n| = |m · n|;
|m| m
3) |n|
= n
.
Solution of an Inequality
Example 6.5.1 Find the solution of the inequality 3x−3 > x+1. By adding
(3 − x) to both sides, we have
3x − 3 + 3 − x > x + 1 + 3 − x.
1+v
Example 6.6.3 Find limv→0 2+v
.
1+v
Since limv→0 (1 + v) = 1 and limv→0 (2 + v) = 2, so limv→0 2+v
= 12 .
Remark 6.6.1 Note that L1 and L2 represent finite numbers; otherwise the-
orems do not apply.
lim a0 + a1 v + a2 v 2 + · · · + an v n = a0 + a1 N + a2 N 2 + · · · + an N n .
v→N
6.7. CONTINUITY AND DIFFERENTIABILITY OF A FUNCTION 95
Thus the term continuous involves no less than three requirements: (1)
the point N must be in the domain of the function; (2) limv→N g(v) exists;
and (3) limv→N g(v) = g(N ).
4v 2
Example 6.7.1 q = v 2 +1
.
Then
4v 2 limv→N 4v 2 4N 2
lim = = .
v→N v2 + 1 limv→N (v 2 + 1) N2 + 1
96CHAPTER 6. COMPARATIVE STATICS AND THE CONCEPT OF DERIVATIVE
v 3 + v 2 − 4v − 4
q=
v2 − 4
v 3 + v 2 − 4v − 4 v(v 2 − 4) + v 2 − 4
q = =
v2 − 4 v2 − 4
(v + 1)(v − 4)
2
= = v + 1 (v ̸= 2, −2).
v2 − 4
∆y
f ′ (x0 ) = lim
∆x→0 ∆x
f (x0 + ∆x) − f (x0 )
≡ lim (differentiability condition).
∆x→0 ∆x
f (x) − f (x0 )
f (x) − f (x0 ) = (x − x0 ).
x − x0
6.7. CONTINUITY AND DIFFERENTIABILITY OF A FUNCTION 97
Taking the limit of each side of the above equation as x → x0 yields the
following results:
f (x) − f (x0 )
Right side = lim lim (x − x0 )
x→x0 x − x0 x→x0
= f ′ (x0 ) lim (x − x0 ) = 0.
x→x0
f (x) − f (0) x
lim+ = lim+ = lim+ 1 = 1.
x→0 x−0 x→0 x x→0
∆y
Thus, limx→0 ∆x
does not exist since the left-side limit and the right-
side limit are not the same, which implies that the derivative of y = |x|
does not exist at x = 0.
98CHAPTER 6. COMPARATIVE STATICS AND THE CONCEPT OF DERIVATIVE
Chapter 7
Constant-Function Rule
dy
≡ y ′ ≡ f ′ = 0.
dx
99
100CHAPTER 7. RULES OF DIFFERENTIATION AND THEIR USE IN COMPARATIVE ST
Proof.
dy df
We can also write dx
= dx
as
d d
y= f.
dx dx
Power-Function Rule
d
f (x) = axa−1 .
dx
(i) If a = 0, then
d 0 d
x = 1 = 0.
dx dx
dx
= 1.
dx
For simplicity, we prove this rule only for the case where a = n, where
n is any positive integer. Since
then
xn − xn0
= xn−1 + x0 xn−2 + x20 xn−3 + · · · + x0n−1 .
x − x0
7.1. RULES OF DIFFERENTIATION FOR A FUNCTION OF ONE VARIABLE101
Thus,
= xn−1
0 + xn−1
0 + xn−1
0 + · · · + xn−1
0
= nxn−1
0 .
√
x. Then y ′ = 12 x− 2 . In particular, we
1
Example 7.1.2 Suppose y = f (x) =
√
can know that f ′ (2) = · 2− 2 =
1 1 2
2 4
.
dy df
= = acxa−1 .
dx dx
dy
= 2x0 = 2.
dx
dy
= 4 · 3x3−1 = 12x2 .
dx
dy
= −6x−2−1 = −6x−3 .
dx
102CHAPTER 7. RULES OF DIFFERENTIATION AND THEIR USE IN COMPARATIVE ST
Common Rules:
We will come back to discuss the exponential function and log func-
tions and their derivatives in Chapter 10.
7.2. RULES OF DIFFERENTIATION INVOLVING TWO OR MORE FUNCTIONS OF THE SAME V
Let f (x) and g(x) be two differentiable functions. We have the following
rules:
Sum-Difference Rule:
d d d
[f (x) ± g(x)] = f (x) ± g(x) = f ′ (x) ± g ′ (x).
dx dx dx
This rule can easily be extended to more functions:
[ ]
d ∑n ∑n
d ∑n
fi (x) = fi (x) = fi′ (x).
dx i=1 i=1 dx i=1
Example 7.2.1
d
(ax2 + bx + c) = 2ax + b.
dx
dC
= 3Q2 − 8Q + 10.
dQ
L’Hopital Rule
Theorem 7.2.1 (L’Hopital Rule) Suppose that f (x) and g(x) are differentiable
on an open interval (a, b), except possibly at c. If limx→c f (x) = limx→c g(x) = 0
or limx→c f (x) = limx→c g(x) = ±∞, g ′ (x) ̸= 0 for all x in (a, b) with x ̸= c,
f ′ (x)
and limx→c ′ exists. Then,
g (x)
f (x) f ′ (x)
lim = lim ′ .
x→c g(x) x→c g (x)
Example 7.2.3
v 3 + v 2 − 4v − 4
q= .
v2 − 4
Note that limv→2 v 3 + v 2 − 4v − 4 = 0 and limv→2 v 2 − 4 = 0. Then, by
L’Hopital Rule, we have
v 3 + v 2 − 4v − 4 + v 2 − 4v − 4)
d
dv
(v 3
lim = lim
v→2 v2 − 4 v→2 d
dv
(v 2 − 4)
3v 2 + 2v − 4
= lim = 3.
v→2 2v
Example 7.2.4
4v + 5
q= .
v2 + 2v − 3
Since limv→∞ 4v + 5 = ∞ and limv→∞ v 2 + 2v − 3 = ∞, by L’Hopital Rule,
we then have
d
4v + 5 dv
(4v + 5)
lim = lim
v→∞ v + 2v − 3
2 v→∞ d
(v 2 + 2v − 3)
dv
4
= lim = 0.
v→∞ 2v + 2
7.2. RULES OF DIFFERENTIATION INVOLVING TWO OR MORE FUNCTIONS OF THE SAME V
Product Rule:
d d d
[f (x)g(x)] = f (x) g(x) + g(x) f (x)
dx dx dx
= f (x)g ′ (x) + g(x)f ′ (x).
Proof.
Example 7.2.5 Suppose y = (2x+3)(3x2 ). Let f (x) = 2x+3 and g(x) = 3x2 .
Then f ′ (x) = 2, g ′ (x) = 6x. Hence,
d
[(2x + 3)(3x2 )] = (2x + 3)6x + 3x2 · 2
dx
= 12x2 + 18x + 6x2
= 18x2 + 18x.
d
[f (x)g(x)h(x)] = f ′ (x)g(x)h(x) + f (x)g ′ (x)h(x) + f (x)g(x)h′ (x).
dx
106CHAPTER 7. RULES OF DIFFERENTIATION AND THEIR USE IN COMPARATIVE ST
AR = 15 − Q.
T R ≡ AR · Q = 15Q − Q2 .
d
MR ≡ T R = 15 − 2Q.
dQ
T R ≡ AR · Q = Qf (Q).
Thus
d
MR ≡ T R = f (Q) + Qf ′ (Q).
dQ
From this, we can tell relationship between M R and AR. Since
M R − AR = Qf ′ (Q),
TR PQ
AR ≡ = = P,
Q Q
we can view AR as the inverse demand function for the product of the
firm. If the market is perfectly competitive, i.e., the firm takes the price as
given, then P = f (Q) =constant. Hence f ′ (Q) = 0. Thus M R − AR = 0
7.2. RULES OF DIFFERENTIATION INVOLVING TWO OR MORE FUNCTIONS OF THE SAME V
Quotient Rule
[ ]
d f (x) f ′ (x)g(x) − f (x)g ′ (x)
= .
dx g(x) g 2 (x)
We will come back to prove this rule after learning the chain rule.
Example 7.2.6
[ ]
d 2x − 3 2(x + 1) − (2x − 3)(1) 5
= = .
dx x + 1 (x + 1)2 (x + 1)2
[ ]
d 5x 5(x2 + 1) − 5x(2x) 5(1 − x2 )
= = .
dx x2 + 1 (x2 + 1)2 (x2 + 1)2
[ ]
d ax2 + b 2ax(cx) − (ax2 + b)c c(ax2 − b) ax2 − b
= = = .
dx cx (cx)2 (cx)2 cx2
C(Q)
AC ≡ (Q > 0),
Q
and
M C ≡ C ′ (Q).
108CHAPTER 7. RULES OF DIFFERENTIATION AND THEIR USE IN COMPARATIVE ST
d
AC > 0 iff M C(Q) > AC(Q);
dQ
d
AC = 0 iff M C(Q) = AC(Q);
dQ
d
AC < 0 iff M C(Q) < AC(Q).
dQ
Chain Rule
dz ∆z ∆z ∆y
= lim = lim .
dx ∆x→0 ∆x ∆x→0 ∆y ∆x
dz dz dy 1 g ′ (x)
= · = − 2 g ′ (x) = − 2 .
dx dy dx y g (x)
Thus,
[ ]
d f (x) d
= [f (x) · g −1 (x)]
dx g(x) dx
d −1
= f ′ (x)g −1 (x) + f (x)[g (x)]
dx
[ ]
′ −1 g(x)
= f (x)g (x) + f (x) − 2
g (x)
′ ′
f (x)g(x) − f (x)g (x)
= . Q.E.D.
g 2 (x)
dz dz dy
= = 6y(2) = 12y = 12(2x + 5).
dx dy dx
dz dz dy
= = 1 · 3x2 = 3x2 .
dx dy dx
The usefulness of this rule can best be appreciated when one must dif-
ferentiate a function such as those below.
dz dz dy
= = 17y 16 · (2x + 3)
dx dy dx
= 17(x2 + 3x − 2)16 (2x + 3).
Once being familiar with the chain rule, it is unnecessary to adopt in-
termediate variables to find the derivative of a function.
We can find the derivative of a more general function by applying the
chain rule repeatedly.
Example 7.3.4 z = [(x3 − 2x + 1)3 + 3x]−2 . Applying the chain rule repeat-
edly, we have.
dz
= −2[(x3 − 2x + 1)3 + 3x]−3 [3(x3 − 2x + 1)2 (3x2 − 2) + 3].
dx
dR dR dQ
M RPL = = = f ′ (Q)g ′ (L) = M R · M PL ,
dL dQ dL
where M RPL is marginal physical product of labor. Thus the result shown
above constitutes the mathematical statement of the well-known result in
economics that M RPL = M R · M PL .
Inverse-Function Rule
function f ; it does not mean the reciprocal of the function f (x). When x
and y refer specifically to numbers, the property of one-to-one mapping is
seen to be unique to the class of function known as monotonic function.
If an inverse function exists, the original and the inverse functions must
be both monotonic. Moreover, if f −1 is the inverse function of f , then f
must be the inverse function of f −1 .
In general, we may not have the explicit inverse function. However,
we can easily find the derivative of an inverse function by the following
inverse function rule:
dx 1
= dy .
dy dx
Proof.
dx ∆x 1 1
= lim = lim ∆y = ′
dy ∆y→0 ∆y ∆x→0 ∆x y
by noting that ∆y → 0 implies ∆x → 0.
dx 1 1
= dy = 4 .
dy dx
5x + 1
7.4. INTEGRATION (THE CASE OF ONE VARIABLE) 113
dx 1 1
= = = x = ey .
dy dy/dx 1/x
Rules of Integration
∫ ∫ ∫
• [af (x) + bg(x)]dx = a f (x)dx + b g(x)dx, where a and b are con-
stants (linearity of the integral);
∫ ∫
• f ′ (x)g(x)dx = f (x)g(x) − f (x)g ′ (x)dx (integration by parts);
∫ ∫
• f (u(t)) du
dt
dt = f (u)du (integration by substitution).
114CHAPTER 7. RULES OF DIFFERENTIATION AND THEIR USE IN COMPARATIVE ST
∫
f ′ (x)
dx = ln |f (x)| + C;
f (x)
∫
1
dx = ln |x| + C;
∫ x
ex dx = ex + C;
∫
f ′ (x)ef (x) dx = ef (x) + C;
∫
xa+1
xa dx = + C, a ̸= −1;
a+1
∫
ax
ax dx = + C, a > 0.
ln a
Example 7.4.1
∫ ∫ ∫ ∫
x2 + 2x + 1 1 x2
dx = xdx + 2dx + dx = + 2x + ln |x| + C.
x x 2
Example 7.4.2
∫
1∫ 1 ∫ −z e−x
2
−x2
(−2x)e−x dx = −
2
xe dx = − e dz = − + C.
2 2 2
Example 7.4.3
∫ ∫
xe dx = xe −
x x
ex dx = xex − ex + C.
for F (x) such that F ′ (x) = f (x) for all x ∈ [a, b].
7.4. INTEGRATION (THE CASE OF ONE VARIABLE) 115
∫ b ∫ b ∫ b
[αf (x) + βg(x)]dx = α f (x)dx + β g(x)dx;
a a a
∫ b ∫ a
f (x)dx = − f (x)dx;
∫aa b
f (x)dx = 0;
a
∫ b ∫ c ∫ b
f (x)dx = f (x)dx + f (x)dx;
a a c
∫ b ∫ b
f (x)dx ≤ |f (x)|dx;
a a
∫ b ∫ g(b)
f (x)g ′ (x)dx = f (u)du, u = g(x) (change of variable);
a g(a)
∫ b ∫ b
f ′ (x)g(x)dx = f (x)g(x)|ba − f (x)g ′ (x)dx,
a a
d ∫ b(λ)
f (x)dx = f (b(λ))b′ (λ) − f (a(λ))a′ (λ).
dλ a(λ)
Example 7.4.4
d ∫x
f (t)dt = f (x).
dx a
116CHAPTER 7. RULES OF DIFFERENTIATION AND THEIR USE IN COMPARATIVE ST
Partial Derivatives
Consider a function
y = f (x1 , x2 , · · · , xn ),
∂y ∆y
= lim .
∂xi ∆xi →0 ∆xi
∂y
≡ fu = (3u + 2v) + (u + 4) · 3
∂u
= 6u + 2v + 12;
∂y
≡ fv = 2(u + 4).
∂v
When u = 2 and v = 1, then fu (2, 1) = 26 and fv (2, 1) = 12.
and
∂y −2(u2 + 3v) − (3u − 2v) · 3 −u(2u + 9)
= 2 2
= .
∂v (u + 3v) (u2 + 3v)2
118CHAPTER 7. RULES OF DIFFERENTIATION AND THEIR USE IN COMPARATIVE ST
Market Model
a+c
p̄ = ;
b+d
ad − bc
Q̄ = .
b+d
∂ p̄ 1
= ;
∂a b+d
∂ p̄ a+c
=− ;
∂b (b + d)2
∂ p̄ 1
= ;
∂c b+d
∂ p̄ a+c
=− .
∂d (b + d)2
∂ p̄ ∂ p̄ ∂ p̄ ∂ p̄
Thus ∂a
= ∂c
> and ∂b
= ∂d
< 0.
National-Income Model
Y = C + I0 + G0 (equilibrium condition);
C = α + β(Y − T ) (α > 0; 0 < β < 1);
T = γ + δY (γ > 0; 0 < δ < 1),
α − βγ + I0 + G0
Ȳ = .
1 − β + βδ
Thus,
∂ Ȳ 1
= > 0 (the government-expenditure multiplier);
∂G0 1 − β + βδ
∂ Ȳ −β
= < 0;
∂γ 1 − β + βδ
∂ Ȳ −β(α − βγ + I0 + G0 ) −β Ȳ
= = < 0.
∂δ (1 − β + βδ) 2 (1 − β + βδ)
7.7. NOTE ON JACOBIAN DETERMINANTS 121
Partial derivatives can also provide a means of testing whether there exists
functional (linear or nonlinear) dependence among a set of n variables.
This is related to the notion of Jacobian determinants.
Consider n differentiable functions in n variables not necessary linear.
y1 = f 1 (x1 , x2 , · · · , xn );
y2 = f 2 (x1 , x2 , · · · , xn );
··· ;
yn = f n (x1 , x2 , · · · , xn ),
where the symbol f i denotes the ith function, we can derive a total of n2
partial derivatives.
∂yi
(i = 1, 2, · · · , n; j = 1, 2, · · · , n).
∂xj
We can arrange them into a square matrix, called the Jacobian matrix
and denoted by J , and then take its determinant, the result will be what
is known as a Jacobian determinant (or a Jacobian, for short), denoted by
|J |:
∂y1
∂x1
∂y1
∂x2
··· ∂y1
∂xn
∂(y1 , y2 , · · · , yn )
∂y2 ∂y2
··· ∂y2
|J | = = ∂x1 ∂x2 ∂xn
.
∂(x1 , x2 , · · · , xn ) ··· ··· ··· ···
∂yn
∂x1
∂yn
∂x2
··· ∂yn
∂xn
y1 = 2x1 + 3x2 ;
122CHAPTER 7. RULES OF DIFFERENTIATION AND THEIR USE IN COMPARATIVE ST
∂y1 ∂y1
2 3
|J | = ∂x1 ∂x2
= = 0.
∂y2 ∂y2
∂x1 ∂x2
(8x1 + 12x2 ) (12x1 + 18x2 )
∂(y1 , y2 )
|J | = = (24x1 + 36x2 ) − (24x1 + 36x2 ) ≡ 0
∂(x1 , x2 )
··· ;
We know that the rows of the coefficient matrix A are linearly depen-
dent if and only if |A| = 0. This result can now be interpreted as a special
7.7. NOTE ON JACOBIAN DETERMINANTS 123
Comparative-Static Analysis of
General-Functions
The study of partial derivatives has enables us, in the preceding chapter,
to handle the simple type of comparative-static problems, in which the e-
quilibrium solution of the model can be explicitly stated in the reduced
form. We note that the definition of the partial derivative requires the ab-
sence of any functional relationship among the independent variables. As
applied to comparative-static analysis, this means that parameters and/or
exogenous variables which appear in the reduced-form solution must be
mutually independent.
Y = C + I0 + G0 (equilibrim condition);
125
126CHAPTER 8. COMPARATIVE-STATIC ANALYSIS OF GENERAL-FUNCTIONS
Y = C(Y, T0 ) + I0 + G0
even though we are unable to determine explicitly the form which this
function takes. Furthermore, in some neighborhood of Ȳ , the following
identical equality will hold:
Ȳ ≡ C(Ȳ , T0 ) + I0 + G0 .
8.1 Differentials
The symbol dy/dx has been regarded as a single entity. We shall now rein-
terpret as a ratio of two quantities, dy and dx.
the magnitude of ∆y can be found, once the ∆y/∆x and the variation ∆x
are known. If we denote the infinitesimal changes in x and y, respectively,
by dx and dy, the identity (8.1) becomes
[ ]
dy
dy ≡ dx. (8.1.2)
dx
or
dy = f ′ (x)dx. (8.1.3)
or
(dy)
≡ f ′ (x).
(dx)
This result shows that the derivative dy/dx ≡ f ′ (x) may be interpreted
128CHAPTER 8. COMPARATIVE-STATIC ANALYSIS OF GENERAL-FUNCTIONS
The following diagram shows the relationship between “∆y" and “dy".
Remark 8.1.1 The purpose of finding the differential dy is called the dif-
ferentiation. Recall that we have also used this term as a synonym for
derivation. To avoid confusion, the word “differentiation" with the phrase
“with respect to x" when we take derivative dy/dx.
S = S(Y, i),
∂S ∂S
dS = dY + di.
∂Y ∂i
That is, the infinitesimal change in S is the sum of the infinitesimal change
in Y and the infinitesimal change in i.
Remark 8.2.1 If i remains constant, the total differential will reduce to the
partial differential: ( )
∂S dS
= .
∂Y dY i constant
∂f ∂f ∂f ∑n
df = dx1 + dx2 + · · · + dxn = fi dxi ,
∂x1 ∂x2 ∂xn i=1
in which each term on the right side indicates the amount of change in y
resulting from an infinitesimal change in one of n variables.
Similar to the case of one variable, the n partial elasticities can be writ-
ten as
∂f xi
ϵf xi = (i = 1, 2, · · · , n).
∂xi f
8.3. RULE OF DIFFERENTIALS 131
Rule I: dc = 0;
Example 8.3.1 Find dy of the function y = 5x21 +3x2 . There are two ways to
find dy. One is the straightforward method by finding ∂f /∂x1 and ∂f /∂x2 :
∂f /∂x1 = 10x1 and ∂f /∂x2 = 3, which will then enable us to write
The other way is to use the rules given above by letting u = 5x21 and
v = 3x2 ;
Example 8.3.2 Find dy of the function y = 3x21 + x1 x22 . Since f1 = 6x1 + x22
and f2 = 2x1 x2 , the desired differential is
x1 + x2
y=
2x21
−(x1 + 2x2 ) 1
f1 = 3
and f2 = 2 ,
2x1 2x1
then
−(x1 + 2x2 ) 1
dy = 3
dx1 + 2 dx2 .
2x1 2x1
The same result can also be obtained by applying the given rules:
1
dy = [2x2 d(x1 + x2 ) − (x1 + x2 )d(2x21 )] [by rule V]
4x41 1
1
= 4 [2x21 (dx1 + dx2 ) − (x1 + x2 )4x1 dx1 ]
4x1
1
= 4 [−2x1 (x1 + 2x2 )dx1 + 2x21 dx2 ]
4x1
x1 + 2x2 1
=− 3
dx1 + 2 dx2 .
2x1 2x1
Consider a function
Here, the variable w can affect y through two channels: (1) indirectly, vi-
a the function g and then f , and (2) directly, via the function. Unlike a
partial derivative, the total derivative does not require the argument x to
main constant as w varies, and can thus allow for the postulated relation-
ship between the two variables. Whereas the partial derivative fw is ade-
quate for expressing the direct effect alone, the total derivative is needed
to express both effects jointly.
To get the total derivative, we first get the total differential
dy = fx dx + fw dw.
Dividing both sides of this equation by dw, we have the total derivative:
dy dx dw
= fx + fw
dw dw dw
∂y dx ∂y
= + .
∂x dw ∂w
dy dx
= fx + fw = 3(4w + 1) − 2w = 10w + 3.
dw dw
As a check, we may substitute the function g into f , to get
dy
= 10w + 3,
dw
U = U (s, c),
dU ∂U dc ∂U
M Us = = +
ds ∂c ds ∂s
∂U 1 ∂U
= ′
+ .
∂c g (c) ∂s
∂U ′
[ ∂U ]
M Uc g (c) + ∂U ′ g ′ (c) + ∂U
M RScs = = ∂s
∂U 1
∂c
∂U = g (c) ∂s
∂U
∂c
′ (c)
= g ′ (c).
M Us ∂c g ′ (c)
+ ∂s ∂c
+ ∂U ∂s
g
For a function
y = f (x1 , x2 , w)
8.4. TOTAL DERIVATIVES 135
dy ∂f dx1 ∂f dx2 ∂f
= + + .
dw ∂x1 dw ∂x2 dw ∂w
Q = Q(K, L, t),
where K is the capital input, L is the labor input, and t is the time which
indicates that the production can shift over time in reflection of technolog-
ical change. Since capital and labor can also change over time, we may
write
K = K(t) and L = L(t).
dQ ∂Q dK ∂Q dL ∂Q
= + + .
dt ∂K dt ∂L dt ∂t
y = f (x1 , x2 , u, v)
with x1 = g(u, v) and x2 = h(u, v), we can find the total derivative of y
with respect to u (while v is held constant). Since
∂f ∂f ∂f ∂f
dy = dx1 + dx2 + du + dv,
∂x1 ∂x2 ∂u ∂v
136CHAPTER 8. COMPARATIVE-STATIC ANALYSIS OF GENERAL-FUNCTIONS
dy ∂y dx1 ∂y dx2 ∂y du ∂y dv
= + + +
du ∂x1 du ∂x2 du ∂u du ∂v du
[ ]
∂y dx1 ∂y dx2 ∂y dv
= + + = 0 since v is constant .
∂x1 du ∂x2 du ∂u du
§y ∂y ∂x1 ∂y ∂x2 ∂y
= + + .
§u ∂x1 ∂u ∂x2 ∂u ∂u
§z §z
Example 8.4.4 Find the partial total derivatives §u
and §v
if
z = 3x2 − 2y 4 + 5uv 2 ,
where
x = u − v 2 + 4.
and
y = 8u3 v + v 2 + 1.
§z ∂z ∂x ∂z ∂y ∂z
= + +
§u ∂x ∂u ∂y ∂u ∂u
= 6x × 1 − 8y 3 × 24u2 v + 5v 2
= 6x − 196y 3 u2 v + 5v 2 .
8.5. IMPLICIT FUNCTION THEOREM 137
and
§z ∂z ∂x ∂z ∂y ∂z
= + +
§v ∂x ∂v ∂y ∂v ∂v
= 6x × −2v − 8y 3 × (8u3 + 2v) + 10uv
= −12xv − 8y 3 (8u3 + 2v) + 10uv.
Remark 8.4.1 In the cases we have discussed, the total derivative formu-
las can be regarded as expressions of the chain rule, or the composite-
function rule. Also the chain of derivatives does not have to be limited to
only two “links"; the concept of the total derivative should be extendible
to cases where there are three or more links in the composite function.
The concept of total differentials can also enable us to find the derivatives
of the so-called “implicit functions." As such, we can still do comparative-
static analysis for general functions.
Implicit Functions
F (y, x1 , x2 , · · · , xn ) = 0.
F (y, x1 , x2 , · · · , xn ) ≡ y − f (x1 , x2 , · · · , xn ) = 0.
(a) the function F has continuous partial derivatives Fy , Fx1 , Fx2 , · · · , Fxn ;
(b) at point (y0 , x10 , x20 , · · · , xn0 ) satisfying F (y0 , x10 , x20 , · · · , xn0 ) =
0, Fy is nonzero.
Differentiating F , we have dF = 0, or
Fy dy + F1 dx1 + · · · + Fn dxn = 0.
Suppose that only y and x1 are allowed to vary. Then the above equa-
8.5. IMPLICIT FUNCTION THEOREM 139
dy ∂y F1
≡ =− .
dx1 other variable constant ∂x1 Fy
In the simple case where the given equation is F (y, x) = 0, the rule
gives
dy Fx
=− .
dx Fy
= − FFxy = − −12x
3
Example 8.5.1 Suppose y − 3x4 = 0. Then dy
dx 1
= 12x3 .
In this particular case, we can easily solve the given equation for y, to
get y = 3x4 so that dy/dx = 12x3 .
dy Fx 2x x
=− =− = − , (y ̸= 0).
dx Fy 2y y
∂y Fx 2y 3 x + yw
=− =− 2 2 .
∂x Fy 3y x + xw
∂Q FK
M PK ≡ =− ;
∂K FQ
∂Q FL
M PL ≡ =− .
∂L FQ
In particular, we can also find the M RT SLK (marginal rate of technical
140CHAPTER 8. COMPARATIVE-STATIC ANALYSIS OF GENERAL-FUNCTIONS
∂K FL
M RT SLK ≡ | |= .
∂L FK
F 1 (y1 , y2 , · · · , yn ; x1 , x2 , · · · , xm ) = 0;
F 2 (y1 , y2 , · · · , yn ; x1 , x2 , · · · , xm ) = 0;
···
F n (y1 , y2 , · · · , yn ; x1 , x2 , · · · , xm ) = 0.
∂F 1 ∂F 1 ∂F 1
∂y1 ∂y2
··· ∂yn
∂F 2 ∂F 2 ∂F 2
∂(F 1 , F 2 , · · · , F n ) ···
|J | = = ∂y1 ∂y2 ∂yn
̸= 0.
∂(y1 , y2 , · · · , yn ) ··· ··· ··· ···
∂F n ∂F n ∂F n
∂y1 ∂y2
··· ∂yn
where
∂F 1 ∂F 1 ∂F 1
∂x1 ∂x2
··· ∂xn
∂F 2
∂F 2
··· ∂F 2
∂x1 ∂x2 ∂xn
Fx = .
··· ··· ··· ···
∂F n ∂F n ∂F n
∂x1 ∂x2
··· ∂xn
∂yj |J i |
= j (j = 1, 2, · · · , n; i = 1, 2, · · · , m),
∂xi |J |
142CHAPTER 8. COMPARATIVE-STATIC ANALYSIS OF GENERAL-FUNCTIONS
∂y
= −J −1 Fxi .
∂xi
Y − C − I0 − G0 = 0;
C − α − β(Y − T ) = 0;
T − γ − δY = 0.
Then
∂F 1 ∂F 1 ∂F 1
∂Y ∂C ∂T
1 −1 0
|J | = ∂F 2
∂Y
∂F 2
∂C
∂F 2
∂T
= −β 1 β = 1 − β + βδ.
∂F 3 ∂F 3 ∂F 3
∂Y ∂C ∂T
−δ 0 1
Suppose that all exogenous variables and parameters are fixed except
8.6. COMPARATIVE STATICS OF GENERAL-FUNCTION MODELS 143
G0 . Then we have
1 −1 0 ∂G
∂ Ȳ
1
0
−β 1 β ∂ C̄
∂G0 = 0 .
−δ 0 1 ∂ T̄
∂G0
0
We can solve the above equation for, say, ∂ Ȳ /∂G0 which comes out to be
1 −1 0
0 1 β
∂ Ȳ 0 0 1 1
= = .
∂G0 |J| 1 − β + βδ
Qd = Qs , [equilibrium condition];
Qd = D(P, Y0 ), [∂D/∂P < 0; ∂D/∂Y0 > 0];
Qs = S(P ), [dS/dP > 0],
D(P, Y0 ) − S(P ) = 0.
Even though this equation cannot be solved explicitly for the equilib-
rium price P̄ , by the implicit-function theorem, we know that there exists
the equilibrium price P̄ that is the function of Y0 :
P̄ = P̄ (Y0 ),
144CHAPTER 8. COMPARATIVE-STATIC ANALYSIS OF GENERAL-FUNCTIONS
such that
F (P̄ , Y0 ) ≡ D(P̄ , Y0 ) − S(P̄ ) = 0.
dQ̄ dS dP̄
= > 0.
dY0 dP dY0
∂
|A| = [Cij ]
∂A
da′ b
= a;
db
db′ a
= a;
db
dM b
= M ′;
db
db′ M b
= (M + M ′ )b.
db
Example 8.7.1 (Find the Least Square Estimator for Multiple Regression Model)
Consider the multiple regression model:
y = Xβ + ϵ,
dE(b)
= 0 ⇒ −2X ′ y + 2X ′ Xb = 0 ⇒ b = (X ′ X)−1 X ′ y.
db
dE 2 (b)
= (2X ′ X)′ = 2X ′ X.
db2
146CHAPTER 8. COMPARATIVE-STATIC ANALYSIS OF GENERAL-FUNCTIONS
147
148CHAPTER 9. OPTIMIZATION: MAXIMA AND MINIMA OF A FUNCTION OF ONE V
y = f (x).
Three specific cases of functions are depicted in Figure 9.1. The point E
and F in (c) are relative (or local) extremum, in the sense that each of these
points represents an extremum in some neighborhood of the point only.
We shall continue our discussion mainly with reference to the search for
relative extreme. Since an absolute (or global) maximum must be either
a relative maxima or one of the ends of the function. Thus, if we know
9.2. EXISTENCE OF EXTREMUM FOR CONTINUOUS FUNCTION 149
Figure 9.1: The extremum for various functions: (a) constant function; (b)
monotonic function, (3) non-monotonic function.
all the relative maxima, it is necessary only to select the largest of these
and compare it with the end points in order to determine the absolute
maximum. Hereafter, the extreme values considered will be relative or
local ones, unless indicated otherwise.
Definition 9.2.2 (Global Optimum) If f (x∗ ) = f (x) (resp. f (x∗ ) > f (x))
for all x in the domain X of the function, then the function is said to have
global (unique) maximum at x∗ ; if f (x∗ ) 5 f (x) (resp. f (x∗ ) < f (x)) for
150CHAPTER 9. OPTIMIZATION: MAXIMA AND MINIMA OF A FUNCTION OF ONE V
all x in the domain of the function, then the function is said to have global
(unique) minimum at x∗ .
f ′ (x) = 0
Figure 9.2: The first derivative test: (a) f ′ (x0 ) doest not exist; and (b)
f ′ (x0 ) = 0.
Note that if the first derivative vanishes at some point, it does not imply
that at this point f possesses an extremum. Such an example is f = x3 . As
such, we can only state that f has a stationary point.
We have some useful results about stationary points.
Figure 9.3: The Mean-Value Theorem implies that there exists some c in
the interval (a, b) such that the secant joining the endpoints of the interval
[a, b] is parallel to the tangent at c.
where [ ]
∂f (x) ∂f (x) ∂f (x)
Df (x) = , ,··· , .
∂x1 ∂x2 ∂xn
An variation of the above mean-value theorem is in form of integral
calculus:
F (b) − F (a)
= F ′ (c) = f (c).
b−a
Therefore, we have
∫ b
f (x)dx = f (c)(b − a).
a
The second variation of the mean-value theorem is the generalized
mean-value theorem:
Proof. The case that g(a) = g(b) is easy. So, assume that g(a) ̸= g(b). Define
f (b) − f (a)
h(x) = f (x) − g(x).
g(b) − g(a)
(a) a relative maximum if f ′ (x) changes its sign from positive to negative from
the immediate left of the point x0 to its immediate right;
(b) a relative minimum if f ′ (x) changes its sign from negative to positive from
the immediate left of the point x0 to its immediate right;
154CHAPTER 9. OPTIMIZATION: MAXIMA AND MINIMA OF A FUNCTION OF ONE V
(c) an inflection (not extreme) point if f ′ (x) has the same sign on some neigh-
borhood.
Its roots are x̄1 = 2 and x̄2 = 6. It is easy to verify that f ′ (x) > 0 for
x < 2 and f ′ (x) < 0 for x > 2. Thus x = 2 is a maximal point and the
corresponding maximum value of the function f (2) = 40. Similarly, we
can verify that x = 6 is a minimal point and f (6) = 8.
AC = f (Q) = Q2 − 5Q + 8.
Since f ′ (2.5) = 0, f ′ (Q) < 0 for Q < 2.5, and f ′ (Q) > 0 for Q > 2.5, so
Q̄ = 2.5 is a minimal point.
or
d3 y d4 y dn y
, , · · · , .
dx3 dx4 dxn
Remark 9.4.1 dn y/dxn can be also written as (dn /dxn )y, where the dn /dxn
part serves as an operator symbol instructing us to take the n-th derivative
with respect to x.
x
y = g(x) = (x ̸= −1).
1+x
156CHAPTER 9. OPTIMIZATION: MAXIMA AND MINIMA OF A FUNCTION OF ONE V
g ′ (x) = (1 + x)−2 ;
g ′′ (x) = −2(1 + x)−3 ;
g ′′′ (x) = 6(1 + x)−4 ;
g (4) (x) = −24(1 + x)−5 .
Remark 9.5.1 Note that when f ′ (x0 ) = 0, f ′′ (x0 ) < 0 (f ′′ (x0 ) > 0) is a
sufficient condition for a relative maximum (resp. minimum) but not a
necessary condition. However, the condition f ′′ (x0 ) ≤ 0 (f ′′ (x0 ) ≥ 0) is a
necessary (even though not sufficient) for a relative maximum (resp. min-
imum).
or
R′ (Q̄) = C ′ (Q̄), or M R(Q̄) = M C(Q̄).
d2 π
≡ π ′′ (Q̄) − C ′′ (Q̄) < 0.
dQ
158CHAPTER 9. OPTIMIZATION: MAXIMA AND MINIMA OF A FUNCTION OF ONE V
1 ′′ 1
ϕ(x) = ϕ(x0 ) + ϕ′ (x0 )(x − x0 ) + ϕ (x0 )(x − x0 )2 + · · · + ϕ(n) (x0 )(x − x0 )n + Rn
2! n!
≡ Pn + Rn ,
ϕ(n+1) (P )
Rn = (x − x0 )n+1
(n + 1)!
with P being a point between x and x0 . Here n! is the "n factorial", defined as
or
ϕ(x) − ϕ(x0 ) = ϕ′ (P )(x − x0 ),
which states that the difference between the value of the function ϕ at x0
and at any other x value can be expressed as the product of the difference
(x − x0 ) and ϕ′ (P ) with P being some point between x and x0 .
1 ′′ 1 1
ϕ(x) = ϕ(0)+ ϕ′ (0)x+ ϕ (0)x2 +· · · + ϕ(n) (0)xn + ϕ(n+1) (P )xn+1 ,
2! n! (n + 1)!
1
ϕ(x) =
1+x
160CHAPTER 9. OPTIMIZATION: MAXIMA AND MINIMA OF A FUNCTION OF ONE V
1
around the point x0 = 1, with n = 4. Since ϕ(1) = 2
and
1 ′′
f (x) − f (x0 ) =f ′ (x0 )(x − x0 ) + f (x0 )(x − x0 )2 + · · ·
2!
1 1
+ f (n) (x0 )(x − x0 )n + f (n+1) (x0 )(x − x0 )n+1 .
n! (n + 1)!
where b denotes a fixed base of the exponent t. Its generalized version has
the form:
y = abct .
163
164 CHAPTER 10. EXPONENTIAL AND LOGARITHMIC FUNCTIONS
y = aert
t = logb y,
and
t = loge y ≡ ln y.
blogb y = bt = y.
Rules:
Properties of Log:
(e) log y → ∞ as y → ∞;
(f) log y → −∞ as y → 0.
d ln t
(a) dt
= 1t ;
det
(b) dt
= et ;
166 CHAPTER 10. EXPONENTIAL AND LOGARITHMIC FUNCTIONS
d f ′ (t)
(d) dt
ln f (t) = f (t)
.
dbt
(a) dt
= bt ln b;
d 1
(b) dt
logb t = t ln b
;
(c) d f (t)
dt
b = f ′ (t)bf (t) ln b;
d f ′ (t) 1
(d) dt
logb f (t) = f (t) ln b
.
t
Proof of (a). Since bt = eln b = et ln b , then (d/dt)bt = (d/dt)et ln b =
(ln b)(et ln b ) = bt ln b.
Proof of (b). Since
An Application
Example 10.3.3 Find dy/dx from y = xa ekx−c . Taking the natural log of
both sides, we have
ln y = a ln x + kx − c.
1 dy a
= + k.
y dx x
dy
Thus dx
= (a/x + k)y = (a/x + k)xa ekx−c .
Use the above technical method, we can similarly find the derivative
of y = ϕ(x)ψ(x) .
168 CHAPTER 10. EXPONENTIAL AND LOGARITHMIC FUNCTIONS
Chapter 11
169
170CHAPTER 11. OPTIMIZATION: MAXIMA AND MINIMA OF A FUNCTION OF TWO
Figure 11.1: The graphical illustrations for extrema of a function with two
choice variables: (a) A is a maximum; and (b) B is a minimum.
First-Order Condition
dz = fx dx + fy dy.
From the function z = f (x, y), we can have two first-order partial
derivatives, fx and fy . Since fx and fy are themselves functions of x, we
can find second-order partial derivatives:
( )
∂ ∂ 2z ∂ ∂z
fxx ≡ fx or 2
≡ ;
∂x ∂x ∂x ∂x
( )
∂ ∂ 2z ∂ ∂z
fyy ≡ fy or 2 ≡ ;
∂y ∂y ∂y ∂y
( )
∂ 2z ∂ ∂z
fxy ≡ ≡ ;
∂x∂y ∂x ∂y
( )
∂ 2z ∂ ∂z
fyx ≡ ≡ .
∂y∂x ∂y ∂x
The last two are called cross (or mixed) partial derivatives.
∂ 2f ∂2f
= , i, j = 1, 2, · · · , n.
∂xj ∂xi ∂xi ∂xj
Remark 11.2.1 Even though fxy and fyx have been separately defined, they
will – according to Young’s theorem, be identical with each other, as long
as the two cross partial derivatives are both continuous. In fact, this the-
orem applies also to functions of three or more variables. Given z =
g(u, v, w), for instance, the mixed partial derivatives will be characterized
by guv = gvu , guw = gwu , etc. provided these partial derivatives are contin-
uous.
Thus, fxx = 6x, fyx = 5, and fyy = −2. As expected, fyx = fxy .
dz = fx dx + fy dy,
∂(dz) ∂(dz)
d2 z ≡ d(dz) = dx + dy
∂x ∂y
∂ ∂
= (fx dx + fy dy)dx + (fx dx + fy dy)dy
∂x ∂y
= [fxx dx + fxy dy]dx + [fyx dx + fyy dy]dy
= fxx dx2 + fxy dydx + fyx dxdy + fyy dy 2
= fxx dx2 + 2fxy dxdy + fyy dy 2 [ if fxy = fyx ].
dz =fx dx + fy dy
=(3x2 + 5y)dx + (5x − 2y)dy.
174CHAPTER 11. OPTIMIZATION: MAXIMA AND MINIMA OF A FUNCTION OF TWO
(a) d2 z is positive definite iff fxx > 0 and |H| = fxx fyy −(fxy )2 >
0;
(b) d2 z is negative definite iff fxx < 0 and |H| = fxx fyy −(fxy )2 >
0.
From the inequality fxx fyy − (fxy )2 > 0, it implies that fxx and fyy are
required to take the same sign.
Example 11.2.4 Give fxx = −2, fxy = 1, and fyy = −1 at a certain point on
a function z = f (x, y), does d2 z have a definite sign at that point regardless
of the values of dx and dy? The Hessian determinant is in this case
−2 1
,
1 −1
11.2. EXTREME VALUES OF A FUNCTION OF TWO VARIABLES 175
−2 1
|H2 | = = 2 − 1 = 1 > 0.
1 −1
Example 11.2.5 Give fxx = −2, fxy = 1, and fyy = −1 at a certain point on
a function z = f (x, y), does d2 z have a definite sign at that point regardless
of the values of dx and dy? The Hessian determinant is in this case
−2 1
,
1 −1
−2 1
|H2 | = = 2 − 1 = 1 > 0.
1 −1
(2) fxx < 0, fyy < 0, and fxx fyy > (fxy )2 .
(2) fxx > 0, fyy > 0, and fxx fyy > (fxy )2 .
24x2 + 2y − 6x = 0;
2y + 2x = 0.
Since fxx (0, 0) = −6 and fyy (0, 0) = 2, it is impossible fxx fyy ≥ (fxy )2 =
4, so the point (x̄1 , ȳ1 ) = (0, 0) is not extreme point. For the solution
(x̄2 , ȳ2 ) = (1/3, −1/3), we find that fxx = 10 > 0, fyy = fxy = 2 > 0, and
fxx fyy − (fxy )2 = 20 − 4 > 0, so (x̄, ȳ, z̄) = (1/3, −1/3, 23/27) is a relative
minimum point.
When there are n choice variables, the objective function may be expressed
as
z = f (x1 , x2 , · · · , xn ).
as
f
11
f12 · · · f1n dx1
[
]
2 f21 f22 · · · f2n dx2
d z = dx1 , dx2 , · · · , dxn
· · · · · · · · · · · ·
···
fn1 fn2 · · · fnn dxn
≡ (dx)′ Hdx.
and the second-order sufficient condition for extremum is, as before, that
all the n principal minors be positive for a minimum in z and that they
duly alternate in sign for a maximum in z, the first one being negative.
In summary, we have the following proposition.
(2) |H1 | < 0, |H2 | > 0, |H3 | < 0, · · · , (−1)n |Hn | > 0. [d2 z is
negative definite].
(2) |H1 | > 0, |H2 | > 0, |H3 | > 0, · · · , |Hn | > 0. [d2 z is positive
definite].
11.4. SECOND-ORDER CONDITIONS IN RELATION TO CONCAVITY AND CONVEXITY179
f1 = 0 : 4x1 + x2 + x3 = 0;
f2 = 0 : x1 + 8x2 + 0 = 0;
f3 = 0 : x1 + 0 + 2x3 = 0,
we can find a unique solution x̄1 = x̄2 = x̄3 = 0. This means that there is
only one stationary value, z̄ = 2. The Hessian determinant of this function
is
f11 f12 f13 4 1 1
|H| = f21 f22 f23 = 1 8 0 .
f31 f32 f33 1 0 2
Since the principal minors of which are all positive: |H1 | = 4, |H2 | =
31, and |H3 | = 54, we can conclude that z̄ = 2 is a minimum.
ly if) d2 z is everywhere negative (resp. positive) definite, i.e., its Hessian matrix
H = D2 f (x) is negative (positive) definite on X.
f11 f12
|H2 | = > 0,
f21 f22
··· ,
(−1)n |Hn | = (−1)n |H| > 0.
∂ 2f
and so on, where fij = . This algebraic condition is very useful for
∂xi ∂xj
testing second-order conditions of optimality. It can easily verify whether
a function is strictly concave (resp. strictly convex) by checking whether
its Hessian matrix is negative (resp. positive) definite.
Since
f11 f12 2 0
|H| = = ,
f21 f22 0 2
Q = f (L, K) = Lα K β ,
fL = αLα−1 K β ,
fK = βLα Lβ−1 ;
fLL = α(α − 1)Lα−2 K β ,
fKK = β(β − 1)Lα K β−2 ,
fLK = αβLα−1 K β−1 ,
thus
tiable, the following position fully characterizes the concavity of the func-
tion.
u(x) − u(x∗ )
5 u′ (x∗ ),
x − x∗
slope=u’(x*)
u’(x)-u’(x*)
slope=
x-x*
x* x
Figure 11.3: The graphical illustration why Proposition 11.4.2 holds for a
concave function.
When there are two or more independent variables, the above propo-
sition becomes:
∑
n
∂f (x)
f (y) 5 f (x) + (yj − xj ). (11.4.2)
j=1 ∂xj
A local optimum is, in general, not equal to the global optimum, but under
certain conditions, these two are consistent with each other.
(1) Df (x∗ ) = 0.
P ROOF . It is clear that (3) ⇒ (2), and it follows from the first-order
condition that (2) ⇒ (1). Therefore, we just need to prove that (1) ⇒ (3).
Suppose Df (x∗ ) = 0. Then f is concave implies that for all x in the
domain, we have:
f (x) 5 f (x∗ ).
Theorem 11.4.3 (The sufficient condition for the uniqueness of global optimum)
Suppose that f (x) is twice continuously differentiable on X ⊆ Rn . We have:
C = 2Q21 + Q1 Q2 + 2Q22 .
π = T R − C = P1 Q1 + P2 Q2 − 2Q21 − Q1 Q2 − 2Q22 .
The firm wants to maximize the profit by choosing the levels of Q1 and
Q2 . For this purpose, setting
∂π
= 0 : P1 − 4Q1 − Q2 = 0;
∂Q1
∂π
= 0 : P2 − Q1 − 4Q2 = 0,
∂Q2
we have
4Q1 + Q2 = P1 ;
Q1 + 4Q2 = P2 ,
188CHAPTER 11. OPTIMIZATION: MAXIMA AND MINIMA OF A FUNCTION OF TWO
and thus
4P1 − P2 4P2 − P1
Q̄1 = and Q̄2 = .
15 15
Also the Hessian matrix is
π11 π12 −4 −1
H= = .
π21 π22 −1 −4
Since |H1 | = −4 < 0 and |H2 | = 16 − 1 > 0, the Hessian matrix (or
d2 z) is negative definite, and the solution does maximize. In fact, since
H is everywhere negative definite, the maximum profit found above is
actually a unique absolute maximum.
Example 11.5.2 Let us now transplant the problem in the above example
into the setting of a monopolistic market.
Suppose that the demands facing the monopolist firm are as follows:
Q1 = 40 − 2P1 + P2 ;
Q2 = 15 + P1 − P2 .
C = Q21 + Q1 Q2 + Q22 .
− 2P1 + P2 = Q1 − 40;
P1 − P2 = Q2 − 15,
11.5. ECONOMIC APPLICATIONS 189
we have
P1 = 55 − Q1 − Q2 ;
P2 = 70 − Q1 − 2Q2 .
T R =P1 Q1 + P2 Q2
=(55 − Q1 − Q2 )Q1 + (70 − Q1 − 2Q2 )Q2 ;
=55Q1 + 70Q2 − 2Q1 Q2 − Q21 − 2Q22 .
π = TR − C
= 55Q1 + 70Q2 − 3Q1 Q2 − 2Q21 − 3Q22 ,
∂π
= 0 : 55 − 4Q1 − 3Q2 = 0;
∂Q1
∂π
= 0 : 70 − 3Q1 − 6Q2 = 0,
∂Q2
2
(Q̄1 , Q̄2 ) = (8, 7 ).
3
1 2 1
P̄1 = 39 , P̄2 = 46 , and π̄ = 488 .
3 3 3
190CHAPTER 11. OPTIMIZATION: MAXIMA AND MINIMA OF A FUNCTION OF TWO
−4 −3
,
−3 −6
we have |H1 | = −4 < 0 and |H2 | = 15 > 0 so that the value of π̄ does rep-
resent the maximum. Also, since Hessian matrix is everywhere negative
definite, it is a unique absolute maximum.
Chapter 12
The last chapter presented a general method for finding the relative ex-
trema of an objective function of two or more variables. One important
feature of that discussion is that all the choice variables are independent
of one another, in the sense that the decision made regarding one vari-
able does not depend on the choice of the remaining variables. However,
in many cases, optimization problems are the constrained optimization
problem. For instance, every consumer maximizes her utility subject to
her budget constraint. A firm minimizes the cost of production with the
constraint of production technique.
191
192 CHAPTER 12. OPTIMIZATION WITH EQUALITY CONSTRAINTS
In general, for a function, say z = f (x, y), the difference between a con-
strained extremum and a free extremum may be illustrated in Figure 12.1.
The free extremum in this particular graph is the peak point of entire
domain, but the constrained extremum is at the peak of the inverse U-
shaped curve situated on top of the constraint line. In general, a constraint
(less freedom) maximum can be expected to have a lower value than the
free (more freedom) maximum, although by coincidence, the two maxi-
ma may happen to have the same value. But the constrained maximum
can never exceed the free maximum. To have certain degrees of freedom
of choices, the number of constraints in general should be less than the
number of choice variables.
12.2. FINDING THE STATIONARY VALUES 193
60 − 4x1
x2 = = 30 − 2x1
2
∂u
= 32 − 4x1 = 0,
∂x1
we get the solution x̄1 = 8 and thus, by the budget constraint, x̄2 =
30 − 2x̄1 = 30 − 16 = 14 since d2 u/dx21 = −4 < 0, that stationary value
constitutes a (constrained) maximum.
Lagrange-Multiplier Method
z = f (x, y)
∂Z
Zλ ≡ = c − g(x, y) = 0;
∂λ
∂Z
Zx ≡ = fx − λgx (x, y) = 0;
∂x
∂Z
Zy ≡ = fy − λgy (x, y) = 0.
∂y
Example 12.2.1 Let us again consider the consumer’s choice problem above.
The Lagrange function is
Zλ = 60 − 4x1 − 2x2 = 0;
Zx1 = x2 + 2 − 4λ = 0;
Zx2 = x1 − 2λ = 0.
Solving the stationary point of the variables, we find that x̄1 = 8, x̄2 = 14,
and λ = 4. As expected, x̄1 = 8 and x̄2 = 14 are the same obtained by the
substitution method.
Zλ = 6 − x − y = 0;
Zx = y − λ = 0;
Zy = x − λ = 0.
196 CHAPTER 12. OPTIMIZATION WITH EQUALITY CONSTRAINTS
Thus, we find λ̄ = 3, x̄ = 3, ȳ = 3.
Zλ = 2 − x1 − 4x2 = 0;
Zx1 = 2x1 − λ = 0;
Zx2 = 2x2 − 4λ = 0.
4 2 8
λ̄ = , x̄1 = , x̄2 = ,
17 17 17
4
is therefore Z̄ = z̄ = 17
.
To tell whether z̄ is a maximum, we need to consider the second-order
condition.
c − g(x̄, ȳ) ≡ 0;
fx (x̄, ȳ) − λ̄gx (x̄, ȳ) ≡ 0;
fy (x̄, ȳ) − λ̄gy (x̄, ȳ) ≡ 0.
Therefore, we have
[ ]
dZ̄ dx̄ dȳ dλ̄ dx̄ dȳ
= fx + fy + [c − g(x̄, ȳ)] + λ 1 − gx − gy
dc dc dc dc dc dc
dx̄ dȳ dx̄
= (fx − λgx ) + (fy − λgy ) + [c − g(x̄, ȳ)] + λ
dc dc dc
= λ.
z = f (x1 , x2 , · · · , xn )
subject to
g(x1 , x2 , · · · , xn ) = c.
Zλ = c − g(x1 , x2 , · · · , xn ) = 0;
Zi = fi (x1 , x2 , · · · , xn ) − λgi (x1 , x2 , · · · , xn ) = 0 [i = 1, 2, · · · , n].
If the objective function has more than one constraint, say, two con-
straints
g(x1 , x2 , · · · , xn ) = c and h(x1 , x2 , · · · , xn ) = d.
Zλ = c − g(x1 , x2 , · · · , xn ) = 0;
Zµ = d − h(x1 , x2 , · · · , xn ) = 0;
Zi = fi (x1 , x2 , · · · , xn ) − λgi (x1 , x2 , · · · , xn ) − µhi (x1 , x2 , · · · , xn ) = 0.
∂L(x∗ , λ∗ ) ∂f (x∗ ) ∑ m
∂g j (x∗ )
= + λ∗j = 0, i = 1, · · · , n.
∂xi ∂xi i=1 ∂xi
12.3. SECOND-ORDER CONDITIONS 199
From the last section, we know that finding the constrained extremum is
equivalent to find the free extremum of the Lagrange function Z and give
the first-order condition. This section gives the second-order sufficient
condition for the constrained extremum of f .
For a constrained extremum of z = f (x, y), subject to g(x, y) = c, the
second-order necessary-and-sufficient conditions still revolve around the
algebraic sign of the second-order total differential d2 z, evaluated at a s-
tationary point. However, there is one important change. In the present
context, we are concerned with the sign definiteness or semidefiniteness
of d2 z, not for all possible values of dx and dy (not both zero), but only
for those dx and dy values (not both zero) satisfying the linear constraint
gx dx + gy dy = 0.
The second-order sufficient conditions are:
Inasmuch as the (dx, dy) pairs satisfying the constraint gx dx+gy dy = 0 con-
stitute merely a subset of the set of all possible dx and dy, the constrained
sign definiteness is less stringent. In other words, the second-order suffi-
cient condition for a constrained-extremum problem is a weaker condition
than that for a free-extremum problem.
In the following, we shall concentrate on the second-order sufficient
conditions.
200 CHAPTER 12. OPTIMIZATION WITH EQUALITY CONSTRAINTS
z = f (x1 , x2 , · · · , xn )
subject to
g(x1 , x2 , · · · , xn ) = c.
0 g1 g2 ··· gn
g1 Z11 Z12 · · · Z1n
|H̄| = g2 Z21 Z22 · · · Z2n
··· ··· ··· ··· ···
gn Zn1 Zn2 · · · Znn
where in the newly introduced symbols, the horizontal bar above H mean-
s bordered, and Zij = fij − λgij .
f1 f2 fn
λ= = = ··· = .
g1 g2 gn
12.3. SECOND-ORDER CONDITIONS 201
0 g1 g2 g3
0 g1 g2
g1 Z11 Z12 Z13
|H̄2 | = g1 Z11 Z12 , |H̄3 | = (etc.)
g2 Z21 Z22 Z23
g2 Z21 Z22
g3 Z31 Z32 Z33
with the last one being |H̄n | = |H̄|, where the subscript indicates the order
of the leading principal minor being bordered. For instance, |H̄2 | involves
the second leading principal minor of the (plain) Hessian, bordered with
0, g1 , and g2 ; and similarly for the others. The conditions for positive and
negative definiteness of d2 z are then:
d2 z is negative definite subject to dg = 0 iff
In the former, all the bordered leading principal minors, starting with
|H̄2 |, must be negative; in the latter, they must alternate in sign. As previ-
ously, a positive definite d2 z is sufficient to establish a stationary value of
z as its minimum, whereas a negative definite d2 z is sufficient to establish
it as a maximum.
Summarizing the above discussions, we have the following conclu-
sions.
(2) |H̄2 | > 0, |H̄3 | < 0, |H̄4 | > 0, · · · , (−1)n |H̄n | > 0.
0 1 4
|H̄| = 1 2 0 = −34 < 0,
4 0 2
Now consider the most general setting of the problem with n variables and
m equality constraints ("Extremize" means to find either the minimum or
the maximum of the objective function f ):
∑
m
L(x1 , . . . , xn , λ1 , . . . , λm ) = f (x1 , . . . , xn ) + λj (bj − g j (x1 , . . . , xn )),
j=1
204 CHAPTER 12. OPTIMIZATION WITH EQUALITY CONSTRAINTS
∂L(x1 , . . . , xn , λ1 , . . . , λm ) ∂f (x1 , . . . , xn ) ∑m
∂g j (x1 , . . . , xn )
= − λj = 0, i = 1, 2, . . . , n,
∂xi ∂xi j=1 ∂xi
∂L(x1 , . . . , xn , λ1 , . . . , λm )
= bj − g j (x1 , . . . , xn ) = 0, j = 1, 2, . . . , m.
∂λj
Solving these equations for x1 , . . . , xn and λ1 , . . . , λm , we will get a set
of stationary points of the Lagrangian. If x∗ = (x∗1 , . . . , x∗n ) is a solution of
the problem (12.4.1), it should be a stationary point of L.
It is important to assume that rank(J ) = m, and the functions are
continuously differentiable.
If we need to check whether a stationary point results in a maximum
or minimum of the object function, the following local sufficient condition
can be applied:
Algebraic Characterization
Note that when f (v) ≥ f (u), the above inequalities imply respectively
Differentiable Functions
When there are two or more independent variables, the above propo-
sition becomes:
∑
n
∂f (x)
f (y) = f (x) ⇒ (yj − xj ) = 0. (12.5.3)
j=1 ∂xj
0 f1 f2 ··· fn
f1 f11 f12 · · · f1n
|B| = f2 f21 f22 · · · f2n .
··· ··· ··· ··· ···
fn fn1 fn2 · · · fnn
Remark 12.5.3 The determinant |B| is different from the bordered Hes-
sian |H|. Unlike |H|, the border in |B| is composed of the first derivatives
of the function f rather than an extraneous constraint function g.
0 f1 f2
0 f1 ,
|B1 | = , |B2 | = f1 f11 f12 , · · · , |Bn | = |B|.
f1 f11
f2 f21 f22
210 CHAPTER 12. OPTIMIZATION WITH EQUALITY CONSTRAINTS
0 x2 x1
0 x2
|B1 | = = −x22 < 0, |B2 | = x2 0 1 = 2x1 x2 > 0.
x2 0
x1 1 0
thus
0 fx
|B1 | = = −(axa−1 y b )2 < 0;
fx fxx
0 fx fy
|B2 | = fx fxx fxy = [2a2 b2 − a(a − 1)b2 − a2 b(b − 1)]x3a−2 y 3b−2
fy fyx fyy
|B| = λ2 |H̄|.
Theorem 12.5.1 (Global Optimum) Suppose that f is concave and the con-
straint function is convex. They are both are twice continuously differentiable
function on X ⊆ Rn , and x∗ is an interior point of X. Then, the following three
statements are equivalent:
Px x + Py y = I
First-Order Condition
Zλ = I − Px x − Py y = 0;
Zx = ux − λPx = 0;
Zy = uy − λPy = 0.
Figure 12.4: The graphical illustration of the conditions for utility maxi-
mization.
ux uy
= = λ,
Px Py
or
ux Px
= .
uy Py
The term ux
uy
≡ M RSxy is the so-called marginal rate of substitution of
Px
x for y. Thus, we obtain the well-known equality: M RSxy = Py
which is
the necessary condition for the interior solution.
214 CHAPTER 12. OPTIMIZATION WITH EQUALITY CONSTRAINTS
Second-Order Condition
0 Px Py
|H̄| = Px uxx uxy = 2Px Py uxy − Py2 uxx − Px2 uyy > 0,
Py uyx uyy
(with all the derivatives evaluated at the stationary point of x̄ and ȳ), then
the stationary value of u will assuredly be maximum.
Since the budget constraint is linear, from the result in the last section,
we have
|B| = λ2 |H̄|.
Px
From M RSxy = Py
, we may solve x or y as a function of another and
then substitute it into the budget line to find the demand function of x or
y.
12.6. UTILITY MAXIMIZATION AND CONSUMER DEMAND 215
ay Px
=
(1 − a)x Py
and then
(1 − a)Px x
y= .
aPy
Substituting the above y into the budget line Px x + Py y = I and solving
for x, we get the demand function for x
aI
x(Px , Py , I) = .
Px
Substituting the above x(Px , Py , I) into the budget line, the demand
function for y is obtained:
(1 − a)I
y(Px , Py , I) = .
Py
216 CHAPTER 12. OPTIMIZATION WITH EQUALITY CONSTRAINTS
Chapter 13
217
218CHAPTER 13. OPTIMIZATION WITH INEQUALITY CONSTRAINTS
max f (x1 , . . . , xn )
s.t. g i (x1 , . . . , xn ) ≤ bi , i = 1, 2, . . . , m;
x1 ≥ 0, . . . , xn ≥ 0.
min f (x1 , . . . , xn )
s.t. g i (x1 , . . . , xn ) ≥ bi , i = 1, 2, . . . , m;
x1 ≥ 0, . . . , xn ≥ 0.
First, note that there are no restrictions on the relative size of m and n,
unlike the case of equality constraints. Second, note that the direction of
the inequalities (≤ or ≥) at the constraints is only a convention, because
the inequality g i ≤ bi can be easily converted to the ≥ inequality by multi-
plying it by -1, yielding −g i ≥ −bi . Third, note that an equality constraint,
say g k = bk , can be replaced by the two inequality constraints, g k ≤ bk and
−g k ≤ −bk .
For the purpose of ruling out certain irregularities on the boundary of the
feasible set, a restriction on the constrained functions is imposed. This re-
striction is the so-called constraint qualification. The following is a strong
version of the constraint qualification, which is much easier to verify.
Definition 13.2.1 Let C be the constraint set. We say that the (linear in-
dependenc) constraint qualification condition is satisfied at x∗ ∈ C if the
constraints that hold at x∗ with equality are independent; that is, if the gra-
dients (the vectors of partial derivatives) of g j -constraints that are valuated
and binding at x∗ are linearly independent for j = 1, . . . , m.
∑
m
L(x1 , . . . , xn , λ1 , . . . , λm ) = f (x1 , . . . , xn ) + λj (bj − g j (x1 , . . . , xn )).
j=1
∂L ∂L
≤ 0, xi ≥ 0 and xi = 0, i = 1, . . . , n;
∂xi ∂xi
∂L ∂L
≥ 0, λj ≥ 0 and λj = 0, j = 1, . . . , m.
∂λj ∂λj
220CHAPTER 13. OPTIMIZATION WITH INEQUALITY CONSTRAINTS
∂L ∂L
≥ 0, xi ≥ 0 and xi = 0, i = 1, . . . , n;
∂xi ∂xi
∂L ∂L
≤ 0, λj ≥ 0 and λj = 0, j = 1, . . . , m.
∂λj ∂λj
∂L
= 10 − 2x1 − 5λ1 − λ2 ≤ 0;
∂x1
∂L
= 20 − 2x2 − 3λ1 − λ2 ≤ 0;
∂x2
13.2. KUHN-TUCKER CONDITIONS 221
∂L
= −(5x1 + 3x2 − 40) ≥ 0;
∂λ1
∂L
= −(x1 − 5) ≥ 0;
∂λ2
∂L
= −(x2 − 10) ≥ 0;
∂λ3
x1 ≥ 0, x2 ≥ 0;
λ1 ≥ 0, λ2 ≥ 0, λ3 ≥ 0;
∂L ∂L
x1 = 0, x2 = 0;
∂x1 ∂x2
∂L
λi = 0, i = 1, 2, 3.
∂λi
Notice that the failure of the constraint qualification signals certain ir-
regularities at the boundary kinks of the feasible set. Only if the optimal
solution occurs in such a kink may the Kuhn-Tucker condition not be sat-
isfied.
If all constraints are linear and functionally independent, the constraint
qualification is always satisfied.
Example 13.2.3 The following example illustrates a case where the Kuhn-
Tucker condition is not satisfied in the solution of an optimization prob-
lem. Consider the problem:
max y
s.t. x + (y − 1)3 ≤ 0;
x ≥ 0, y ≥ 0.
222CHAPTER 13. OPTIMIZATION WITH INEQUALITY CONSTRAINTS
The solution to this problem is (0, 1). (If y > 1, then the restriction
x + (y + 1)3 ≤ 0 implies x < 0.) The Lagrangian function is:
L = y + λ[−x − (y − 1)3 ].
∂L
≤ 0,
∂y
or
1 − 3λ(y − 1)2 ≤ 0.
As can be observed, this condition is not verified at the point (0, 1).
∂f (x∗ )
(c.i) there exists j such that ∂xj
< 0 (resp. > 0);
∂f (x∗ )
(c.ii) there exists j such that ∂xj
> 0 (resp. < 0)
and xj can take a positive value without violating
the constraints;
Proposition 13.2.4 If (x∗ , λ∗ ) solves the saddle point problem then (x∗ , λ∗ )
solves the problem (13.1.1).
224CHAPTER 13. OPTIMIZATION WITH INEQUALITY CONSTRAINTS
Since the marginal rate of substitution of x for y is a/b and the economic
rate of substitution of x for y is px /py are both constant, they cannot be
in general equal. So the first-order condition cannot hold with equality as
long as a
b
̸= px
py
. In this case the answer to the utility-maximization problem
typically involves a boundary solution: only one of the two goods will
be consumed. It is worthwhile presenting a more formal solution since
it serves as a nice example of the Kuhn-Tucker theorem in action. The
Kuhn-Tucker theorem is the appropriate tool to use here, since we will
almost never have an interior solution.
The Lagrange function is
and thus
∂L
= a − λpx ; (13.3.1)
∂x
∂L
= b − λpt ; (13.3.2)
∂y
∂L
= m − px − py . (13.3.3)
∂λ
x x x
x x x
Remark 13.3.1 In fact, it is easily found out the optimal solutions by com-
paring relatives steepness of the indifference curves and the budget line.
a px
For instance, as shown in Figure 13.1 below, when b
> py
, the indiffer-
ence curves become steeper, and thus the optimal solution is the one the
226CHAPTER 13. OPTIMIZATION WITH INEQUALITY CONSTRAINTS
a px
consumer spends his all income on good x . When b
< py
, the indiffer-
ence curves become flatter, and thus the optimal solution is the one the
consumer spends his all income on good y. When a
b
̸= px
py
, the indiffer-
ence curves and the budget line are parallel and coincide at the optimal
solutions, and thus the optimal solutions are given by all the points on the
budget line.
• fj = ∂f
∂xj
is the marginal profit (revenue) of product j;
∂g i
• gji = ∂xj
is the amount of resource i used in producing a marginal
unit of product j;
∑
m
The condition ∂L
∂xj
≤ 0 can be written as fj ≤ λi gji and it says that the
i=1
marginal profit of the jth product cannot exceed the aggregate marginal
imputed cost of the jth product.
∂L
The Kuhn-Tucker condition xj ∂x j
= 0 implies that, in order to produce
good j (xj > 0), the marginal profit of good j must be equal to the ag-
∂L
gregate marginal imputed cost ( ∂x j
= 0). The same condition shows that
∂L
good j is not produced (xj = 0) if there is an excess imputation xj ∂x j
< 0.
The Kuhn-Tucker condition ∂L
∂λi
≥ 0 is simply a restatement of con-
straint i, which states that the total amount of resource i used in producing
all the n goods should not exceed total amount available ri .
∂L
The condition ∂λi
= 0 indicates that if a resource is not fully used in
∂L
the optimal solution ( ∂λ i
> 0), then its shadow price will be 0 (λi = 0). On
∂L
the other hand, a fully used resource ( ∂λ i
= 0) has a strictly positive price
(λi > 0).
A firm has to produce two goods using three kinds of resources avail-
able in the amounts 40, 5, 10 respectively. The first resource is used in the
production of both goods: five units are necessary to produce one unit of
good 1, and three units to produce one unit of good 2. The second re-
source is used only in producing good 1 and the third resource is used
only in producing good 2.
The sale prices of the two goods are given by the linear inverse demand
equations p1 = 10 − x1 and p2 = 20 − x2 . The problem the firm faces is
how much to produce of each good in order to maximize revenue R =
x1 p1 + x2 p2 . The solution (2, 10) gives the optimal amounts the firm should
produce.
Chapter 14
Differential Equations
229
230 CHAPTER 14. DIFFERENTIAL EQUATIONS
C 1 4 dy y
+ x is the solution of the ordinary differential equation + = x3 ,
x 5 dx x
where C is an arbitrary constant. Next we introduce the concept of general
solutions and particular solutions of ordinary differential equations.
∂ψ ∂ψ ∂ψ
···
∂C1 ∂C2 ∂Cn
∂ψ (1) ∂ψ (1)
∂ψ (1)
D[ψ, ψ (1) , · · · , ψ (n−1) ] def ···
= ∂C1 ∂C2 ∂Cn
D[C1 , · · · , Cn ] .. .. .. ..
. . . .
(n−1) (n−1) (n−1)
∂ψ ∂ψ ∂ψ
···
∂C1 ∂C2 ∂Cn
(1) (n−1)
y(x0 ) = y0 , y (1) (x0 ) = y0 , · · · , y (n−1) (x0 ) = y0 , (14.0.3)
then the ordinary differential equation (14.0.1) and the initial value con-
ditions (14.0.3) are said to be the Cauchy problem or initial value prob-
14.1. EXISTENCE AND UNIQUENESS THEOREM OF SOLUTIONS FOR ORDINARY DIFFEREN
lem for the nth-order ordinary differential equations. Then the question is
what conditions the function F should satisfy so that the above ordinary
differential equations are uniquely solvable. This problem is the existence
and uniqueness of solutions for ordinary differential equations.
The following is the theorem on the uniqueness of the solution for dif-
ferential equations.
For example, for (x2 + 1)y ′ + 2xy 2 = 0, y(0) = 1, using the above solving
procedure, we get the solution as
1
y(x) = .
ln(x2 + 1) + 1
dy
+ p(x)y = q(x). (14.2.4)
dx
Suppose that
∫
y = c(x)e− p(x)dx
,
∫ ∫
y ′ = c′ (x)e− p(x)dx
+ c(x)p(x)e− p(x)dx
,
then substituting this back into the original differential equation, we have
∫ ∫ ∫
c′ (x)e− p(x)dx
+ c(x)p(x)e− p(x)dx
= p(x)c(x)e− p(x)dx
+ q(x),
and thus
∫
′ p(x)dx
c (x) = q(x)e .
We have
∫ ∫
p(x)dx
c(x) = q(x)e dx + C.
∫ (∫ ∫ )
y(x) = e− p(x)dx
q(x)e p(x)dx
dx + C .
Bernoulli Equation
dy
+ p(x)y = q(x)y n , (14.2.5)
dx
dy
(1 − n)y (−n) + (1 − n)y (1−n) p(x) = (1 − n)q(x).
dx
236 CHAPTER 14. DIFFERENTIAL EQUATIONS
dz
+ (1 − n)zp(x) = (1 − n)q(x),
dx
The differential equations with explicit solutions have other forms, such
as some special forms of Ricatti equations, and the equations similar to
M (x, y)dx + N (x, y)dy = 0, but not satisfying
∂M (x, y) ∂N (x, y)
≡ .
∂y ∂x
isfying
λn + a1 λn−1 + · · · + an−1 λ + an = 0.
yi1 (x) = eλi x , yi2 (x) = xeλi x , · · · , yik (x) = xk−1 eλi x .
yj1 = eαj x cos βj x, yj2 = xeαj x cos βj x, · · · , yjl = xl−1 eαj x cos βj x;
yjl+1 = eαj x sin βj x, yjl+2 = xeαj x sin βj x, · · · , yj2l = xl−1 eαj x sin βj x.
yp (x) = xs Qk (x)ebx ,
(4) If f (x) = f1 (x) + f2 (x) + · · · + fr (x), and yp1 (x), · · · , ypr (x) are the
particular solutions corresponding to f1 (x), · · · , fr (x), then
6a = 1, −5 × 2a + 6b = 0, 2a − 5b + 6c = −5, d − 5d + 6d = 1,
varying vector.
Consider the case that A is a constant coefficient matrix and b is a con-
stant vector, also called the system of differential equations with con-
stant coefficients:
x(t) = xc (t) + xp .
There are two methods for solving the system of homogeneous differential
equations (14.4.8).
The first one is that we can eliminate n − 1 dependent variables so that
the system of differential equations becomes the differential equation of
order n, such as the following example.
ẋ = 2x + y,
ẏ = 3x + 4y.
thus the general solution is x(t) = C1 et + C2 e5t . Since y(t) = ẋ − 2x, y(t) =
−C1 et + 3C2 e5t .
x(t) = eAt x0 ,
where
A2 t2
eAt = I + At + + ··· .
2!
Now we solve eAt in three different cases.
Matrix A has different real eigenvalues, which means that its eigenvec-
tors are linearly independent. Thus A can be diagonalized, that is,
A = P ΛP −1 ,
x(t) = P eΛt P −1 x0
= P eΛt c
= c 1 v 1 e λ1 t + · · · + c n v n e λn t ,
First, consider a simple case that A has only one eigenvalue of multiplic-
ity m. In this case, there are at most m linearly independent eigenvectors,
which means that the matrix P can not be constructed as a matrix consist-
ing of linearly independent eigenvectors, so A can not be diagonalized.
Thus, the solution has the following form:
∑
m
x(t) = ci hi (t),
i=1
where hi (t), ∀i, are quasi-polinomials, and ci , ∀i, are determined by initial
conditions. For example, when m = 3, we have:
h1 (t) = eλt v1 ,
h2 (t) = eλt (tv1 + v2 ),
h3 (t) = eλt (t2 v1 + 2tv2 + 3v3 ),
(A − λI)vi = vi−1 , v0 = 0.
If A has more than one multiple real eigenvalues, then the solution of the
244 CHAPTER 14. DIFFERENTIAL EQUATIONS
Now consider a simple case: A has only one pair of complex eigenval-
ues, λ1 = α + βi and λ2 = α − βi.
x(t) = eAt x0
= P eΛt P −1 x0
= P eΛt c
= c1 v1 e(α+βi)t + c2 v2 e(α−βi)t
= c1 v1 eαt (cos βt + i sin βt) + c2 v2 eαt (cos βt − i sin βt)
= (c1 v1 + c2 v2 )eαt cos βt + i(c1 v1 − c2 v2 )eαt sin βt
= h1 eαt cos βt + h2 eαt sin βt,
def
V̇ (x) = ▽V (x)f (t, x) ≤ 0, ∀x ∈ Q, (14.6.10)
Theorem 14.6.1 If there exists a Lyapunov function V for the dynamic system
(14.5.9), then the equilibrium point x∗ is globally stable.
If the Lyapunov function (14.6.10) of the dynamic system satisfies V̇ (x) <
0, ∀x ∈ Q, x ̸= x∗ , then the equilibrium point x∗ is asymptotically globally
stable.
Chapter 15
Difference Equations
249
250 CHAPTER 15. DIFFERENCE EQUATIONS
If the coefficients of y0 (k), yn (k) are not zero, and the highest correspond-
ing order is n, then it is called an nth-order difference equation.
y(t) = Y + y ∗ .
y(t + 1) + ay(t) = 0,
r
y∗ = , a ̸= −1,
1+a
y ∗ = rt, a = −1.
∑
t−1
y∗ = (−a)t−1−i r(i),
i=0
∑
t−1
y(t) = (−a)t y0 + (−a)t−1−i r(i), t = 1, 2, · · · .
i=0
y(t + 1) − 3y(t) = t2 + t + 2.
y(t + 1) − 3y(t) = 0,
or
−2At2 + 2(A − B)t + A + B − 2D = t2 + t + 2.
1 3
which gives A = − , B = −1 and D = − , thus we have a particular
2 4
∗ 1 2 3
solution: y = − t − t − . Therefore, a particular solution of the nonho-
2 4
1 3
mogeneous equation is y(t) = Y + y ∗ = C3t − t2 − t − .
2 4
254 CHAPTER 15. DIFFERENCE EQUATIONS
We can also solve the case with an exponential function by using the
method of undetermined-coefficients.
Here are some of the common ways for finding particular solutions:
Then, its general solution depends on the roots of the following linear
equation:
m2 + a1 m + a2 = 0,
Case 3: m1 and m2 are two complex roots, namely r(cos θ±i sin θ) with r >
0, θ ∈ (−π, π]. The general solution of the homogeneous equation is
Y = C1 rt cos(tθ + C2 ).
mn + a1 mn−1 + · · · + an−1 m + an = 0.
C1 rt cos(tθ + C2 ).
rt [C1,1 cos(tθ + C1,2 ) + C2,1 t cos(tθ + C2,2 ) + · · · + Cp,1 tp−1 cos(tθ + Cp,2 )].
∑
n ∞
∑
y∗ = θs mis r(t − i),
s=1 i=0
where
ms
θs = .
Πj̸=s (ms − mj )
Theorem 15.4.1 Suppose that the modulus of all eigenvalues of the character-
istic equation are less than 1. Then, the difference equation (15.4.9) is asymptoti-
cally stable.
When the following inequality conditions are satisfied, the modulus of all
eigenvalues of the characteristic equation are less than 1.
1 an
> 0,
an 1
1 0 an an−1
a1 1 0 an
> 0,
an 0 1 a1
an−1 an 0 1