0% found this document useful (0 votes)
16 views13 pages

Joint Probability Distribution Reference 1

Uploaded by

jiero.fernandez
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
16 views13 pages

Joint Probability Distribution Reference 1

Uploaded by

jiero.fernandez
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 13

6.

2 Distribution Functions of Two Random Variables 229

6.1 INTRODUCTION
In Chapters 4 and 5, we discussed various phenomena by enlisting a (single) random
variable and studying its distribution and characteristics. However, in many situations,
experiments are performed that involve two or more random variables. For example, we
may be interested in the diameter and length of rods, the number of dots on two dice when
rolled simultaneously, say (X, Y ), where 1 ≤ X ≤ 6, 1 ≤ Y ≤ 6, or the composition of a
Monel (70% nickel, 30% copper) alloy, where we may focus on solid contents, say X, and
liquid content Y, which again we would quote as a joint pair (X, Y ).
In this chapter, then, we will study the joint distribution functions of two or more
discrete and continuous random variables.

6.2 DISTRIBUTION FUNCTIONS OF TWO


RANDOM VARIABLES
6.2.1 Case of Two Discrete Random Variables
If, for each element e in a finite sample space S, we make two measurements on e,
say (X(e), Y (e)), and if (xi , yj ), i = 1, 2, . . . , m and j = 1, 2, . . . , n, are possible values
of (X(e), Y (e)), and if we let

pij = p(xi , yj ) = P (X(e) = xi , Y (e) = yj ) (6.2.1)

then the set of all possible values {(xi , yj )} of (X(e), Y (e)) is called the sample space of
(X(e), Y (e)), while the set of associated probabilities pij is the joint probability function
(p.f.) of the pair of discrete random variables (X, Y ).
Thus, we may think of k = mn points (xi , yj ) in the xy-plane in which the probabilities
pij are located and are all positive and sum to 1. If we define pi· and p·j such that


n 
m
pi. = pij and p.j = pij (6.2.2)
j=1 i=1

then
pi. = P (X(e) = xi ) and p.j = P (Y (e) = yj ) (6.2.3)

The possible values xi , i = 1, 2, . . . , m, of X(e) together with their probabilities pi. consti-
tute the marginal distribution of the random variable X. This gives rise to the probability
function of X, ignoring Y, and is therefore merely the probability function of X. In a simi-
lar manner, the yj , j = 1, 2, . . . , n, of Y (e) together with their probabilities p.j constitute
the marginal distribution of the random variable Y.
Geometrically, if x is the usual horizontal axis and y the vertical axis and if we project
the sum of the probabilities pi1 , . . . , pij , . . . , pin located at the points [(xi , y1 ), . . . , (xi , yj ),
. . . , (xi , yn )], vertically onto the x-axis, we obtain the marginal distribution pi. of the ran-
dom variable X. If instead we project sum of these probabilities p1j , . . . , pij , . . . , pmj hori-
zontally onto the y-axis, we obtain the marginal distribution p.j of the random variable Y.
The mean μ1 and variance σ12 of X are defined by applying (4.2.1) and (4.2.2) to the
probability distribution pi. . Similarly, the mean μ2 and variance σ22 of Y are defined by
applying those formulas to p.j .
230 6 Distribution of Functions of Random Variables

When the probability function pij factors into the product of the two marginal
probability functions, that is, if for all possible (xi , yj ) in the sample space of (X, Y ),
we have
pij = pi. p.j (6.2.4)

then X and Y are said to be independent random variables.

Example 6.2.1 (Probability function of two random variables) Roll a pair of fair dice,
of which one die is green and the other is red. Let the random variables X and Y denote
the outcomes on the green and red dies, respectively. Then, the sample space of (X, Y ) is
S = {(1, 1), (1, 2), . . . , (1, 6), . . . , (6, 6)}. Each of the 36 sample points has the probability
1/36. Then, the joint probability function of the random variables X and Y can be written
in tabular form as follows:

X 1 2 3 4 5 6 Total (p·j )
Y
1 1/36 1/36 1/36 1/36 1/36 1/36 1/6
.. .. .. ..
2 1/36 . . . . 1/6
3 1/36 ··· 1/36 ··· 1/36 1/36 1/6
.. .. .. ..
4 1/36 . . . . 1/6
5 1/36 ··· 1/36 ··· 1/36 1/36 1/6
6 1/36 ··· 1/36 ··· 1/36 1/36 1/6
Total (pi· ) 1/6 1/6 1/6 1/6 1/6 1/6 1

This table shows the probabilities assigned to each sample point. Using (6.2.2) for the
probabilities, we easily find the marginal distributions pi· and p·j of the random variables
X and Y, respectively, as shown in the table. The probability function in this example can
also be expressed as

1/36, x = 1, 2, 3, 4, 5, 6 and y = 1, 2, 3, 4, 5, 6
p(x, y) =
0, otherwise

We give a graphical representation of the p.f. of X and Y in Figure 6.2.1.

Example 6.2.2 (Marginal probability functions) Let the joint probability function of
random variables X and Y be defined as
x+y
P (x, y) = , x = 1, 2, 3, 4; y = 1, 2, 3
54
Find the marginal probability functions of X and Y and also examine whether X and Y
are independent.
6.2 Distribution Functions of Two Random Variables 231

0.033

0.030
p(x, y)
0.027

6
0.024 5
4
3 y
2
1 2 1
3 4 5 6
x

Figure 6.2.1 Graphical representation of the p.f. in Example 6.2.1.

Solution: From equation (6.2.2), it follows that the probability of X, say p1 (x), is given by

3
x+y x+1 x+2 x+3 3x + 6
P (X = x) = p1 (x) = = + + = , for x = 1, 2, 3, 4
y=1
54 54 54 54 54

Similarly,

4
x+y 10 + 4y
P (Y = y) = p2 (y) = = , for y = 1, 2, 3
x=1
54 54

For (x, y) belonging to the sample space of (X, Y ), say Sxy = {(x, y)|x = 1, 2, 3, 4; y =
1, 2, 3}, we have that
p(x, y) = p1 (x) × p2 (y)

so that the random variables X and Y are not independent.

Example 6.2.3 (Joint probability function and its marginals) In dealing a hand of 13
cards from a deck of ordinary playing cards, let X1 and X2 be random variables denoting
the numbers of spades and of hearts, respectively. Obviously, 0 ≤ X1 ≤ 13, 0 ≤ X2 ≤ 13,
and 0 ≤ X1 + X2 ≤ 13. Then, we see that p(x1 , x2 ), the p.f. of (x1 , x2 ), is given by
   
13 13 26
x1 x2 13−x1 −x2
p(x1 , x2 ) =  52 
13

where the sample space of (X1 , X2 ) is all pairs of nonnegative integers (x1 , x2 ) for which
0 ≤ x1 , x2 ≤ 13 and 0 ≤ x1 + x2 ≤ 13. That is, the sample space {(x1 , x2 )} consists of the
105 points:
{(0, 0), . . . , (0, 13), . . . , (12, 0), (12, 1), (13, 0)}

Now, it is possible by a direct probability argument to find the marginal distribution


of X1 , for the probability of x1 spades in a hand of 13 is clearly given by
  
13 39
x1 13−x1
P (X1 = x1 ) = p1 (x1 ) =  52 
13

where 0 ≤ x1 ≤ 13.
232 6 Distribution of Functions of Random Variables

In a similar manner, it is easy to find p2 (x2 ) and to show that the random variables
X1 and X2 are not independent.

6.2.2 Case of Two Continuous Random Variables


If the sample space S consists of a continuum of elements and if for any point (x1 , x2 ) in
the x1 x2 -plane we let

F (x1 , x2 ) = P [X1 (e) ≤ x1 , X2 (e) ≤ x2 ] (6.2.5)

then F (x1 , x2 ) is called the cumulative distribution function (c.d.f.) of the pair of random
variables (X1 , X2 ) (dropping e). If there exists a nonnegative function f (x1 , x2 ) such that
 x1  x2
F (x1 , x2 ) = f (t1 , t2 )dt2 dt1 (6.2.6)
−∞ −∞

then
∂ 2 F (x1 , x2 )
f (x1 , x2 ) =
∂x1 ∂x2
and f (x1 , x2 ) is called the joint probability density function (p.d.f.) of the pair of random
variables (X1 , X2 ). The probability that this pair of random variables represents a point
in a region E, that is, the probability that the event E occurs, is given by

P ((X1 , X2 ) ∈ E) = f (x1 , x2 )dx2 dx1 (6.2.7)
E

Note that if E = {(X1 , X2 )|X1 < x1 , X2 < x2 }, then (6.2.7) equals F (x1 , x2 ). Also, if we let

 ∞
f1 (x1 ) = f (x1 , x2 )dx2 (6.2.8)
−∞

f2 (x2 ) = f (x1 , x2 )dx1 (6.2.9)
−∞

then f1 (x1 ) and f2 (x2 ) are called the marginal probability density functions of X1 and X2 ,
respectively. This means that f1 (x1 ) is the p.d.f. of X1 (ignoring X2 ), and f2 (x2 ) is the
p.d.f. of X2 (ignoring X1 ).
Geometrically, if we think of f (x1 , x2 ) as a function describing the manner in which
the total probability 1 is continuously “smeared” in the x1 x2 -plane, then the integral in
(6.2.7) represents the amount of probability contained in the region E. Also, f1 (x1 ) is the
p.d.f. one obtains by projecting the probability density in the x1 x2 -plane orthogonally onto
the x1 -axis, and f2 (x2 ) is similarly obtained by orthogonal projection of the probability
density onto the x2 -axis.

If f (x1 , x2 ) factors into the product of the two marginal p.d.f.’s, that is, if

f (x1 , x2 ) = f1 (x1 )f2 (x2 ) (6.2.10)

for all (x1 , x2 ) in the sample space of (X1 , X2 ), then X1 and X2 are said to be
independent continuous random variables.
6.2 Distribution Functions of Two Random Variables 233

Example 6.2.4 (Marginal probability functions) Let the joint probability density function
of the random variables X1 and X2 be defined as

f (x1 , x2 ) = 2e−(2x1 +x2 ) ; x1 > 0, x2 > 0

Find the marginal probability density functions of X1 and X2 and examine whether or not
X1 and X2 are independent.

Solution: From equations (6.2.8) and (6.2.9), it follows that for x1 > 0
 ∞  ∞
f1 (x1 ) = 2e−(2x1 +x2 ) dx2 = 2e−2x1 e−x2 dx2
0 0

= 2e−2x1 [−e−x2 ]∞
0 = 2e
−2x1
, x1 > 0

while for x2 > 0,


 ∞  ∞
f2 (x2 ) = 2e−(2x1 +x2 ) dx1 = e−x2 2e−2x1 dx1
0 0

2e−2x1
= e−x2 = e−x2 , x2 > 0
−2 0

Here, we clearly have that f (x1 , x2 ) = f1 (x1 )f2 (x2 ), which implies that the random vari-
ables X1 and X2 are independent.
Finally, note that the joint distribution function satisfies the properties given below:

1. 0 ≤ F (x1 , x2 ) ≤ 1 for all (x1 , x2 ) belong to the sample space of (X1 , X2 ).


2. F (−∞, x2 ) = F (x1 , −∞) = F (−∞, ∞) = F (∞, −∞) = 0, F (∞, ∞) = 1.
3. F is nondecreasing.
4. For every pair of (X1 , X2 ) values, say xi1 and xi2 where xi1 < xi2 for i = 1, 2,
the following inequality holds:

F (x12 , x22 ) − F (x12 , x21 ) − F (x11 , x22 ) + F (x11 , x21 ) ≥ 0 (6.2.11)

The reader should verify that the left-hand side of (6.2.11) gives the value of P (x11 <
X1 < x12 , x21 < X2 < x22 ).

6.2.3 The Mean Value and Variance of Functions


of Two Random Variables
Suppose that (X1 , X2 ) is a pair of discrete random variables and g(X1 , X2 ) is a function of
(X1 , X2 ). Then, the mean value or expectation of g(X1 , X2 ), say E(g(X1 , X2 )), is given by

E(g(X1 , X2 )) = g(x1i , x2j )p(x1i , x2j ) (6.2.12)
234 6 Distribution of Functions of Random Variables

where the summation is over all pairs (x1i , x2j ) in the sample space of (X1 , X2 ), and
for the continuous case we have that
 ∞ ∞
E(g(X1 , X2 )) = g(x1 , x2 )f (x1 , x2 )dx1 dx2 (6.2.13)
−∞ −∞

We may now state, and the reader should verify equation (6.2.14) in Theorem 6.2.1
stated below.

Theorem 6.2.1 If X1 and X2 are independent random variables and if g1 (X1 )


and g2 (X2 ) depend only on X1 and X2 , respectively, then

E(g1 (X1 ) g2 (X2 )) = E(g1 (X1 )) E(g2 (X2 )) (6.2.14)

If we choose g(X1 , X2 ) as (X1 − μ1 )(X2 − μ2 ), we obtain the covariance, which is a


measure of the relationship between two random variables X1 and X2 , that is,

Cov(X1 , X2 ) = E[(X1 − μ1 )(X2 − μ2 )] (6.2.15)

In the case where X1 and X2 are independent, we find that

Cov(X1 , X2 ) = E(X1 − μ1 )E(X2 − μ2 ) = 0 (6.2.16)

In many problems, however, we deal with linear functions of two or even more independent
random variables. The following theorem is of particular importance in this connection:

Theorem 6.2.2 Let X1 and X2 be independent random variables such that the
mean and variance of X1 are μ1 and σ12 , and the mean and variance of X2 are μ2
and σ22 . Then, if c1 and c2 are constants, c1 X1 + c2 X2 is a random variable having
mean value c1 μ1 + c2 μ2 and variance c21 σ12 + c22 σ22 .

Proof: To prove this theorem, it is sufficient to consider the case of continuous random
variables. (The proof for discrete random variables is obtained by replacing integral signs
by signs of summation.) For the mean value of c1 X1 + c2 X2 , we have, since X1 and X2 are
independent, that
 ∞ ∞
E(c1 X1 + c2 X2 ) = (c1 x1 + c2 x2 )f1 (x1 )f2 (x2 )dx1 dx2
−∞ −∞
 ∞  ∞  ∞  ∞
= c1 x1 f1 (x1 )dx1 f2 (x2 )dx2 + c2 x2 f2 (x2 )dx2 f1 (x1 )dx1
−∞ −∞ −∞ −∞

= c1 E(X1 ) + c2 E(X2 )
= c1 μ1 + c2 μ2
6.2 Distribution Functions of Two Random Variables 235

For the variance of c1 X1 + c2 X2 , we have similarly (omitting some straightforward


details), that

V ar(c1 X1 + c2 X2 ) = E[c1 (X1 − μ1 ) + c2 (X2 − μ2 )]2


= c21 E(X1 − μ1 )2 + c22 E(X2 − μ2 )2 + 2c1 c2 E[(X1 − μ1 )(X2 − μ2 )]
= c21 E(X1 − μ1 )2 + c22 E(X2 − μ2 )2 + 0
= c21 σ12 + c22 σ22

since X1 and X2 are independent, so that E[(X1 − μ1 )(X2 − μ2 )] = 0, as stated in (6.2.16).


We remark that if X1 , X2 are not independent (and the reader should verify), then we
have that

E(c1 X1 + c2 X2 ) = c1 μ1 + c2 μ2
V ar(c1 X1 + c2 X2 ) = c21 σ12 + c22 σ22 + 2c1 c2 Cov(X1 , X2 )


In a straightforward manner, it is easy to prove the following theorem, which extends


the results of this section:

Theorem 6.2.3 Let X1 , X2 , . . . , Xn be n random variables such that the mean and
variance of Xi are μi and σi2 respectively, and where the covariance of Xi and Xj
is σij , that is, E[(Xi − μi )(Xj − μj )] = σij , i = j. If c1 , c2 , . . . , cn are constants,
then the random variable L = c1 X1 + · · · + cn Xn has mean value and variance that
are given by
E(L) = c1 μ1 + · · · + cn μn (6.2.17)

V ar(L) = c21 σ12 + · · · + c2n σn2 + 2c1 c2 σ12 + 2c1 c3 σ13 + · · · + 2cn−1 cn σn−1,n (6.2.18)

Further, if X1 , X2 , . . . , Xn are mutually independent, then σij = 0, so that the


mean of L is as in (6.2.17). However, the variance of L is

V ar(L) = c21 σ12 + · · · + c2n σn2 (6.2.19)

6.2.4 Conditional Distributions


Suppose that a pair of discrete random variables (X1 , X2 ) has joint p.f. p(x1 , x2 ) and
marginal probability functions p1 (x1 ) and p2 (x2 ), as defined in Section 6.2.1. Suppose
that we assign to one of the random variables, say X1 , a value x1 such that p1 (x1 ) = 0,
and we want to find the probability that the other random variable X2 has a particular
value, say x2 . The required probability is a conditional probability that we may denote by
p(X2 = x2 |X1 = x1 ), or, more briefly, by p(x2 |x1 ), and is defined as follows:

p(x1 , x2 )
p(x2 |x1 ) = (6.2.20)
p1 (x1 )

where p1 (x1 ) = 0.
236 6 Distribution of Functions of Random Variables

Note that p(x2 |x1 ) has all the properties of an ordinary probability function; that is,
as the reader should verify, the sum of p(x2 |x1 ) over all possible values of x2 , for fixed x1 ,
is 1. Thus, p(x2 |x1 ), x2 = x21 , . . . , x2k2 , is a p.f. and is called the conditional probability
function of X2 , given that X1 = x1 .
Note that we can write (6.2.20) as

p(x1 , x2 ) = p1 (x1 ) · p(x2 |x1 ) (6.2.21)

to provide a two-step procedure for finding p(x1 , x2 ) by first determining p1 (x1 ), then
p(x2 |x1 ), and by multiplying the two together.

Example 6.2.5 (Conditional probability function) In Example 6.2.3, suppose that we


want to find the conditional probability function of X2 given X1 = x1 , that is, p(x2 |x1 ).

Solution: The probability function of X1 is given by


  
13 39
x1 13−x1
p1 (x1 ) =  52 
13

Hence, as is easily verified, p(x2 |x1 ) is given by


  
13 26
x2 13−x1 −x2
p(x2 |x1 ) =  
39
13−x1

where the sample space of X2 , given X1 = x1 , is {0, 1, . . . , 13 − x1 }. The interpretation of


p(x2 |x1 ) is that if we are given that a hand of 13 cards contains x1 spades, then the value of
p(x2 |x1 ) as given previously is the probability that the hand also contains X2 = x2 hearts.
In the case of a pair of continuous random variables (X1 , X2 ) having probability den-
sity function f (x1 , x2 ) and marginal probability density functions f1 (x1 ) and f2 (x2 ), the
conditional probability density function f (x2 |x1 ) of X2 given X1 = x1 is defined as

f (x1 , x2 )
f (x2 |x1 ) = (6.2.22)
f1 (x1 )

where f1 (x1 ) = 0, which is the analogue of (6.2.20), now for a pair of continuous random
variables. Note that f (x2 |x1 ) has all the properties of an ordinary probability density
function.
Now from (6.2.22), we have the result that is given below.

f (x1 , x2 ) = f1 (x1 ) f (x2 |x1 ) (6.2.22a)


6.2 Distribution Functions of Two Random Variables 237

We now have the analogue of (6.2.21) for obtaining the probability density function of a
pair of continuous random variables in two steps, as given in (6.2.22a).

Example 6.2.6 (Determination of marginal probability functions) Suppose that we are


dealing with a pair of continuous random variables (X1 , X2 ) whose sample space S is given
by S = {(x1 , x2 )|0 ≤ x1 , x2 ≤ 1; 0 ≤ x1 + x2 ≤ 1}. Suppose further that the probability den-
sity function f (x1 , x2 ) of (X1 , X2 ) is given by

⎨2, if (x1 , x2 ) ∈ S
f (x1 , x2 ) =
⎩0, otherwise

Solution: Because the probability density function is constant over the triangle defined
by S in the x1 , x2 -plane (see Figure 6.2.2), we sometimes say that (X1 , X2 ) is uniformly
distributed over S.

f(x1, x2)

(0, 0)
x1
S (1, 0)

x1 + x2 = 1
(0, 1)

x2

Figure 6.2.2 Graphical representation of the p.d.f. in Example 6.2.6.

The marginal probability density function of X1 for 0 < x1 < 1 is given by


 ∞  1−x1
f1 (x1 ) = f (x1 , x2 )dx2 = 2 dx2 = 2(1 − x1 )
−∞ 0

Hence, ⎧
⎨ 2
= 1
0 < x2 < 1 − x1
2(1−x1 ) (1−x1 ) , for
f (x2 |x1 ) =
⎩0, otherwise

Note that if (X1 , X2 ) is a pair of discrete random variables, then the conditional mean
and variance of X2 given X1 = x1 are defined as given at (6.2.23) and (6.2.24).
238 6 Distribution of Functions of Random Variables


E(X2 |X1 = x1 ) = x2 p(x2 |x1 ) (6.2.23)

x2

V ar(X2 |X1 = x1 ) = [x2 − E(x2 |x1 )]2 p(x2 |x1 ) (6.2.24)


x2

where p(x2 |x1 ) is the conditional probability function of the random variable X2 given
X1 = x1 . The mean and variance for other functions of X2 given X1 = x1 can be defined
in the same manner.
Similarly, for the case of a pair of continuous random variables, we have the following
results.

 ∞
E(X2 |X1 = x1 ) = x2 f (x2 |x1 )dx2 (6.2.25)
 ∞ −∞
V ar(X2 |X1 = x1 ) = [x2 − E(x2 |x1 )]2 f (x2 |x1 )dx2 (6.2.26)
−∞

6.2.5 Correlation between Two Random Variables


The reader will note from (6.2.15) that the covariance between the random variables
X1 and X2 is a quantity measured in [(units of X1 ) × (units of X2 )]. A somewhat more
convenient measure of how X1 and X2 “co-vary,” or are dependent on each other, is the the-
oretical or population correlation coefficient ρ. This dimensionless measure of dependence
is defined by

ρ = Cov(X1 , X2 )/σ1 σ2 (6.2.27)

where σ1 and σ2 are the population standard deviations of X1 and X2 , respectively. It can
be shown that −1 ≤ ρ ≤ 1, and hence, we have that −σ1 σ2 ≤ Cov(X1 , X2 ) ≤ σ1 σ2 .
Now, from (6.2.16) and using (6.2.17), we have that if X1 and X2 are independent
random variables, then ρ = 0. The converse need not be true however, as the following
example shows.

Example 6.2.7 (Independence and correlation coefficient) Two random variables X1 and
X2 have joint probability function given by

1/3, if , (x1 , x2 ) = (0, 0), (1, 1), (2, 0)
p(x1 , x2 ) =
0, otherwise

It is easy to see that 


1/3, if x1 = 0, 1, 2
p1 (x1 ) =
0, otherwise
6.2 Distribution Functions of Two Random Variables 239

and that 
2/3, if x2 = 0
p2 (x2 ) =
1/3, if x2 = 1

Hence, p(0, 0) = p1 (0)p2 (0), and so on, and X1 and X2 are not independent. Simple calcu-
lations further show that

μ1 = E(X1 ) = 1, μ2 = E(X2 ) = 1/3

σ12 = E(X1 − μ1 )2 = 2/3, σ22 = E(X2 − μ2 )2 = 2/9

Also, since (X1 − μ1 )(X2 − μ2 ) = (X1 − 1)(X2 − 1/3), we have that



Cov(X1 , X2 ) = (x1 − 1)(x2 − 1/3)p(x1 , x2 )
= 1/3[(0 − 1)(0 − 1/3) + (1 − 1)(1 − 1/3) + (2 − 1)(0 − 1/3)]
= 1/3[1/3 + 0 − 1/3]
=0

Therefore, the correlation coefficient has value ρ = 0, yet X1 and X2 are not independent.

Example 6.2.8 (Joint probability density function) Let the length of life (in years) of both
an operating system and the hard drive of a computer be denoted by the random variables
X1 and X2 , respectively. Suppose that the joint distribution of the random variables of X1
and X2 is given by

x x e−(x1 +x2 )/2 , if x1 > 0, x2 > 0
1 2
f (x1 , x2 ) = 64 1 2
0, otherwise

The probability density function is graphed in Figure 6.2.3.

(a) Find the marginal distributions of the random variables X1 and X2 .


(b) Find the mean and variance of the random variables X1 and X2 .
(c) Examine whether the random variables X1 and X2 are independent.
(d) Find Cov(X1 , X2 ).

Solution:
(a) The marginal probability density function of X is given by
 ∞  ∞ 
1 2 −(x1 +x2 )/2 1 2 −x1 /2 ∞
f1 (x1 ) = f (x1 , x2 )dx2 = x1 x2 e dx2 = x1 e x2 e−x2 /2 dx2
0 0 64 64 0

Integrating by parts, we have that


 ∞
1 2 −x1 /2 e−x2 /2 ∞ e−x2 /2
f1 (x1 ) = x1 e x2 | − 1× dx
64 −1/2 0 0 −1/2 2

1 2 −x1 /2 ∞ −x2 /2
= x1 e e dx2
32 0
1 2 −x1 /2
= x e , x1 > 0
16 1
240 6 Distribution of Functions of Random Variables

x2
f (x1, x2)

x1

Figure 6.2.3 Graphical representation of the joint p.d.f. in Example 6.2.8.

Similarly, it can be shown (as the reader should verify) that


1
f2 (x2 ) = x e−x2 /2 , x2 > 0
4 2
Comparing the marginal distribution of X1 and X2 with the gamma distribution given
in equation (5.9.10), we can see that the random variables X1 and X2 are distributed
as gamma with parameters γ = 3, λ = 1/2, and γ = 2, λ = 1/2, respectively.
(b) Since the random variables X1 and X2 are distributed marginally as gamma, using
equation (5.9.11), we have

μ1 = 6, σ12 = 12 and μ2 = 4, σ22 = 8

(c) We also have that


1 2 −x1 /2 1 2 −x2 /2 1 2 −(x1 +x2 )/2
f1 (x1 ) × f2 (x2 ) = x1 e × x2 e = x xe = f (x1 , x2 )
16 4 64 1 2
so that the random variables X1 and X2 are independent.
(d) Since the random variables X1 and X2 are independent, we have that Cov(X1 , X2 ) = 0.
We now state an important result about the expected value of sum of functions of
random variables X1 and X2 , similar to the one for a single variable.

Theorem 6.2.4 Let X1 and X2 be random variables and gi (X1 , X2 ), i =


1, 2, . . . , m be m functions of X1 and X2 . Then,
m 
 
m
E gi (X1 , X2 ) = E(gi (X1 , X2 )) (6.2.28)
i=1 i=1

This result can be proved in exactly the same manner as in single random-variable
(univariate) case.
6.2 Distribution Functions of Two Random Variables 241

Theorem 6.2.5 Let X1 and X2 be random variables with means μ1 and μ2 , respec-
tively. Then,
Cov(X1 , X2 ) = E(X1 X2 ) − μ1 μ2 (6.2.29)

From equation (6.2.15), we have

Cov(X1 , X2 ) = E[(X1 − μ1 )(X2 − μ2 )]


= E[X1 X2 − X1 μ2 − μ1 X2 + μ1 μ2 ]

Using Theorem 6.2.4 with g1 (X1 , X2 ) = X1 X2 , g2 (X1 , X2 ) = −μ2 X1 , g3 (X1 , X2 ) =


−μ1 X2 , and g4 (X1 , X2 ) = μ1 μ2 , we find that

Cov(X1 , X2 ) = E(X1 X2 ) − μ2 E(X1 ) − μ1 E(X2 ) + μ1 μ2


= E(X1 X2 ) − μ1 μ2 − μ1 μ2 + μ1 μ2
= E(X1 X2 ) − μ1 μ2

The reader should now use the result of Theorem 6.2.1 together with equation (6.2.29)
to show the following corollary:

Corollary 6.2.1 If X1 and X2 are two independent random variables, Cov(X1 ,


X2 ) = 0.

6.2.6 Bivariate Normal Distribution


Consider a pair of continuous random variables (X1 , X2 ). These random variables (X1 , X2 )
are said to be distributed as the bivariate normal if their joint p.d.f. f (x1 , x2 ) is given
below.

1
f (x1 , x2 ) = 
2πσ1 σ2 (1 − ρ2 )
 
1 (x1 − μ1 )2 (x1 − μ1 )(x2 − μ2 ) (x2 − μ2 )2
× exp − − 2ρ +
2(1 − ρ2 ) σ12 σ1 σ2 σ22
(6.2.30)

where −∞ ≤ xi ≤ ∞, −∞ ≤ μi ≤ ∞, σi2 > 0, i = 1, 2, and −1 < ρ < 1.


When plotted in three dimensions, a typical bivariate normal probability density func-
tion takes the form given in Figure 6.2.4. We say that this probability density function
has parameters μ1 , μ2 , σ1 , σ2 , and ρ, and it can be shown that μ1 , μ2 and σ1 , σ2 are means
and standard deviations of the random variables X1 and X2 , respectively. Further, ρ is
the correlation coefficient, where −1 < ρ < 1. By integrating f (x1 , x2 ) over −∞ ≤ x1 ≤ ∞

You might also like

pFad - Phonifier reborn

Pfad - The Proxy pFad of © 2024 Garber Painting. All rights reserved.

Note: This service is not intended for secure transactions such as banking, social media, email, or purchasing. Use at your own risk. We assume no liability whatsoever for broken pages.


Alternative Proxies:

Alternative Proxy

pFad Proxy

pFad v3 Proxy

pFad v4 Proxy