05 Random Variables
05 Random Variables
S Vijayakumar
Indian Institute of Information Technology,
Design & Manufacturing, Kancheepuram
In this Module
I Random Variables
I Discrete Random Variables and Probability Mass Functions
I Standard Discrete Random Variables
I Continuous Random Variables and Probability Density Functions
I Standard Continuous Random Variables
I Cumulative Distribution Function
I Function of a Random Variable
Random Variables
Motivation:
Random Variables
Motivation:
Motivation:
Definition
Let S be a sample space. Then a random variable X is a function X : S → R from the sample
space S to the set of real of numbers.
Example
Consider the experiment of tossing 3 fair coins.
Example
Consider the experiment of tossing 3 fair coins. Let X denote the number of heads obtained.
Example
Consider the experiment of tossing 3 fair coins. Let X denote the number of heads obtained.
(Thus X defines a function from the sample space to the set of real numbers.)
Example
Consider the experiment of tossing 3 fair coins. Let X denote the number of heads obtained.
(Thus X defines a function from the sample space to the set of real numbers.) Then X is a
random variable.
Example
Consider the experiment of tossing 3 fair coins. Let X denote the number of heads obtained.
(Thus X defines a function from the sample space to the set of real numbers.) Then X is a
random variable. It assumes one of 0, 1, 2, 3.
Example
Consider the experiment of tossing 3 fair coins. Let X denote the number of heads obtained.
(Thus X defines a function from the sample space to the set of real numbers.) Then X is a
random variable. It assumes one of 0, 1, 2, 3.
Thus the event {X = 0} occurs if and only if the event {TTT } occurs.
Example
Consider the experiment of tossing 3 fair coins. Let X denote the number of heads obtained.
(Thus X defines a function from the sample space to the set of real numbers.) Then X is a
random variable. It assumes one of 0, 1, 2, 3.
Thus the event {X = 0} occurs if and only if the event {TTT } occurs.
The event {X = 1} occurs if and only if the event {HTT , THT , TTH} occurs.
Example
Consider the experiment of tossing 3 fair coins. Let X denote the number of heads obtained.
(Thus X defines a function from the sample space to the set of real numbers.) Then X is a
random variable. It assumes one of 0, 1, 2, 3.
Thus the event {X = 0} occurs if and only if the event {TTT } occurs.
The event {X = 1} occurs if and only if the event {HTT , THT , TTH} occurs. Indeed
1
P({X = 0}) = P({TTT }) =
8
3
P({X = 1}) = P({HTT , THT , TTH}) =
8
3
P({X = 2}) = P({HHT , HTH, THH}) =
8
1
P({X = 3}) = P({HHH}) =
8
Example Contd.
Let A = {0, 1}. Then {X ∈ A} = {X = 0} ∪ {X = 1}.
Example Contd.
Let A = {0, 1}. Then {X ∈ A} = {X = 0} ∪ {X = 1}.
Thus {X ∈ A} occurs if and only if {TTT , HTT , THT , TTH} occurs. So,
4
P({X ∈ A}) = P({TTT , HTT , THT , TTH}) = .
8
Example Contd.
Let A = {0, 1}. Then {X ∈ A} = {X = 0} ∪ {X = 1}.
Thus {X ∈ A} occurs if and only if {TTT , HTT , THT , TTH} occurs. So,
4
P({X ∈ A}) = P({TTT , HTT , THT , TTH}) = .
8
1 3 4
Equivalently, P({X ∈ A}) = P({X = 0}) + P({X = 1}) = + = .
8 8 8
Example Contd.
Let A = {0, 1}. Then {X ∈ A} = {X = 0} ∪ {X = 1}.
Thus {X ∈ A} occurs if and only if {TTT , HTT , THT , TTH} occurs. So,
4
P({X ∈ A}) = P({TTT , HTT , THT , TTH}) = .
8
1 3 4
Equivalently, P({X ∈ A}) = P({X = 0}) + P({X = 1}) = + = .
8 8 8
Consider the experiment of tossing n fair coins. Let X denote the number of heads obtained.
Then X is a random variable. It assumes one of 0, 1, 2, . . . , n.
Example
Consider the experiment of tossing n fair coins. Let X denote the number of heads obtained.
Then X is a random variable. It assumes one of 0, 1, 2, . . . , n.
Here {X = k} is the event that k heads are obtained. Thus the probability of this event is
Example
Consider the experiment of tossing n fair coins. Let X denote the number of heads obtained.
Then X is a random variable. It assumes one of 0, 1, 2, . . . , n.
Here {X = k} is the event that k heads are obtained. Thus the probability of this event is
n 1
P(X = k) = .
k 2n
Example
Consider the experiment of tossing n fair coins. Let X denote the number of heads obtained.
Then X is a random variable. It assumes one of 0, 1, 2, . . . , n.
Here {X = k} is the event that k heads are obtained. Thus the probability of this event is
n 1
P(X = k) = .
k 2n
We note that
n n
X X n 1
P(X = k) = = 1.
k 2n
k=0 k=0
Definition
A random variable X that assumes only a finite or countably infinite number of values is called
a discrete random variables.
Discrete Random Variables
Definition
A random variable X that assumes only a finite or countably infinite number of values is called
a discrete random variables.
Example 1: A binomial random variable X with parameters (n, 21 ) is a discrete random variable
as it assumes only n + 1 values.
Discrete Random Variables
Definition
A random variable X that assumes only a finite or countably infinite number of values is called
a discrete random variables.
Example 1: A binomial random variable X with parameters (n, 21 ) is a discrete random variable
as it assumes only n + 1 values.
Example 2: Consider the experiment of tossing a fair coin until a head is obtained.
Discrete Random Variables
Definition
A random variable X that assumes only a finite or countably infinite number of values is called
a discrete random variables.
Example 1: A binomial random variable X with parameters (n, 21 ) is a discrete random variable
as it assumes only n + 1 values.
Example 2: Consider the experiment of tossing a fair coin until a head is obtained. Let X
denote the number of tosses made.
Discrete Random Variables
Definition
A random variable X that assumes only a finite or countably infinite number of values is called
a discrete random variables.
Example 1: A binomial random variable X with parameters (n, 21 ) is a discrete random variable
as it assumes only n + 1 values.
Example 2: Consider the experiment of tossing a fair coin until a head is obtained. Let X
denote the number of tosses made. Then X is random variable that assumes one of 1, 2, 3, . . ..
Discrete Random Variables
Definition
A random variable X that assumes only a finite or countably infinite number of values is called
a discrete random variables.
Example 1: A binomial random variable X with parameters (n, 21 ) is a discrete random variable
as it assumes only n + 1 values.
Example 2: Consider the experiment of tossing a fair coin until a head is obtained. Let X
denote the number of tosses made. Then X is random variable that assumes one of 1, 2, 3, . . ..
Thus X is a discrete random variable as it assumes only countably infinite number of values.
Discrete Random Variables
Definition
A random variable X that assumes only a finite or countably infinite number of values is called
a discrete random variables.
Example 1: A binomial random variable X with parameters (n, 21 ) is a discrete random variable
as it assumes only n + 1 values.
Example 2: Consider the experiment of tossing a fair coin until a head is obtained. Let X
denote the number of tosses made. Then X is random variable that assumes one of 1, 2, 3, . . ..
Thus X is a discrete random variable as it assumes only countably infinite number of values.
Definition
A random variable X that assumes only a finite or countably infinite number of values is called
a discrete random variables.
Example 1: A binomial random variable X with parameters (n, 21 ) is a discrete random variable
as it assumes only n + 1 values.
Example 2: Consider the experiment of tossing a fair coin until a head is obtained. Let X
denote the number of tosses made. Then X is random variable that assumes one of 1, 2, 3, . . ..
Thus X is a discrete random variable as it assumes only countably infinite number of values.
Note: The probability mass function p(a) is positive for at most a countable number of values
of a: If X assumes one of the values x1 , x2 , x3 . . ., then
p(xi ) ≥ 0 for i = 1, 2, 3, . . ..
The Probability Mass Function
Definition
The probability mass function p(a) of a discrete random variable X is given by
Note: The probability mass function p(a) is positive for at most a countable number of values
of a: If X assumes one of the values x1 , x2 , x3 . . ., then
p(xi ) ≥ 0 for i = 1, 2, 3, . . ..
p(x) = 0 for all other values of x.
The Probability Mass Function
Definition
The probability mass function p(a) of a discrete random variable X is given by
Note: The probability mass function p(a) is positive for at most a countable number of values
of a: If X assumes one of the values x1 , x2 , x3 . . ., then
p(xi ) ≥ 0 for i = 1, 2, 3, . . ..
p(x) = 0 for all other values of x.
Note: The probability mass function p(a) is positive for at most a countable number of values
of a: If X assumes one of the values x1 , x2 , x3 . . ., then
p(xi ) ≥ 0 for i = 1, 2, 3, . . ..
p(x) = 0 for all other values of x.
Note: A discrete random variable is completely specified by its probability mass function (pmf).
The Discrete Uniform Random Variable
Definition
A random variable X is called a discrete uniform random variable if it is equally likely to
assume any of the n values 1, 2, 3, . . . , n.
The Discrete Uniform Random Variable
Definition
A random variable X is called a discrete uniform random variable if it is equally likely to
assume any of the n values 1, 2, 3, . . . , n. It probability mass function is
1
p(k) = P(X = k) = , k = 1, 2, . . . , n.
n
The Discrete Uniform Random Variable
Definition
A random variable X is called a discrete uniform random variable if it is equally likely to
assume any of the n values 1, 2, 3, . . . , n. It probability mass function is
1
p(k) = P(X = k) = , k = 1, 2, . . . , n.
n
Definition
A random variable X is called a discrete uniform random variable if it is equally likely to
assume any of the n values 1, 2, 3, . . . , n. It probability mass function is
1
p(k) = P(X = k) = , k = 1, 2, . . . , n.
n
Definition
A random variable X is called a discrete uniform random variable if it is equally likely to
assume any of the n values 1, 2, 3, . . . , n. It probability mass function is
1
p(k) = P(X = k) = , k = 1, 2, . . . , n.
n
Example: The random number generators in computers are almost uniform over the given
range of values.
The Bernoulli Random Variable
Definition
A random variable X is called a Bernoulli random variable with parameter p, 0 ≤ p ≤ 1, if its
probability mass function (pmf) is given by
p(1) = P(X = 1) = p
p(0) = P(X = 0) = 1 − p
That is it always assumes one of the values 0 and 1 with the above probabilities.
Definition
A random variable X is called a Bernoulli random variable with parameter p, 0 ≤ p ≤ 1, if its
probability mass function (pmf) is given by
p(1) = P(X = 1) = p
p(0) = P(X = 0) = 1 − p
That is it always assumes one of the values 0 and 1 with the above probabilities.
Definition
A random variable X is called a Bernoulli random variable with parameter p, 0 ≤ p ≤ 1, if its
probability mass function (pmf) is given by
p(1) = P(X = 1) = p
p(0) = P(X = 0) = 1 − p
That is it always assumes one of the values 0 and 1 with the above probabilities.
Definition
A random variable X is called a Bernoulli random variable with parameter p, 0 ≤ p ≤ 1, if its
probability mass function (pmf) is given by
p(1) = P(X = 1) = p
p(0) = P(X = 0) = 1 − p
That is it always assumes one of the values 0 and 1 with the above probabilities.
Definition
A random variable X is called a binomial random variable with parameters (n, p), 0 ≤ p ≤ 1, if
its probability mass function (pmf) is given by
n k
p(k) = P(X = k) = p (1 − p)n−k , k = 0, 1, 2, . . . , n.
k
Definition
A random variable X is called a binomial random variable with parameters (n, p), 0 ≤ p ≤ 1, if
its probability mass function (pmf) is given by
n k
p(k) = P(X = k) = p (1 − p)n−k , k = 0, 1, 2, . . . , n.
k
Example: Suppose that an experiment consists of n independent trials, where each trial will be
a success with the same probability p.
The Binomial Random Variable
Definition
A random variable X is called a binomial random variable with parameters (n, p), 0 ≤ p ≤ 1, if
its probability mass function (pmf) is given by
n k
p(k) = P(X = k) = p (1 − p)n−k , k = 0, 1, 2, . . . , n.
k
Example: Suppose that an experiment consists of n independent trials, where each trial will be
a success with the same probability p. If we let X denote the number of successes, then
X ∼ Binomial(n, p).
Example
Screws produced by a certain company will be defective with probability 0.01 independently of
one another.
Example
Screws produced by a certain company will be defective with probability 0.01 independently of
one another. The company sells the screws in packages of 10 and offers a money-back
guarantee that at most 1 of the 10 screws is defective.
Example
Screws produced by a certain company will be defective with probability 0.01 independently of
one another. The company sells the screws in packages of 10 and offers a money-back
guarantee that at most 1 of the 10 screws is defective. What proportion of packages sold must
the company replace?
Example
Screws produced by a certain company will be defective with probability 0.01 independently of
one another. The company sells the screws in packages of 10 and offers a money-back
guarantee that at most 1 of the 10 screws is defective. What proportion of packages sold must
the company replace?
P(X ≥ 2) =
Example
Screws produced by a certain company will be defective with probability 0.01 independently of
one another. The company sells the screws in packages of 10 and offers a money-back
guarantee that at most 1 of the 10 screws is defective. What proportion of packages sold must
the company replace?
Thus approximately 4 out of each 100 packages sold may have to be replaced.
Note
I Let X a binomial random variable with parameters (n, p). Suppose n is large.
Note
I Let X a binomial random variable with parameters (n, p). Suppose n is large.
I Such random variables often arise in applications.
Note
I Let X a binomial random variable with parameters (n, p). Suppose n is large.
I Such random variables often arise in applications.
In this case, computing kn will be difficulty for many values of k.
I
Note
I Let X a binomial random variable with parameters (n, p). Suppose n is large.
I Such random variables often arise in applications.
In this case, computing kn will be difficulty for many values of k.
I
I Let X a binomial random variable with parameters (n, p). Suppose n is large.
I Such random variables often arise in applications.
In this case, computing kn will be difficulty for many values of k.
I
I Let X a binomial random variable with parameters (n, p). Suppose n is large.
I Such random variables often arise in applications.
In this case, computing kn will be difficulty for many values of k.
I
Definition
A random variable X that assumes one of the values 0, 1, 2, 3, . . . . . . is called a Poisson random
variable with parameter λ, for some λ > 0, if its probability mass function (pmf) is given by
λi
p(i) = P(X = i) = e −λ , i = 0, 1, 2, . . . .
i!
The Poisson Random Variable
Definition
A random variable X that assumes one of the values 0, 1, 2, 3, . . . . . . is called a Poisson random
variable with parameter λ, for some λ > 0, if its probability mass function (pmf) is given by
λi
p(i) = P(X = i) = e −λ , i = 0, 1, 2, . . . .
i!
Definition
A random variable X that assumes one of the values 0, 1, 2, 3, . . . . . . is called a Poisson random
variable with parameter λ, for some λ > 0, if its probability mass function (pmf) is given by
λi
p(i) = P(X = i) = e −λ , i = 0, 1, 2, . . . .
i!
n(n − 1) . . . (n − i + 1) 1 2 i −1
= 1 · 1 − 1 − . . . 1 −
ni n n n
The Poisson Random Variable Approximates the Binomial Random Variable
Let X be a binomial random variable with parameters (n, p). Suppose n is large and p is small
so that λ = np is moderate. Then
n i
P(X = i) = p (1 − p)n−i
i
n!
= p i (1 − p)n−i
i!(n − i)!
i n−i
n(n − 1) . . . (n − i + 1) λ λ
= 1−
i! n n
λ n
i
n(n − 1) . . . (n − i + 1) λ 1− n
= i
ni i! 1 − λn
For n large and λ moderate, n
λ
1− ≈ e −λ ,
n
and
i
n(n − 1) . . . (n − i + 1) 1 2 i −1 λ
= 1 · 1 − 1 − . . . 1 − ≈ 1, and 1 −
ni n n n n
The Poisson Random Variable Approximates the Binomial Random Variable
Let X be a binomial random variable with parameters (n, p). Suppose n is large and p is small
so that λ = np is moderate. Then
n i
P(X = i) = p (1 − p)n−i
i
n!
= p i (1 − p)n−i
i!(n − i)!
i n−i
n(n − 1) . . . (n − i + 1) λ λ
= 1−
i! n n
λ n
i
n(n − 1) . . . (n − i + 1) λ 1− n
= i
ni i! 1 − λn
For n large and λ moderate, n
λ
1− ≈ e −λ ,
n
and
i
n(n − 1) . . . (n − i + 1) 1 2 i −1 λ
= 1 · 1 − 1 − . . . 1 − ≈ 1, and 1 − ≈ 1.
ni n n n n
The Poisson Random Variable Approximates the Binomial Random Variable
So,
λ n
λi 1− λi
n(n − 1) . . . (n − i + 1)
P(X = i) = n
≈ e −λ .
ni i! λ i i!
1− n
The Poisson Random Variable Approximates the Binomial Random Variable
So,
λ n
λi 1− λi
n(n − 1) . . . (n − i + 1)
P(X = i) = n
≈ e −λ .
ni i! λ i i!
1− n
This is the probability P(Y = i) of the Poisson random variable Y with parameter λ. (Here
λ = np.)
The Poisson Random Variable Approximates the Binomial Random Variable
So,
λ n
λi 1− λi
n(n − 1) . . . (n − i + 1)
P(X = i) = n
≈ e −λ .
ni i! λ i i!
1− n
This is the probability P(Y = i) of the Poisson random variable Y with parameter λ. (Here
λ = np.)
Thus the following random variables follow Poisson distribution, each with a specific parameter
λ.
Applications
Thus the following random variables follow Poisson distribution, each with a specific parameter
λ.
I The number of misprints on a page of a book.
Applications
Thus the following random variables follow Poisson distribution, each with a specific parameter
λ.
I The number of misprints on a page of a book.
I The number of people in a community who live for 100 years.
Applications
Thus the following random variables follow Poisson distribution, each with a specific parameter
λ.
I The number of misprints on a page of a book.
I The number of people in a community who live for 100 years.
I The number earthquakes on a given day.
Applications
Thus the following random variables follow Poisson distribution, each with a specific parameter
λ.
I The number of misprints on a page of a book.
I The number of people in a community who live for 100 years.
I The number earthquakes on a given day.
I The number of α-particles emitted by a radioactive material in a fixed time interval.
Applications
Thus the following random variables follow Poisson distribution, each with a specific parameter
λ.
I The number of misprints on a page of a book.
I The number of people in a community who live for 100 years.
I The number earthquakes on a given day.
I The number of α-particles emitted by a radioactive material in a fixed time interval.
I The number of customers entering a post office on a given day.
Examples
Example 1: The number of typographical errors on a single page of a book is a Poisson random
variable with parameter λ = 12 . Calculate the probability that there is at least one error on
page number 10.
Examples
Example 1: The number of typographical errors on a single page of a book is a Poisson random
variable with parameter λ = 12 . Calculate the probability that there is at least one error on
page number 10.
Solution: Let X denote the number of defective items in the sample of 10 items. Then
X ∼ Binomial(10, 0.1).
Examples
Example 1: The number of typographical errors on a single page of a book is a Poisson random
variable with parameter λ = 12 . Calculate the probability that there is at least one error on
page number 10.
Solution: Let X denote the number of defective items in the sample of 10 items. Then
X ∼ Binomial(10, 0.1).
Solution: Let X denote the number of defective items in the sample of 10 items. Then
X ∼ Binomial(10, 0.1).
Solution: Let X denote the number of defective items in the sample of 10 items. Then
X ∼ Binomial(10, 0.1).
Solution: Let X denote the number of defective items in the sample of 10 items. Then
X ∼ Binomial(10, 0.1).
Solution: Let X denote the number of defective items in the sample of 10 items. Then
X ∼ Binomial(10, 0.1).
Solution: Let X denote the number of defective items in the sample of 10 items. Then
X ∼ Binomial(10, 0.1).
Consider an experiment that consists of counting the number of α particles given off in a
1-second interval by 1 gram of radioactive material.
Problem
Consider an experiment that consists of counting the number of α particles given off in a
1-second interval by 1 gram of radioactive material. From the past experience, it is known that
on the average, 3.2 such α particles are given off.
Problem
Consider an experiment that consists of counting the number of α particles given off in a
1-second interval by 1 gram of radioactive material. From the past experience, it is known that
on the average, 3.2 such α particles are given off. Find the probability that at most 2 α
particles will appear.
The Geometric Random Variable
Definition
A random variable X that assumes one of the values 1, 2, 3, . . . is called a geometric random
variable with parameter p, 0 ≤ p ≤ 1, if its pmf is given by
Definition
A random variable X that assumes one of the values 1, 2, 3, . . . is called a geometric random
variable with parameter p, 0 ≤ p ≤ 1, if its pmf is given by
Example: Suppose independent trials, where each trial has probability p, 0 ≤ p ≤ 1, of being a
success, are performed until a success occurs.
The Geometric Random Variable
Definition
A random variable X that assumes one of the values 1, 2, 3, . . . is called a geometric random
variable with parameter p, 0 ≤ p ≤ 1, if its pmf is given by
Example: Suppose independent trials, where each trial has probability p, 0 ≤ p ≤ 1, of being a
success, are performed until a success occurs. Let X denote the number of trials made. Then
X ∼ Geometric(p).
The Geometric Random Variable
Definition
A random variable X that assumes one of the values 1, 2, 3, . . . is called a geometric random
variable with parameter p, 0 ≤ p ≤ 1, if its pmf is given by
Example: Suppose independent trials, where each trial has probability p, 0 ≤ p ≤ 1, of being a
success, are performed until a success occurs. Let X denote the number of trials made. Then
X ∼ Geometric(p).
Definition
A random variable X that assumes one of the values r , r + 1, r + 2, . . . (r ≥ 1) is called a
negative binomial random variable with parameters (r , p), where 0 ≤ p ≤ 1, if its pmf is given
by
n−1 r
p(n) = P(X = n) = p (1 − p)n−r , n = r , r + 1, r + 2, . . . .
r −1
The Negative Binomial Random Variable
Definition
A random variable X that assumes one of the values r , r + 1, r + 2, . . . (r ≥ 1) is called a
negative binomial random variable with parameters (r , p), where 0 ≤ p ≤ 1, if its pmf is given
by
n−1 r
p(n) = P(X = n) = p (1 − p)n−r , n = r , r + 1, r + 2, . . . .
r −1
Example: Suppose independent trials, where each trial has probability p, 0 ≤ p ≤ 1, of being a
success, are performed until r successes occur. Let X denote the number of trials made. Then
X ∼ Negative Binomial(r , p).
Problems
Problem 1: Independent trials, each resulting in a success with probability p, are performed.
Problems
Problem 1: Independent trials, each resulting in a success with probability p, are performed.
What is the probability that r successes occur before m failures?
Problems
Problem 1: Independent trials, each resulting in a success with probability p, are performed.
What is the probability that r successes occur before m failures?
Problem 1: Independent trials, each resulting in a success with probability p, are performed.
What is the probability that r successes occur before m failures?
Problem 1: Independent trials, each resulting in a success with probability p, are performed.
What is the probability that r successes occur before m failures?
Problem 1: Independent trials, each resulting in a success with probability p, are performed.
What is the probability that r successes occur before m failures?
Definition
Let X be a random variable. Then the function F (x) given by
Definition
Let X be a random variable. Then the function F (x) given by
Definition
Let X be a random variable. Then the function F (x) given by
Example: Let X ∼ Bernoulli( 21 ). Its cumulative distribution function (cdf) is given in the
following figure:
Example
Let X be a random variable that assumes one of the three values s1 , s2 , s3 with probabilities
p1 , p2 , p3 , respectively.
Example
Let X be a random variable that assumes one of the three values s1 , s2 , s3 with probabilities
p1 , p2 , p3 , respectively. Then its cdf is given by in the following figure:
Properties of the Cumulative Distribution Function
Let F (x) = P(X ≤ x) be the cumulative distribution function (cdf) of a random variable X .
Properties of the Cumulative Distribution Function
Let F (x) = P(X ≤ x) be the cumulative distribution function (cdf) of a random variable X .
Then
1. F (x) is a nondecreasing function: If a < b, then F (a) ≤ F (b).
Properties of the Cumulative Distribution Function
Let F (x) = P(X ≤ x) be the cumulative distribution function (cdf) of a random variable X .
Then
1. F (x) is a nondecreasing function: If a < b, then F (a) ≤ F (b).
2. lim F (b) = 1.
b→∞
Properties of the Cumulative Distribution Function
Let F (x) = P(X ≤ x) be the cumulative distribution function (cdf) of a random variable X .
Then
1. F (x) is a nondecreasing function: If a < b, then F (a) ≤ F (b).
2. lim F (b) = 1.
b→∞
3. lim F (a) = 0.
a→−∞
Properties of the Cumulative Distribution Function
Let F (x) = P(X ≤ x) be the cumulative distribution function (cdf) of a random variable X .
Then
1. F (x) is a nondecreasing function: If a < b, then F (a) ≤ F (b).
2. lim F (b) = 1.
b→∞
3. lim F (a) = 0.
a→−∞
4. F (x) is right continuous: lim+ F (x) = F (a).
x→a
Proof.
1. F (x) is a nondecreasing function: This is true because for a < b, the event {X ≤ a} is
contained in the event {X ≤ b}. So, F (a) ≤ F (b).
Proof.
1. F (x) is a nondecreasing function: This is true because for a < b, the event {X ≤ a} is
contained in the event {X ≤ b}. So, F (a) ≤ F (b).
2. lim F (b) = 1: this follows from the continuity property of the probability function.
b→∞
Proof.
1. F (x) is a nondecreasing function: This is true because for a < b, the event {X ≤ a} is
contained in the event {X ≤ b}. So, F (a) ≤ F (b).
2. lim F (b) = 1: this follows from the continuity property of the probability function.
b→∞
3. lim F (a) = 0: this also follows from the continuity property of the probability function.
a→−∞
Proof.
1. F (x) is a nondecreasing function: This is true because for a < b, the event {X ≤ a} is
contained in the event {X ≤ b}. So, F (a) ≤ F (b).
2. lim F (b) = 1: this follows from the continuity property of the probability function.
b→∞
3. lim F (a) = 0: this also follows from the continuity property of the probability function.
a→−∞
4. F (x) is right continuous: this too follows from the continuity property of the probability
function.
Why Cumulative Distribution Function?
Let X be a random variable with cdf F (x). Then F (x) contains all the information about the
random variable X .
Why Cumulative Distribution Function?
Let X be a random variable with cdf F (x). Then F (x) contains all the information about the
random variable X .
I P(X ≤ a) = F (a).
Why Cumulative Distribution Function?
Let X be a random variable with cdf F (x). Then F (x) contains all the information about the
random variable X .
I P(X ≤ a) = F (a).
I P(X = a) = P(X ≤ a) − P(X < a) = F (a) − lim− F (x).
x→a
Why Cumulative Distribution Function?
Let X be a random variable with cdf F (x). Then F (x) contains all the information about the
random variable X .
I P(X ≤ a) = F (a).
I P(X = a) = P(X ≤ a) − P(X < a) = F (a) − lim− F (x).
x→a
I For a < b, P(a < X ≤ b) = P(X ≤ b) − P(X ≤ a) = F (b) − F (a).
Why Cumulative Distribution Function?
Let X be a random variable with cdf F (x). Then F (x) contains all the information about the
random variable X .
I P(X ≤ a) = F (a).
I P(X = a) = P(X ≤ a) − P(X < a) = F (a) − lim− F (x).
x→a
I For a < b, P(a < X ≤ b) = P(X ≤ b) − P(X ≤ a) = F (b) − F (a).
Note: Cumulative distribution functions are very important in the study of continuous random
variables.
Example
The cumulative distribution function F (x) of a random variable X (X is not purely discrete) is
given by
0 x <0
F (x) =
Example
The cumulative distribution function F (x) of a random variable X (X is not purely discrete) is
given by
0 x <0
x
2 0≤x <1
F (x) =
Example
The cumulative distribution function F (x) of a random variable X (X is not purely discrete) is
given by
0 x <0
x
2 0≤x <1
2
F (x) = 3 1≤x <2
Example
The cumulative distribution function F (x) of a random variable X (X is not purely discrete) is
given by
0 x <0
x
2 0≤x <1
2
F (x) = 3 1≤x <2
11
2≤x <3
12
Example
The cumulative distribution function F (x) of a random variable X (X is not purely discrete) is
given by
0 x <0
x
2 0≤x <1
2
F (x) = 3 1≤x <2
11
2≤x <3
12
1 3≤x
Example
The cumulative distribution function F (x) of a random variable X (X is not purely discrete) is
given by
0 x <0
x
2 0≤x <1
2
F (x) = 3 1≤x <2
11
2≤x <3
12
1 3≤x
The cumulative distribution function F (x) of a random variable X (X is not purely discrete) is
given by
0 x <0
x
2 0≤x <1
2
F (x) = 3 1≤x <2
11
2≤x <3
12
1 3≤x
The cumulative distribution function F (x) of a random variable X (X is not purely discrete) is
given by
0 x <0
x
2 0≤x <1
2
F (x) = 3 1≤x <2
11
2≤x <3
12
1 3≤x
Draw the graph of F (x). Compute (a) P(X < 3), (b) P(X = 1),
Example
The cumulative distribution function F (x) of a random variable X (X is not purely discrete) is
given by
0 x <0
x
2 0≤x <1
2
F (x) = 3 1≤x <2
11
2≤x <3
12
1 3≤x
Draw the graph of F (x). Compute (a) P(X < 3), (b) P(X = 1), (c) P(X > 21 )
Example
The cumulative distribution function F (x) of a random variable X (X is not purely discrete) is
given by
0 x <0
x
2 0≤x <1
2
F (x) = 3 1≤x <2
11
2≤x <3
12
1 3≤x
Draw the graph of F (x). Compute (a) P(X < 3), (b) P(X = 1), (c) P(X > 21 )
and (d) P(2 < X ≤ 4).
Solution:
1
11
12
2
3
1
2
1 2 3 4
Solution:
1
11
12
2
3
1
2
1 2 3 4
1
11
12
2
3
1
2
1 2 3 4
1
11
12
2
3
1
2
1 2 3 4
11
(a) P(X < 3) = lim F (x) = .
x→3− 12
Solution:
1
11
12
2
3
1
2
1 2 3 4
11
(a) P(X < 3) = lim F (x) = .
x→3− 12
(b) P(X = 1) =
Solution:
1
11
12
2
3
1
2
1 2 3 4
11
(a) P(X < 3) = lim F (x) = .
x→3− 12
(b) P(X = 1) = F (1) − lim F (x) =
x→1−
Solution:
1
11
12
2
3
1
2
1 2 3 4
11
(a) P(X < 3) = lim F (x) = .
x→3− 12
2
(b) P(X = 1) = F (1) − lim F (x) = −
x→1− 3
Solution:
1
11
12
2
3
1
2
1 2 3 4
11
(a) P(X < 3) = lim F (x) = .
x→3− 12
2 1
(b) P(X = 1) = F (1) − lim F (x) = − =
x→1− 3 2
Solution:
1
11
12
2
3
1
2
1 2 3 4
11
(a) P(X < 3) = lim F (x) = .
x→3− 12
2 1 1
(b) P(X = 1) = F (1) − lim F (x) = − = .
x→1− 3 2 6
Solution:
1
11
12
2
3
1
2
1 2 3 4
11
(a) P(X < 3) = lim F (x) = .
x→3− 12
2 1 1
(b) P(X = 1) = F (1) − lim F (x) = − = .
x→1− 3 2 6
(c) P(X > 12 ) =
Solution:
1
11
12
2
3
1
2
1 2 3 4
11
(a) P(X < 3) = lim F (x) = .
x→3− 12
2 1 1
(b) P(X = 1) = F (1) − lim F (x) = − = .
x→1− 3 2 6
(c) P(X > 12 ) = 1 − P(X ≤ 12 ) =
Solution:
1
11
12
2
3
1
2
1 2 3 4
11
(a) P(X < 3) = lim F (x) = .
x→3− 12
2 1 1
(b) P(X = 1) = F (1) − lim F (x) = − = .
x→1− 3 2 6
(c) P(X > 12 ) = 1 − P(X ≤ 12 ) = 1 − F ( 12 ) =
Solution:
1
11
12
2
3
1
2
1 2 3 4
11
(a) P(X < 3) = lim F (x) = .
x→3− 12
2 1 1
(b) P(X = 1) = F (1) − lim F (x) = − = .
x→1− 3 2 6
(c) P(X > 12 ) = 1 − P(X ≤ 12 ) = 1 − F ( 12 ) = 1 − 1
4 =
Solution:
1
11
12
2
3
1
2
1 2 3 4
11
(a) P(X < 3) = lim F (x) = .
x→3− 12
2 1 1
(b) P(X = 1) = F (1) − lim F (x) = − = .
x→1− 3 2 6
(c) P(X > 12 ) = 1 − P(X ≤ 12 ) = 1 − F ( 12 ) = 1 − 1
4 = 34 .
Solution:
1
11
12
2
3
1
2
1 2 3 4
11
(a) P(X < 3) = lim F (x) = .
x→3− 12
2 1 1
(b) P(X = 1) = F (1) − lim F (x) = − = .
x→1− 3 2 6
(c) P(X > 12 ) = 1 − P(X ≤ 12 ) = 1 − F ( 12 ) = 1 − 1
4 = 34 .
(d) P(2 < X ≤ 4) =
Solution:
1
11
12
2
3
1
2
1 2 3 4
11
(a) P(X < 3) = lim F (x) = .
x→3− 12
2 1 1
(b) P(X = 1) = F (1) − lim F (x) = − = .
x→1− 3 2 6
(c) P(X > 12 ) = 1 − P(X ≤ 12 ) = 1 − F ( 12 ) = 1 − 1
4 = 34 .
(d) P(2 < X ≤ 4) = F (4) − F (2) =
Solution:
1
11
12
2
3
1
2
1 2 3 4
11
(a) P(X < 3) = lim F (x) = .
x→3− 12
2 1 1
(b) P(X = 1) = F (1) − lim F (x) = − = .
x→1− 3 2 6
(c) P(X > 12 ) = 1 − P(X ≤ 12 ) = 1 − F ( 12 ) = 1 − 1
4 = 34 .
(d) P(2 < X ≤ 4) = F (4) − F (2) = 1 − 11 1
12 = 12 .
Continuous Random Variables
Definition
A random variable X is said to be a continuous random variable if there exists a non-negative
function f (x), defined for all real x, −∞ < x < ∞, such that for any set of B of real numbers
Z
P(X ∈ B) = f (x)dx.
B
The function f (x) is called the probability density function (pdf) of the random variable X .
Continuous Random Variables: Note
Z ∞
I P(X ∈ (−∞, ∞)) = f (x)dx = 1.
−∞
Continuous Random Variables: Note
Z ∞
I P(X ∈ (−∞, ∞)) = f (x)dx = 1.
−∞
Z ∞
I P(X ∈ (−∞, ∞)) = f (x)dx = 1.
−∞
Z b
I P(X ∈ [a, b]) = f (x)dx.
a
Continuous Random Variables: Note
Z ∞
I P(X ∈ (−∞, ∞)) = f (x)dx = 1.
−∞
Z b
I P(X ∈ [a, b]) = f (x)dx.
a
I P(X = a) =
Continuous Random Variables: Note
Z ∞
I P(X ∈ (−∞, ∞)) = f (x)dx = 1.
−∞
Z b
I P(X ∈ [a, b]) = f (x)dx.
Z a a
I P(X = a) = f (x)dx = 0.
a
Continuous Random Variables: Note
Z ∞
I P(X ∈ (−∞, ∞)) = f (x)dx = 1.
−∞
Z b
I P(X ∈ [a, b]) = f (x)dx.
Z a a
I P(X = a) = f (x)dx = 0.
a
Z a
I P(X < a) = P(X ≤ a) = F (a) = f (x)dx.
−∞
Example
The amount of time in hours that a computer functions before breaking down is a continuous
random variable with probability density
function given by
λe −x/100 x ≥0
f (x) =
0 0≤x <0
Example
The amount of time in hours that a computer functions before breaking down is a continuous
random variable with probability density
function given by
λe −x/100 x ≥0
f (x) =
0 0≤x <0
What is the probability that
(a) the computer will function between 50 and 150 hours before breaking down?
Example
The amount of time in hours that a computer functions before breaking down is a continuous
random variable with probability density
function given by
λe −x/100 x ≥0
f (x) =
0 0≤x <0
What is the probability that
(a) the computer will function between 50 and 150 hours before breaking down?
(b) it will function for fewer than 100 hours?
Example
The amount of time in hours that a computer functions before breaking down is a continuous
random variable with probability density
function given by
λe −x/100 x ≥0
f (x) =
0 0≤x <0
What is the probability that
(a) the computer will function between 50 and 150 hours before breaking down?
(b) it will function for fewer than 100 hours?
Solution: Let X denote the life time of the computer.
Example
The amount of time in hours that a computer functions before breaking down is a continuous
random variable with probability density
function given by
λe −x/100 x ≥0
f (x) =
0 0≤x <0
What is the probability that
(a) the computer will function between 50 and 150 hours before breaking down?
(b) it will function for fewer than 100 hours?
Solution: Let X denote the life time of the computer. We are given the probability density
function f (x) (pdf)
Z ∞ of X .
Z ∞ h i∞
f (x)dx = λe −x/100 dx = λ(−100) e −x/100 = 100λ.
−∞ 0 0
Example
The amount of time in hours that a computer functions before breaking down is a continuous
random variable with probability density
function given by
λe −x/100 x ≥0
f (x) =
0 0≤x <0
What is the probability that
(a) the computer will function between 50 and 150 hours before breaking down?
(b) it will function for fewer than 100 hours?
Solution: Let X denote the life time of the computer. We are given the probability density
function f (x) (pdf)
Z ∞ of X .
Z ∞ h i∞
f (x)dx = λe −x/100 dx = λ(−100) e −x/100 = 100λ.
−∞Z 0 0
∞
But f (x)dx = 1.
−∞
Example
The amount of time in hours that a computer functions before breaking down is a continuous
random variable with probability density
function given by
λe −x/100 x ≥0
f (x) =
0 0≤x <0
What is the probability that
(a) the computer will function between 50 and 150 hours before breaking down?
(b) it will function for fewer than 100 hours?
Solution: Let X denote the life time of the computer. We are given the probability density
function f (x) (pdf)
Z ∞ of X .
Z ∞ h i∞
f (x)dx = λe −x/100 dx = λ(−100) e −x/100 = 100λ.
−∞Z 0 0
∞
But f (x)dx = 1. So, 100λ = 1
−∞
Example
The amount of time in hours that a computer functions before breaking down is a continuous
random variable with probability density
function given by
λe −x/100 x ≥0
f (x) =
0 0≤x <0
What is the probability that
(a) the computer will function between 50 and 150 hours before breaking down?
(b) it will function for fewer than 100 hours?
Solution: Let X denote the life time of the computer. We are given the probability density
function f (x) (pdf)
Z ∞ of X .
Z ∞ h i∞
f (x)dx = λe −x/100 dx = λ(−100) e −x/100 = 100λ.
−∞Z 0 0
∞
1
But f (x)dx = 1. So, 100λ = 1 ⇒ λ = 100 .
−∞
Example
The amount of time in hours that a computer functions before breaking down is a continuous
random variable with probability density
function given by
λe −x/100 x ≥0
f (x) =
0 0≤x <0
What is the probability that
(a) the computer will function between 50 and 150 hours before breaking down?
(b) it will function for fewer than 100 hours?
Solution: Let X denote the life time of the computer. We are given the probability density
function f (x) (pdf)
Z ∞ of X .
Z ∞ h i∞
f (x)dx = λe −x/100 dx = λ(−100) e −x/100 = 100λ.
−∞Z 0 0
∞
1
But f (x)dx = 1. So, 100λ = 1 ⇒ λ = 100 .
−∞
Property 1:
Z a+ 2
P(a − ≤ X ≤ a + ) = ≈ f (a).
2 2 a− 2
Properties of the Probability Density Function
Let X be a continuous random variable with pdf f (x).
Property 1:
Z a+ 2
P(a − ≤ X ≤ a + ) = ≈ f (a).
2 2 a− 2
That is, the probability that X assumes a value in an interval of length around the point a is
approximately f (a).
Properties of the Probability Density Function
Let X be a continuous random variable with pdf f (x).
Property 1:
Z a+ 2
P(a − ≤ X ≤ a + ) = ≈ f (a).
2 2 a− 2
That is, the probability that X assumes a value in an interval of length around the point a is
approximately f (a).
Property 2:
Let F (a) be the cumulative distribution of X :
Properties of the Probability Density Function
Let X be a continuous random variable with pdf f (x).
Property 1:
Z a+ 2
P(a − ≤ X ≤ a + ) = ≈ f (a).
2 2 a− 2
That is, the probability that X assumes a value in an interval of length around the point a is
approximately f (a).
Property 2:
Let F (a) be the cumulative distribution of X :
F (a) = P(X ≤ a) =
Properties of the Probability Density Function
Let X be a continuous random variable with pdf f (x).
Property 1:
Z a+ 2
P(a − ≤ X ≤ a + ) = ≈ f (a).
2 2 a− 2
That is, the probability that X assumes a value in an interval of length around the point a is
approximately f (a).
Property 2:
Let F (a) be the cumulative distribution of X :
F (a) = P(X ≤ a) = P(X ∈ (−∞, a]) =
Properties of the Probability Density Function
Let X be a continuous random variable with pdf f (x).
Property 1:
Z a+ 2
P(a − ≤ X ≤ a + ) = ≈ f (a).
2 2 a− 2
That is, the probability that X assumes a value in an interval of length around the point a is
approximately f (a).
Property 2:
Let F (a) be the cumulative distribution of X :
Z a
F (a) = P(X ≤ a) = P(X ∈ (−∞, a]) = f (x)dx.
−∞
Properties of the Probability Density Function
Let X be a continuous random variable with pdf f (x).
Property 1:
Z a+ 2
P(a − ≤ X ≤ a + ) = ≈ f (a).
2 2 a− 2
That is, the probability that X assumes a value in an interval of length around the point a is
approximately f (a).
Property 2:
Let F (a) be the cumulative distribution of X :
Z a
F (a) = P(X ≤ a) = P(X ∈ (−∞, a]) = f (x)dx.
−∞
Note: Because f (x) > 0 only when x ∈ [a, b], X always assumes a value in the interval [a, b].
Standard Continuous Distributions: Uniform Random Variables
Definition
A random variable X is said to be a uniform random variable on the interval [a, b] if its
probability density function is given by
1
b−a a≤x ≤b
f (x) =
0 otherwise
Note: Because f (x) > 0 only when x ∈ [a, b], X always assumes a value in the interval [a, b].
Also X is just as likely to be near any value in [a, b] as it is to be near any other value.
Standard Continuous Distributions: Uniform Random Variables
Definition
A random variable X is said to be a uniform random variable on the interval [a, b] if its
probability density function is given by
1
b−a a≤x ≤b
f (x) =
0 otherwise
Note: Because f (x) > 0 only when x ∈ [a, b], X always assumes a value in the interval [a, b].
Also X is just as likely to be near any value in [a, b] as it is to be near any other value.
Note: Because f (x) > 0 only when x ∈ [a, b], X always assumes a value in the interval [a, b].
Also X is just as likely to be near any value in [a, b] as it is to be near any other value.
That is, the probability is proportional to the length of the subinterval [c, d].
Standard Continuous Distributions:
Standard Continuous Distributions: Uniform Random Variables
Definition
A random variable X is said to be a uniform random variable on the interval [a, b] if its
probability density function is given by
1
b−a a≤x ≤b
f (x) =
0 otherwise
Note: Because f (x) > 0 only when x ∈ [a, b], X always assumes a value in the interval [a, b].
Also X is just as likely to be near any value in [a, b] as it is to be near any other value.
That is, the probability is proportional to the length of the subinterval [c, d].
Homework: Plot the graph of the pdf of X ∼ Uniform([a, b]). Also find its cdf and plot it.
Examples
Example 1: The random variable X is uniformly distributed over the interval [0, 10].
Examples
Example 1: The random variable X is uniformly distributed over the interval [0, 10]. Calculate
the probability that (a) X < 3,
Examples
Example 1: The random variable X is uniformly distributed over the interval [0, 10]. Calculate
the probability that (a) X < 3, (b) X > 6
Examples
Example 1: The random variable X is uniformly distributed over the interval [0, 10]. Calculate
the probability that (a) X < 3, (b) X > 6 and (c) 3 < X < 8.
Examples
Example 1: The random variable X is uniformly distributed over the interval [0, 10]. Calculate
the probability that (a) X < 3, (b) X > 6 and (c) 3 < X < 8.
Example 1: The random variable X is uniformly distributed over the interval [0, 10]. Calculate
the probability that (a) X < 3, (b) X > 6 and (c) 3 < X < 8.
Example 2: Buses arrive at a specified stop at 15-minute intervals starting at 7 AM. That is,
they arrive at 7, 7:15, 7:30, and so on.
Examples
Example 1: The random variable X is uniformly distributed over the interval [0, 10]. Calculate
the probability that (a) X < 3, (b) X > 6 and (c) 3 < X < 8.
Example 2: Buses arrive at a specified stop at 15-minute intervals starting at 7 AM. That is,
they arrive at 7, 7:15, 7:30, and so on. If a passenger arrives at the stop at a time that is
uniformly distributed between 7 and 7:30, find the probability that he waits (a) less than 5 five
minutes for a bus;
Examples
Example 1: The random variable X is uniformly distributed over the interval [0, 10]. Calculate
the probability that (a) X < 3, (b) X > 6 and (c) 3 < X < 8.
Example 2: Buses arrive at a specified stop at 15-minute intervals starting at 7 AM. That is,
they arrive at 7, 7:15, 7:30, and so on. If a passenger arrives at the stop at a time that is
uniformly distributed between 7 and 7:30, find the probability that he waits (a) less than 5 five
minutes for a bus; (b) more than 10 minutes for a bus.
Examples
Example 1: The random variable X is uniformly distributed over the interval [0, 10]. Calculate
the probability that (a) X < 3, (b) X > 6 and (c) 3 < X < 8.
Example 2: Buses arrive at a specified stop at 15-minute intervals starting at 7 AM. That is,
they arrive at 7, 7:15, 7:30, and so on. If a passenger arrives at the stop at a time that is
uniformly distributed between 7 and 7:30, find the probability that he waits (a) less than 5 five
minutes for a bus; (b) more than 10 minutes for a bus.
Solution: Let X denote the arrival time of the passenger at the bus stop.
Examples
Example 1: The random variable X is uniformly distributed over the interval [0, 10]. Calculate
the probability that (a) X < 3, (b) X > 6 and (c) 3 < X < 8.
Example 2: Buses arrive at a specified stop at 15-minute intervals starting at 7 AM. That is,
they arrive at 7, 7:15, 7:30, and so on. If a passenger arrives at the stop at a time that is
uniformly distributed between 7 and 7:30, find the probability that he waits (a) less than 5 five
minutes for a bus; (b) more than 10 minutes for a bus.
Solution: Let X denote the arrival time of the passenger at the bus stop. Then
R 151
R 301
(a) P(10 < X < 15) + P(25 < X < 30) = 10 30
dx + 25 30
dx = 31 .
Example
Consider a random chord of a circle of radius r .
Example
Consider a random chord of a circle of radius r . What is the probability that the length of the
chord will be greater than the side of an equilateral triangle inscribed in that circle?
Example
Consider a random chord of a circle of radius r . What is the probability that the length of the
chord will be greater than the side of an equilateral triangle inscribed in that circle?
Solution: Formulation I. Consider a chord of the circle obtained by choosing the distance D of
the chord from the center uniformly at random from 0 to r .
Example
Consider a random chord of a circle of radius r . What is the probability that the length of the
chord will be greater than the side of an equilateral triangle inscribed in that circle?
Solution: Formulation I. Consider a chord of the circle obtained by choosing the distance D of
the chord from the center uniformly at random from 0 to r . Then the required probability is
r r /2 1
P(D < )= = .
2 r 2
Example
Consider a random chord of a circle of radius r . What is the probability that the length of the
chord will be greater than the side of an equilateral triangle inscribed in that circle?
Solution: Formulation I. Consider a chord of the circle obtained by choosing the distance D of
the chord from the center uniformly at random from 0 to r . Then the required probability is
r r /2 1
P(D < )= = .
2 r 2
Formulation II. Consider a chord of the circle obtained by fixing one end A and choosing the
other end B randomly:
Example
Consider a random chord of a circle of radius r . What is the probability that the length of the
chord will be greater than the side of an equilateral triangle inscribed in that circle?
Solution: Formulation I. Consider a chord of the circle obtained by choosing the distance D of
the chord from the center uniformly at random from 0 to r . Then the required probability is
r r /2 1
P(D < )= = .
2 r 2
Formulation II. Consider a chord of the circle obtained by fixing one end A and choosing the
other end B randomly: This means that the angle θ made by the chord with the tangent at A
varies uniformly from 0◦ to 180◦ .
Example
Consider a random chord of a circle of radius r . What is the probability that the length of the
chord will be greater than the side of an equilateral triangle inscribed in that circle?
Solution: Formulation I. Consider a chord of the circle obtained by choosing the distance D of
the chord from the center uniformly at random from 0 to r . Then the required probability is
r r /2 1
P(D < )= = .
2 r 2
Formulation II. Consider a chord of the circle obtained by fixing one end A and choosing the
other end B randomly: This means that the angle θ made by the chord with the tangent at A
varies uniformly from 0◦ to 180◦ . Thus the required probability is
120 − 60 1
P(60◦ < θ < 120◦ ) = = .
180 3
Example
Consider a random chord of a circle of radius r . What is the probability that the length of the
chord will be greater than the side of an equilateral triangle inscribed in that circle?
Solution: Formulation I. Consider a chord of the circle obtained by choosing the distance D of
the chord from the center uniformly at random from 0 to r . Then the required probability is
r r /2 1
P(D < )= = .
2 r 2
Formulation II. Consider a chord of the circle obtained by fixing one end A and choosing the
other end B randomly: This means that the angle θ made by the chord with the tangent at A
varies uniformly from 0◦ to 180◦ . Thus the required probability is
120 − 60 1
P(60◦ < θ < 120◦ ) = = .
180 3
A paradox?
Exponential Random Variables
Definition
A continuous random variable whose probability density function (pdf) is given, for some
λ > 0, by
λe −λx
x ≥0
f (x) =
Exponential Random Variables
Definition
A continuous random variable whose probability density function (pdf) is given, for some
λ > 0, by
λe −λx
x ≥0
f (x) =
0 x <0
is called an exponential random variable with parameter λ.
Exponential Random Variables
Definition
A continuous random variable whose probability density function (pdf) is given, for some
λ > 0, by
λe −λx
x ≥0
f (x) =
0 x <0
is called an exponential random variable with parameter λ.
F (a) = P(X ≤ a)
Exponential Random Variables
Definition
A continuous random variable whose probability density function (pdf) is given, for some
λ > 0, by
λe −λx
x ≥0
f (x) =
0 x <0
is called an exponential random variable with parameter λ.
F (a) = P(X ≤ a)
Z a
= λe −λx dx
0
Exponential Random Variables
Definition
A continuous random variable whose probability density function (pdf) is given, for some
λ > 0, by
λe −λx
x ≥0
f (x) =
0 x <0
is called an exponential random variable with parameter λ.
F (a) = P(X ≤ a)
Z a
= λe −λx dx
0
−λx a
= −e 0
Exponential Random Variables
Definition
A continuous random variable whose probability density function (pdf) is given, for some
λ > 0, by
λe −λx
x ≥0
f (x) =
0 x <0
is called an exponential random variable with parameter λ.
F (a) = P(X ≤ a)
Z a
= λe −λx dx
0
−λx a
= −e 0
= 1 − e −λa
Exponential Random Variables
Definition
A continuous random variable whose probability density function (pdf) is given, for some
λ > 0, by
λe −λx
x ≥0
f (x) =
0 x <0
is called an exponential random variable with parameter λ.
F (a) = P(X ≤ a)
Z a
= λe −λx dx
0
−λx a
= −e 0
= 1 − e −λa
Suppose the length of a phone call in minutes is an exponential random variable with
1
parameter λ = 10 .
Example
Suppose the length of a phone call in minutes is an exponential random variable with
1
parameter λ = 10 . Find the probability that (a) it lasts more than 10 minutes;
Example
Suppose the length of a phone call in minutes is an exponential random variable with
1
parameter λ = 10 . Find the probability that (a) it lasts more than 10 minutes; (b) it gets over
between 10 and 20 minutes.
Example
Suppose the length of a phone call in minutes is an exponential random variable with
1
parameter λ = 10 . Find the probability that (a) it lasts more than 10 minutes; (b) it gets over
between 10 and 20 minutes.
Suppose the length of a phone call in minutes is an exponential random variable with
1
parameter λ = 10 . Find the probability that (a) it lasts more than 10 minutes; (b) it gets over
between 10 and 20 minutes.
Solution: Let X denote the length of the phone call. Then the required probabilities are
(a) P(X > 10) =
Example
Suppose the length of a phone call in minutes is an exponential random variable with
1
parameter λ = 10 . Find the probability that (a) it lasts more than 10 minutes; (b) it gets over
between 10 and 20 minutes.
Solution: Let X denote the length of the phone call. Then the required probabilities are
(a) P(X > 10) = 1 − F (10) =
Example
Suppose the length of a phone call in minutes is an exponential random variable with
1
parameter λ = 10 . Find the probability that (a) it lasts more than 10 minutes; (b) it gets over
between 10 and 20 minutes.
Solution: Let X denote the length of the phone call. Then the required probabilities are
(a) P(X > 10) = 1 − F (10) = 1 − (1 − e −1 ) =
Example
Suppose the length of a phone call in minutes is an exponential random variable with
1
parameter λ = 10 . Find the probability that (a) it lasts more than 10 minutes; (b) it gets over
between 10 and 20 minutes.
Solution: Let X denote the length of the phone call. Then the required probabilities are
(a) P(X > 10) = 1 − F (10) = 1 − (1 − e −1 ) = e −1 ≈ 0.368.
Example
Suppose the length of a phone call in minutes is an exponential random variable with
1
parameter λ = 10 . Find the probability that (a) it lasts more than 10 minutes; (b) it gets over
between 10 and 20 minutes.
Solution: Let X denote the length of the phone call. Then the required probabilities are
(a) P(X > 10) = 1 − F (10) = 1 − (1 − e −1 ) = e −1 ≈ 0.368.
(b) P(10 < X < 20) =
Example
Suppose the length of a phone call in minutes is an exponential random variable with
1
parameter λ = 10 . Find the probability that (a) it lasts more than 10 minutes; (b) it gets over
between 10 and 20 minutes.
Solution: Let X denote the length of the phone call. Then the required probabilities are
(a) P(X > 10) = 1 − F (10) = 1 − (1 − e −1 ) = e −1 ≈ 0.368.
(b) P(10 < X < 20) =
Example
Suppose the length of a phone call in minutes is an exponential random variable with
1
parameter λ = 10 . Find the probability that (a) it lasts more than 10 minutes; (b) it gets over
between 10 and 20 minutes.
Solution: Let X denote the length of the phone call. Then the required probabilities are
(a) P(X > 10) = 1 − F (10) = 1 − (1 − e −1 ) = e −1 ≈ 0.368.
(b) P(10 < X < 20) = F (20) − F (10) =
Example
Suppose the length of a phone call in minutes is an exponential random variable with
1
parameter λ = 10 . Find the probability that (a) it lasts more than 10 minutes; (b) it gets over
between 10 and 20 minutes.
Solution: Let X denote the length of the phone call. Then the required probabilities are
(a) P(X > 10) = 1 − F (10) = 1 − (1 − e −1 ) = e −1 ≈ 0.368.
(b) P(10 < X < 20) = F (20) − F (10) = e −1 − e −2 ≈ 0.233.
The Memoryless Property of Exponetial Random Variables
Definition
A nonnegative random variable X is said to be memoryless if
Suppose X denotes the lifetime of some instrument. What does the above equation mean?
The Memoryless Property of Exponetial Random Variables
Definition
A nonnegative random variable X is said to be memoryless if
Suppose X denotes the lifetime of some instrument. What does the above equation mean?
Note:
P(X > s + t|X > t) =
The Memoryless Property of Exponetial Random Variables
Definition
A nonnegative random variable X is said to be memoryless if
Suppose X denotes the lifetime of some instrument. What does the above equation mean?
Note: P((X > s + t) ∩ (X > t))
P(X > s + t|X > t) = =
P(X > t)
The Memoryless Property of Exponetial Random Variables
Definition
A nonnegative random variable X is said to be memoryless if
Suppose X denotes the lifetime of some instrument. What does the above equation mean?
Note: P((X > s + t) ∩ (X > t)) P(X > s + t)
P(X > s + t|X > t) = = .
P(X > t) P(X > t)
The Memoryless Property of Exponetial Random Variables
Definition
A nonnegative random variable X is said to be memoryless if
Suppose X denotes the lifetime of some instrument. What does the above equation mean?
Note: P((X > s + t) ∩ (X > t)) P(X > s + t)
P(X > s + t|X > t) = = .
P(X > t) P(X > t)
Suppose X denotes the lifetime of some instrument. What does the above equation mean?
Note: P((X > s + t) ∩ (X > t)) P(X > s + t)
P(X > s + t|X > t) = = .
P(X > t) P(X > t)
Suppose X denotes the lifetime of some instrument. What does the above equation mean?
Note: P((X > s + t) ∩ (X > t)) P(X > s + t)
P(X > s + t|X > t) = = .
P(X > t) P(X > t)
Suppose X denotes the lifetime of some instrument. What does the above equation mean?
Note: P((X > s + t) ∩ (X > t)) P(X > s + t)
P(X > s + t|X > t) = = .
P(X > t) P(X > t)
x−µ
Hint: Use the substitution y = σ to reduce the integral
Z ∞
1 2 2
√ e −(x−µ) /2σ dx
2πσ −∞
to the form
Z ∞
1 2
√ e −y /2
dy .
2π −∞
Z ∞
2
Let I = e −y /2
dy .
−∞
Homework
2 2
Prove that f (x) = √ 1 e −(x−µ) /2σ , −∞ < x < ∞ is indeed a probability density function.
2πσ
x−µ
Hint: Use the substitution y = σ to reduce the integral
Z ∞
1 2 2
√ e −(x−µ) /2σ dx
2πσ −∞
to the form
Z ∞
1 2
√ e −y /2
dy .
2π −∞
Z ∞
2
Let I = e −y /2
dy . Then show that
−∞
∞ Z Z ∞
−x 2 /2 2
2
I = e dx e −y /2 dy =
−∞ −∞
Homework
2 2
Prove that f (x) = √ 1 e −(x−µ) /2σ , −∞ < x < ∞ is indeed a probability density function.
2πσ
x−µ
Hint: Use the substitution y = σ to reduce the integral
Z ∞
1 2 2
√ e −(x−µ) /2σ dx
2πσ −∞
to the form
Z ∞
1 2
√ e −y /2
dy .
2π −∞
Z ∞
2
Let I = e −y /2
dy . Then show that
−∞
∞ Z Z ∞ Z ∞
−x 2 /2 −y 2 /2 2
/2 −y 2 /2
2
I = e dx e dy = e −x e dxdy =
−∞ −∞ −∞
Homework
2 2
Prove that f (x) = √ 1 e −(x−µ) /2σ , −∞ < x < ∞ is indeed a probability density function.
2πσ
x−µ
Hint: Use the substitution y = σ to reduce the integral
Z ∞
1 2 2
√ e −(x−µ) /2σ dx
2πσ −∞
to the form
Z ∞
1 2
√ e −y /2
dy .
2π −∞
Z ∞
2
Let I = e −y /2
dy . Then show that
−∞
∞ Z Z ∞ Z ∞
−x 2 /2 −y 2 /2 2
/2 −y 2 /2
2
I = e dx e dy = e −x e dxdy = 2π.
−∞ −∞ −∞
Fact
Note: Here in the place of ‘µ’ we have aµ + b and in the place of ‘σ 2 ’, we have
Proof...
Note: Here in the place of ‘µ’ we have aµ + b and in the place of ‘σ 2 ’, we have a2 σ 2 . Thus
aX + b is normal with parameters (aµ + b, a2 σ 2 ).
Proof...
Note: Here in the place of ‘µ’ we have aµ + b and in the place of ‘σ 2 ’, we have a2 σ 2 . Thus
aX + b is normal with parameters (aµ + b, a2 σ 2 ).
X −µ
Note: In particular, if X is a normal random variable with parameters (µ, σ 2 ), then Z = σ is
a normal random variable with parameters
Proof...
Note: Here in the place of ‘µ’ we have aµ + b and in the place of ‘σ 2 ’, we have a2 σ 2 . Thus
aX + b is normal with parameters (aµ + b, a2 σ 2 ).
Definition
If Z is a normal random variable with parameters (0, 1), it is called the standard normal or unit
normal random variable.
The Standard Normal Random Variable
Definition
If Z is a normal random variable with parameters (0, 1), it is called the standard normal or unit
normal random variable.
Observations
1. If X is normal with parameters (µ, σ 2 ), then Z = X −µ
σ is a normal random variable with
parameters (0,1) and so it is the standard normal random variable.
The Standard Normal Random Variable
Definition
If Z is a normal random variable with parameters (0, 1), it is called the standard normal or unit
normal random variable.
Observations
1. If X is normal with parameters (µ, σ 2 ), then Z = X −µ
σ is a normal random variable with
parameters (0,1) and so it is the standard normal random variable.
2. Conversely, if Z is the standard normal random variable, then X = σZ + µ is a normal
random variable with parameters (µ, σ 2 ).
The cdf of the Standard Normal and The Normal Table
Φ(z)
The cumulative distribution function of the standard normal random variable Z is usually
denoted by Φ(z):
The cdf of the Standard Normal and The Normal Table
Φ(z)
The cumulative distribution function of the standard normal random variable Z is usually
denoted by Φ(z): Z z
1 2
Φ(z) = P(Z ≤ z) = √ e −x /2 dx.
2π −∞
The cdf of the Standard Normal and The Normal Table
Φ(z)
The cumulative distribution function of the standard normal random variable Z is usually
denoted by Φ(z): Z z
1 2
Φ(z) = P(Z ≤ z) = √ e −x /2 dx.
2π −∞
Fact
Φ(z)
The cumulative distribution function of the standard normal random variable Z is usually
denoted by Φ(z): Z z
1 2
Φ(z) = P(Z ≤ z) = √ e −x /2 dx.
2π −∞
Fact
Note: The “Normal Table” usually gives the values of Φ(z) for z ≥ 0.
The cdf of the Standard Normal and The Normal Table
Φ(z)
The cumulative distribution function of the standard normal random variable Z is usually
denoted by Φ(z): Z z
1 2
Φ(z) = P(Z ≤ z) = √ e −x /2 dx.
2π −∞
Fact
Note: The “Normal Table” usually gives the values of Φ(z) for z ≥ 0. For negative z, the
above identity can be used.
Example
If X is a normal random variable with parameters (µ, σ 2 ) = (3, 9), find (a) P(2 < X < 5); (b)
P(X > 0); (c) P(|X − 3| > 6).
Example
If X is a normal random variable with parameters (µ, σ 2 ) = (3, 9), find (a) P(2 < X < 5); (b)
P(X > 0); (c) P(|X − 3| > 6).
Solution: (a)
2−3 X −3 5−3
P(2 < X < 5) = P( < < )
3 3 3
Example
If X is a normal random variable with parameters (µ, σ 2 ) = (3, 9), find (a) P(2 < X < 5); (b)
P(X > 0); (c) P(|X − 3| > 6).
Solution: (a)
2−3 X −3 5−3
P(2 < X < 5) = P( < < )
3 3 3
−1 2
= P( <Z < )
3 3
Example
If X is a normal random variable with parameters (µ, σ 2 ) = (3, 9), find (a) P(2 < X < 5); (b)
P(X > 0); (c) P(|X − 3| > 6).
Solution: (a)
2−3 X −3 5−3
P(2 < X < 5) = P( < < )
3 3 3
−1 2
= P( <Z < )
3 3
2 −1
= Φ( ) − Φ( )
3 3
Example
If X is a normal random variable with parameters (µ, σ 2 ) = (3, 9), find (a) P(2 < X < 5); (b)
P(X > 0); (c) P(|X − 3| > 6).
Solution: (a)
2−3 X −3 5−3
P(2 < X < 5) = P( < < )
3 3 3
−1 2
= P( <Z < )
3 3
2 −1
= Φ( ) − Φ( )
3 3
≈ 0.3779
Example
If X is a normal random variable with parameters (µ, σ 2 ) = (3, 9), find (a) P(2 < X < 5); (b)
P(X > 0); (c) P(|X − 3| > 6).
Solution: (a)
2−3 X −3 5−3
P(2 < X < 5) = P( < < )
3 3 3
−1 2
= P( <Z < )
3 3
2 −1
= Φ( ) − Φ( )
3 3
≈ 0.3779
(b)
X −3 0−3
P(X > 0) = P( > ) = P(Z > −1)
3 3
Example
If X is a normal random variable with parameters (µ, σ 2 ) = (3, 9), find (a) P(2 < X < 5); (b)
P(X > 0); (c) P(|X − 3| > 6).
Solution: (a)
2−3 X −3 5−3
P(2 < X < 5) = P( < < )
3 3 3
−1 2
= P( <Z < )
3 3
2 −1
= Φ( ) − Φ( )
3 3
≈ 0.3779
(b)
X −3 0−3
P(X > 0) = P( > ) = P(Z > −1)
3 3
= 1 − Φ(−1)
Example
If X is a normal random variable with parameters (µ, σ 2 ) = (3, 9), find (a) P(2 < X < 5); (b)
P(X > 0); (c) P(|X − 3| > 6).
Solution: (a)
2−3 X −3 5−3
P(2 < X < 5) = P( < < )
3 3 3
−1 2
= P( <Z < )
3 3
2 −1
= Φ( ) − Φ( )
3 3
≈ 0.3779
(b)
X −3 0−3
P(X > 0) = P( > ) = P(Z > −1)
3 3
= 1 − Φ(−1)
= Φ(1)
Example
If X is a normal random variable with parameters (µ, σ 2 ) = (3, 9), find (a) P(2 < X < 5); (b)
P(X > 0); (c) P(|X − 3| > 6).
Solution: (a)
2−3 X −3 5−3
P(2 < X < 5) = P( < < )
3 3 3
−1 2
= P( <Z < )
3 3
2 −1
= Φ( ) − Φ( )
3 3
≈ 0.3779
(b)
X −3 0−3
P(X > 0) = P( > ) = P(Z > −1)
3 3
= 1 − Φ(−1)
= Φ(1)
≈ 0.8411
The DeMoivre-Laplace Theorem
Theorem
Let 0 ≤ p ≤ 1 be fixed. Let Xn be a binomial random variable with parameters (n, p). Then for
any a < b,
The DeMoivre-Laplace Theorem
Theorem
Let 0 ≤ p ≤ 1 be fixed. Let Xn be a binomial random variable with parameters (n, p). Then for
any a < b,
!
Xn − np
P a≤ p ≤ b → Φ(b) − Φ(a)
np(p − 1)
as n → ∞.
The DeMoivre-Laplace Theorem
Theorem
Let 0 ≤ p ≤ 1 be fixed. Let Xn be a binomial random variable with parameters (n, p). Then for
any a < b,
!
Xn − np
P a≤ p ≤ b → Φ(b) − Φ(a)
np(p − 1)
as n → ∞.
Note: This is a special case of the Central Limit Theorem that will be proved later.
The Gamma Function
Definition
The gamma function Γ(α) is given by
Z ∞
Γ(α) = e −x x α−1 dx, α > 0.
0
The Gamma Function
Definition
The gamma function Γ(α) is given by
Z ∞
Γ(α) = e −x x α−1 dx, α > 0.
0
Properties
1. Γ(α) = (α − 1)Γ(α − 1) for α > 1. (Prove!)
The Gamma Function
Definition
The gamma function Γ(α) is given by
Z ∞
Γ(α) = e −x x α−1 dx, α > 0.
0
Properties
1. Γ(α) = (α − 1)Γ(α − 1) for α > 1. (Prove!)
2. Γ(1) = 1. (Prove!)
The Gamma Function
Definition
The gamma function Γ(α) is given by
Z ∞
Γ(α) = e −x x α−1 dx, α > 0.
0
Properties
1. Γ(α) = (α − 1)Γ(α − 1) for α > 1. (Prove!)
2. Γ(1) = 1. (Prove!)
3. Γ(n + 1) = n! for any interger n ≥ 0:
The Gamma Function
Definition
The gamma function Γ(α) is given by
Z ∞
Γ(α) = e −x x α−1 dx, α > 0.
0
Properties
1. Γ(α) = (α − 1)Γ(α − 1) for α > 1. (Prove!)
2. Γ(1) = 1. (Prove!)
3. Γ(n + 1) = n! for any interger n ≥ 0:
Γ(n + 1) =
The Gamma Function
Definition
The gamma function Γ(α) is given by
Z ∞
Γ(α) = e −x x α−1 dx, α > 0.
0
Properties
1. Γ(α) = (α − 1)Γ(α − 1) for α > 1. (Prove!)
2. Γ(1) = 1. (Prove!)
3. Γ(n + 1) = n! for any interger n ≥ 0:
Γ(n + 1) = nΓ(n) = n
The Gamma Function
Definition
The gamma function Γ(α) is given by
Z ∞
Γ(α) = e −x x α−1 dx, α > 0.
0
Properties
1. Γ(α) = (α − 1)Γ(α − 1) for α > 1. (Prove!)
2. Γ(1) = 1. (Prove!)
3. Γ(n + 1) = n! for any interger n ≥ 0:
Γ(n + 1) = nΓ(n) = n(n − 1)Γ(n − 1)
The Gamma Function
Definition
The gamma function Γ(α) is given by
Z ∞
Γ(α) = e −x x α−1 dx, α > 0.
0
Properties
1. Γ(α) = (α − 1)Γ(α − 1) for α > 1. (Prove!)
2. Γ(1) = 1. (Prove!)
3. Γ(n + 1) = n! for any interger n ≥ 0:
Γ(n + 1) = nΓ(n) = n(n − 1)Γ(n − 1) = . . . = n(n − 1) . . . 2 · Γ(1)
The Gamma Function
Definition
The gamma function Γ(α) is given by
Z ∞
Γ(α) = e −x x α−1 dx, α > 0.
0
Properties
1. Γ(α) = (α − 1)Γ(α − 1) for α > 1. (Prove!)
2. Γ(1) = 1. (Prove!)
3. Γ(n + 1) = n! for any interger n ≥ 0:
Γ(n + 1) = nΓ(n) = n(n − 1)Γ(n − 1) = . . . = n(n − 1) . . . 2 · Γ(1) = n!.
Gamma Random Variables
Definition
A random variable X is called a gamma random variable with parameters (α, λ), where λ > 0
and α > 0, if its density function is given by
( −λx
λe (λx)α−1
x ≥0
f (x) = Γ(α)
Gamma Random Variables
Definition
A random variable X is called a gamma random variable with parameters (α, λ), where λ > 0
and α > 0, if its density function is given by
( −λx
λe (λx)α−1
x ≥0
f (x) = Γ(α)
0 x <0
Gamma Random Variables
Definition
A random variable X is called a gamma random variable with parameters (α, λ), where λ > 0
and α > 0, if its density function is given by
( −λx
λe (λx)α−1
x ≥0
f (x) = Γ(α)
0 x <0
Note
1. If α = 1,
Gamma Random Variables
Definition
A random variable X is called a gamma random variable with parameters (α, λ), where λ > 0
and α > 0, if its density function is given by
( −λx
λe (λx)α−1
x ≥0
f (x) = Γ(α)
0 x <0
Note
1. If α = 1, X is simply the exponential random variable with parameter λ.
Gamma Random Variables
Definition
A random variable X is called a gamma random variable with parameters (α, λ), where λ > 0
and α > 0, if its density function is given by
( −λx
λe (λx)α−1
x ≥0
f (x) = Γ(α)
0 x <0
Note
1. If α = 1, X is simply the exponential random variable with parameter λ.
2. The special case for which α = n (a positive integer) is important in Statistics.
Functions of Random Variables
Note: If X is a random variable, then Y = g (X ) is also a random variable for any (real-valued)
function g of X .
Functions of Random Variables
Note: If X is a random variable, then Y = g (X ) is also a random variable for any (real-valued)
function g of X .
Example: Let X be a discrete random variable assuming one of the three values −1, 0 and 1
with probability 31 .
Functions of Random Variables
Note: If X is a random variable, then Y = g (X ) is also a random variable for any (real-valued)
function g of X .
Example: Let X be a discrete random variable assuming one of the three values −1, 0 and 1
with probability 31 . Then Y = X 2 is a random variable assuming the values 0 and 1 with
probabilities 31 and 23 , respectively.
Functions of Random Variables
Note: If X is a random variable, then Y = g (X ) is also a random variable for any (real-valued)
function g of X .
Example: Let X be a discrete random variable assuming one of the three values −1, 0 and 1
with probability 31 . Then Y = X 2 is a random variable assuming the values 0 and 1 with
probabilities 31 and 23 , respectively. Indeed
1
P(Y = 0) = P(X = 0) =
3
Functions of Random Variables
Note: If X is a random variable, then Y = g (X ) is also a random variable for any (real-valued)
function g of X .
Example: Let X be a discrete random variable assuming one of the three values −1, 0 and 1
with probability 31 . Then Y = X 2 is a random variable assuming the values 0 and 1 with
probabilities 31 and 23 , respectively. Indeed
1
P(Y = 0) = P(X = 0) =
3
and
2
P(Y = 1) = P(X 2 = 1) = P(X ∈ {−1, 1}) = P(X = −1) + P(X = 1) = .
3
Question: Given the pmf pX of a random variable X , what is the pmf of pY of Y = g (X )?
Question: Given the pmf pX of a random variable X , what is the pmf of pY of Y = g (X )?
Answer: Suppose that X assumes the values x1 , x2 , . . . with positive probabilities and that
Y = g (X ) assumes the values y1 , y2 , . . ..
Question: Given the pmf pX of a random variable X , what is the pmf of pY of Y = g (X )?
Answer: Suppose that X assumes the values x1 , x2 , . . . with positive probabilities and that
Y = g (X ) assumes the values y1 , y2 , . . .. Then
X X
pY (yj ) = P(Y = yj ) = P({xi | g (xi ) = yj }) = P(X = xi ) = pX (xi ).
xi :g (xi )=yj xi :g (xi )=yj
Functions of Random Variables: Continuous Case
Example: Let X be the uniform random variable on the interval [0, 1]. Find the density
function (pdf) of the random variable Y = X 2 .
Functions of Random Variables: Continuous Case
Example: Let X be the uniform random variable on the interval [0, 1]. Find the density
function (pdf) of the random variable Y = X 2 .
FY (y ) = P(Y ≤ y )
Functions of Random Variables: Continuous Case
Example: Let X be the uniform random variable on the interval [0, 1]. Find the density
function (pdf) of the random variable Y = X 2 .
FY (y ) = P(Y ≤ y )
= P(X 2 ≤ y )
Functions of Random Variables: Continuous Case
Example: Let X be the uniform random variable on the interval [0, 1]. Find the density
function (pdf) of the random variable Y = X 2 .
FY (y ) = P(Y ≤ y )
= P(X 2 ≤ y )
√
= P(X ≤ y )
Functions of Random Variables: Continuous Case
Example: Let X be the uniform random variable on the interval [0, 1]. Find the density
function (pdf) of the random variable Y = X 2 .
FY (y ) = P(Y ≤ y )
= P(X 2 ≤ y )
√
= P(X ≤ y )
√
= FX ( y )
Functions of Random Variables: Continuous Case
Example: Let X be the uniform random variable on the interval [0, 1]. Find the density
function (pdf) of the random variable Y = X 2 .
FY (y ) = P(Y ≤ y )
= P(X 2 ≤ y )
√
= P(X ≤ y )
√
= FX ( y )
√
= y
Functions of Random Variables: Continuous Case
Example: Let X be the uniform random variable on the interval [0, 1]. Find the density
function (pdf) of the random variable Y = X 2 .
FY (y ) = P(Y ≤ y )
= P(X 2 ≤ y )
√
= P(X ≤ y )
√
= FX ( y )
√
= y
1
Thus the density function of Y = X 2 is fY (y ) = √ for 0 < y < 1 and 0 otherwise.
2 y