Probability: Event Spaces and Sample Spaces
Probability: Event Spaces and Sample Spaces
It is that time in the quarter (it is still Week 1) when we get to talk about probability. We are again
going to build this up from first principles. We will heavily use the rules of counting that we learned
earlier this week.
Probability
In the 20th century, people figured out one way to define what a probability is:
n(E)
P(E) = lim
n→∞ n
In English this reads: say you perform n trials of an experiment. The probability of a desired event
E is the ratio of trials that result in an outcome in E to the number of trials performed (in the limit
as your number of trials approaches infinity).
You can also give other meanings to the concept of a probability, however. One common meaning
ascribed is that P(E) is a measure of the chance of E occurring.
I often think of a probability in another way: I don’t know everything about the world. As a result I
have to come up with a way of expressing my belief that E will happen given my limited knowledge.
This interpretation acknowledges that there are two sources of probabilities: natural randomness
and our own uncertainty.
– 2–
Axioms of Probability
Here are some basic truths about probabilities:
Axiom 1: 0 ≤ P(E) ≤ 1
Axiom 2: P(S) = 1
Axiom 3: If E and F mutually exclusive (E ∩ F = ∅), then P(E) + P(F) = P(E ∪ F)
You can convince yourself of the first axiom by thinking about the definition of probability above:
when performing some number of trials of an actual experiment, it is not possible to get more
occurrences of the event than there are trials (so probabilities are at most 1), and it is not possible
to get less than 0 occurrences of the event (so probabilities are at least 0).
The second axiom makes intuitive sense as well: if your event space is the same as the sample space,
then each trial must produce an outcome from the event space. Of course, this is just a restatement
of the definition of the sample space; it is sort of like saying that the probability of you eating cake
(event space) if you eat cake (sample space) is 1.
Identity 2:
Identity 3:
This last rule is somewhat complicated, but the notation makes it look far worse than it is. What we
are trying to find is the probability that any of a number of events happens. The outer sum loops
over the possible sizes of event subsets (that is, first we look at all single events, then pairs of events,
then subsets of events of size 3, etc.). The “-1” term tells you whether you add or subtract terms
with that set size. The inner sum sums over all subsets of that size. The less-than signs ensure that
you don’t count a subset of events twice, by requiring that the indices i 1, . . . , ir are in ascending
order.
Here’s how that looks for three events (E1 , E2 , E3 ):
P(E1 ∪ E2 ∪ E3 ) = P(E1 ) + P(E2 ) + P(E3 )
− P(E1 E2 ) − P(E1 E3 ) − P(E2 E3 )
+ P(E1 E2 E3 )
– 3–
Probability with equally likely outcomes: For a sample space S in which all outcomes are equally
likely,
1
P(Each outcome) =
|S|
and for any event E ⊆ S,
number of outcomes in E |E|
P(E) = =
number of outcomes in S |S|