AP Stat Review
AP Stat Review
Chapter 2: Distributions
𝑥−µ
○ map N(µ, σ) to standard Normal distribution: 𝑧 = σ
is
distributed N(0,1)
● How to assess normality: graphing, Empirical Rule, Normal Probability
Plot (linear → normal, turned right = skewed right)
Chapter 3: Correlations
○ residual plot (x: exp, y: resid): checks if LSR line was a good model
of the data
■ no pattern → LSR line was a good model of data
○ standard deviation of residuals (RSE): s=__ tells us that when we
try to predict (res) from (exp), we will typically be off by __ (units
of exp). (same as in 12.1)
● coefficient of determination: correlation squared
2
○ About (𝑟 · 100%) of the variation in (res) is accounted for by the
linear model relating (res) to (exp).
Chapter 4: Studies
4.2: Experiments
● observational study: no manipulation
● experiment: cause/effect
○ controls (does not eliminate, rather attempts to average out
differences) confounding: comparison (most important; even
voluntary sample compared with control group can show
Chapter 5: Probability
○ σ𝑋 = 𝑛𝑝(1 − 𝑝)
○ SRS lacks replacement: not independent trials
■ 10% Condition: binomial distribution can approximately
1
model counts of successes when 𝑛 ≤ 10
𝑁 (population is
much larger than sample)
○ Large Counts Condition: When 𝑛𝑝 ≥ 10, 𝑛(1 − 𝑝) ≥ 10 (at least
10 expected successes and failures each; recall 6.1), the
distribution of X is approximately Normal
● Geometric random variable Y: counts of TRIALS taken to get ONE
success
○ BINS without N
○ geometric distribution has parameter n
7.1: Overview
● Parameter (µ, 𝑝) describes Population, Statistic (𝑥, 𝑝) describes Sample
○ statistic (𝑥, 𝑝) estimates corresponding parameter (µ. 𝑝)
○ unbiased estimator: statistic whose mean is equal to value of
parameter being estimated (does NOT consistently either
over/underestimate value of parameter)
● take every possible sample of size n & graph some statistic → sampling
distribution: frequency distribution of ALL possible values of a statistic
(we can only approximate)
8.1: Overview
● confidence interval: estimate population parameter from sample
statistic
● C% confidence interval: statistic (unbiased point estimator) ± margin of
error, where margin of error = critical value * standard deviation of
population (if don’t know standard deviation of population, must use
standard error as estimate)
○ critical value depends on confidence level
○ confidence level does NOT tell the chance that a particular interval
captures the parameter; the parameter is fixed already, so one you
make a sample, the interval either contains or does not contain
the parameter
^
❖Choosing sample size to bound margin of error: to guess 𝑝, use past
^
study if given, otherwise use 𝑝 = 0. 5 (maximizes margin of error)
^ ^
* 𝑝(1−𝑝)
➢solve 𝑧 · 𝑛
≤ 𝑚𝑎𝑥 𝑀𝐸 and return ⌈n⌉ (nearest greater
integer)
9.1: Overview
● significance test: use sample statistic to decide between competing
hypotheses about population parameter
○ null hypothesis (H0): original claim (no-change claim)
■ parameter = value
○ alternative hypothesis (Ha): claim to try to prove (aka find
convincing evidence for)
■ parameter (<, >, ≠) value
● ≠ is a two-sided test
● P-value: assuming H0 were true, the probability that the statistic would
take a value at least as extreme as statistic (in the direction(s) of Ha)
○ think proof by contradiction: getting a very unlikely (in the
direction(s) of Ha) statistic according to H0 suggests that Ha is true
○ P-value < α (significance level): statistically significant at level α
■ reject H0, convincing evidence of Ha (otherwise fail to reject
H0, no convincing evidence of Ha)
■ if α not given, set α = 0. 05
■ draw appropriate distribution alongside calculation
○ interpretation: “Assuming (H0), there would be a (P-value)
probability of getting a (statistic definition) as (extreme or more
extreme; specify direction) than (sample statistic value).”
● Type I error: rejected H0 when H0 was actually true
○ significance level α = 𝑃(𝑇𝑦𝑝𝑒 𝐼) (= 100% - Confidence%)
● Type II error: failed to reject H0 when Ha was actually true (H0 was
actually false)
9.2: Proportion
FRQ Response: (memorize entire thing below)
[State] We want to test at the 5% significance level.
H0: p = (p0)
Ha: p ≠ (p0) (operator depends on problem)
where p is the proportion…[define]
[Plan] If conditions are met, we will conduct a one-sample z-test for p at the
α = 0. 05 level.
Conditions: see 8.2
[Do: either manual or technology]
^
^ 𝑝−𝑝0
1. 𝑝 =..., test statistic 𝑧 = , then using technology, P-value =
𝑝0(1−𝑝0)
𝑛
9.3: Mean
FRQ Response:
● same outline as 9.2, except:
○ procedure: one-sample t-test for µ at the α = 0. 05 level
○ Conditions: same as 8.3
○ Do (2 ways):
𝑥‾−µ0
1. 𝑥‾ =..., 𝑠𝑥 =..., 𝑑𝑓 = 𝑛 − 1, test statistic 𝑡 = 𝑠𝑥 , then using
𝑛
10.1: Proportions
^ ^
● sampling distribution of 𝑝1 − 𝑝2 has mean 𝑝1 − 𝑝2 and standard
10.2: Means
● sampling distribution of 𝑥‾1 − 𝑥‾2 has mean µ1 − µ2 and standard
2
○ df = # categories - 1 (larger df = flatter χ distribution)
2
2 (𝑜𝑏𝑠𝑒𝑟𝑣𝑒𝑑−𝑒𝑥𝑝𝑒𝑐𝑡𝑒𝑑) 2
1. chi-square statistic χ = ∑ 𝑒𝑥𝑝𝑒𝑐𝑡𝑒𝑑
, then P-value = χ cdf
2
a. draw χ distribution with all three statistics
2 2
2. Using technology, χ GOF-Test (report df, χ , and P-value, and draw
2
χ distribution with all three statistics)
2
● either method, suggested to write out formula for χ with few
components and … (partial credit in case of error)
● follow-up analysis (if asked): CNTRB list → “The largest contributors to
the statistic are (category 1) (component value) and (category 2)
(component value). There are fewer/more (category 1) and fewer/more
(category 2) than expected.” (must look back at lists to see if
fewer/more than expected since all components are positive)
2
○ either method, suggested to write out formula for χ with
few components and … (partial credit in case of error)
● follow-up analysis (if asked): same as 11.1 but name cells (by two
labels); usually given computer output (otherwise, manual
2
method is χ -Test → manually plug in expected counts from matrix
into list)
● convincing evidence of association between TWO categorical variables
within ONE population: chi-square test for independence
○ H0: There is no association between (categorical var 1) and
(categorical var 2).
○ Ha: There is an association between (categorical var 1) and
(categorical var 2).
○ everything else is the same as homogeneity (except for conclusion
wording, of course)
● inference
○ Conditions: LINER
■ Linear: linear association (see 3.1), no leftover pattern in
residual plot
■ Independent: observations are independent (one
individual’s response does not affect another’s), 10%
Condition
2
● given r2, 𝑟 =± 𝑟 (depends on slope b)
● interpretations:
○ SEb: “If we take many identical samples of size n, the sample slope
will typically differ from the true slope by (SEb).” (b vs β)
○ s: “The predicted (explanatory) will typically differ from the
^
population regression line by (s).” (𝑦 vs y; same as in 3.1)
^
■ transformed equation is in form log(𝑦) = # + # log(𝑥)
𝑥
● exponential model: original relationship is 𝑦 = 𝐶 + 𝑎𝑏
^ ^
○ apply logarithm onto 𝑦: log 𝑦 = # + #𝑥
^
● What if both methods (log both, log 𝑦) result in a linear graph?
○ whichever residual plot has more random scatter
■ if residual plots are roughly same, check for smaller s or
bigger r2