0% found this document useful (0 votes)
61 views5 pages

WWS 507c Assignment 4 Solutions: October 27, 2015

This document provides solutions to problems from a statistics assignment. Problem 1 defines sample proportions and shows that sample proportions are unbiased estimators of population proportions. It also derives the variance of a sample proportion. Problem 2 works through hypothesis tests on a sample proportion. It calculates a t-statistic and p-value for a two-sided test and a one-sided test, and explains why the p-values differ between the two cases. Problem 3 calculates a t-statistic and p-value for another data set and hypothesis test, and determines whether to reject the null hypothesis based on a significance level of 5%.

Uploaded by

jake frei
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
61 views5 pages

WWS 507c Assignment 4 Solutions: October 27, 2015

This document provides solutions to problems from a statistics assignment. Problem 1 defines sample proportions and shows that sample proportions are unbiased estimators of population proportions. It also derives the variance of a sample proportion. Problem 2 works through hypothesis tests on a sample proportion. It calculates a t-statistic and p-value for a two-sided test and a one-sided test, and explains why the p-values differ between the two cases. Problem 3 calculates a t-statistic and p-value for another data set and hypothesis test, and determines whether to reject the null hypothesis based on a significance level of 5%.

Uploaded by

jake frei
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 5

WWS 507c

Assignment 4 Solutions

October 27, 2015

Problem 1

Part a.

Y1 2 {0 or 1}, Y2 2 {0 or 1}, . . . , Yn 2 {0 or 1}

so taking the sum Y1 + · · · + Yn simply counts the number of Yi ’s that are equal to 1 in the sample.
Hence, the sample proportion of 1’s is

#10 s in sample
p̂ =
n
Y1 + · · · + Yn
=
n
= Ȳn

Part b.

Unbiasedness means E[p̂] = p.

E[p̂] = E[Ȳn ]
n
X
1
= E[n Yi ]
i=1
X
1
= n E[Yi ]
i=1
1
= n (p + · · · + p)
| {z }
n
= p

1
Part c.

Suppose 2
y = E[Yi ] is the population variance of a single observation Yi . Then 2
y = p(1 p).

V ar[p̂] = V ar[Ȳn ]
n
X
1
= V ar[n Yi ]
i=1
n
X
1 2
= (n ) V ar[ Yi ]
i=1

Because Yi ’s are independent, the variance of a sum is the sum of variances. Then
n
X
1 2
V ar[p̂] = (n ) V ar[ Yi ]
i=1
n
X
1 2
= (n ) · V ar[Yi ]
i=1
1 2 2
= (n ) · (n · y)
1 2
= n y
1
= n p(1 p)

Problem 2

Part a.

215
p̂ = Ȳn = = .5375
400

Part b.
p p
By the central limit theorem, t stat = n Ȳn y .5 = n p̂ .5
⇠a N (0, 1), where y is the population
y p
standard deviation of a single observation. But y is very close to the value p̂(1 p̂), so we substitute
p
p̂(1 p̂) in the denominator.

p p̂ .5
np ⇠a N (0, 1)
p̂(1 p̂)

The p-value asks how “rare” your observation is, in the sense of: what is the probability of observing
a t-stat which is more “extreme” than what you see in-sample? In a two-sided hypothesis test

H0 : p = .5 vs. H1 : p 6= .5

this means the probability of observing a t-stat with absolute value greater than what you see.

Well, in-sample, we see a t-stat of


p .5375 .5
400 p = 1.5
.5375(1 .5375)

2
and the probability of seeing something more extreme is

P r(|N (0, 1)| > 1.5) = P r(N (0, 1) > 1.5) + P r(N (0, 1) < 1.5)
= .1336

Part c.
p p
By the central limit theorem, t stat = n Ȳn y .5 = n p̂ .5
y
⇠a N (0, 1), where y is the population
standard deviation of a single observation. We do not know the true population value y , but y
p p
should be very close to the value p̂(1 p̂), so we substitute p̂(1 p̂) in the denominator.

p p̂ .5
np ⇠a N (0, 1)
p̂(1 p̂)

The p-value asks how “rare” your observation is, in the sense of: what is the probability of observing
a t-stat which is more “extreme” than what you see in-sample? In a one-sided hypothesis test

H0 : p = .5 vs. H1 : p > .5

this means the probability of observing a t-stat greater than what you see.

Well, in-sample, we see a t-stat of


p .5375 .5
400 p = 1.5
.5375(1 .5375)

and the probability of seeing something more extreme on one side is

P r(N (0, 1) > 1.5) = P r(N (0, 1) > 1.5)


= .0668

Part d.

Our answers differ because we have different definitions of what it means to be more “extreme.” In a
two-sided hypothesis test, we have to allow for the possibility that the t-statistic is more extreme in
absolute value.

In a one-sided test, because we are only concerned about alternate hypotheses

H0 : p = .5 vs. H1 : p > .5

we only care about t-statistics being more extreme in the positive direction.

Part e.

Depends on what level tests you’re considering. For a two-sided 10% test, we do not reject because
the p-value is .1336 > .10. For a one-sided 10% test, we reject because the p-value is .0668. Clearly,
neither test rejects at the 5% level or lower.

3
Problem 3

Part a.

p p̂ .5
t stat = np
p̂(1 p̂)
p .54 .5
= 1055 p
.54(1 .54)
= 2.6068

The two-sided p-value is P r(|N (0, 1)| > 2.6068) = .0091 < .05. We reject the null.

Part b.

p p̂ .5
t stat = np
p̂(1 p̂)
p .54 .5
= 1055 p
.54(1 .54)
= 2.6068

The one-sided p-value is P r(N (0, 1) > 2.6068) = .0046 < .05. We reject the null.

Problem 4

The survey is drawn from some population of people, which may or may not be the “given” population.
We want to see if p is equal to the value .11 or not.

H0 : p = .11 vs. H1 : p 6= .11

Then
p p̂ .11
t stat = np
p̂(1 p̂)
p .24 .11
= 600 p
.24(1 .24)
= 7.456

The p-value is ⇡ 0 (notice that this is true for both one and two-sided alternative hypoth)

Problem 5

Part a.

4
We are asking if E[Ȳn ] = µY ...

1
E[Ȳn ] = E[n (Y1 + · · · + Yn )]
1
= n E[Y1 + · · · + Yn ]
1
= n (E[Y1 ] + . . . E[Yn ])
1
= n (µY + · · · + µY )
| {z }
n
1
= n · nµY
= µY

so yes, it is unbiased

Part b.

We’re asking ifE[Ȳn2 ] = µ2Y ...

Suppose n = 2, and Y1 , Y2 ⇠iid N (0, 1), so that µ2y = 0 and 2


y = 1. Then

E[Ȳ22 ] = E[((Y1 + Y2 )/2)2 ]


= (1/4) · E[(Y12 + Y22 + 2Y1 Y2 )]
= (1/4) · (E[Y12 ] + E[Y2 ]2 + E[2Y1 Y2 ])
2 2
= (1/4) · ( y + y + 0)
2
= y /2

= 1/2
6= µ2y

so no, it is not unbiased

Part c.

We’re asking if the distribution of Ȳn gets closer to µy . This follows from the law of large numbers,
since Y1 , . . . , Yn are iid. So yes, it is consistent.

Part d.

We’re asking if the distribution of Ȳn2 gets closer to µ2y .

Part c. tells us that the distribution of Ȳn gets closer to µy , and the continuous mapping theorem tells
us that any continuous function of Ȳn gets closer to µy , so long as the function is continuous at µy
itself. That is, for any continuous function G(·),

Ȳn !p µy ) G(Ȳn ) !p G(µy )

Clearly the function G(y) = y 2 is continuous, so

Ȳn2 !p µ2y

and yes, Ȳn2 is consistent for µy .

You might also like

pFad - Phonifier reborn

Pfad - The Proxy pFad of © 2024 Garber Painting. All rights reserved.

Note: This service is not intended for secure transactions such as banking, social media, email, or purchasing. Use at your own risk. We assume no liability whatsoever for broken pages.


Alternative Proxies:

Alternative Proxy

pFad Proxy

pFad v3 Proxy

pFad v4 Proxy