A Review of Basic Statistical Concepts: Answers To Odd Numbered Problems 1
A Review of Basic Statistical Concepts: Answers To Odd Numbered Problems 1
1. Descriptive Statistics
a. X = 21.32
b. S = 13.37
c. S2 = 178.76
d. If the policy is successful, smaller orders will be eliminated and the mean will
increase.
e. If the change causes all customers to consolidate a number of small orders into
large orders, the standard deviation will probably decrease. Otherwise, it is very
difficult to tell how the standard deviation will be affected.
d. We see that the 95% confidence intervals in b and c are not much different.
This explains why a sample of size n = 30 is often taken as the cutoff between
large and small samples.
1
H1: > 12.1 S = 1.7 X = 13.5
13.5 12.1
Z = 1.7 = 8.235
100
Reject H0 since the computed Z (8.235) is greater than the critical Z (1.645). The mean has
increased.
Since |2.67| = 2.67 > 1.96, reject H 0 at the 5% level. The mean satisfaction rating is
different from 5.9.
p-value: P(Z < 2.67 or Z > 2.67) = 2 P(Z > 2.67) = 2(.0038) = .0076, very strong
evidence against H 0
715 700
Z = 50 = 2.12
50
Since the calculated Z is greater than the critical Z (2.12 > 1.96), reject the null hypothesis.
The forecast does not appear to be reasonable.
p-value: P(Z < 2.12 or Z > 2.12) = 2 P(Z > 2.12) = 2(.017) = .034, strong evidence
against H 0
11. a.
2
b. Positive
13. This is a good population for showing how random samples are taken. If three-digit
random numbers are generated from Minitab as demonstrated in Problem 10, the selected
3
items for the sample can be easily found. In this population, = 0.06, so most students
will not reject the null hypothesis which so states this. The few students that erroneously
reject H0 demonstrate Type I error.
Hypothesis test:
H 0 : 44
two-sided test, = .02, critical value: |Z|= 2.33
H 1 : 44
X 44 45.2 44
Test statistic: Z 1.54
S/ n 10.3 / 175
As expected, the results of the hypothesis test are consistent with the confidence
interval for ; = 44 is not ruled out by either procedure.
CHAPTER 3
3. The secular trend of a time series is the long-term component that represents the growth or
decline in the series over an extended period of time. The cyclical component is the wave-
like fluctuation around the trend. The seasonal component is a pattern of change that
repeats
itself year after year. The irregular component measures the variability of the time series
after the other components have been removed.
5. The autocorrelation coefficient measures the correlation between a variable, lagged one or
more periods, and itself.
7. a. nonstationary series
b. stationary series
c. nonstationary series
d. stationary series
11. Classical decomposition, census II, Winter's exponential smoothing, time series multiple
regression, and Box-Jenkins methods. Examples are: electrical consumption,
summer/winter activities (sports like skiing), clothing, and agricultural growing seasons,
retail sales influenced by holidays, three-day weekends, and school calendars.
15. a. MPE
b. MAPE
c. MSE
50,056,722
r1 = = .895
55,912 ,891
H0: ρ1 = 0H1: ρ1 0
Reject if t < -2.069 or t > 2.069
k 1 11
1 2 ri 2 1 2 r1
2
1
SE( r k ) = i 1
= i 1
= = .204
24
n 24
r1 1 .895 0
t = = 4.39
SE(rk) .204
Since the computed t (4.39) is greater than the critical t (2.069), reject the null.
H0: ρ2 = 0H1: ρ2 0
Reject if t < -2.069 or t > 2.069
44,081,598
r2 = 55,912,891 = .788
k 1 2 1
1 2 ri 2 1 2 .895
2
2.6
SE( r k )= i 1
= i 1
= = .33
24
n 24
r1 1 .788 0
t = 2.39
SE(r1) .33
Since the computed t (4.39) is greater than the critical t (2.069), reject the null.
1.0
0.8
0.6
0.4
0.2
0.0
-0.2
-0.4
-0.6
-0.8
-1.0
1 2 3 4 5 6
7
21. a. Time series plot follows
The sample autocorrelations die out rapidly. This behavior is consistent with a
stationary series. Note that the sales data are not random. Sales in adjacent weeks
tend to be positively correlated.
Since, in this case, the residuals differ from the original observations by the constant
CHAPTER 4
8
MOVING AVERAGES AND SMOOTHING METHODS
1. Exponential smoothing
3. Moving average
7.
Price AVER1 FITS1 RESI1
19.39 * * *
18.96 * * *
18.20 18.8500 * *
17.89 18.3500 18.8500 -0.96000
18.43 18.1733 18.3500 0.08000
19.98 18.7667 18.1733 1.80667
19.51 19.3067 18.7667 0.74333
20.63 20.0400 19.3067 1.32333
19.78 19.9733 20.0400 -0.26000
21.25 20.5533 19.9733 1.27667
21.18 20.7367 20.5533 0.62667
22.14 21.5233 20.7367 1.40333
Accuracy Measures
MAPE: 4.63186 MAD: 0.94222 MSD: 1.17281
9. 3-month moving-average
9
1 9.29 * * *
2 9.99 * * *
3 10.16 9.8133 * *
4 10.25 10.1333 9.8133 0.436667
5 10.61 10.3400 10.1333 0.476667
6 11.07 10.6433 10.3400 0.730000
7 11.52 11.0667 10.6433 0.876667
8 11.09 11.2267 11.0667 0.023333
9 10.80 11.1367 11.2267 -0.426667
10 10.50 10.7967 11.1367 -0.636667
11 10.86 10.7200 10.7967 0.063333
12 9.97 10.4433 10.7200 -0.750000
Accuracy Measures
MAPE: 4.58749 MAD: 0.49111 MSD: 0.31931 MPE: .6904
Moving Average
Actual
11.5
Predicted
Forecast
Actual
Predicted
Forecast
10.5
Yield
Moving Average
Length: 3
MAPE: 4.58749
9.5
MAD: 0.49111
MSD: 0.31931
0 2 4 6 8 10 12 14
Time
b. 5-month moving-average
Accuracy Measures
MAPE: 5.58295 MAD: 0.60400 MSD: 0.52015 MPE: .7100
Moving Average
Actual
12
Predicted
Forecast
Actual
Predicted
11 Forecast
Yield
Moving Average
Length: 5
10
MAPE: 5.58295
MAD: 0.60400
MSD: 0.52015
0 2 4 6 8 10 12 14
Time
11.
Time Demand Smooth Predict Error
11
1 205 205.000 205.000 0.0000
2 251 228.000 205.000 46.0000
3 304 266.000 228.000 76.0000
4 284 275.000 266.000 18.0000
5 352 313.500 275.000 77.0000
6 300 306.750 313.500 -13.5000
7 241 273.875 306.750 -65.7500
8 284 278.938 273.875 10.1250
9 312 295.469 278.938 33.0625
10 289 292.234 295.469 -6.4688
11 385 338.617 292.234 92.7656
12 256 297.309 338.617 -82.6172
Accuracy Measures
MAPE: 14.67 MAD: 43.44 MSD: 2943.24
Actual
400
Predicted
Forecast
Actual
Predicted
Forecast
300
Alpha: 0.500
MAPE: 14.67
MSD: 2943.24
0 2 4 6 8 10 12 14
Time
12
13. a. = .4
Accuracy Measures
MAPE: 14.05 MAD: 24.02 MSD: 1174.50
b. = 0.6
Accuracy Measures
MAPE: 14.68 MAD: 24.56 MSD: 1080.21
13
d. No! When the residual autocorrelations shown above are examined, some of
them were found to be significant.
15. The autocorrelation function shows that the data are seasonal with a slight trend.
Therefore, the Winters’ model is used to forecast revenues.
14
An examination of the autocorrelation coefficients for the residuals of this
Winters’ model shown below indicates that none of them are significantly
different from zero.
15
CHAPTER 5
1. The purpose of decomposing a time series variable is to observe its various elements
in isolation. By doing so, insights into the causes of the variability of the series are
frequently gained. A second important reason for isolating time series components
is to facilitate the forecasting process.
3. The basic forces that affect and help explain the trend-cycle of a series are
population growth, price inflation, technological change, and productivity increases.
5. Weather and the calendar year such as holidays affect the seasonal component.
7. a.
YEAR Y T C
1980 11424 11105.2 102.871
1981 12811 12900.6 99.306
1982 14566 14695.9 99.116
1983 16542 16491.3 100.307
1984 19670 18286.7 107.565
1985 20770 20082.1 103.426
1986 22585 21877.4 103.234
16
1987 23904 23672.8 100.977
1988 25686 25468.8 100.855
1989 26891 27263.6 98.633
1990 29073 29059.0 100.048
1991 28189 30854.3 91.362
1992 30450 32649.7 93.263
1993 31698 34445.1 92.025
1994 35435 36240.5 97.777
1995 37828 38035.8 99.454
1996 42484 39831.2 106.660
1997 44580 41626.6 107.095
b. Y = 9309.81 + 1795.38X
9. Y = TS = 850(1.12) = $952
13. a. The regression equation and seasonal indexes are shown below. The trend is
the most important but both trend and seasonal should be used to forecast.
Seasonal Indices
Period Index
1 0.96392
2 1.02559
3 1.00323
4 1.00726
Accuracy of Model
MAPE: 3.2
MAD: 87.3
MSD: 11114.7
b. Forecasts
c. The forecast for third quarter is very accurate (3,349.5 versus 3,340)
The forecast for fourth quarter is high compared to Value Line (3,388.7
versus 3,300). Following are computer generated graphs for this problem:
Actual
3400
Predicted
3200 Forecast
Actual
3000 Predicted
Forecast
2800
Sales
2600
2400
0 5 10 15 20 25 30 35 40 45
Time
18
15. a. Addivitive Decomposition Ln(Sales)
19
Data LnCavSales
Length 77.0000
NMissing 0
Yt = 4.61322 + 2.19E-02*t
Seasonal Indices
Period Index
1 0.33461
2 -0.01814
3 -0.40249
4 -0.63699
5 -0.71401
6 -0.57058
7 -0.27300
8 -0.00120
9 0.46996
10 0.72291
11 0.74671
12 0.34220
c. Forecasts
20
Date Period Forecasts(Ln(Sales)) Forecasts(Sales)
17. a.
600
19. a. JAN = = 500
1.2
500(1.37) = 685 people estimated for FEB
c. 5
CHAPTER 6
REGRESSION ANALYSIS
23
1. Option b is inconsistent because the regression coefficient and the correlation coefficient
must have the same sign.
Analysis of Variance
Source DF SS MS F P
Regression 1 92432 92432 20.47 0.000
Error 8 36121 4515
Total 9 128552
b. Y = 828 + 10.8X
5.
. The regression equation is
Cost = 208 + 70.9 Age
Analysis of Variance
Source DF SS MS F P
Regression 1 634820 634820 50.96 0.000
Error 7 87197 12457
Total 8 722017
a.
25
1100
1000
900
800
700
Cost
600
500
400
300
200
0 5 10
Age
b. Positive
d. Y = 208.2033 + 70.9181X
e. H0 : = 0
H1: 0
Reject H0 if t < -2.365 or t > 2.365
sb1 s y x / (X X ) 2
111 .6 / 126.22 9.93
b1 70.9181
t 7.14
sb1 9.93
Reject the null hypothesis. Age and maintenance cost are linearly related in
the population.
26
7. The regression equation is
Orders = 15.8 + 1.11 Catalogs
b. Y = 15.846 + 1.1132X
c. sy.x = 5.57
d. Analysis of Variance
Source DF SS MS F P
Regression 1 317.53 317.53 9.58 0.011
Error 10 331.38 33.14
Total 11 648.92
f. H0 : = 0
H1: 0
Reject H0 if t < -3.169 or t > 3.169
sb1 s y x / (X X ) 2
5.757 / 256.25 .36
b1 1.1132
t 3.09
sb1 .36
Fail to reject the null hypothesis. Orders received and catalogs distributed
are not linearly related in the population.
g. H0 : = 0
H1: 0
Reject H0 if F > 10.04
MSR 317.53
F 9.58
MSE 33.14
27
9. a. The two firms seem to be using very similar rationale since r = .959.
c. If the data constitute a sample, the point estimate for Y would be the same but the
interval estimate would use the standard error of the forecast instead of S y.x, since
the variability of sample regression lines around the population regression line
would have to be considered along with the scatter of sample Y values around the
calculated line. The probability of winning the bid would then involve the same
problem as in c. above except that sf would be used in the calculation:
Analysis of Variance
Source DF SS MS F P
Regression 1 559607 559607 26.88 0.000
Error 7 145753 20822
Total 8 705360
a.
28
b. Y = 5375 Y2 = 3,915,429 X = 100.6
X2 = 1151.12 XY = 56,219.8
r = -0.8907
d. 44.9474
e. r2 = .7934
Source DF SS MS F P
Regression 1 14331 14331 231.77 0.000
29
Error 11 680 62
Total 12 15011
a.
d.
30
Examination of the residuals shows a curvilinear pattern.
e. The best model transforms the predictor variable and uses X2 as its
independent variable.
Analysis of Variance
Source DF SS MS F P
Regression 1 14951 14951 2727.00 0.000
Error 11 60 5
Total 12 15011
Unusual Observations
Obs Sizesqr Defects Fit StDev Fit Residual St Resid
3 5625 6.000 10.367 0.920 -4.367 -2.03R
g.
c. F 72.6 , p-value = .000 < .10. The regression is clearly significant at the
.10 level.
f. Unusual Observations
Obs Play Cost OpExpens Fit SE Fit Residual St Resid
7 18.0 60.00 42.31 1.64 17.69 3.45R
32