Sample Size Determination
Sample Size Determination
net/publication/321769343
CITATIONS READS
0 387
1 author:
Alessandro Vasta
Baxter International
2 PUBLICATIONS 14 CITATIONS
SEE PROFILE
Some of the authors of this publication are also working on these related projects:
All content following this page was uploaded by Alessandro Vasta on 18 March 2019.
Alessandro Vastaa
This is a Manuscript of an article submitted to Taylor & Francis in Quality Engineering
Journal on 09-Mar-2019 , for possible online publication at: http://www.tandfonline.com/
ARTICLE HISTORY
Compiled March 19, 2019
ABSTRACT
In traditional industrial verification process, when the aim is the conformance to
specifications, it’s difficult finding an affordable statistical method reaching this
purpose.
Most of data tables included in industrial procedures and standards could not
be suitable for verification tests or otherwise they could lead to unknown or very
high values of Type I error. The practitioner could be tempted referring distribution
free intervals leading to very high sample sizes. A two-sided tolerance interval in a
normal universe, combined with a bivariate statistical hypothesis test can be used
to address this problem. The proposed approach leads to a determination of the
sample size with pre established probabilities of Type I and Type II errors, essential
for risk and reliability estimations.
Finally, a novel approach for the tolerance interval testing factors determination is
proposed. This approach calculates the testing factors based on the deviation of the
mean and the variance from the null hypothesis, when a specified value of Type II
error is met. The deviations of the mean and variance are determined in such a way
that a certain proportion of the population falls within the specification limit rather
than a specific proportion of the population within the sampled tolerance interval.
KEYWORDS
Tolerance Interval; Sample Size; Confidence; Population; Significance; Power;
Verification; Risk Analysis; Reliability
1. Introduction
2
and setting the Type I error to 1%, the sample size will be n=25, (this values can
be calculated with the attached R program ”Sample size by spec.r” setting all input
parameters according the above example and the parameter mode sample_find = T).
The R library ”EnvStats” [12] with the function tolIntNormN() compute the sample
size necessary to achieve a specified half-width of a tolerance interval for a normal
distribution, given the estimated standard deviation, coverage, and confidence level.
Nevertheless, if these half-width values will be used to set specification limits running
a verification test, as in the previous example, high values of Type I error probability
should be expected.
Regarding hypothesis testing, Ryan [10] points out that the null hypothesis are usu-
ally defined in comparing two means that are equal, which implies that the difference
between them is zero. In practice the difference almost certainly won’t be zero, so the
difference between means in the null hypothesis shouldn’t be specified as zero. Equiv-
alence testing should be formalized in place of equality test. So, if the difference of the
two means is less than ∆µ = µ1 -µ0 (i.e. with µ1 > µ0 ), then the means are equivalent.
µ1 could be an alternative hypothesis, with the null hypothesis µ0 such that the dif-
ference between the two means is lower or equal to ∆µ. This equivalence concept is
one of two specific cases that will be considered in the next sections (see sect. 2.2 and
2.6) about null and alternative hypothesis definition H0 , Haµ1 & Haσ1 .
Many references Meeker [11], Wilks, Kirkpatrick [3, 6] and Faulkenberry [14] are still
based on a population proportion precision criteria where the sample size is chosen
large enough such that:
- The probability is γ (large) that at least a proportion P of the population will be
included within tolerance interval and
0 0
- The probability is δ (small) that more than a proportion P with P > P will be
included.
It’s worth noting that the probability δ is a decreasing function of the sample size n, so
higher the sample size lower is the uncertainty determining the population proportion.
Another recent reference, Young [15], propose a criteria in choosing P’ and δ avoiding
subjectivity, in particular P’ is determined as a function of the margin between the
specification limits and the expected mean µ̃ derived from historical data; thus guiding
the practitioner achieving an acceptable result concerning Type I error and minimum
sample size.
The aim of this paper, is to achieve an exact solution, starting from pre established
values of Type I (α) and also Type II (β) error probabilities. This solution is expect-
ed to find the minimum sample size so that at least an assigned proportion of the
population falls within the specification limits with the specified power (1 - β).
3
1.1. Basic formulation
Let’s suppose we verify a requirement where a measurable entity x, represented by
a random variable with a normal probability distribution with known mean µ and
standard deviation σ, shall be found inside a proper specification range
[LSL; USL]. The Figure 1 shows a particular situation where the above condition is
met with a probability 100γ% = 95%.
LT L = x̄ − k2 ∗ s (1)
U T L = x̄ + k2 ∗ s
4
Where:
s
n−1
k2 = r (2)
χ2n−1,γ
where:
n is the number of observations (the sample size)
χ2 n−1,γ denotes the number for which the probability that χ2 distribution with n-1
degrees of freedom will exceed this number is γ
r is the solution of the following equation:
1
Z √ +r
1 n −t2
√ e 2 dt = P
2π √
1
n
−r
Each UTL and LTL are random processes, then the proposed methodology discussed
in the next sections, will deal with the determination of the probability distribution
of LTL and UTL.
2. Methodology
5
max(abs(LTL), UTL) = ATL ≤ ASL
In the relationships above we derived the one-sided tolerance limit ATL starting
from the original two-sided interval; therefore equations (1) can be rewritten in this
way:
AT L = |x̄| + k2 ∗ s (3)
The Figure 2 shows that the probability density is not normal and strongly asym-
metrical.
The Figure 3 below shows instead the histogram with a large sample size n = 1000.
This figure suggests that the probability distribution is almost normally distributed.
This can be evaluated applying the Anderson Darling normality test; the results for
6
n=1000 are reported below:
σ
f|x̄| (x, n) = 2N orm(x, 0, √ ); x ≥ 0 (4)
n
Let’s call instead fk2s (x, n) the PDF of the second term k2*s.
Starting from the working hypothesis, with xi the independent observation from a
normal distribution, Cochran’s theorem [13] shows that s2 follows a scaled chi-square
distribution where: √
(n−1) 2 2 (n − 1), that implies : n−1 s ∼ χ(n − 1)
σ2 s ∼ χ σ
Then, we can write:
√ √
n−1 n−1
fk2s (x, n) = χ(x , n − 1) (5)
k2 ∗ σ k2 ∗ σ
√
n−1
It’s worth noting that the scaling factor k2∗σ included in χ() function argument, in
order to refer the term k2*s in the abscissa axis, shall be appliedR also to the ordinate
+∞
axis to maintain the following distribution probability property: 0 fk2s (x, n)dx = 1
Finally the PDF of ATL is the convolution of the two above mentioned functions
f|x̄| (x, n) and fk2s (x, n), that is:
Z ∞
T IH0 (AT L, n) = fk2s (AT L − x, n)f|x̄| (x, n)dx (6)
0
Figure 4 shows in the dashed line the plot of PDF function defined for ATL accord-
7
ing (6), overlapped to histogram of equation (3) computation with N=100,000 and n
= 3. Both of them have been calculated with R program ”Sample size by spec.r” in
attachment.
Z ASL
(1 − α) = T IH0 (z, n)dz (7)
0
The equation (7) provide the value of ASL depending on the set of parameters (n,
µ0 , σ0 , P, γ, α). It’s worth noting that in equation (7) there is a dependency on n, P
and γ embedded in factor k2 in the term fk2s (x, n) .
Equation (7) has been solved numerically in R and, the related statistical results to
evaluate the experimental values of significance level α, are reported in next section 2.4.
8
2.3. Extension to general case µ 6= 0
Thanks to the independency of µ from k2 as already discussed in section 1.2, it follows
that:
U SL = µ + ASL (8)
LSL = µ − ASL
Where nout is the number of occurrences that doesn’t satisfy in N reiterations the
following relationships:
USL = ASL ≥ UTL
LSL = -ASL ≤ LTL
The additional parameters used in this experimental test are:
9
Table 1. Simulation results and ASL calculation for P=0.95; γ=0.95; α=0.05
n UTLmean ASL s(UTL) UTLmax αe % MAX Pxout % k2
3 8.8 17.6 4.65 35.8 5.11 4.34 0 9.92
4 5.86 10.7 2.53 18.5 4.93 4.45 0 6.37
5 4.77 8.21 1.79 14.2 5.06 4.41 0 5.08
6 4.21 6.93 1.42 10.9 4.99 4.82 0 4.41
7 3.84 6.14 1.19 9.4 4.89 4.52 0 4.01
8 3.6 5.61 1.04 8.41 5.01 4.76 0 3.73
9 3.43 5.22 0.93 7.8 5.02 4.76 0 3.53
18 2.78 3.81 0.534 5.23 4.96 5.03 0.252 2.82
42 2.41 3.01 0.307 3.91 5.06 5.13 10.4 2.43
100 2.31 2.59 0.169 3.14 4.97 5.35 61.7 2.23
Let’s consider now the results in the column αe %. The expected theoretical value
is 5%, the results are close to this value but never equal. Nevertheless, if we perform
an hypothesis test of one binomial proportion for an event with 5% probability of
occurrence over a number of trials N=100,000, we obtain the following 95% confidence
interval:
In Table 2 the same results as Table 1 are shown, but with the following parameters
changes:
Table 2. Simulation results and ASL calculation for P=0.99; γ=0.99; α=0.01
n UTLmean ASL s(UTL) UTLmax αe % MAX Pxout % k2
3 25.8 62.8 13.5 104 1.04 4.54 0 29.1
4 13.4 28.7 5.66 42.5 0.971 4.5 0 14.5
5 9.63 19.1 3.52 27.9 1.01 4.57 0 10.3
6 7.9 14.8 2.58 21.6 0.994 4.79 0 8.3
7 6.9 12.4 2.06 16.8 1.06 4.64 0 7.19
8 6.24 10.8 1.73 14.4 0.977 4.64 0 6.47
9 5.78 9.75 1.5 13.8 1 6.08 0 5.97
18 4.24 6.26 0.769 7.62 1.02 5 0 4.31
42 3.46 4.54 0.415 5.66 1.01 4.9 0.03 3.49
100 3.17 3.71 0.229 4.18 1.04 5.52 1.99 3.1
10
2.5. Sample Size Determination
In ”Sample Size by Spec.r” R program, an iterative algorithm determines the sam-
ple size n assuming the function ASLtest (n), solution of equation (9), a monotone
decreasing function.
Z ASLtest (n)
(1 − α) = T IH0 (z, n)dz (9)
0
Where USL and LSL are the desired specification limit and µ0 = (USL + LSL) / 2
the expected mean.
Being n an integer value, the above relationship (10) cannot be generally satisfied
with = operator, this means that the specified significance level is correspondent to
ASLtest . If instead ASL is used for the test limits, a value ≤ α for the significance level
should be expected.
As an example, Table 3 reports the results of ASLtest , k2 and sample size n, ( in the
rightmost columns gruoup), for some specific values of input data, ( in the leftmost
columns gruoup), with σ0 = 1.
Table 3. Sample size determination
ASL Population Confidence Significance ASLtest k2 Sample Size
P% γ% α% n
4 95 95 5 3.983 2.903 16
4 95 95 1 3.956 2.723 21
4 99 99 5 3.997 3.344 54
4 99 99 1 4 3.231 69
4 99.9 99.9 1 3.999 3.666 446
5 95 95 5 4.923 3.379 10
5 95 95 1 4.828 3.081 13
5 99 99 5 4.99 3.904 25
5 99 99 1 5 3.706 31
5 99.9 99.9 1 4.991 4.202 101
Finally, the proposed sample size determination approach requires to define a Null
Hypothesis H0 :(µ0 ; σ0 ) with respectively µ0 , σ0 mean and standard deviation of a
normal process that, combined with statistical parameter (P, γ, α), provide the sample
size for null hypothesis testing.
It’s worth noting that a preliminary estimation of µ0 and σ0 is required. This can
be performed with a pilot study with data external to the data batch under test or
also internal; for instance, Ryan [10] in various application, such as medical, deals with
sequential sample size selection because more than one phase can be involved, such as
in clinical trials, and sequential sample size determination can also be used within a
given phase. Thus, sample size reestimation can occur as a study progresses.
11
2.6. Considerations about the Power of the Test
This section will take into consideration two particular alternative hypotheses Haµ1
and Haσ1 , where in the first case: Haµ1 :(µ 6= µ0 ; σ0 ). That is, only the mean µ deviates
from the null hypothesis.
So, we would like finding that µ1 = µ0 ± ∆µ leading to an assigned Type II error
probability β. According what already discussed in sect. 1.2, without loss of generality
we can assume µ0 = 0 and µ1 = ∆µ > 0.
In order to calculate µ1 , leading to a β% probability to accept the null hypothesis,
we need to find the probability distribution as already done in sect. 2.1.
In this particular case the equation (4) becomes:
σ0 σ0
f|x̄| (x, n, µ1 ) = N orm(x, µ1 , √ ) + N orm(−x, µ1 , √ ); x ≥ 0 (11)
n n
While for the term k2*s, the previous function discussed in 2.1 fk2s (x,n) is still valid:
√ √
n−1 n−1
fk2s (x, n) = χ(x , n − 1) (12)
k2 ∗ σ0 k2 ∗ σ0
Finally, the PDF of ATL is the convolution of the two above mentioned functions
f|x̄| (x, n, µ1 ) and fk2s (x, n), that is:
Z +∞
T IHaµ1 (AT L, n, µ1 ) = fk2s (AT L − x, n)f|x̄| (x, n, µ1 )dx (13)
0
Let’s now introduce the parameter β as the maximum probability of accepting the
null hypothesis H0 when in fact Haµ1 is true.
µ1 is the solution of the following equation:
Z ASL
β= T IHaµ1 (z, n, µ1 )dz (14)
0
With a similar approach we can take into consideration another particular alter-
native hypothesis Haσ1 where: Haσ1 :(µ0 = 0 ; σ1 > σ0 ). That is, only the standard
deviation σ deviates from the null hypothesis.
More in detail we would like finding that σ1 = σ0 *∆σ with ∆σ >1 leading to an
assigned Type II error probability β.
In order to calculate σ1 , leading to a β% probability to accept the null hypothesis,
we need to follow the same approach above described.
The equation (11) becomes:
12
σ1
f|x̄| (x, n, σ1 ) = 2N orm(x, 0, √ ); x ≥ 0 (15)
n
√ √
n−1 n−1
fk2s (x, n, σ1 ) = χ(x , n − 1) (16)
k2 ∗ σ1 k2 ∗ σ1
Then:
Z +∞
T IHaσ1 (AT L, n, σ1 ) = fk2s (AT L − x, n, σ1)f|x̄| (x, n, σ1 )dx (17)
0
Z ASL
β= T IHaσ1 (z, n, σ1 )dz (18)
0
Finally the test will have a power = 1 - β when the alternative hypothesis will
imply a standard deviation ratio ∆σ = σ1 /σ0
Every sample iteration is always obtained calculating tolerance limits with equa-
tion (1) for each sample size n.
In Table 4 and 5 are reported output parameters and simulation results for all
sample sizes and data input combinations already taken in consideration for Table 3,
where the following additional parameters are shown :
• Pwr = (1 - β) = the probability to reject the null hypthesis when the alteranative
hypothesis Haµ1 or Haσ1 are true.
• µ1 = the mean deviation correspondent to the specified Power.
• σ1 = the σ deviation correspondent to the specified Power.
• Pµ1 = the proportion of population falling within the specification limits when
the alternative hypothesis Haµ1 is true.
13
• Pσ1 = the proportion of population falling within the specification limits when
the alternative hypothesis Haσ1 is true.
• αe is the experimental % probability to reject the Null hypothesis H0 when H0
in fact is true
• Pwrµ1 = 100 - β(%) is the experimental % probability to reject the null hypoth-
esis H0 when in fact Haµ1 is true.
• Pwrσ1 = 100 - β(%) is the experimental % probability to reject the null hypoth-
esis H0 when in fact Haσ1 is true.
In Tables 4 the testing factors in the rightmost columns group are calculated solving
the equations (7), (14) and (18), while in the rightmost columns group of Table 5 are
reported the simulation results always obtained by N=100,000 trials. All of these
results are obtained by the attached R program ”Tolerance interval hypothesis test
sigma.r”. The population Pµ1 and Pσ1 are calcuated according the following equations:
U SL − µ1 LSL − µ1
Pµ1 = Φ( ) − Φ( ) (19)
σ0 σ0
U SL LSL
Pσ1 = Φ( ) − Φ( )
σ1 σ1
where Φ(·) indicates the the standard normal cumulative distribution function.
Table 4. Testing factors and alterantive hypotheis deviations
ASL P% γ% α% Pwr% ASLtest k2 n µ1 σ1 Pµ1 % Pσ1 %
4 95 95 5 95 3.983 2.903 16 2.061 1.822 97.4 97.2
4 95 95 1 95 3.956 2.723 21 2.041 1.841 97.5 97
4 99 99 5 99 3.997 3.344 54 1.465 1.488 99.4 99.3
4 99 99 1 99 4 3.231 69 1.467 1.498 99.4 99.2
4 99.9 99.9 1 99.9 3.999 3.666 446 0.737 1.205 99.94 99.91
5 95 95 5 95 4.923 3.379 10 2.978 2.17 97.8 97.9
5 95 95 1 95 4.828 3.081 13 2.899 2.176 98.2 97.8
5 99 99 5 99 4.99 3.904 25 2.453 1.81 99.5 99.4
5 99 99 1 99 5 3.706 31 2.466 1.83 99.4 99.4
5 99.9 99.9 1 99.9 4.991 4.202 101 1.739 1.481 99.94 99.93
14
From Table 4 and 5 we can observe that the populations Pµ1 and Pσ1 falling within
the specification limits LSL USL, in relation to the alternative hypothesis Haµ1 and
Haσ1 , are always greater than the correspondent population of the tolerance interval.
Furthermore these alternative hypothesis occurrences impliy an high rejection proba-
bility, the test power. Based on these considerations we can derive that the tolerance
interval coverage P and the confidence γ are no more crucial parameters for our aim,
but rather the minimum population proportion falling inside the specification limits
when the correspondent alterantive hypothesis can be detected with a specified power.
It’s worth noting that in this section the specification limit ASLtest , µ1 and σ1 are
output results of input data ASL, n, P, γ α and β.
ASLtest limit, that is always lower or equal the initial specification, can be used for
the verification test based on the following limits:
U SLtest = µ0 + ASLtest
LSLtest = µ0 − ASLtest
The R program ”Tolerance interval hypothesis test sigma.r”, starting from inputs
n, P, γ α an β, calculates µ1 , σ1 , ASLtest and k2 for test execution.
3. A new approach
Starting from the considerations of the previous sections, we can derive that we are no
more interested about the tolerance interval population coverage and confidence of the
tolerance interval itself, but we are looking for testing factors determination starting
from different constraints.
What we are looking for is the set of testing factors: ASLtest , k2 and the minimum
sample size n, satisfying the followings desired constraints:
• When the null hypothesis H0 is true the probability to reject it is equal to the
assigned value α (the Type I error).
• When the alternative hypothesis Haσ1 is true, we want the populations Pσ1 falling
within the specification limits LSL, USL not lower an assigned value Pσ1set , with
a probability not higher than the assigned value β (the Type II error), to accept
the null hypothesis with Pσ1 < Pσ1set .
• When the alternative hypothesis Haµ1 is true, we want the populations Pµ1 falling
within the specification limits LSL, USL not lower an assigned value Pµ1set , with
a probability not higher than the assigned value β (the Type II error), to accept
the null hypothesis with Pµ1 < Pµ1set .
Given Pσ1set and Pµ1set data input, we can obtain σ1set and µ1set as deviation
parameters, solving the following equations:
The testing factors ASLtest , k2 and n are calculated finding the solutioon of the
following equations system:
15
Z ASLtest
(1 − α) = T IH0 (z, n, µ0 , σ0 , k2)dz (21)
0
Z ASLtest
β= T IHaσ1 (z, n, µ0 , σ1set , k2)dz (22)
0
Z ASLtest
β= T IHaµ1 (z, n, µ1set , σ0 , k2)dz (23)
0
where the probability density functions T IH0 , T IHaσ1 and T IHaµ1 are calculated
according the former (6), (17) and (13). The above equantions system is solved by a
numerical reiterated algorithm with three nested loops. In the innermost loop ASLtest
is changed in order to fulfill the equation (21), in the middle loop the sample size n
is changed to fulfill the equation (22), while in the outermost loop the k2 factor is
changed to fulfill the equation (23). In the attached R program ”Testing factors with
sym.r” the solutions accuracy is controlled as desired setting an appropriate number
of iterations for each loop. It’s worth noting that for the middle loop (22), solving
the sample size n, the n solution accuracy implies an error , due to n rounded by
excess to the nearest integer, with 0 ≥ < 1. So, the correspondent deviation σ1 will
be generally lower than σ1 set, moreover the difference σ1set − σ1 and consequently
Pσ1 − Pσ1set , increases with decreasing of n.
Concerning instead the remaining solutions ASLtest and k2, the related errors can be
reduced as far as needed, properly increasing the specific loop iterations counter.
According this new approach, we refer in Table 6 a similar set of input parameter
as in Table 4, where we replaced the tolerance interval population P and confidence γ,
with populations Pµ1set and Pσ1set , both of them assumed equal to the correspondent
P value of Table 4.
Pµ1set and Pσ1set fall within the specification limits when the related alternative hy-
potheses Haµ1 and Haσ1 imply a power Pwr = 1 - β to reject the null hypothesis.
The Table 6 below shows the solution parameters solving the above equations sys-
tem (21), (22) and (23)
Table 6. Testing factors according the new approach
ASL Pµ1set % Pσ1set % α% Pwr% ASLtest k2 n µ1 σ1 Pµ1 % Pσ1 %
4 95 95 5 95 4.071 2.841 12 2.355 2.013 95.00 95.31
4 95 95 1 95 4.148 2.726 16 2.355 2.013 95.00 95.31
4 99 99 5 99 4.263 3.519 45 1.674 1.547 99.00 99.03
4 99 99 1 99 4.286 3.409 58 1.674 1.553 99.00 99.00
4 99.9 99.9 1 99.9 4.812 4.402 402 0.91 1.215 99.90 99.90
5 95 95 5 95 5.114 3.379 8 3.355 2.412 95.00 96.18
5 95 95 1 95 5.122 3.101 10 3.355 2.443 95.00 95.93
5 99 99 5 99 5.079 3.895 21 2.674 1.92 99.00 99.08
5 99 99 1 99 5.071 3.667 26 2.674 1.938 99.00 99.01
5 99.9 99.9 1 99.9 5.247 4.38 90 1.91 1.517 99.90 99.90
16
Now, examining Table 6, we can notice that ASLtest is no more constrained to
be lower than ASL, but is always greater, while the sample sizes n are lower than
the values seen in Table 4. On the other hand, now the popuations Pµ1 and Pσ1
falling inside the specification limits are very close to the correspondent expected
values Pµ1set and Pσ1set . In particular only Pσ1 shows a slight mismatch due to the
considerations described above about the sample size rounding.
As done in Table 5 of the previous section, in Table 7 below are shown the simulation
results with the testing factors of Table 6
Table 7. Simulation results for H0 , Haµ1 and Haσ1 for the new approach
ASL Pµ1set % Pσ1set % α% Pwr% αe % Pwrµ1 % Pwrσ1 %
4 95 95 5 95 5.01 95.0 95.0
4 95 95 1 95 0.95 95.2 95 1
4 99 99 5 99 4.99 99.1 98.9
4 99 99 1 99 1.03 99.0 99.0
4 99.9 99.9 1 99.9 0.98 99.9 99.9
5 95 95 5 95 4.98 95.0 95.0
5 95 95 1 95 1.02 95.0 95.0
5 99 99 5 99 4.94 99.0 99.1
5 99 99 1 99 1.01 99.0 99.0
5 99.9 99.9 1 99.9 1.04 99.9 99.9
In the next Table 8 a significant sample size reduction is obtained applying asym-
metric values for Pµ1set and Pσ1set , where in particular is Pσ1set slightly lower than
Pµ1set .
Table 8. Testing factors with asymmetric deviation sensitivity
ASL Pµ1set % Pσ1set % α% Pwr% ASLtest k2 n µ1 σ1 Pµ1 % Pσ1 %
4 95 90 5 95 3.35 2.117 8 2.355 2.398 95.00 90.47
4 95 90 1 95 3.507 2.086 11 2.355 2.344 95.00 91.21
4 99 95 5 99 2.706 1.953 19 1.674 2.012 99.00 95.32
4 99 95 1 99 2.746 1.867 24 1.674 2.028 99.00 95.14
4 99.9 99 1 99.9 2.068 1.654 92 0.908 1.549 99.90 99.02
5 95 90 5 95 4.516 2.779 6 3.355 2.828 95.00 92.29
5 95 90 1 95 4.444 2.422 7 3.355 2.952 95.00 90.97
5 99 95 5 99 3.773 2.576 11 2.674 2.55 99.00 95.01
5 99 95 1 99 3.874 2.463 14 2.674 2.515 99.00 95.32
5 99.9 99 1 99.9 3.476 2.598 38 1.91 1.935 99.90 99.02
Now, ASLtest is alawys lower than ASL. This asymmetric approach could be at-
tractive when a lower sensitivity to σ deviation is acceptable. In Table 9 below, the
simulation results for this case are shown.
17
Table 9. Simulation results for H0 , Haµ1 and Haσ1 in asymmetric case
ASL Pµ1set % Pσ1set % α% Pwr% αe % Pwrµ1 % Pwrσ1 %
4 95 90 5 95 5.08 95.0 95.0
4 95 90 1 95 1.06 94.9 95 1
4 99 95 5 99 4.96 99.1 99.0
4 99 95 1 99 1.06 99.0 99.0
4 99.9 99 1 99.9 1.01 99.9 99.9
5 95 90 5 95 4.96 95.0 94.9
5 95 90 1 95 1.00 95.0 95.1
5 99 95 5 99 5.09 99.0 99.0
5 99 95 1 99 1.01 99.0 99.0
5 99.9 99 1 99.9 0.97 99.9 99.9
4. Conclusions
When a verification program requires testing a bivariate null hypothesis referring the
mean µ and standard deviation σ of a normal process, a novel tolerance interval based
approach is proposed.
The sample size is determined such that the desired Type I = α and Type II = β
errors probabilities, however always taking into account the borderline populations
Pµ1 and Pσ1 falling within the specification limits, with the assigned statistical power
(1 - β) to reject the null hypothesis.
Finally the input data are:
• Null hypothesis H0 by (µ0 , σ0 ). Where µ0 and σ0 are preliminary estimations
obtained with a pilot study or theoretical considerations.
• The borderline population Pµ1 falling within the specification limits LSL, USL.
This is consequence of the deviation µ1 in case of alternative hypothesis
Haµ1 :(µ0 + µ1; σ0 ).
• The borderline population Pσ1 falling within the specification limits LSL, USL.
This is consequence of the deviation σ1 /σ0 in case of alternative hypothesis
Haσ1 :(µ0 ; σ1 ).
• Significance level α (Type I error). That is the probability to reject the null
hypothesis when in fact it is true.
• Power = 1 -β (1 - Type II error). That is the probability to reject the null
hypotheis when in fact one of the alternative hypothesis Haµ1 or Haσ1 is true.
The population covered by the specification limits will be always greater than Pµ1
and Pσ1 when the null hypothesis is true or when the alternative hypotheses will be
met with deviation lower than µ1 and σ1.
From a risk management perspective, we can summarize the problem in this way:
- There is a manufacturer risk that is provided by the combination of significance level
α probability and the verification cost.
- There is a design (or customer) risk, based on the combination of β probability and
the potential harms due to the alternative hypothesis deviations µ1 and σ1. The risk
management of the two issues above mentioned can be an essential element for the
sample size selection. Finally, the claim of this paper is: a tolerance interval with a
confidence γ to cover the population P inside its random limits LTL UTL, unlikely
leads to an optimal criterion to verify if a minimum expected population is falling
inside the assigned specification limits LSL USL.
18
The proposed approach instead provides the following advantages:
• sets a minimum population that falls within the specifications, implying the test
failing with the pre established power if this requirement is not met
• provides the parameters to gain full control about costs (n) and risks (n ∗ α;
µ1 ∗ β; σ1 ∗ β).
• provides accurate results when specific level of safety or reliability are required.
References
19
5. Appendices
20