Unit-3 Biostatistics Descriptive
Unit-3 Biostatistics Descriptive
Objectives
Type-I & Type-II Error
P value and its applications
1. Define type-I & II errors
2. Explain power of the test
3. Define P-value and real its application in research
4. Interpret different p-values
Example:
Medical Test Scenario
Let’s consider a medical test scenario where you are testing whether a new diagnostic tool is better
than a placebo.
Null Hypothesis (H0): The diagnostic tool has no effect (i.e., it performs no better than the placebo).
Alternative Hypothesis (HA): The diagnostic tool performs better than the placebo.
If you conduct the test and reject H0 (based on the data), this means you believe the diagnostic tool is
effective. However, a Type-I error would occur if H0 is actually true (the tool does no better than the
placebo), but you incorrectly reject H0 and conclude that the tool is effective.
False positive: The test indicates that the tool is effective when, in fact, it isn't.
Consequences: This might lead to unnecessary adoption of the tool, patients being treated based on
false results, or even resources being wasted on further studies and treatments that are not beneficial.
The true state of nature is that the null hypothesis is correct (i.e., the drug has no effect).
But your data suggests otherwise, leading you to reject H0 and conclude that the drug has an effect
(even though it doesn’t).
P(Type-I error)=α
Unit 8: Type I and Type II errors and Power of the test
Example: If α = 0.05, you are allowing a 5% risk of incorrectly rejecting the null hypothesis when it is
true.
Lower α (e.g., 0.01): More stringent, fewer Type-I errors, but increases the risk of a Type-II error (failing
to detect a true effect).
Higher α (e.g., 0.10): Less stringent, higher risk of Type-I error, but might reduce Type-II errors
(increasing power).
Use a larger sample size: Larger samples generally provide more reliable data and reduce the variability
of test statistics, which in turn can help avoid making a Type-I error.
Replicate the study: Repeating experiments and ensuring that results are consistent across different
samples and studies can help confirm whether the effect (if any) is real or just due to random chance.
Adjust for multiple comparisons: If you perform multiple tests, the chances of making a Type-I error
across all tests increase. Researchers use correction methods (like the Bonferroni correction) to adjust
the significance level when performing multiple tests.
Key Characteristics:
Null hypothesis (H0) is false, but we fail to reject it.
In practical terms, this means accepting the null hypothesis when it should have been rejected.
The probability of committing a Type-II error is denoted by beta (β).
Example:
Consider a clinical trial where you are testing a new drug:
Null hypothesis (H0): The new drug has no effect (it is the same as the placebo).
Alternative hypothesis (HA): The new drug has an effect (it is different from the placebo).
Unit 8: Type I and Type II errors and Power of the test
If the new drug does have an effect, but your test fails to detect it and you fail to reject the null
hypothesis (incorrectly concluding that the drug has no effect), this would be a Type-II error.
Impact:
Type-II errors are significant because they can lead to missed opportunities to detect real effects,
which can result in incorrect conclusions and poor decision-making.
Power of a Test
The power of a statistical test is a critical concept in hypothesis testing. It is defined as the probability
of correctly rejecting the null hypothesis when it is false (i.e., detecting a true effect when it exists). In
other words, power reflects the ability of a test to avoid Type-II errors (false negatives).
Formula for Power
The power of a test is defined as:
Power = 1− β
Where:
β is the probability of a Type-II error (failing to reject the null hypothesis when it is false).
Power represents the probability of rejecting the null hypothesis correctly when it is false, i.e., correctly
identifying a true effect.
Thus, the power of the test is 1 minus the probability of making a Type-II error.
Interpretation of Power
Power = 0.80 (80%) means there is an 80% chance of correctly rejecting the null hypothesis when it is
false (detecting a true effect).
Power = 0.90 (90%) means there is a 90% chance of correctly rejecting the null hypothesis when it is
false.
Sample Size (n): Larger sample sizes generally increase the power of a test. A larger sample provides
more reliable data, making it easier to detect a true effect if it exists.
Effect Size: The larger the true effect (i.e., the larger the difference between the null hypothesis and the
alternative hypothesis), the greater the power. A larger effect is easier to detect.
Significance Level (α): The power of a test is also related to the chosen significance level (α). By
increasing α (e.g., from 0.01 to 0.05), you increase the power of the test (since it becomes easier to
reject H0), but you also increase the risk of a Type-I error (false positive).
Variability of the Data: Less variability (or noise) in the data leads to higher power. High variability
makes it harder to detect a true effect.
Power = 0.90: Some fields, particularly those involving high-stakes decisions (like medical research), aim
for even higher power, such as 90%. This means there is a 90% chance of detecting an effect if one truly
exists.
Power Analysis
Before conducting a study, researchers often perform a power analysis to determine the sample size
needed to achieve a desired power level (typically 0.80 or 0.90). This helps ensure that the study has
enough power to detect a meaningful effect, if one exists.
Unit 8: Type I and Type II errors and Power of the test
P value
The area that falls in the tail beyond the value of the test statistic. A p value tells you the chance of
getting a statistics as extreme or more extreme that the one calculated for the sample.
-Z cal
Unit 8: Type I and Type II errors and Power of the test
Z cal
For right tail test
Z cal
-Z cal
Null hypothesis (H0): The new drug has no effect on patients' recovery time.
Alternative hypothesis (HA): The new drug reduces recovery time.
Test Decision:
Significance level (α): 0.05
Understanding p value
The p value can also be thought of as the probability of obtaining a result as extreme as or more
extreme than the actual sample value obtained given that the null hypothesis is true.
Example 1
A topic of recent clinical interest is the possibility of using drugs to reduce infarct size in those patients
who have had a MI within the past 24 hrs. Suppose we know that in untreated patients the mean
infarct size is 25 with a standard deviation of 10. Furthermore, in 50 patients treated with the drug, the
mean infarct size is 16. Is the drug effective in reducing infarct size?
Unit 8: Type I and Type II errors and Power of the test
Solution:
Given Data:
Mean infarct size in untreated patients (μ0) = 25 (population mean)
Standard deviation (σ) = 10
Sample size (n) = 50 (treated patients)
Sample mean (Xˉ) = 16 (treated patients)
We need to determine if the drug is effective in reducing infarct size, which means testing if the mean
infarct size for treated patients is less than 25.
Where:
Xˉ=16 (sample mean)
μ0=25 (population mean)
σ=10 (standard deviation)
n=50 (sample size)
Substitute the given values into the formula:
Z = 16−26/ 10 /√ 50
Step 5: Conclusion
Calculated Z-value = -6.36
Critical Z-value = -1.645
Since -6.36 < -1.645, we reject the null hypothesis.
There is enough evidence at the 5% significance level to conclude that the drug is effective in reducing
infarct size. The mean infarct size in treated patients is significantly smaller than the mean infarct size in
untreated patients, suggesting that the drug has a statistically significant effect in reducing infarct size.
Unit 8: Type I and Type II errors and Power of the test
So;
Null Hypothesis: H0:μ≥25
Alternative Hypothesis: HA:μ<25
Significance level: α=0.05
Z-value: -6.36
Critical Z-value: -1.645
Conclusion: Reject H0, indicating the drug is effective in reducing infarct size.
The p-value plays a crucial role in various applications across different fields such as medicine, social
sciences, economics, business, and engineering.
In this case, researchers want to determine if the new drug has a statistically significant effect on
patient recovery time. If the p-value is low (e.g., p-value=0.02), it suggests the observed effect is
unlikely to have occurred by chance, and they would reject H0, concluding the drug is effective.
Marketers may use hypothesis testing to evaluate if a new advertising campaign has had an impact
on product sales. If the p-value is below 0.05, it suggests that the observed increase in sales is
statistically significant, and the company may decide to continue with or scale up the campaign.
Engineers might test a new process or change to improve manufacturing efficiency or quality. If the p-
value from the testing is low (e.g., p-value=0.01, this suggests the process change led to significant
improvements, prompting adoption.
In environmental studies, researchers test the effects of new regulations, conservation strategies, or
pollution-control measures. A small p-value (e.g., p-value=0.03 might suggest that the regulation
significantly reduces pollution levels, supporting its implementation.
Context Implications
A p-value helps determine if a treatment is statistically effective, but must
Clinical Trials
be interpreted alongside clinical significance and practical considerations.
P-values help assess associations between exposures and outcomes, but
Epidemiology
must account for confounding factors and potential biases.
Multiple The p-value can increase the risk of Type I errors when multiple tests are
Comparisons conducted, requiring adjustments (e.g., Bonferroni correction).
P-values alone do not measure the size or precision of an effect; they must
Effect Size and CI be considered with effect size and confidence intervals for a complete
interpretation.
Type I and Type II The p-value is directly related to the probability of Type I errors, and must
Errors be balanced with Type II error (power of the test).
Limitations of P-Value
While p-values are powerful tools in statistical testing, they do have limitations and can sometimes
lead to misinterpretations:
Unit 8: Type I and Type II errors and Power of the test
Multiple Comparisons:
When multiple hypotheses are tested, the chance of finding a significant result by chance increases.
This is known as the multiple testing problem.