Stats CHP 1 Notes
Stats CHP 1 Notes
The purpose of inferential statistics is to draw a conclusion (an inference) about conditions that exist in a
population (the complete set of observations) by studying a sample (a subset) drawn from the population.
Inferential statistics is a branch of statistics used to make predictions or generalizations about a
population based on sample data. Unlike descriptive statistics, which simply summarize data, inferential
statistics allow you to make informed decisions or conclusions with a degree of certainty.
IMPORTANT TERMINOLOGIES
1. Assumption of Normality
The assumption of normality is fundamental in parametric statistics and dictates that the data being
analyzed should follow a normal distribution. A normal distribution, often referred to as a bell curve, is
symmetrical about the mean, where most observations cluster around the central peak and frequencies
taper off equally on both sides. This distribution is crucial because many statistical tests, like the t-test and
ANOVA, rely on this pattern to make accurate inferences. When data significantly deviates from normality,
statistical results may become misleading, potentially leading to false conclusions.
Outliers are data points that deviate significantly from the rest of the dataset and can skew the results of
parametric tests, making them unreliable. These extreme values can inflate or deflate measures of central
tendency and dispersion, such as the mean and standard deviation, leading to inaccurate statistical
interpretations. Detecting outliers is typically done using box plots, scatter plots, or calculating Z-scores to
identify values that lie beyond three standard deviations from the mean. Once identified, it’s essential to
assess whether these outliers result from errors, rare events, or valid but unusual observations.
Depending on the context, outliers may be transformed, removed, or analyzed separately to ensure they
do not disproportionately affect the analysis.
Why These Assumptions Matter
• Parametric tests are generally more powerful than non-parametric tests because they make
specific assumptions about the data.
• Violating these assumptions can increase the chances of Type I or Type II errors.
• If the assumptions are not met, consider using non-parametric tests (like the Mann-Whitney U test
or the Kruskal-Wallis test), which make fewer assumptions about the data distribution.
NON-PARAMETRIC TESTS do not rely on strict assumptions about the data distribution. They are
suitable for ordinal or nominal data and are robust when data is non-normally distributed or when sample
sizes are small. Examples include the Chi-square test, Mann-Whitney U test, and Kruskal-Wallis test.
Although less powerful than parametric tests, they provide a reliable alternative when parametric
• α=0.05\alpha = 0.05α=0.05 (5% significance level): The most commonly used level.
• α=0.01\alpha = 0.01α=0.01 (1% significance level): Used when stronger evidence is required.
• α=0.10\alpha = 0.10α=0.10 (10% significance level): Used in exploratory research or when less
precision is acceptable.
Region of Rejection
The region of rejection is the portion of the sampling distribution that contains values of the test
statistic that are unlikely to occur by chance if the null hypothesis is true.
• If the calculated test statistic falls into this region, the null hypothesis (H0H_0H0) is rejected.
• If it does not fall into the rejection region, we fail to reject the null hypothesis.
When deciding whether to retain or reject the null hypothesis, researchers follow a systematic approach
based on the level of significance and the evidence gathered from the data. The null hypothesis (H₀) is a
statement that assumes no effect, no difference, or no relationship between variables. On the other hand,
the alternative hypothesis (Hₐ) is the statement that contradicts the null hypothesis, indicating that there is
an effect or difference.
The decision to retain or reject the null hypothesis is based on a criterion known as the level of
significance. This level of significance represents the probability of making a Type I error, which means
rejecting the null hypothesis when it is actually true. The most commonly used level of significance is 0.05
(5%), although stricter levels like 0.01 (1%) can also be used in cases where more precision is required.
Researchers calculate a test statistic from the sample data and compare it to a critical value determined
by the chosen level of significance. If the test statistic falls within the region of rejection (the area of the
sampling distribution where outcomes are considered unlikely under the null hypothesis), the null
hypothesis is rejected. This indicates that the observed difference or effect is statistically significant and
unlikely to have occurred by chance.
However, if the test statistic does not fall within the region of rejection, the null hypothesis is retained,
meaning there is not enough evidence to conclude that a difference or effect exists. It is important to
understand that retaining the null hypothesis does not prove it true, but simply indicates that the data
does not provide strong enough evidence to reject it.
Why Use 0.05 as the Level of Significance?
The 0.05 level of significance is widely used because it provides a balance between risk and
reliability. It implies that there is only a 5% chance of rejecting a true null hypothesis. In some cases
where higher accuracy is needed, 0.01 may be chosen to minimize errors.
where σ is the standard deviation of the population, and n is the sample size. As sample
size increases, the SE decreases, making the sample mean a more accurate estimate of the population
mean.
4.Less Variability than Population
The sampling distribution is less spread out than the population distribution because the SE is smaller
than the population standard deviation. This is because averages tend to be more stable than individual
scores.
5.Symmetry
If the population distribution is normal or the sample size is large, the sampling distribution of means will
also be symmetrical, centered around the population mean.
Assumptions Associated with Inference about the Difference Between Two Independent Means
1. Random Sampling:
Each sample must be drawn randomly from its respective population. This ensures that every individual
has an equal chance of being selected, which leads to more reliable and unbiased results that are
generalizable to the population.
2. Independence of Samples:
The two samples must be independently selected, meaning the selection of one does not influence the
other. Independence ensures that the results from one sample do not affect or bias the results of the other
sample.
3. Sampling with Replacement:
Samples should be drawn with replacement, meaning that after selecting an individual, it is put back into
the population before selecting the next one. This keeps the probability of selection the same for each
observation, maintaining sample representativeness.
4. Normality of Sampling Distribution:
The sampling distribution of the difference between the two means should follow a normal distribution.
This assumption is important for applying normal-based statistical tests, which require the data to
approximate normality to make valid inferences.
5. Use of the t-Statistic:
When population standard deviations are unknown, the t-distribution is used instead of the normal
distribution. This allows us to estimate the population standard deviations from the sample data, adjusting
for this uncertainty when calculating test statistics.
6. Homogeneity of Variance (Equal Variance Assumption):
The assumption is that the variances of both populations are roughly equal. If this assumption is violated,
the results may not be reliable, but the effect is smaller with large sample sizes or when the sample sizes
are equal between the groups.
7. Central Limit Theorem (CLT):
The CLT states that as sample size increases, the sampling distribution of the mean approaches a normal
distribution, even if the population is not normally distributed. For sample sizes over 25, moderate
skewness can usually be tolerated without a significant effect on the results.
8. When to Use Nonparametric Tests:
If the parent population is highly non-normal or if the sample size is small, nonparametric tests are
preferred. These tests do not rely on normality assumptions, making them suitable for data that doesn't
meet the conditions for parametric tests.
ONE-TAILED AND TWO-TAILED HYPOTHESES
When conducting hypothesis testing, researchers choose between two types of tests: one-tailed tests
and two-tailed tests. The choice between these tests depends on the nature of the research question
and the expected direction of the effect.
ONE-TAILED HYPOTHESIS (DIRECTIONAL TEST)
A one-tailed hypothesis is used when the researcher expects the difference or effect to occur in a specific
direction. This means that the alternative hypothesis (Hₐ) states that the population parameter is either
greater than or less than the hypothesized value, but not both. In other words, the region of rejection (the
area where the null hypothesis is rejected) is entirely located on one side (tail) of the sampling
distribution.
One-tailed tests are typically used when there is a clear and justified expectation about the direction of
the outcome. For example:
• A psychologist studying aggression might hypothesize that children show more aggressive behavior
after watching violent TV shows. Here, the hypothesis is one-directional because it expects an
increase in aggression.
• A fitness test for schoolchildren might hypothesize that their physical fitness levels are lower than the
national standard. Again, the expectation is one-directional (lower).
The advantage of a one-tailed test is that it is more statistically powerful for detecting an effect in the
specified direction. However, its major disadvantage is that it does not account for the possibility of an
effect occurring in the opposite direction. If the effect turns out to be opposite to what was predicted, a
one-tailed test fails to detect it, even if the difference is significant.
TWO-TAILED HYPOTHESIS (NONDIRECTIONAL TEST)
A two-tailed hypothesis is used when the researcher does not predict the direction of the effect or
difference, only that there is an effect. The alternative hypothesis (Hₐ) states that the parameter may be
either greater than or less than the hypothesized value. In this case, the region of rejection is divided
equally between both tails of the sampling distribution.
Two-tailed tests are appropriate when it is important to detect any difference, regardless of direction. For
example:
• A study comparing student performance between two teaching methods might not predict
whether one method is better or worse than the other but simply tests for a difference.
• An experiment testing a new drug might not specify whether it will improve or worsen a patient’s
condition, only that it will have an impact compared to the standard treatment.
The advantage of a two-tailed test is that it is more conservative and less likely to lead to false positives
because it accounts for effects in both directions. However, it is generally less statistically powerful than a
one-tailed test since the level of significance (alpha) is split between the two tails.
Choosing Between One-Tailed and Two-Tailed Tests
The choice between a one-tailed and two-tailed test should be made before data collection and should be
based on the logic and objective of the study. It is not appropriate to decide on the type of test after
analyzing the data, as this could introduce bias and inflate error rates.
• Use a one-tailed test if you have a specific directional prediction and are confident that an effect will
occur only in that direction.
• Use a two-tailed test if you want to detect any difference, regardless of whether it is positive or
negative.
• Making a premature or unjustified choice can lead to inaccurate conclusions, so it’s crucial to
understand the research context and hypothesis before deciding.
ASSUMPTIONS OF Z TEST
A Z-test is a statistical test used to determine whether there is a significant difference between a sample
mean and a population mean when the population variance is known, or to compare the means of two
large samples. It helps assess whether the observed data is likely to have occurred by chance under the
assumption that the null hypothesis is true.
1. Random Sampling:
The first assumption of the Z test is that the sample drawn from the population is random. This means
that every individual in the population has an equal chance of being selected. Random sampling helps
ensure that the sample is representative of the population, reducing bias and increasing the reliability of
the results. If the sample is not random, it can lead to skewed results that do not accurately reflect the
population. However, achieving truly random samples can be challenging in practice.
2. Sampling with Replacement:
The second assumption is that sampling is done with replacement. This means that after selecting an
individual, they are placed back into the population, so they can be chosen again. Sampling with
replacement helps maintain the same probability of selection throughout the process. Although sampling
without replacement is more common in practice, the error introduced is minimal if the sample size is
small compared to the population size.
3. Normal Distribution of Sampling Means:
The third assumption is that the sampling distribution of the sample mean follows a normal distribution.
This assumption is crucial because the Z test relies on the properties of the normal distribution to
calculate probabilities and critical values. If the population data is approximately normal, the sampling
distribution will also be normal, especially when the sample size is large (usually 25 or more), thanks to
the Central Limit Theorem. This theorem states that the distribution of sample means will be
approximately normal, regardless of the population’s distribution, when the sample size is sufficiently
large.
4. Known Population Standard Deviation (σ):
The fourth assumption is that the population standard deviation (σ) is known. In most real-world
scenarios, however, this information is rarely available. Knowing the population standard deviation means
that we have precise information about the variability within the population. If σ is unknown, the t-test is
usually preferred over the Z test. The formula for the Z test includes σ, which is used to calculate the
standard error of the mean.
Conclusion:
Meeting these assumptions is essential to obtain valid and reliable results from a Z test. If any of these
assumptions are violated, the conclusions drawn from the test may be inaccurate or misleading.
Therefore, it is important to carefully assess whether the data and sampling methods meet these criteria
before applying the Z test.
STUDENT’S DITRIBUTION OF t
The student’s t-distribution is a probability distribution used in statistics when the sample size is small,
and the population standard deviation is unknown. It is similar to the normal distribution but has heavier
tails, making it useful for estimating population parameters and conducting hypothesis tests when data is
limited.
William S. Gosset (1876–1937) was a British mathematician who developed the t-test. It was originally
called Student’s distribution of t because Gosset published under the name ‘Student’ while working for the
Guinness Brewing Company, which did not allow their employees to publish papers.
Characteristics of Student’s t-Distribution
1. Family of Distributions:
The Student’s t-distribution is not a single distribution but a family of distributions. Each distribution within
this family is defined by its degrees of freedom (df), which is closely related to the sample size. The shape
of the t-distribution changes with different degrees of freedom, becoming more similar to the normal
distribution as the sample size increases.
2. Relation to Sample Size:
o When the sample size is large, the sample standard deviation closely approximates the population
standard deviation, making the t-distribution almost identical to the normal (z) distribution.
o When the sample size is small, standard deviation may vary significantly from population standard
deviation, causing the t-distribution to deviate from the normal distribution.
3. Symmetry and Shape:
o The Student’s t-distribution is symmetrical and unimodal, meaning it has one peak and is mirrored
on either side of the mean.
o The mean of the distribution is zero, similar to the normal distribution.
4. Platykurtic Nature:
The t-distribution is platykurtic, meaning it has a flatter peak and thicker tails compared to the normal
distribution. This characteristic indicates a higher probability of obtaining extreme values, making it more
suitable for small sample sizes or situations where the population standard deviation is unknown.
5. Larger Standard Deviation:
The standard deviation of the t-distribution is larger than that of the normal distribution, reflecting the
greater variability expected when the sample size is small.
6. Effect of Degrees of Freedom:
o The shape and spread of the t-distribution are influenced by the degrees of freedom.
o As the degrees of freedom increase, the t-distribution becomes more similar to the normal
distribution.
o When the degrees of freedom are infinite, the t-distribution and normal distribution are identical.
o As the degrees of freedom decrease, the tails become thicker and the peak becomes lower,
emphasizing greater variability.
7. Critical Values and Significance Levels:
Due to the thicker tails, critical values for hypothesis testing are larger for the t-distribution compared to
the z-distribution. This means that to reject the null hypothesis with the same level of significance (e.g.,
α=0.05), a t-value must be more extreme than a corresponding z-value.
o For example, with infinite degrees of freedom, the critical t-value equals the critical z-value of
±1.96.
o For smaller degrees of freedom, the critical t-value is significantly larger, indicating a stricter
criterion for rejecting the null hypothesis.
In summary, the Student’s t-distribution is designed to handle situations where the sample size is small
and the population standard deviation is unknown. It accounts for additional uncertainty by having thicker
tails and greater variability, making it a robust choice for estimating population parameters from limited
data.
ASSUMPTIONS OF THE T-TEST:
1. Random Sampling:
The data should be collected using a random sampling method to ensure that every individual in the
population has an equal chance of being selected. This helps minimize sampling bias and makes the
results more generalizable. Without random sampling, the results may not accurately represent the
population.
2. Normality:
The population from which the sample is drawn should be approximately normally distributed. This
assumption is crucial when the sample size is small (usually n < 30). If the data are not normally
distributed, the test results may be inaccurate, but the Central Limit Theorem helps when the sample
size is large.
3. Equal Variance (for Independent t-Test):
When performing an independent t-test, the variances of the two populations being compared should
be equal, a condition known as homogeneity of variance. This is often tested using Levene’s Test.
If variances are unequal, a modified version of the t-test (Welch’s t-test) is used.
4. Independence:
The observations in the sample must be independent of each other, meaning that the value of one
observation should not influence another. This assumption is crucial for ensuring the accuracy of the
test statistics. Violating this assumption can lead to misleading conclusions.
5. Unknown Population Standard Deviation:
The standard deviation of the population is generally unknown, and instead, the sample standard
deviation is used as an estimate. This makes the t-test more applicable in real-world situations, as
population parameters are rarely known.
When to Use Each Test:
• Z-Test: When the sample size is large (n>30) and the population variance is known.
• T-Test: When the sample size is small (n≤30n) and the population variance is unknown.
DEGREES OF FREEDOM
1. Definition:
Degrees of freedom (df) in psychological statistics refer to the number of independent values that can
vary in an analysis while estimating a statistical parameter. It represents the flexibility or freedom to
vary when calculating statistical measures.
2. Significance:
Degrees of freedom are crucial because they influence the shape of statistical distributions, like the t-
distribution. They help determine the accuracy and reliability of hypothesis testing, making it possible
to make valid inferences from sample data.
df=n−1 where n is the sample size.
3. Why Important:
o Ensures accurate hypothesis testing by adjusting for sample size and variability.
o Helps choose the correct distribution (like t or F distribution) when making inferences.
o Reduces bias by accounting for the number of estimated parameters.
4. Impact on Distribution:
The fewer the degrees of freedom, the more the t-distribution deviates from the normal curve, making
it flatter with thicker tails. As the degrees of freedom increase, the t-distribution becomes more similar
to the normal distribution.
LEVELS OF SIGNIFICANCE VS. P-VALUES
1. Level of Significance (Alpha, 𝛼):
o The level of significance (usually denoted as 𝛼) is the probability of rejecting the null hypothesis
(H₀) when it is actually true.
o It represents the risk of making a Type I error, meaning falsely concluding that there is an effect
when none exists.
o Commonly used values of 𝛼 are 0.05, 0.01, and 0.001, where a lower value indicates stricter
criteria for rejecting H₀.
o Researchers decide the significance level before conducting the test to control the risk of error.
2. p-Value:
o The p-value measures the probability of obtaining a sample result as extreme as, or more
extreme than, the one observed, assuming that H₀ is true.
o It indicates the rarity of the observed outcome under the null hypothesis.
o A smaller p-value suggests stronger evidence against H₀, while a larger p-value indicates weak
evidence.
o It is not pre-set like the significance level but is calculated after data analysis.
3. Comparison of Level of Significance and p-Value:
o If the p-value ≤ 𝛼, the result is considered statistically significant, and H₀ is rejected.
o If the p-value > 𝛼, the result is not statistically significant, and H₀ is not rejected.
o For example, if 𝛼 = 0.05 and the p-value is 0.012, the p-value is less than 0.05, so H₀ is
rejected.
4. Reporting p-Values:
o Researchers often do not report exact p-values but rather indicate whether they are below
critical levels like 0.05, 0.01, or 0.001.
o A statement like "p < 0.05" or "p < 0.01" is commonly used to indicate statistical significance.
In summary, the level of significance is a pre-determined threshold for decision-making, while the p-value
is a post-analysis measure of how likely the observed data would occur if the null hypothesis were true.