0% found this document useful (0 votes)
16 views87 pages

Al2 Report

The document discusses the concepts of validity and reliability in test construction, detailing various types of validity (content, criterion-related, and construct) and methods to establish reliability (test-retest, equivalent forms, split-half, and Kuder-Richardson formulas). It emphasizes the importance of validity coefficients and reliability coefficients in assessing test quality, alongside factors affecting both validity and reliability. The document also includes examples and calculations to illustrate how to determine the validity and reliability of tests.

Uploaded by

danielkon25
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PPTX, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
16 views87 pages

Al2 Report

The document discusses the concepts of validity and reliability in test construction, detailing various types of validity (content, criterion-related, and construct) and methods to establish reliability (test-retest, equivalent forms, split-half, and Kuder-Richardson formulas). It emphasizes the importance of validity coefficients and reliability coefficients in assessing test quality, alongside factors affecting both validity and reliability. The document also includes examples and calculations to illustrate how to determine the validity and reliability of tests.

Uploaded by

danielkon25
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PPTX, PDF, TXT or read online on Scribd
You are on page 1/ 87

ESTABLISHING

VALIDITY AND
RELIABILTY OF
TESTS
Table of contents
 Validity of a Test
 Types of Validity
 Important Things to Remember
about Validity
 Factors Affecting the Validity of a
Test Item
 Reasons that Reduce the Validity of
the Test Item
 Guide Questions to Improve Validity
 Validity Coefficient
 Reliability of a Test
 Factors Affecting Reliability of a Test
 Four Methods of Establishing of a
Test
 Reliability Coefficient
 Description of Reliability Coefficient
 Interpreting Reliability Coefficient
Learning Outcomes
At the end of this chapter, the students should
be able to:
1. define the following terms: validity,
reliability, content validity, construct validity,
criterion-related validity, predictive validity,
concurrent validity, test-retest method,
equivalent/parallel method, split-half method,
Kuder-Richardson formula, validity coefficient,
reliability coefficient;
2. discuss the different approaches of validity;
3. present and discuss the different methods
of solving the reliability of a tests;
4. identify the different factors affecting the
validity of the test;
5. identify the factors affecting the reliability
of the test;
6. compute the validity coefficient and
reliability coefficient; and
7. interpret the reliability coefficient and
validity coefficient of the test.
INTRODUCTION
Test constructors believed that every
assessment tool should possess good qualities.
Most literatures consider the most common
technical concepts in assessment are the
validity and reliability. For any type of
assessment whether traditional or authentic it
should be carefully developed so that it may
serve whatever purpose it may have. In this
chapter, we shall discuss the different ways of
establishing validity and establishing reliability.
VALIDITY OF A
TEST
Validity (Airasian, 2000) is concerned
whether the information obtained from
an assessment permits the teacher to
make a correct decision about a
student's learning.
TYPES OF
VALIDITY
1. CONTENT
VALIDITY
A type of validation that
refers to the relationship
between a test and the
instructional objectives,
establishes content so that
the test measures what it is
supposed to measure.
Things to remember about
validity:
a. The evidence of the content validity of
a test is found in the Table of
Specification.
b. This is the most important type of
validity for a classroom teacher.
C. There is no coefficient for content
validity. It is determined by experts
judgmentally, not empirically.
2. CRITERION-RELATED
VALIDITY
A type of validation that refers to the
extent to which scores from a test
relate to theoretically similar measures.
It is a measure of how accurately a
student's current test score can be
used to estimate a score on a criterion
measure.
a. Concurrent validity. The criterion
and the predictor data are collected at
the same time. This type of validity is
appropriate for tests designed to assess
a student's current criterion status or
when you want to diagnose student's
status; it is a good diagnostic screening
test.
b. Predictive validity. A type of
validation that refers to a measure of
the extent to which a student's current
test result can be used to estimate
accurately the outcome of the student's
performance at later time. It is
appropriate for tests designed to assess
student's future status on a criterion
Predictive validity is very important in psychological testing, like if the
psychologists want to predict responses, behaviors, outcomes,
performances and others.
3. CONSTRUCT VALIDITY
A type of validation that refers
to the measure of the extent to
which a test measures a
theoretical and unobservable
variable qualities such as
intelligence, math
achievement, performance.
TEST OR MEASUREMENT
INSTRUMENT

a. Convergent validity is a type of


construct validation wherein a test has a
high correlation with another test that
measures the same construct.
b. Divergent validity is a type of
construct validation wherein a test has
low correlation with a test that
measures a different construct. In this
case, a high validity occurs only when
there is a low correlation coef-ficient
between the tests that measure
different traits.
C. Factor analysis is another method
of assessing the construct validity of a
test using complex statistical
procedures conducted with different
procedures.

There are other ways of assessing construct validity


like test’s internal consistency, development change
and experimental intervention.
IMPORTANT
THINGS TO
REMEMBER
ABOUT
VALIDITY
1. Validity refers to the decisions we make, and
not to the test itself or to the measurement.
2. Like reliability, validity is not an all-or-nothing
concept; it is never totally absent or absolutely
perfect.
3. A validity estimate, called a validity
coefficient, refers to specific type of validity. It
ranges between 0 and 1.
4. Validity can never be finally determined; it is
specific to each administration of the test.
FACTORS
AFFECTING THE
VALIDITY OF A
TEST ITEM
1.The test itself.
2. The administration and scoring of a
test.
3. Personal factors influencing how
students response to the test.
4. Validity is always specific to a
particular group.
REASONS THAT
REDUCE VALIDITY
OF A TEST ITEM
1. Poorly constructed test items
2. Unclear directions
3. Ambiguous test items
4. Too difficult vocabulary
5. Complicated syntax
6. Inadequate time limit
7. Inappropriate level of difficulty
8. Unlimited clues
9. Improper arrangement of test items
GUIDE
QUESTIONS TO
IMPROVE
VALIDITY
1. What is the purpose of the test?
2. How well do the instructional objectives
selected for the test represent the instructional
goals?
3.Which test item format will best measure the
achievement of each objective?
4. How many test items will be required to
measure the performance adequately to each
objective?
VALIDITY
COEFFICIENT
The validity coefficient is the computed
value of the rxy. In theory, the validity
coefficient has values like the correlation
that ranges from 0 to 1. In practice, most of
the validity scores are usually small and
they range from 0.3 to 0.5, few exceed 0.6
to 0.7. Hence, there is a lot of improvement
in most of our psychological measurement.
Another way of interpreting the findings
is to consider the squared correlation
coefficient
2 (rxy) ,this is called coefficient
of determination. Coefficient of
determination indicates how much
variation in the criterion can be
accounted for by the predictor (teacher
test).
Example, if the computed value of ray=
0.75. The coefficient of determination is
= 0.5625 or 56.25% of the variance in
the student performance can be
attributed to the test or 43.75% of the
student performance cannot be
attributed to the test results.
Example:
Teacher Benjamin James develops a
45-item test and he wants to
determine if his test is valid. He takes
another test that is already
acknowledged for its validity and
uses it as criterion. He conducted
these two sets of test to his 15
students. The following table shows
the results of the two tests. Is the
test developed by Mr. Benjamin
James valid? Find the Validity
coefficient using Pearson r and the
coefficient of determination.
INTERPRETATION:
The correlation coefficient is 0.94,which means that the
validity of the test is high, or 88.36% of the variance in the
students’ performance can be attributed to the test.
RELIABILITY OF A
TEST
Reliability refers to the consistency with
which it yields the same rank for individuals
who take the test more than once (Kubiszyn
and Borich, 2007).

The reliability of a test can be determined


by means of Pearson product correlation
coefficient, Spearman-Brown formula and
Kuder-Richardson formulas.
FACTORS AFFECTING
RELIABILITY OF A TEST
1. Length of the test
2. Moderate item difficulty
3. Objectives scoring
4. Heterogeneity of the student
group
5. Limited time
FOUR METHODS
OF ESTABLISHING
RELIABILITY OF A
TEST
1. TEST-RETEST METHOD
A type of reliability determined by administering the
same test twice to the same group of students with any
time interval between the tests. The results of the test
scores are correlated using the Pearson product
correlation coefficient (r) and this correlation coefficient
provides a measure of stability. This indicates how
stable the test result over a period of time. The formula
is
2. EQUIVALENT FORM
A type of reliability determined by administering two
different but equivalent forms of the test (also called
parallel or alternate forms)to the same group of students
in close succession. The equivalent forms are
constructed to the same set of specifications that is
similar in content, type of items and difficulty. The
results of the test scores are correlated using the
Pearson product correlation coefficient (r) and this
correlation coefficient provides a measure of the degree
to which generalization about the performance of
students from one assessment to another assessment is
justified. It measures the equivalence of the tests.
3. SPLIT-HALF
METHOD
Administer test once and score two equivalent halves of
the test. To split the test into halves that are equivalent,
the usual procedure is to score the even-numbered and
the odd-numbered test item separately. This provides
two scores for each student. The results of the test
scores are correlated using the Spearman-Brown formula
and this correlation coefficient provides a measure of
internal consistency. It indicates the degree to which
consistent results are obtained from two halves of the
test. The formula is
4. KUDER-RICHARDSON
FORMULA
Administer the test once. Score the total test and apply the Kuder-
Richardson formula. The Kuder-Richardson 20 formula is applicable
only in situations where students' responses are scored
dichotomously, and therefore, is most useful with traditional test
items that are scored as right or wrong, true or false, and yes or no
type. KR-20 formula estimates of reliability provide information
whether the degree to which the items in the test measure is of the
same characteristic, it is an assumption that all items are of equal in
difficulty. (A statistical procedure used to estimate coefficient alpha, a
correlation coefficient is given.) Another formula for testing the
internal consistency of a test is the KR-21 formula, which is not
limited to test items that are scored dichotomously.
RELIABILITY
COEFFICIENT
Reliability coefficient is a
measure of the amount of
error associated with the
test scores.
DESCRIPTION OF RELIABILITY
COEFFICIENT
a. The range of the reliability coefficient is
from 0 to 1.0.
b. The acceptable range value 0.60 or higher.
c. The higher the value of the reliability
coefficient the more reliable the overall test
scores.
d. Higher reliability indicates that the test
items measures the same thing,
1. PEARSON PRODUCT MOMENT
CORRELATION COEFFICIENT (rxy)
2. SPEARMAN-BROWN
FORMULA
3. KR-20 AND KR-21
FORMULAS
INTERPRETING
RELIABILITY
COEFFICIENT
1. The group variability will affect the size of
the reliability coefficient. Higher coefficient
results from heterogeneous groups than from
the homogeneous groups. As group variability
increases, reliability goes up.

2. Scoring reliability limits test score


reliability. If tests are scored unreliable, error
is introduced. This will limit the reliability of
the test scores.
3. Test length affects test score
reliability. As the length increases, the
test's reliability tends to go up.

4. Item difficulty affects test score


reliability. As test items become very
easy or very hard, the test's reliability
goes down.
EXAMPLE 1
Prof. Henry Joel conducted a test to
his 10 students in Elementary
Statistics class twice after one-day
interval. The test given after one day
is exactly the same test given the
first time. Scores below were
gathered in the first test (FT) and
second test (ST). Using test-retest
method, is the test reliable? Show the
complete solution.”
Analysis:
The reliability coefficient using the Pearson
r = 0.91 means that it has a very high
reliability. The scores of the 10 students
conducted twice with one-day interval are
consistent. Hence, the test has a very high
reliability.

Note: Compute the reliability coefficient of


the same date using Spearman rho formula.
Is the test reliable?
EXAMPLE 2
Prof. Vinci Glenn conducted a test to his 10
students in his Biology class two times
after one-week interval. The test given
after one week is the parallel form of the
test during the first time the test was
conducted. Scores below were gathered in
the first test (FT) and second test or
parallel test (PT). Using equivalent or
parallel form method, is the test reliable?
Show the complete solution, using the
Pearson r formula.”
Student FT PT
1 12 20
2 20 22
3 19 23
4 17 30
5 25 25
6 22 20
7 15 19
8 16 18
9 23 25
10 21 24
Analysis:
The reliability coefficient using the Pearson
r = 0.76 means that it has a high reliability.
The scores of the 10 students conducted
twice with one-week interval are consistent.
Hence, the test has a high reliability.

Note: Compute the reliability coefficient of


the same date using Spearman rho formula.
Is the test reliable?
EXAMPLE 3

Prof. Glenn Lord conducted a test to his 10


students in his Chemistry class. The test
was given only once. The scores of the
students in odd and even items below were
gathered, (O) odd items and (E) even items.
Using split-half method, is the test reliable?
Show the complete solution.
ANALYSIS

The reliability coefficient using Brown


formula is 0.50, which is questionable
reliability. Hence, the test items should be
revised.
EXAMPLE 4

Ms. Gauat administered a 40-item test


in english for her Grade VI pupils in
Malanao Elementary School. Below
are the scores of 15 pupils, find the
reliability using the Kuder-Richardson
formula.
ANALYSIS

The reliability coefficient using KR-21


formula is 0.90 which means that the test
has a very good reliability. Meaning, the
test is very good for a classroom test.
Steps in Solving the Reliability
Coefficient
1. Solve the Using
difficulty index KR20
of each item(p).
2. Solve the value of q in each item.
3. Find the product of p and q columns.
4. Find the summation of pq.
5. Solve the variance of the scores.
6. Solve the reliability coefficient using KR-20
formula.

The first thing to do is solve the difficulty


index in each and the variance of the total
EXAMPLE 5

Mr. Mark Anthony administered a 20-item


true or false test for his English IV class.
Below are the scores of 40 students. Find the
reliability coefficient using the KR-20 formula
and interpret the computed value, solve also
the coefficient of determination
Interpretation:
The reliability coefficient using the KR20 =
0.98 means that it has a very high reliability
or excellent reliability.

2
Coefficient of determination = (0.98)
=0.9604
= 96.04%
Interpretation:
96.04% of the variance in the students'
performance can be attributed to the test.
THANK
YOU

You might also like

pFad - Phonifier reborn

Pfad - The Proxy pFad of © 2024 Garber Painting. All rights reserved.

Note: This service is not intended for secure transactions such as banking, social media, email, or purchasing. Use at your own risk. We assume no liability whatsoever for broken pages.


Alternative Proxies:

Alternative Proxy

pFad Proxy

pFad v3 Proxy

pFad v4 Proxy