PSYCHOLOGICAL ASSESSMENT CHAPTER 3-6 (Summary)
PSYCHOLOGICAL ASSESSMENT CHAPTER 3-6 (Summary)
Measures of Variability
Variability is an indication of how scores
in a distribution are scattered or
dispersed.
Meta-Analysis
- techniques used to statistically
combine information across studies
- Effect size: estimates derived from meta-
analysis
Criterion-Referenced (Domain-Referenced):
method of evaluation and a way of deriving meaning
from test scores by evaluating an
Assumptions about Psychological individual's test score with reference to a set
Testing and Assessment standard.
1. Psychological traits and states exist
2. Psychological traits and states can be Norm-Referenced: a method of evaluation and a
quantified and
way of deriving meaning from test scores by
measured
evaluating an individual's test score and comparing it
3. Test-related behavior predicts non-test
to scores of group of testtakers on the same test.
related behavior
4. Tests and other measurement
techniques have strengths SAMPLING TO DEVELOP NORMS
and weaknesses Population- set of individuals with at least one
5. Various sources of error are part of the common, observable characteristic.
assessment process
6. Testing and assessment can be Sample- representative of the whole population.
conducted fair and unbiased
manner Sampling- the process of selection the portion
7. Testing and assessment benefit society of the universe deemed to be representative of
the whole population.
Trait- any distinguishable, relatively
enduring way in which one individual Types of sampling:
varies from another. 1. Stratified sampling- researchers divide subjects
into
States- relatively less enduring subgroups called strata based on characteristics that
Psychological traits exist as constructs they share.
from overt
behavior. 2. Convenience/ Incidental sampling- is a non-
probability sampling method where units are
Norms- used as a frame of reference for
selected for inclusion in the sample because they are
interpreting test scores.
the easiest for the researcher to access.
Normative sample or norm group-
group of people presumed to be 3. Purposive sampling- is a non-probability sampling
representative of the universe of people in which researchers rely on their own judgment
when choosing members of the population to
Norming- process of deriving norms. participate in their surveys.
Types of error:
1. Random error: a source of error in measuring a targeted variable caused by unpredictable fluctuations
and inconsistencies of other variables in the measurement process.
2. Systematic error: a source of error in measuring a variable that is typically constant or proportionate to
what is presumed to be true value of the variable being measured.
Sources of Error Variance:
• Test Construction- item sampling/ content sampling
• Test Administration
- test environment
- Testtaker variables
- Examiner-related variables
• Test Scoring and Interpretation- objectivity/ subjectivity
• Other sources: sampling error, methodological error
Reliability Estimates
1. Test-Retest Reliability- obtained by correlating a pair of scores from the same people on two different
administrations of the test.
- As the time interval between administrations of the same test increases, the correlation between the scores
obtained on each testing decreases.
- Passage of time can be a source of variance.
- Coefficient of stability- refers to the estimate of test-retest reliability when the interval between testing is
greater than six
months.
2. Parallel/ Alternate Forms Reliability- obtained y correlating scores various forms of a test.
Coefficient of equivalence- degree of the relationship between various forms of a test can be evaluated by
means of an alternate-forms or parallel-forms.
- Parallel Form: two forms of the same test and the means and variances of observed test scores are
equal.
- Alternate Form: different versions of a test that have been constructed so as to be parallel.
3. Internal Consistency/ Inter-item Consistency- refers to the degree of correlation among all the items on a
scale.
- It its useful in assessing the homogeneity of a test.
- Homogeneity- refers to the extent to which items in a scale measure a single trait.
- Heterogeneity- the degree to which a test measures different factors.
Methods:
4. Split- Half Reliability- obtained by correlating two pairs from equivalent halves of a single test
administered once.
•odd-even reliability
• Spearman-brown formula
Other methods of estimating Internal Consistency:
Kuder-Richardson formulas (KR-20/ KR-21)
- By G. Frederic Kuder and M.W. Richardson
- KR-20- highly homogenous & dichotomous items
- KR-21- simplified version KR-20, assuming items have the same difficulty
CHAPTER 6: VALIDITY
VALIDITY- refers to how well a test measures what it purports to measure in a particular context.
Validation- process of gathering and evaluating evidence about validity.
Validation studies- are used in research to compare the accuracy of a measure with a gold standard measure
to identify and eliminate bias.
Local validation studies- when the test user alters a tool in some way, for example, the format, instructions,
language, or content for a particular population of testtakers.
Types of Validity
Face Validity- a judgement concerning how relevant test items appear to be.
Lack of face validity: lack confidence in the perceived effectiveness of the test and decrease in the testtakers
cooperation and motivation.
Content Validity- a judgement of how adequately a test samples behavior representative of the universe of
behavior that the test was designed to sample.
oTable of Specifications/ Test blueprint- a plan regarding the types of information to be covered by the
items, the number of items tapping each area of coverage, the organization of items in the test.
Criterion-Related Validity- judgement of how adequately a test score can be used to infer an individual's most
probable standing on some measure of a criterion.
Criterion- is the standard against which a test or a test score is evaluated.
Characteristics of a criterion:
1. Relevant 2. Valid 3. Uncontaminated