0% found this document useful (0 votes)
70 views2 pages

Assessment

This document discusses principles of high quality assessment. It outlines eight key principles: 1) clarity of learning targets, 2) appropriateness of assessment methods, 3) validity, 4) reliability, 5) fairness, 6) positive consequences, 7) practicality and efficiency, and 8) ethics. It describes different types of assessment methods like essays, checklists, and oral questioning. It also explains the importance of an assessment having validity, reliability, fairness, positive impact, practical application, and adherence to ethical standards. The overall purpose is to provide guidance on designing assessments that accurately measure student learning.

Uploaded by

Ma Elena Llunado
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
70 views2 pages

Assessment

This document discusses principles of high quality assessment. It outlines eight key principles: 1) clarity of learning targets, 2) appropriateness of assessment methods, 3) validity, 4) reliability, 5) fairness, 6) positive consequences, 7) practicality and efficiency, and 8) ethics. It describes different types of assessment methods like essays, checklists, and oral questioning. It also explains the importance of an assessment having validity, reliability, fairness, positive impact, practical application, and adherence to ethical standards. The overall purpose is to provide guidance on designing assessments that accurately measure student learning.

Uploaded by

Ma Elena Llunado
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
You are on page 1/ 2

PRINCIPLES OF HIGH QUALITY ASSESSMENT  Essays – can test the students’ grasp of the higher level cognitive skills

1. Clarity of learning targets(knowledge, reasoning, skills, products, affects)  Checklists – list of several characteristics or activities presented to the subjects of a
2. Appropriateness of Assessment Methods study, where they will analyze and place a mark opposite to the characteristics.
3. Validity b. Product Rating Scales
4. Reliability  Used to rate products like book reports, maps, charts, diagrams, notebooks, creative
5. Fairness endeavors
6. Positive Consequences  Need to be developed to assess various products over the years
7. Practicality and Efficiency c. Performance Tests - Performance checklist
8. Ethics  Consists of a list of behaviors that make up a certain type of performance
CLARITY OF LEARNING TARGETS (knowledge, reasoning, skills, products, affects)  Used to determine whether or not an individual behaves in a certain way when asked
Assessment can be made precise, accurate and dependable only if what are to be to complete a particular taskd.
achieved are clearly stated and feasible. The learning targets, involving knowledge, d. Oral Questioning – appropriate assessment method when the objectives are to:
reasoning, skills, products and effects, need to be stated in behavioural terms which  Assess the students’ stock knowledge and/or
denote something which can be observed through the behaviour of the students.  Determine the students’ ability to communicate ideas in coherent verbal sentences.
1. CLARITY OF LEARNING TARGETS e. Observation and Self Reports
Cognitive Targets  Useful supplementary methods when used in conjunction with oral questioning and
Benjamin Bloom (1954) proposed a hierarchy of educational objectives at the performance tests
cognitive level. These are: PROPERTIES OF ASSESSMENT METHODS
• Knowledge – acquisition of facts, concepts and theories
 Validity
• Comprehension - understanding, involves cognition or awareness of the
 Reliability
interrelationships
 Fairness
• Application – transfer of knowledge from one field of study to another of from one
 Positive Consequences
concept to another concept in the same discipline
 Practicality and Efficiency
• Analysis – breaking down of a concept or idea into its components and explaining
 Ethics
g the concept as a composition of these concepts
3. VALIDITY
• Synthesis – opposite of analysis, entails putting together the components in order
to summarize the concept  Something valid is something fair.
• Evaluation and Reasoning – valuing and judgment or putting the “worth” of a  A valid test is one that measures what it is supposed to measure.
concept or principle. Types of Validity
SKILLS, COMPETENCIES AND ABILITIES TARGETS  Face: What do students think of the test?
 Skills – specific activities or tasks that a student can proficiently do  Construct: Am I testing in the way I taught?
 Competencies – cluster of Skills  Content: Am I testing what I taught?
 Abilities – made up of relate competencies categorized as:  Criterion-related: How does this compare with the existing valid test? Tests can be
i. Cognitive made more valid by making them more subjective (open items).
ii. Affective Validity – appropriateness, correctness, meaningfulness and usefulness of the
iii. Psychomotor specific conclusions that a teacher reaches regarding the teaching-learning situation.
 Products,Outputs and Project Targets - tangible and concrete evidence of a  Content validity – content and format of the instrument
student’s ability - need to clearly specify the level of workmanship of projects i. Students’ adequate experience
i. expert ii. Coverage of sufficient material
ii. skilled iii. Reflect the degree of emphasis
iii. novice  Face validity – outward appearance of the test, the lowest form of test validity
2. APPROPRIATENESS OF ASSESSMENT METHODS  Criterion-related validity – the test is judge against a specific criterion
a. Written - Response Instruments  Construct validity – the test is loaded on a “construct” or factor
 Objective tests – appropriate for assessing the various levels of hierarchy of 4. RELIABILITY
educational objectives  Something reliable is something that works well and that you can trust.
 A reliable test is a consistent measure of what it is supposed to measure. - Positive consequence on students, teachers, parents, and other stakeholders
Questions:
 Can we trust the results of the test? 7. PRACTICALITY AND EFFECIENCY
 Would we get the same results if the tests were taken again and scored by a different  Something practical is something effective in real situations.
person?  A practical test is one which can be practically administered.
Tests can be made more reliable by making them more objective (controlled items). Questions:
 Reliability is the extent to which an experiment, test, or any measuring procedure  Will the test take longer to design than apply?
yields the same result on repeated trials.  Will the test be easy to mark?
 Equivalency reliability is the extent to which two items measure identical concepts Tests can be made more practical by making it more objective (more controlled
at an identical level of difficulty. Equivalency reliability is determined by relating two items)
sets of test scores to one another to highlight the degree of relationship or  Teacher Familiarity with the Method
association.  Time required
 Stability reliability (sometimes called test, re-test reliability) is the agreement of  Complexity of Administration
measuring instruments over time. To determine stability, a measure or test is  Ease of scoring
repeated on the same subjects at a future date.  Ease of Interpretation
 Internal consistency is the extent to which tests or procedures assess the same  Cost
characteristic, skill or quality. It is a measure of the precision between the observers Teachers should be familiar with the test,
or of the measuring instruments used in a study. - does not require too much time
 Interrater reliability is the extent to which two or more individuals (coders or raters) - implementable
agree. Interrater reliability addresses the consistency of the implementation of a RELIABILITY, VALIDITY & PRACTICALITY
rating system. The problem:
RELIABILITY – CONSISTENCY, DEPENDABILITY, STABILITY WHICH CAN BE  The more reliable a test is, the less valid.
ESTIMATED BY:  The more valid a test is, the less reliable.
 Split-half method  The more practical a test is, (generally) the less valid.
 Calculated using the The solution:
i. Spearman-Brown prophecy formula As in everything, we need a balance (in both exams and exam items)
ii. Kuder-Richardson – KR 20 and KR21 8. ETHICS
 Consistency of test results when the same test is administered at two different time  Informed consent
periods  Anonymity and Confidentiality
i. Test-retest method 1. Gathering data
ii. Correlating the two test results 2. Recording Data
3. Reporting Data
5. FAIRNESS ETHICS IN ASSESSMENT – “RIGHT AND WRONG”
The concept that assessment should be 'fair' covers a number of aspects.  Conforming to the standards of conduct of a given profession or group
 Student Knowledge and learning targets of assessment  Ethical issues that may be raised
 Opportunity to learn i. Possible harm to the participants.
 Prerequisite knowledge and skills ii. Confidentiality.
 Avoiding teacher stereotype iii. Presence of concealment or deception.
 Avoiding bias in assessment tasks and procedures iv. Temptation to assist students.

6. POSITIVE CONSEQUENCES
Learning assessments provide students with effective feedback and potentially
improve their motivation and/or self-esteem. Moreover, assessments of learning
gives students the tools to assess themselves and understand how to improve.

You might also like

pFad - Phonifier reborn

Pfad - The Proxy pFad of © 2024 Garber Painting. All rights reserved.

Note: This service is not intended for secure transactions such as banking, social media, email, or purchasing. Use at your own risk. We assume no liability whatsoever for broken pages.


Alternative Proxies:

Alternative Proxy

pFad Proxy

pFad v3 Proxy

pFad v4 Proxy