Professional Documents
Culture Documents
1. 2. 3. 4.
5.
is not Excessively Expensive It stays within appropriate time constraints It is relatively easy to administer, and It has a scoring/evaluation procedure that is specific and time-efficient
2) RELIABILITY
A Reliable
It has several fluctuations: o Student-Related Reliability o Rater Reliability o Test Administrator Reliability, and o Test Reliability
Student-Related Reliability
categories
by physical or psychological factor such as temporary illness, fatigue, a bad day and anxiety. Also such factors as a test-takers test-wiseness or strategies for efficient test taking
caused
Rater Reliability Human error, subjectivity and bias may occur in the process of scoring. In rater reliability there are two categories: inter-rater reliability; it occurs when two or
Sometimes the test itself can cause measurement error Such as: the time limitations or ambiguous test items
3) Validity
This term refers to the extent to which a test measures what it says it measures. (Coomble and Hubbley, 2009,5)
According to McNamara(2000): it is the reltionhsip between evidence from the test performance and infernces about candidates capacity to perform in the criterion that are drwn from that evidence. In other words, test what you teach, how you teach it!
VALIDITY
CONTENT VALIDITY
CONSTRUCT VALIDITY
FACE VALIDITY
Content Validity
For classroom teashers it means that the test assesses the course content and outcomes using format familiar to the student.
Construct validity
Refers to the fit between the underlying theories and methedology of language learning and the type of assessment. Example: a communicative language learning approach must be matched by communication language testing
Face Validity
It means that the test looks as though it measures what it is supposed to measure
Face vallidity will likely be high if leaners encounter:
Face Validity
Criterion-related validiy
Consequential Validity
Criterion-related validiy:
The
Consequential Validity:
4)Authenticity
5) Washback
Coombe nad Hubley (2009:8) Proposed proposed two other criteria before constructing test: Trasparency
and Security
Planning
a) Establish purpose of the test
Place student program Acheivement of course outcomes Diognosis of the strengths and areas improvements International benchmark b) Identify objectives Operationalizes outcomes c) Decide on cut-offs Grade, mastery d) Inventory course content and materials Consider approriate formats Establishe overall weighting e) Scheduling f) Write test specifications
Map the exam Decide on sections, formats, weighting Construct items according to test specifications Establish grading criteria Prepare an answer key Vett the exam Pilot the exam
Test Administration
a). Decide on test conditions and procedures b) Organize equipment needed c) Establish makeup policy d) Inform student about availability of result
Learn from each exam Did it serve its purpose? What was the fit with curricular outcomes? Was it valid and reliable test? Was it part of the studentss learning experience? What future changes would you make?