Purpose of Oral Assessment Tests2 Types: • Diagnostic Tests • Placement Tests... Designing Assessment ToolsMust have: • Purpose: Oral assessment must be tailor-made for a particular pur
Trang 1Barbara Thompson
Trang 2Teaching Oral English Communicatively
Barbara Thompson
Trang 3A Holistic Approach to Assessing Levels
of Spoken English
Barbara Thompson
Trang 4Purpose of Oral Assessment Tests
2 Types:
• Diagnostic Tests
• Placement Tests
Trang 5Placement Tests
Purpose:
• For program entry/exit or placement within a course of study
• Proficiency threshold or cutoff - students can pass or fail
• Tests can be quantifiable or holistic
Trang 6Administrative Challenges
• Time consuming to design and administer
• Scheduling constraints
• Expensive
• Can be subjective
• Rater training and turnover
Trang 7Designing Assessment Tools
Must have:
• Purpose: Oral assessment must be tailor-made
for a particular purpose and population
• Reliability: Accuracy and consistency of
measurement
• Validity: Justification or interpretation of test
results
Trang 8Designing Assessment Tools (cont.)
Must have:
Instructional authenticity: Engagement of test
taker’s communicative language ability
Practicality: Constraints imposed by time, money,
personnel, equipment, institutional or
governmental policies
Face validity: Assessment tests what was taught
Trang 9Rating of Performance Tests
Rating scales generally have 2 features:
1 Each aspect of language to be rated is listed
(oral grammar, phrasal stress, etc.)
2 Definition of each level in terms of mastery
of language features
Trang 10Designing an Assessment
What tasks will the test takers perform?
• Sentence repetition or completion: lower
levels
• Oral interview: face to face, telephone, skype
• Role play: situations relevant to test takers
Trang 11Designing an Assessment (cont.)
What tasks will the test takers perform?
• Small group discussions
• Answer questions or give opinions
• Describe a picture, graph, or tell a story
• Oral presentation – more formal
Trang 12Developing Test Questions
• List your objectives or content area
• Develop items that meet these objectives
• Pilot with groups of students to see if they elicit the expected language, and are
scoreable
Trang 13Developing Test Questions (cont.)
• Decide on the number of test items
• Avoid yes/no questions
• To raise difficulty level, create questions that test multiple things
Trang 14Standardize Test Administration
• Will the test be live, taped, or on computer? (test should be administered the same way each time)
• Is repetition or rephrasing by the tester
allowed?
Trang 15Standardize Rater Training
Testers must administer tests in a standard way, and be able to score reliably
• Training must always cover the same points
• Trainers and trainees must have a thorough
understanding of the rubric
Trang 16Standardize Test Administration (cont.)
• What should the tester do if the student asks
a question or doesn’t respond?
• Is there a time limit for each item?
• Can students read the questions during the
test?
Trang 17Test Security
• Test items are never practiced in class
• Do not discuss items with students before,
during, or after the test
• If possible have several versions of tests (pilot
to standardize versions)
• Video or audio tape each test
Trang 18Standardize Rater Training (cont.)
• Trainees should do many practice exercises
• Trainees should do certification exercises
• Inter-rater reliability
• Periodic re-norming