Writing and Reviewing Assessment Items: Guidelines and Tips

Similar documents
Analysis: Evaluation: Knowledge: Comprehension: Synthesis: Application:

How to Judge the Quality of an Objective Classroom Test

Creating Student-Friendly Tests

West s Paralegal Today The Legal Team at Work Third Edition

Maximizing Learning Through Course Alignment and Experience with Different Types of Knowledge

The College Board Redesigned SAT Grade 12

Taxonomy of the cognitive domain: An example of architectural education program

Study Group Handbook

Beyond the Blend: Optimizing the Use of your Learning Technologies. Bryan Chapman, Chapman Alliance

PAGE(S) WHERE TAUGHT If sub mission ins not a book, cite appropriate location(s))

Full text of O L O W Science As Inquiry conference. Science as Inquiry

Instructional Supports for Common Core and Beyond: FORMATIVE ASSESMENT

RETURNING TEACHER REQUIRED TRAINING MODULE YE TRANSCRIPT

Create Quiz Questions

Developing True/False Test Sheet Generating System with Diagnosing Basic Cognitive Ability

Language Acquisition Chart

The Task. A Guide for Tutors in the Rutgers Writing Centers Written and edited by Michael Goeller and Karen Kalteissen

Loughton School s curriculum evening. 28 th February 2017

WORK OF LEADERS GROUP REPORT

Stimulating Techniques in Micro Teaching. Puan Ng Swee Teng Ketua Program Kursus Lanjutan U48 Kolej Sains Kesihatan Bersekutu, SAS, Ulu Kinta

Introducing the New Iowa Assessments Language Arts Levels 15 17/18

Facing our Fears: Reading and Writing about Characters in Literary Text

Grade 11 Language Arts (2 Semester Course) CURRICULUM. Course Description ENGLISH 11 (2 Semester Course) Duration: 2 Semesters Prerequisite: None

Stacks Teacher notes. Activity description. Suitability. Time. AMP resources. Equipment. Key mathematical language. Key processes

Achievement Level Descriptors for American Literature and Composition

Teachers Guide Chair Study

Guidelines for Writing an Internship Report

November 2012 MUET (800)

Introducing the New Iowa Assessments Mathematics Levels 12 14

Ohio s New Learning Standards: K-12 World Languages

Prentice Hall Literature: Timeless Voices, Timeless Themes Gold 2000 Correlated to Nebraska Reading/Writing Standards, (Grade 9)

HISTORY COURSE WORK GUIDE 1. LECTURES, TUTORIALS AND ASSESSMENT 2. GRADES/MARKS SCHEDULE

Grade 4. Common Core Adoption Process. (Unpacked Standards)

Lecturing Module

a) analyse sentences, so you know what s going on and how to use that information to help you find the answer.

Encoding. Retrieval. Forgetting. Physiology of Memory. Systems and Types of Memory

CAAP. Content Analysis Report. Sample College. Institution Code: 9011 Institution Type: 4-Year Subgroup: none Test Date: Spring 2011

Prentice Hall Literature: Timeless Voices, Timeless Themes, Platinum 2000 Correlated to Nebraska Reading/Writing Standards (Grade 10)

Comprehension Recognize plot features of fairy tales, folk tales, fables, and myths.

Common Core Exemplar for English Language Arts and Social Studies: GRADE 1

The role of the first language in foreign language learning. Paul Nation. The role of the first language in foreign language learning

Writing for the AP U.S. History Exam

SETTING STANDARDS FOR CRITERION- REFERENCED MEASUREMENT

Rubric for Scoring English 1 Unit 1, Rhetorical Analysis

Running head: LISTENING COMPREHENSION OF UNIVERSITY REGISTERS 1

Table of Contents. Introduction Choral Reading How to Use This Book...5. Cloze Activities Correlation to TESOL Standards...

ENGLISH. Progression Chart YEAR 8

South Carolina English Language Arts

Smarter Balanced Assessment Consortium:

Scoring Guide for Candidates For retake candidates who began the Certification process in and earlier.

An Analysis of the Early Assessment Program (EAP) Assessment for English

YMCA SCHOOL AGE CHILD CARE PROGRAM PLAN

Kindergarten Lessons for Unit 7: On The Move Me on the Map By Joan Sweeney

success. It will place emphasis on:

Prentice Hall Literature Common Core Edition Grade 10, 2012

Literature and the Language Arts Experiencing Literature

MANAGERIAL LEADERSHIP

Protocol for using the Classroom Walkthrough Observation Instrument

Rendezvous with Comet Halley Next Generation of Science Standards

5. UPPER INTERMEDIATE

AGENDA LEARNING THEORIES LEARNING THEORIES. Advanced Learning Theories 2/22/2016

Changing User Attitudes to Reduce Spreadsheet Risk

Innovative Methods for Teaching Engineering Courses

PREVIEW LEADER S GUIDE IT S ABOUT RESPECT CONTENTS. Recognizing Harassment in a Diverse Workplace

The Extend of Adaptation Bloom's Taxonomy of Cognitive Domain In English Questions Included in General Secondary Exams

Be aware there will be a makeup date for missed class time on the Thanksgiving holiday. This will be discussed in class. Course Description

Alberta Police Cognitive Ability Test (APCAT) General Information

PEDAGOGICAL LEARNING WALKS: MAKING THE THEORY; PRACTICE

Essay on importance of good friends. It can cause flooding of the countries or even continents..

Evaluating Statements About Probability

CONSTRUCTION OF AN ACHIEVEMENT TEST Introduction One of the important duties of a teacher is to observe the student in the classroom, laboratory and

Copyright Corwin 2015

TABE 9&10. Revised 8/2013- with reference to College and Career Readiness Standards

Diagnostic Test. Middle School Mathematics

Common Core State Standards for English Language Arts

TEKS Resource System. Effective Planning from the IFD & Assessment. Presented by: Kristin Arterbury, ESC Region 12

Khairul Hisyam Kamarudin, PhD 22 Feb 2017 / UTM Kuala Lumpur

Build on students informal understanding of sharing and proportionality to develop initial fraction concepts.

Using Proportions to Solve Percentage Problems I

AGS THE GREAT REVIEW GAME FOR PRE-ALGEBRA (CD) CORRELATED TO CALIFORNIA CONTENT STANDARDS

Introducing the New Iowa Assessments Reading Levels 12 14

Arizona s English Language Arts Standards th Grade ARIZONA DEPARTMENT OF EDUCATION HIGH ACADEMIC STANDARDS FOR STUDENTS

TU-E2090 Research Assignment in Operations Management and Services

TROY UNIVERSITY MASTER OF SCIENCE IN INTERNATIONAL RELATIONS DEGREE PROGRAM

OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS

Providing student writers with pre-text feedback

Missouri Mathematics Grade-Level Expectations

CEFR Overall Illustrative English Proficiency Scales

Florida Interim Assessment Item Bank and Test Platform

Teaching Task Rewrite. Teaching Task: Rewrite the Teaching Task: What is the theme of the poem Mother to Son?

NATIONAL CENTER FOR EDUCATION STATISTICS RESPONSE TO RECOMMENDATIONS OF THE NATIONAL ASSESSMENT GOVERNING BOARD AD HOC COMMITTEE ON.

Math Pathways Task Force Recommendations February Background

1 3-5 = Subtraction - a binary operation

Testing for the Homeschooled High Schooler: SAT, ACT, AP, CLEP, PSAT, SAT II

Author: Justyna Kowalczys Stowarzyszenie Angielski w Medycynie (PL) Feb 2015

Philosophy in Literature: Italo Calvino (Phil. 331) Fall 2014, M and W 12:00-13:50 p.m.; 103 PETR. Professor Alejandro A. Vallega.

MADERA SCIENCE FAIR 2013 Grades 4 th 6 th Project due date: Tuesday, April 9, 8:15 am Parent Night: Tuesday, April 16, 6:00 8:00 pm

Organizing Comprehensive Literacy Assessment: How to Get Started

Sample Performance Assessment

Linking the Ohio State Assessments to NWEA MAP Growth Tests *

Transcription:

Writing and Reviewing Assessment Items: Guidelines and Tips by John Painter, Ph.D. September, 2004 Family and Children s Resource Program Jordan Institute for Families School of Social Work University of North Carolina at Chapel Hill 1

The focus of this document is on the practical aspects of planning and writing test items. A well-designed test should represent the instructional objectives of the curriculum; assessment items should be clearly written and fair. There are three basic steps to test construction: planning, development, and review. 1. Step 1: Planning 3 2. Step 2: Development. 6 3. Step 3: Guidelines for review of test items. 9 2

Step 1: Planning. Specifying What is to be Assessed 1. Clarify instructional objectives. Decide what it is students should learn. Include all the objectives that you think are important. Remember that broad, general objectives are of little value in developing test items. 2. Define the purpose of the test. The types of items developed for a test will depend on the purpose of the test. Three common types of test are: a. Pre-instructional assessments. Tend to cover a broad range of topics and skills, and does not attempt to make fine discriminations about individual performance. b. Interim mastery tests. Administered before moving from one instructional unit to another. These tests tend to be relatively brief, but composed of highly specific questions designed to sample a range of cognitive objectives in proportion to what is important for final outcomes. c. Mastery tests. Used for assigning grades, deciding on promotion, or selection for special recognition. These tests are administered infrequently, but carry disproportionate importance because the scores can have longterm implications. It is very important that these tests are thoroughly scrutinized and validated. 3. Specify what will be tested and at what level. Any given objective can be tested at many different levels. Decide whether a student s ability to recall facts, apply principles, or demonstrate mastery is the expected outcome. 4. Develop a test blueprint. After deciding upon important objectives and the specifics of what will be tested develop a plan to guide the number and difficulty of test item construction. See Table 2 for an example. 3

Table 1: Summary of Bloom s Taxonomy of Educational Objectives for the Cognitive Domain. Ordered from most to least frequently used. Level Definition Question Cues: Knowledge Observation and recall of information Knowledge of dates, events, places List, define, tell, describe, identify, show, label, collect, examine, tabulate, quote, name, who, when, where, etc. Knowledge of major ideas Mastery of subject matter Comprehension Understanding information Grasp meaning Translate knowledge into new context Interpret facts, compare, contrast Order, group, infer causes Predict consequences Application Use information Use methods, concepts, theories in new situations Solve problems using required skills or knowledge Analysis Seeing patterns Organization of parts Recognition of hidden meanings Identification of components Synthesis Use old ideas to create new ones Generalize from given facts Relate knowledge from several areas Predict, draw conclusions Evaluation Compare and discriminate between ideas Assess value of theories, presentations Make choices based on reasoned argument Verify value of evidence Recognize subjectivity Summarize, describe, interpret, contrast, predict, associate, distinguish, estimate, differentiate, discuss, extend Apply, demonstrate, calculate, complete, illustrate, show, solve, examine, modify, relate, change, classify, experiment, discover Analyze, separate, order, explain, connect, classify, arrange, divide, compare, select, explain, infer Combine, integrate, modify, rearrange, substitute, plan, create, design, invent, what if?, compose, formulate, prepare, generalize, rewrite Assess, decide, rank, grade, test, measure, recommend, convince, select, judge, explain, discriminate, support, conclude, compare, summarize 4

Planning: Example of a Test Blueprint Here is an example of a test blueprint for a select competencies from the curriculum Table 2: Test Blueprint Foster Home Licensing Competency The worker understands the historical, philosophical, and legal basis of child welfare practice. The worker knows the individual and family characteristics and dynamics that contribute to effective foster and adoptive parenting. The worker knows the values of family-centered practice, including family preservation, permanence for children, preservation of parent s and children s rights, client self-determination, reasonable efforts, and respect for individual and cultural differences. Knowledge Comprehension and Application Analysis, Synthesis, and Evaluation 1 1 2 2 2 2 6 2 3 2 7 5 6 4 15 Total Number of Items Decisions embodied via a test blueprint: How many items are to be constructed for a specific Competency? What cognitive level are questions going to address? The Statewide Partnership has specified that the number of items per competency be determined by the amount of time spent on that competency during training. 5

Step 2: Development: Basic Tenants of Test Item Writing Maxims of Item Writing in General 1. Keep the reading level difficulty of test items appropriate in relation to the individuals who are taking the test, unless of course the purpose is to measure verbal and reading abilities. You do not want language difficulties to interfere with a student s opportunity to show what she knows. 2. Write items as simply as possible, making sure that students know exactly what information is being requested. Irrelevant details, grammatically or logically sloppy construction, or unnecessarily sophisticated vocabulary can contribute to confusion. Often the writer has such a clear idea of what is being solicited that she does not realize that there are alternative interpretations to the question posed. For example: Poor: The United States Constitution was written in. Better: The United States Constitution was written in the year. 3. Test for important ideas, information and skills not trivial detail. It is often easiest to write questions about specific details, but is that information you want to test? While knowledge about specific facts is important on the way to mastering higher-level cognitive skills, it is usually not necessary to test each detail on the path to that higher-level skill. 4. Avoid using direct quotations or verbatim statements from a test or article to form a question. It is not unusual to save time by constructing a test item from a passage in a book or article. Taken out of context such material can be ambiguous or confusing, and it also encourages students to focus primarily on memorization rather than understanding. However, it is acceptable to present a specific passage then refer to it in subsequent questions. 5. If an item is based on opinion or authority, indicate Whose opinion or What authority so that the item clearly has only one correct or best answer. 6. Avoid using interrelated items. If getting the correct answer to one item depends on correctly answering a previous item then undue weight is given to the first item. It is ok to refer to the same stimulus material, as long as getting the correct answer to one item does not depend on the correct answer to another item. 7. Avoid irrelevant clues and give-away questions. Unintentional clues include a tendency to reserve / avoid a certain position for the correct answer (e.g., response D is usually correct / incorrect) or having subject-verb agree between stem and correct response only. Also avoid use of words such as never, and always. 6

8. Avoid trick questions. A sure way to alienate students is to develop questions that confuse trivial detail with major issues. Consider the following True-False question: T F John L. Kennedy was the youngest person to serve as President of the United States. The answer is technically false because President Kennedy s middle initial is F. Even if the student recognizes the difference she would not know if it is a typographical error. The student should not be expected to assess the test writer s motivation or typing skill when answering a question. 9. Ask questions about single ideas; avoid double-barreled questions. 10. Have someone else review your items. Writing Multi-Choice Items A Multiple-choice question consists of a statement or question (the stem), followed by a number of possible responses (the options, distractors, or alternatives). Terms: Multiple-Choice Item: This is the most common objective-type item. The multiplechoice item is a test question which has a number of alternative choices from which the examinee is to select the correct answer. It is generally recommended that one use 4 or 5 choices per question, whenever possible. Using fewer alternatives often results in items with inferior characteristics. The item choices are typically identified on the test copy by the letters A through E. Stem: This is the part of the item in which the problem is stated for the examinee. It can be a question, a set of directions or a statement with an embedded blank. Options/Alternatives: These are the choices given for the item. Key: This is the correct choice for the item. Distractors: These are the incorrect choices for the item. Difficulty: The difficulty level of a multiple choice item will depend on two things: A. The thought process called for by the stem. This can range from recalling factual information to evaluating new information. B. Similarity of the options. Items with options with a high degree of conceptual similarity require greater understanding in order to identify a correct response. 7

1. Distractors should portray a single concept that appears plausible to students who have not mastered the material. Here are some guidelines for writing good distractors: a. Base distractors on the most frequent errors made during class discussions or on assignments related to the concept. b. Use words in the options that are associated with words in the stem. c. Use concepts from the instructional material that have similar vocabulary or were used in the same context as the correct answer. d. Use distractors that are similar in content or form to the correct answer e. Make the distractors similar to the correct response in terms of complexity, sentence structure, and length. 2. Most multiple-choice items should have three to five options. Although more options will reduce the likelihood of guessing the correct answer, the quality of the distractors is most important. 3. Place repetitious words in the stem, instead of in each option. 4. Use None of the above and All of the above sparingly. 5. In general the stem of an item should be meaningful without having to read all the options. In other words, the stem should be a complete sentence or idea. Poor: The President of the United States a. is chosen by the electoral college b. can serve a maximum of two four-year terms Better: The President of the United States is chosen by the a. electoral college b. members of the Senate 6. Avoid elaborate explanations or including background information in an effort to make an item clearer. 7. Options should come at the end of the statement, not in the middle. Although this is not always possible, items written this way are usually easier to understand. Poor: The city of is the capital of California. Better: The capital of California is. 8

Step 3: Guidelines for Review of Test Items The following guidelines are recommended for reviewing individual test items. When you review an item, write your comments on a copy of the item indicating your suggested changes. If you believe an item is not worth retaining, suggest it be deleted. 1. Consider the item as a whole and whether a. it measures knowledge or a skill component which is worthwhile and appropriate for the examinees who will be tested; b. there is a markedly better way to test what this item tests; c. it is of the appropriate level of difficulty for the examinees who will be tested. 2. Consider the stem and whether it a. presents a clearly defined problem or task to the examinee; b. contains unnecessary information; c. could be worded more simply, clearly or concisely. 3. Consider the alternatives and whether a. they are parallel in structure; b. they fit logically and grammatically with the stem; c. they could be worded more simply, clearly or concisely; d. any are so inclusive that they logically eliminate another more restricted option from being a possible answer. 4. Consider the key and whether it a. is the best answer among the set of options for the item; b. actually answers the question posed in the stem; c. is too obvious relative to the other alternatives (i.e., should be shortened, lengthened, given greater numbers of details, made less concrete). 5. Consider the distractors and whether a. there is any way you could justify one or more as an acceptable correct answer; b. they are plausible enough to be attractive to examinees who are misinformed or ill-prepared; c. any one calls attention to the key (e.g., no distractor should merely state the reverse of the key or resemble the key very closely unless another pair of choices is similarly parallel or involves opposites). 9

References Bloom, B.S. (Ed.) (1956). Taxonomy of educational objectives: The classification of educational goals: Handbook I, cognitive domain. New York ; Toronto: Longman, Green. Krathwokl, D. (1993). Methods of Educational and Social Science Research. New York: Longman. Thorndike, R. and Hagen, E. (1955). Measurement and Evaluation in Psychology and Education. New York: Chapman & Hall, Limited. Worthen, B., Borg, W., and White, K. (1993) Measurement and Evaluation in the Schools. New York: Longman. Wollack, James A. and Cohen, Allan S.. Handbook on Test Development: Helpful Tips for Creating Reliable and Valid Classroom Tests. Testing & Evaluation Services University of Wisconsin-Madison available on line at: <http://wiscinfo.doit.wisc.edu/exams/handbook%20on%20test%20construction.pdf> 10