PLEASE SCROLL DOWN FOR ARTICLE

Similar documents
To link to this article: PLEASE SCROLL DOWN FOR ARTICLE

Philip Hallinger a & Arild Tjeldvoll b a Hong Kong Institute of Education. To link to this article:

MMOG Subscription Business Models: Table of Contents

PLEASE SCROLL DOWN FOR ARTICLE. Full terms and conditions of use:

Zealand Published online: 16 Jun To link to this article:

PLEASE SCROLL DOWN FOR ARTICLE

PLEASE SCROLL DOWN FOR ARTICLE

ScienceDirect. Noorminshah A Iahad a *, Marva Mirabolghasemi a, Noorfa Haszlinna Mustaffa a, Muhammad Shafie Abd. Latif a, Yahya Buntat b

A Note on Structuring Employability Skills for Accounting Students

ASSESSMENT REPORT FOR GENERAL EDUCATION CATEGORY 1C: WRITING INTENSIVE

Andrew S. Paney a a Department of Music, University of Mississippi, 164 Music. Building, Oxford, MS 38655, USA Published online: 14 Nov 2014.

Writing a Basic Assessment Report. CUNY Office of Undergraduate Studies

A Study of Metacognitive Awareness of Non-English Majors in L2 Listening

Published online: 26 Mar 2010.

1GOOD LEADERSHIP IS IMPORTANT. Principal Effectiveness and Leadership in an Era of Accountability: What Research Says

Psychometric Research Brief Office of Shared Accountability

Concept mapping instrumental support for problem solving

Show and Tell Persuasion

Further, Robert W. Lissitz, University of Maryland Huynh Huynh, University of South Carolina ADEQUATE YEARLY PROGRESS

MASTER S THESIS GUIDE MASTER S PROGRAMME IN COMMUNICATION SCIENCE

Online publication date: 07 June 2010

A Coding System for Dynamic Topic Analysis: A Computer-Mediated Discourse Analysis Technique

Evidence for Reliability, Validity and Learning Effectiveness

PLEASE SCROLL DOWN FOR ARTICLE. Full terms and conditions of use:

How to Judge the Quality of an Objective Classroom Test

Management of time resources for learning through individual study in higher education

TCH_LRN 531 Frameworks for Research in Mathematics and Science Education (3 Credits)

Helma W. Oolbekkink Marchand a, Jan H. van Driel b & Nico Verloop b a Radboud University Nijmegen, The Netherlands. Published online: 24 Jan 2007.

Understanding and Interpreting the NRC s Data-Based Assessment of Research-Doctorate Programs in the United States (2010)

Content Teaching Methods: Social Studies. Dr. Melinda Butler

Conceptual and Procedural Knowledge of a Mathematics Problem: Their Measurement and Their Causal Interrelations

Developing an Assessment Plan to Learn About Student Learning

Kelso School District and Kelso Education Association Teacher Evaluation Process (TPEP)

English Language Arts Summative Assessment

Development of a scoring system to assess mind maps

ACADEMIC AFFAIRS GUIDELINES

Journal Article Growth and Reading Patterns

Available online: 03 Nov 2011

Learning Objectives by Course Matrix Objectives Course # Course Name Psyc Know ledge

Delaware Performance Appraisal System Building greater skills and knowledge for educators

OVERVIEW OF CURRICULUM-BASED MEASUREMENT AS A GENERAL OUTCOME MEASURE

Georgetown University School of Continuing Studies Master of Professional Studies in Human Resources Management Course Syllabus Summer 2014

EVALUATING MATH RECOVERY: THE IMPACT OF IMPLEMENTATION FIDELITY ON STUDENT OUTCOMES. Charles Munter. Dissertation. Submitted to the Faculty of the

Using Moodle in ESOL Writing Classes

SY 6200 Behavioral Assessment, Analysis, and Intervention Spring 2016, 3 Credits

Running head: LISTENING COMPREHENSION OF UNIVERSITY REGISTERS 1

Last Editorial Change:

Graduate Program in Education

The Good Judgment Project: A large scale test of different methods of combining expert predictions

Instructor: Mario D. Garrett, Ph.D. Phone: Office: Hepner Hall (HH) 100

Maximizing Learning Through Course Alignment and Experience with Different Types of Knowledge

UCLA Issues in Applied Linguistics

Saeed Rajaeepour Associate Professor, Department of Educational Sciences. Seyed Ali Siadat Professor, Department of Educational Sciences

Planning a research project

Navitas UK Holdings Ltd Embedded College Review for Educational Oversight by the Quality Assurance Agency for Higher Education

OPAC and User Perception in Law University Libraries in the Karnataka: A Study

Physics Experimental Physics II: Electricity and Magnetism Prof. Eno Spring 2017

International Conference on Current Trends in ELT

ADDIE MODEL THROUGH THE TASK LEARNING APPROACH IN TEXTILE KNOWLEDGE COURSE IN DRESS-MAKING EDUCATION STUDY PROGRAM OF STATE UNIVERSITY OF MEDAN

On-the-Fly Customization of Automated Essay Scoring

ATW 202. Business Research Methods

CHAPTER III RESEARCH METHOD

Update on Standards and Educator Evaluation

Approaches for analyzing tutor's role in a networked inquiry discourse

PREDISPOSING FACTORS TOWARDS EXAMINATION MALPRACTICE AMONG STUDENTS IN LAGOS UNIVERSITIES: IMPLICATIONS FOR COUNSELLING

George Mason University Graduate School of Education Program: Special Education

SCHEMA ACTIVATION IN MEMORY FOR PROSE 1. Michael A. R. Townsend State University of New York at Albany

Enhancing Students Understanding Statistics with TinkerPlots: Problem-Based Learning Approach

Language Arts Methods

Practical Research. Planning and Design. Paul D. Leedy. Jeanne Ellis Ormrod. Upper Saddle River, New Jersey Columbus, Ohio

TAIWANESE STUDENT ATTITUDES TOWARDS AND BEHAVIORS DURING ONLINE GRAMMAR TESTING WITH MOODLE

APA Basics. APA Formatting. Title Page. APA Sections. Title Page. Title Page

Standards-Based Bulletin Boards. Tuesday, January 17, 2012 Principals Meeting

Department of Geography Bachelor of Arts in Geography Plan for Assessment of Student Learning Outcomes The University of New Mexico

GDP Falls as MBA Rises?

Procedia - Social and Behavioral Sciences 98 ( 2014 ) International Conference on Current Trends in ELT

Greek Teachers Attitudes toward the Inclusion of Students with Special Educational Needs

SCIENCE DISCOURSE 1. Peer Discourse and Science Achievement. Richard Therrien. K-12 Science Supervisor. New Haven Public Schools

Designing Case Study Research for Pedagogical Application and Scholarly Outcomes

Aronson, E., Wilson, T. D., & Akert, R. M. (2010). Social psychology (7th ed.). Upper Saddle River, NJ: Prentice Hall.

New Ways of Connecting Reading and Writing

Kendriya Vidyalaya Sangathan

IMPROVING STUDENTS READING COMPREHENSION USING FISHBONE DIAGRAM (A

Youth Mental Health First Aid Instructor Application

Statistical Analysis of Climate Change, Renewable Energies, and Sustainability An Independent Investigation for Introduction to Statistics

Daniel Reinholz a a Center for STEM Learning, University of Colorado, Boulder, CO, USA Published online: 11 Feb 2015.

Digital Media Literacy

Young Enterprise Tenner Challenge

What Makes Professional Development Effective? Results From a National Sample of Teachers

Pragmatic Use Case Writing

1 3-5 = Subtraction - a binary operation

A Retrospective Study

New Venture Financing

Research Design & Analysis Made Easy! Brainstorming Worksheet

KENTUCKY FRAMEWORK FOR TEACHING

Room: Office Hours: T 9:00-12:00. Seminar: Comparative Qualitative and Mixed Methods

Use of the Kalamazoo Essential Elements Communication Checklist (Adapted) in an Institutional Interpersonal and Communication Skills Curriculum

Sheila M. Smith is Assistant Professor, Department of Business Information Technology, College of Business, Ball State University, Muncie, Indiana.

Foothill College Fall 2014 Math My Way Math 230/235 MTWThF 10:00-11:50 (click on Math My Way tab) Math My Way Instructors:

Procedia - Social and Behavioral Sciences 191 ( 2015 ) WCES Why Do Students Choose To Study Information And Communications Technology?

Transcription:

This article was downloaded by: [University at Buffalo, the State University of New York (SUNY)] On: 30 October 2009 Access details: Access Details: [subscription number 784375718] Publisher Routledge Informa Ltd Registered in England and Wales Registered Number: 1072954 Registered office: Mortimer House, 37-41 Mortimer Street, London W1T 3JH, UK Teaching of Psychology Publication details, including instructions for authors and subscription information: http://www.informaworld.com/smpp/title~content=t775653707 Developing a Rubric to Assess Student Learning Outcomes Using a Class Assignment Nicholas Thaler a ; Ellie Kazemi a ; Crystal Huscher a a California State University, Northridge Online Publication Date: 01 April 2009 To cite this Article Thaler, Nicholas, Kazemi, Ellie and Huscher, Crystal(2009)'Developing a Rubric to Assess Student Learning Outcomes Using a Class Assignment',Teaching of Psychology,36:2,113 116 To link to this Article: DOI: 10.1080/00986280902739305 URL: http://dx.doi.org/10.1080/00986280902739305 PLEASE SCROLL DOWN FOR ARTICLE Full terms and conditions of use: http://www.informaworld.com/terms-and-conditions-of-access.pdf This article may be used for research, teaching and private study purposes. Any substantial or systematic reproduction, re-distribution, re-selling, loan or sub-licensing, systematic supply or distribution in any form to anyone is expressly forbidden. The publisher does not give any warranty express or implied or make any representation that the contents will be complete or accurate or up to date. The accuracy of any instructions, formulae and drug doses should be independently verified with primary sources. The publisher shall not be liable for any loss, actions, claims, proceedings, demand or costs or damages whatsoever or howsoever caused arising directly or indirectly in connection with or arising out of the use of this material.

Teaching of Psychology, 36: 113 116, 2009 Copyright C Taylor & Francis Group, LLC ISSN: 0098-6283 print / 1532-8023 online DOI: 10.1080/00986280902739305 FACULTY FORUM Downloaded By: [University at Buffalo, the State University of New York (SUNY)] At: 03:00 30 October 2009 Developing a Rubric to Assess Student Learning Outcomes Using a Class Assignment Nicholas Thaler, Ellie Kazemi, and Crystal Huscher California State University, Northridge We developed a rubric to assess several of our department s undergraduate student learning outcomes (SLOs). Target SLOs include applications of principles of research methodology, using appropriate statistics, adherence to the Publication Manual of the American Psychological Association, and written communication skills. We randomly sampled 20 percent (N = 55) of the final written manuscripts from several sections of a research methods course and trained 2 graduate-level raters to use the rubric to score the students papers. We found statistically significant interrater reliability and convergent validity coefficients. These findings are discussed to encourage the development and evaluation of such rubrics to be used across colleges and universities. In the last decade, undergraduate psychology programs have responded to the need for assessment as a legitimate force in higher education (see Dunn, Mehrotra, & Halonen, 2004). As such, there is an increasing interest in multimethod assessment of student learning outcomes and the development of reliable and valid rubrics to directly measure student learning (e.g., Stellmack, Konheim-Kalkstein, Manor, Massey, & Schmitz, 2009). However, it is quite challenging to develop rubrics that reliably assess student learning based on oral and visual presentations, group work and discussions, final projects, written reports, and other culminated student work. Halonen et al. (2003) created a rubric specifically designed for psychology departments to assess undergraduate students scientific inquiry skills. Scientific inquiry is defined as forming hypotheses, designing experiments, interpreting outcomes, and communicating results. These four domains converge with McGovern and Hawks s (1986) conclusions on what defines psychology student learning objectives (SLOs) specifically that scientific reasoning and proficiencies in research methods and communication are their principal components. The rubric that Halonen and colleagues (2003) designed is comprehensive and broad, incorporating elements of communication, collaboration, and selfassessment, as well as scientific inquiry skills. Each domain has five levels of proficiency, ranging from before training to professional graduate and beyond. They designed the rubric to measure a student s progress throughout his or her academic experience, from as early as high school up to college graduation. One limitation Halonen and colleagues (2003) acknowledged is the time and energy required to effectively use a comprehensive multidimensional rubric. Consequently, some educators might lose enthusiasm in this rubric s implementation. This limitation can be addressed by focusing on a simpler rubric and using itasonepartofamultimethodmodelofassessment. Furthermore, Halonen and colleagues rubric was not empirically tested for reliability and validity. The authors hoped that their article would encourage other researchers to investigate the potential of developing their own rubrics to quantify and assess psychology education achievement. It is possible that designing a more specific rubric for a single assignment or final project might provide educators with an easier way of evaluating their students achievement, provided that Vol. 36, No. 2, 2009 113

the assignment or final project is representative and appropriately selected. The purpose of this study was to follow Halonen et al. s (2003) footsteps and to empirically develop a rubric that assesses achievement of certain learning outcomes using psychology students research manuscripts. We constructed a rubric based on the criteria of the Publication Manual of the American Psychological Association (APA, 2001). This rubric was designed to specifically measure the degree to which our students achieved some of the SLOs of our department. Our rubric has a strong basis for its design in that most of its content was directly lifted from the APA Manual. We hope that our findings will demonstrate the potential of designing rubrics to assess psychology SLOs in both an objective and reliable manner. Participants Method For this study, we randomly selected the names of 20% of the enrolled students in each section of a research methods course and sent the names of the selected students to the course instructors. All instructors offered copies of the selected students final papers which resulted in a total of 55 manuscripts (women = 52, men = 3). Because the data contained no identifying information, the California State University, Northridge (CSUN) Institutional Review Board for Protection of Human Subjects approved this study. Table 1. Materials The Psychology Department at CSUN has identified eight specific SLOs adopting the Task List goals established by the American Psychological Association Task Force (2007). CSUN s SLOs encompass many of the skills required to conceptualize and design an experiment, use the appropriate statistical tests, and write a manuscript in accordance with the APA guidelines. The rubric has 10 items that assess four of our SLOs: the use and interpretation of statistical techniques, critical thinking skills and skeptical inquiry in evaluating their own and others research, competencies in electronic and information technologies, and effective written communication skills. Each of the 10 categories has a 6-point Likert scale. A copy of the rubric can be viewed at http://docs.google.com/doc?id=df6b863n 0dw8dm3gj. Procedure A focus group of faculty members initially developed the rubric. We recruited two raters and revised the rubric twice based on the raters feedback. Of the 55 manuscripts, both raters coded 22 (i.e., 40%) for assessment of interrater reliability. We also compared the final ratings of the manuscripts that were based on the rubric with each individual instructor s evaluation methods. Results and Discussion The interrater reliability between the two raters was analyzed using Spearman s correlations (see Table 1 ). Mean, Medians, Standard Deviation, Inter-Rater Reliability Between the Two Coders, and Convergent Validity M Mdn SD IRR r S CV rp 1. Adherence to APA Style 3.9 4 1.0.58.34 2. Quality of Abstract 4.2 5 1.1.65.43 3. Relevant Literature 4.1 4 1.0.59.74 4. Hypotheses Oper. Defined 3.6 4 1.3.35.46 5. Quality of Methods 3.2 3 1.1.69.51 6. Quality of Results 3.4 3 1.4.89.29 7. Quality of Discussion 3.8 4 1.0.70.70 8. Statement Pros/Cons 3.9 4 1.1.64.23 9. Use of Charts/Figures 1 2.8 3 1.5.79.18 10. Written Communication 3.5 4 1.1.92.71 Note. p <.05 Note. p <.01 1 Note. Data does not include missing charts and represents 91% of the data. (50 out of 55)M 114 Teaching of Psychology

Average interrater reliability was r s (22) =.68, p <.01. The Quality of Results, r s (22) =.89, p <.01, and the Overall Written Communication items, r s (22) =.92, p <.01, had the strongest correlations, and the correlation coefficient of the Hypotheses Operationally Defined item was not statistically significant, r s (22) =.35, p <.11. All other correlations between individual items ranged from.58 to.79 and were large (see Cohen, 1988, for interpretations of correlation coefficients as related to reliability). We analyzed convergent validity using Pearson correlations (see Table 1). Each of the 10 scores assigned on the rubric as well as an overall average score was compared to the instructors given scores of the papers from the course. Instructors used their own methods for grading the papers that were unrelated to this developed rubric. Not all instructors were available to provide their past scores of the individual assignments. Results show that overall the rubric correlated with actual scores with a mean of r (17) =.56, p <.01. Individual items varied, but the Quality of Introduction item, r (17) =.74, p <.01, the Quality of Methods item, r (17) =.51, p <.05, the Quality of Discussion item, r (17) =.70, p <.01, and the Written Communication Skills item, r (17) =.71, p <.01, all yielded significant correlations. All other individual item correlations were not significant. We constructed this assessment rubric under the guidelines of previous papers on rubric development and standardization. Although it has only been tested on one particular sample, it appears to be an effective measure for assessing certain SLOs of the CSUN undergraduate Psychology Department. More specifically, the rubric can be used as a direct assessment of written communication skills, use of appropriate statistics, use of previous literature to support ideas and hypotheses, adherence to the APA Manual, and ability to discuss key concepts of research methodology. The strong interrater reliability suggests that the meaning of each item was adequately transmitted to the raters, although there was a great deal of difficulty in explaining the qualities that a strong manuscript s statements of hypotheses entail. This item s low correlations can be attributed to the relatively obscure nature of the item itself, as guidelines were not available in the APAManual. All other items have moderate to strong correlations between the two coders. We found that items that depended strongly on writing ability were most similar to the actual grade of the manuscript, and items that focused on statistics and research methodologies were dissimilar. It is possible that professors valued writing ability above all other SLOs when judging the manuscripts due to a halo effect. It is also possible that the rubric has difficulty in capturing the professors interpretations of the students research methodologies and knowledge of statistics. Limitations and Future Research One limitation of our study is a lack of intrarater reliability, as the raters only coded each manuscript once. Another limitation is the relatively small sample of papers analyzed for convergent validity, as some faculty members were unavailable to provide permission in accessing their actual final paper assigned scores. The possible halo effect of a strongly written paper might have interfered with the true quality of the manuscripts. It is difficult to ascertain whether a manuscript received a high score for demonstrating mastery of our SLOs, or merely for being well written. Conversely, it could be that instructors, raters, or both underestimated a poorly written paper and consequently assigned it lower scores in other content areas. This project provided an example of constructing an item-specific rubric with explicit guidelines on what is expected on each section of a psychology manuscript. Future studies should describe the process and challenges of other researchers in developing their own assessment rubrics for SLOs. Although we designed the rubric for use in one psychology department, we hope that these results will provide an addition to the relatively sparse information on the empirical development of reliable and valid rubrics designed to measure undergraduate psychology students academic performances. References American Psychological Association. (2001). Publication manual of the American Psychological Association (5th ed.). Washington, DC: Author. American Psychological Association Task Force on Psychology Major Competencies. (2007). APA guidelines for the undergraduate psychology major. Washington, DC: Author. Cohen, J. (1988). Statistical power analysis for the behavioral sciences (2nd ed.). Hillsdale, NJ: Lawrence Erlbaum Associates, Inc. Dunn, D. S., Mehrotra, C. M., & Halonen, J. S. (Eds.). (2004). Measuring up: Educational assessment challenges and practices for psychology. Washington, DC: American Psychological Association. Halonen, J. S., Bosack, T., Clay, S., & McCarthy, M. (with Dunn, D. S., Hill, G. W., IV, McEntarffer, R., Mehrotra, Vol. 36, No. 2, 2009 115

C., Nesmith, R., Weaver, K. A., & Whitlock, K.). (2003). A rubric for learning, teaching, and assessing scientific inquiry in psychology. Teaching of Psychology, 30, 196 208. McGovern, T. V., & Hawks, B. K. (1986). The varieties of undergraduate experience. Teaching of Psychology, 31, 174 181. Stellmack, M. A., Konheim-Kalkstein, Y. L., Manor, J. E., Massey, A. R., & Schmitz, J. P. (2009). An assessment of reliability and validity of a rubric for grading APA-style introductions. Teaching of Psychology. Notes 1. We thank Dr. Shannon Morgan for her detailed and helpful comments on an earlier draft of this article. We also thank the CSUN Psychology Department faculty for their participation, input, and support. 2. Send correspondence to Ellie Kazemi, Department of Psychology, California State University, Northridge, CA 91330 8255; e-mail: ellie.kazemi@csun.edu. 116 Teaching of Psychology