BIAS AND RANDOM ERROR IN CLASSROOM SGPS 1. Investigating the Amount of Systematic and Random. Error in Mean Classroom-Level SGPs. Joshua J.

Similar documents
Probability and Statistics Curriculum Pacing Guide

Multiple regression as a practical tool for teacher preparation program evaluation

STA 225: Introductory Statistics (CT)

On-the-Fly Customization of Automated Essay Scoring

learning collegiate assessment]

Algebra 1, Quarter 3, Unit 3.1. Line of Best Fit. Overview

Universityy. The content of

NCEO Technical Report 27

w o r k i n g p a p e r s

Working Paper: Do First Impressions Matter? Improvement in Early Career Teacher Effectiveness Allison Atteberry 1, Susanna Loeb 2, James Wyckoff 1

Introduction. Educational policymakers in most schools and districts face considerable pressure to

Linking the Ohio State Assessments to NWEA MAP Growth Tests *

Longitudinal Analysis of the Effectiveness of DCPS Teachers

A Guide to Adequate Yearly Progress Analyses in Nevada 2007 Nevada Department of Education

Hierarchical Linear Modeling with Maximum Likelihood, Restricted Maximum Likelihood, and Fully Bayesian Estimation

An Empirical Analysis of the Effects of Mexican American Studies Participation on Student Achievement within Tucson Unified School District

PROFESSIONAL TREATMENT OF TEACHERS AND STUDENT ACADEMIC ACHIEVEMENT. James B. Chapman. Dissertation submitted to the Faculty of the Virginia

Grade Dropping, Strategic Behavior, and Student Satisficing

Chapters 1-5 Cumulative Assessment AP Statistics November 2008 Gillespie, Block 4

The lab is designed to remind you how to work with scientific data (including dealing with uncertainty) and to review experimental design.

Cross-Year Stability in Measures of Teachers and Teaching. Heather C. Hill Mark Chin Harvard Graduate School of Education

Psychometric Research Brief Office of Shared Accountability

CHAPTER 4: REIMBURSEMENT STRATEGIES 24

Lecture 1: Machine Learning Basics

School Size and the Quality of Teaching and Learning

Statewide Framework Document for:

A Comparison of Charter Schools and Traditional Public Schools in Idaho

Technical Manual Supplement

Assignment 1: Predicting Amazon Review Ratings

Peer Influence on Academic Achievement: Mean, Variance, and Network Effects under School Choice

Proficiency Illusion

THE PENNSYLVANIA STATE UNIVERSITY SCHREYER HONORS COLLEGE DEPARTMENT OF MATHEMATICS ASSESSING THE EFFECTIVENESS OF MULTIPLE CHOICE MATH TESTS

College Pricing. Ben Johnson. April 30, Abstract. Colleges in the United States price discriminate based on student characteristics

Do First Impressions Matter? Predicting Early Career Teacher Effectiveness

Radius STEM Readiness TM

Gender and socioeconomic differences in science achievement in Australia: From SISS to TIMSS

Why Did My Detector Do That?!

A Model to Predict 24-Hour Urinary Creatinine Level Using Repeated Measurements

Teacher Supply and Demand in the State of Wyoming

Grade 6: Correlated to AGS Basic Math Skills

Math-U-See Correlation with the Common Core State Standards for Mathematical Content for Third Grade

TIMSS ADVANCED 2015 USER GUIDE FOR THE INTERNATIONAL DATABASE. Pierre Foy

Evidence for Reliability, Validity and Learning Effectiveness

Understanding and Interpreting the NRC s Data-Based Assessment of Research-Doctorate Programs in the United States (2010)

FOUR STARS OUT OF FOUR

Karla Brooks Baehr, Ed.D. Senior Advisor and Consultant The District Management Council

Mathematics subject curriculum

CS Machine Learning

CAAP. Content Analysis Report. Sample College. Institution Code: 9011 Institution Type: 4-Year Subgroup: none Test Date: Spring 2011

Hierarchical Linear Models I: Introduction ICPSR 2015

Math 96: Intermediate Algebra in Context

On the Distribution of Worker Productivity: The Case of Teacher Effectiveness and Student Achievement. Dan Goldhaber Richard Startz * August 2016

Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition

ACADEMIC AFFAIRS GUIDELINES

Comparing Teachers Adaptations of an Inquiry-Oriented Curriculum Unit with Student Learning. Jay Fogleman and Katherine L. McNeill

Python Machine Learning

Certified Six Sigma Professionals International Certification Courses in Six Sigma Green Belt

Norms How were TerraNova 3 norms derived? Does the norm sample reflect my diverse school population?

BENCHMARK TREND COMPARISON REPORT:

The elimination of social loafing behavior (i.e., the tendency for individuals

How to Judge the Quality of an Objective Classroom Test

AGS THE GREAT REVIEW GAME FOR PRE-ALGEBRA (CD) CORRELATED TO CALIFORNIA CONTENT STANDARDS

Redirected Inbound Call Sampling An Example of Fit for Purpose Non-probability Sample Design

Kansas Adequate Yearly Progress (AYP) Revised Guidance

PIRLS. International Achievement in the Processes of Reading Comprehension Results from PIRLS 2001 in 35 Countries

School Competition and Efficiency with Publicly Funded Catholic Schools David Card, Martin D. Dooley, and A. Abigail Payne

Effectiveness of McGraw-Hill s Treasures Reading Program in Grades 3 5. October 21, Research Conducted by Empirical Education Inc.

CONSTRUCTION OF AN ACHIEVEMENT TEST Introduction One of the important duties of a teacher is to observe the student in the classroom, laboratory and

Shelters Elementary School

Standards-based Mathematics Curricula and Middle-Grades Students Performance on Standardized Achievement Tests

Mathematics. Mathematics

Review of Student Assessment Data

VOL. 3, NO. 5, May 2012 ISSN Journal of Emerging Trends in Computing and Information Sciences CIS Journal. All rights reserved.

Role Models, the Formation of Beliefs, and Girls Math. Ability: Evidence from Random Assignment of Students. in Chinese Middle Schools

Instructor: Mario D. Garrett, Ph.D. Phone: Office: Hepner Hall (HH) 100

The Good Judgment Project: A large scale test of different methods of combining expert predictions

The Efficacy of PCI s Reading Program - Level One: A Report of a Randomized Experiment in Brevard Public Schools and Miami-Dade County Public Schools

ACTL5103 Stochastic Modelling For Actuaries. Course Outline Semester 2, 2014

Guide to the Uniform mark scale (UMS) Uniform marks in A-level and GCSE exams

OVERVIEW OF CURRICULUM-BASED MEASUREMENT AS A GENERAL OUTCOME MEASURE

Individual Differences & Item Effects: How to test them, & how to test them well

Evolutive Neural Net Fuzzy Filtering: Basic Description

Corpus Linguistics (L615)

GDP Falls as MBA Rises?

Massachusetts Department of Elementary and Secondary Education. Title I Comparability

Numeracy Medium term plan: Summer Term Level 2C/2B Year 2 Level 2A/3C

University of Groningen. Systemen, planning, netwerken Bosman, Aart

ASSESSMENT REPORT FOR GENERAL EDUCATION CATEGORY 1C: WRITING INTENSIVE

Probability Therefore (25) (1.33)

Detailed course syllabus

Statistical Analysis of Climate Change, Renewable Energies, and Sustainability An Independent Investigation for Introduction to Statistics

Montana Content Standards for Mathematics Grade 3. Montana Content Standards for Mathematical Practices and Mathematics Content Adopted November 2011

Analysis of Enzyme Kinetic Data

Principal vacancies and appointments

Edexcel GCSE. Statistics 1389 Paper 1H. June Mark Scheme. Statistics Edexcel GCSE

Linking the Common European Framework of Reference and the Michigan English Language Assessment Battery Technical Report

Measures of the Location of the Data

Running head: DELAY AND PROSPECTIVE MEMORY 1

Accountability in the Netherlands

Transcription:

BIAS AND RANDOM ERROR IN CLASSROOM SGPS 1 Running Head: BIAS AND RANDOM ERROR IN CLASSROOM SGPS Investigating the Amount of Systematic and Random Error in Mean Classroom-Level SGPs Joshua J. Marland Craig S. Wells Stephen G. Sireci Katherine Furgol Castellano

BIAS AND RANDOM ERROR IN CLASSROOM SGPS 2 Abstract Aggregate student growth percentiles (SGPs) are increasingly being used for educator and institutional accountability throughout the country, and research on their statistical properties is necessary to ensure appropriate use. In this study, true and observed SGPs were simulated, and the amount of systematic and random error was estimated to determine if aggregated SGPs can support their intended purposes across classrooms of differences sizes. Overall, the amount of systematic error was relatively small, while random error was substantially larger across all classroom sizes. Bias is more a function of true SGP, where random error is a function of both classroom size and true SGP. Classification was affected to a moderate extent across all classroom sizes because of the amount of error in the aggregate SGPs, but was most impacted for those in the top and bottom rating categories. Considerations should be made for random error when considering classifying teachers into rating categories, especially for those with small classrooms and who are close to the rating category cuts.

BIAS AND RANDOM ERROR IN CLASSROOM SGPS 3 Investigating the Amount of Systematic and Random Error in Classroom-Level SGPs Introduction Student growth percentiles (SGPs; Betebenner, 2009) were initially developed to provide students with a normative measure about their growth, but have since expanded into being used for evaluating teachers, schools, and school districts for several federal and state accountability initiatives. According to Collins and Amrein-Beardsley (2012), during the 2011/2012 school year, 13 states were using or piloting SGPs for the purpose of evaluating teachers. Soto (2013) stated SGPs are being used in 22 states for various purposes. Although SGPs are used by several states, the amount of random error present in studentlevel SGPs has been disconcerting (Sireci, Wells, & Bahry, 2013; Wells, Sireci, & Bahry, 2014). For example, Wells, Sireci, and Bahry (2014) examined the systematic and random error of student-level SGPs when conditioning on one, two, and three years of test data via a simulation study. They found that although SGPs exhibited small systematic error, the amount of random error was substantial (e.g., confidence intervals for students with SGPs around 50 ranged from 29 to 78), which calls into question their utility for interpreting students normative growth. SGPs are also aggregated across students, for example, within a classroom for the purpose of evaluating teacher effectiveness, or across all students in a school for institutional accountability. As of 2012, the weight aggregate SGPs carried in educator evaluations ranged from 20 to 50 percent (Hull, 2013), which makes it important that the aggregated SGPs exhibit reasonably small systematic and random error to support inferences drawn regarding educator or institutional effectiveness 1. Although the amount of random error is expected to be smaller when 1 It is important to note that having a small amount of random and systematic error is not a sufficient condition to support valid inferences regarding teacher effectiveness. Additional evidence would need to be gathered as part of the validity argument supporting the valid use of SGPs for evaluating teacher effectiveness.

BIAS AND RANDOM ERROR IN CLASSROOM SGPS 4 aggregating SGPs within a classroom, it is unclear if the amount of error is sufficiently small to support valid inferences regarding teacher effectiveness. Shang, VanIwaarden, and Betebenner (2015) found there to be greater random error in aggregate SGPs than bias, but both were nonnegligible with and without a measurement error correction. The authors reported a squared bias of 10.41, variance of 12.97 and mean square error of 23.38 when calculating SGPs without a measurement error correction, and aggregating using a mean approach. McCaffrey, Castellano and Lockwood (2015) found similar results to Shang, et al, with greater random error in aggregate SGPs than bias with no substantive changes to results across estimation methods. The purpose of the current study is to quantify the amount of random and systematic error exist in classroom SGPs, partition that error, and understand the implications of error on classifications of those being evaluated using aggregate SGPs. To accomplish these goals we simulate true SGPs at the student level and aggregate them at the classroom level across classes of different sizes. The details of our methodology are described next. Method A simulation study was conducted to examine the random and systematic error of classroom-level SGPs. True and observed scale scores were simulated to represent students test scores on a typical statewide assessment for grades 4 and 5. Observed scores were simulated using operational conditional standard errors of measurement for each scale score. Grade 5 true and observed SGPs were then calculated using the simulated data using grade 4 as the conditioning year. The data were simulated using a multilevel model to produce a nested structure that is observed in real data; that is, students were nested within classroom. Furthermore, to simulate realistic data, the parameters in the simulation were based on real test data. Bias, random error, and root mean square error (RMSE) were investigated across varying

BIAS AND RANDOM ERROR IN CLASSROOM SGPS 5 classroom sizes to better understand the extent to which error is a function of true SGPs and classroom size. One state s recommended operational classification scheme was used to determine rates of agreement between observed and true SGPs across 100 replications for every classroom. Data Generation Generating true scale scores and SGP values. To generate the simulees true scores, a two-level hierarchical linear model (HLM) with random-intercepts and slopes was used, where level 1 represents the student-level scores, level 2 represents the classroom effect. The studentlevel model (level 1) is presented in equation (1). Yij 0 j ij *Grade4 ij rij (1) Y ij represents student i s 5 th grade scale score in classroom j; 0 j and 1 j represent the intercept and slope for students in classroom j and when regressing 5 th grade scores on 4 th grade scores; Grade4ij represents the grade 4 scale score for student i in classroom j; and r ij represents the level-1 residual for students in classroom j. Along with values for student level residual, denoted 0 j and 1 j 2, will be used in the simulation to generate scores. The classroom-level model (level 2) is presented in equations (2) and (3). 0 j 00 0 j, the variance of the u (2) u (3) 1 j 10 1 j 00 represents the intercept for all students and u 0 j represents the random intercepts for students in classroom j. The variance of the intercepts, denoted, was used in the simulation. In 0 equation (3), 10 represents the slope for students at the mean of classroom j when regressing 5 th

BIAS AND RANDOM ERROR IN CLASSROOM SGPS 6 grade scores on 4 th grade scores. In addition, the correlation between the intercept and slopes, denoted,, 0 1, was used in the simulation. The full model is shown in equation (4). Y *Grade4 u u *Grade4 r (4) ij 00 10 0 j 1 j 1 j Using real data from a large-scale, statewide assessment, we fit the previously described two-level multilevel 2 model. Table 1 contains the parameter estimates that will be used to 2 generate true scale scores. However, the variance at level-1 ( ) was manipulated so that the correlation between the grade 4 and grade 5 scale scores is approximately 0.85, which equals the disattenuated correlation coefficient in the real data. Generating true scores with a nested structure is a two-step process. In step one, we sampled 5,000 intercepts and slopes for level 2 (i.e., classroom level) from a bivariate normal distribution with a mean vector defined by 00 and 10 and a covariance matrix defined by 0, 1, and, 0 1 given in equation (5).. For this study, the mean vector was 238 and 0.73 and the covariance matrix is 118.35 0.015 L2 (5) In step two, we first sampled 4 th grade scale scores from a normal distribution with a mean of 240 and standard deviation of 15 (to mimic the mean and standard deviation of the state data we are emulating). We used the 4 th grade scale score to predict the 5 th grade scale score for classroom j using a modification to equation (1), adjusting for the fact that the coefficient estimates are group-centered. We found a deviation for each student that represents the

BIAS AND RANDOM ERROR IN CLASSROOM SGPS 7 difference between their Grade 4 score and the intercept for their classroom ( ). We used that deviation score in equation (1) to calculate Grade 5 scores. 0k Deviation Grade4 (6) ij ij 0 j Y Deviation r (7) * ij 0j 1 j ij ij We then sampled n students for a classroom from a normal distribution with the mean equal to the predicted value for classroom j with a variance of 30. A variance of 30 was selected so that the correlation between the grade 4 and grade 5 scale scores was approximately 0.85, which equals the disattenuated correlation observed in the real data. Classroom sizes were sampled from a normal distribution with a mean of 20 and standard deviation of 5. The minimum n for a classroom was 10 with a maximum of 36. The total n of students in each data set was 100,158. True scale scores were rounded to the nearest integer and bounded between 200 and 280. Generating observed scale scores. Observed scale scores were sampled from a normal distribution with the mean equal to the simulee s true scale score and the standard deviation equal to the standard error of measurement conditioned on the true scale score (i.e., CSEM). The same CSEM was used for 4 th and 5 th grade, and was based on real test data (Figure 1). One hundred replications were conducted. The true and observed student-level SGPs were determined via quantile regression using the true and observed scale scores. The R package SGP (Betebenner, VanIwaarden, & Domingue, 2013) was used to implement quantile regression. The true and observed classroomlevel SGPs were based on the mean of the student-level SGPs within a classroom because it is used in practice. Data Analysis

BIAS AND RANDOM ERROR IN CLASSROOM SGPS 8 The relationship between the classroom-level true and observed SGPs was examined using the Spearman rho correlation coefficient. The amount of systematic error (bias) was examined by comparing the mean classroom-level SGP to the true classroom-level SGP as a function of the true SGP and classroom size. We also examined the amount of random error across replications by calculating the standard deviation of SGPs, using the difference between classroom-level observed SGPs and true SGPs for each replication. Lastly, we calculated the root mean square error for each classroom to determine the combined amount of random and systematic error. To determine the amount of systematic error (or bias) present in observed SGPs, we use the following calculation: Bias k100 SGP k k SGP True (8) Where we calculate the difference between mean observed SGPs and true SGPs for each of k replications, and then find the mean across all replications. To investigate the amount of error in our observed SGPs across replications, we use the following equation: SD SGP k100 ( SGP SGP) k n 1 2 (9) Where we calculate for each replication the difference between the observed SGP ( SGP ) and the average of SGPs across replications ( SGP ). This gives us the standard deviation of observed SGPs within a classroom. To calculate root mean square error, we use the following formula: 2 2 RMSE SD Bias (10)

BIAS AND RANDOM ERROR IN CLASSROOM SGPS 9 Where we find the square root of the sum of the squared standard deviation (random error) and bias (systematic error). Lastly, to investigate the practical implications of random and systematic error in aggregated SGPs, we classified teachers into four rating categories based on a method offered as guidance from a state education agency. The approach classifies teachers into rating categories based on numeric cuts along the classroom-level SGP scale. Table 2 contains the cut scores that define each of the performance categories. For the analysis, teachers were classified into the four rating categories based on their classroom-level true SGP, as well as for each of the 100 observed SGPs they received in the simulation. The number and proportion of classifications in agreement between true and observed SGPs across all replications was calculated, which results in a 5000 x 1 column vector (one proportion agreement value for each classroom). Results Bias The Spearman rho correlation between classroom-level true and observed SGPs was.917. The mean bias within classrooms across all replications is.04 SGPs. As can be seen in Figure 2, bias is really a function of true SGP rather than classroom size with small differences between class size categories. Classrooms with fewer than 19 students have the greatest bias, with an average bias of.11 for classrooms with between 10 and 14 students, and.07 for those with 15 19 students. For classrooms with 20 24 students, the average bias is.02, and.01 for classrooms with greater than 24 students (Table 2). Class size differences in bias exists mostly at the extremes. However, we can see that the mean SGPs for classrooms with lower true SGPs are

BIAS AND RANDOM ERROR IN CLASSROOM SGPS 10 over-predicted by about seven SGPs, where those with higher true SGPs are under-predicted by about the same amount. In Figure 3, bias is plotted as a function of true SGPs deciles and classroom size, so that variability in bias between and within each decile is more apparent. There appears to be a negative linear relationship between true SGP and bias across all four classroom size categories, with a decreasing amount of variability as class size gets larger. Random Error In Figure 4 (and Table 3), we can see that classrooms with the greatest number of students have the least amount of random error on average. The average standard deviation for classrooms with greater than 24 students is 6.8 SGPs, where it is 15.5 SGPs for classrooms with between 10 and 14 students. In practical terms, a teacher in a small classroom and an aggregate SGP of 50 could actually have an SGP between 20 and 80. In Figure 5, random error is plotted as a function of true SGP and classroom size. Again, classrooms with fewer students have the greatest amount of random error throughout the true SGP scale. Random error is somewhat smaller as SGPs move out toward the extremes on the scale. In Figure 6, random error is again plotted as a function of true SGP deciles and class size, with substantially more variability in the amount of random error within a decile across class sizes. For instance, random error ranges from about 10 20 SGPs for the lowest true SGP decile for teachers with between 10 14 students, where it ranges from about 4 8 SGPs for classrooms in the lowest true SGP decile for teachers with greater than 24 students. Root mean square error (RMSE) RMSE is greatest for smallest classrooms, with an average RMSE of 16 for classrooms with fewer than 15 students, where it is only 7.7 for classrooms with greater than 24 students. In

BIAS AND RANDOM ERROR IN CLASSROOM SGPS 11 Figure 6, the distribution of RMSE is almost bifurcated, where classrooms with true SGPs between 40 and 60 with more than 24 students have the lowest RMSE. Classrooms with SGPs toward the higher end and the fewest students have the highest RMSE. Classification Overall, mean agreement between true classroom-sgps and each observed SGP across replications is 72.4 percent. In Figure 9, that average agreement is higher for true SGPs that are further away from the SGP cuts set forth by the state, each of which is plotted as a reference line. The proportion agreement is close to 100 percent for true SGPs below 22, after which agreement starts to decline all the way down to 20 percent for true SGPs near 35. For true SGPs just above the 35 cut, agreement increases slightly and then begins to decline again as true SGPs approach 50. However, classrooms in the top and bottom rating categories had the lowest average agreement across replications, compared with the two middle categories. Those with true SGPs in the top rating category had an average agreement across all replications of 58.1 percent, where those in the bottom rating category had an agreement of 63.4 percent (Table 6). This results is compared with average agreement of 75.5 and 79.5 percent for the second and third categories, respectively. Similar patterns held when results are broken down by classroom size category, with greatest average agreement for the middle two rating categories. The lowest average agreement was for those in the highest true rating category with more than 24 students (55.1 percent), while the highest average agreement was for the same classroom size category, but for the third true SGP rating category (83.1 percent). To further understand the magnitude of classification changes, we also calculated the number of categories classrooms change across all replications. As mentioned, 72.4 percent of teachers do not change categories at all, while 27.5 percent move up or down one category, and

BIAS AND RANDOM ERROR IN CLASSROOM SGPS 12 just.1 percent move up or down two categories (Table 5). Classrooms with fewer students have the lowest classification agreement across replications at 68.3 percent for classrooms with 10 14 students, compared with 74.3 percent agreement across replications for classrooms with greater than 24 students. Discussion SGPs are widely used to evaluate teachers, but there has been little study of their appropriateness for this purpose. In this study, we investigated the amounts of systematic and random error in mean SGP scores at the classroom level across classes of various sizes. We generated 4 th and 5 th grade true and observed Math scale scores using parameter estimates from a two-level random intercepts and slopes model. One hundred observed score data sets and one true score data set were created with five thousand classrooms with a total of 100,000 students in each. Student-level true and observed SGPs were calculated using each of the data sets, and then aggregated to the classroom level to investigate the amount of random and systematic error present in mean SGPs. As a final step, classrooms were classified into one of four rating categories to determine the extent to which error impacts stability of ratings across replications. The results suggest random error in aggregate SGPs poses more of a threat to the effective use of the measures in educator evaluation than systematic error does. In this study, however, systematic error was greatest for classrooms with really low or high true SGPs, with bias averaging to about seven SGPs for those classrooms near 20 or 80 on the true SGP scale. Across all classrooms, bias averaged to just.04, where classrooms with the fewest students had slightly higher bias with an average of.11 SGPs compared with about.01 for classrooms with greater than 24 students. Random error was about 10 SGPs across all classrooms, which means that a classroom with an average SGP of 50 could be as low as about 30 and as high as 70. The

BIAS AND RANDOM ERROR IN CLASSROOM SGPS 13 range grows large for small classrooms, where those with between 10 and 14 students at 50 could have an average SGP between 20 and 80. The practical implications of this level of error means many educators may be misclassified, which can potentially impact their overall evaluation. In this study, 27.4 percent of educators were misclassified by one category, meaning they could have been one category higher or lower on their aggregate SGP measure. A misclassification rate this high argues against using aggregated SGPs for high-stakes purposes such as rewarding or sanctioning teachers. A negligible proportion of classrooms moved more than one category either way (~.1 percent). The variability in SGPs could be mitigated somewhat through the use of confidence intervals, where educators are classified into a no-stakes category when there is too much uncertainty to make a meaningful decision. Policymakers advocating for the use of SGPs should also consider the extent to which misclassification is a function of where the educator is along the SGP scale. In this study, those in the top and bottom rating category had lower average agreement across replications, where those in the middle two had higher average agreement. This could be due to the RMSE being greater toward the ends of the true SGP scale. Limitations This study utilized simulated data, albeit simulated based on empirical data from one state. There were aspects to the data generation process that could have more closely mirrored the realities of operational data collected from a state. Fourth grade true scale scores were generated using a random normal distribution with a mean of 240 and standard deviation of 15, rather than with a nested structure like 5 th grade true scale scores. The number of students in each classroom, or attributable to an educator, was also not perfectly realistic, in that some educators may be responsible for more than 36 students. This is

BIAS AND RANDOM ERROR IN CLASSROOM SGPS 14 often the case in realistic settings. In addition, the minimum number of students in a classroom was ten, so the results are really only generalizable to those classrooms. There are teachers in 8 students, 1 teacher, 1 paraprofessional settings, for which states may choose to provide an aggregate SGP (if combined with ELA scores, for instance). Lastly, classification results were based on one state s approach that used four specific classification categories. Given that states are taking many different approaches to using SGPs in evaluation systems, classification consistency will be affected by the number of categories and the cut points that define them. However, this study does closely mirror the practical realities that exist in many states choosing to use aggregate SGPs for evaluating educators. Summary More research is needed on the reliability and validity of SGPs. The present study represents one investigation of the reliability of aggregated SGPs and is modeled after the way many states are using them to evaluate teachers. Previous research has suggested the amount of error in student-level SGPs is substantial and expressed concern about reporting and interpreting them at the student level (e.g., Wells et al., 2014). The present study raises similar concerns for aggregated SGPs, which are being used for higher-stake decisions. Our results suggest more research is needed on SGPs, with particular attention paid to the amount of random error and its implications for classification decisions.

BIAS AND RANDOM ERROR IN CLASSROOM SGPS 15 References Betebenner, D. (2009). Norm- and criterion-referenced growth. Educational Measurement Issues and Practice, 28(4) 42-51. Collins, C. & Amrein-Beardsley, A. (2012, April). Putting growth and value-added on the map: A national overview. Paper presented at the 2012 annual meeting of the American Educational Research Association, Vancouver, British Columbia, Canada. McCaffrey, D., Castellano, K., & Lockwood, J.R. (2015). The impact of measurement error on the accuracy of individual and aggregate SGP. Educational Measurement Issues and Practice, 34(1) 15-21. Shang, Y., VanIwaarden, A., Betebenner, D. (2015). Covariate measurement error correction for student growth percentiles using the SIMEX method. Educational Measurement Issues and Practice, 34(1) 4-14. Sireci, S.G., Wells, C.S., Bahry, L. (2013, April). Student growth percentiles: More noise than signal? Paper presented at the 2013 annual meeting of the American Educational Research Association, San Francisco, CA. Wells, C.S., Sireci, S.G., Bahry, L. (2014, April). The effect of conditioning years on the reliability of SGPs. Paper presented at the 2014 annual meeting of the National Council of Measurement in Education, Philadelphia, PA.

BIAS AND RANDOM ERROR IN CLASSROOM SGPS 16 Appendix Tables & Figures Table 1 Two-Level Model, 4 th grade Math score predicting 5 th grade Math score Estimate 00 238.02 10 0.729 0 118.41 1.349 0, 1.262 2 51.14 Figure 1 CSEM by scale score (same for both Grade 4 and 5)

BIAS AND RANDOM ERROR IN CLASSROOM SGPS 17 Figure 2 Systematic error (bias) as a function of true SGP and classroom size Figure 3 Systematic error as a function of true SGP decile and classroom size

BIAS AND RANDOM ERROR IN CLASSROOM SGPS 18 Figure 4 Random Error as a Function of Classroom Size Figure 5 Random Error as a Function of Average True SGP and Classroom Size (mspline smoothing)

BIAS AND RANDOM ERROR IN CLASSROOM SGPS 19 Figure 6 Random Error as a function of true SGP decile and classroom size Figure 7 Root mean square error as a function of true SGP and classroom size (mspline smoothing)

BIAS AND RANDOM ERROR IN CLASSROOM SGPS 20 Figure 8 Root mean square error as a function of true SGP decile and classroom size Table 3 Random Error, Bias and RMSE as a function of classroom size Random Error (SD) Bias RMSE 10-14 students 15.5 0.11 16.0 15-19 11.0 0.07 11.6 20-24 8.6 0.02 9.3 > 24 students 6.9-0.01 7.7 Total 10.0 0.04 10.7 Table 4 Classification Approach Effectiveness Category SGP Cut Score Bottom Rating Category 1 34 Category 2 35-49 Category 3 50 64 Top Rating Category 65-99

BIAS AND RANDOM ERROR IN CLASSROOM SGPS 21 Table 5: Classification changes as a function of classroom size Category No Change +/- One Category +/- Two Categories 10-14 Students 68.0 31.7 0.3 15-19 71.6 28.3 0.1 20-24 73.9 26.1 0.0 > 24 74.4 25.6 0.0 All Classrooms 72.4 27.5 0.1 Table 6: Average agreement across replications by true SGP rating category and classroom size Bottom Category Category 2 Category 3 Top Category 10-14 Students 62.7 70.6 72.3 60.6 15-19 63.2 74.2 78.1 58.7 20-24 64.3 77.5 81.0 57.5 > 24 62.7 78.3 83.1 55.9 All Classrooms 63.4 75.5 79.5 58.1

BIAS AND RANDOM ERROR IN CLASSROOM SGPS 22 Figure 9 Proportion agreement as a function of true SGP and classroom size Figure 10 Proportion agreement as a function of true SGP and classroom size (mspline smoothing)