An Investigation of Gender, Income, and Special Education Status Bias on Curriculum-Based Measurement Slope in Reading

Similar documents
OVERVIEW OF CURRICULUM-BASED MEASUREMENT AS A GENERAL OUTCOME MEASURE

Using CBM to Help Canadian Elementary Teachers Write Effective IEP Goals

The My Class Activities Instrument as Used in Saturday Enrichment Program Evaluation

QUESTIONS ABOUT ACCESSING THE HANDOUTS AND THE POWERPOINT

Using CBM for Progress Monitoring in Reading. Lynn S. Fuchs and Douglas Fuchs

Progress Monitoring & Response to Intervention in an Outcome Driven Model

ISD 2184, Luverne Public Schools. xcvbnmqwertyuiopasdfghjklzxcv. Local Literacy Plan bnmqwertyuiopasdfghjklzxcvbn

Psychometric Research Brief Office of Shared Accountability

An Empirical Analysis of the Effects of Mexican American Studies Participation on Student Achievement within Tucson Unified School District

The Oregon Literacy Framework of September 2009 as it Applies to grades K-3

Effectiveness of McGraw-Hill s Treasures Reading Program in Grades 3 5. October 21, Research Conducted by Empirical Education Inc.

NCEO Technical Report 27

Peer Influence on Academic Achievement: Mean, Variance, and Network Effects under School Choice

Conceptual and Procedural Knowledge of a Mathematics Problem: Their Measurement and Their Causal Interrelations

Recent advances in research and. Formulating Secondary-Level Reading Interventions

Shelters Elementary School

Evaluation of Teach For America:

Sector Differences in Student Learning: Differences in Achievement Gains Across School Years and During the Summer

Practices Worthy of Attention Step Up to High School Chicago Public Schools Chicago, Illinois

School Competition and Efficiency with Publicly Funded Catholic Schools David Card, Martin D. Dooley, and A. Abigail Payne

Review of Student Assessment Data

Interdisciplinary Journal of Problem-Based Learning

Instructional Intervention/Progress Monitoring (IIPM) Model Pre/Referral Process. and. Special Education Comprehensive Evaluation.

Cooper Upper Elementary School

Technical Report #1. Summary of Decision Rules for Intensive, Strategic, and Benchmark Instructional

EFFECTS OF MATHEMATICS ACCELERATION ON ACHIEVEMENT, PERCEPTION, AND BEHAVIOR IN LOW- PERFORMING SECONDARY STUDENTS

Multiple regression as a practical tool for teacher preparation program evaluation

A Guide to Adequate Yearly Progress Analyses in Nevada 2007 Nevada Department of Education

Wonderworks Tier 2 Resources Third Grade 12/03/13

teacher, peer, or school) on each page, and a package of stickers on which

Iowa School District Profiles. Le Mars

On-the-Fly Customization of Automated Essay Scoring

Chapters 1-5 Cumulative Assessment AP Statistics November 2008 Gillespie, Block 4

Predicting the Performance and Success of Construction Management Graduate Students using GRE Scores

Cooper Upper Elementary School

The Effects of Super Speed 100 on Reading Fluency. Jennifer Thorne. University of New England

Gender and socioeconomic differences in science achievement in Australia: From SISS to TIMSS

Getting Results Continuous Improvement Plan

Essentials of Ability Testing. Joni Lakin Assistant Professor Educational Foundations, Leadership, and Technology

PROGRESS MONITORING FOR STUDENTS WITH DISABILITIES Participant Materials

ACADEMIC AFFAIRS GUIDELINES

Travis Park, Assoc Prof, Cornell University Donna Pearson, Assoc Prof, University of Louisville. NACTEI National Conference Portland, OR May 16, 2012

Orleans Central Supervisory Union

Legacy of NAACP Salary equalization suits.

Process Evaluations for a Multisite Nutrition Education Program

Port Jefferson Union Free School District. Response to Intervention (RtI) and Academic Intervention Services (AIS) PLAN

RtI: Changing the Role of the IAT

Confirmatory Factor Structure of the Kaufman Assessment Battery for Children Second Edition: Consistency With Cattell-Horn-Carroll Theory

The Relationship Between Tuition and Enrollment in WELS Lutheran Elementary Schools. Jason T. Gibson. Thesis

Standards-based Mathematics Curricula and Middle-Grades Students Performance on Standardized Achievement Tests

Data-Based Decision Making: Academic and Behavioral Applications

Student-led IEPs 1. Student-led IEPs. Student-led IEPs. Greg Schaitel. Instructor Troy Ellis. April 16, 2009

Tools and. Response to Intervention RTI: Monitoring Student Progress Identifying and Using Screeners,

Hierarchical Linear Modeling with Maximum Likelihood, Restricted Maximum Likelihood, and Fully Bayesian Estimation

Probability and Statistics Curriculum Pacing Guide

New Ways of Connecting Reading and Writing

ROA Technical Report. Jaap Dronkers ROA-TR-2014/1. Research Centre for Education and the Labour Market ROA

George Mason University Graduate School of Education Program: Special Education

Linking the Ohio State Assessments to NWEA MAP Growth Tests *

Hierarchical Linear Models I: Introduction ICPSR 2015

Volume 19 Number 2 THE JOURNAL OF AT-RISK ISSUES JARI NATIONAL DROPOUT PREVENTION CENTER/NETWORK

Data Diskette & CD ROM

Miami-Dade County Public Schools

Learning By Asking: How Children Ask Questions To Achieve Efficient Search

Comparing Teachers Adaptations of an Inquiry-Oriented Curriculum Unit with Student Learning. Jay Fogleman and Katherine L. McNeill

Kahului Elementary School

Running head: DELAY AND PROSPECTIVE MEMORY 1

Pyramid. of Interventions

VIEW: An Assessment of Problem Solving Style

Early Warning System Implementation Guide

PSIWORLD Keywords: self-directed learning; personality traits; academic achievement; learning strategies; learning activties.

Teacher intelligence: What is it and why do we care?

IS FINANCIAL LITERACY IMPROVED BY PARTICIPATING IN A STOCK MARKET GAME?

PROFESSIONAL TREATMENT OF TEACHERS AND STUDENT ACADEMIC ACHIEVEMENT. James B. Chapman. Dissertation submitted to the Faculty of the Virginia

Wisconsin 4 th Grade Reading Results on the 2015 National Assessment of Educational Progress (NAEP)

A Pilot Study on Pearson s Interactive Science 2011 Program

Race, Class, and the Selective College Experience

The Effect of Extensive Reading on Developing the Grammatical. Accuracy of the EFL Freshmen at Al Al-Bayt University

Aimsweb Fluency Norms Chart

Social Emotional Learning in High School: How Three Urban High Schools Engage, Educate, and Empower Youth

5 Programmatic. The second component area of the equity audit is programmatic. Equity

Running Head GAPSS PART A 1

Algebra 1, Quarter 3, Unit 3.1. Line of Best Fit. Overview

THEORY OF PLANNED BEHAVIOR MODEL IN ELECTRONIC LEARNING: A PILOT STUDY

Mathematical learning difficulties Long introduction Part II: Assessment and Interventions

TIMSS ADVANCED 2015 USER GUIDE FOR THE INTERNATIONAL DATABASE. Pierre Foy

THE EFFECTS OF TEACHING THE 7 KEYS OF COMPREHENSION ON COMPREHENSION DEBRA HENGGELER. Submitted to. The Educational Leadership Faculty

PSYC 620, Section 001: Traineeship in School Psychology Fall 2016

Evaluation of the. for Structured Language Training: A Multisensory Language Program for Delayed Readers

Sheila M. Smith is Assistant Professor, Department of Business Information Technology, College of Business, Ball State University, Muncie, Indiana.

learning collegiate assessment]

Understanding Games for Teaching Reflections on Empirical Approaches in Team Sports Research

Newburgh Enlarged City School District Academic. Academic Intervention Services Plan

The Relationship of Grade Span in 9 th Grade to Math Achievement in High School

Kansas Adequate Yearly Progress (AYP) Revised Guidance

Implementing Response to Intervention (RTI) National Center on Response to Intervention

Gridlocked: The impact of adapting survey grids for smartphones. Ashley Richards 1, Rebecca Powell 1, Joe Murphy 1, Shengchao Yu 2, Mai Nguyen 1

Unraveling symbolic number processing and the implications for its association with mathematics. Delphine Sasanguie

Deborah Simmons Texas A&M University, College Station, Texas, USA

Transcription:

School Psychology Quarterly 2011 American Psychological Association 2011, Vol. 26, No. 2, 119 130 1045-3830/11/$12.00 DOI: 10.1037/a0023021 An Investigation of Gender, Income, and Special Education Status Bias on Curriculum-Based Measurement Slope in Reading Seungsoo Yeo Inje University Jamie Fearrington Appalachian State University Theodore J. Christ University of Minnesota This study investigated slope bias on student background variables for both Curriculum Based Measurement of Oral Reading (CBM-R) and Curriculum Based Measurement Maze Reading (Maze). Benchmark scores from 1,738 students in Grades 3 through 8 were used to examine potential slope bias in CBM-R and Maze. Latent growth modeling was used to both estimate growth rates and examine the extent to which demographic variables affected the estimated growth rates. Results indicate a significant CBM-R slope bias on special education status at Grade 3 and on gender at Grade 7. For Maze, slope bias on gender was associated with Maze slope estimates at Grades 5 and 7. Slope bias on various demographic variables was not consistent across CBM measures and grades. Results and implications are discussed. Keywords: curriculum based measurement, oral reading, maze, slope bias This article was published Online First April 18, 2011. Seungsoo Yeo, Department of Special Education, Inje University; Jamie Fearrington, Department of Psychology, Appalachian State University; and Theodore J. Christ, Department of Educational Psychology, University of Minnesota. Correspondence concerning this article should be addressed to Seungsoo Yeo, Department of Special Education, Inje University, 607 Obang-dong, Kimhae, Kyungnam 621-749, South Korea. E-mail: yeoxx008@inje.ac.kr Curriculum Based Measurement (CBM) is a set of standardized procedures that were initially designed to index the level and rate of student achievement within the basic skill areas of reading, mathematics, written expression, and spelling (Deno, 1985, Deno, 2003). CBM has now expanded to include measures of early literacy (Good & Kaminski, 2002) and early numeracy (Clarke & Shinn, 2004). Extensive research and substantial evidence provide support for the reliability and validity of CBM for a variety of applications, which include screening, benchmarking, and progress monitoring (Jiban & Deno, 2007; Fuchs, 2004; Marston, 1989; Wayman, Wallace, Wiley, Ticha, & Espin, 2007; Yeo, 2010). One distinctive feature of CBM that distinguishes it from other measures of academic achievement is its utility. That is, CBM is quick, simple, and repeatable. These characteristics allow teachers to measure student skills on a frequent basis, which might be weekly, biweekly, monthly, or triannually. Frequent and repeated administrations over time are useful to estimate the rate of student achievement over relatively brief periods of instruction or intervention. In the area of reading, two types of CBM measures have been used in research and practice: reading aloud and maze selection. On the CBM-R (reading aloud) measure, student performance is measured by requiring students to read aloud passages of meaningful text for 1 min. The number of words read correctly is scored as the reading rate (Deno, 1985). On CBM-Maze tasks, student performance is measured by having students read silently from a passage where every seventh word has been deleted and replaced with three response choices inside parentheses: one is the correct word and the other two are distracters. The number of correct selections is scored. Recently, CBM-Maze has been receiving more attention due to the fact that it can be administered to a group of students at one time, whereas CBM-R is individually administered. 119

120 YEO, FEARRINGTON, AND CHRIST Importance and Relevance of CBM in Relation to Response to Intervention Recent educational reform policies have placed an increased importance on accountability and standards (No Child Left Behind Act, 2002), as well as use of response to intervention (RTI) as a means to identify children with specific learning disabilities (SLDs). The recently reauthorized Individuals with Disabilities Education Improvement Act (2004) stipulates that schools may choose to use RTI to guide special education eligibility decisions. This shift in federal policy highlights the need for assessments that are useful to index both the level and rate of student achievement through a continuous process of screening/benchmarking and progress monitoring. In addition, RTI requires that educators track student progress over time as they are exposed to varying degrees of instructional intervention (Batsche et al., 2005; Vaughn & Fuchs, 2003). In most RTI models, the majority of students receive instruction in Tier 1, which consists of classroom instruction within the general education curriculum. Tier-2 instruction is delivered to students who are identified as at-risk through universal screenings of academic skills with CBM. Tier 2 relies on small-group instruction with scientifically based programs coupled with frequent progress monitoring. Students that do not respond to interventions delivered at Tier 2 may be moved to Tier 3, which is the most intensive and typically consists of individual instruction. CBM is the assessment tool of choice for many RTI models, and many school districts use CBM as part of their RTI model. CBM is used as a universal screening device to identify those children who need more intensive instruction (Tier 1), to evaluate instructional effects (Tiers 2 and 3), and to make special education eligibility decisions, which include the diagnosis of SLD. There are various commercial forms of CBM that are available for purchase by school districts who are implementing RTI. AIMSweb is one such commercial benchmarking- and progress-monitoring system (retrieved from www.aimsweb.com). AIMSweb provides its users with CBM measures in the academic areas of early literacy, early numeracy, reading, reading comprehension, math computation, math reasoning, and writing. AIMSweb users also have access to web-based data management that enables educators to organize and report outcomes for students at all three RTI tiers. Slope on CBM Measures Within a RTI Model Within an RTI framework, estimates of slope from frequent assessment are more crucial than estimates of level because slope is a key element for examining whether students are properly responding to a particular set of instructional conditions or interventions. Moreover, CBM is described in the literature as uniquely suited for repeated, ongoing assessment, which functions to establish time-series progressmonitoring data that are useful to evaluate growth (Deno, 1985, 1986, 2003). For example, progress monitoring at Tier 1 estimates the growth rate for all students in the general education classroom. If a student s slope or rate of growth is sufficiently below that of peers in Tier 1, the student moves to Tier 2 for an intervention that is more intensive. Those students who respond to Tier 2 intervention are soon served again within Tier 1, but those students who do not respond are later served within Tier 3 (Vaughn & Fuchs, 2003). The slope estimates derived from CBM data provide the evidence of sufficient or insufficient response that determines continuation or discontinuation of services at Tiers 1, 2, or 3. Slope estimates within and across tiers also provide the evidence to support special education eligibility decisions and SLD identification. As a result, the trustworthiness of the growth data used in a RTI decision-making framework is an important consideration. Group Differences on CBM Slope Given the importance of assessing growth rate within a RTI model, it is important to examine whether CBM slopes are consistent among a variety of groups, including gender, ethnicity, special education status, and socioeconomic status (SES). In other words, estimates of growth derived from CBM data might systematically increase or decrease as a function of group membership or demographic characteristics (Betts et al., 2008). Systematic over- or

AN INVESTIGATION OF BIAS AND CBM SLOPE 121 underestimation might function as a source of bias and thereby bias school-based decision making. This is an important concern and a subject worthy of examination. Until recently, many of the early CBM studies examined the extent to which scores at a single time point equally predict concurrent or future performance on criterion measures, which include statewide tests and assessment batteries, across groups and demographic characteristics. For example, a study by Hintze, Callahan, Matthews, Williams, and Tobin (2002) investigated ethnic bias on CBM-R across African American and Caucasian students by controlling for age, gender, and SES. This study found no predictive bias in the use of CBM-R to predict reading comprehension scores on the Woodcock-Johnson Psychoeducational Battery, Revised, meaning that the CBM score is not a biased predictor of future reading performance for either group of students. In fact, conducting research on predictive bias for CBM measures is meaningful to provide evidence that CBM can be broadly used for various groups. The RTI context establishes that it is important to examine the extent to which the estimates of growth based on CBM data are consistent and nonbiased across demographics and group memberships; however, relatively few studies have examined the issue. A review of the research yielded only one study that examined potential slope bias in CBM-R and Maze. Using hierarchical linear modeling (HLM), Shin, Deno, and Espin (2000) evaluated the technical features of the CBM maze task, including reliability, sensitivity, and validity, for measuring growth rates. In this study, group differences in slopes were examined between general and special education students. Results indicated that there was no significant difference in slope between the two groups; however, that study did not include other characteristics, such as gender, SES, and ethnicity. In addition, the analysis in that study did not control for the initial status (i.e., achievement level). Given the possibility that slope estimates might be either positively or negatively related to initial status, it seems necessary to control the initial level of performance. With regard to research examining CBM-R slope bias on group differences, recently, a longitudinal study by Chard et al. (2008) examined the effect of demographic variables, such as ethnicity, gender, and language, on a standardized achievement test in reading. Participants were 668 kindergarten and 1st graders who were identified as in need of intensive intervention in reading through Grade 3. Results revealed that none of the demographic variables included in the analyses were significant predictors of CBM slope. Although CBM growth rate in this study was evaluated in the context of initial status, this study did not examine whether or not demographic variables continued to be insignificant predictors across increasing grade levels. Given that CBM measures are often viewed as a seamless system of measuring students growth rate (Wayman et al., 2007), it would be necessary to investigate whether these demographic variables were significant predictors of CBM growth rate for students from elementary to middle school levels. Aims of the Current Study There is a paucity of research and evidence to support conclusions about potential slope bias in CBM-R and Maze; moreover, there are limitations inherent in the analyses of the research published in the literature. The first aim of this study was to examine whether estimates of slope from CBM data are likely to be an unbiased index of growth across groups and demographic characteristics such as gender, ethnicity, SES, and special education status. As an improvement upon prior analyses, the analytic design for this study controlled for the initial status of achievement by using latent growth modeling (LGM). The second aim of this study was to examine whether estimates of slope from CBM data are differentially biased across grades. The third, and final, aim of this study was to examine whether slope bias on various subgroups is consistent between CBM-R and Maze. Although both CBM-R and Maze are common progress-monitoring measures in reading, research has yet to examine whether slope bias on the demographic variables changes as a function of the differences between CBM-R and CBM-Maze. Therefore, it is important to investigate whether slope bias on the various variables is consistent between the two measures.

122 YEO, FEARRINGTON, AND CHRIST Method Participants and Setting The study took place in a rural school district located in the Southeastern United States. Students (N 1,738) in Grades 3 through 8 attending two elementary schools, and three middle schools participated in this study. The sample included 850 females (49.1%) and 880 males (50.9%) enrolled in both general and special education. The percentage of students that received special education services was 14.6%. The sample was predominantly Caucasian (94.5%). The ethnic breakdown for the remainder of the sample was 3.2% Hispanic, 1.4% African American, 0.5% Asian, and 0.2% Native American. Further, 1.33% of the students were classified as English as Second Language (ESL). Approximately half of the students (55.4%) were eligible to receive free or reduced-price lunches. Procedures All assessments used in this study were part of systematic universal screenings that were conducted three times a year by the school district. CBM-R and Maze data that were collected during the 2006 to 2007 school year were used in this study. CBM-R and Maze probes from AIMSweb were administered by a benchmark team that consisted of school psychologists, teaching assistants, and graduate students from a nearby school psychology doctoral training program. All persons on the benchmark team attended training sessions prior to testing. Assessments were conducted in a quiet area of the classroom or in the hallway outside the classroom. Each CBM-R benchmark score represented the median score from a series of three grade-appropriate passages. Selecting the median score from three probes is the commonly used method in practice and research on CBM (Yeo, 2008). The CBM-R passages were administered on an individual basis. Maze passages were administered to the entire classroom simultaneously and took approximately 5 min to administer. Each Maze passage was scored by the administration team directly after administration. Both scores were then entered into a database. Measurements CBM-R scores reflect the number of words correctly read out loud in 1 min (Shinn & Shinn, 2002). Over two decades of research have established the technical adequacy of CBM-R (Wayman et al., 2007). CBM-R is typically considered to be the AIMSweb reading measure with the most psychometric support (see Marston, 1989; Shinn, Good, Knutson, Tilly, & Collins, 1992). Test-retest and interrater reliability coefficients range from.82 to.99, respectively (Marston, 1989). Concurrent validity with other standardized reading achievement tests is also well established, with coefficients ranging from.58 to.86 (Jenkins & Jewell, 1993). To specifically measure reading comprehension, AIMSweb also provides Maze-reading probes. Maze requires students to complete a multiple-choice cloze task while they read a 150- to 400-word passage silently. The first sentence of each passage is complete. After, every seventh word is replaced by three plausible selections, and the student must circle the word that correctly completes the sentence. The Maze score represents the number of words circled correctly during a 3-min time period (Shinn & Shinn, 2002). Maze also has satisfactory evidence of technical adequacy. Criterionrelated validity with CBM-R is strong, with coefficients ranging from.77 to.86 (Espin, Deno, Maruyama, & Cohen, 1989). The concurrent validity of Maze is also well established with group-administered standardized tests of reading achievement (Jenkins & Jewell, 1993). Results Descriptive statistics for all study variables are presented in Table 1. Data are depicted as a proportion for all demographic characteristics and as a mean of CBM scores for each point in time. It seems that though the mean level of CBM-R scores increased over time for all grade levels, the mean level of Maze scores increased for only Grades 3, 5, and 7. Most students in all grade levels belonged to the Caucasian ethnic group. For this reason, the ethnic group was not used as a potential covariate in this study. Before proceeding with the substantive analyses, all study variables were screened for the potential problems of skewness and

AN INVESTIGATION OF BIAS AND CBM SLOPE 123 Table 1 Descriptive Statistics for All Variables Variables Grade 3rd 4th 5th 6th 7th 8th Mean or proportion (SD) Mean or proportion (SD) Mean or proportion (SD) Mean or proportion (SD) Mean or proportion (SD) Mean or proportion (SD) CBM-R a Fall 68.80 (35.51) 91.50 (34.43) 105.28 (45.99) 120.66 (45.57) 128.53 (42.44) 126.72 (42.13) Winter 92.67 (40.54) 111.88 (36.01) 117.31 (48.31) 130.87 (43.51) 137.42 (42.62) 141.60 (40.69) Spring 106.37 (41.50) 123.94 (38.68) 130.32 (47.07) 142.49 (43.62) 150.71 (40.44) 149.56 (40.51) Maze b Fall 10.30 (5.67) 11.12 (4.83) 14.92 (7.05) 17.69 (8.13) 19.54 (7.85) 20.92 (7.13) Winter 12.18 (5.42) 16.80 (5.92) 17.77 (8.37) 21.82 (8.38) 21.60 (9.75) 18.39 (7.13) Spring 12.34 (6.01) 15.22 (6.24) 20.93 (9.23) 21.42 (9.52) 24.37 (8.97) 23.36 (9.14) Caucasian.95 (.23).95 (.21).94 (.23).93 (.26).94 (.23).95 (.21) Female.48 (.50).51 (.50).50 (.50).49 (.50).48 (.50).48 (.50) Free lunch.43 (.50).46 (.50).41 (.49).44 (.50).44 (.50).49 (.50) Special education.13 (.34).13 (.34).13 (34).19 (.39).12 (.23).15 (.36) Note. Missing data for each grade ranged from 10.2% to 13.2%. a Results reported as words read correctly per minute. b Results reported as words selected correctly per 3 min. kurtosis. The skewness and kurtosis coefficients of CBM measures varied (.19 for CBM-R scores to 1.29 for Maze scores), indicating that all CBM scores approximated a normal distribution. Testing Shape of Growth In the first step, latent growth modeling (LGM) techniques were conducted with the AMOS 4.0 program, using full information maximum likelihood (FIML), which provides efficient estimates from data with missing values (Ming, Conger, & Lorenz, 2005). An alpha level of.05 was set for all statistical tests. To determine the developmental growth trajectories, the following two models were initially fit to CBM measures. The first model, a no growth model, provided estimates that included initial status but not growth rate. Second, a linear growth model with both initial status and linear growth trajectory was fitted for change in CBM measures over time. The loadings for the slope factors were specified as 0, 5, and 9 (August 0, January 5, and May 9) over three time points. This set of factor loading allows for representing a liner growth rate. Especially, the first factor loading on the slope factor (August) was coded as zero so that the intercept is directly interpreted as the starting point of the growth trajectory (Bollen & Curran, 2006). Although it would be appropriate to model quadratic or cubic parameters before selecting a linear model, with only three data points per grade, nonlinearity of slopes could not be evaluated adequately as a potential growth model (Fitzmaurice, Laird, & Ware, 2004). As shown in Tables 2 and 3, the chi-square difference tests ( 2 ) show that the linear growth model fit the data best for CBM-R and Maze scores. In addition, the linear growth models were evaluated by using RMSEA (root mean square error of approximation; good model.08) and CFI (comparative fit index; good model.95; Hu & Bentler, 1999). Our findings indicated that, overall, the linear growth models are considered acceptable for the data from the CBM-R and Maze scores for all grade levels. Testing Variance in Slope and Intercept Tables 2 and 3 also provide information on whether there is significant variability in the linear-slope and intercept estimates for the CBM-R and Maze scores. If significant variance in the slope exists, it is necessary to add predictors that explain the significant variance. In contrast, if there is no significant variance in the slope, no further test is required, indicating that the CBM

Table 2 Fit Indices and Unstandardized Parameter Estimates for CBM-R RMSEA a Slope Intercept Model label 2 df (90%CI) CFI b 2 df p (d) M (SE) Var M (SE) Var Grade 3 No growth 421.11 4.69 (.63.74).79 91.99 (2.90) 1,871.16 Linear growth.81 1.00 (.00.11).99 419.31 3.00 3.59 (.12) 2.75 69.06 (2.47) 1,280.35 Conditional model 12.64 5.08 (.03.14) 1.00 3.83 (.58) 44.33 (11.01) Grade 4 No growth 396.16 4.63 (.58.68).84 110.93 (2.39) 1388.55 Linear growth 2.99 1.00 (.00.15).99 393.17 3.00 3.18 (.11) 1.03 90.97 (2.21) 995.25 Conditional model Grade 5 No growth 408.21 4.56 (.51.60).87 116.93 (2.72) 2,357.99 Linear growth.09 1.00 (.00.02).99 408.12 3.00 2.56 (.09) 2.23 102.61 (2.64) 2,201.47 Conditional model 6 Grade 6 No growth 208.92 4.43 (.38.48).93 129.30 (2.68) 1,940.27 Linear growth 2.45 1.03 (.00.13).99 201.30 3.00 1.89 (.11) 1.37 118.84 (2.73) 1,854.90 Conditional model Grade 7 No growth 266.23 4.45 (.40.50).91 137.33 (2.44) 1,829.92 Linear growth 1.48 1.03 (.00.15) 1.00 264.75 3.00 1.93 (.10) 4.31 128.30 (2.34) 1,930.01 Conditional model 6.95 5.03 (.00.09) 1.00 1.87 (.40) 99.05 (8.23) Grade 8 No growth 255.66 4.44 (.40.49).91 138.14 (2.35) 1,615.88 Linear growth.12 1.00 (.00.02).99 245.62 3.00 2.01 (.10).67 127.63 (2.35) 1,633.37 Conditional model Note. CBM-R results reported as words read correctly per minute. 2 chi-square; df degree of freedom; RMSEA root mean square error of approximation; CFI comparative fit index; 2 difference in chi-square tests; df difference in df; p (d) probability of the difference tests. The best fitting models are in italics. a RMSEA.08 indicates good fit. b CFI.95 indicates good fit. p.05. 124 YEO, FEARRINGTON, AND CHRIST

Table 3 Fit Indices and Unstandardized Parameter Estimates for Maze Model label 2 df RMSEA CFI 2 df p (d) M (SE) Var M (SE) Var Grade 3 No growth 43.56 4.21 (.16.27).97 11.54 (.35) 21.53 Linear growth 5.18 1.08 (.00.18).99 38.48 3.00.21 (.03).09 12.56 (.38) 17.81 Conditional model Grade 4 No growth 258.88 4.51 (.46.56).85 14.63 (.37) 19.98 Linear growth 1.74 1.06 (.00.19).94 257.14 3.00.43 (.03).11 11.82 (.31) 7.81 Conditional model Grade 5 No growth 274.51 4.46 (.41.50).88 17.60 (.45) 55.68 Linear growth 6.64 1.08 (.02.16).99 267.87 3.00.53 (.03).29 20.51 (.51) 46.74 Conditional model 18.65 5.09 (.05.14) 1.00.59 (.16) 12.65 (1.62) Grade 6 No growth 106.21 4.30 (.25.35).94 20.32 (.49) 47.71 Linear growth 2.63 1.07 (.00.19).99 84.05 3.00.41 (.04).31 22.33 (.54) 26.59 Conditional model 30.90 5.14 (.09.18).99.28 (.17) 15.36 (1.50) Grade 7 No growth 118.61 4.30 (.25.34).94 21.17 (.46) 45.78 Linear growth.27 1.00 (.00.13).99 118.34 3.00.45 (.04).40 19.32 (.43) 15.38 Conditional model 3.88 5.00 (.00.07) 1.00.16 (.17) 15.44 (1.60) Grade 8 No growth 128.32 4.31 (.26.36).93 19.46 (.46) 45.34 Linear growth 1.97 1.06 (.00.19).94 126.35 3.00.13 (.05).57 18.25 (.48) 24.69 Conditional model 130.00 5.28 (.24.32).96.33 (.25) 20.22 (2.11) Note. Maze results are reported as words selected correctly per 3 min. 2 chi-square; df degree of freedom; RMSEA root mean square error of approximation; CFI comparative fit index; 2 difference in chi-square tests; df difference in df; p (d) probability of the difference tests. The best fitting models are italicized. a RMSEA.08 indicates good fit. b CFI.95 indicates good fit. p.05. Slope Intercept AN INVESTIGATION OF BIAS AND CBM SLOPE 125

126 YEO, FEARRINGTON, AND CHRIST slope is consistent among students, regardless of the characteristics of the population (Ming et al., 2005). In other words, significant variance in the slope would suggest that there are significant individual variances in development trajectories for CBM measures that should be explained by additional variables. The results in Tables 2 and 3 show that there are significant variances in the linear slope for CBM-R and Maze scores in the 3rd and 7th grades and in the 5th, 6th, 7th, and 8th grades, respectively. Interestingly, the CBM-R slope is consistent for all grade levels, except for the 3rd and 7th grade levels. This suggests that the CBM-R slope can be used as an unbiased predictor of improved performance in reading in Grades 4, 5, 6, and 8. For only upper grade level students, it appears that the estimated Maze slope is inconsistent among various subgroups of students. Conditional LGM CBM Fall 1 Free Lunch 1 1 CBM Intercept CBM Winter 5 9 Given the substantial variance in the slope coefficient of the CBM-R and Maze scores in the 3rd and 7th grades and in the 5th, 6th, 7th, and 8th grades, respectively, the next step involved conducting conditional linear growth modeling with predictors, such as demographic characteristics, to account for the significant variance in the slopes and intercepts (see Figure 1). Effects of the predictors on the slopes and intercepts are presented in Table 4. For Grade 3, special education status was negatively related to the slope, controlling for the initial status and other demographic variables, meaning that students receiving special education services had lower growth rates than did general education students. In addition, for Grade 7, the CBM slope was affected by gender, controlling for the initial status and other demographic characteristics, meaning that female students are more likely to have a higher slope than male students. For Grade 3, special education status influenced the initial performance on the CBM-R measure, and for Grade 7, significant variance in the intercept estimates was accounted for by all demographic characteristics. With regard to Maze scores in Grades 5 and 7, there was a significant difference in the slope estimates of boys and girls, meaning that female students showed a faster rate of increase in Maze scores than did male students, controlling for the rate of change and other demographic variables. Gender was a significant predictor of the Maze slopes at all grade levels, meaning that female students were more likely to have higher initial Maze scores than male students, controlling for the rate of change and other demographic variables. At all grade levels (except for Grade 6), SES was related to higher initial Maze scores, controlling for the rate of change and other demographic variables, indicating that students with high SES showed higher initial Maze scores than those with lower SES. Finally, students receiving special education services had lower initial Maze scores than those not receiving special education services, controlling for other demographic variables. Comparison of CBM-R and Maze Scores Interestingly, slope bias on various demographic variables is not consistent between CBM-R and Maze scores. For example, although the slope estimate for CBM-R scores in Grade 3 is homogeneous among participants, the slope estimate for Maze scores among Grade 3 participants is heterogeneous. In addition, for 7th grade students, gender was a significant predictor of increased growth rates for CBM-R scores, whereas the gender variable 0 Gender CBM Slope CBM Spring Special Education Figure 1. Estimated latent growth model (LGM) for demographic characteristics.

AN INVESTIGATION OF BIAS AND CBM SLOPE 127 Table 4 Unstandardized Parameter Estimates and Standard Errors for Conditional Latent Growth Models CBM-R a Maze b Grade 3 Grade 7 Grade 5 Grade 6 Grade 7 Grade 8 Est SE Est SE Est SE Est SE Est SE Est SE Regression effects on intercept Female3I 6.83 4.72 10.26 3.82 1.56.74 2.22.79 2.62.74 1.78.86 Free lunch3i 9.95 4.65 17.01 3.83 1.96.76 1.18.81 2.32.74 2.35.87 Special education3i 36.57 7.78 62.34 5.73 7.01 1.14 10.55 1.04 9.49 1.12 9.61 1.27 Regression effects on slope Female3S.19.24.42.19.29.07.04.09.25.08.00.10 Free lunch3s.21.24.26.26.11.07.07.09.00.08.12.10 Special education3s 1.15.42.34.35.19.11.05.12.08.13.02.16 Note. I intercept; S slope. a Results reported as words read correctly per minute. b Results reported as words selected correctly per 3 min. p.05. was a significant predictor of decreased growth rates for Maze scores. Discussion The focus of the current study was to examine whether the CBM slope is an unbiased indicator of improved performance in reading. The present study extends previous work on bias in CBM measures. As mentioned above, because most previous studies have only highlighted bias in CBM scores at one point in time among subgroups of participants, it was necessary to examine whether the CBM slope estimates are consistent among various characteristics of the population. To our knowledge, this is the first study to examine the potential slope bias of CBM-R and Maze data as a function of demographic variables. This is an important research question, because both CBM-R and Maze are commonly employed to monitor student progress and evaluate instructional effects within RTI models of service delivery. The results of the current study show that there were significant variances in CBM-R slopes in Grades 3 and 7 only, meaning that the growth rate of CBM-R scores from fall to spring is likely to be homogenous among subgroups, except for Grades 3 and 7. Unlike the CBM-R measurement conditions, there were significant variances in the growth rates derived from Maze scores for only the higher grade levels (Grades 5, 6, 7, and 8). This indicates that students from Grades 5 to 8 did not appear to hold a similar growth rate from fall to spring, but that the 3rd and 4th grade readers growth rates were identical from fall to spring. Given that significant variance in CBM-R slopes only existed for Grades 3 and 7, it is evident that the use of the different CBM measures affected the different shape of the growth trajectories of the participants. With regard to student background variables leading to slope bias, those within the CBM-R measurement condition were influenced by group membership with regard to special education and gender. In the case of Maze scores, gender was the only variable that affected slope estimates. Interestingly, for all grade levels (with the exception of Grade 3), special education status was not a significant predictor of slope within the CBM-R measurement conditions. This means that the CBM-R slopes for students receiving special education services are similar to those of general education students. Given that prior published studies (e.g., Deno, Fuchs, Marston, & Shin, 2001) have shown that general education students had larger growth rates than did students with disabilities, it is somewhat unexpected that special education status had no negative impact on the slopes of CBM-R scores in the current sample. One possible explanation for the discrepancy in findings is methodological differences. That is, unlike the previous studies, other demographic variables as well as initial status were controlled for when examining the relationship between special education status and slope. Another reason is that the number of students with disabil-

128 YEO, FEARRINGTON, AND CHRIST ities is very small, comparing the number of students without disabilities. Finally, it is possible that the only absolute difference in CBM-R slope estimates between students with and without disabilities existed during the early grades (e.g., Grade 3). As mentioned above, gender was the only significant variable associated with Maze slope bias. Interestingly, although female students in Grade 7 had higher levels of Maze score growth than male students, CBM-R slope estimates were higher for male than female students. That is, depending on the use of different CBM measures, female students in Grade 7 were either positively or negatively associated with growth rate on CBM measures. Why does gender in Grade 7 lead to inconsistent results? One possible explanation for this result involves the AIMSweb reading passages used in this study. That is, it is quite plausible the AIMSweb passages for the Maze test are biased in favor of male students at the Grade 7 level. Such a result implies that slope bias on various demographic variables may be attributable to effects of progress-monitoring measurement materials used for the two measures. Although both CBM-R and Maze reading measures have been viewed as valid, reliable indices of overall reading competence, the two measures pose different technical adequacies in terms of effects of text materials. Thus, future research should examine the possibility that slope bias on student background variables is due, in part, to differences in the use of CBM measures. We were also interested in the extent to which grade level is associated with significant background variables affecting the magnitude of growth rate. Consistent with prior findings by Chard et al. (2008), we found that regardless of grade level, growth rate of CBM-R scores did not vary as a function of SES background variables. This result suggests that the CBM-R slope may be not biased in terms of SES. However, as previously mentioned, gender and special education status varied with grade level. Limitations and Directions for Future Research Although this study provides meaningful insights into the slope bias derived from CBM measures in reading, some key limitations remain. First, in this study, slope bias of demographic variables was examined with universal benchmark screenings (i.e., fall, winter, and spring) rather than frequent CBM progress monitoring that would be administered weekly or biweekly. As such, it is impossible to generalize outcomes from the current study to other CBM research in which it has been used as a means of frequently measuring the progress of students in the problem-solving model (Hintze et al., 2002). The schedule of data collection also precluded evaluation of alternate nonlinear models for which research provides support (Christ, Silberglitt, Yeo, & Cormier, 2010). Replication and extension of this work, especially using more frequent progress monitoring data, is needed. Second, as mentioned above, the current sample was primarily Caucasian (more than 90%). Thus, future research should consider utilizing similar sample sizes for each ethnic/ racial group in order to evaluate the generalizability of the findings from this study. Third, although this study shows that special education status was not a significant predictor of slope within the CBM measures, except for CBM-R measurement at the Grade 3 level, we did not provide information on the special education sample in this study. It is possible that the percentages of students associated with severe or mild disabilities could be one of potential factors affecting the magnitude of growth rate. Fourth, it should be also noted that the nature of the CBM reading passages used in this study might result in very different outcomes. In fact, some studies have reported evidence that rates of growth change as a function of the effects of text material (e.g., Fuchs & Deno, 1992; Hintze & Shapiro, 1997). In this study, we only used the CBM measures produced by AIMSweb. To advance understanding of the effects of curriculum on CBM reading measures, additional research is needed. Such research should directly examine whether the CBM slope bias of demographic variables is dependent upon the effects of text materials, indicating that the generalizability of our findings should be limited to studies with AIMSweb passage sets. Finally, given the fact that CBM has been used as a valid, reliable indicator of global proficiency in basic academic skills, such as math and written expression, it would be meaningful to conduct similar studies focusing on CBM in other academic areas.

AN INVESTIGATION OF BIAS AND CBM SLOPE 129 Implications for Practice and Research Despite the limitations described above, the present study has significant implications for practice and research. The first implication is that a student s background variables may be associated with a significant amount of variance in the student s rate of growth. Although a previous study by Chard et al. (2008) showed that student demographic variables were not significant predictors of CBM-R slope, this study provides evidence that special education status and gender lead to biased slope estimates at some grade levels. Given that CBM slope estimates serve as a means of identifying students who are struggling to read with researchbased interventions as part of RTI (Fuchs, Fuchs, & Compton, 2004), practitioners who administer CBM measures within an RTI model should realize that average rates of growth might be significantly influenced by students background variables, regardless of the quality of intervention delivered to students at risk. However, more research is needed to provide firm support for the conclusions of this study. The second implication drawn from our study is that schools and teachers should use multiple progress monitoring measures within an RTI framework. Using a single progress monitoring measure is not appropriate for estimating the mean rate of growth to make high-stakes decisions about which students are eligible for special education services providing more intensive interventions (Ardoin & Christ, 2008). This study suggests that choosing which CBM measures to administer may result in substantial slope bias on demographic variables. For instance, it is possible that using CBM-R slopes as indicators of overall reading proficiency leads to overestimations of growth rates for a specific background variable, such as gender, whereas the use of Maze slopes is associated with underestimations of growth rates. Therefore, we encourage schools and practitioners who use CBM measures in reading to administer multiple progress-monitoring measures. The final implication involves previous research in which the rate of progress was estimated without considering student background factors affecting variance in the estimated slope or controlling for initial level of performance. Surprisingly, most CBM studies have ignored the effects of demographic variables and initial levels of performance. As Baker and his colleagues (2008) mentioned, to correctly evaluate CBM slope estimates, multiple factors resulting in their variance should be controlled before interpreting growth rate. References Ardoin, S. P., & Christ, T. J. (2008). Evaluating curriculum-based measurement slope estimates using data from triannual universal screenings. School Psychology Review, 37, 109 125. Baker, S. K., Smolkowski, K., Katz, R., Fien, H., Seeley, J. R., Kame enui, E. J., & Beck, C. T. (2008). Reading fluency as a predictor of reading proficiency in low-performing, high-poverty schools. School Psychology Review, 37, 18 37. Batsche, G., Elliot, J., Graden, J. L., Grimes, J., Kovaleski, J. F., Prasse, D.,... Tilly, W. D., III. (2005). Response to intervention: Policy considerations and implementation. Alexandria, VA: National Association of State Directors of Special Education. Betts, J., Reschley, A., Pickart, M., Heistad, D., Sheran, C., & Marston, D. (2008). An examination of the predictive bias for second grade reading outcomes from measures of early literacy skills in kindergarten with respect to English language learners and ethnic subgroups. School Psychology Quarterly, 23, 553 570. doi:10.1037/1045 3830.23.4.553 Bollen, K. A., & Curran, P. J. (2006). Latent curve models: A structural equation perspective. Hoboken, NJ: Wiley-Interscience. Chard, D. J., Stoolmiller, M., Harn, B. A., Wanzek, J., Vaughn, S., Linan-Thompson, S., & Kame enui, E. J. (2008). Predicting reading success in a multilevel schoolwide reading model: A retrospective analysis. Journal of Learning Disabilities, 41, 174 188. doi:10.1177/ 0022219407313588 Christ, T. J., Silberglitt, B., Yeo, S., & Cormier, D. (2010). Curriculum-Based Measurement of Oral Reading (CBM-R): An evaluation of growth rates and seasonal effects among students served in general and special education. School Psychology Review, 39, 343 349. Clarke, B., & Shinn, M. R. (2004). A preliminary investigation into the identification and development of early mathematics curriculum-based measurement. School Psychology Review, 33, 234 248. Deno, S. L. (1985). Curriculum-based measurement: The emerging alternative. Exceptional Children, 52, 219 232. Deno, S. L. (1986). Formative evaluation of individual student programs: A new role for school

130 YEO, FEARRINGTON, AND CHRIST psychologists. School Psychology Review, 15, 358 374. Deno, S. L. (2003). Developments in curriculumbased measurement. Journal of Special Education, 37, 184 192. Deno, S. L., Fuchs, L. S., Marston, D., & Shin, J. (2001). Using curriculum-based measurement to establish growth standards for students with learning disabilities. School Psychology Review, 30, 507 524. Espin, C. A., Deno, S. L., Maruyama, G., & Cohen, C. (1989, April). The basic academic skills samples: An instrument for screening and identifying children at risk for failure in the regular education classroom. Paper presented at the American Educational Research Association Meeting, San Francisco. Fitzmaurice, G. M., Laird, N. M., & Ware, J. H. (2004). Applied longitudinal analysis. Hoboken, NJ: Wiley-Interscience. Fuchs, L. S. (2004). The past, present, and future of curriculum-based measurement research. School Psychology Review, 33, 188 192. Fuchs, L. S., & Deno, S. L. (1992). Effects of curriculum within curriculum-based measurement. Exceptional Children, 58, 232 243. Fuchs, L. S., Fuchs, D., & Compton, D. L. (2004). Monitoring early reading development in first grade: Word identification fluency versus nonsense word fluency. Exceptional Children, 71, 7 21. Good, R. H., & Kaminski, R. A. (Eds.). (2002). Dynamic indicators of basic early literacy skills (6th ed.). Eugene, OR: Institute for the Development of Educational Achievement. Retrieved from http://dibels.uoregon.edu Hintze, J. M., Callahan, J. E., Matthews, W. J., Williams, S. A., & Tobin, K. G. (2002). Oral reading fluency and prediction of reading comprehension in African American and Caucasian elementary school children. School Psychology Review, 31, 540 553. Hintze, J. M., & Shapiro, E. S. (1997). Curriculumbased measurement and literature-based reading: Is curriculum-based measurement meeting the needs of changing reading curricula? Journal of School Psychology, 35, 351 375. Hu, L.-T., & Bentler, P. M. (1999). Cutoff criteria for fit indexes in covariance structure analysis: Conventional criteria versus. Structural Equation Modeling, 6, 1 55. Individuals with Disabilities Education Improvement Act, Pub. L. No. 108 446, 118 Stat. 2647 (2004). Jenkins, J. R., & Jewell, M. (1993). Examining the validity of two measures for formative teaching: Reading aloud and maze. Exceptional Children, 59, 421 432. Jiban, C., & Deno, S. L. (2007). Using math and reading curriculum-based measurements to predict state mathematics test performance: Are simple one-minute measures technically adequate? Assessment for Effective Intervention, 32, 78 89. Marston, D. B. (1989). A curriculum-based measurement approach to assessing academic performance: What it is and why do it. In M. R. Shinn (Ed.), Curriculum-based measurement: Assessing special children (pp. 18 78). New York: Guilford Press. Ming, C., Conger, R. D., & Lorenz, F. O. (2005). Predicting change in adolescent adjustment from change in marital problems. Developmental Psychology, 41, 812 823. No Child Left Behind Act, Pub. L. No. 107 110, 115 Stat. 1425 (2002). Shin, J., Deno, S. L., & Espin, C. (2000). Technical adequacy of the maze task for curriculum-based measurement of reading growth. Journal of Special Education, 34, 164 172. Shinn, M. R., Good, R. H., Knutson, N., Tilly, W. D., & Collins, V. L. (1992). Curriculum-based measurement of oral reading fluency: A confirmatory analysis of its relation to reading. School Psychology Review, 21, 459 479. Shinn, M. R., & Shinn, M. M. (2002). AIMSweb training workbook. Eden Prairie, MN: EdInformation, Inc. Vaughn, S., & Fuchs, L. S. (2003). Redefining learning disabilities as inadequate response to instruction: The promise and potential problems: Learning Disabilities Research & Practice, 18, 137 146. Wayman, M. M., Wallace, T., Wiley, H. I., Ticha, R., & Espin, C. A. (2007). Literature synthesis on curriculum-based measurement in reading. Journal of Special Education, 41, 85 120. Yeo, S. (2008). Relation between 1-minute CBM reading aloud measure and reading comprehension tests: A multilevel meta-analysis (Unpublished doctoral dissertation). University of Minnesota, Minneapolis and St. Paul. Yeo, S. (2010). Predicting performance on state achievement tests using curriculum-based measurement in reading: A multilevel meta-analysis. Remedial and Special Education, 31, 412 422. doi:10.1177/0741932508327463