Hold my calls: an activity for introducing the statistical process

Similar documents
ICTCM 28th International Conference on Technology in Collegiate Mathematics

Informal Comparative Inference: What is it? Hand Dominance and Throwing Accuracy

Mathematics subject curriculum

Probability and Statistics Curriculum Pacing Guide

Characterizing Mathematical Digital Literacy: A Preliminary Investigation. Todd Abel Appalachian State University

Research Design & Analysis Made Easy! Brainstorming Worksheet

AGS THE GREAT REVIEW GAME FOR PRE-ALGEBRA (CD) CORRELATED TO CALIFORNIA CONTENT STANDARDS

Algebra 1, Quarter 3, Unit 3.1. Line of Best Fit. Overview

Designing a Rubric to Assess the Modelling Phase of Student Design Projects in Upper Year Engineering Courses

Lesson M4. page 1 of 2

Extending Place Value with Whole Numbers to 1,000,000

Analysis of Enzyme Kinetic Data

Number of students enrolled in the program in Fall, 2011: 20. Faculty member completing template: Molly Dugan (Date: 1/26/2012)

Mathematics Program Assessment Plan

Edexcel GCSE. Statistics 1389 Paper 1H. June Mark Scheme. Statistics Edexcel GCSE

Montana Content Standards for Mathematics Grade 3. Montana Content Standards for Mathematical Practices and Mathematics Content Adopted November 2011

South Carolina English Language Arts

What is Thinking (Cognition)?

Lecturing Module

Spinners at the School Carnival (Unequal Sections)

Science Fair Project Handbook

Full text of O L O W Science As Inquiry conference. Science as Inquiry

ESTABLISHING A TRAINING ACADEMY. Betsy Redfern MWH Americas, Inc. 380 Interlocken Crescent, Suite 200 Broomfield, CO

KENTUCKY FRAMEWORK FOR TEACHING

re An Interactive web based tool for sorting textbook images prior to adaptation to accessible format: Year 1 Final Report

What s in a Step? Toward General, Abstract Representations of Tutoring System Log Data

The Editor s Corner. The. Articles. Workshops. Editor. Associate Editors. Also In This Issue

Ohio s Learning Standards-Clear Learning Targets

Professional Learning Suite Framework Edition Domain 3 Course Index

Copyright Corwin 2015

Unit 3. Design Activity. Overview. Purpose. Profile

What is beautiful is useful visual appeal and expected information quality

Technical Manual Supplement

Study Group Handbook

VOL. 3, NO. 5, May 2012 ISSN Journal of Emerging Trends in Computing and Information Sciences CIS Journal. All rights reserved.

CHAPTER 4: REIMBURSEMENT STRATEGIES 24

The lab is designed to remind you how to work with scientific data (including dealing with uncertainty) and to review experimental design.

Maximizing Learning Through Course Alignment and Experience with Different Types of Knowledge

Cal s Dinner Card Deals

Politics and Society Curriculum Specification

Missouri Mathematics Grade-Level Expectations

Contact: For more information on Breakthrough visit or contact Carmel Crévola at Resources:

Introduction to the Practice of Statistics

Scoring Guide for Candidates For retake candidates who began the Certification process in and earlier.

PEDAGOGICAL LEARNING WALKS: MAKING THE THEORY; PRACTICE

Tun your everyday simulation activity into research

Summary results (year 1-3)

Shockwheat. Statistics 1, Activity 1

Kentucky s Standards for Teaching and Learning. Kentucky s Learning Goals and Academic Expectations

DEVELOPING YOUNG STUDENTS INFORMAL INFERENCE SKILLS IN DATA ANALYSIS 7

Functional Skills Mathematics Level 2 assessment

Experience College- and Career-Ready Assessment User Guide

Content Language Objectives (CLOs) August 2012, H. Butts & G. De Anda

WE GAVE A LAWYER BASIC MATH SKILLS, AND YOU WON T BELIEVE WHAT HAPPENED NEXT

Integrating simulation into the engineering curriculum: a case study

Laboratory Notebook Title: Date: Partner: Objective: Data: Observations:

Developing an Assessment Plan to Learn About Student Learning

Proficiency Illusion

Self Study Report Computer Science

Case study Norway case 1

Qualitative Site Review Protocol for DC Charter Schools

Math 96: Intermediate Algebra in Context

Learning Disabilities and Educational Research 1

NORTH CAROLINA VIRTUAL PUBLIC SCHOOL IN WCPSS UPDATE FOR FALL 2007, SPRING 2008, AND SUMMER 2008

Mathematics Scoring Guide for Sample Test 2005

Document number: 2013/ Programs Committee 6/2014 (July) Agenda Item 42.0 Bachelor of Engineering with Honours in Software Engineering

Kindergarten Lessons for Unit 7: On The Move Me on the Map By Joan Sweeney

Grade 6: Module 2A Unit 2: Overview

Exemplar 6 th Grade Math Unit: Prime Factorization, Greatest Common Factor, and Least Common Multiple

What is PDE? Research Report. Paul Nichols

Application of Virtual Instruments (VIs) for an enhanced learning environment

BIOH : Principles of Medical Physiology

An Empirical Analysis of the Effects of Mexican American Studies Participation on Student Achievement within Tucson Unified School District

Facing our Fears: Reading and Writing about Characters in Literary Text

Lab 1 - The Scientific Method

On-the-Fly Customization of Automated Essay Scoring

EQuIP Review Feedback

SPECIALIST PERFORMANCE AND EVALUATION SYSTEM

Call Center Assessment-Technical Support (CCA-Technical Support)

Alpha provides an overall measure of the internal reliability of the test. The Coefficient Alphas for the STEP are:

Georgetown University School of Continuing Studies Master of Professional Studies in Human Resources Management Course Syllabus Summer 2014

Association Between Categorical Variables

Observing Teachers: The Mathematics Pedagogy of Quebec Francophone and Anglophone Teachers

Characteristics of the Text Genre Informational Text Text Structure

Characteristics of Functions

Conceptual Framework: Presentation

FIGURE IT OUT! MIDDLE SCHOOL TASKS. Texas Performance Standards Project

Math-U-See Correlation with the Common Core State Standards for Mathematical Content for Third Grade

Modified Systematic Approach to Answering Questions J A M I L A H A L S A I D A N, M S C.

How to Judge the Quality of an Objective Classroom Test

NCEO Technical Report 27

Providing Feedback to Learners. A useful aide memoire for mentors

THEORY OF PLANNED BEHAVIOR MODEL IN ELECTRONIC LEARNING: A PILOT STUDY

Author's response to reviews

Inquiry Learning Methodologies and the Disposition to Energy Systems Problem Solving

MENTORING. Tips, Techniques, and Best Practices

International Business BADM 455, Section 2 Spring 2008

ACCOMMODATIONS FOR STUDENTS WITH DISABILITIES

CEFR Overall Illustrative English Proficiency Scales

Transcription:

Original Article Hold my calls: an activity for introducing the statistical process Todd Abel Department of Mathematical Sciences, Appalachian State University, Boone, North Carolina, USA e-mail: abelta@appstate.edu Lisa Poling Reich College of Education, Appalachian State University, Boone, North Carolina, USA e-mail: polingll@appstate.edu Summary Keywords: Working with practicing teachers, this article demonstrates, through the facilitation of a statistical activity, how to introduce and investigate the unique qualities of the statistical process including: formulate a question, collect data, analyze data, and interpret data. Teaching; Statistical literacy; Problem solving cycle; Statistical process; Teaching statistics; Data investigation. INTRODUCTION Statistical reasoning defined by Gal (2000, 2002) is the ability to problem solve and reason with data in a way that makes data succinct and understandable. To be able to reason statistically, statisticians and statistical educators have developed standard phases of the statistical process that can be described as follows: (1) specify the problem and outline a plan; (2) collect the required data; (3) analyse and represent the data; and (4) interpret and discuss the findings (Marriott et al. 2009; Guidelines for Assessment and Instruction in Statistics Education (GAISE), 2005; Wild and Pfannkuch 1999). Teachers of statistics should understand the statistical process not only as a way of doing statistics but as a primary means of motivating, introducing, investigating and applying statistical concepts. Indeed, to be engaged in the conceptual understanding of statistics, both learner and instructor must value and understand the process of statistical thinking. Unfortunately, inappropriate reasoning about statistical ideas is widespread and persistent, similar at all age levels, and quite difficult to change (Garfield and Ben- Zvi 2007, p. 374). The purpose of this article is to describe a statistical activity used with practicing teachers to introduce and investigate the statistical process. While the activity is intended for secondary level learners and could be used to teach particular statistical content, the primary intent is to highlight and discuss the statistical process. The statistical process Learning experiences are bound by how individuals navigate information and what they deem to be pertinent and essential. According to Wild and Pfannkuch (1999), the ultimate goal of statistical investigation is learning in the context sphere (p 225). When selecting tasks for the classroom, the significance of the activity to the individual and the focus on statistical reasoning skills become paramount to the learning experience. Designing an investigation based on data requires that teachers identify issues of interest and frame the questions, together with considerations of practicalities in collecting the data (MacGillivray 2007, p. 53), in a manner that adheres to the most appropriate measure of statistical analysis (MacGillivray 2007). Using a statistical process model provides a framework that allows individuals to proceed through an exploration that promotes uniformity and reliability. The problem-solving approach in teaching statistics is considered to improve students skills, particularly as they interact with real data (Marriott et al. 2009, p. 3). Although various models of the problem-solving approach exist, for this study, we used a four-step model, the GAISE framework (2005). Designed by statisticians and statistics educators in the USA, the 96

Hold my calls 97 GAISE framework provides a conceptual framework for K 12 statistics education. The GAISE framework outlines a statistical process consisting of four components: (1) formulate questions; (2) collect data; (3) analyse data; and (4) interpret results (GAISE, 2005). The GAISE framework is then organized into three sequential developmental levels. As students progress across levels, learning becomes more student directed and less teacher directed, and student engagement becomes more nuanced and sophisticated (figure 1). Importance for teachers Ben-Zvi and Garfield (2004) identify various factors that challenge both learners and teachers of statistics: statistical ideas are complex, lack of mathematical knowledge impedes the computational aspect of statistics, the context in which statistical problems are presented is confusing and individuals transfer limited understanding of mathematics to statistics focusing on getting the correct answer but not interpreting results. To create environments that scaffold learning and engage students, instructional strategies must provide support so that students can experience the full range of the statistical process: enter the problem, allow for dissonance and make connections to existing concepts (Garfield and Ben-Zvi 2009). Students, as constructors of knowledge, must be actively engaged as they explore probability and statistics. Teachers therefore must provide consistent feedback, giving students opportunities to practice using statistical techniques and address misconceptions as sources of knowledge (Garfield 1995; Garfield and Ben-Zvi 2007). Just as students need experience to understand a concept, instructors need to first experience a concept in order to fully understand and more effectively teach it (Polly and Hannafin 2011). Thus, in order to help teachers experience and understand the statistical process, we developed and implemented a task during a professional development workshop for grades 6 12 mathematics teachers that used the GAISE framework to engage participants in that process. The underpinning and consistency of using the GAISE framework allow learners to organize and enter into a statistical problem, providing the scaffold that learners need in order to make sense of specific statistical content. Creating such carefully designed sequences helps students improve reasoning and understanding (Garfield and Ben-Zvi 2007). Moreover, it emphasizes the nature of statistics and the practices and processes involved in statistical thought. CONTEXT In subsequent sections, an activity is outlined that was used with practicing teacher, showing the four steps in the problem-solving approach to teaching statistics. Four different groups of practicing teachers engaged in the activity at various times during 2 days of professional development. As a result, the data displayed in the examples differ depending on the group. Fig. 1. Guidelines for Assessment and Instruction in Statistics Education framework

98 Todd Abel and Lisa Poling INTRODUCING THE STATISTICAL PROCESS WITH AN ACTIVITY Adapting the Cell Phone Impairment? activity from the Statistics Education Web (STEW) (nd) of the American Statistics Association, the lesson investigated the relationship between cell phone usage and reaction time. Changes in the activity were made to conform to the particular context and emphasize the statistical process, permitting multiple levels of engagement. Formulate questions The first phase of the statistics process involves formulating questions to be investigated. Participants were given the following prompt: According to the National Safety Council, cell phones are a factor in 1.3 million traffic accidents each year, resulting in thousands of deaths and injuries. Many have conjectured that this is due to decreased reaction time and attentiveness. What questions does this prompt you to ask? The open-ended nature of the prompt was useful for eliciting a variety of possible responses. This highlighted the questioning habits of teachers in the groups and motivated a discussion of characteristics of good statistics questions. Initial questions tended to either address the reliability of the statement or propose possible contributing variables. For instance, how do they know cell phones are a factor? where did the number come from define traffic accident? how is cell phone use reported? how does this compare to other distracting activities as a factor? is it just cell phones or any distraction? what are the demographics of drivers? how will you measure reaction time? hands-free versus not? does type of phone matter? how is phone being use? does age matter? were there other possible reasons? Is cell phone the only/leading factor? men or women? does type of driving (setting) make difference? These questions lack specificity and do not anticipate variability in a way that can be systematically investigated. Many of them are, however, important questions to consider and were useful for considering sources of variation and possible avenues for investigation. Listing such questions encouraged the groups to consider the characteristics of good statistical questions and prompted a discussion about what could or should be investigated. In table groups, participants then proposed possible statistics questions refined from the initial list. A direct investigation of the reasonableness of the statistics from the prompt was unrealistic, so the focus was placed on the conjectured relationship between cell phone usage and reaction time. Ultimately, each group settled on the following question: Does cell phone usage have an impact on reaction time? Although it does not consider any effects on driving, the final question could guide a classroom investigation and was compelling to the participants. The nature of the initial questions highlights that teachers, as with students, need opportunities to practice and develop the statistical habit of mind of questioning. In the classroom, as in our example, it may be necessary for instructors to use questioning to guide students to a particular well-defined statistical question. If the goal of the lesson is to understand the statistical process, however, it is important that students have opportunities to propose questions, and that viable alternatives are recognized. Collect data Data collection procedures must account for potential sources of variation in data. While students need opportunities to design data collection procedures, allowing a class to design an experiment from scratch is often impractical. Implementing ideal data collection procedures can also be difficult. In many cases, a classroom pilot experiment may highlight the necessary content while permitting students to be actively involved in data collection. In this activity, basic data collection procedures were given, but participants were asked to analyse the procedures for possible sources of variability and then specify ways to minimize them. The following data collection procedures were proposed to the group: 1. Participants worked with a partner, and two sets of partners formed a table group of four. 2. Reaction time was quantified using a ruler drop method. One partner (the dropper) held a ruler and dropped it at random, while the other partner (the catcher) held their hand at the bottom end of the ruler and caught it as it fell. Reaction

Hold my calls 99 Table 1. Types of analysis Individualized Aggregate Separate conditions Type I C: graph compares each individual s reaction time for each condition: with and without a cell phone. Type A C: graph compares the aggregate data for each condition: with and without a cell phone. Difference Type I D: graph illustrates the difference in two reaction times for each individual. Type A D: graph illustrates the difference in reaction time for the overall data set, instead of by individual. time was measured by recording the distance the ruler travelled before being caught. 1 3. A coin flip determined whether the table group would measure reaction time with a cell phone or without a cell phone first. a. Each partner took a turn as catcher and as dropper without a cell phone. 2 b. Each partner took a turn as catcher and as dropper. During a particular turn, the two catchers in a table group talked to each other on cell phones. To simulate a conversation, each participant created a list of ten words, which they read to the other catcher, who had to try and repeat as many back as they could. While doing that, the dropper would drop the ruler. 4. Each table group input their data into a collaborative spreadsheet to make it available to the whole group. When prompted, participants identified and clarified several other possible sources of variability, such as the following: How will the dropper hold the ruler? Where and how should the catcher hold their hand? Where should the measurement be taken? What finger will locate the measurement? What level of precision is expected in the length measurements (nearest inch, quarter inch, etc.)? How many trials with and without the cell phone? 1 Note that the measurement of reaction time using these data collection procedures is actually a length measurement. Because the ruler will accelerate due to gravity, the length and time are not proportional. During data analysis, one could convert the lengths to times using the equation d = 192t 2, where t is time in seconds and d is distance in inches. However, the question can be addressed without attending to that conversion, as it was in the instance presented here. Participants did acknowledge the length-to-time conversion issue, and one group modelled the conversion. The results were close enough to proportional, and the times short enough, that this group decided to ignore the conversion to time and work strictly with the recorded lengths. Thus, in what follows, reaction time refers to the point on the ruler where the ruler was caught. 2 If available, reaction time apps are freely available for laptops, tablets and smartphones and could be used in place of the ruler drop method. Using a think, pair, share protocol, data collection procedures were standardized as a group. Participants were also asked to identify limitations of the data collection plan, such as the limited population (limited to the classroom), the minimal number of trials and possible confounding of data from subjects being observed by other subjects. Despite the quasi-experimental nature of data collection, these procedures were sufficient to highlight key characteristics of experimental design and illustrate the nature of the data collection phase of the statistical process. Issues of data entry, handling and preparation were not emphasized here. Analyse data During the data analysis phase of the statistical process, data are explored and represented. In instructional settings, the data analysis phase may serve as an opportunity to motivate and implement new techniques, to further explore or apply previously developed techniques, or to compare different analyses. The goal of this particular activity was to introduce and investigate the statistical process rather than any particular content, so each table group was asked to examine the data by creating a visual representation of their choice, an important first step in data analysis. This permitted engagement at different levels, thereby creating opportunities to discuss the statistical process and differentiate levels of understanding. Although this group did not explore further, more sophisticated analysis techniques would certainly be appropriate given different learning goals. It was observed that table groups organized their choice of visual representation by two main categorizations: individualized versus aggregate and separate versus difference as shown in table 1. Below, examples of each type are shown, and conceptual underpinnings (or misunderstandings) underlying each are discussed. Note that each table group of four participants created a graph for the data from their entire class. Four different classes took part in the activity, so the data differ somewhat among the graphs shown below. Type I C table groups placed an emphasis on representing each individual time, both with and

100 Todd Abel and Lisa Poling Fig. 2. Examples of type I C graphs without the cell phone. These graphs attempted to give an overall sense of which time was generally higher without regard to the magnitude of the differences for each individual or attempt to summarize the data. A variety of type I C representations was used. Some table groups created double bar graphs showing each individual s reaction time with and without the cell phone. Other table groups created a scatterplot, with each ordered pair (without cell phone and with cell phone) representing an individual. Effect on reaction time was visualized by comparing points to the line y = x. Other table groups created line graphs that attempted to show the same person-by-person comparison. While each of these graphs gives the overall impression that the cell phone reaction times were typically longer, they do not give any specific information about the group as a whole, attending instead to individuals. In addition, the line graph is particularly misleading since the data on the horizontal axis (individuals) are not ordered (figure 2). Only one table group created a graph of type I D. This group calculated the change in reaction time (with cell phone) and (no cell phone) for each individual, then plotted each one using a type of bar graph. This does attend to the differences in reaction time, but still emphasizes each Fig. 3. An example of a type I D bar graph individual rather than giving any sense of the data as a whole. The graph itself is problematic it is not initially clear that each bar represents an individual (on first glance, it looks more like a histogram) and makes the horizontal axis appear to be ordered when it is not (figure 3). The most popular type of graph was type A C. These groups compared the aggregated data with and without a cell phone and used summary statistics to display information about each treatment on the entire group. One table group simply found the mean of the reaction times with and without the cell phone and created a bar graph to show the two means. More common, however, was a comparative box plot one box plot for the aggregate data within each condition. Such graphs are moving beyond looking at each individual subject and instead taking advantage of aggregate data. However, they are not taking account of the experimental situation and hence not conveying any specific information about the differences. This problem is emphasized by an incorrect interpretation described below (figure 4). Type A D graphs considered the difference (with cell phone and without cell phone) for each individual then created graphs that looked at the data as a whole instead of individually. For instance, one group created a box plot using the differences, and several groups created histogram plots for the set of data of the differences. Such graphs illustrated the most sophisticated representation of the data, moving beyond individual responses and giving a sense of the typical difference in reaction times. Note how only the A D graphs can pick up both the overall effects and the nuances of effects and variation in individuals of using cell phones (figure 5). Allowing participants to choose and create their own displays of data provided opportunities to compare the different data displays, highlighting the strengths and weakness, and possible misapplications, of each. It also highlighted that the choice of appropriate graphs is an important topic

Hold my calls 101 Fig. 4. Examples of type A C graphs Fig. 5. Examples of type A D graphs for teachers to consider, and one with which they may need to gain experience. Comparing different graphs also prompted discussions about what attributes of the data were important for addressing the original question. Interpret results It can be challenging to generalize from data without making unsupported claims. Each table group was asked to write the conclusions that they could draw from their graphs. Several examples are represented in the pictures above. The participants conclusions often reflected the type of representations and the sophistication of their understanding of data. For instance, a very general conclusion such as the reaction time is greater when using a cell phone reflected the nonspecific graphical analysis undertaken in the task. Type I C and type I D graphs did not illustrate any specific measurements of differences in reaction time, permitting only general overall impressions of the effect. Table groups creating type A C and type A D graphs, however, could use summary statistics to support their conclusions. Although type A C graphs ignored the paired nature of the data, in this particular case, the contrast between using and not using cell phones was far greater than variation across individuals, allowing participants who created comparative box plots to make statements such as the following: [The] median is higher for use of the cell phone. Maximum data is much higher for reaction time with cell phone. All measurements of quartiles without the cell phone are less than the quartile measures of with a cell phone. Asking participants to draw and support conclusions from their graphs also highlighted some misconceptions and problems with choice of graphical representation. For instance, the table group shown below created a comparative box plot and concluded that the graph shows that over 75% can react faster without the distraction of a cell phone! While it is true that 75% of the cell phone reaction times were greater than at least 75% of the without cell phone reaction times, this does not imply that 75% of individuals had faster reaction times without cell phones. Since differences on individuals were not used, thegraphcannotbeusedtomakestatements (figure 6). To emphasize the importance of interpreting results in a meaningful and well-supported way, participants were asked if they were confident enough in their conclusions to 1. change their own behaviour, 2. use it as evidence to influence the behaviour of others or 3. present it to other statistically proficient people.

102 Todd Abel and Lisa Poling Fig. 6. A misleading representation of data with misconceptions in the conclusion The subsequent discussion allowed the group to not just propose conclusions but also support them and analyse the strength of those conclusions. teacher participants became students. Participants were asked to identify the key ideas they would like to carry forward into their classrooms. All groups identified the nature of statistical investigation as amajorpointofemphasis.forinstance,onegroup noted that they realized they needed to let students be flexible [it s] not a linear path,whileanother described it as not a set pattern. Theideaof different levels of sophistication was also meaningful to many groups. Opportunities such as this are important for both teachers and students. Using the statistical investigation process as a framework for the activity structured the experience and supported participant growth in statistical understanding. This activity illustrated how intentionally designed learning opportunities can support understanding of the statistical process and the nature of statistical work. DISCUSSION Summarizing the activity Since the learning goal of this activity was the statistical process, the groups identified the portion of the activity that corresponded to each phase of the statistical process, identifying the key characteristics. For instance, standardizing data collection procedures in order to isolate a single source of variability (as much as possible) was noted as a key characteristic of the collect data phase. Making choices about how to organize, summarize and display the data was identified as key characteristics of the analyse data phase. The reaction time task is typical of many activities in that it permits adaptations that would increase or decrease the statistical sophistication of any given component. For instance, the question could be posed for the class, and only individualized data might be considered. Such adaptations may be appropriate for an initial exposure to the statistical process if the issues of variability involved in each component are discussed. For a more sophisticated approach to the activity, students could design data collection procedures that incorporate random sampling outside the classroom, and data analysis could include tests of significance and measures of variability. CONCLUSION To promote the statistical process, the implementation of the problem-solving approach for statistics, References Ben-Zvi, D. and Garfield, J. (Eds) (2004). The Challenge of Developing Statistical Literacy, Reasoning, and Thinking, Dordrecht, Netherlands: Kluwer. Gal, I. (Ed) (2000). Adult Numeracy Development: Theory, Research, Practice, Cresskill, NJ: Hampton Press. Gal, I. (2002). Adults statistical literacy: meaning, components, responsibilities. International Statistical Review, 70(1), 1 5. Garfield, J. (1995). How students learn statistics. International Statistical Review, 63, 25 34. Garfield, J. and Ben-Zvi, D. (2007). How students learn statistics revisited: a current review of research on teaching and learning statistics. International Statistical Review, 75, 372 396. Garfield, J. and Ben-Zvi, D. (2009). Helping students develop statistical reasoning: implementing a statistical reasoning learning environment. Teaching Statistics, 31,72 77. Franklin,C.,Kader,G.,Mewborn,D.,Moreno,J., Peck, R., Perry, M. and Scheaffer, R. (2005). Guidelines for Assessment and Instruction in Statistics Education (GAISE) Report; A Pre-K Curriculum Framework, Alexandria, VA: American Statistical Association. MacGillivray, H. (2007). Clasping hands and folding arms: a data investigation. Teaching Statistics, 29, 49 53. Marriott, J., Davies, N. and Gibson, L. (2009). Teaching, learning and assessing statistical problem solving. Journal of Statistics Education, 17(1). Retrieved from http: www.amstat.org/

Hold my calls 103 publications/jse/v17n1/marriott.html (accessed 28 January 2015). Polly, D. and Hannafin, M.J. (2011). Examining how learner-centered professional development influences teachers/espouse and enacted practices. The Journal of Education Research, 104, 120 130. Statistics Education Web (STEW). (nd) Retrieved from http://www.amstat.org/education/stew/ (accessed May 2014). Wild, C.J. and Pfannkuch, M. (1999). Statistical thinking in empirical enquiry. International Statistical Review, 67(3), 223 265.