Characterizing Research in Computing Education: A preliminary analysis of the literature

Similar documents
TU-E2090 Research Assignment in Operations Management and Services

Identifying Novice Difficulties in Object Oriented Design

Inquiry Learning Methodologies and the Disposition to Energy Systems Problem Solving

Maximizing Learning Through Course Alignment and Experience with Different Types of Knowledge

THE ROLE OF TOOL AND TEACHER MEDIATIONS IN THE CONSTRUCTION OF MEANINGS FOR REFLECTION

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

Note: Principal version Modification Amendment Modification Amendment Modification Complete version from 1 October 2014

A Case Study: News Classification Based on Term Frequency

Activities, Exercises, Assignments Copyright 2009 Cem Kaner 1

10.2. Behavior models

Logical Soft Systems Methodology for Education Programme Development

Graduate Program in Education

Linking the Common European Framework of Reference and the Michigan English Language Assessment Battery Technical Report

The Impact of Instructor Initiative on Student Learning: A Tutoring Study

Ontologies vs. classification systems

Document number: 2013/ Programs Committee 6/2014 (July) Agenda Item 42.0 Bachelor of Engineering with Honours in Software Engineering

Preliminary Report Initiative for Investigation of Race Matters and Underrepresented Minority Faculty at MIT Revised Version Submitted July 12, 2007

WE GAVE A LAWYER BASIC MATH SKILLS, AND YOU WON T BELIEVE WHAT HAPPENED NEXT

A cognitive perspective on pair programming

Research as Design-Design as Research

A Coding System for Dynamic Topic Analysis: A Computer-Mediated Discourse Analysis Technique

Designing a Case Study Protocol for Application in IS research. Hilangwa Maimbo and Graham Pervan. School of Information Systems, Curtin University

Ph.D. in Behavior Analysis Ph.d. i atferdsanalyse

ACADEMIC AFFAIRS GUIDELINES

Strategies that Students Use to Trace Code: An Analysis Based in Grounded Theory

Pair Programming: When and Why it Works

A pilot study on the impact of an online writing tool used by first year science students

AGENDA LEARNING THEORIES LEARNING THEORIES. Advanced Learning Theories 2/22/2016

A Note on Structuring Employability Skills for Accounting Students

WHY SOLVE PROBLEMS? INTERVIEWING COLLEGE FACULTY ABOUT THE LEARNING AND TEACHING OF PROBLEM SOLVING

On-the-Fly Customization of Automated Essay Scoring

Analysis: Evaluation: Knowledge: Comprehension: Synthesis: Application:

On Human Computer Interaction, HCI. Dr. Saif al Zahir Electrical and Computer Engineering Department UBC

Further, Robert W. Lissitz, University of Maryland Huynh Huynh, University of South Carolina ADEQUATE YEARLY PROGRESS

Planning a Dissertation/ Project

GCSE English Language 2012 An investigation into the outcomes for candidates in Wales

Software Maintenance

What Different Kinds of Stratification Can Reveal about the Generalizability of Data-Mined Skill Assessment Models

VOL. 3, NO. 5, May 2012 ISSN Journal of Emerging Trends in Computing and Information Sciences CIS Journal. All rights reserved.

THE PENNSYLVANIA STATE UNIVERSITY SCHREYER HONORS COLLEGE DEPARTMENT OF MATHEMATICS ASSESSING THE EFFECTIVENESS OF MULTIPLE CHOICE MATH TESTS

CS Machine Learning

Automating Outcome Based Assessment

Abstractions and the Brain

HEROIC IMAGINATION PROJECT. A new way of looking at heroism

NATIONAL CENTER FOR EDUCATION STATISTICS RESPONSE TO RECOMMENDATIONS OF THE NATIONAL ASSESSMENT GOVERNING BOARD AD HOC COMMITTEE ON.

Characterizing Mathematical Digital Literacy: A Preliminary Investigation. Todd Abel Appalachian State University

Rule Learning With Negation: Issues Regarding Effectiveness

ASSESSMENT REPORT FOR GENERAL EDUCATION CATEGORY 1C: WRITING INTENSIVE

Is operations research really research?

Observing Teachers: The Mathematics Pedagogy of Quebec Francophone and Anglophone Teachers

A Characterization of Calculus I Final Exams in U.S. Colleges and Universities

Last Editorial Change:

Content analysis (qualitative, thematic) (Last updated: 9/4/06, Yan Zhang)

Khairul Hisyam Kamarudin, PhD 22 Feb 2017 / UTM Kuala Lumpur

Handbook for Graduate Students in TESL and Applied Linguistics Programs

Implementing a tool to Support KAOS-Beta Process Model Using EPF

A Metacognitive Approach to Support Heuristic Solution of Mathematical Problems

AQUA: An Ontology-Driven Question Answering System

CHAPTER 4: RESEARCH DESIGN AND METHODOLOGY

THEORY OF PLANNED BEHAVIOR MODEL IN ELECTRONIC LEARNING: A PILOT STUDY

School Leadership Rubrics

GUIDE TO EVALUATING DISTANCE EDUCATION AND CORRESPONDENCE EDUCATION

STUDENT LEARNING ASSESSMENT REPORT

Metadiscourse in Knowledge Building: A question about written or verbal metadiscourse

The IDN Variant Issues Project: A Study of Issues Related to the Delegation of IDN Variant TLDs. 20 April 2011

Practical Research. Planning and Design. Paul D. Leedy. Jeanne Ellis Ormrod. Upper Saddle River, New Jersey Columbus, Ohio

Master Program: Strategic Management. Master s Thesis a roadmap to success. Innsbruck University School of Management

Higher education is becoming a major driver of economic competitiveness

EQuIP Review Feedback

Master s Programme in European Studies

Case study Norway case 1

CONCEPT MAPS AS A DEVICE FOR LEARNING DATABASE CONCEPTS

HDR Presentation of Thesis Procedures pro-030 Version: 2.01

Success Factors for Creativity Workshops in RE

MASTER S THESIS GUIDE MASTER S PROGRAMME IN COMMUNICATION SCIENCE

b) Allegation means information in any form forwarded to a Dean relating to possible Misconduct in Scholarly Activity.

A Study of Successful Practices in the IB Program Continuum

Developing a Language for Assessing Creativity: a taxonomy to support student learning and assessment

A cautionary note is research still caught up in an implementer approach to the teacher?

What is Thinking (Cognition)?

On-Line Data Analytics

The Political Engagement Activity Student Guide

Mathematics Program Assessment Plan

An application of student learner profiling: comparison of students in different degree programs

Stimulating Techniques in Micro Teaching. Puan Ng Swee Teng Ketua Program Kursus Lanjutan U48 Kolej Sains Kesihatan Bersekutu, SAS, Ulu Kinta

WORK OF LEADERS GROUP REPORT

IS USE OF OPTIONAL ATTRIBUTES AND ASSOCIATIONS IN CONCEPTUAL MODELING ALWAYS PROBLEMATIC? THEORY AND EMPIRICAL TESTS

Strategies for Solving Fraction Tasks and Their Link to Algebraic Thinking

DYNAMIC ADAPTIVE HYPERMEDIA SYSTEMS FOR E-LEARNING

STA 225: Introductory Statistics (CT)

1 3-5 = Subtraction - a binary operation

Feature-oriented vs. Needs-oriented Product Access for Non-Expert Online Shoppers

Leveraging MOOCs to bring entrepreneurship and innovation to everyone on campus

General study plan for third-cycle programmes in Sociology

LEAD 612 Advanced Qualitative Research Fall 2015 Dr. Lea Hubbard Camino Hall 101A

Deploying Agile Practices in Organizations: A Case Study

Department of Sociology Introduction to Sociology McGuinn 426 Spring, 2009 Phone: INTRODUCTION TO SOCIOLOGY AS A CORE COURSE

South Carolina English Language Arts

Systematic reviews in theory and practice for library and information studies

Transcription:

Characterizing Research in Computing Education: A preliminary analysis of the literature Lauri Malmi Lauri.Malmi@tkk.fi Roman Bednarik University of Eastern Finland roman.bednarik@cs.joensuu.fi Niko Myller University of Eastern Finland nmyller@cs.joensuu.fi Judy Sheard Monash University, Australia judy.sheard@infotech.monash.edu.au Juha Helminen juha.helminen@cs.hut.fi Juha Sorva Juha.Sorva@iki.fi Simon University of Newcastle, Australia simon@newcastle.edu.au Ari Korhonen archie@cs.hut.fi Ahmad Taherkhani ahmad@cc.hut.fi ABSTRACT This paper presents a preliminary analysis of research papers in computing education. While previous analysis has explored what research is being done in computing education, this project explores how that research is being done. We present our classification system, then the results of applying it to the papers from all five years of ICER. We find that this subset of computing education research has more in common with research in information systems than with that in computer science or software engineering; and that the papers published at ICER generally appear to conform to the specified ICER requirements. Categories and Subject Descriptors K.3.2 [Computers and education]: Computer and Information Science Education computer science education; A.0 [General]: conference proceedings. General Terms Measurement. Keywords Classifying publications, computing education, research methods. 1. INTRODUCTION Computing education is a relatively new area of research with a growing body of literature. It is multifaceted and draws on a range of disciplines with more established traditions of research (e.g. cognitive psychology, education, and behavioral science). As computing education research matures as a discipline it is important to understand what characterizes this area of research and where it is positioned in the research milieu. This will help establish computing education as an identifiable and acknowledged field of research, increasing the credibility and acceptance of its research findings. For this to happen we must be aware of the type of research that is Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. ICER 2010, August 9 10, 2010, Aarhus, Denmark. Copyright 2010 ACM 978-1-4503-0257-9/10/08...$10.00. conducted in computing education and the theoretical bases upon which this research is founded. One way to achieve this awareness is through analysis of the computing education research literature. In recent years a number of studies have examined the literature of computing education research (CER). Some of them have concentrated on identifying the subfields of CER [13, 23]. Later work investigated the characteristics of CER, forming an overview of what kind of work is being carried out in the field and commenting on the quality of work [16, 27-29, 33-36]. Some metaanalysis has focused more narrowly on the research literature on the teaching and learning of programming [24, 32, 38]. All of these studies have provided valuable overviews of research in computing education and have contributed to our understanding of the current state of the art of CER. An important motivation for this kind of work is to improve the quality of research by recognizing good practices and exploring what might qualify as rigorous research in different contexts. Analysis of this sort also assists in the making of a clearer distinction between practice reports and research papers. Practice reports focus on disseminating new ideas, tools and approaches among computing educators. Some research papers aim at rigorously testing such novel ideas to evaluate their effectiveness and generalizability, while others set out to build a deeper understanding of students and teachers conceptions and working processes. In this paper we continue this classification work by looking more deeply at the research processes documented in CER papers. We present a purpose-built classification scheme and demonstrate its use with an analysis of papers from five years of ICER workshops. Specifically, we ask: What theories, models, frameworks, instruments, technologies or tools have been used, built upon, or extended by the research? What other disciplines does the research link to? What was the general purpose of the research: describing something new, formulating new tools or methods, or evaluating them? What research frameworks have been used? What kind of data has been collected for the research and how has it been analyzed? 3

Through this analysis we hope to get a better understanding of how computing education research builds on previous work, what features it uses from other disciplines, what purposes the researchers have, and how they address their research questions. 2. Research Method and Analysis Approach Our investigation of the CER literature entailed the development of a purpose built classification scheme. We based our scheme on an existing one designed by Vessey, Ramesh and Glass [40]; even so, developing and implementing it proved to be a non-trivial task. The two principal authors developed a draft classification scheme of seven dimensions and presented it to the other authors at a workshop. Particular papers were discussed in depth at the workshop, culminating in consensus classifications of those papers in all seven dimensions. Participants then classified a trial set of papers which were discussed at two further workshops, resulting in a number of clarifications and several modifications to the scheme. A second trial set of papers was then classified and discussed at a fourth workshop. All classifications after the initial round were done in pairs, with each pair classifying a set of papers individually, meeting to compare and discuss, and then submitting an agreed list of classifications. Previous work by Simon et al [35] found classifying in pairs to be more reliable than classifying singly, as it helped to eliminate simple errors and focused discussion on papers that were genuinely difficult to classify. Following the initial classifications we conducted an inter-rater reliability test, using the Fleiss-Davies kappa [11] for the case where all of the raters are rating all of the items. To be amenable to such a test, a dimension must consist of a fixed set of categories and each item being classified should be placed into just one category. Not all of our dimensions meet these requirements, but we were able to adjust four of the seven dimensions for the purposes of the test. On this and other kappa measurements, an agreement of less than 40% is generally considered to be poor; between 40% and 75% is considered fair to good; and more than 75% is rated excellent [3]. With four pairs of raters rating the same set of 28 papers, our agreements on four dimensions were 68%, 60%, 55%, and 65%. These values are generally closer to the top than to the bottom of the fair to good range and suggest that in these dimensions we are classifying the papers fairly reliably. In the absence of a formal measure for the other three dimensions, we can only assume that our classification in those dimensions is as reliable as in the ones that we were able to test. Simon et al [35] reported inter-rater reliability of 55% to 80% on different dimensions when classifying in pairs. Vessey et al, classifying individually, reported reliability roughly at the 60% level in 2002 [15], improving to agreements in the 75%-85% range in 2005 [40]. Our results are comparable with those of Simon et al and with the early results of Vessey et al, suggesting that our system is comparable in reliability with theirs. It seems reasonable, too, to expect that our inter-rater reliability will improve as we become more familiar with our scheme and its application. We now considered that the classification scheme was stable and we had sufficient agreement and understanding of the scheme to begin our study. From this point onward, each pair classified a different set of papers. 2.1 Selection of Papers for Analysis While CER papers are found in many journals and conference proceedings, for this study we decided to focus on long papers, as opposed to the 5-page papers found in SIGCSE and ITiCSE. The extra length of an 8-to-12-page paper encourages discussion of theoretical background and methodological issues as well as presentation of experimental results. Longer papers can also allow reporting of qualitative studies that typically use more space than quantitative papers to explain their background and approach. This decision limited our initial pool of data sources to journals, ICER, the Australasian Computing Education Conference, and Koli Calling. We decided to apply our classification scheme to an analysis of papers from the International Computing Education Research workshop (ICER). In the five years it has been running, ICER has clearly established itself as a leading forum for CER. A previous analysis of ICER papers [35] found that ICER had the highest percentage of research papers among computing education conferences. The call for papers for ICER states that papers for the ICER workshop should, as appropriate, display A clear theoretical basis, drawing on existing literature in computing education or related disciplines and a strong empirical basis, drawing on relevant research methods. Such expectations are not made explicit for other computing education conferences such as SIGCSE and ITiCSE. Our study will thus offer some insight into how these instructions have been followed in the papers accepted to ICER. Obviously we cannot say anything about the papers not accepted to the conference, though analyzing such data would be interesting from the point of view of investigating the review process and what kind of impact the above instructions have on the acceptance decisions. 3. CLASSIFICATION SCHEME The classification scheme we have devised has seven dimensions, enabling the classification of computing education papers based on Theory/Model/Framework/Instrument, Technology/Tool, Reference Discipline, Research Purpose, Research Framework, Data Source and Analysis Method. In developing our classification scheme we drew upon the work of Vessey, Ramesh and Glass [40], who devised a scheme for the classification of research literature across the different disciplines in the computing field [15, 26, 39]. Our intended focus had a great deal in common with this scheme, so we adopted it, but modified it to give specific details of linkages to other disciplines and more detail of the research methods. From the Vessey et al [40] scheme of five dimensions we used the dimensions of reference discipline, research approach and research method. We modified slightly their categories for research approach and renamed this dimension Research Purpose as we felt this better describes the categories in this dimension. We considered their research method dimension too broad and so divided it into Research Framework, Analysis Method and Data Source. To identify specific ways in which the reported work connects to other disciplines we added the dimensions of Technology/Tool and Theory/Model/Framework/Instrument. Our scheme was designed to be used in conjunction with that of Simon et al [35], whose dimensions of Theme and Scope were very similar to the Topic and Level of Analysis of Vessey et al. Therefore we chose not to use the latter two dimensions. However, 4

Table 1: Descriptive, Evaluative, and Formulative categories of the Research Purpose dimension Descriptive-information/ human system (DI) Descriptive-technical system (DT) Descriptive-other (DO) Evaluative-positivist (EP) Evaluative-interpretive (EI) Evaluative-critical (EC) Evaluative-other (EO) Formulative-concept (FC) Formulative-model (FM) Formulative-process, method, algorithm (FP) Formulative-standards (FS) Describing an existing information system or human system, e.g. a classroom environment, teaching approach or course offering. Describing an existing technical system or implementation. This may include rationale for development, explanation of functionality and technical implementation. Describing something other than an existing technical/information/human system, e.g. an opinion piece, a description of a proposed system or a paper whose purpose is to review or summarize the literature. Research that takes the positivist perspective of an objective reality that can be measured and determined. Typically employing quantitative methods and deductive in nature; however, data mining, which is often inductive in nature, will also be included in this category. Research that takes the interpretive perspective that reality is subjective. Typically employing qualitative methods, interpretive research attempts to understand phenomena from the subjective meanings that participants assign to them. May be inductive or deductive in nature. Research that takes the critical perspective. The purpose of critical research is to understand and expose false beliefs or ideas for the purpose of explaining and perhaps transforming social systems. An example of a paper with this research purpose is Schwartzman [31]. Research with an evaluative component that does not belong in any of the other Evaluative subcategories above, e.g. an opinion piece survey that does not use a construct-based questionnaire [40]. The paper formulates a novel concept. The paper develops a model, which is a simplified view of a complex reality. The model may be usable as a framework (a basic conceptual structure used to solve or address complex issues) or as a classification or taxonomy. The paper formulates a process, method or algorithm for accomplishing something (other than quality evaluation, which is covered by FS below). The paper formulates guidelines, standards or criteria for evaluating the quality of something. It may also formulate a process for using the guidelines, standards or criteria. as most of the ICER papers have already been classified according to Simon s scheme [35], these two dimensions and two more of Simon s have been omitted from the analysis presented here. The remainder of this section explains the dimensions that we used and the specific categories within each dimension, and concludes with illustrative classifications of a handful of papers. 3.1 Theory / Model / Framework / Instrument This dimension is used to show linkages to prior work. Here we list the TMFI theories, models, frameworks and instruments (established questionnaires) used in the research reported in the paper. Note that we include only those TMFI that are used explicitly rather than just referred to in motivating or positioning the work. For example, we include TMFI that are applied to a research design or used in an analysis, and we include TMFI that are modified or extended by the reported work. We do not include TMFI that are developed in the paper itself. TMFI may range from those that are well known and identifiable by name to those that are unnamed and the result of a single study, such as a theory generated from a phenomenographical study. Note that we do not include technical frameworks (e.g. a Model-View- Controller GUI framework) or research frameworks (e.g. phenomenography) in this dimension. The former will often be represented in the Technology/Tool dimension (section 3.2) and the latter in the Research Framework dimension (section 3.5). 3.2 Technology / Tool Here we list any technologies or tools (TT) used in the work reported in the paper; for example, a visualization tool, a class library, software or hardware. As with the TMFI dimension, we are looking for an explicit level of use which is more than just a reference to the technology or tool, and we do not include TT that are developed in the paper itself. We only include technologies whose use is significant for the purpose of the paper. For example, if a paper investigates the effects of teaching Java in CS1 using BlueJ as the IDE, BlueJ will be listed as a TT if its features are explicitly relevant to the research, but not if the use of BlueJ appears to be incidental. 3.3 Reference Discipline In this dimension we list the disciplines that the reported work is linked to through use of theories, models, frameworks, instruments, technologies or tools. These are commonly termed reference disciplines [40]. All of the papers we analyze are positioned in computing education research, so we only note reference disciplines outside of computing education. 3.4 Research Purpose The research purpose is concerned with the goals of the research. This dimension is based on the research approach dimension of Vessey et al [39], which was developed from the work of Morrison and George [20]. The three categories are: Descriptive description of a tool, technology or system. This may involve detailed explanation of features, functionality and rationale for development. Evaluative assessment of a tool, method or situation, typically through a systematic process involving data gathering, analysis and reporting. This may involve hypothesis testing and may be exploratory or investigative in nature. Formulative development and/or refinement of a theory, model, standard, or process, or proposition of a new concept. 5

Table 2: Research Framework dimension Action Research (AR) A self-reflective systematic inquiry undertaken by participants to improve practice. Typically conducted as an iterative cycle of planning, action, change, reflection [13, 41]. Case Study (CS) In-depth, descriptive examination conducted in situ, usually of a small number of cases/examples [13]. Constructive Research (CR) Research that aims to demonstrate and/or evaluate the feasibility of a proposed idea (concept implementation; proof-of-concept research). Revolves around the development of, e.g., software, technology, a teaching approach, or an evaluation instrument. Delphi Seeking consensus by showing a group of raters a summary of their ratings, with justifications, then iteratively inviting them to reconsider their ratings in the light of what the others have said [35]. Ethnography (Eth) A branch of anthropology that deals with the scientific description of individual cultures [41]. Experimental Research (Exp) Quantitative research based on manipulating some variables while varying and measuring others. This requires formation of control and experimental groups of participants with random assignment of participants or use of naturally formed groups. Grounded Theory (GT) Qualitative, data-driven research in the tradition of Glaser and/or Strauss [37] which aims to formulate theories or hypotheses based on data. Phenomenography (PhG) Investigation of the significant variations in the ways that people experience a phenomenon (2 nd order perspective). Phenomenology (PhL) Investigation of the richness and essence of a phenomenon by studying one s own or others experiences of it (1 st order perspective). Survey Research (Survey) Quantitative research based on exploring the incidence, distribution and/or relationships of variables in nonexperimental settings. Typically this will involve synthesis and integration of information and argumentation. The subcategories we use within each category are slightly modified from those specified in Vessey et al [39]. We have expanded the subcategories of the Descriptive category to distinguish between technical systems and information/human systems. We have also consolidated three subcategories in the Formulative category as we found it difficult to distinguish between the formulative categories of model, framework and taxonomy/classification. The use of these terms depends on the perspective of the researcher: a taxonomy may be considered a model and may also be used as a framework for further research. The categories of this dimension are explained in Table 1. For every paper we identify at least one research purpose. As many papers report studies with several parts, we also list further research purposes where appropriate. 3.5 Research Framework A research framework (see Table 2) is an overall orientation or approach that guides or describes the research, as opposed to a specific method or technique. A research framework may have associated theoretical, epistemological, and/or ontological assumptions (e.g. phenomenography), may prescribe or suggest the use of particular methods (e.g. grounded theory), or may simply be a descriptive term for a kind of research activity that has certain characteristics (e.g. action research, case study). Not all papers will have a research framework. 3.6 Data Source The data source dimension describes the nature of the data and how it was gathered in the reported research. Most research papers will have at least one data source (see Table 3). 3.7 Analysis Method An analysis method describes how empirical data was analyzed or what other means were used to draw conclusions. Practically all papers will have at least one kind of analysis method. Table 4 lists the categories in this dimension. If a paper has a research framework, that framework might well direct the analysis method that is used; but the same analysis method can of course be found in a paper that is not applying a specified research framework. 3.8 Sample classifications A brief illustration of the application of our scheme will examine five papers from ICER 2008, the first set of papers that we classified. Bennedsen et al [4] is a paper with three distinct TMFI: the theory of cognitive development, the pendulum test, and SOLO. It has a reference discipline of psychology, and a single research purpose of evaluative positivist. Its research framework is survey, its data sources are research specific and naturally occurring data, and its analysis method is statistical analysis. Chinn et al [9] has no TMFI, does not report on a technology/tool, and has no reference discipline. It has research purposes of evaluative positivist and evaluative interpretive. Its framework is experimental research, its data source is research specific data, and it uses three analysis methods: interpretive qualitative analysis, statistical analysis, and interpretive classification. Denny et al [12] has no TFMI and no reference discipline, but reports on the technology/tool of Peerwise. Its single research purpose is evaluative positivist, its research framework is survey, its data source is naturally occurring data, and its analysis method is statistical analysis. Moström et al [21] has the TMFI of threshold concepts, with education as reference discipline. Its research purpose is evaluative interpretive. It has no research framework, its data source is research specific data, and its analysis method is interpretive classification. 6

Table 3: Data Source dimension Naturally occurring data (Nat) Research specific data (Res) Reflection (Ref) Software (Sw) Data pertaining to human subjects that would be available regardless of whether the research was carried out. This includes examination results, course grades, task submissions, student/teacher emails, course documentation, and published literature. It also includes data from previous research that is re-analyzed. Data collection pertaining to human subjects tailored specifically for the needs of the research. This includes interviews, questionnaires, observation data and any other data collected from assignments or tasks designed specifically for the target research. The researchers own reflections and experiences of a phenomenon serve as data. These reflections are what the paper deals with (they are analyzed and/or described); e.g. phenomenological research. Data collected about a software system. The data primarily describes the software rather than the humans using it or their activities; e.g. performance benchmarking, software metrics. Simon et al [35] has Simon s system as TMFI, with no technology/tool or reference discipline. Its single research purpose is descriptive other, it uses the Delphi research framework, its data source is naturally occurring (in the literature), and its analysis methods are descriptive statistics and interpretive classification. 4. RESULTS In total we analyzed all 72 ICER papers published to date, with the following numbers per year: 2005: 16 papers; 2006: 13; 2007: 14; 2008: 16; 2009: 13 papers. In this section we present the results for each dimension. 4.1 Theories, Models, Frameworks and Instruments We deliberately limited the TMFI to those explicitly mentioned and applied in the paper we were classifying. A secondary reference, a TMFI mentioned in the papers referenced by the paper we were considering, might have been useful to note, but might also misrepresent the research we were classifying. We did not distinguish between theories, models and frameworks, as the definitions tend to overlap, and the terms themselves can be used inconsistently in different papers. We also decided not to question how faithfully a TMFI had been used in the work being reported, as such criticism would be beyond the scope of this work. This also holds for methodological issues, discussed later. In 41 of the 72 papers we identified that the work had been built on some previous work by others in the sense that some explicit theories, models, frameworks and instruments (TMFI) were used in the work. In two further cases the work was built on such TMFI developed by one or more of the authors themselves [32, 35]. The remaining 29 papers, 40% of the data, did not apply any TMFI either from the authors or from others. These included wholly standalone works presenting novel research and a fair number of papers building on previous technical work, such as evaluating the effect on learning outcomes of using some tool. Many papers also extended or built on such results from their authors previous work or from other authors results. Identifying such work was rather challenging, as few papers emphasize the TMFI they are using. While some papers might mention them explicitly in the methodology or discussion sections, others might discuss them just briefly in a section on related work, then proceed to use them with no further mention. A general observation was that most TMFI identified in the data were not classical theories from education, such as constructivism, behaviorism or cognitive learning. They were much more specific, and in many cases referred to the theory by reference to an author rather than by its recognized name. There was great diversity. We listed in total 78 instances of TMFI, of which 68 were distinct. These numbers are clearly subject to interpretation, as the decision to list a TMFI was in many cases not clear-cut. However, they clearly indicate that a large and diverse pool of external work is being used in CER research. Argumentation (Arg) Conceptual Analysis (CA) Descriptive Statistics (DS) Exploratory Statistical Analysis (ESA) Interpretive Classification (IC) Interpretive Qualitative Analysis (IQA) Mathematical Proof (MP) Statistical Analysis (SA) Table 4: Analysis Method dimension Conclusions are reached through arguments presented by the author(s). As all research papers include some argumentation, this method is recorded only for works where the discussion forms a substantial part of the presentation of results. Breaking down or analyzing concepts into their constituent parts in order to gain knowledge or a better understanding of the concepts themselves. Only single-variable descriptive statistics (distributions, means, medians, etc.) are presented without checking for relationships between variables. (When intervariable relationships beyond cross-tabulations are investigated in a paper, we code it as ESA or SA rather than DS.) Statistics exploring variables and relationships but without checking statistical significance. Includes factor analysis, statistical regression, data mining techniques such as clustering, and correlation. Papers with ESA will also often include DS (see above), which we do not report separately. Classifying data based on an existing categorization (which might be refined during the analysis). Systematic data-driven formation of qualitatively different categories (e.g. interpretive content analysis or a typical phenomenographic or grounded theory study). Deductive reasoning that produces a proof without using empirical data. Statistical tests are used to check for relationships or differences between variables. This will include some form of statistical analysis beyond descriptives. Control and experimental groups may be used, but are not necessary. Papers with SA will also often include DS (see above), which we will not report separately 7

Some of the well-known theories, models and frameworks that we identified are: Bloom s taxonomy [7] self-efficacy theory [2] cognitive apprenticeship [10] situated learning [17] cognitive load theory [1, 22] SOLO taxonomy [6] general systems theory [5] threshold concepts [19] schema theory [18] Very few papers used instruments, established questionnaires that measure specific behaviors or characteristics. Examples of those found include the pendulum test [8] to measure abstraction ability, the test of emotional intelligence [30], and the motivated strategies for learning questionnaire (MSQL) [25]. 4.2 Technologies and Tools Tools research is an important part of CER. We found 12 tools or technologies that played an important role in the research of 12 papers. These were Agar, ALVIS Live!, Aropä, BlueJ, Classroom presenter, Jeliot, Marksense, Peerwise, PlanAni, SALSA, tablet PCs, and WebCAT. 4.3 Reference Disciplines The reference disciplines were determined by noting the TMFI used that were not developed in computing education research. The following reference disciplines emerged from the data (the number in parentheses is the number of papers linking to each reference discipline). Biology (1) Philosophy of science (1) Computing (11) Psychology (12) Education (11) Sociology (1) Human-computer interaction (1) Systems theory (1) Mathematics (1) A further 12 papers were linked to the broader discipline of computing through use of a technology or tool. We conclude that computing education research does indeed make wide use of works from other disciplines, with the key disciplines being computing, education and psychology. 4.4 Research Purpose For each paper we identified one or more research purposes. In 35 papers we found one research purpose, in 33 papers, two research purposes, and in four papers, three purposes. In some papers where we had identified a secondary purpose, it was not clear whether it was sufficiently substantial to warrant inclusion. For example, a large quantitative study might include a brief qualitative mention of student feedback. Table 5: Distribution of Research Purposes (72 papers) Research Purpose Papers Descriptive-information/human system (DI) 10 (14%) Descriptive-technical (DT) 2 (3%) Descriptive-other (DO) 4 (6%) Evaluative-positivist (EP) 36 (50%) Evaluative-interpretive (EI) 29 (40%) Evaluative-critical (EC) 1 (1%) Evaluative-other (EO) 5 (7%) Formulative-model (FM) 15 (21%) Formulative-process,method,algorithm (FP) 8 (11%) Formulative-standards (FS) 3 (4%) The decision was therefore necessarily subjective. In Table 5 we list the distribution of all research purposes found in the papers. Evaluative purposes are found in a clear majority of the papers: 62 papers (86%) involve some evaluative aspect. More than a third of the papers (36%) had a formulative aspect, and 22% of papers had some descriptive purpose. Considering these more closely, we note that all DI papers (describing information / human systems) also had a secondary purpose. In seven cases this was evaluating the system (EI/EP), and in three cases formulating a process (FP). Using the classification scheme developed by Simon [35], two of the latter were considered position papers or proposals rather than research papers. There was one pure literature summary [35]. Of the two technical descriptive papers (DT), one also formulated a process and one was a pure technical research work. We conclude that ICER papers have a clear focus on evaluative work. 4.5 Research Framework We identified a research framework in 57 of the papers (79%). Of these, 47 had a single framework, nine had two, and one had three frameworks. Fifteen papers did not use a framework. The distribution of frameworks was very uneven, as we expected. In greatest numbers were survey frameworks (28 papers, 39% of all papers), experimental frameworks (11, 15%), constructive research (10, 14%) and grounded theory (9, 13%). Smaller counts included phenomenography (4 papers), case studies (3), Delphi method (1), ethnography (1) and phenomenology (1). The papers with no clear framework tended to use some qualitative analysis method or content analysis. 4.6 Data Source As might be expected, the majority of papers (79%) had research as one data source, indicating that the data was specifically collected for the research being reported. Twelve papers (16%) used naturally occurring data in combination with research-specific data, while six used natural data alone. Eight papers had literature as their data source, and one paper used reflection. In one technical paper we could not identify any data source, because the paper simply described a tool. 4.7 Analysis Method With the exception of one technical paper, all papers applied some method of analysis. In 24 papers we identified the use of two different analysis methods; in four papers, three methods; and in one paper, four methods. The most common method used was some kind of statistical analysis (SA) with tests of significance (30 papers, 42%). Exploratory statistical analysis (ESA) was used in 12 papers (17%), and purely descriptive statistics (DS) in eight papers (11%). Qualitative methods were widely used. In 25 papers (35%) we identified interpretive qualitative analysis (IQA), meaning some data-driven categorization with no predefined categories. Some form of interpretive classification or content analysis (IC) was used in 19 papers (26%). Here some predefined categorization scheme was either applied as such or refined on the basis of the data. Finally, in 12 papers we considered that argumentation (Arg) was an analysis method. We recognize that all papers have some form of argumentation, but we listed Arg only in cases where argumentation clearly played a strong role in reaching the conclusion, perhaps supported by data and/or theory. 8

4.8 Validity and Reliability Issues in the Classification Several of the authors had conducted analysis and classification of literature in earlier studies [23, 24, 28, 32-36]. However, during the process of developing this classification scheme we soon realized the complexity of the task, as we were looking more deeply into the research process than previous studies had done. There were several difficulties. We wanted to identify which TMFI had really been used in a paper, rather than just referred to as related work. Many papers did not report this clearly, and we had to read the papers carefully to decide whether something should be included. In a number of cases the decision had to be negotiated between the classifiers before consensus was found. Working in pairs was a clear advantage in this regard. We recognize that we may still find individual papers for which we have wrongly decided whether to include some TMFI, but we are confident that the big picture is reasonably accurate, and this is our goal. In this kind of research, the scale of numbers is more important than absolute precision. The same observation holds to some extent for the other dimensions. Reference disciplines were deduced from where the TMFI was published, and in some cases we were unable to check the background of the originators of a TMFI. The main research purpose was clear in most cases, but for some papers the decision to include additional research purposes had to be negotiated. Research framework was quite clear in most cases, as were data source and analysis method. The inclusion of argumentation as a research method was invariably a matter for negotiation. But notwithstanding these difficult issues, we are confident that the big picture is correct, and that our classification scheme is a useful and usable tool. 5. DISCUSSION Previous literature studies, as mentioned above, have mapped the CER field in some detail, answering questions such as what has been done and what kind of work has been carried out in CER. Our goal in this study was to delve deeper, exploring how research in CER is being carried out. Huge numbers of papers have been published in computing education conferences and journals in the past 40 years, so we would expect much to have been learnt about teaching and learning in computing. However, after decades of research, we still have only a vague understanding of why it is so difficult for many students to learn programming, the basis of the discipline, and consequently of how it should be taught. Sheard et al [32] analyzed a wide pool of papers on programming education and expressed a... need to explore the process of learning in relation to theories and models of learning. It was a concern that we didn t find many studies that investigated learning within such a theoretical framework. Therefore it is natural to ask whether there might be unknown problems in the research itself that could explain our slow progress. Does it have enough focus on evaluation, or is the dominating purpose describing or formulating new teaching practices and tools? Does it apply knowledge from related fields, which have a much longer history of research, rather than reinventing the wheel? In their wide analysis of computing research, Glass, Ramesh and Vessey [14] found that in Computer Science (CS) 79% of papers had a formulative research purpose (which they called research approach), 11% had an evaluative purpose and 10% had a descriptive purpose. In Software Engineering (SE) and Information Systems (IS) the numbers were 55%, 14%, 28% and 24%, 67%, 9% respectively. As many CER researchers have a background in computer science, we might expect the strong tradition of formulative work to have some influence on their work in CER. However, our analysis of ICER papers fails to support this expectation. The corresponding results were 18% formulative, 71% evaluative, and 11% descriptive work. We conclude that the CER presented at ICER is much closer in purpose to IS research than to CS/SE research. Does CER build on previous work by disciplines such as Education and Psychology? We examined the reference disciplines on which the research was based. Glass et al [14] found that only 11% of CS papers and 2% of SE papers built on theories from other disciplines, whereas in IS 73% of papers used theories outside computing. In our findings, 44% of CER papers in our data pool built their research on TMFI from other disciplines. If we count only disciplines outside Computing, we still have 35% of papers applying TMFI from these disciplines. Thus the CER presented at ICER seems to lie about halfway between CS/SE and IS. We should be cautious about generalizing these observations. We have analyzed only papers from ICER, a small and young conference, which, according to previous research [35], has a larger percentage of research papers than many other CER conferences. To get a broader picture of CER we will need to look at journals and other conferences, which may give us a different picture. Casting back to the ICER CFP requirement that papers should have a clear theoretical basis, drawing on existing literature in computing education or related disciplines, we conclude that this requirement has been well met. In a clear majority of papers some TMFI from related disciplines was identified. This is a very good sign, considering that we required the TMFI to be used or extended, not just referred to. Moreover, we were impressed by the diversity of TMFI used, which suggests that the authors really seem to look for TMFI that fit their research contexts. We also investigated papers for which we did not find any TMFI, but we found no clear characteristics to explain this lack. A few of them built on some Tool/Technology, and thus on previous work from computing, while others simply presented novel work. It should be recalled that we did not record as a TMFI a theory, model, framework, or instrument actually developed within the paper we were classifying, so some of these papers could well have been developing TMFI. In addition, it is presumably possible to have a clear theoretical basis, drawing on existing literature, without applying an explicit theory, model, framework, or instrument from that literature. The other ICER requirement was that papers should have a strong empirical basis, drawing on relevant research methods. Again we could confirm that the goal has been well met. Empirical evaluation has been widely applied, with 86% of papers having at least one evaluative research purpose. The richness of the conference is demonstrated by the observation that both evaluative positivist and evaluative interpetive purposes were widely applied, indicating general acceptance of both quantitative and qualitative evaluation papers. In summary it seems that ICER does indeed achieve its stated goals. We should now look at other publishing forums to form a broader picture of the CER field. 6. ACKNOWLEDGMENTS Our thanks to Dr Tony Clear for his advice on classification of critical research papers. 9

REFERENCES [1] Atkinson, R. K., Derry, S. J., Renkl, A., and Wortham, D., Learning from Examples: Instructional Principles from the Worked Examples Research, Review of Educational Research, vol. 70, pp. 181-214, 2000. [2] Bandura, A., Self-efficacy, in Encyclopedia of Human Behavior. vol. 4, V. S. Ramachaudran, Ed. NY: Academic Press, 1994, pp. 71-81. [3] Banerjee, M., Capozzoli, M., McSweeney, L., and Sinha, D., Beyond kappa: a review of interrater agreement measures, Canadian Journal of Statistics, vol. 27, pp. 3-23, 1999. [4] Bennedsen, J. and Caspersen, M. E., Abstraction ability as an indicator of success for learning computing science, in 4th Sydney, Australia, 2008, pp. 15-25. [5] Bertalanffy, L. V., The History and Status of General Systems Theory, The Academy of Management Journal, vol. 15 pp. 407-426, 1972. [6] Biggs, J. B., Teaching for quality learning at university. what the student does. Maidenhead, UK: Open University Press, 2003. [7] Bloom, B. S., Mesia, B. B., and Krathwohl, D. R., Taxonomy of Educational Objectives (two vols: The Affective Domain & The Cognitive Domain) Addison-Wesley, 1956. [8] Bond, T. G., Piaget and the Pendulum, Science and Education, vol. 13, pp. 389-399, 2004. [9] Chinn, D. and VanDeGrift, T., Gender and Diversity in Hiring Software Professionals: What Do Students Say?, in 4th Sydney, Australia, 2008, pp. 39-50. [10] Collins, A. M., Brown, J. S., and Holum, A., Cognitive apprenticeship: Making thinking visible, American Educator, vol. 15, pp. 6-11, 38-46, 1991. [11] Davies, M. and Fleiss, J. L., Measuring agreement for multinomial data., Biometrics, vol. 38, pp. 1047-1051, 1982. [12] Denny, P., Hamer, J., and Luxton-Reilly, A., PeerWise: Students Sharing their Multiple Choice Questions, in 4th Sydney, Australia, 2008, pp. 51-58. [13] Fincher, S. and Petre, M., Computer Science Education Research, Netherlands: Taylor & Francis 2004, p. 239. [14] Glass, R. L., Ramesh, V., and Vessey, I., An analysis of research in computing disciplines, Communications of the ACM, vol. 47, pp. 89-94, 2004. [15] Glass, R. L., Vessey, I., and Ramesh, V., Research in software engineering: An analysis of the literature, Information and Software Technology, vol. 44, pp. 491-506, 2002. [16] Joy, M., Sinclair, J., Sun, S., Sitthiworachart, J., and López- González, J., Categorising computer science education research, Education and Information Technologies, vol. 14, pp. 105-126, 2009. [17] Lave, J. and Wenger, E., Situated Learning: Legitimate Peripheral Participation. Cambridge, UK: Cambridge University Press, 1991. [18] Marshall, S. P., Schemas in problem solving. New York: Cambridge University Press, 1995. [19] Meyer, J. H. and Land, R., Threshold concepts and troublesome knowledge (2): Epistemological considerations and a conceptual framework for teaching and learning, Higher Education, vol. 49, pp. 373-388, 2005. [20] Morrison, J. and George, J. F., Exploring the software engineering component in MIS research, Communications of the ACM, vol. 38, pp. 80-91, 1995. [21] Moström, J. E., Boustedt, J., Eckerdal, A., McCartney, R., Sanders, K., Thomas, L., and Zander, C., Concrete examples of abstraction as manifested in students' transformative experiences, in 4th International Workshop on Computing Education Research, Sydney, Australia, 2008, pp. 125-136. [22] Paas, F., Renkl, A., and Sweller., J., Cognitive load theory and instructional design: recent developments, Educational Psychologist, vol. 38, pp. 1-4, 2003. [23] Pears, A., Seidman, S., Eney, C., Kinnunen, P., and Malmi, L., Constructing a core literature for computing education research, SIGCSE Bulletin, vol. 37, pp. 152-161, 2005. [24] Pears, A., Seidman, S., Malmi, L., Mannila, L., Adams, E., Bennedsen, J., Devlin, M., and Paterson, J., A survey of literature on the teaching of introductory programming, SIGCSE Bulletin, vol. 39, pp. 204-223, 2007. [25] Pintrich, P. R., Smith, D., Garcia, T., and McKeachie, W., A manual for the use of the motivated strategies for learning questionnaire, University of Michigan, technical report 91-b- 0041991. [26] Ramesh, V., Glass, R. L., and Vessey, I., Research in computer science: An empirical study, The Journal of Systems and Software, vol. 70, pp. 165-176, 2004. [27] Randolph, J., Multidisciplinary methods in educational technology research and development: Published by HAMK Press/Justus Randolph, 2008. [28] Randolph, J., Bednarik, R., and Myller, N., A methodological review of the articles published in the proceedings of Koli Calling 2001-2004, in 5th Baltic Sea Conference on Computer Science Education, 2005, pp. 103-109. [29] Randolph, J. J., Findings from A Methodological Review of the Computer Science Education Research: 2000-2005, in SIGCSE Bulletin, 2007, pp. 130-130. [30] Schutte, N. S., Malouff, J. M., Hall, L. E., Haggerty, D. J., Cooper, J. T., Golden, C. J., and Dornheim, L., Development and validation of a measure of emotional intelligence, 1998. [31] Schwartzman, L., On the nature of student defensiveness: theory and feedback from a software design course, in 5th 2009, pp. 81-91. [32] Sheard, J., Simon, Hamilton, M., and Lönnberg, J., Analysis of research into the teaching and learning of programming, in 5th 2009, pp. 93-104. [33] Simon, A classification of recent Australasian computing education publications, Computer Science Education, vol. 17, pp. 155-170, 2007. [34] Simon, Koli Calling comes of age: An analysis, in 7th Baltic Sea Conference on Computing Education Research, 2008, pp. 119-126. [35] Simon, Carbone, A., de Raadt, M., Lister, R., Hamilton, M., and Sheard, J., Classifying computing education papers: Process and results, in 4th International Workshop on Computing Education Research, Sydney, Australia, 2008, pp. 161-172. 10

[36] Simon, Sheard, J., Carbone, A., de Raadt, M., Hamilton, M., Lister, R., and Thompson, E., Eight years of computing education papers at NACCQ, in 21st Annual Conference of the National Advisory Committee on Computing Qualifications, 2008, pp. 101-107. [37] Suddaby, R., From the editors: What grounded theory is not, Academy of Management Journal, vol. 4, pp. 633-642, 2006. [38] Valentine, D. W., CS educational research: A meta-analysis of SIGCSE technical symposium proceedings, in 35th SIGCSE Technical Symposium on Computer Science Education, 2004, pp. 255-259. [39] Vessey, I., Ramesh, V., and Glass, R. L., Research in information systems: An empirical study of diversity in the discipline and its journals, Journal of Management Information Systems, vol. 19, pp. 129-174, 2002. [40] Vessey, I., Ramesh, V., and Glass, R. L., A unified classification system for research in the computing disciplines, Information and Software Technology, vol. 47, pp. 245-255, 2005. [41] Wiersma, W., Research Methods in Education: An Introduction, 8th ed. MassachuseTT: Allyn and Bacon, 2005. 11