Identifying Threats to Validity and Control Actions in the Planning Stages of Controlled Experiments

Similar documents
Running head: METACOGNITIVE STRATEGIES FOR ACADEMIC LISTENING 1. The Relationship between Metacognitive Strategies Awareness

Research Design & Analysis Made Easy! Brainstorming Worksheet

Towards a Collaboration Framework for Selection of ICT Tools

Practical Research. Planning and Design. Paul D. Leedy. Jeanne Ellis Ormrod. Upper Saddle River, New Jersey Columbus, Ohio

School Size and the Quality of Teaching and Learning

MASTER OF ARTS IN APPLIED SOCIOLOGY. Thesis Option

CS Machine Learning

TU-E2090 Research Assignment in Operations Management and Services

12- A whirlwind tour of statistics

Instructor: Mario D. Garrett, Ph.D. Phone: Office: Hepner Hall (HH) 100

The Effect of Written Corrective Feedback on the Accuracy of English Article Usage in L2 Writing

A cognitive perspective on pair programming

An Empirical Analysis of the Effects of Mexican American Studies Participation on Student Achievement within Tucson Unified School District

Experiences Using Defect Checklists in Software Engineering Education

Experience and Innovation Factory: Adaptation of an Experience Factory Model for a Research and Development Laboratory

The Implementation of Interactive Multimedia Learning Materials in Teaching Listening Skills

Improving software testing course experience with pair testing pattern. Iyad Alazzam* and Mohammed Akour

The digital copy of this thesis is protected by the Copyright Act 1994 (New Zealand).

Tun your everyday simulation activity into research

Systematic reviews in theory and practice for library and information studies

CHALLENGES FACING DEVELOPMENT OF STRATEGIC PLANS IN PUBLIC SECONDARY SCHOOLS IN MWINGI CENTRAL DISTRICT, KENYA

What is Thinking (Cognition)?

BSM 2801, Sport Marketing Course Syllabus. Course Description. Course Textbook. Course Learning Outcomes. Credits.

Quantitative analysis with statistics (and ponies) (Some slides, pony-based examples from Blase Ur)

Preparing a Research Proposal

Strategic Planning for Retaining Women in Undergraduate Computing

STUDENT SATISFACTION IN PROFESSIONAL EDUCATION IN GWALIOR

GUIDE FOR THE WRITING OF THE DISSERTATION

PROJECT MANAGEMENT AND COMMUNICATION SKILLS DEVELOPMENT STUDENTS PERCEPTION ON THEIR LEARNING

Motivation to e-learn within organizational settings: What is it and how could it be measured?

Intro to Systematic Reviews. Characteristics Role in research & EBP Overview of steps Standards

Activities, Exercises, Assignments Copyright 2009 Cem Kaner 1

How to read a Paper ISMLL. Dr. Josif Grabocka, Carlotta Schatten

Deploying Agile Practices in Organizations: A Case Study

Thesis-Proposal Outline/Template

Word Segmentation of Off-line Handwritten Documents

University of Groningen. Systemen, planning, netwerken Bosman, Aart

Lecture 2: Quantifiers and Approximation

Unit 3. Design Activity. Overview. Purpose. Profile

Self Study Report Computer Science

Number of students enrolled in the program in Fall, 2011: 20. Faculty member completing template: Molly Dugan (Date: 1/26/2012)

Entrepreneurial Discovery and the Demmert/Klein Experiment: Additional Evidence from Germany

What is beautiful is useful visual appeal and expected information quality

Maximizing Learning Through Course Alignment and Experience with Different Types of Knowledge

Lecture 1: Machine Learning Basics

HAZOP-based identification of events in use cases

What is PDE? Research Report. Paul Nichols

USER ADAPTATION IN E-LEARNING ENVIRONMENTS

Patterns for Adaptive Web-based Educational Systems

A COMPARATIVE STUDY BETWEEN NATURAL APPROACH AND QUANTUM LEARNING METHOD IN TEACHING VOCABULARY TO THE STUDENTS OF ENGLISH CLUB AT SMPN 1 RUMPIN

ScienceDirect. Noorminshah A Iahad a *, Marva Mirabolghasemi a, Noorfa Haszlinna Mustaffa a, Muhammad Shafie Abd. Latif a, Yahya Buntat b

THE INFORMATION SYSTEMS ANALYST EXAM AS A PROGRAM ASSESSMENT TOOL: PRE-POST TESTS AND COMPARISON TO THE MAJOR FIELD TEST

The Impact of Test Case Prioritization on Test Coverage versus Defects Found

School Leadership Rubrics

On-the-Fly Customization of Automated Essay Scoring

An ICT environment to assess and support students mathematical problem-solving performance in non-routine puzzle-like word problems

Software Security: Integrating Secure Software Engineering in Graduate Computer Science Curriculum

Rule discovery in Web-based educational systems using Grammar-Based Genetic Programming

A Model to Detect Problems on Scrum-based Software Development Projects

PREVIEW LEADER S GUIDE IT S ABOUT RESPECT CONTENTS. Recognizing Harassment in a Diverse Workplace

Note: Principal version Modification Amendment Modification Amendment Modification Complete version from 1 October 2014

Task Tolerance of MT Output in Integrated Text Processes

Shank, Matthew D. (2009). Sports marketing: A strategic perspective (4th ed.). Upper Saddle River, NJ: Pearson/Prentice Hall.

ABET Criteria for Accrediting Computer Science Programs

Running head: DELAY AND PROSPECTIVE MEMORY 1

Probability and Statistics Curriculum Pacing Guide

Science Fair Project Handbook

The Impact of Formative Assessment and Remedial Teaching on EFL Learners Listening Comprehension N A H I D Z A R E I N A S TA R A N YA S A M I

SETTING STANDARDS FOR CRITERION- REFERENCED MEASUREMENT

The Effects of Jigsaw and GTM on the Reading Comprehension Achievement of the Second Grade of Senior High School Students.

CWIS 23,3. Nikolaos Avouris Human Computer Interaction Group, University of Patras, Patras, Greece

Analyzing the Usage of IT in SMEs

Language Acquisition Chart

Inquiry Learning Methodologies and the Disposition to Energy Systems Problem Solving

Learning Disabilities and Educational Research 1

CHEM 6487: Problem Seminar in Inorganic Chemistry Spring 2010

Match or Mismatch Between Learning Styles of Prep-Class EFL Students and EFL Teachers

The Round Earth Project. Collaborative VR for Elementary School Kids

South Carolina English Language Arts

MASTER S THESIS GUIDE MASTER S PROGRAMME IN COMMUNICATION SCIENCE

IMPORTANT GUIDELINE FOR PROJECT/ INPLANT REPORT. FOSTER DEVELOPMENT SCHOOL OF MANAGEMENT, DR.BABASAHEB AMBEDKAR MARATHWADA UNIVERSITY,AURANGABAD...

Linking the Common European Framework of Reference and the Michigan English Language Assessment Battery Technical Report

MSE 5301, Interagency Disaster Management Course Syllabus. Course Description. Prerequisites. Course Textbook. Course Learning Objectives

Mathematics Program Assessment Plan

Coordination Challenges in Global Software Development

This Performance Standards include four major components. They are

GDP Falls as MBA Rises?

Procedia - Social and Behavioral Sciences 98 ( 2014 ) International Conference on Current Trends in ELT

A Study of Metacognitive Awareness of Non-English Majors in L2 Listening

THE EFFECTS OF CREATIVE TEACHING METHOD ON MOTIVATION AND ACADEMIC ACHIEVEMENT OF ELEMENTARY SCHOOL STUDENTS IN ACADEMIC YEAR

State Parental Involvement Plan

TABLE OF CONTENTS TABLE OF CONTENTS COVER PAGE HALAMAN PENGESAHAN PERNYATAAN NASKAH SOAL TUGAS AKHIR ACKNOWLEDGEMENT FOREWORD

Space Travel: Lesson 2: Researching your Destination

Monitoring Metacognitive abilities in children: A comparison of children between the ages of 5 to 7 years and 8 to 11 years

NCEO Technical Report 27

ATW 202. Business Research Methods

and secondary sources, attending to such features as the date and origin of the information.

On the Combined Behavior of Autonomous Resource Management Agents

Georgetown University School of Continuing Studies Master of Professional Studies in Human Resources Management Course Syllabus Summer 2014

Transcription:

Identifying Threats to Validity and Control Actions in the Planning Stages of Controlled Experiments Amadeu Anderlin Neto and Tayana Uchôa Conte Instituto de Computação Universidade Federal do Amazonas (UFAM) Manaus, AM, Brazil {neto.amadeu, tayana}@icomp.ufam.edu.br Abstract During the planning phase of an experiment, we must identify threats to validity in order to assess their impact over the analyzed data. Furthermore, actions to address these threats must be defined, if possible, so that we can minimize their impact over the experiment s results. In this paper, we propose the Threats to Validity Assistant (TVA), which is a tool to assist novice researchers in identifying and addressing threats to validity of controlled experiments in Software Engineering. The TVA was evaluated through a controlled experiment which aimed at measuring the effectiveness and efficiency in identifying threats to validity and actions to address them. Sixteen subjects were asked to identify threats to validity and control actions related to an empirical study reported in literature. Our results indicate that the subjects who used TVA were significantly more effective and efficient than the subjects who did not use it. Keywords validity, controlled experiment, empirical study I. INTRODUCTION Empirical studies have long been applied to provide confidence in assertions about what is true or not true in the Software Engineering (SE) domain [12]. According to Wohlin et al. [14], empirical research is responsible for the maturation of SE. There are growing numbers of empirical studies to evaluate different SE practices and techniques [9]. New technologies should be compared with existing technologies through experimentation. Thus, we can collect evidence related to the performance of a technology (e.g., benefits, limitations, costs, risks). However, conducting experiments is a complex task [8] because researchers can cause changes in the experiment environment which could influence its outcomes and introduce threats that may compromise its validity. According to Biffl and Halling [4], every empirical study presents threats to validity (TTVs). These TTVs must be identified and addressed [11]. The identification and mitigation of TTVs are critical activities that require a lot of effort from researchers. TTVs are potential risks that may arise during the planning and execution of empirical studies [14]. Thus, experiments must be carefully planned and executed. Researchers in Empirical Software Engineering (ESE) have proposed guidelines, checklists, and summaries to support the identification and mitigation of TTVs [7]. However, none of these present relationships between TTVs and control actions (CAs). Novice researchers have difficulty in identifying and addressing TTVs during the experiment planning. Thus, such problem motivated us to build a conceptual model presenting relationships between TTVs and CAs [1]. The proposed conceptual model shows which CAs can be applied to address a TTV. In addition, it also shows the new TTV that may occur when applying a CA to address a certain TTV. As a way to facilitate the use of the conceptual model, we developed a tool, the Threats to Validity Assistant (TVA). The main idea is to assist novice researchers in identifying and mitigating TTVs. Therefore, we aim at enabling researchers to: (a) reduce the effect of the TTVs that can compromise the results of empirical studies; and (b) increase the degree of confidence in the obtained conclusions by enhancing the quality of the empirical studies. The scope of this work focused only in TTVs of controlled experiments. Kampenes [8] defines controlled experiment as a randomized experiment or a quasi-experiment in which individuals or teams (the experimental units) conduct one or more software engineering tasks for the sake of comparing different populations, processes, methods, techniques, languages, or tools (the treatments). In this paper, we present a controlled experiment which evaluated the TVA. This experiment was conducted in order to measure the effectiveness and efficiency in identifying TTVs and CAs. The quantitative results indicated that the experimental group had better performance than the control group. This paper is organized as follows. Our approach to assist the identification and mitigation of TTVs is described in Section II, while its evaluation follows in Section III. Then, we present the experiment s results in Section IV and, in Section V, we explore its TTVs. Finally, Section VI concludes the paper and describes future work. II. THE APPROACH TO IDENTIFY AND ADDRESS TTVS We followed a few stages to develop our approach which assists the identification and mitigation of TTVs. In the following subsections, we describe the conceptual model and the TVA tool. A. The Conceptual Model The conceptual model shows TTVs and CAs which were described in experiments reported in literature. We selected 256

these TTVs and CAs through a Systematic Literature Review (SLR). The steps for performing the SLR are described in [1]. We organized the conceptual model as follows. First, we grouped the TTVs with their respective CAs. After that, we grouped only TTVs and CAs which present trade-off. The trade-off is characterized when a CA is applied in order to control a TTV and, by applying such CA, new TTVs may arise. Figure 1 shows part of the conceptual model. The codes presented in each entity correspond to the ID of the TTV or CAs. These codes follow the structure: ValidityType- TypeOfEntity-Number. For instance, EXT-T01 means the entity is the first (01) TTV (T) to external validity (EXT) identified within the results from the SLR. This TTVs can be controlled by CA EXT-C21 (in which C means control). However, this CA may cause TTV COS-T03 (in which COS means conclusion validity). In order to address this TTV, COS- C04 can be applied. Nevertheless, CA COS-C04 may cause TTV EXT-T01. Figure 1. Trade-off between threats to external and conclusion validities. As shown in Figure 1, there is a trade-off between validity types [3]. When increasing one type, another type may be decreased [14]. The researchers are responsible for prioritizing the validity types according to their needs. Depending on the experiment, some TTVs are more critical than others [2]. The conceptual model allows a visualization of the ways to mitigate TTVs and consequences that may occur when the researcher chooses a CA, since relationships (trade-offs) between TTVs and CAs are mapped. As way to facilitate the use of the conceptual model, we have developed a tool, which is described below. Figure 2. (a) Part of the checklist page in the TVA; (b) Example of the threat list in the TVA. B. The Threats to Validity Assistant The Threats to Validity Assistant (TVA) was developed in order to facilitate the use of the conceptual model. We used the Grails 1 framework and the MySQL database to develop the tool. All relationships presented in the conceptual model were mapped to the database within the TVA. From the list of TTVs obtained from SLR, we derived a checklist. The checklist contains 65 items which aim to capture the experiment context and identify TTVs that may occur. Such items are related to one or more TTVs. Likewise, the TTVs are related to one or more items. Thus, there are relationships between items and TTVs. Besides the relationships between items and TTVs, there are two relationships between TTVs and CAs. The first one is a relationship in which a CA can be applied to address a TTV. Each TTV can has none, one or more CAs. Otherwise, a CA can control one or more TTVs. The second relationship characterize the trade-offs. A new TTV may arise when a specific CA is applied to control a TTV. Thus, one CA can cause none, one or more TTVs. Figure 2a shows part of the checklist page where we have selected the items The data collected can be inaccurate and There is the possibility of communication between the subjects during the experiment. Once the researchers fill in 1 http://www.grails.org/ the checklist, the next screen of the TVA shows the list of possible TTVs. For example, Figure 2b shows a list of TTVs which was produced according to the selected items in the previous page (Figure 2a). Researchers may select one or more CAs to address a TTV, or even none, according to the purpose of the experiment. For instance, Figure 3a shows the CAs to address the TTV Communication among subjects during the experiment. When saving the selected CAs, if a CA may cause a new TTV, a warning message appears for the researcher. If the researcher confirms the use of the CA, the new TTV will be included in the list of TTVs on the previous screen (Figure 2b). In our example, we selected Introduce observers in experimental environment to address the TTV Communication among subjects during the experiment. However, this CA can cause the TTVs Experimental environment representativeness and Subjects act differently when being observed. Thus, these TTVs will be included in the TTVs list as shown in Figure 3b. III. EXPERIMENT DESCRIPTION We have designed and executed a controlled experiment to answer the following research question: Is the identification of TTVs and CAs more efficient and effective when the TVA is used? To address this question, we have conducted a controlled experiment in order to measure the effectiveness and efficiency in identifying TTVs and CAs, with or without the use of the TVA. 257

Figure 3. (a) Example of the control actions list in the TVA; (b) Example of the new threats list after being modified by the researcher s actions. The definition, the planning (context, variables, hypotheses, selection of subjects, design, and instrumentation) and the operation of the experiment are explained in the following subsections. A. Experiment Definition The experiment was defined based on the template proposed in previous work [14] as: Analyze For the purpose of With respect to From the point of view In the context of Threat to Validity Assistant (TVA) characterize its effectiveness and efficiency of the researchers the identification of TTVs and CAs by students from an Empirical Software Engineering course. B. Context Selection The experiment context was an Empirical Software Engineering (ESE) course offered at UFAM (Federal University of Amazonas) at the north of Brazil. Although participating in the experiment is a mandatory part of the course, the subjects could request that their data were removed from the analysis. The subjects who agreed in participating in the experiment were 10 senior-level (fourth-year) undergraduate Computer Science students, and 6 graduated Computer Science students. The senior-level graduated students enrolled in the course aiming to be prepared at a masters or doctorate degree (more details about the subjects are presented in Subsection E). The controlled experiment used in this study was selected from the SLR and was based on the following criteria: (1) to describe more than 10 TTVs; (2) to contain TTVs of the four validity types; (3) to present different context used in the pretest activity; and, (4) to follow the experiment process described by Wohlin et al. [14]. Using these criteria, we selected the controlled experiment described by Madeyski [10] (hereafter referred to as Madeyski s experiment), which contains 23 TTVs and 12 CAs. The context of Madeyski s experiment was related to testing in software systems. C. Variables Selection The independent variable is the treatment applied by the groups in order to identify and address TTVs. Thus, the experimental group used the TVA while the control group used the bibliographical material of the Empirical Software Engineering course, mainly the guidelines described by Wohlin et al. [14]. We chose this book since it was the most cited reference (with about 49 citations within the SLR presented in [1]) in empirical studies to identify and address TTVs. The dependent variables are the effectiveness and efficiency in identifying TTVs and CAs. Effectiveness and efficiency are based on Conte et al. [5], and are defined as follows. Effectiveness in identifying threats to validity: Effectiveness = the ratio between the number of real threats found and the total number of threats known. Effectiveness in identifying control actions: Effectiveness = the ratio between the number of real actions found and total number of actions known. Efficiency in identifying threats to validity: Efficiency = the ratio between the number of real threats found and time spent. Efficiency in identifying control actions: Efficiency = the ratio between the number of real actions found and time spent. D. Hypotheses Using the variables described above, we defined the following null hypotheses: H 01 there is no difference in the effectiveness indicator when identifying TTVs with or without using the TVA; H 02 there is no difference in the effectiveness indicator when identifying CAs with or without using the TVA; H 03 there is no difference in the efficiency indicator when identifying TTVs with or without using the TVA; H 04 there is no difference in the efficiency indicator when identifying CAs with or without using the TVA. 258

E. Selection of Subjects The selected subjects were 16 students taking the ESE course. All subjects had lessons about the experiment process described by Wohlin et al. [14]. In addition, all subjects made presentations about validity evaluation in SE experiments and validity types. All subjects signed a consent form and filled out a characterization form that measured their expertise with controlled experiments. We also applied a pretest activity to measure more directly the subjects expertise. The pretest activity consisted in the identification of TTVs and CAs from a paper reporting a controlled experiment. In this activity, the subjects did not use any support. We selected the controlled experiment described by Thelin et al. [13]. This paper was retrieved by the SLR [1], which described a controlled experiment in the context of software inspection. The pretest activity was conducted simultaneously with all subjects in a classroom. The pretest activity consisted in identifying and addressing TTVs without any support. The subjects had 90 minutes to perform the task. At the end, a researcher had analyzed the answers and assigned a grade to the results of the pretest activity. Based on the results of the characterization form and the pretest activity, the subjects were divided into groups of: low (L), medium-low (ML), and medium (M) knowledge. Low knowledge means that a person identified one to four real TTVs and CAs, and does not have experience. Medium-low knowledge means that a person identified five to eight real TTVs and CAs, and has little experience (i. e., participated in an experiment as subject). Finally, medium knowledge means that a person identified more than eight real TTVs and CAs, and has little experience. In order to reduce the bias of having more experienced subjects in one or another treatment (with or without using the TVA), we equally and randomly distributed the subjects, balancing both groups. F. Design of the experiment The design is one factor, with two treatments. The assignment of the treatments was randomized. Each group consisted of 8 subjects and each used only one treatment, either the TVA or the guidelines proposed by Wohlin et al. [14]. G. Instrumentation The instruments used in this experiment were: characterization and consent forms, guidelines described by Wohlin et al. [14], data collection form (control group, since the TVA automates such process), TVA, instructions to use TVA (presented on the initial screen of the tool), follow-up questionnaire (experimental group, since it was used to gather improvement opportunities), excerpts from Madeyski s experiment in the original language (English) and a translated version (Portuguese). These excerpts were: the experiment context, treatments details, metrics details, and experiment planning. The translation was carried out by a researcher and reviewed by another researcher with an advanced English level. All instruments were validated by the authors. H. Experiment Operation The subjects were divided into two classrooms, one for the experimental group and another for the control group. All subjects had up to 90 minutes to carry out the experiment. During the experiment, the control group had to fill out the data collection form as they found TTVs or CAs. In addition, the control group used the checklist presented in [14]. On the other hand, the experimental group used the TVA. When the subjects from the experimental group finished the task, they had to fill out the follow-up questionnaire. After finishing the experiment, we created a list containing all different TTVs without duplicated ones. Likewise, we created a unique CAs list. Both lists were based on TTVs and CAs described in the validity section of Madeyski s experiment. A total of 23 TTVs and 12 CAs were found in the TTVs section of the paper. However, there were TTVs that were not described in the validity section. Thus, an independent experienced researcher evaluated Madeyski s experiment and found 23 new TTVs and 29 new CAs. Therefore, a total of 46 TTVs and 41 CAs were identified. Other TTVs and CAs that were found by the subjects and that were not already included in the lists, were classified as false positives (in order words, they were not present in Madeyski s experiment). In the next section, we present the results of the controlled experiment. IV. RESULTS We obtained the quantitative data from the data collection form resulting from the experimental task. After creating the list of TTVs and the list of CAs, we counted the number of TTVs and CAs. We also considered the time used by each subject to carry out the experimental task. Table I presents both the results per subject and per treatment. We have performed a statistical analysis using the statistical tool SPSS v 20.0.0 and α = 0.05. This choice of statistical significance was motivated by the small sample size used in this experiment [6]. In order to compare the effectiveness and efficiency of both samples, we used the Mann-Whitney nonparametrical statistic method. This test is equivalent of the t- Student parametrical test (for further information, please see [14]). It was used because we had two groups to compare, different subjects in each condition, and no assumption about the data distribution. We also used the boxplots graph to facilitate the visualization of results. The boxplot graph which compares the effectiveness indicator for the identification of TTVs and CAs is shown in Figure 4. When analyzing the graph, we can see that the mean from the experimental group is higher than the mean from the control group for both indicators. Thus, the group that used the TVA was more effective than the group that did not use it. Also, the comparison using the Mann-Whitney statistic method showed that there was significant difference between the groups (p = 0.001 for TTVs; and, p = 0.005 for CAs). These results suggest that the use of the TVA provided different effectiveness related to the identification of TTVs and CAs in controlled experiments; and that this difference is significant from a statistical point of view. These results reject the null hypothesis H 01 and H 02. 259

TABLE I. QUANTITATIVE RESULTS PER SUBJECT AND TREATMENT. Control group (guidelines) Experimental group (TVA) 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 Expertise Level M ML M L ML ML ML ML L ML M M ML ML ML ML Time (min) 90 88 75 45 87 87 80 82 70 60 47 62 53 63 52 40 Total TTVs Found 6 4 7 12 14 8 4 6 18 42 33 50 28 15 35 14 Real TTVs Found 3 3 7 6 8 8 3 4 13 35 22 37 20 12 25 13 False Positive TTVs 3 1 0 6 6 0 1 2 5 7 11 13 8 3 10 1 Effectiveness in Identifying TTVs % 6.52 6.52 15.22 13.04 17.39 17.39 6.52 8.70 28.26 76.09 47.83 80.43 43.48 26.09 54.35 28.26 Efficiency in Identifying TTVs 2.00 2.05 5.60 8.00 5.52 5.52 2.25 2.93 11.14 35.00 28.09 35.81 22.64 11.43 28.85 19.50 Total CAs Found 6 4 7 12 10 8 4 5 10 32 17 49 18 6 20 8 Real CAs Found 1 3 5 3 6 8 1 2 5 26 12 29 12 5 18 6 False Positive CAs 5 1 2 9 4 0 3 3 5 6 5 20 6 1 2 2 Effectiveness in Identifying CAs % 2.44 7.32 12.20 7.32 14.63 19.51 2.44 4.88 12.20 63.41 29.27 70.73 29.27 12.20 43.90 14.63 Efficiency in Identifying CAs 0.67 2.05 4.00 4.00 4.14 5.52 0.75 1.46 4.29 26.00 15.32 28.06 13.58 4.76 20.77 9.00 The boxplot graph which compares the efficiency indicator for the identification of TTVs and CAs is shown in Figure 5. According to the graph, the experimental group was more efficient than the control group, finding more TTVs and CAs in lesser time. Also, the Mann-Whitney statistical test shows that the difference between the groups is statistically significant (p = 0.001 for both indicators). Thus, the use of the TVA provides better efficiency regarding the identification of TTVs and CAs. These results reject the null hypothesis H 03 and H 04. Figure 4. Effectiveness indicator for identification of threats and actions. V. THREATS TO VALIDITY The following subsections present the TTVs considered in this controlled experiment. These TTVs have been grouped in four main categories: internal, external, conclusion and construct. A. Internal Validity Individual differences among the subjects. In order to address this TTV, we divided the subjects in balanced groups, according to the results from the characterization form and the pretest activity. Language of the study object was different from the native language of the subjects. In order to address this TTV, we translated the study object to the native language of the subjects. The translation was done by a researcher and reviewed by another researcher with high proficiency in English. Both documents (original and translated) were given to the subjects. B. External Validity Representativeness of subjects. The students are representative of the target population (novice researchers), but the results cannot be generalized. This experiment should be replicated for other samples and contexts. Representativeness of artifact. In order to avoid this TTV, we selected an experiment reported in literature. However, new experiments should be conducted with study objects from other contexts. Figure 5. Efficiency indicator for identification of threats and actions. 260

C. Conclusion Validity The small number of data points. However, small sample sizes are a known problem in Software Engineering which is difficult to overcome [5]. Thus, the data extracted from this controlled experiment can only be considered indicators and not conclusive. Heterogeneous characteristic of the subjects. This TTV was not considered because all students do not have experience in the planning of controlled experiments. Thus, the sample was homogeneous. D. Construct Validity Biased judgment of real TTVs and CAs. In order to mitigate this TTV, an independent researcher evaluated Madeyski s experiment and judged which were real TTVs or CAs. However, only one independent researcher might not be enough to classify TTVs and CAs. Only one study object was used. Therefore, the results may depend on the study object used. In order to mitigate this TTV, we plan to conduct new controlled experiments with different study objects, to enhance the validity of our results. Subjects may behave differently, because they were observed during the experiment execution. In order to avoid this TTV, subjects were observed during the pretest activity. Thus, it is expected that the subjects adapted to the environment. However, the observation impacts cannot be measured. In summary, we conclude that the existing TTVs were not considered large in this study. Also, we carried out CAs in order to reduce their impact. VI. CONCLUSION AND FUTURE WORK This paper has presented a controlled experiment to answer the following research question: Is the identification of TTVs and CAs more efficient and effective when the TVA is used? In order to answer this question we have measured the results of the TVA in terms of effectiveness and efficiency. Results showed that the use of the TVA provided better effectiveness and efficiency related to the identification of TTVs and CAs. The quantitative analysis using the Mann- Whitney statistic method showed that the effectiveness and efficiency indicators are better, when using the TVA, and that there is significant statistical difference. Due to the small sample used and the number of carried out experiments (only one), these results are not conclusive but indicate that the use of the TVA enhance the effectiveness and efficiency of the identification of TTVs and their CAs. We intend to perform new empirical studies to validate the results, including subjects with high experience in planning controlled experiments. Thus, we will obtain evidence if the TVA is useful for experienced researchers, while evaluating if experienced researchers agree with our categorization and relationships of TTVs and CAs. We hope that our findings will be useful in the promotion and improvement of the current practice of experiments planning. We also hope that the proposed tool aids researchers in the validity evaluation of their empirical studies, improving the confidence in their results and reducing the effort during planning stages. ACKNOWLEDGMENT We would like to acknowledge the financial support granted by FAPEAM process PAPE 032/2013. REFERENCES [1] A. Anderlin-Neto, and T. Conte, Threats to validity and their control actions results of a systematic literature review, TR-USES-2014-0002, UFAM, March 2014, Available at http://uses.icomp.ufam.edu.br/attachments/article/42/tr-uses-2014-0002.pdf. [2] C. Andersson, T. Thelin, P. Runeson, and N. Dzamashvili, An experimental evaluation of inspection and testing for detection of design faults, ISESE, pp. 174-184, 2003. [3] E. Arisholm, H. Gallis, T. Dyba, and D. I. K. Sjoberg, Evaluating pair programming with respect to system complexity and programmer expertise, TSE 33 (2), pp. 65-86, 2007. [4] S. Biffl, and M. Halling, Investigating the defect detection effectiveness and cost benefit of nominal inspection teams, TSE 29 (5), pp. 385-397, 2003. [5] T. Conte, J. Massollar, E. Mendes, and G. H. Travassos, Usability evaluation based on web design perspectives, ESEM, pp. 146-155, 2007. [6] T. Dyba, V. B. Kampenes, and D. I. K. Sjoberg, A systematic review of statistical power in Software Engineering experiments, IST 48 (8), pp. 745-755, 2006. [7] R. Feldt, and A. Magazinius, Validity threats in empirical Software Engineering research an initial survey, SEKE, pp. 374-379, 2010. [8] V. B. Kampenes, Quality of design, analysis and reporting of Software Engineering experiments, a systematic review, Doctoral Thesis, University of Oslo, 2007. [9] S. MacDonell, M. Shepperd, B. Kitchenham, and E. Mendes, How reliable are systematic reviews in empirical software engineering?, TSE 36 (5), pp. 676-687, 2010. [10] L. Madeyski, The impact of Test-First programming on branch coverage and mutation score indicator of unit tests an experiment, IST 52 (2), pp. 169-184, 2010. [11] J. C. Maldonado, J. Carver, F. Shull, S. Fabbri, E. Dória, L. Martimiano, M. Mendonça, and V. Basili, Perspective-Based Reading: a replicated experiment focused on individual reviewer effectiveness, EMSE 11 (1), pp. 119-142, 2006. [12] F. Shull, D. Cruzes, V. Basili, and M. Mendonça, Simulating families of studies to build confidence in defect hypotheses, IST 47 (15), pp. 1019-1032, 2005. [13] T. Thelin, P. Runeson, and C. Wöhlin, An experimental comparison of Usage-Based and Checklist-Based Reading, TSE 29 (8), pp. 687-704, 2003. [14] C. Wohlin, P. Runeson, M. Host, M. C. Ohlsson, B. Regnell, and A. Wessl, Experimentation in Software Engineering, Kluwer Academic Publishers, 2012. 261