Classifying Breast Cancer By Using Decision Tree Algorithms

Similar documents
Learning From the Past with Experiment Databases

Rule Learning With Negation: Issues Regarding Effectiveness

CS Machine Learning

Rule Learning with Negation: Issues Regarding Effectiveness

Experiment Databases: Towards an Improved Experimental Methodology in Machine Learning

Python Machine Learning

The 9 th International Scientific Conference elearning and software for Education Bucharest, April 25-26, / X

Impact of Cluster Validity Measures on Performance of Hybrid Models Based on K-means and Decision Trees

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

Applications of data mining algorithms to analysis of medical data

Content-based Image Retrieval Using Image Regions as Query Examples

Numeracy Medium term plan: Summer Term Level 2C/2B Year 2 Level 2A/3C

ScienceDirect. A Framework for Clustering Cardiac Patient s Records Using Unsupervised Learning Techniques

Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition

Comparison of EM and Two-Step Cluster Method for Mixed Data: An Application

An Empirical Comparison of Supervised Ensemble Learning Approaches

Machine Learning from Garden Path Sentences: The Application of Computational Linguistics

AGS THE GREAT REVIEW GAME FOR PRE-ALGEBRA (CD) CORRELATED TO CALIFORNIA CONTENT STANDARDS

Probability and Statistics Curriculum Pacing Guide

(Sub)Gradient Descent

THE ROLE OF DECISION TREES IN NATURAL LANGUAGE PROCESSING

Radius STEM Readiness TM

Predicting Students Performance with SimStudent: Learning Cognitive Skills from Observation

Predicting Student Attrition in MOOCs using Sentiment Analysis and Neural Networks

Analysis of Hybrid Soft and Hard Computing Techniques for Forex Monitoring Systems

Reducing Features to Improve Bug Prediction

Universidade do Minho Escola de Engenharia

Ordered Incremental Training with Genetic Algorithms

Rule discovery in Web-based educational systems using Grammar-Based Genetic Programming

Montana Content Standards for Mathematics Grade 3. Montana Content Standards for Mathematical Practices and Mathematics Content Adopted November 2011

Math-U-See Correlation with the Common Core State Standards for Mathematical Content for Third Grade

A Decision Tree Analysis of the Transfer Student Emma Gunu, MS Research Analyst Robert M Roe, PhD Executive Director of Institutional Research and

Iterative Cross-Training: An Algorithm for Learning from Unlabeled Web Pages

Maximizing Learning Through Course Alignment and Experience with Different Types of Knowledge

WE GAVE A LAWYER BASIC MATH SKILLS, AND YOU WON T BELIEVE WHAT HAPPENED NEXT

Machine Learning and Data Mining. Ensembles of Learners. Prof. Alexander Ihler

A NEW ALGORITHM FOR GENERATION OF DECISION TREES

Grade 6: Correlated to AGS Basic Math Skills

Evaluating and Comparing Classifiers: Review, Some Recommendations and Limitations

Improving Simple Bayes. Abstract. The simple Bayesian classier (SBC), sometimes called

Assignment 1: Predicting Amazon Review Ratings

Page 1 of 11. Curriculum Map: Grade 4 Math Course: Math 4 Sub-topic: General. Grade(s): None specified

Lecture 1: Machine Learning Basics

Optimizing to Arbitrary NLP Metrics using Ensemble Selection

Mining Student Evolution Using Associative Classification and Clustering

Dublin City Schools Mathematics Graded Course of Study GRADE 4

Evolutive Neural Net Fuzzy Filtering: Basic Description

Ph.D in Advance Machine Learning (computer science) PhD submitted, degree to be awarded on convocation, sept B.Tech in Computer science and

Fourth Grade. Reporting Student Progress. Libertyville School District 70. Fourth Grade

Issues in the Mining of Heart Failure Datasets

Cooperative evolutive concept learning: an empirical study

Algebra 1, Quarter 3, Unit 3.1. Line of Best Fit. Overview

An Empirical Analysis of the Effects of Mexican American Studies Participation on Student Achievement within Tucson Unified School District

COMPUTER-ASSISTED INDEPENDENT STUDY IN MULTIVARIATE CALCULUS

Modeling user preferences and norms in context-aware systems

Innovative Methods for Teaching Engineering Courses

Experiments with SMS Translation and Stochastic Gradient Descent in Spanish Text Author Profiling

Twitter Sentiment Classification on Sanders Data using Hybrid Approach

Julia Smith. Effective Classroom Approaches to.

Probability estimates in a scenario tree

Mining Association Rules in Student s Assessment Data

Using Genetic Algorithms and Decision Trees for a posteriori Analysis and Evaluation of Tutoring Practices based on Student Failure Models

CS 1103 Computer Science I Honors. Fall Instructor Muller. Syllabus

Linking Task: Identifying authors and book titles in verbose queries

Extending Place Value with Whole Numbers to 1,000,000

CHAPTER 4: REIMBURSEMENT STRATEGIES 24

Pre-AP Geometry Course Syllabus Page 1

Word Segmentation of Off-line Handwritten Documents

Generation of Attribute Value Taxonomies from Data for Data-Driven Construction of Accurate and Compact Classifiers

Decision Analysis. Decision-Making Problem. Decision Analysis. Part 1 Decision Analysis and Decision Tables. Decision Analysis, Part 1

CSL465/603 - Machine Learning

CS4491/CS 7265 BIG DATA ANALYTICS INTRODUCTION TO THE COURSE. Mingon Kang, PhD Computer Science, Kennesaw State University

OFFICE SUPPORT SPECIALIST Technical Diploma

The Method of Immersion the Problem of Comparing Technical Objects in an Expert Shell in the Class of Artificial Intelligence Algorithms

On-Line Data Analytics

have to be modeled) or isolated words. Output of the system is a grapheme-tophoneme conversion system which takes as its input the spelling of words,

STA 225: Introductory Statistics (CT)

TOPICS LEARNING OUTCOMES ACTIVITES ASSESSMENT Numbers and the number system

Constructive Induction-based Learning Agents: An Architecture and Preliminary Experiments

Activity Recognition from Accelerometer Data

Learning Methods in Multilingual Speech Recognition

Introduction to Causal Inference. Problem Set 1. Required Problems

A Version Space Approach to Learning Context-free Grammars

Learning goal-oriented strategies in problem solving

NCSC Alternate Assessments and Instructional Materials Based on Common Core State Standards

University of Groningen. Systemen, planning, netwerken Bosman, Aart

Australian Journal of Basic and Applied Sciences

GACE Computer Science Assessment Test at a Glance

Analyzing sentiments in tweets for Tesla Model 3 using SAS Enterprise Miner and SAS Sentiment Analysis Studio

Edexcel GCSE. Statistics 1389 Paper 1H. June Mark Scheme. Statistics Edexcel GCSE

Linking the Ohio State Assessments to NWEA MAP Growth Tests *

Learning Methods for Fuzzy Systems

Introducing the New Iowa Assessments Mathematics Levels 12 14

MULTIPLE CHOICE. Choose the one alternative that best completes the statement or answers the question.

Using dialogue context to improve parsing performance in dialogue systems

Digital Fabrication and Aunt Sarah: Enabling Quadratic Explorations via Technology. Michael L. Connell University of Houston - Downtown

CS 446: Machine Learning

Switchboard Language Model Improvement with Conversational Data from Gigaword

GCSE Mathematics B (Linear) Mark Scheme for November Component J567/04: Mathematics Paper 4 (Higher) General Certificate of Secondary Education

Notes on The Sciences of the Artificial Adapted from a shorter document written for course (Deciding What to Design) 1

Transcription:

Classifying Breast Cancer By Using Decision Tree Algorithms

Nusaibah AL-SALIHY, Turgay IBRIKCI (Presenter) Cukurova University, TURKEY

What Is A Decision Tree? Why A Decision Tree? Why Decision TreeClassification? Dataset And Features What Is A Classification? Classify By Decision Tree? A Basic Decision Tree Algorithms Decision Tree Algorithms Some Measures Applied In B.C Dataset Conclusion Selected References

What is a Decision Tree? An characterized by the inference of general laws from particular instances learning task. Use specific facts to draw more popularize conclusions A predictive model according on a branching series of logical and arithmetic test. These smaller tests are less complex than a one-stage classifier because, decision tree break-down a dataset into smaller and smaller sets whilst keeping in the same time linked decision tree is gradually progressing.

Block diagram for decision tree C Root node Where:- A. Nodes of the tree A A B. Leaves nodes of the tree / (terminal nodes) B B B B C. Branches of the tree / (decision point) Figure 1. Simple model for decision tree

Why Decision Tree Classification Decision trees are a simple and soft, also very powerful form of diversified variable analysis. Obtain similar and sometimes better accuracy compared to other models Their outcomes are interpretable. They do not need any special parameters. The construction process is comparatively fast. Decision Tree is excessively used by many researchers in healthcare field.

Dataset and Features The dataset had taken from Wisconsin Breast Cancer Data from the UCI Machine Learning Repository. 569 patients with the type of diagnosis illnesses (B, Benign or M, Malignant). The rest of 30 features are properties of cells with Mean, Standard errors and Worst values of the radius, texture, perimeter, area, smoothness, compactness, concavity, concave points, symmetry and the fractional dimension to each cell nucleus

Dataset Dataset Characteristics Attribute Characteristics Multivariate Real No. of Attributes: 31 No. of Instances: 569 Associated Tasks Classification No. of Classes: 2 Description of dataset included types, no. of attributes, instances and classes.

By using 10-fold cross validation can be done in following ways: 1. Create 10 instances of 90-10 split randomly for data set, some of the data may always only occur in train/test set. This is called 10-fold random cross validation 2. Create 10 equal sized partitions of data randomly and at 10 instances of learning, use 9 of them for training and 1 for testing. Make sure every partition is used only once of testing. Training Data B.C Dataset with classes Classification Algorithms Classifier (Model) Classifier outcome Figure 2. Training data for decision tree

J48 Algorithm J48 it s one of widespread decision tree algorithms, because it s actives with specific attributes, lost or missed values of the dataset By J48 can be increased a precision by pruning. To classify a dataset by a perfectly as possible, J48 algorithm builds decision trees for data from a set of training data. Decision Tree Algorithms Figure 3 Result J48 via Weka

Functional Tree algorithm FT is one of form of multivariate tree. Classifier for build up functional tree that classification trees that may have logistic regression tasks on the internal nodes / leaf nodes. The functional tree can handle with binary and nonbinary trees which called multi-ways or multi- classes targeted variables, missing or null values, numeric and nominal attributes Best First tree algorithms BF form of decision tree learning, and may do almost of the characteristics of standard decision learning. The naming best-node is saying for node that divides leads to the utmost limit of pollution among all nodes that helpful for splitting. It can deal with categorical and numerical variables.

Alternating Decision Tree Algorithm AD created by Yoav Freund and Llew Mason. AD includes two types nodes are decision / forecast nodes. Each decision-node contains a division test, whereas all forecast nodes contain a real number for valued. Decision Stump Algorithm A D.S is fundamentally a decision tree with only one single split. A tree in this algorithm divided at the root-level, which depend on a specified feature value pair. Sometimes D.S called 1-rule. Random Forest Tree Algorithm RF called random decision forest. RF classifier is data mining method developed by Beriman that is fast in learning and, runs efficiently in big datasets. RF increasingly for ML because RF offers two aspects that are very useful in data mining: high forecasting accuracy and new information on variable importance for classification methods.

Evaluation Metrics Applied in B.C Data Set

Precision(Specificity): is a measure of true postive rate accuracy. It is ration of the true positive-number to total number for forecasted positive samples Recall(Sensivity) : is measure of accuracy. It is the proportion of positive samples that were correctly identified to the total number of positive samples. It is also known as sensitivity or true positive rate (TPR). (1) Accuracy: is a measure of ratio for correctly forecasting. (2) F-Measure: is the mean of precision t and recall. It is an important measure as it gives equal importance to accuracy and recall. (3) (4)

Receiver Operating Characteristic (ROC) Curve is effective method of evaluating the quality or performance of diagnostic tests. TPR is plotted along the y axis and FP is plotted along the x axis. Figure 4. Comparison of D.T by Line chart according to Precision, Recall, F-Measure, and ROC Curves Decision Tree Precision Recall F- ROC Measure Curve FT 0.977 0.977 0.977 0.990 J48 0.934 0.933 0.933 0.931 RF 0.967 0.967 0.967 0.989 DS 0.891 0.889 0.887 0.874 AD 0.940 0.940 0.940 0.985 BF 0.930 0.930 0.930 0.938 Table 2. Comparison of D.T classification algorithms

Figure 5.Performance correct instances analysis (Accuracy) in percentage values by Line chart Algorithms Accuracy FT 97.7% J48 93.1% RF 96.6% DS 88.0% AD 94.0% Table 3. Accuracy percentage of test instances D.T algorithms to and best / worst percentage BF 92.0%

. Figure 6. The taken time to test model on training/testing data for each D.T algorithms Algorithms Time (sec. ) FT 0.03 J48 0.22 RF 0.15 DS 0.13 AD 0.11 BF 0.04 Table 4 Shows time taken to test model on training / testing data for each D.T algorithms

Conclusions Decision Trees could be deal with multidimensional dataset. Depending on precision for J48, FT, AD, Random Forest, BF and Decision stump algorithms using different precision measures part. The experiential results showed that highest precision 97.7% found in FT classifier with highest correctly numbers instance (550) but, precision 88% is found in Decision stump with lowest correctly numbers instance (509).

Selected References Quinlan, J. R.1986. Induction of Decision Trees.Mach. Learn., 81-106. Cobain EF, Hayes DF. 2015. Indications for prognostic gene expression profiling in early breast cancer. Curr. Treat Options Oncol.16(5):23. Sharma,P., Ratnoo,S. 2014, A review on Discovery of Classification Rules using Pittsburgh Approach, Int. J. Artif. Intell. Know l. Discov., 4(3): 16. Han, J. and Kamber,M. 2001. Data Mining: Concepts and Techniques, Morgan Kaufmann. Breiman, L., 2001. Random forests, Machine Learning 45(1), 5-32. Iba, W.; and Langley, P. 1992; Induction of One-Level Decision Trees, in ML92: Proceedings of the Ninth Int. Conf. on Mach. Learn, 233 240. Holte, Robert C. 1993. Very Simple Classification Rules Perform Well on Most Commonly Used Datasets, Machine Learning. 11:63-91. Frank, E., Hal, M. A., and Witten I. 2016. The WEKA Workbench. Online Appendix for Data Mining: Practical Machine Learning Tools and Techniques, Morgan Kaufman. UCI Machine Learning Repository: 1995. Center for Mach. Learn. and Intelligent Systems. Breast cancer Wisconsin (diagnostic) dataset. Buck,Carol J. 2016. Step-by-Step Medical Coding, Elsevier Health Sciences (Book).