Inductive learning. Learning Algorithm. hypothesis. dataset. Dataset must be sufficiently large

Similar documents
Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

A Version Space Approach to Learning Context-free Grammars

Innovative Methods for Teaching Engineering Courses

AQUA: An Ontology-Driven Question Answering System

Lecture 1: Basic Concepts of Machine Learning

Rule Learning With Negation: Issues Regarding Effectiveness

Python Machine Learning

Axiom 2013 Team Description Paper

Rule-based Expert Systems

MYCIN. The MYCIN Task

Learning goal-oriented strategies in problem solving

Word Segmentation of Off-line Handwritten Documents

Decision Analysis. Decision-Making Problem. Decision Analysis. Part 1 Decision Analysis and Decision Tables. Decision Analysis, Part 1

The Strong Minimalist Thesis and Bounded Optimality

Cognitive Modeling. Tower of Hanoi: Description. Tower of Hanoi: The Task. Lecture 5: Models of Problem Solving. Frank Keller.

Lecture 1: Machine Learning Basics

Chapter 2 Rule Learning in a Nutshell

Some Principles of Automated Natural Language Information Extraction

Artificial Neural Networks written examination

Introduction to Simulation

Lecture 10: Reinforcement Learning

Developing Students Research Proposal Design through Group Investigation Method

LEGO MINDSTORMS Education EV3 Coding Activities

Practical Research. Planning and Design. Paul D. Leedy. Jeanne Ellis Ormrod. Upper Saddle River, New Jersey Columbus, Ohio

QuickStroke: An Incremental On-line Chinese Handwriting Recognition System

Rule Learning with Negation: Issues Regarding Effectiveness

Knowledge based expert systems D H A N A N J A Y K A L B A N D E

Transfer Learning Action Models by Measuring the Similarity of Different Domains

Proof Theory for Syntacticians

Designing a Computer to Play Nim: A Mini-Capstone Project in Digital Design I

OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS

WiggleWorks Software Manual PDF0049 (PDF) Houghton Mifflin Harcourt Publishing Company

Active Learning. Yingyu Liang Computer Sciences 760 Fall

PH.D. IN COMPUTER SCIENCE PROGRAM (POST M.S.)

KLI: Infer KCs from repeated assessment events. Do you know what you know? Ken Koedinger HCI & Psychology CMU Director of LearnLab

Getting Started with Deliberate Practice

Designing A Computer Opponent for Wargames: Integrating Planning, Knowledge Acquisition and Learning in WARGLES

Neuro-Symbolic Approaches for Knowledge Representation in Expert Systems

On-Line Data Analytics

Integrating derivational analogy into a general problem solving architecture

Abstractions and the Brain

FUZZY EXPERT. Dr. Kasim M. Al-Aubidy. Philadelphia University. Computer Eng. Dept February 2002 University of Damascus-Syria

An OO Framework for building Intelligence and Learning properties in Software Agents

Objectives. Chapter 2: The Representation of Knowledge. Expert Systems: Principles and Programming, Fourth Edition

Word learning as Bayesian inference

INPE São José dos Campos

Evolution of Collective Commitment during Teamwork

Planning with External Events

An Introduction to Simio for Beginners

A Neural Network GUI Tested on Text-To-Phoneme Mapping

Stacks Teacher notes. Activity description. Suitability. Time. AMP resources. Equipment. Key mathematical language. Key processes

Machine Learning from Garden Path Sentences: The Application of Computational Linguistics

Using computational modeling in language acquisition research

Case Acquisition Strategies for Case-Based Reasoning in Real-Time Strategy Games

Shared Mental Models

P. Belsis, C. Sgouropoulou, K. Sfikas, G. Pantziou, C. Skourlas, J. Varnas

An Online Handwriting Recognition System For Turkish

Truth Inference in Crowdsourcing: Is the Problem Solved?

Knowledge-Based - Systems

Action Models and their Induction

Model Ensemble for Click Prediction in Bing Search Ads

Issues in the Mining of Heart Failure Datasets

NAME OF ASSESSMENT: Reading Informational Texts and Argument Writing Performance Assessment

Testing A Moving Target: How Do We Test Machine Learning Systems? Peter Varhol Technology Strategy Research, USA

(Sub)Gradient Descent

Contents. Foreword... 5

Unit: Human Impact Differentiated (Tiered) Task How Does Human Activity Impact Soil Erosion?

AGS THE GREAT REVIEW GAME FOR PRE-ALGEBRA (CD) CORRELATED TO CALIFORNIA CONTENT STANDARDS

University of Groningen. Systemen, planning, netwerken Bosman, Aart

Cooperative evolutive concept learning: an empirical study

Laboratorio di Intelligenza Artificiale e Robotica

Ohio s Learning Standards-Clear Learning Targets

GACE Computer Science Assessment Test at a Glance

have to be modeled) or isolated words. Output of the system is a grapheme-tophoneme conversion system which takes as its input the spelling of words,

Designing a Rubric to Assess the Modelling Phase of Student Design Projects in Upper Year Engineering Courses

arxiv: v1 [cs.cv] 10 May 2017

An Effective Framework for Fast Expert Mining in Collaboration Networks: A Group-Oriented and Cost-Based Method

Self Study Report Computer Science

How do adults reason about their opponent? Typologies of players in a turn-taking game

Knowledge Elicitation Tool Classification. Janet E. Burge. Artificial Intelligence Research Group. Worcester Polytechnic Institute

An Investigation into Team-Based Planning

Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition

Learning to Schedule Straight-Line Code

George Mason University Graduate School of Education Education Leadership Program. Course Syllabus Spring 2006

A Reinforcement Learning Variant for Control Scheduling

Human Emotion Recognition From Speech

Analysis of Hybrid Soft and Hard Computing Techniques for Forex Monitoring Systems

ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY DOWNLOAD EBOOK : ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY PDF

We are strong in research and particularly noted in software engineering, information security and privacy, and humane gaming.

Probability and Game Theory Course Syllabus

Knowledge Transfer in Deep Convolutional Neural Nets

Learning and Transferring Relational Instance-Based Policies

Predicting Student Attrition in MOOCs using Sentiment Analysis and Neural Networks

How to Do Research. Jeff Chase Duke University

Laboratorio di Intelligenza Artificiale e Robotica

UML MODELLING OF DIGITAL FORENSIC PROCESS MODELS (DFPMs)

An Experimental Comparison of Knowledge Representation Schemes

Focused on Understanding and Fluency

Mining Association Rules in Student s Assessment Data

Transcription:

Inductive learning Learning algorithms like neural networks, decision trees, inductive logic programming, etc. all require a good number of examples to be able to do good predictions. dataset Learning Algorithm hypothesis Dataset must be sufficiently large

Analytical learning We don t need a large dataset if besides taking examples as input, the learning algorithm can take prior knowledge. dataset Prior knowledge Learning Algorithm hypothesis Dataset does not need to be large

Explanation-based learning Prior knowledge is used to reduce the size of the hypothesis space. Hypothesis Space HS Prior knowledge Reduced HS It analyzes each example to infer which features are relevant and which ones are irrelevant.

Example, learning to play chess Suppose we want to learn a concept like what is a board position in which black will lose the queen in x moves?. Chess is a complex game. Each piece can occupy many positions. We would need many examples to learn this concept. But humans can learn these type of concepts with very few examples. Why?

Example, learning to play chess Humans can analyze an example and use prior knowledge related to legal moves. From there it can generalize with only few examples. Because the white king is attacking both king and queen; black must avoid check, letting white capture the queen reasoning

Example, learning to play chess What is the prior knowledge involved in playing chess? It is knowledge about the rules of chess: Legal moves for the pieces. Players alternate moves in games. To win you must capture the opponent s king.

Inductive and Analytical Learning Inductive Learning Analytical Learning Input: HS, D, Input: HS, D, B Output: hypothesis h Output: hypothesis h h is consistent with D h is consistent with D and B ( B h ) HS: Hypothesis Space D: Training Set B: Background knowledge (domain theory)

Example, Analytical Learning Given: Dataset where each instance is a pair of objects represented by the following predicates: Color, Volume, Owner, Material, Density, On. SafeToStack(x,y) is the target concept

Example, Analytical Learning Hypothesis space: set of Horn clause rules. The head of each rule has the predicate SafeToStack. The body of each rule is based on the instances and the predicates LessThan, Equal, Greater and plus, minus and times. Example: SafeToStack(x,y) Volume(x,vx) ^ Volume(y,vy) ^ LessThan(vx,vy) Training examples: On(Obj1,Obj2) Type(Obj1,Box) Type(Obj2,Endtable) Color(Obj1,Red) Color(Obj2,Blue) Volume(Obj1,2) Owner(Obj1,Fred) Owner(Obj2,Louise) Density(Obj1,0.3) Material(Obj1,Cardboard) Material(Obj2,Wood)

Example, Analytical Learning Domain theory SafeToStack(x,y) Fragile(y) SafeToStack(x,y) Lighter(x,y) Lighter(x,y) Weight(x,wx) Λ Weight(y,wy) Λ LessThan(wx,wy) Fragile(x) Material(x,Glass) Determind: Note: Hypothesis consistent with both the training examples and the domain theory The domain theory refers to predicates not contained in the examples. The domain theory is sufficient to prove the example is true.

Perfect Domain Theories A domain theory is correct if each statement is true. A domain theory is complete if it covers every positive example of the instance space (w.r.t a target concept and instance space). A perfect domain theory is correct and complete.

Perfect Domain Theories Examples of where to find perfect domain theories: Rules of chess Examples of where not to find perfect domain theories: SafetoStack problem In this chapter, only learning problems with perfect domain theories is considered.

Explanation Based Learning Algorithm We consider an algorithm that has the following properties: It is a sequential covering algorithm considering the data incrementally For each positive example not covered by the current rules it forms a new rule by: o o o Explain how training example satisfies target concept, in terms of domain theory. Analyze the explanation to find a the most general conditions under which this explanation holds. Refine the current hypothesis by adding a new Horn Clause rule to cover the example.

Explanation Based Learning Algorithm PROLOG-EBG Learning a single Horn clause rule Remove the positive training examples covered by this rule Until no further positive examples remain uncovered The explanation is a proof that the example belongs to the target (if the theory is perfect)

Explanation

Explanation There might be more than one explanation to the example. In that case one or all explanations may be used. An explanation is obtained using a backward chaining search as is done by PROLOG. PROLOG-EBG stops when it finds the first proof.

Analyze Many features appear in an example. Of them, how many are truly relevant? We consider as relevant those features that show in the explanation. Example: Relevant feature: Density Irrelevant feature: Owner

Analyze Taking the leaf nodes of the explanation and substituting variables x and y for Obj1, and Obj2: SafeToStack(x,y) Volume(x,2) Λ Density(x,0.3) Λ Type(y,Endtable) Remove features that are independent of x and y such as Equal(0.6,times(2,0.3)) and LessThan(0.6,5). The rule is now more general and can serve to explain other instances matching the rule.

Computing the weakest preimage of explanation

Refine The current hypothesis is the set of Horn clauses that have been learned to this point. By using sequential covering more rules are added, thus refining our hypothesis. A new instance is negative if it is not covered by any rule.

Discovering New Features The PROLOG-EBG system described can formulate new features that are not in the training examples. Example: Volume * Density > 5 (derived from the domain theory) Similar to features represented by hidden neurons in ANN. Difference: ANN is a statistical method that requires many training examples to obtain the hidden neuron features PROLOG-EBG use an analytical process to obtain new features from analysis of single training examples

Inductive Bias in Explanation-Based Learning What is the inductive bias of explanation based learning? The hypothesis h follows deductively from D and B D: database B: Background knowledge Bias: Prefer small sets of maximally general Horn Clauses

Search Control Knowledge Problem: learning to speed up search programs ( speedup learning ) Examples include: playing chess scheduling and optimization problems. Problem formulation: S: set of possible search states O: set of legal operators (transform one state into another state) G: predicate over S indicating the goal states

Prodigy Prodigy is a planning system. Input: state space S and operators O. Output: A sequence of operators that lead from the initial state to the final state. Prodigy uses a means-end planner: we decompose goals into subgoals: Goal Subproblems

Prodigy Example: Goal: accommodate blocks to read UNIVERSAL On(U,N) On(N,I) On(I,V) On(V,E) On(A,L) Question: what subgoal should be attacked first? Answer is given by search control knowledge

Prodigy and Explanation Based Learning Prodigy defines a set of target concepts to learn, e.g., which operator given the current state takes you to the goal state? An example of a rule learned by Prodigy in the blockstacking problem is: IF One subgoal to be solved is On(x,y) AND One subgoal to be solved is On(y,z) THEN Solve the subgoal On(y,z) before On(x,y)

Prodigy and Explanation Based Learning The rationale behind the rule is that it would avoid a conflict when stacking blocks. Prodigy learns by first encountering a conflict, then explaining the reason for the conflict and creating a rule like the one above. Experiments show an improvement in efficiency by a factor of two to four.

Problems with EBL The number of control rules that must be learned is very large. If the control rules are many, much time will be spent looking for the best rule. Utility analysis is used to determine what rules to keep and what rules to forget.

Problems with EBL Another problem with EBL is that it is sometimes difficult to create an explanation for the target concept. For example, in chess, learning a concept like: states for which operator A leads to a solution The search here grows exponentially.

Summary Different from inductive learning, analytical learning looks for a hypothesis that fit the background knowledge and covers the training examples. Explanation based learning is one kind of analytical learning that divides into three steps: Explain the target value for the current example Analyze the explanation (generalize) Refine the hypothesis

Summary PROLOG-EBG constructs intermediate features after analyzing examples. Explanation based learning can be used to find search control rules. Depend on a perfect domain theory.