Today s Lecture. Clustering Methods. Lecture 19: Uncertainty 4. Victor R. Lesser. Example of Multiply Connected BN

Similar documents
Introduction to Simulation

Rule-based Expert Systems

Planning with External Events

The Good Judgment Project: A large scale test of different methods of combining expert predictions

The Evolution of Random Phenomena

stateorvalue to each variable in a given set. We use p(x = xjy = y) (or p(xjy) as a shorthand) to denote the probability that X = x given Y = y. We al

COMPUTER-ASSISTED INDEPENDENT STUDY IN MULTIVARIATE CALCULUS

Probability and Statistics Curriculum Pacing Guide

ACTL5103 Stochastic Modelling For Actuaries. Course Outline Semester 2, 2014

CS Machine Learning

Lecture 1: Machine Learning Basics

Objectives. Chapter 2: The Representation of Knowledge. Expert Systems: Principles and Programming, Fourth Edition

Radius STEM Readiness TM

Mathematics process categories

Malicious User Suppression for Cooperative Spectrum Sensing in Cognitive Radio Networks using Dixon s Outlier Detection Method

Decision Analysis. Decision-Making Problem. Decision Analysis. Part 1 Decision Analysis and Decision Tables. Decision Analysis, Part 1

Generative models and adversarial training

Seminar - Organic Computing

Mathematics Success Grade 7

A Model of Knower-Level Behavior in Number Concept Development

Rendezvous with Comet Halley Next Generation of Science Standards

A Reinforcement Learning Variant for Control Scheduling

Grade 6: Correlated to AGS Basic Math Skills

An OO Framework for building Intelligence and Learning properties in Software Agents

Algebra 1 Summer Packet

Learning Probabilistic Behavior Models in Real-Time Strategy Games

Diagnostic Test. Middle School Mathematics

Probabilistic Mission Defense and Assurance

Short Text Understanding Through Lexical-Semantic Analysis

Algebra 2- Semester 2 Review

Dublin City Schools Mathematics Graded Course of Study GRADE 4

Discriminative Learning of Beam-Search Heuristics for Planning

ENME 605 Advanced Control Systems, Fall 2015 Department of Mechanical Engineering

Learning and Transferring Relational Instance-Based Policies

STT 231 Test 1. Fill in the Letter of Your Choice to Each Question in the Scantron. Each question is worth 2 point.

TOKEN-BASED APPROACH FOR SCALABLE TEAM COORDINATION. by Yang Xu PhD of Information Sciences

On-Line Data Analytics

Algebra 1, Quarter 3, Unit 3.1. Line of Best Fit. Overview

CHAPTER 4: REIMBURSEMENT STRATEGIES 24

A Stochastic Model for the Vocabulary Explosion

South Carolina College- and Career-Ready Standards for Mathematics. Standards Unpacking Documents Grade 5

An Introduction to Simio for Beginners

Maximizing Learning Through Course Alignment and Experience with Different Types of Knowledge

AUTOMATED TROUBLESHOOTING OF MOBILE NETWORKS USING BAYESIAN NETWORKS

University of Groningen. Systemen, planning, netwerken Bosman, Aart

POLA: a student modeling framework for Probabilistic On-Line Assessment of problem solving performance

FF+FPG: Guiding a Policy-Gradient Planner

Proof Theory for Syntacticians

Reinforcement Learning by Comparing Immediate Reward

Montana Content Standards for Mathematics Grade 3. Montana Content Standards for Mathematical Practices and Mathematics Content Adopted November 2011

Simple Random Sample (SRS) & Voluntary Response Sample: Examples: A Voluntary Response Sample: Examples: Systematic Sample Best Used When

The 9 th International Scientific Conference elearning and software for Education Bucharest, April 25-26, / X

Lecture 15: Test Procedure in Engineering Design

Fault tree analysis for maintenance needs

Mathematics subject curriculum

Artificial Neural Networks written examination

AGS THE GREAT REVIEW GAME FOR PRE-ALGEBRA (CD) CORRELATED TO CALIFORNIA CONTENT STANDARDS

STA 225: Introductory Statistics (CT)

Lahore University of Management Sciences. FINN 321 Econometrics Fall Semester 2017

Uncertainty concepts, types, sources

Executive Guide to Simulation for Health

Word learning as Bayesian inference

Henry Tirri* Petri Myllymgki

Level 6. Higher Education Funding Council for England (HEFCE) Fee for 2017/18 is 9,250*

Graphical Data Displays and Database Queries: Helping Users Select the Right Display for the Task

Knowledge based expert systems D H A N A N J A Y K A L B A N D E

Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition

Finding Your Friends and Following Them to Where You Are

Active Learning. Yingyu Liang Computer Sciences 760 Fall

Evaluating Statements About Probability

An Efficient Implementation of a New POP Model

GACE Computer Science Assessment Test at a Glance

University of Cincinnati College of Medicine. DECISION ANALYSIS AND COST-EFFECTIVENESS BE-7068C: Spring 2016

The Use of Statistical, Computational and Modelling Tools in Higher Learning Institutions: A Case Study of the University of Dodoma

BAYESIAN ANALYSIS OF INTERLEAVED LEARNING AND RESPONSE BIAS IN BEHAVIORAL EXPERIMENTS

Continuing Competence Program Rules

PM tutor. Estimate Activity Durations Part 2. Presented by Dipo Tepede, PMP, SSBB, MBA. Empowering Excellence. Powered by POeT Solvers Limited

Probability estimates in a scenario tree

Extending Place Value with Whole Numbers to 1,000,000

Georgetown University at TREC 2017 Dynamic Domain Track

"f TOPIC =T COMP COMP... OBJ

A Semantic Imitation Model of Social Tag Choices

Grade 2: Using a Number Line to Order and Compare Numbers Place Value Horizontal Content Strand

Analysis of Enzyme Kinetic Data

Clouds = Heavy Sidewalk = Wet. davinci V2.1 alpha3

Edexcel GCSE. Statistics 1389 Paper 1H. June Mark Scheme. Statistics Edexcel GCSE

12- A whirlwind tour of statistics

Software Maintenance

Python Machine Learning

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

BADM 641 (sec. 7D1) (on-line) Decision Analysis August 16 October 6, 2017 CRN: 83777

The Effect of Collaborative Partnerships on Interorganizational

A Bootstrapping Model of Frequency and Context Effects in Word Learning

Evolutive Neural Net Fuzzy Filtering: Basic Description

The Socially Structured Possibility to Pilot One s Transition by Paul Bélanger, Elaine Biron, Pierre Doray, Simon Cloutier, Olivier Meyer

Machine Learning and Data Mining. Ensembles of Learners. Prof. Alexander Ihler

***** Article in press in Neural Networks ***** BOTTOM-UP LEARNING OF EXPLICIT KNOWLEDGE USING A BAYESIAN ALGORITHM AND A NEW HEBBIAN LEARNING RULE

Semi-Supervised Face Detection

Regret-based Reward Elicitation for Markov Decision Processes

Developing an Assessment Plan to Learn About Student Learning

Transcription:

Lecture 19: Uncertainty 4 Victor R. Lesser CMPSCI 683 Fall 2010 Today s Lecture Inference in Multiply Connected BNs Clustering methods transform the network into a probabilistically equivalent polytree. Also called Join tree algorithms Conditioning methods instantiate certain variables and evaluate a polytree for each possible instantiation. Stochastic simulation approximate the beliefs by generating a large number of concrete models that are consistent with the evidence and CPTs. Example of Multiply Connected BN Clustering Methods P(C=T)=.5 Creating meganodes until the network becomes a polytree. C P(S=T) T.10 F.50 S R P(W=T) T T.99 T F.90 F T.90 F F.00 C P(R=T) T.80 F.20 Most effective approach for exact evaluation of multiply connected BNs. The tricky part is choosing the right meganodes. Q. What happens to the NP-hardness of the inference problem?

Clustering Example* Cutset Conditioning Methods S+R P(W) T T.99 T F.90 F T.90 F F.00 P(C)=.5 Spr+ P(S+R) C TT TF FT FF T.08.02.72.18 F.10.40.10.40 How do you still answer P(=True =False)? How do you create meganode? What are the disadvantages? Once a variable is instantiated it can be duplicated and thus break a cycle. A cutset is a set of variables whose instantiation makes the graph a polytree. Each polytree s likelihood is used as a weight when combining the results. Networks Created by Instantiation Eliminate from BN; Sum(%+,%Cloud-) + P(S)=.1 P(R)=.8 + P(S)=.5 - C P(S) T.10 F.50 P(R)=.2 C P(R) T.80 F.20 - Stochastic Simulation -- Direct Sampling Assign each root node (without parents) a value based on prior probability. Assign all other nodes a NULL value. Pick a node X with no value, but whose parents have values, and randomly assign a value to X using P(X Parents(X)) as the distribution. Repeat until there is no such X. After N trials, P(X E) can be estimated by occurrences (X and E) / occurrences (E). Approximate P(X,E)/P(E) Does not focus on generating occurrences of E

Example P(WetGrass )

Stochastic Simulation cont. Problem with very unlikely events. Likelihood weighting can be used to fix problem. Likelihood weighting converges much faster than logic sampling and works well for very large networks.

Example of Likelihood Weighting P(WetGrass ) Choose a value for with prior P() = 0.5. Assume we choose cloudy = false. Choose a value for. We see that P( ) = 0.5, so we randomly choose a value given that distribution. Assume we choose =True. Look at. This is an evidence variable that has been set to True, so we look at the table to see that P( ) = 0.2. This run therefore counts as 0.2 of a complete run. Example of Likelihood Weighty cont d Look at WetGrass. Choose randomly with P (WetGrass =T =T) =0.99; assume we choose WetGrass = True. We now have completed a run with likelihood 0.2 that says WetGrass = True given = True. The next run will result in a different likelihood, and (possibly) a different value for WetGrass. We continue until we have accumulated enough runs, and then add up the evidence for each value, weighted by the likelihood score. Likelihood weighting usually converges much faster than logic sampling Still takes a long time to reach accurate probabilities for unlikely events Stochastic Simulation Likelihood Weighting Likelihood weighting example P( =T, WetGrass=T) ; for all nodes in the network ordered by parents ; if you are at the node that you have evidence for ; adjust likelihood of this run based on the likelihood of evidence given parents ;otherwise randomly choose based on value of parents chosen in previous steps w = 1.0

w = 1.0 0.1 w = 1.0 w = 1.0 0.1 W = 1.0 0.1 0.99 = 0.099

Stochastic Simulation Markov Chain Monte Carlo A node is conditionally independent of all other nodes in the network given its parents, children, and children s parents that is, given its Markov blanket. The MCMC algorithm MCMC generates each event by making a random change to the preceding event. It is therefore helpful to think of the network being in a particular current state specifying a value for every variable. The next state is generated by randomly sampling a value for one of the non-evidence variables X i, conditioned on the current values of the variables in the Markov blanket of X i. Don t need to look at any other variables MCMC therefore wanders randomly around the state space the space of possible complete assignments flipping one variable at a time but keeping the evidence variables fixed. The Markov chain Markov blanket sampling

MCMC example cont. Summary of a Belief Networks Conditional independence information is a vital and robust way to structure information about an uncertain domain. Belief networks are a natural way to represent conditional independence information. The links between nodes represent the qualitative aspects of the domain, and the conditional probability tables represent the quantitative aspects. A belief network is a complete representation for the joint probability distribution for the domain, but is often exponentially smaller in size. Summary of a Belief Networks, cont d Inference in belief networks means computing the probability distribution of a set of query variables, given a set of evidence variables. Belief networks can reason causally, diagnostically, in mixed mode, or intercausally. No other uncertain reasoning mechanism can handle all these modes. The complexity of belief network inference depends on the network structure. In polytrees (singly connected networks), the computation time is linear in the size of the network. Summary of a Belief Networks, cont d There are various inference techniques for general belief networks, all of which have exponential complexity in the worst case. In real domains, the local structure tends to make things more feasible, but care is needed to construct a tractable network with more than a hundred nodes. It is also possible to use approximation techniques, including stochastic simulation, to get an estimate of the true probabilities with less computation.

Next Lecture Introduction to Decision Theory Making Single-Shot Decisions Utility Theory