Final Exam: 1:00-3:30 pm, August 8, 2003 Room 265 Materials Sciences Building

Similar documents
Lecture 1: Machine Learning Basics

Radius STEM Readiness TM

Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition

Python Machine Learning

CS Machine Learning

Semi-supervised methods of text processing, and an application to medical concept extraction. Yacine Jernite Text-as-Data series September 17.

Word Segmentation of Off-line Handwritten Documents

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

Lecture 1: Basic Concepts of Machine Learning

Maximizing Learning Through Course Alignment and Experience with Different Types of Knowledge

Assignment 1: Predicting Amazon Review Ratings

QuickStroke: An Incremental On-line Chinese Handwriting Recognition System

CSL465/603 - Machine Learning

Chinese Language Parsing with Maximum-Entropy-Inspired Parser

Lecture 10: Reinforcement Learning

Learning From the Past with Experiment Databases

Rule Learning With Negation: Issues Regarding Effectiveness

Knowledge-Based - Systems

SARDNET: A Self-Organizing Feature Map for Sequences

Softprop: Softmax Neural Network Backpropagation Learning

A Case Study: News Classification Based on Term Frequency

Rule Learning with Negation: Issues Regarding Effectiveness

Artificial Neural Networks written examination

Axiom 2013 Team Description Paper

Learning Methods in Multilingual Speech Recognition

(Sub)Gradient Descent

Objectives. Chapter 2: The Representation of Knowledge. Expert Systems: Principles and Programming, Fourth Edition

Measurement. When Smaller Is Better. Activity:

FUZZY EXPERT. Dr. Kasim M. Al-Aubidy. Philadelphia University. Computer Eng. Dept February 2002 University of Damascus-Syria

Laboratorio di Intelligenza Artificiale e Robotica

Evolution of Symbolisation in Chimpanzees and Neural Nets

Learning Methods for Fuzzy Systems

Discriminative Learning of Beam-Search Heuristics for Planning

Machine Learning from Garden Path Sentences: The Application of Computational Linguistics

Machine Learning and Data Mining. Ensembles of Learners. Prof. Alexander Ihler

INPE São José dos Campos

Twitter Sentiment Classification on Sanders Data using Hybrid Approach

Switchboard Language Model Improvement with Conversational Data from Gigaword

Universidade do Minho Escola de Engenharia

Generative models and adversarial training

Evolutive Neural Net Fuzzy Filtering: Basic Description

Human Emotion Recognition From Speech

Chapter 2 Rule Learning in a Nutshell

Exploration. CS : Deep Reinforcement Learning Sergey Levine

Abstractions and the Brain

Proof Theory for Syntacticians

A Version Space Approach to Learning Context-free Grammars

The 9 th International Scientific Conference elearning and software for Education Bucharest, April 25-26, / X

Reducing Features to Improve Bug Prediction

Self Study Report Computer Science

Dublin City Schools Mathematics Graded Course of Study GRADE 4

have to be modeled) or isolated words. Output of the system is a grapheme-tophoneme conversion system which takes as its input the spelling of words,

CS4491/CS 7265 BIG DATA ANALYTICS INTRODUCTION TO THE COURSE. Mingon Kang, PhD Computer Science, Kennesaw State University

Learning to Schedule Straight-Line Code

CS 446: Machine Learning

Version Space. Term 2012/2013 LSI - FIB. Javier Béjar cbea (LSI - FIB) Version Space Term 2012/ / 18

Standard 1: Number and Computation

Semi-Supervised Face Detection

Course Outline. Course Grading. Where to go for help. Academic Integrity. EE-589 Introduction to Neural Networks NN 1 EE

Experiment Databases: Towards an Improved Experimental Methodology in Machine Learning

Speech Recognition at ICSI: Broadcast News and beyond

An OO Framework for building Intelligence and Learning properties in Software Agents

arxiv: v1 [cs.lg] 15 Jun 2015

CSC200: Lecture 4. Allan Borodin

Applications of data mining algorithms to analysis of medical data

MYCIN. The MYCIN Task

Probability and Statistics Curriculum Pacing Guide

MULTIPLE CHOICE. Choose the one alternative that best completes the statement or answers the question.

Notes on The Sciences of the Artificial Adapted from a shorter document written for course (Deciding What to Design) 1

Edexcel GCSE. Statistics 1389 Paper 1H. June Mark Scheme. Statistics Edexcel GCSE

System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks

Numeracy Medium term plan: Summer Term Level 2C/2B Year 2 Level 2A/3C

OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS

Ensemble Technique Utilization for Indonesian Dependency Parser

Grade 6: Correlated to AGS Basic Math Skills

ReinForest: Multi-Domain Dialogue Management Using Hierarchical Policies and Knowledge Ontology

UNIVERSITY OF CALIFORNIA SANTA CRUZ TOWARDS A UNIVERSAL PARAMETRIC PLAYER MODEL

University of Groningen. Systemen, planning, netwerken Bosman, Aart

Sample Problems for MATH 5001, University of Georgia

AGS THE GREAT REVIEW GAME FOR PRE-ALGEBRA (CD) CORRELATED TO CALIFORNIA CONTENT STANDARDS

Learning Distributed Linguistic Classes

Netpix: A Method of Feature Selection Leading. to Accurate Sentiment-Based Classification Models

The Good Judgment Project: A large scale test of different methods of combining expert predictions

COMPUTER-ASSISTED INDEPENDENT STUDY IN MULTIVARIATE CALCULUS

PHYSICS 40S - COURSE OUTLINE AND REQUIREMENTS Welcome to Physics 40S for !! Mr. Bryan Doiron

Firms and Markets Saturdays Summer I 2014

Introduction to Causal Inference. Problem Set 1. Required Problems

Content Language Objectives (CLOs) August 2012, H. Butts & G. De Anda

Large-Scale Web Page Classification. Sathi T Marath. Submitted in partial fulfilment of the requirements. for the degree of Doctor of Philosophy

Model Ensemble for Click Prediction in Bing Search Ads

Software Maintenance

ScienceDirect. A Framework for Clustering Cardiac Patient s Records Using Unsupervised Learning Techniques

Functional Maths Skills Check E3/L x

Data Fusion Through Statistical Matching

DOCTOR OF PHILOSOPHY HANDBOOK

Beyond the Pipeline: Discrete Optimization in NLP

Montana Content Standards for Mathematics Grade 3. Montana Content Standards for Mathematical Practices and Mathematics Content Adopted November 2011

Time series prediction

While you are waiting... socrative.com, room number SIMLANG2016

Introducing the New Iowa Assessments Mathematics Levels 12 14

Transcription:

Final Exam: 1:003:30 pm, August 8, 2003 Room 265 Materials Sciences Building CLOSED BOOK (twosided sheet of handwritten notes and a calculator allowed) Write your answers on these pages and show your work. If you feel that a question is not fully specified, state any assumptions you need to make in order to solve the problem. You may use the backs of these sheets for scratch work. Budget your time wisely. Before you begin, write your name on every page of the exam and read through all the questions (as some have multiple parts and are more involved than others). Make sure your exam contains seven (7) problems on eleven (11) pages. Name Student ID Problem Score Max Score 1 20 2 15 3 25 4 15 5 30 6 10 7 10 TOTAL 125

a) Answer each of the following questions true or false: i. Tabu search with a horizon of 1 behaves the same as a greedy hillclimbing search. ii. Simulated annealing with a temperature T = 0 also behaves identically to a greedy hillclimbing search. iii. Breadthfirst search is always a complete search method, even if all of the actions have different costs. iv. When hillclimbing and greedy best first search use the exact same admissible heuristic function, they will expand the same set of search nodes. v. If two admissible heuristic functions evaluate the same search node n as h 1 (n) = 6 and h 2 (n) = 8, we say h 1 dominates h 2, because it is less likely to overestimate the actual cost. b) Provide a short answer to the following questions: i. What are the three basic components of any genetic algorithm? ii. Compare and contrast genetic algorithms to beam search. 2

c) Consider the following partial search tree, where edges are labeled with actual costs of the associated action, and each node is labeled with its heuristic evaluation. Which node will be expanded next by each of the following search methods? A i. Greedy bestfirst search: ii. Uniform cost search: iii. A* search: B h=2 18 6 3 C D h=15 4 8 9 5 E F G H h=10 h=7 h=9 h=11 d) Your good friend is the groundskeeper at the mansion of mean old Mr. Mathis. In the backyard, there is a huge fountain with a complex network of pipes controlled by over 100 valves. One weekend, Mr. Mathis announces he s going on vacation, and when he returns he wants the fountain to spray as high as it can but the plans for the pipe network have been lost! Plus, since she only has the weekend, your friend can t possibly try all of the valve combinations to find the optimal setting. Which local optimization search method might she want to use in real life (since the fountain can t be simulated on a computer) to maximize the height of the fountain? You may assume that a valve is either on or off, and the water height is easily measured. If you need to make any other assumptions, state them clearly. 3

e) Provide a short answer to each of the following questions: i. Are the literals P(F(y), y, x) and P(x, F(A), F(v)) unifiable? If so, show the unifying substitution θ. If not, explain why. ii. Convert this FOL sentence to conjunctive normal form: x y Owns(x,y) Like(x,y) iii. What is the frame problem in situation calculus? How do we deal with it? f) Use forward chaining to show that Bigger(Smaug, Bilbo) is entailed by the following KB. Indicate which sentences were used with GMP, and show each substitution θ. 1. Hobbit(Bilbo) 4. Dragon(x) Wizard(y) Bigger(x,y) 2. Wizard(Gandalf) 5. Wizard(x) Hobbit(y) Bigger(x,y) 3. Dragon(Smaug) 6. Bigger(x,y) Bigger(y,z) Bigger(x,z) 4

Consider the task of learning to identify mushrooms that are SAFE or POISONOUS to eat based on a set of physical features. Four Boolean and discrete valued features that you could use are: STEM = {short, long}, BELL = {rounded, flat}, TEXTURE = {plain, spots, bumpy, ruffles}, and NUMBER = {single, multiple}. Consider using these features on the following training data: SAFE: POISONOUS: a) How would the naïve Bayes classifier label the following example: b) How would 3nearest neighbors, using hamming distance and unweighted voting, classify the same example from part (a)? 5

c) Use information gain to choose between TEXTURE and NUMBER to use in a decision stump (a decision tree with only one internal node at the root). Draw a diagram of the learned stump, and break ties at classification nodes by labeling them as POISONOUS (just to be on the safe side). Show all your work for partial credit. d) Estimate the future accuracy of this decision stump given the following test set: SAFE POISONOUS e) Draw a perceptron that might be used to learn this problem (just the structure is fine, you do not need to label with weights). f) What is the dimensionality of a perceptron s hypothesis space for this problem? 6

Consider the following Bayesian network where all variables are Boolean: A C P(A) 0.2 P(C) 0.5 P(E) 0.7 E B D A,C A,C A, C A, C P(B) 0.9 0.6 0.4 0.5 C,E C,E C, E C, E P(D) 0.3 0.1 0.8 0.7 a) How large would the fulljoint probability table (FJPT) have to be for this problem? b) What is the joint probability that all five variables are simultaneously true? c) Compute the probability P( C A, B, D, E) 7

Provide a short answer to each of the following questions: a) What s the difference between eager and lazy learning algortihms? Give an example of each. b) Give an example of a regression learning task. Which of the machine learning algorithms we ve discussed can learn this kind of concept well? c) Give one advantage and one disadvantage of Bayesian belief networks over naïve Bayes. d) Would the hypotheses learned by an inductive logic programming (ILP) system be considered symbolic or connectionist AI? Breifly explain. 8

e) What is the curse of dimensionality? Which learning methods are most affected by it? Describe one approach to dealing with this problem. f) What relationship does the minimum description length principle have with the ID3 learning algorithm? What about backpropagation learning for neural networks? g) Ensemble learning algorithms often benefit from the combined knowledge of its constituent agents, which learn from slightly different subsets of examples. Imagine that we want to create and ensemble learner where each constituent agent learns from subsets that are very different. How might we do this? How might the ensemble then classify new examples? Clearly state any assumptions that you need to. 9

The following Venn diagrams represent training examples for a Booleanvalued concept function plotted in feature space. Show how each of the following machine learning algorithms might partition the space based on these examples. Briefly explain to the right of each diagram why that algorithm would partition the data that particular way. (No need for calculations just give a qualitative answer.) Naïve Bayes 1Nearest Neighbor ID3 Neural Network with 1 hidden layer of 2 units Neural Network with unlimited hidden layers/units 10

Briefly define the following concepts and explain the significance of each to A.I. (Write your answer below the phrase.) αβ Search Occam s Razor Tuning Sets Promotion / Demotion The Markov Assumption 11 Have a good rest of the summer!!