OHJ-2556 ARTIFICIAL INTELLIGENCE Spring 2012

Similar documents
Intelligent Agents. Chapter 2. Chapter 2 1

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

Agents and environments. Intelligent Agents. Reminders. Vacuum-cleaner world. Outline. A vacuum-cleaner agent. Chapter 2 Actuators

Chapter 2. Intelligent Agents. Outline. Agents and environments. Rationality. PEAS (Performance measure, Environment, Actuators, Sensors)

Lecture 10: Reinforcement Learning

Seminar - Organic Computing

MYCIN. The MYCIN Task

Reinforcement Learning by Comparing Immediate Reward

On Human Computer Interaction, HCI. Dr. Saif al Zahir Electrical and Computer Engineering Department UBC

Laboratorio di Intelligenza Artificiale e Robotica

University of Groningen. Systemen, planning, netwerken Bosman, Aart

Action Models and their Induction

Introduction to Simulation

Laboratorio di Intelligenza Artificiale e Robotica

Innovative Methods for Teaching Engineering Courses

Lecture 1: Machine Learning Basics

An OO Framework for building Intelligence and Learning properties in Software Agents

Modeling user preferences and norms in context-aware systems

A GENERIC SPLIT PROCESS MODEL FOR ASSET MANAGEMENT DECISION-MAKING

Axiom 2013 Team Description Paper

Rover Races Grades: 3-5 Prep Time: ~45 Minutes Lesson Time: ~105 minutes

Testing A Moving Target: How Do We Test Machine Learning Systems? Peter Varhol Technology Strategy Research, USA

Radius STEM Readiness TM

Lecture 1: Basic Concepts of Machine Learning

LEGO MINDSTORMS Education EV3 Coding Activities

Visual CP Representation of Knowledge

COMPUTER-ASSISTED INDEPENDENT STUDY IN MULTIVARIATE CALCULUS

BMBF Project ROBUKOM: Robust Communication Networks

Toward Probabilistic Natural Logic for Syllogistic Reasoning

Firms and Markets Saturdays Summer I 2014

Rule-based Expert Systems

MATH 205: Mathematics for K 8 Teachers: Number and Operations Western Kentucky University Spring 2017

TD(λ) and Q-Learning Based Ludo Players

Knowledge based expert systems D H A N A N J A Y K A L B A N D E

The Strong Minimalist Thesis and Bounded Optimality

Contents. Foreword... 5

Learning Optimal Dialogue Strategies: A Case Study of a Spoken Dialogue Agent for

THE UNIVERSITY OF SYDNEY Semester 2, Information Sheet for MATH2068/2988 Number Theory and Cryptography

Self Study Report Computer Science

Purdue Data Summit Communication of Big Data Analytics. New SAT Predictive Validity Case Study

AGENDA LEARNING THEORIES LEARNING THEORIES. Advanced Learning Theories 2/22/2016

Speeding Up Reinforcement Learning with Behavior Transfer

Designing a Rubric to Assess the Modelling Phase of Student Design Projects in Upper Year Engineering Courses

Story Problems with. Missing Parts. s e s s i o n 1. 8 A. Story Problems with. More Story Problems with. Missing Parts

ECE-492 SENIOR ADVANCED DESIGN PROJECT

AGS THE GREAT REVIEW GAME FOR PRE-ALGEBRA (CD) CORRELATED TO CALIFORNIA CONTENT STANDARDS

Data Modeling and Databases II Entity-Relationship (ER) Model. Gustavo Alonso, Ce Zhang Systems Group Department of Computer Science ETH Zürich

Specification and Evaluation of Machine Translation Toy Systems - Criteria for laboratory assignments

Learning goal-oriented strategies in problem solving

Decision Analysis. Decision-Making Problem. Decision Analysis. Part 1 Decision Analysis and Decision Tables. Decision Analysis, Part 1

ProFusion2 Sensor Data Fusion for Multiple Active Safety Applications

EECS 571 PRINCIPLES OF REAL-TIME COMPUTING Fall 10. Instructor: Kang G. Shin, 4605 CSE, ;

What is Thinking (Cognition)?

Causal Link Semantics for Narrative Planning Using Numeric Fluents

The Good Judgment Project: A large scale test of different methods of combining expert predictions

Exploration. CS : Deep Reinforcement Learning Sergey Levine

The Task. A Guide for Tutors in the Rutgers Writing Centers Written and edited by Michael Goeller and Karen Kalteissen

How to Judge the Quality of an Objective Classroom Test

Evolution of Collective Commitment during Teamwork

A Context-Driven Use Case Creation Process for Specifying Automotive Driver Assistance Systems

TABLE OF CONTENTS TABLE OF CONTENTS COVER PAGE HALAMAN PENGESAHAN PERNYATAAN NASKAH SOAL TUGAS AKHIR ACKNOWLEDGEMENT FOREWORD

On the Combined Behavior of Autonomous Resource Management Agents

Software Maintenance

High-level Reinforcement Learning in Strategy Games

Probability estimates in a scenario tree

Major Milestones, Team Activities, and Individual Deliverables

ENME 605 Advanced Control Systems, Fall 2015 Department of Mechanical Engineering

Notes on The Sciences of the Artificial Adapted from a shorter document written for course (Deciding What to Design) 1

Designing Autonomous Robot Systems - Evaluation of the R3-COP Decision Support System Approach

Numeracy Medium term plan: Summer Term Level 2C/2B Year 2 Level 2A/3C

Lecture 1.1: What is a group?

Generating Test Cases From Use Cases

Objectives. Chapter 2: The Representation of Knowledge. Expert Systems: Principles and Programming, Fourth Edition

What is a Mental Model?

Spring 2015 IET4451 Systems Simulation Course Syllabus for Traditional, Hybrid, and Online Classes

Test How To. Creating a New Test

Science Olympiad Competition Model This! Event Guidelines

Using focal point learning to improve human machine tacit coordination

Reference to Tenure track faculty in this document includes tenured faculty, unless otherwise noted.

Team Dispersal. Some shaping ideas

ENEE 302h: Digital Electronics, Fall 2005 Prof. Bruce Jacob

Unit 13 Assessment in Language Teaching. Welcome

Evolutive Neural Net Fuzzy Filtering: Basic Description

Intensive English Program Southwest College

An Introduction to Simio for Beginners

IAT 888: Metacreation Machines endowed with creative behavior. Philippe Pasquier Office 565 (floor 14)

Welcome to the session on ACCUPLACER Policy Development. This session will touch upon common policy decisions an institution may encounter during the

STA 225: Introductory Statistics (CT)

Evidence for Reliability, Validity and Learning Effectiveness

Artificial Neural Networks written examination

Activities, Exercises, Assignments Copyright 2009 Cem Kaner 1

Algebra 1, Quarter 3, Unit 3.1. Line of Best Fit. Overview

AUTOMATED TROUBLESHOOTING OF MOBILE NETWORKS USING BAYESIAN NETWORKS

FONDAMENTI DI INFORMATICA

Facing our Fears: Reading and Writing about Characters in Literary Text

Learning Cases to Resolve Conflicts and Improve Group Behavior

Level 6. Higher Education Funding Council for England (HEFCE) Fee for 2017/18 is 9,250*

First Grade Standards

ReinForest: Multi-Domain Dialogue Management Using Hierarchical Policies and Knowledge Ontology

Alberta Police Cognitive Ability Test (APCAT) General Information

Transcription:

OHJ-2556 ARTIFICIAL INTELLIGENCE Spring 2012 1 OHJ-2556 ArtificialIntelligence, Spring 2012 12.1.2012 2 General 6 credit units Can be included in post-graduate studies Lectures (4h per week), 8 + 6½ weeks Student presentations instead of lectures towards the end Weekly exercises (2h per week) Programming exercise An essay + voluntary presentation You may have a chance to influence the contents now Lectures based on a new edition of the course book, not significantly different from the previous one OHJ-2556 ArtificialIntelligence, Spring 2012 12.1.2012 1

3 Organization & timetable Lectures: prof. Tapio Elomaa Tue 12 14 TB219 & Thu 12 14 TB223 Jan. 10 Apr. 26 Period break: Mar. 5 11 Easter break: Apr. 5 11 Traveling?: Apr. 18 20 Student presentations after Vappu (in May) Weekly exercises starting on week 4 (Jan. 23 ) M.Sc. Teemu Heinimäki Day, Room? Exam: Fri May 18, 2012 1 4 PM 4 Topics Part I Artificial Intelligence 1 Introduction 2 Intelligent Agents Part II Problem Solving 3 Solving Problems by Searching 4 Beyond Classical Search 5 Adversarial Search 6 Constraint Satisfaction Problems Part III Knowledge and Reasoning 7 Logical Agents 8 First-Order Logic 9 Inference in First-Order Logic 10 Classical Planning 11 Planning and Acting in the Real World 12 Knowledge Representation Part IV Uncertain Knowledge and Reasoning 13 Quantifying Uncertainty 14 Probabilistic Reasoning 15 Probabilistic Reasoning over Time 16 Making Simple Decisions 17 Making Complex Decisions Part V Learning 18 Learning from Examples 19 Knowledge in Learning 20 Learning Probabilistic Models 21 Reinforcement Learning Part VII Communicating, Perceiving, and Acting 22 Natural Language Processing 23 Natural Language for Communication 24 Perception 25 Robotics Part VIII Conclusions 26 Philosophical Foundations 27 AI: The Present and Future 2

5 Exercise = Essay, details still open The course has one compulsory exercise. To pass the course you need to pass it The exercise is graded on scale 0 10, in addition the voluntary presentation yields up to 4 points Grade 1 means passing Topic of your own choice from the area of AI. Group/single? Returned by date? - Written report, 10 15 pp. - Oral presentation 20 min 6 Grading In grading there are several components: The exam (max 30 p.) May 18 An essay (max 10 p.) DL? The max points altogether is 40 p. Oral presentation (max 4 p.) After Vappu Weekly exercises earn extra points (max 6 p.) Most probably the grade is decided as follows: points 20 24 28 32 36 grade 1 2 3 4 5 3

7 Material The textbook of the course is S. Russell, P. Norvig: Artificial Intelligence, A Modern Approach, Third ed., Pearson, 2010 There is no prepared material, the slides appear in the web as the lectures proceed http://www.cs.tut.fi/kurssit/ohj-2550/2556.html http://www.cs.tut.fi/~elomaa/teach/2556.html/ The exam is based on the lectures (i.e., not on the slides only) 8 Weekly Exercises It is most advisable to take part in the weekly exercises The exercise questions appear in the web on Thu of previous week the latest Being ready to present one s own solution to a question publicly yields one mark. Each session has c. 5 questions you may gather altogether c. 5 12 = 60 marks Marks Extra points 30% 1 40% 2 50% 3 60% 4 70% 5 80% 6 4

9 1. INTRODUCTION Artificial intelligence is a wide and far-reaching concept. It also keeps changing over time Nowadays maybe more of a playground of philosophers and cognitive scientists From the point of view of computer science AI comprises of a set of more focused research fields that have already drifted quite far apart from each other The common goal in different subfields is to raise the intelligence of computers/machines I.e., to make the use of software easier As a result one gets ready-to-use software and theory charting out the boundaries of mechanical computation 10 1.1 What is AI? The only comparison to an intelligent machine that we are aware of is ourselves On the other hand, comparison to the human intelligence limits out other (better) alternatives Ideal intelligence is called rationality One can view intelligence from the point of view of though and behavior As combinations we get four distinct views to artificial intelligence: thinking humanly, acting humanly, thinking rationally, and acting rationally 5

11 Turing Test English mathematician Alan Turing proposed in 1950 the following criterion for the intelligence of a machine: a human interrogator cannot differentiate whether s/he is communicating with another human or a computer using text messages An example of a test of acting human-like In the so-called total Turing test the machine also has to be able to observe and manipulate its physical environment Time-limited Turing test competitions are organized annually The best performance against knowledgeable organizers is recorded by programs that try to fool the interrogator Human experts have the highest probability of being judged as non-humans 12 Rationality Study of rational thought is essentially study of formal logic and logical deduction Methods that are based (only) on logics suffer from computational complexity issues and of the difficulty of expressing uncertain knowledge In the model of rational acting we examine agents An agent is something that acts A software agent is distinguished from a program e.g. by its autonomous control, capability to perceive its environment, ability to adapt to change, persistence over a prolonged time, ability to take on another s goals, and so forth 6

13 A rational agent works to reach the best possible outcome given its observations and knowledge Under uncertainty one aims at maximizing expectation over the outcome Rational performance on the long run may require one to perform seemingly irrationally for shorter periods In the real world total rationality is not usually possible (due to the lack of time) 14 1.4 The State of the Art Different activities in many subfields: Robotic vehicles: Driverless robotic cars are being developed in closed environments and more and more in daily traffic. Modern cars recognize speed limits, adapt to the traffic, take care of pedestrian safety, can park themselves, have intelligent light systems, wake up the driver, Speech recognition: Many devices and services nowadays understand spoken words (even dialogs) Autonomous planning and scheduling: E.g. space missions are tomorrow planned autonomously Game playing: Computers defeat human world champions in chess systematically and convincingly. Watson that won human players in Jeopardy was today named by HS as the science achievement of 2011 7

15 Spam fighting: Learning algorithms reliably filter away 80% or 90% of all messages saving us time for more important tasks Logistics planning: E.g. military operations are helped by automated logistics planning and scheduling for transportation Robotics: Autonomous vacuum cleaners, lawn movers, toys, and special (hazardous) environment robots are common these days Machine translation: Translation programs based on statistics and machine learning are in ever increasing demand (in particular in EU) There are of course many other interesting AI applications some of them taking advantage of the Web 16 2 INTELLIGENT AGENTS An agent perceives its environment through sensors and acts upon the environment through actuators Our sensors include eyes, ears, nose, and other organs Our actuators include hands, legs, mouth, and other body parts The sensors of a robot (or a car) can include e.g. cameras, infrared and laser range finders and various motors as actuators A software agent receives keystrokes, file contents, and network packets as sensory inputs It acts on the environment by displaying on the screen, writing files, and sending network packets 8

17 We assume that every agent will perceive its own actions, but not always the effects In general, an agent s choice of action at any given instant can depend on the entire percept sequence observed to date The agent function maps any given percept sequence to an action The table of all possible input-output pairs of the function is a complete external characterization of the agent Of course such a table is infinite in most cases not applicable Internally the agent function for an artificial agent will be implemented by an agent program 18 Measuring the Performance of an Agent The rational agent that we are aiming at should be successful in the task it is performing To assess the success we need to have a performance measure What is rational at any given time depends on The performance measure that defines the criterion of success. The agent s prior knowledge of the environment. The actions that the agent can perform. The agent s percept sequence to date. For each possible percept sequence a rational agent should select an action that is expected to maximize its performance measure, given the evidence provided by the percept sequence and whatever built-in knowledge the agent has. 9

19 Junk mail filtering has to classify e-mail messages as junk or relevant messages A physician has to decide whether to operate on a patient or not The number of correctly classified instances is not the best possible measure of performance, because right and wrong decisions have different weights Moving an important message to the junk folder is worse than letting through some junk mails occasionally True Positive False negative False positive True negative 20 Properties of Environments Task environments can be classified at least by the following properties Fully vs. partially observable In a fully observable environment the agent s sensors give it all relevant aspects affecting its choice of performance Hence, the agent does not need to maintain any internal state (understanding of the state of the world) An environment might be partially observable because of noisy and inaccurate sensors or simply due to its basic nature Deterministic vs. stochastic If the next state of the environment is completely determined by the current state and the action executed, then we say that it is deterministic. Otherwise it is stochastic. A deterministic environment may appear stochastic if it is partially observable 10

21 Task Environments /2 Episodic vs. sequential In an episodic task environment, the next episode does not depend on the actions taken in previous episodes In sequential environments the current decision could affect all future decisions Static vs. dynamic In a static environment the agent may stop to deliberate its actions without fearing that the world changes In a dynamic environment the agent has to keep looking for the state of the environment In a semidynamic environment the environment itself does not change with the passage of time, but the agent s performance score does (the agent is penalized for the time required to plan its actions) 22 Task Environments /3 Discrete vs. continuous The distinction can be made with respect to the state of the environment, time, and the percepts and actions of the agent Single agent vs. multiagent Depends on which entities one wants to view as agents or just simply as stochastically behaving objects In the multiagent environment one can compete or cooperate Known vs. unknown Whether the laws of physics of the environment are known In a known environment the outcomes (or outcome probabilities in a stochastic environment) for all actions are given A known environment can be partially observable 11

23 Real World from the Perspective of a Robot Partially observable Sensors are not perfect and perceive only the close environment (holds also for humans) Stochastic Wheels slip, batteries run out, parts break one can never be certain that the intended action is fulfilled Sequential The effects of actions change over time a robot has to manage sequential decision problems and be able to learn Dynamic When to deliberate, when to act Continuous/Infinite Algorithms must work is this environment, not, e.g., in a finite discrete search space Single/multiagent environment Depending on whether one wants to look at other objects as agents or stochastic parts of the environment 24 2.4 The Structure of Agents Because the agent s history of percept-action pairs stored in a table describes the external behavior of the agent, in principle the control program of the agent could be based on table lookup Obviously, this only works for very small environments In more realistic situations tabulating percept-action pairs is not a viable solution Instead, the agent program has to be able to decide the desired action on any percept history without tabulating all possible alternatives The following agent types are the most common solutions to this problem 12

25 Simple / Model-based Reflex Agents The simplest possible control program makes the agent operate solely on the current percept discarding the percept history Reflexes are used in emergency situations by humans as well as by robots Reflexive behavior yields correct decisions only when the task environment is fully observable One can choose a set of current rules based on the agents internal state Hence one can maintain a model of the world covering some information that is not directly observable 26 2.4.4 Goal-based Agents In addition to its percepts the agent possesses knowledge of its goal The goal is some assertion concerning the environment which should be satisfied By combining the goal and knowledge of the effects of available actions the agent can try to satisfy the goal If the goal cannot be satisfied directly through one action, one has to find out a sequence of actions to satisfy it The agent may resort to search algorithms (next section) or planning (later on) 13

27 2.4.5 Utility-based Agents The agent may achieve its goal in many different ways different solutions may have differences in quality Setting a goal alone does not suffice to express more complex settings If the possible states of the environment are assigned an order through an utility function, then the agent can try to improve its value In partially observable and stochastic environments we choose the action that maximizes the expected utility of the outcomes The utility function maps a state (or a sequence of states) onto a real number In order to take advantage of this approach, the agent does not necessarily have to possess a utility function explicitly 28 2.4.6 Learning Agents Programming agents for all possible tasks by hand appears to be a hopeless task Already Turing (1950) proposed machine learning as a method of creating intelligent systems Learning allows the agent to operate in initially unknown environments and become more competent than its initial knowledge alone might allow The learning element of an agent has to be kept distinct from the actual performance element We come back to the techniques of machine learning towards the end of the course 14

29 3 SOLVING PROBLEMS BY SEARCHING A goal-based agent aims at solving problems by performing actions that lead to desirable states Let us first consider the uninformed situation in which the agent is not given any information about the problem other than its definition In blind search only the structure of the problem to be solved is formulated The agent aims at reaching a goal state The world is static, discrete, and deterministic The possible states of the world define the state space of the search problem 30 In the beginning the world is in the initial state s 1 The agent aims at reaching one of the goal states G Quite often one uses a successor function S: P ) to define the agent s possible actions Each state s has a set of legal successor states S(s) that can be reached in one step Paths have a non-negative cost, most often the sum of costs of steps in the path 15

31 The definitions above naturally determine a directed and weighted graph The simplest problem in searching is to find out whether any goal state can be reached from the initial state s 1 The actual solution to the problem, however, is a path from the initial state to a goal state Usually one takes also the costs of paths into account and tries to find a cost-optimal path to a goal state Many tasks of a goal-based agent are easy to formulate directly in this representation 32 7 2 4 1 2 5 6 3 4 5 8 3 1 6 7 8 For example, the states of the world of 8-puzzle (a sliding-block puzzle) are all 9!/2 = 181 440 reachable configurations of the tiles Initial state is one of the possible states The goal state is the one given on the right above Possible values of the successor function are moving the blank to left, right, up, or down Each move costs one unit and the path cost is the total number of moves 16

33 Donald Knuth s (1964) illustration of how infinite state spaces can arise Conjecture: Starting with the number 4, a sequence of factorial, square root, and floor operations will reach any desired positive integer 4!! 5 States : Positive numbers Initial state s 1 : 4 Actions: Apply factorial, square root, or floor operation Transition model: As given by the mathematical definitions of the operations Goal test: State is the desired positive integer 34 Search Tree When the search for a goal state begins from the initial state and proceeds by steps determined by the successor function, we can view the progress of search in a tree structure When the root of the search tree, which corresponds to the initial state, is expanded, we apply the successor function to it and generate new search nodes to the tree as children of the root corresponding to the successor states The search continues by expanding other nodes in the tree respectively The search strategy (search algorithm) determines in which order the nodes of the tree are expanded 17

35 7 2 4 5 6 8 3 1 7 2 4 7 4 7 2 4 7 2 4 5 6 5 2 6 5 6 5 3 6 8 3 1 8 3 1 8 3 1 8 1 36 The node is a data structure with five components: The state to which the node corresponds, Link to the parent node, The action that was applied to the parent to generate this node, The cost of the path from the initial state to the node, and The depth of the node Global parameters of a search tree include: b (average or maximum) branching factor, d the depth of the shallowest goal, and m the maximum length of any path in the state space 18