Intelligent Agents. Chapter 2. Chapter 2 1

Similar documents
Agents and environments. Intelligent Agents. Reminders. Vacuum-cleaner world. Outline. A vacuum-cleaner agent. Chapter 2 Actuators

Chapter 2. Intelligent Agents. Outline. Agents and environments. Rationality. PEAS (Performance measure, Environment, Actuators, Sensors)

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

Lecture 10: Reinforcement Learning

LEGO MINDSTORMS Education EV3 Coding Activities

Modeling user preferences and norms in context-aware systems

Reinforcement Learning by Comparing Immediate Reward

Standards-Based Bulletin Boards. Tuesday, January 17, 2012 Principals Meeting

MYCIN. The MYCIN Task

Testing A Moving Target: How Do We Test Machine Learning Systems? Peter Varhol Technology Strategy Research, USA

Seminar - Organic Computing

What is a Mental Model?

Innovative Methods for Teaching Engineering Courses

The Good Judgment Project: A large scale test of different methods of combining expert predictions

The One Minute Preceptor: 5 Microskills for One-On-One Teaching

Robot manipulations and development of spatial imagery

A. Permission. All students must have the permission of their parent or guardian to participate in any field trip.

EECS 571 PRINCIPLES OF REAL-TIME COMPUTING Fall 10. Instructor: Kang G. Shin, 4605 CSE, ;

Rule-based Expert Systems

Readyman Activity Badge Outline -- Community Group

The Tutor Shop Homework Club Family Handbook. The Tutor Shop Mission, Vision, Payment and Program Policies Agreement

SSE - Supervision of Electrical Systems

Speeding Up Reinforcement Learning with Behavior Transfer

ISFA2008U_120 A SCHEDULING REINFORCEMENT LEARNING ALGORITHM

B.S/M.A in Mathematics

TD(λ) and Q-Learning Based Ludo Players

CSL465/603 - Machine Learning

A Context-Driven Use Case Creation Process for Specifying Automotive Driver Assistance Systems

AUTOMATED TROUBLESHOOTING OF MOBILE NETWORKS USING BAYESIAN NETWORKS

Medical Complexity: A Pragmatic Theory

Evaluation of Usage Patterns for Web-based Educational Systems using Web Mining

Evaluation of Usage Patterns for Web-based Educational Systems using Web Mining

Attendance. St. Mary s expects every student to be present and on time for every scheduled class, Mass, and school events.

Specification and Evaluation of Machine Translation Toy Systems - Criteria for laboratory assignments

USER ADAPTATION IN E-LEARNING ENVIRONMENTS

ESIC Advt. No. 06/2017, dated WALK IN INTERVIEW ON

Lecture 1: Machine Learning Basics

Diagnostic Test. Middle School Mathematics

WHAT ARE VIRTUAL MANIPULATIVES?

Action Models and their Induction

PART C: ENERGIZERS & TEAM-BUILDING ACTIVITIES TO SUPPORT YOUTH-ADULT PARTNERSHIPS

Introduction to Simulation

Surprise-Based Learning for Autonomous Systems

Teaching a Laboratory Section

Individual Component Checklist L I S T E N I N G. for use with ONE task ENGLISH VERSION

Parent Information Welcome to the San Diego State University Community Reading Clinic

CS 1103 Computer Science I Honors. Fall Instructor Muller. Syllabus

Computers Change the World

Functional Skills Mathematics Level 2 assessment

Alberta Police Cognitive Ability Test (APCAT) General Information

Taking Kids into Programming (Contests) with Scratch

Designing Autonomous Robot Systems - Evaluation of the R3-COP Decision Support System Approach

Kindergarten - Unit One - Connecting Themes

A MULTI-AGENT SYSTEM FOR A DISTANCE SUPPORT IN EDUCATIONAL ROBOTICS

MAKING YOUR OWN ALEXA SKILL SHRIMAI PRABHUMOYE, ALAN W BLACK

Aviation English Solutions

Strategic Management (MBA 800-AE) Fall 2010

PowerTeacher Gradebook User Guide PowerSchool Student Information System

Learning Microsoft Publisher , (Weixel et al)

Objectives. Chapter 2: The Representation of Knowledge. Expert Systems: Principles and Programming, Fourth Edition

Disability Resource Center St. Philip's College ensures Access. YOU create Success. Frequently Asked Questions

Firms and Markets Saturdays Summer I 2014

Emergency Management Games and Test Case Utility:

TOURISM ECONOMICS AND POLICY (ASPECTS OF TOURISM) BY LARRY DWYER, PETER FORSYTH, WAYNE DWYER

Public Speaking Rubric

1. Answer the questions below on the Lesson Planning Response Document.

Probability Therefore (25) (1.33)

An Interactive Intelligent Language Tutor Over The Internet

MAE Flight Simulation for Aircraft Safety

Economics 201 Principles of Microeconomics Fall 2010 MWF 10:00 10:50am 160 Bryan Building

Are You a Left- or Right-Brain Thinker?

Reduce the Failure Rate of the Screwing Process with Six Sigma Approach

Entrepreneurial Discovery and the Demmert/Klein Experiment: Additional Evidence from Germany

Why Pay Attention to Race?

Procedia - Social and Behavioral Sciences 237 ( 2017 )

Software Maintenance

POLA: a student modeling framework for Probabilistic On-Line Assessment of problem solving performance

Executive Guide to Simulation for Health

Different Requirements Gathering Techniques and Issues. Javaria Mushtaq

Fundraising 101 Introduction to Autism Speaks. An Orientation for New Hires

Knowledge based expert systems D H A N A N J A Y K A L B A N D E

Blank Table Of Contents Template Interactive Notebook

An Introduction to Simulation Optimization

Disciplinary Literacy in Science

A process by any other name

Consultation skills teaching in primary care TEACHING CONSULTING SKILLS * * * * INTRODUCTION

Solution Focused Methods RAYYA GHUL 2017

The Foundations of Interpersonal Communication

On-Line Data Analytics

Lecture 1: Basic Concepts of Machine Learning

Rule Learning With Negation: Issues Regarding Effectiveness

On the Combined Behavior of Autonomous Resource Management Agents

Renaissance Learning P.O. Box 8036 Wisconsin Rapids, WI (800)

Probabilistic Latent Semantic Analysis

Using Blackboard.com Software to Reach Beyond the Classroom: Intermediate

Dr. Karen L. Pennington Vice President, Student Development & Campus Life Dr. Michele Campagna Executive Director, Center for Advising & Student

Contents. Foreword... 5

DEVELOPMENT OF AN INTELLIGENT MAINTENANCE SYSTEM FOR ELECTRONIC VALVES

Science Olympiad Competition Model This! Event Guidelines

Mathematics Scoring Guide for Sample Test 2005

Transcription:

Intelligent Agents Chapter 2 Chapter 2 1

Outline Agents and environments Rationality PEAS (Performance measure, Environment, Actuators, Sensors) Environment types The structure of agents Chapter 2 2

Agents and environments sensors environment percepts actions? agent actuators Agents include humans, robots, softbots, thermostats, etc. The agent function maps from percept histories to actions: f : P A The agent program runs on the physical architecture to produce f Chapter 2 3

Vacuum-cleaner world A B Percepts: location and contents, e.g., [A, Dirty] Actions: Left, Right, Suck, NoOp Chapter 2 4

A vacuum-cleaner agent Percept sequence Action [A, Clean] Right [A, Dirty] Suck [B, Clean] Lef t [B, Dirty] Suck [A, Clean], [A, Clean] Right [A, Clean], [A, Dirty]. Suck. function Reflex-Vacuum-Agent([location,status]) returns an action if status = Dirty then return Suck else if location = A then return Right else if location = B then return Left What is the right function? Can it be implemented in a small agent program? Chapter 2 5

Outline Agents and environments Rationality PEAS (Performance measure, Environment, Actuators, Sensors) Environment types The structure of agents Chapter 2 6

Rationality Fixed performance measure evaluates any sequence of environment states Chapter 2 7

Rationality Fixed performance measure evaluates any sequence of environment states the amount of dirt cleaned in time T one point per each square cleaned up at each time step T? one point per clean square per time step, minus one per move? penalize for > k dirty squares, electricity, noise? having a clean floor Design performance measures according to what one actually wants in the environment, rather than to how one thinks the agent should behave. A rational agent chooses whichever action maximizes the expected value of the performance measure given the percept sequence to date Chapter 2 8

Rationality What is rational at any given time depends on four things: The performance measure that defines the criterion of success. The agents prior knowledge of the environment. The actions that the agent can perform. The agents percept sequence to date. The performance measure awards one point for each clean square at each time step, over a lifetime of 1000 time steps. The geography of the environment is known a priori but the dirt distribution and the initial location of the agent are not. Clean squares stay clean and sucking cleans the current square. The Left and Right actions move the agent left and right except when this would take the agent outside the environment, in which case the agent remains where it is. The only available actions are Left, Right, and Suck. The agent correctly perceives its location and whether it contains dirt. Chapter 2 9

Omniscience, learning, autonomy Rational omniscient percepts may not supply all relevant information action outcomes may not be as expected assume the agent does not look both ways - is it rational to cross? information gathering (exploration): doing actions in order to modify future percepts Hence, rational successful (rationality maximises expected performace), given the percept sequence to date. Autonomy: A rational agent should be learn what it can to compensate for partial or incorrect prior knowledge initial knowlege + abilitity to learn Rational exploration, learning, autonomy Chapter 2 10

Outline Agents and environments Rationality PEAS (Performance measure, Environment, Actuators, Sensors) Environment types The structure of agents Chapter 2 11

PEAS To design a rational agent, we must specify the task environment Consider, e.g., the task of designing an automated taxi: Performance measure?? Environment?? Actuators?? Sensors?? Chapter 2 12

PEAS To design a rational agent, we must specify the task environment Consider, e.g., the task of designing an automated taxi: Performance measure?? safety, destination, profits, legality, comfort,... Environment?? streets/highways, traffic, pedestrians,,pathholes, puddles, weather,... Actuators?? steering, accelerator, brake, horn, speaker/display,... Sensors?? video, speedometer, accelerometer, odometer, engine sensors, GPS,... Chapter 2 13

Performance measure?? Environment?? Actuators?? Sensors?? Internet shopping agent Chapter 2 14

Internet shopping agent Performance measure?? price, quality, appropriateness, efficiency Environment?? current and future WWW sites, vendors, shippers Actuators?? display to user, follow URL, fill in form Sensors?? HTML pages (text, graphics, scripts) Chapter 2 15

Performance measure?? Environment?? Actuators?? Sensors?? Medical diagnosis system Chapter 2 16

Medical diagnosis system Performance measure??: Healthy patient, reduced costs Environment??: Patient, hospital, staff Actuators??: Display of questions, tests, diagnoses, treatments Sensors??: Keyboard entry of symptoms, findings, patient s answers Chapter 2 17

Performance measure?? Environment?? Actuators?? Sensors?? Interactive English tutor Chapter 2 18

Interactive English tutor Performance measure??: Student s score on test Environment??: Set of students, testing agency Actuators??: Display of exercises, suggestions, corrections Sensors??: Keyboard entry Chapter 2 19

Outline Agents and environments Rationality PEAS (Performance measure, Environment, Actuators, Sensors) Environment types The structure of agents Chapter 2 20

Environment types Fully observable: sensors give access to the complete state of the environment at each point in time. partially obervable: noise, no sensors, Single/Multi agent Does an agent A (the taxi driver for example) have to treat an object B (another vehicle) as an agent, or can it be treated merely as an object behaving according to the laws of physics? whether Bs behavior is best described as maximizing a performance measure whose value depends on agent As behavior. competitve/cooperative (chess,taxy,tenis) Deterministic/stochastic: the next state of the environment is completely determined by the current state and the action of the agent Uncertain: not fully observable or not deterministic. Nondeterministic: actions are characterized by their possible outcomes (no probabilities attached as in case of stochastic). Chapter 2 21

Episodic vs. sequential: the next episode does not depend on the actions in previous episodes. Static vs. dynamic: environment can change while an agent is deliberating semi-dynamic: the environment does not change, but the performance of the agent does (chess with clock) Discrete vs. continuous: applies to the state of the environment, to the way time is handled, and to the percepts and actions of the agent (chess vs. taxi driving) Known vs. unknown: not to the environment but to agent s knowledge (solitaire vs. video game - know vs. observable) Chapter 2 22

Environment types Task Env. Observable Agents Deterministic Episodic Static Discrete Chapter 2 23

Environment types Task Env. Observable Agents Deterministic Episodic Static Discrete Crossword puzzle Fully Single Deterministic Sequential Static Discrete Chapter 2 24

Environment types Task Env. Observable Agents Deterministic Episodic Static Discrete Crossword puzzle Fully Single Deterministic Sequential Static Discrete Chess with clock Fully Multi Deterministic Sequential Semi Discrete Chapter 2 25

Environment types Task Env. Observable Agents Deterministic Episodic Static Discrete Crossword puzzle Fully Single Deterministic Sequential Static Discrete Chess with clock Fully Multi Deterministic Sequential Semi Discrete Poker Partially Multi Stochastic Sequential Static Discrete Chapter 2 26

Environment types Task Env. Observable Agents Deterministic Episodic Static Discrete Crossword puzzle Fully Single Deterministic Sequential Static Discrete Chess with clock Fully Multi Deterministic Sequential Semi Discrete Poker Partially Multi Stochastic Sequential Static Discrete Backgammon Fully Multi Stochastic Sequential Static Discrete Chapter 2 27

Environment types Task Env. Observable Agents Deterministic Episodic Static Discrete Crossword puzzle Fully Single Deterministic Sequential Static Discrete Chess with clock Fully Multi Deterministic Sequential Semi Discrete Poker Partially Multi Stochastic Sequential Static Discrete Backgammon Fully Multi Stochastic Sequential Static Discrete Taxi driving Partially Multi Stochastic Sequential Dynamic Continous Chapter 2 28

Environment types Task Env. Observable Agents Deterministic Episodic Static Discrete Crossword puzzle Fully Single Deterministic Sequential Static Discrete Chess with clock Fully Multi Deterministic Sequential Semi Discrete Poker Partially Multi Stochastic Sequential Static Discrete Backgammon Fully Multi Stochastic Sequential Static Discrete Taxi driving Partially Multi Stochastic Sequential Dynamic Continous Medical diagnosis Partially Multi Stochastic Sequential Dynamic Continous Chapter 2 29

Environment types Task Env. Observable Agents Deterministic Episodic Static Discrete Crossword puzzle Fully Single Deterministic Sequential Static Discrete Chess with clock Fully Multi Deterministic Sequential Semi Discrete Poker Partially Multi Stochastic Sequential Static Discrete Backgammon Fully Multi Stochastic Sequential Static Discrete Taxi driving Partially Multi Stochastic Sequential Dynamic Continous Medical diagnosis Partially Multi Stochastic Sequential Dynamic Continous Image analysis Fully Single Deterministic Episodic Semi Continous Chapter 2 30

Environment types Task Env. Observable Agents Deterministic Episodic Static Discrete Crossword puzzle Fully Single Deterministic Sequential Static Discrete Chess with clock Fully Multi Deterministic Sequential Semi Discrete Poker Partially Multi Stochastic Sequential Static Discrete Backgammon Fully Multi Stochastic Sequential Static Discrete Taxi driving Partially Multi Stochastic Sequential Dynamic Continous Medical diagnosis Partially Multi Stochastic Sequential Dynamic Continous Image analysis Fully Single Deterministic Episodic Semi Continous Part picking Partially Single Stochastic Episodic Dynamic Continous robot Chapter 2 31

Environment types Task Env. Observable Agents Deterministic Episodic Static Discrete Crossword puzzle Fully Single Deterministic Sequential Static Discrete Chess with clock Fully Multi Deterministic Sequential Semi Discrete Poker Partially Multi Stochastic Sequential Static Discrete Backgammon Fully Multi Stochastic Sequential Static Discrete Taxi driving Partially Multi Stochastic Sequential Dynamic Continous Medical diagnosis Partially Multi Stochastic Sequential Dynamic Continous Image analysis Fully Single Deterministic Episodic Semi Continous Part picking Partially Single Stochastic Episodic Dynamic Continous robot Refinery controller Partially Single Stochastic Sequential Dynamic Continous Chapter 2 32

Environment types Task Env. Observable Agents Deterministic Episodic Static Discrete Crossword puzzle Fully Single Deterministic Sequential Static Discrete Chess with clock Fully Multi Deterministic Sequential Semi Discrete Poker Partially Multi Stochastic Sequential Static Discrete Backgammon Fully Multi Stochastic Sequential Static Discrete Taxi driving Partially Multi Stochastic Sequential Dynamic Continous Medical diagnosis Partially Multi Stochastic Sequential Dynamic Continous Image analysis Fully Single Deterministic Episodic Semi Continous Part picking Partially Single Stochastic Episodic Dynamic Continous robot Refinery controller Partially Single Stochastic Sequential Dynamic Continous Interactive English Partially Multi Stochastic Sequential Dynamic Discrete tutor Chapter 2 33

Outline Agents and environments Rationality PEAS (Performance measure, Environment, Actuators, Sensors) Environment types The structure of agents Chapter 2 34

agent= architecture + program Agent types Four basic types in order of increasing generality: simple reflex agents reflex agents with state goal-based agents utility-based agents All these can be turned into learning agents Chapter 2 35

The table driven agent function TABLE-DRIVEN-AGENT(percept) returns an action persistent: percepts, a sequence, initially empty table, a table of actions, indexed by percept sequences, initially fully specified append percept to the end of percepts action LOOKUP( percepts, table) return action Let P the set of possible percepts T lifetime of the agent The lookup table will contain: Chapter 2 36

Simple reflex agents Agent Sensors Condition action rules What the world is like now What action I should do now Environment Actuators Chapter 2 37

Example function Reflex-Vacuum-Agent([location,status]) returns an action if status = Dirty then return Suck else if location = A then return Right else if location = B then return Left (setq joe (make-agent :name joe :body (make-agent-body) :program (make-reflex-vacuum-agent-program))) (defun make-reflex-vacuum-agent-program () # (lambda (percept) (let ((location (first percept)) (status (second percept))) (cond ((eq status dirty) Suck) ((eq location A) Right) ((eq location B) Left))))) Ex: car in front is braking (fully observable, one frame) Chapter 2 38

Reflex agents with state State How the world evolves What my actions do Condition action rules Sensors What the world is like now What action I should do now Environment Agent Actuators Ex: best guess for what the world is like now. Chapter 2 39

Example function Reflex-Vacuum-Agent([location,status]) returns an action static: last A, last B, numbers, initially if status = Dirty then... (defun make-reflex-vacuum-agent-with-state-program () (let ((last-a infinity) (last-b infinity)) # (lambda (percept) (let ((location (first percept)) (status (second percept))) (incf last-a) (incf last-b) (cond ((eq status dirty) (if (eq location A) (setq last-a 0) (setq last-b 0)) Suck) ((eq location A) (if (> last-b 3) Right NoOp)) ((eq location B) (if (> last-a 3) Left NoOp))))))) Chapter 2 40

Goal-based agents State How the world evolves What my actions do Goals Sensors What the world is like now What it will be like if I do action A What action I should do now Environment Agent Actuators Chapter 2 41

Utility-based agents State How the world evolves What my actions do Utility Sensors What the world is like now What it will be like if I do action A How happy I will be in such a state What action I should do now Environment Agent Actuators Ex: not only happy/unhappy, handle conflicting goals, uncertain goals. Chapter 2 42

Performance standard Learning agents Critic Sensors feedback learning goals Learning element changes knowledge Performance element Environment Problem generator Agent Actuators Chapter 2 43

Summary Agents interact with environments through actuators and sensors The agent function describes what the agent does in all circumstances The performance measure evaluates the environment sequence A perfectly rational agent maximizes expected performance Agent programs implement (some) agent functions PEAS descriptions define task environments Environments are categorized along several dimensions: observable? deterministic? episodic? static? discrete? single-agent? Several basic agent architectures exist: reflex, reflex with state, goal-based, utility-based, learning Chapter 2 44