CS343 Artificial Intelligence

Similar documents
Lecture 10: Reinforcement Learning

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

Laboratorio di Intelligenza Artificiale e Robotica

Exploration. CS : Deep Reinforcement Learning Sergey Levine

Lecture 1: Basic Concepts of Machine Learning

Laboratorio di Intelligenza Artificiale e Robotica

Python Machine Learning

Welcome to. ECML/PKDD 2004 Community meeting

CSL465/603 - Machine Learning

The 9 th International Scientific Conference elearning and software for Education Bucharest, April 25-26, / X

Lecture 1: Machine Learning Basics

Axiom 2013 Team Description Paper

Reinforcement Learning by Comparing Immediate Reward

Robot Learning Simultaneously a Task and How to Interpret Human Instructions

FF+FPG: Guiding a Policy-Gradient Planner

DOCTOR OF PHILOSOPHY HANDBOOK

Ph.D in Advance Machine Learning (computer science) PhD submitted, degree to be awarded on convocation, sept B.Tech in Computer science and

Agents and environments. Intelligent Agents. Reminders. Vacuum-cleaner world. Outline. A vacuum-cleaner agent. Chapter 2 Actuators

An OO Framework for building Intelligence and Learning properties in Software Agents

Task Completion Transfer Learning for Reward Inference

Chapter 2. Intelligent Agents. Outline. Agents and environments. Rationality. PEAS (Performance measure, Environment, Actuators, Sensors)

Rule-based Expert Systems

Deep search. Enhancing a search bar using machine learning. Ilgün Ilgün & Cedric Reichenbach

Learning Methods for Fuzzy Systems

Computational Data Analysis Techniques In Economics And Finance

Learning and Transferring Relational Instance-Based Policies

Georgetown University at TREC 2017 Dynamic Domain Track

Master s Programme in Computer, Communication and Information Sciences, Study guide , ELEC Majors

Speeding Up Reinforcement Learning with Behavior Transfer

Knowledge-Based - Systems

Task Completion Transfer Learning for Reward Inference

Regret-based Reward Elicitation for Markov Decision Processes

ISFA2008U_120 A SCHEDULING REINFORCEMENT LEARNING ALGORITHM

AMULTIAGENT system [1] can be defined as a group of

ReinForest: Multi-Domain Dialogue Management Using Hierarchical Policies and Knowledge Ontology

Testing A Moving Target: How Do We Test Machine Learning Systems? Peter Varhol Technology Strategy Research, USA

Evolutive Neural Net Fuzzy Filtering: Basic Description

Challenges in Deep Reinforcement Learning. Sergey Levine UC Berkeley

Using focal point learning to improve human machine tacit coordination

Generative models and adversarial training

A survey of multi-view machine learning

Knowledge Transfer in Deep Convolutional Neural Nets

Math-U-See Correlation with the Common Core State Standards for Mathematical Content for Third Grade

Evolution of Symbolisation in Chimpanzees and Neural Nets

IAT 888: Metacreation Machines endowed with creative behavior. Philippe Pasquier Office 565 (floor 14)

Human Emotion Recognition From Speech

Seven Keys to a Positive Learning Environment in Your Classroom. Study Guide

INNOWIZ: A GUIDING FRAMEWORK FOR PROJECTS IN INDUSTRIAL DESIGN EDUCATION

An Investigation into Team-Based Planning

Australian Journal of Basic and Applied Sciences

A Genetic Irrational Belief System

Specification and Evaluation of Machine Translation Toy Systems - Criteria for laboratory assignments

Machine Learning from Garden Path Sentences: The Application of Computational Linguistics

A student diagnosing and evaluation system for laboratory-based academic exercises

CS Machine Learning

arxiv: v1 [cs.lg] 15 Jun 2015

Course Outline. Course Grading. Where to go for help. Academic Integrity. EE-589 Introduction to Neural Networks NN 1 EE

Detecting Wikipedia Vandalism using Machine Learning Notebook for PAN at CLEF 2011

Graphical Data Displays and Database Queries: Helping Users Select the Right Display for the Task

Natural Language Processing: Interpretation, Reasoning and Machine Learning

A Case-Based Approach To Imitation Learning in Robotic Agents

Automatic Discretization of Actions and States in Monte-Carlo Tree Search

Learning Prospective Robot Behavior

Semi-supervised methods of text processing, and an application to medical concept extraction. Yacine Jernite Text-as-Data series September 17.

Seminar - Organic Computing

Information System Design and Development (Advanced Higher) Unit. level 7 (12 SCQF credit points)

Time series prediction

Tutor Coaching Study Research Team

Undergraduate Program Guide. Bachelor of Science. Computer Science DEPARTMENT OF COMPUTER SCIENCE and ENGINEERING

Artificial Neural Networks written examination

Coaching Others for Top Performance 16 Hour Workshop

Financial Accounting Concepts and Research

Reducing Features to Improve Bug Prediction

Probabilistic Latent Semantic Analysis

Intelligent Agents. Chapter 2. Chapter 2 1

Learning Human Utility from Video Demonstrations for Deductive Planning in Robotics

Pod Assignment Guide

Neuro-Symbolic Approaches for Knowledge Representation in Expert Systems

Eduroam Support Clinics What are they?

Truth Inference in Crowdsourcing: Is the Problem Solved?

ALL-IN-ONE MEETING GUIDE THE ECONOMICS OF WELL-BEING

Universal Design for Learning Lesson Plan

Continual Curiosity-Driven Skill Acquisition from High-Dimensional Video Inputs for Humanoid Robots

An investigation of imitation learning algorithms for structured prediction

Universidade do Minho Escola de Engenharia

A Case Study: News Classification Based on Term Frequency

UDL AND LANGUAGE ARTS LESSON OVERVIEW

Learning Optimal Dialogue Strategies: A Case Study of a Spoken Dialogue Agent for

Proposal of Pattern Recognition as a necessary and sufficient principle to Cognitive Science

Language Acquisition Fall 2010/Winter Lexical Categories. Afra Alishahi, Heiner Drenhaus

Semi-Supervised Face Detection

Indicators Teacher understands the active nature of student learning and attains information about levels of development for groups of students.

Computerized Adaptive Psychological Testing A Personalisation Perspective

AUTOMATED TROUBLESHOOTING OF MOBILE NETWORKS USING BAYESIAN NETWORKS

A Bayesian Model of Imitation in Infants and Robots

The whole school approach and pastoral care

Preliminary AGENDA. Practical Applications of Load Resistance Factor Design for Foundation and Earth Retaining System Design and Construction

Pp. 176{182 in Proceedings of The Second International Conference on Knowledge Discovery and Data Mining. Predictive Data Mining with Finite Mixtures

Student Assessment Policy: Education and Counselling

Iterative Cross-Training: An Algorithm for Learning from Unlabeled Web Pages

Transcription:

CS343 Artificial Intelligence Prof: Department of Computer Science The University of Texas at Austin

Good Morning, Colleagues

Good Morning, Colleagues Are there any questions?

Some Context First weeks: search (BFS, A*, minimax, alpha-beta) Find an optimal plan (or solution)

Some Context First weeks: search (BFS, A*, minimax, alpha-beta) Find an optimal plan (or solution) Best thing to do from the current state Know transition and cost (reward) functions

Some Context First weeks: search (BFS, A*, minimax, alpha-beta) Find an optimal plan (or solution) Best thing to do from the current state Know transition and cost (reward) functions Either execute complete solution (deterministic) or search again at every step

Some Context First weeks: search (BFS, A*, minimax, alpha-beta) Find an optimal plan (or solution) Best thing to do from the current state Know transition and cost (reward) functions Either execute complete solution (deterministic) or search again at every step Know current state

Some Context First weeks: search (BFS, A*, minimax, alpha-beta) Find an optimal plan (or solution) Best thing to do from the current state Know transition and cost (reward) functions Either execute complete solution (deterministic) or search again at every step Know current state Next: MDPs

Some Context First weeks: search (BFS, A*, minimax, alpha-beta) Find an optimal plan (or solution) Best thing to do from the current state Know transition and cost (reward) functions Either execute complete solution (deterministic) or search again at every step Know current state Next: MDPs towards reinforcement learning

Some Context First weeks: search (BFS, A*, minimax, alpha-beta) Find an optimal plan (or solution) Best thing to do from the current state Know transition and cost (reward) functions Either execute complete solution (deterministic) or search again at every step Know current state Next: MDPs towards reinforcement learning Still know transition and reward function

Some Context First weeks: search (BFS, A*, minimax, alpha-beta) Find an optimal plan (or solution) Best thing to do from the current state Know transition and cost (reward) functions Either execute complete solution (deterministic) or search again at every step Know current state Next: MDPs towards reinforcement learning Still know transition and reward function Looking for a policy optimal action from every state

Some Context First weeks: search (BFS, A*, minimax, alpha-beta) Find an optimal plan (or solution) Best thing to do from the current state Know transition and cost (reward) functions Either execute complete solution (deterministic) or search again at every step Know current state Next: MDPs towards reinforcement learning Still know transition and reward function Looking for a policy optimal action from every state Action learning: Reinforcement learning

Some Context First weeks: search (BFS, A*, minimax, alpha-beta) Find an optimal plan (or solution) Best thing to do from the current state Know transition and cost (reward) functions Either execute complete solution (deterministic) or search again at every step Know current state Next: MDPs towards reinforcement learning Still know transition and reward function Looking for a policy optimal action from every state Action learning: Reinforcement learning Policy without knowing transition or reward functions

Some Context First weeks: search (BFS, A*, minimax, alpha-beta) Find an optimal plan (or solution) Best thing to do from the current state Know transition and cost (reward) functions Either execute complete solution (deterministic) or search again at every step Know current state Next: MDPs towards reinforcement learning Still know transition and reward function Looking for a policy optimal action from every state Action learning: Reinforcement learning Policy without knowing transition or reward functions Still know state

Some Context (cont.) Probabilistic Reasoning: Now state is unknown Bayesian networks state estimation/inference

Some Context (cont.) Probabilistic Reasoning: Now state is unknown Bayesian networks state estimation/inference Prior, net structure, and CPT s known

Some Context (cont.) Probabilistic Reasoning: Now state is unknown Bayesian networks state estimation/inference Prior, net structure, and CPT s known Week 4: Utilities

Some Context (cont.) Probabilistic Reasoning: Now state is unknown Bayesian networks state estimation/inference Prior, net structure, and CPT s known Week 4: Utilities Week 7: Conditional independence and inference (exact and approximate)

Some Context (cont.) Probabilistic Reasoning: Now state is unknown Bayesian networks state estimation/inference Prior, net structure, and CPT s known Week 4: Utilities Week 7: Conditional independence and inference (exact and approximate) Week 9: State estimation over time

Some Context (cont.) Probabilistic Reasoning: Now state is unknown Bayesian networks state estimation/inference Prior, net structure, and CPT s known Week 4: Utilities Week 7: Conditional independence and inference (exact and approximate) Week 9: State estimation over time Week 9: Utility-based decisions

Some Context (cont.) Probabilistic Reasoning: Now state is unknown Bayesian networks state estimation/inference Prior, net structure, and CPT s known Week 4: Utilities Week 7: Conditional independence and inference (exact and approximate) Week 9: State estimation over time Week 9: Utility-based decisions Week 10: What if they re not known?

Some Context (cont.) Probabilistic Reasoning: Now state is unknown Bayesian networks state estimation/inference Prior, net structure, and CPT s known Week 4: Utilities Week 7: Conditional independence and inference (exact and approximate) Week 9: State estimation over time Week 9: Utility-based decisions Week 10: What if they re not known? Also Bayesian networks for classification

Some Context (cont.) Probabilistic Reasoning: Now state is unknown Bayesian networks state estimation/inference Prior, net structure, and CPT s known Week 4: Utilities Week 7: Conditional independence and inference (exact and approximate) Week 9: State estimation over time Week 9: Utility-based decisions Week 10: What if they re not known? Also Bayesian networks for classification A type of machine learning

Some Context (cont.) After that: More machine learning Week 11: Neural nets and Deep Learning Week 12: SVMs, Kernels, and Clustering Week 13: Classical planning Reasoning with first order representations

Some Context (cont.) After that: More machine learning Week 11: Neural nets and Deep Learning Week 12: SVMs, Kernels, and Clustering Week 13: Classical planning Reasoning with first order representations So far we ve dealt with propositions

Some Context (cont.) After that: More machine learning Week 11: Neural nets and Deep Learning Week 12: SVMs, Kernels, and Clustering Week 13: Classical planning Reasoning with first order representations So far we ve dealt with propositions Back to known transitions, known state, etc.

Some Context (cont.) After that: More machine learning Week 11: Neural nets and Deep Learning Week 12: SVMs, Kernels, and Clustering Week 13: Classical planning Reasoning with first order representations So far we ve dealt with propositions Back to known transitions, known state, etc. Week 14: Philosophical foundations and ethics

Some Context (cont.) After that: More machine learning Week 11: Neural nets and Deep Learning Week 12: SVMs, Kernels, and Clustering Week 13: Classical planning Reasoning with first order representations So far we ve dealt with propositions Back to known transitions, known state, etc. Week 14: Philosophical foundations and ethics It s all about building agents Sense, decide, act

Some Context (cont.) After that: More machine learning Week 11: Neural nets and Deep Learning Week 12: SVMs, Kernels, and Clustering Week 13: Classical planning Reasoning with first order representations So far we ve dealt with propositions Back to known transitions, known state, etc. Week 14: Philosophical foundations and ethics It s all about building agents Sense, decide, act Maximize expected utility

Topics not covered Knowledge representation and reasoning. (Chapters 7-9, 11, 12) Game theory and auctions (Sections 17.5, 17.6) Aspects of learning (Chapters 18, 19) Natural language (Chapters 22, 23) Vision (Chapter 24) Robotics (Chapter 25)