POMDP Learning using Qualitative Belief Spaces

Size: px
Start display at page:

Download "POMDP Learning using Qualitative Belief Spaces"

Transcription

1 POMDP Learning using Qualitative Belief Spaces Bruce D Ambrosio Computer Science Dept. Oregon State University Corvallis, OR dambrosi@research.cs.orst.edu Abstract We present Κ-abstraction as a method for automatically generating small discrete belief spaces for partially observable Markov decision problems (POMDPs). This permits direct application of existing reinforcement learning methods to POMDPs. We show results from applying these methods to a 256 state POMDP, and discuss the types of problems for which the method is suitable. Topic: Algorithms and Architectures Introduction Many ongoing problems, such as monitoring and repair of on-line systems, are naturally formulated as partially observable Markov decision problems (POMDPs). Informally, a Markov decision problem (MDP) model includes a state space model; an action model; a transition model showing how the state space evolves in response to actions; a reward model which reflects (typically) the cost of actions and the rewards (or costs) associated with various states; and a performance model which describes how agent performance is to be scored. A Partially Observable MDP (POMDP) is one in which we assume the agent does not have direct access to the current state, but has only limited (perhaps noisy) evidence (see [Cassandra et al, 94] for a good overview of the POMDP problem and solution methods). Unfortunately, exact solution remains an elusive goal for all but the most trivial POMDPs [Littman et al, 95], and even approximation methods have achieved only limited success [Parr & Russell, 95]. In general, the solution to a Markov decision problem is a policy, a mapping from each state in the state space to the optimal action given that state. The policy can be represented as a valuation across system states, so often solution methods compute this value function, rather than the policy mapping. For a POMDP, either of these is of limited use, since the agent generally does not precisely know the current state. A standard approach is to transform the problem into that of finding a policy which takes as its argument the belief state of the agent, rather than the actual system state. Under the assumption that the agent uses appropriate Bayesian belief updating procedures, it can be shown that an optimal policy for the (presumably fully observable) belief state is also optimal with respect to the underlying POMDP. Unfortunately, even when the underlying state space is discrete, belief state is continuous. As a result, it would seem that methods developed for solving discrete state MDPs would not apply to POMDPs. It has been shown that the value function for a POMDP must be piecewise linear convex. As a result, many POMDP solution methods, exact and approximate, have focused on building representations of the value function as the max of a set of planes in belief X value space 1. This representation has the advantage that it can approximate arbitrarily closely the exact solution (and in some cases can exactly represent the optimal value function). However, it suffers from two disadvantages. First, the number of planes needed tends to grow very rapidly, restricting the approach to very small problems. Second, the size of each vector is linear in the size of the state space. This is a severe limitation in many problems. 1 However, see [Singh et al, 94] for an exception.

2 We are investigating an alternate approach, in which we compute a discrete approximation to the belief space and then use standard reinforcement learning methods to compute optimal Q-value with respect to the discretized state space (A Q-Value is a function mapping from belief X action to value). The primitive element of our representation is a Κ-abstracted belief state. While the representation for a plane in PCL-based methods grows linearly with state-space size, Κ- abstracted belief-state representations grow only with the log of the state space. We have successfully applied this abstraction method to a small test problem (256 states) which is, to our knowledge, the largest POMDP solved to date. In the remainder of this paper we first describe our algorithm, Κ-RL. Next, we present an experimental domain we have been studying, that of on-line maintenance, and the results of some experiments applying Κ-RL in this domain. We close with a discussion of the results and a review of related work. The Κ-RL algorithm Kappa abstraction of belief states The Kappa calculus [Goldszmidt, 95] is a qualitative representation of probability in which a probability is represented as Κ( p) = logε p. Several elegant algorithms have been developed based on this representation, algorithms which we believe have applicability to the POMDP policy computation problem. In this work, however, we merely use the representation itself as an intermediate stage in our qualitative abstraction process. A belief state is a probability distribution over the underlying state space. Assuming states are ordered by decreasing probability, we define a Kappa abstraction of a belief state as: Κ k = s S i iε Κ( p( s ) k That is, an abstraction to a given Κ level is an integer whose value is dependent only the ordering of states down to belief level Κ=k. The transformation of a probability to a Κ value requires an epsilon to serve as base of the logarithm. We use.95*posterior-probability of the highest probability state as our epsilon. That is, the epsilon used is different for each time step. As an example, consider a system with 4 states (s0 - s3), and an agent belief state of [.125,.5,.25,.125], that is, p(s0) =.125, p(s1) =.5, and so on. Then epsilon is.475, and K0 = 1 K1 = 1*1+2*4 = 9 K2 =1*1+2*4+3*16+0*16=57 Alternate encodings, could of course, be developed. In fact, the above encoding unnecessarily collapses non-equivalent states, and perhaps makes some unnecessary distinctions. We will discuss these points later. Under this definition, Kappa abstraction maps a belief state into an integer in the range: 0 < < S k+ 1 Κ k Since the abstraction is discrete, we can directly apply standard reinforcement learning methods, such as Q-learning, to this representation. There are two issues worth discussing before we present experimental results. First, how many Kappa states are there? Second, the computation of belief state from observation data can be complex. Is a method that relies on this computation feasible? The size of the Kappa-abstracted belief space is quite large; therefore, the proposed abstraction does not lend itself to MDP solution techniques which require explicit representation of the entire state-space. Luckily, there exists a class of methods, namely reinforcement learning methods, which only require that we represent explicitly those states actually encountered in the solution process. We have, at this time, no strong theoretical arguments why the number of Κ-states encountered should be small, even for relatively small problems. However, we will show experimental evidence that it remains quite small for surprisingly large problems (e.g., a POMDP with a state space of 256 elements). We will discuss later the circumstances under which we believe this likely to occur. i i

3 We do have strong reasons for believing that the observation-to-κ-abstraction transformation is tractable, at least in our domain of interest. For our domain we use a belief net model, which decomposes the overall state space into a set of component states and adds a set of probabilistic relationships among component states, other aspects of internal system state, and observables. Given this model, incremental inference procedures exist (e.g., IPI, [D Ambrosio, 93]) which can enumerate state space elements in decreasing probability order. The time to obtain the each statespace element is linear in the number of components, or log in the state-space size. Thus, the overall time to compute a Κ-abstraction will be nlog( S ), where n is the number of states with Κ level k or lower. We refer to the worst case n for a problem as the confusabilty of a POMDP. As n approaches 1, the problem looks more and more like a fully observable MDP. Computing the projected belief state We use the Κ-abstraction of the current belief state as our Q-value table index. We must also, however, compute an updated prior for the next stage, based on current beliefs, selected action, and the state transition matrix. Again, this computation could easily grow intractable as the statespace size grows. We again apply the same techniques: we use a belief net model of state transition, and use the same incremental computation methods to compute an approximation of the projected belief. We use the same Κ methods for ranking and selecting state space elements to keep in our representation, but we keep the actual probabilities of these states, rather than Kappa abstracting them. Since we will only be maintaining one Κ belief state in memory at a time, there is no particular advantage to performing the full Κ abstraction on the projected belief. Note that this abstraction process is essentially assigning a zero probability to most elements of the state space. Assuming again that the projected belief state is [.125,.5,.25,.125], we compute the following Κ- projected belief states: Κ 0 = [δ, 1.0-δ, δ, δ] Κ 1 = [δ.66,.33, δ] Κ 2 = [.125,.5,.25,.125] Notice that, in order to allow for the possibility that the actual system state is one of the states below the Kappa cutoff, we insert small non-zero probabilities for all states below the cutoff. Again, we only explicitly represent those states with belief above the cutoff. As long as the number of such states is small, computation is fast and representation size remains tractable. We refer to the K threshold used in table lookup at Kt, and the threshold used in computing the projection as Kp. We use a standard reinforcement learning algorithm, Q-learning, with this representation. We use Q-learning, rather than a possibly more efficient learning of the value function, because of the difficulty of using a value function representation at run time. In our domain the simple one-step lookahead required to compute the optimal act, given a state and a value function, is quite costly 2. The basic update function is, then: Qb (, a) = ( 1 α) Qb (, α) + α( R+ Qb (, α )) k k k Since we expected b k to be sparsely distributed over a very large range, we use a hash table for storing Q-values. Experiments The Task: On-Line Maintenance Our study problem is the in situ diagnosis and repair of simple digital systems. Diagnosis is often formulated as a static, detached process, the goal of which is the assessment of the exact (or most probable) state of some external system. In contrast, we view diagnosis as a dynamic, practical activity by an agent engaged with a changing and uncertain world. Further, we extend the task to 2 Using a value table requires one step look-ahead at run time, but look-ahead requires taking an expectation over possible observations. This requires performing the observation to belief transformation for every possible next observation.

4 include the repair task to focus diagnostic activity. Our formulation of embedded diagnosis has the following characteristics: (1) the equipment 3 under diagnosis continues to operate while being diagnosed, and multiple faults can occur (and can continue to occur after an initial fault is detected); (2) the agent has only limited observational ability: it cannot directly observe component state; (3) the agent senses equipment operation through a set of fixed sensors and one or more movable probes; (4) there is a known fixed cost per unit time while the equipment is malfunctioning (i.e., any component is in a faulted state); (5) action alternatives include probing test points, replacing individual components, and simply waiting for the next sense report - each action has a corresponding cost; (6) the agent can only perform one action at a time; and (7) the overall task is to minimize total cost over some extended time period during which several failures can be expected to occur. We term this task the On-Line Maintenance task, and an agent intended for performing such a task an On-Line Maintenance Agent (OLMA) [D Ambrosio, 92 & 96]. An interesting aspect of this formulation, from the perspective of the machine diagnosis community, is that diagnosis is not a direct goal. A precise diagnosis is neither always obtainable nor necessary. Indeed, it is not obvious a-priori what elements of a diagnosis are even relevant to the decision at hand. One final comment: the problem is surprisingly complex. The simple problem instance studied in this paper is well beyond the capability of current exact POMDP solution methods (the MDP state space for the simple 4 gate problem studied in this paper has 256 states, ignoring the stochastic behavior of our model of the unknown mode!). Deterministic policies which only consider current observations perform quite poorly, failing to ever repair some faults. Stochastic policies can perform reasonably well, but that is the subject of another paper. The four gate circuit Our first study problem in this domain was a simple four gate digital circuit known as a half adder. The circuit diagram is shown in figure 2, and the corresponding belief network is shown in figure 3. Each component is modeled as having four possible states, ok, stuck-at0 (in which the output is always 0, regardless of the inputs), stuck-at1, and unknown (in which the output is a stochastic function, independent of the inputs, and uniformly distributed over {0,1}). Since each gate has four possible states, the overall state space has 4 4 or 256 states (multiple faults are possible, thus all 256 states are reachable, and even likely over the length of the long training runs we used). We used uniform failure probabilities, (.002 for each failure state), chosen to produce an interesting number of failures over a reasonable length simulated test run. The agent was given the values of I1, I2, O, and Carry as its standard observation set. Possible actions included the replacement of any one component, the probe of either P1 or P2 (in which case the respective value was added to the observation set for the next cycle), or no action. The reward was -1 for each cycle in which at least one component was faulted, plus -6 for a replacement action or -1 for a probe action. We trained using an initial alpha of.2, decreasing to 0 over the course of an epoch. After each epoch we ran a short evaluation run in which we re-initialized the simulation to a known good state, and measured the total reward over the evaluation run. We then reset alpha to.2 and ran another epoch (without resetting the Q-value table). The first epoch was 8000 iterations. Subsequent epochs doubled the length of the previous epoch (i.e., 16000, 32000, 64000, ). We terminated training after the iteration for Κt=0,1 and for Κt=2,3. 3 We will use system or agent to refer to our diagnostic system, and equipment to refer to the target physical system.

5 Nor1 Nor2 I1 Nor1 P1 Nor2 O I1 P1 O I2 I2 P2 Carry Nand P2 Not Carry Nand Not Figure 1: 4 Gate Circuit and Bayes Net 2500 Cost/Fail & TableSize Tablesize Cost/Fail Training cycles Figure 2: Typical Cost/Failure(*60) and Table Size during training Figure 2 above shows behavior over time for a typical training run. Cost/Failure is multiplied by 60 to put it on the same scale as Q table size. The algorithm very rapidly learns to repair single faults. As a result, the first cost/failure data point (at 8000 cycles) is already reasonably low. Table 1 below shows the number of entries in the Q-value table (state X action) at the end of training for the twelve combinations of Κp and Κt we explored. We considered the possibility that the problem is such that the actual number of belief states visited is low. We tested this possibility by training using the exact belief state rather than the K- abstracted belief state. Up to a limit of 12,500 iterations (at which point the program ran out of swap space - exact belief states are large objects!) the number of revisited exact belief states remained insignificant. 4 Table 1 also shows our performance metric, cost/failure, for each of these combinations. These values are averages over failures per cell. As a comparison, our best finite-lookahead online algorithm achieves a cost/failure of 21.0 on this same problem when it is not charged for computation time (the on-line algorithm takes about 8 seconds per decision, whereas the Κ-rl policy can be executed in under ½ second, including computing the Κ-abstracted belief state and computed the Κ-projection for the next belief state. We see that the policy (especially when Κp = 4) significantly outperforms the on-line algorithm (see [D Ambrosio, UAI-96] for details on the on-line approaches we have tried). The difference between row 0 and rows 2 and 4 is statistically significant. 4 We truncated at 5 significant digits for comparison only to reduce numerical stability problems.

6 Kp\Kt Row Avg / / / / / / / / / / / / / / /18.25 Col Avg 2207/ / / /21.6 Table 1: Table Size/Cost-Per-Failure for 4 gate circuit The seven gate circuit The confusability for the half-adder is relatively low. An examination of the circuit reveals that we have enough information to distinguish between faults occurring in the upper and lower halves of the circuit, so at most two components can be the cause of any single fault. Any Κ level that excludes multiple simultaneous faults (as do many of those we tested), will encounter a maximum of four states at Κ 0. To make the problem more challenging we increased the confusability index of our second test problem, a seven gate circuit. Unfortunately, Κ-RL is not tractable for the full seven gate problem. The problem is not that the number of Κ-abstracted states would grow too large to represent, but rather that learning would take too long to converge. The most difficult Κ- states to learn Q-values for are the intermediate states which occur shortly after a state change is detected and before it has been isolated. These states tend to be transitory and infrequently occurring. As a result, very long training runs would be required (we estimate at least 10 8 cycles) before convergence. As a warm up to attacking the full problem, we ran an experiment in which we restricted the simulation to producing only one fault at any one time. That is, once one component had faulted, another fault could not occur until the first was repaired. While this might seem to make the problem trivial (the single fault state space has only 22 states), two complications make this an interesting problem. First, we constructed the experiment such that the confusability was much higher (as many as 5 gates could be confused for some single faults). Second, while the simulation was restricted to single fault mode, the learner was not restricted to considering only single faults. This caused it to enter a subset of the multiple fault belief space: Situations occurred when the learner was not sure it had repaired a fault; that is, a number of states other than the all ok state were in the Κ-abstracted projected belief state. However, the fault actually had been repaired, and a second fault occurred. As far as the agent can tell, this may in fact be a multiplefault state. We achieved similar results on this problem to those on the 4 gate problem. Q-table size at convergence ranged from 2800 (Kt=0, Kp=4) to 15,300 (Kt=4, Kp=4), and the performance metric (cost/failure) showed only small performance differences with Kt or Kp variation over the range tested (0, 2, 4). Space precludes further discussion, details are available in the full paper. Discussion Table size remains quite tractable for both the 4 gate and 7 gate problems. We believe further training would not significantly increase the number of states encountered: plots of table-size vs length of training clearly show asymptotic behavior, and performance over a large number of failures indicates convergence has been attained. There are two characteristics of our study domain that, together, enabled our approach. First, the domain has low confusability. That is, at any time the agent has a pretty good idea where it is in the state space. Second, the belief models are such that the high probability states can be identified quickly. The characteristic which enables this is skewness of the probability distributions (see [D Ambrosio: UAI-93] for details. We discovered rather late several problems with our Κ-abstraction process. The current abstraction distinguishes between states with same Κ values, but different ordering. Conversely, it fails to distinguish between states with or without state-space element zero. Also, it maps

7 1(0)+3(0) (Κ 0 ) into the same value as 0(0)+2(1)(Κ 1 ). Presumably a better mapping would only improve the performance of the algorithm. We do not believe that the approach as it stands scales well. As discussed earlier, the primary problem is not space, but rather training time. There are simply too many possible multiple faults to visit each enough times to assure convergence over the reachable Κ-abstracted belief space. We are exploring further abstraction methods to cope with this problem. Related Work Jordan and Singh [Singh et al, 94] have studied stochastic policies for POMDPs. We have not yet applied their methods to our domain, but expect that the resulting on-line performance will not match that of Κ-rl generated policies. Kaebling et al [Littman et al, 95] have developed exact algorithms for POMDPs. Russell and Parr [Parr & Russell, 95] have developed an elegant approximation method for compactly representing value functions. We attempted to apply their method to the 4 gate problem, but were unable to obtain convergence with up to 8 planes. We are continuing to study their approach, and may have more to report in the final version of the paper. However, we note that even 8 planes takes more space than the largest K-abstracted Q-values table in table 1. Conclusion We have presented Κ-abstraction, a method for automatically generating qualitative belief spaces for POMDPs. We have shown successful application to a 256 state problem, and discussed problem characteristics required for successful application of the method (the key being low confusability). Finally, we have concluded that further scaling up will require even more powerful abstraction methods. Acknowledgments This work done with the support of NSF grants IRI and CDA It benefitted from many useful discussions with Robert Fung, Thomas Dietterich, and Prasad Tadepalli. References [Cassandra et al, 94] Acting Optimally in Partially Observable Stochastic Domains. Brown Univ. Tech. Report CS [D Ambrosio, 92]. B. D Ambrosio. Value-driven real-time diagnosis. In Proceedings, Third International Workshop on the Principles of Diagnosis, October [D Ambrosio, 93] Incremental Probabilistic Inference. In Proceedings of the Ninth Annual Conference on Uncertainty in Artificial Intelligence, pp , July Morgan Kaufmann, Publishers. [D Ambrosio, 96] Some Experiments with Real-time Decision Algorithms. To appear In Proceedings of the Twelfth Annual Conference on Uncertainty in Artificial Intelligence, July Morgan Kaufmann, Publishers. [Goldszmidt, 95] M. Goldszmidt. Fast Belief Updating Using Order of Magnitude Probabilities. In Uncertainty in Artificial Intelligence, Proceedings of the Eleventh Conference. pp Morgan Kaufmann, Publishers, July, [Littman et al, 95] Littman, M., Cassandra, A. and Kaebling, L. Learning Policies for Partially Observable Environments: Scaling Up. In Machine Learning, 12: [Parr & Russell, 95] Approximating Optimal Policies for Partially Observable Stochastic Domains. Proceedings IJCAI95, pp [Singh et al, 94] Singh, S., Jaakola, T. and Jordan, M. Learning without State-Estimation in Partially Observable Markovian Decision Processes. In Machine Learning: Proceedings of the Eleventh International Conference. (1994):

Lecture 1: Machine Learning Basics

Lecture 1: Machine Learning Basics 1/69 Lecture 1: Machine Learning Basics Ali Harakeh University of Waterloo WAVE Lab ali.harakeh@uwaterloo.ca May 1, 2017 2/69 Overview 1 Learning Algorithms 2 Capacity, Overfitting, and Underfitting 3

More information

Lecture 10: Reinforcement Learning

Lecture 10: Reinforcement Learning Lecture 1: Reinforcement Learning Cognitive Systems II - Machine Learning SS 25 Part III: Learning Programs and Strategies Q Learning, Dynamic Programming Lecture 1: Reinforcement Learning p. Motivation

More information

Reinforcement Learning by Comparing Immediate Reward

Reinforcement Learning by Comparing Immediate Reward Reinforcement Learning by Comparing Immediate Reward Punit Pandey DeepshikhaPandey Dr. Shishir Kumar Abstract This paper introduces an approach to Reinforcement Learning Algorithm by comparing their immediate

More information

Software Maintenance

Software Maintenance 1 What is Software Maintenance? Software Maintenance is a very broad activity that includes error corrections, enhancements of capabilities, deletion of obsolete capabilities, and optimization. 2 Categories

More information

Exploration. CS : Deep Reinforcement Learning Sergey Levine

Exploration. CS : Deep Reinforcement Learning Sergey Levine Exploration CS 294-112: Deep Reinforcement Learning Sergey Levine Class Notes 1. Homework 4 due on Wednesday 2. Project proposal feedback sent Today s Lecture 1. What is exploration? Why is it a problem?

More information

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur Module 12 Machine Learning 12.1 Instructional Objective The students should understand the concept of learning systems Students should learn about different aspects of a learning system Students should

More information

Introduction to Simulation

Introduction to Simulation Introduction to Simulation Spring 2010 Dr. Louis Luangkesorn University of Pittsburgh January 19, 2010 Dr. Louis Luangkesorn ( University of Pittsburgh ) Introduction to Simulation January 19, 2010 1 /

More information

Regret-based Reward Elicitation for Markov Decision Processes

Regret-based Reward Elicitation for Markov Decision Processes 444 REGAN & BOUTILIER UAI 2009 Regret-based Reward Elicitation for Markov Decision Processes Kevin Regan Department of Computer Science University of Toronto Toronto, ON, CANADA kmregan@cs.toronto.edu

More information

The Good Judgment Project: A large scale test of different methods of combining expert predictions

The Good Judgment Project: A large scale test of different methods of combining expert predictions The Good Judgment Project: A large scale test of different methods of combining expert predictions Lyle Ungar, Barb Mellors, Jon Baron, Phil Tetlock, Jaime Ramos, Sam Swift The University of Pennsylvania

More information

Planning with External Events

Planning with External Events 94 Planning with External Events Jim Blythe School of Computer Science Carnegie Mellon University Pittsburgh, PA 15213 blythe@cs.cmu.edu Abstract I describe a planning methodology for domains with uncertainty

More information

A Reinforcement Learning Variant for Control Scheduling

A Reinforcement Learning Variant for Control Scheduling A Reinforcement Learning Variant for Control Scheduling Aloke Guha Honeywell Sensor and System Development Center 3660 Technology Drive Minneapolis MN 55417 Abstract We present an algorithm based on reinforcement

More information

AUTOMATED TROUBLESHOOTING OF MOBILE NETWORKS USING BAYESIAN NETWORKS

AUTOMATED TROUBLESHOOTING OF MOBILE NETWORKS USING BAYESIAN NETWORKS AUTOMATED TROUBLESHOOTING OF MOBILE NETWORKS USING BAYESIAN NETWORKS R.Barco 1, R.Guerrero 2, G.Hylander 2, L.Nielsen 3, M.Partanen 2, S.Patel 4 1 Dpt. Ingeniería de Comunicaciones. Universidad de Málaga.

More information

Discriminative Learning of Beam-Search Heuristics for Planning

Discriminative Learning of Beam-Search Heuristics for Planning Discriminative Learning of Beam-Search Heuristics for Planning Yuehua Xu School of EECS Oregon State University Corvallis,OR 97331 xuyu@eecs.oregonstate.edu Alan Fern School of EECS Oregon State University

More information

OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS

OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS Václav Kocian, Eva Volná, Michal Janošek, Martin Kotyrba University of Ostrava Department of Informatics and Computers Dvořákova 7,

More information

Artificial Neural Networks written examination

Artificial Neural Networks written examination 1 (8) Institutionen för informationsteknologi Olle Gällmo Universitetsadjunkt Adress: Lägerhyddsvägen 2 Box 337 751 05 Uppsala Artificial Neural Networks written examination Monday, May 15, 2006 9 00-14

More information

ReinForest: Multi-Domain Dialogue Management Using Hierarchical Policies and Knowledge Ontology

ReinForest: Multi-Domain Dialogue Management Using Hierarchical Policies and Knowledge Ontology ReinForest: Multi-Domain Dialogue Management Using Hierarchical Policies and Knowledge Ontology Tiancheng Zhao CMU-LTI-16-006 Language Technologies Institute School of Computer Science Carnegie Mellon

More information

Axiom 2013 Team Description Paper

Axiom 2013 Team Description Paper Axiom 2013 Team Description Paper Mohammad Ghazanfari, S Omid Shirkhorshidi, Farbod Samsamipour, Hossein Rahmatizadeh Zagheli, Mohammad Mahdavi, Payam Mohajeri, S Abbas Alamolhoda Robotics Scientific Association

More information

On the Combined Behavior of Autonomous Resource Management Agents

On the Combined Behavior of Autonomous Resource Management Agents On the Combined Behavior of Autonomous Resource Management Agents Siri Fagernes 1 and Alva L. Couch 2 1 Faculty of Engineering Oslo University College Oslo, Norway siri.fagernes@iu.hio.no 2 Computer Science

More information

Speeding Up Reinforcement Learning with Behavior Transfer

Speeding Up Reinforcement Learning with Behavior Transfer Speeding Up Reinforcement Learning with Behavior Transfer Matthew E. Taylor and Peter Stone Department of Computer Sciences The University of Texas at Austin Austin, Texas 78712-1188 {mtaylor, pstone}@cs.utexas.edu

More information

Learning Optimal Dialogue Strategies: A Case Study of a Spoken Dialogue Agent for

Learning Optimal Dialogue Strategies: A Case Study of a Spoken Dialogue Agent for Learning Optimal Dialogue Strategies: A Case Study of a Spoken Dialogue Agent for Email Marilyn A. Walker Jeanne C. Fromer Shrikanth Narayanan walker@research.att.com jeannie@ai.mit.edu shri@research.att.com

More information

Intelligent Agents. Chapter 2. Chapter 2 1

Intelligent Agents. Chapter 2. Chapter 2 1 Intelligent Agents Chapter 2 Chapter 2 1 Outline Agents and environments Rationality PEAS (Performance measure, Environment, Actuators, Sensors) Environment types The structure of agents Chapter 2 2 Agents

More information

A Case Study: News Classification Based on Term Frequency

A Case Study: News Classification Based on Term Frequency A Case Study: News Classification Based on Term Frequency Petr Kroha Faculty of Computer Science University of Technology 09107 Chemnitz Germany kroha@informatik.tu-chemnitz.de Ricardo Baeza-Yates Center

More information

FF+FPG: Guiding a Policy-Gradient Planner

FF+FPG: Guiding a Policy-Gradient Planner FF+FPG: Guiding a Policy-Gradient Planner Olivier Buffet LAAS-CNRS University of Toulouse Toulouse, France firstname.lastname@laas.fr Douglas Aberdeen National ICT australia & The Australian National University

More information

Probabilistic Latent Semantic Analysis

Probabilistic Latent Semantic Analysis Probabilistic Latent Semantic Analysis Thomas Hofmann Presentation by Ioannis Pavlopoulos & Andreas Damianou for the course of Data Mining & Exploration 1 Outline Latent Semantic Analysis o Need o Overview

More information

Litterature review of Soft Systems Methodology

Litterature review of Soft Systems Methodology Thomas Schmidt nimrod@mip.sdu.dk October 31, 2006 The primary ressource for this reivew is Peter Checklands article Soft Systems Metodology, secondary ressources are the book Soft Systems Methodology in

More information

Action Models and their Induction

Action Models and their Induction Action Models and their Induction Michal Čertický, Comenius University, Bratislava certicky@fmph.uniba.sk March 5, 2013 Abstract By action model, we understand any logic-based representation of effects

More information

ISFA2008U_120 A SCHEDULING REINFORCEMENT LEARNING ALGORITHM

ISFA2008U_120 A SCHEDULING REINFORCEMENT LEARNING ALGORITHM Proceedings of 28 ISFA 28 International Symposium on Flexible Automation Atlanta, GA, USA June 23-26, 28 ISFA28U_12 A SCHEDULING REINFORCEMENT LEARNING ALGORITHM Amit Gil, Helman Stern, Yael Edan, and

More information

Thesis-Proposal Outline/Template

Thesis-Proposal Outline/Template Thesis-Proposal Outline/Template Kevin McGee 1 Overview This document provides a description of the parts of a thesis outline and an example of such an outline. It also indicates which parts should be

More information

Seminar - Organic Computing

Seminar - Organic Computing Seminar - Organic Computing Self-Organisation of OC-Systems Markus Franke 25.01.2006 Typeset by FoilTEX Timetable 1. Overview 2. Characteristics of SO-Systems 3. Concern with Nature 4. Design-Concepts

More information

POLA: a student modeling framework for Probabilistic On-Line Assessment of problem solving performance

POLA: a student modeling framework for Probabilistic On-Line Assessment of problem solving performance POLA: a student modeling framework for Probabilistic On-Line Assessment of problem solving performance Cristina Conati, Kurt VanLehn Intelligent Systems Program University of Pittsburgh Pittsburgh, PA,

More information

Continual Curiosity-Driven Skill Acquisition from High-Dimensional Video Inputs for Humanoid Robots

Continual Curiosity-Driven Skill Acquisition from High-Dimensional Video Inputs for Humanoid Robots Continual Curiosity-Driven Skill Acquisition from High-Dimensional Video Inputs for Humanoid Robots Varun Raj Kompella, Marijn Stollenga, Matthew Luciw, Juergen Schmidhuber The Swiss AI Lab IDSIA, USI

More information

University of Groningen. Systemen, planning, netwerken Bosman, Aart

University of Groningen. Systemen, planning, netwerken Bosman, Aart University of Groningen Systemen, planning, netwerken Bosman, Aart IMPORTANT NOTE: You are advised to consult the publisher's version (publisher's PDF) if you wish to cite from it. Please check the document

More information

Knowledge based expert systems D H A N A N J A Y K A L B A N D E

Knowledge based expert systems D H A N A N J A Y K A L B A N D E Knowledge based expert systems D H A N A N J A Y K A L B A N D E What is a knowledge based system? A Knowledge Based System or a KBS is a computer program that uses artificial intelligence to solve problems

More information

SARDNET: A Self-Organizing Feature Map for Sequences

SARDNET: A Self-Organizing Feature Map for Sequences SARDNET: A Self-Organizing Feature Map for Sequences Daniel L. James and Risto Miikkulainen Department of Computer Sciences The University of Texas at Austin Austin, TX 78712 dljames,risto~cs.utexas.edu

More information

Learning Cases to Resolve Conflicts and Improve Group Behavior

Learning Cases to Resolve Conflicts and Improve Group Behavior From: AAAI Technical Report WS-96-02. Compilation copyright 1996, AAAI (www.aaai.org). All rights reserved. Learning Cases to Resolve Conflicts and Improve Group Behavior Thomas Haynes and Sandip Sen Department

More information

Self Study Report Computer Science

Self Study Report Computer Science Computer Science undergraduate students have access to undergraduate teaching, and general computing facilities in three buildings. Two large classrooms are housed in the Davis Centre, which hold about

More information

Rule Learning With Negation: Issues Regarding Effectiveness

Rule Learning With Negation: Issues Regarding Effectiveness Rule Learning With Negation: Issues Regarding Effectiveness S. Chua, F. Coenen, G. Malcolm University of Liverpool Department of Computer Science, Ashton Building, Ashton Street, L69 3BX Liverpool, United

More information

ENME 605 Advanced Control Systems, Fall 2015 Department of Mechanical Engineering

ENME 605 Advanced Control Systems, Fall 2015 Department of Mechanical Engineering ENME 605 Advanced Control Systems, Fall 2015 Department of Mechanical Engineering Lecture Details Instructor Course Objectives Tuesday and Thursday, 4:00 pm to 5:15 pm Information Technology and Engineering

More information

Lecture 1: Basic Concepts of Machine Learning

Lecture 1: Basic Concepts of Machine Learning Lecture 1: Basic Concepts of Machine Learning Cognitive Systems - Machine Learning Ute Schmid (lecture) Johannes Rabold (practice) Based on slides prepared March 2005 by Maximilian Röglinger, updated 2010

More information

An empirical study of learning speed in backpropagation

An empirical study of learning speed in backpropagation Carnegie Mellon University Research Showcase @ CMU Computer Science Department School of Computer Science 1988 An empirical study of learning speed in backpropagation networks Scott E. Fahlman Carnegie

More information

Python Machine Learning

Python Machine Learning Python Machine Learning Unlock deeper insights into machine learning with this vital guide to cuttingedge predictive analytics Sebastian Raschka [ PUBLISHING 1 open source I community experience distilled

More information

Iterative Cross-Training: An Algorithm for Learning from Unlabeled Web Pages

Iterative Cross-Training: An Algorithm for Learning from Unlabeled Web Pages Iterative Cross-Training: An Algorithm for Learning from Unlabeled Web Pages Nuanwan Soonthornphisaj 1 and Boonserm Kijsirikul 2 Machine Intelligence and Knowledge Discovery Laboratory Department of Computer

More information

Notes on The Sciences of the Artificial Adapted from a shorter document written for course (Deciding What to Design) 1

Notes on The Sciences of the Artificial Adapted from a shorter document written for course (Deciding What to Design) 1 Notes on The Sciences of the Artificial Adapted from a shorter document written for course 17-652 (Deciding What to Design) 1 Ali Almossawi December 29, 2005 1 Introduction The Sciences of the Artificial

More information

AGENDA LEARNING THEORIES LEARNING THEORIES. Advanced Learning Theories 2/22/2016

AGENDA LEARNING THEORIES LEARNING THEORIES. Advanced Learning Theories 2/22/2016 AGENDA Advanced Learning Theories Alejandra J. Magana, Ph.D. admagana@purdue.edu Introduction to Learning Theories Role of Learning Theories and Frameworks Learning Design Research Design Dual Coding Theory

More information

Test Effort Estimation Using Neural Network

Test Effort Estimation Using Neural Network J. Software Engineering & Applications, 2010, 3: 331-340 doi:10.4236/jsea.2010.34038 Published Online April 2010 (http://www.scirp.org/journal/jsea) 331 Chintala Abhishek*, Veginati Pavan Kumar, Harish

More information

AMULTIAGENT system [1] can be defined as a group of

AMULTIAGENT system [1] can be defined as a group of 156 IEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS PART C: APPLICATIONS AND REVIEWS, VOL. 38, NO. 2, MARCH 2008 A Comprehensive Survey of Multiagent Reinforcement Learning Lucian Buşoniu, Robert Babuška,

More information

Math-U-See Correlation with the Common Core State Standards for Mathematical Content for Third Grade

Math-U-See Correlation with the Common Core State Standards for Mathematical Content for Third Grade Math-U-See Correlation with the Common Core State Standards for Mathematical Content for Third Grade The third grade standards primarily address multiplication and division, which are covered in Math-U-See

More information

Improving Action Selection in MDP s via Knowledge Transfer

Improving Action Selection in MDP s via Knowledge Transfer In Proc. 20th National Conference on Artificial Intelligence (AAAI-05), July 9 13, 2005, Pittsburgh, USA. Improving Action Selection in MDP s via Knowledge Transfer Alexander A. Sherstov and Peter Stone

More information

Knowledge Transfer in Deep Convolutional Neural Nets

Knowledge Transfer in Deep Convolutional Neural Nets Knowledge Transfer in Deep Convolutional Neural Nets Steven Gutstein, Olac Fuentes and Eric Freudenthal Computer Science Department University of Texas at El Paso El Paso, Texas, 79968, U.S.A. Abstract

More information

Mathematics subject curriculum

Mathematics subject curriculum Mathematics subject curriculum Dette er ei omsetjing av den fastsette læreplanteksten. Læreplanen er fastsett på Nynorsk Established as a Regulation by the Ministry of Education and Research on 24 June

More information

CS Machine Learning

CS Machine Learning CS 478 - Machine Learning Projects Data Representation Basic testing and evaluation schemes CS 478 Data and Testing 1 Programming Issues l Program in any platform you want l Realize that you will be doing

More information

Firms and Markets Saturdays Summer I 2014

Firms and Markets Saturdays Summer I 2014 PRELIMINARY DRAFT VERSION. SUBJECT TO CHANGE. Firms and Markets Saturdays Summer I 2014 Professor Thomas Pugel Office: Room 11-53 KMC E-mail: tpugel@stern.nyu.edu Tel: 212-998-0918 Fax: 212-995-4212 This

More information

Georgetown University at TREC 2017 Dynamic Domain Track

Georgetown University at TREC 2017 Dynamic Domain Track Georgetown University at TREC 2017 Dynamic Domain Track Zhiwen Tang Georgetown University zt79@georgetown.edu Grace Hui Yang Georgetown University huiyang@cs.georgetown.edu Abstract TREC Dynamic Domain

More information

Task Completion Transfer Learning for Reward Inference

Task Completion Transfer Learning for Reward Inference Task Completion Transfer Learning for Reward Inference Layla El Asri 1,2, Romain Laroche 1, Olivier Pietquin 3 1 Orange Labs, Issy-les-Moulineaux, France 2 UMI 2958 (CNRS - GeorgiaTech), France 3 University

More information

Laboratorio di Intelligenza Artificiale e Robotica

Laboratorio di Intelligenza Artificiale e Robotica Laboratorio di Intelligenza Artificiale e Robotica A.A. 2008-2009 Outline 2 Machine Learning Unsupervised Learning Supervised Learning Reinforcement Learning Genetic Algorithms Genetics-Based Machine Learning

More information

The 9 th International Scientific Conference elearning and software for Education Bucharest, April 25-26, / X

The 9 th International Scientific Conference elearning and software for Education Bucharest, April 25-26, / X The 9 th International Scientific Conference elearning and software for Education Bucharest, April 25-26, 2013 10.12753/2066-026X-13-154 DATA MINING SOLUTIONS FOR DETERMINING STUDENT'S PROFILE Adela BÂRA,

More information

Visual CP Representation of Knowledge

Visual CP Representation of Knowledge Visual CP Representation of Knowledge Heather D. Pfeiffer and Roger T. Hartley Department of Computer Science New Mexico State University Las Cruces, NM 88003-8001, USA email: hdp@cs.nmsu.edu and rth@cs.nmsu.edu

More information

TD(λ) and Q-Learning Based Ludo Players

TD(λ) and Q-Learning Based Ludo Players TD(λ) and Q-Learning Based Ludo Players Majed Alhajry, Faisal Alvi, Member, IEEE and Moataz Ahmed Abstract Reinforcement learning is a popular machine learning technique whose inherent self-learning ability

More information

CSL465/603 - Machine Learning

CSL465/603 - Machine Learning CSL465/603 - Machine Learning Fall 2016 Narayanan C Krishnan ckn@iitrpr.ac.in Introduction CSL465/603 - Machine Learning 1 Administrative Trivia Course Structure 3-0-2 Lecture Timings Monday 9.55-10.45am

More information

IAT 888: Metacreation Machines endowed with creative behavior. Philippe Pasquier Office 565 (floor 14)

IAT 888: Metacreation Machines endowed with creative behavior. Philippe Pasquier Office 565 (floor 14) IAT 888: Metacreation Machines endowed with creative behavior Philippe Pasquier Office 565 (floor 14) pasquier@sfu.ca Outline of today's lecture A little bit about me A little bit about you What will that

More information

WE GAVE A LAWYER BASIC MATH SKILLS, AND YOU WON T BELIEVE WHAT HAPPENED NEXT

WE GAVE A LAWYER BASIC MATH SKILLS, AND YOU WON T BELIEVE WHAT HAPPENED NEXT WE GAVE A LAWYER BASIC MATH SKILLS, AND YOU WON T BELIEVE WHAT HAPPENED NEXT PRACTICAL APPLICATIONS OF RANDOM SAMPLING IN ediscovery By Matthew Verga, J.D. INTRODUCTION Anyone who spends ample time working

More information

Given a directed graph G =(N A), where N is a set of m nodes and A. destination node, implying a direction for ow to follow. Arcs have limitations

Given a directed graph G =(N A), where N is a set of m nodes and A. destination node, implying a direction for ow to follow. Arcs have limitations 4 Interior point algorithms for network ow problems Mauricio G.C. Resende AT&T Bell Laboratories, Murray Hill, NJ 07974-2070 USA Panos M. Pardalos The University of Florida, Gainesville, FL 32611-6595

More information

Executive Guide to Simulation for Health

Executive Guide to Simulation for Health Executive Guide to Simulation for Health Simulation is used by Healthcare and Human Service organizations across the World to improve their systems of care and reduce costs. Simulation offers evidence

More information

10.2. Behavior models

10.2. Behavior models User behavior research 10.2. Behavior models Overview Why do users seek information? How do they seek information? How do they search for information? How do they use libraries? These questions are addressed

More information

AN EXAMPLE OF THE GOMORY CUTTING PLANE ALGORITHM. max z = 3x 1 + 4x 2. 3x 1 x x x x N 2

AN EXAMPLE OF THE GOMORY CUTTING PLANE ALGORITHM. max z = 3x 1 + 4x 2. 3x 1 x x x x N 2 AN EXAMPLE OF THE GOMORY CUTTING PLANE ALGORITHM Consider the integer programme subject to max z = 3x 1 + 4x 2 3x 1 x 2 12 3x 1 + 11x 2 66 The first linear programming relaxation is subject to x N 2 max

More information

Specification and Evaluation of Machine Translation Toy Systems - Criteria for laboratory assignments

Specification and Evaluation of Machine Translation Toy Systems - Criteria for laboratory assignments Specification and Evaluation of Machine Translation Toy Systems - Criteria for laboratory assignments Cristina Vertan, Walther v. Hahn University of Hamburg, Natural Language Systems Division Hamburg,

More information

Level 6. Higher Education Funding Council for England (HEFCE) Fee for 2017/18 is 9,250*

Level 6. Higher Education Funding Council for England (HEFCE) Fee for 2017/18 is 9,250* Programme Specification: Undergraduate For students starting in Academic Year 2017/2018 1. Course Summary Names of programme(s) and award title(s) Award type Mode of study Framework of Higher Education

More information

(Sub)Gradient Descent

(Sub)Gradient Descent (Sub)Gradient Descent CMSC 422 MARINE CARPUAT marine@cs.umd.edu Figures credit: Piyush Rai Logistics Midterm is on Thursday 3/24 during class time closed book/internet/etc, one page of notes. will include

More information

Reducing Features to Improve Bug Prediction

Reducing Features to Improve Bug Prediction Reducing Features to Improve Bug Prediction Shivkumar Shivaji, E. James Whitehead, Jr., Ram Akella University of California Santa Cruz {shiv,ejw,ram}@soe.ucsc.edu Sunghun Kim Hong Kong University of Science

More information

Learning to Schedule Straight-Line Code

Learning to Schedule Straight-Line Code Learning to Schedule Straight-Line Code Eliot Moss, Paul Utgoff, John Cavazos Doina Precup, Darko Stefanović Dept. of Comp. Sci., Univ. of Mass. Amherst, MA 01003 Carla Brodley, David Scheeff Sch. of Elec.

More information

Learning Methods for Fuzzy Systems

Learning Methods for Fuzzy Systems Learning Methods for Fuzzy Systems Rudolf Kruse and Andreas Nürnberger Department of Computer Science, University of Magdeburg Universitätsplatz, D-396 Magdeburg, Germany Phone : +49.39.67.876, Fax : +49.39.67.8

More information

Switchboard Language Model Improvement with Conversational Data from Gigaword

Switchboard Language Model Improvement with Conversational Data from Gigaword Katholieke Universiteit Leuven Faculty of Engineering Master in Artificial Intelligence (MAI) Speech and Language Technology (SLT) Switchboard Language Model Improvement with Conversational Data from Gigaword

More information

Computer Science 141: Computing Hardware Course Information Fall 2012

Computer Science 141: Computing Hardware Course Information Fall 2012 Computer Science 141: Computing Hardware Course Information Fall 2012 September 4, 2012 1 Outline The main emphasis of this course is on the basic concepts of digital computing hardware and fundamental

More information

Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models

Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models Stephan Gouws and GJ van Rooyen MIH Medialab, Stellenbosch University SOUTH AFRICA {stephan,gvrooyen}@ml.sun.ac.za

More information

CPS122 Lecture: Identifying Responsibilities; CRC Cards. 1. To show how to use CRC cards to identify objects and find responsibilities

CPS122 Lecture: Identifying Responsibilities; CRC Cards. 1. To show how to use CRC cards to identify objects and find responsibilities Objectives: CPS122 Lecture: Identifying Responsibilities; CRC Cards last revised February 7, 2012 1. To show how to use CRC cards to identify objects and find responsibilities Materials: 1. ATM System

More information

2/15/13. POS Tagging Problem. Part-of-Speech Tagging. Example English Part-of-Speech Tagsets. More Details of the Problem. Typical Problem Cases

2/15/13. POS Tagging Problem. Part-of-Speech Tagging. Example English Part-of-Speech Tagsets. More Details of the Problem. Typical Problem Cases POS Tagging Problem Part-of-Speech Tagging L545 Spring 203 Given a sentence W Wn and a tagset of lexical categories, find the most likely tag T..Tn for each word in the sentence Example Secretariat/P is/vbz

More information

Software Security: Integrating Secure Software Engineering in Graduate Computer Science Curriculum

Software Security: Integrating Secure Software Engineering in Graduate Computer Science Curriculum Software Security: Integrating Secure Software Engineering in Graduate Computer Science Curriculum Stephen S. Yau, Fellow, IEEE, and Zhaoji Chen Arizona State University, Tempe, AZ 85287-8809 {yau, zhaoji.chen@asu.edu}

More information

Abstractions and the Brain

Abstractions and the Brain Abstractions and the Brain Brian D. Josephson Department of Physics, University of Cambridge Cavendish Lab. Madingley Road Cambridge, UK. CB3 OHE bdj10@cam.ac.uk http://www.tcm.phy.cam.ac.uk/~bdj10 ABSTRACT

More information

Experiments with SMS Translation and Stochastic Gradient Descent in Spanish Text Author Profiling

Experiments with SMS Translation and Stochastic Gradient Descent in Spanish Text Author Profiling Experiments with SMS Translation and Stochastic Gradient Descent in Spanish Text Author Profiling Notebook for PAN at CLEF 2013 Andrés Alfonso Caurcel Díaz 1 and José María Gómez Hidalgo 2 1 Universidad

More information

Evolutive Neural Net Fuzzy Filtering: Basic Description

Evolutive Neural Net Fuzzy Filtering: Basic Description Journal of Intelligent Learning Systems and Applications, 2010, 2: 12-18 doi:10.4236/jilsa.2010.21002 Published Online February 2010 (http://www.scirp.org/journal/jilsa) Evolutive Neural Net Fuzzy Filtering:

More information

Chapter 2. Intelligent Agents. Outline. Agents and environments. Rationality. PEAS (Performance measure, Environment, Actuators, Sensors)

Chapter 2. Intelligent Agents. Outline. Agents and environments. Rationality. PEAS (Performance measure, Environment, Actuators, Sensors) Intelligent Agents Chapter 2 1 Outline Agents and environments Rationality PEAS (Performance measure, Environment, Actuators, Sensors) Agent types 2 Agents and environments sensors environment percepts

More information

CS 1103 Computer Science I Honors. Fall Instructor Muller. Syllabus

CS 1103 Computer Science I Honors. Fall Instructor Muller. Syllabus CS 1103 Computer Science I Honors Fall 2016 Instructor Muller Syllabus Welcome to CS1103. This course is an introduction to the art and science of computer programming and to some of the fundamental concepts

More information

Task Completion Transfer Learning for Reward Inference

Task Completion Transfer Learning for Reward Inference Machine Learning for Interactive Systems: Papers from the AAAI-14 Workshop Task Completion Transfer Learning for Reward Inference Layla El Asri 1,2, Romain Laroche 1, Olivier Pietquin 3 1 Orange Labs,

More information

Financing Education In Minnesota

Financing Education In Minnesota Financing Education In Minnesota 2016-2017 Created with Tagul.com A Publication of the Minnesota House of Representatives Fiscal Analysis Department August 2016 Financing Education in Minnesota 2016-17

More information

Maximizing Learning Through Course Alignment and Experience with Different Types of Knowledge

Maximizing Learning Through Course Alignment and Experience with Different Types of Knowledge Innov High Educ (2009) 34:93 103 DOI 10.1007/s10755-009-9095-2 Maximizing Learning Through Course Alignment and Experience with Different Types of Knowledge Phyllis Blumberg Published online: 3 February

More information

Oregon Institute of Technology Computer Systems Engineering Technology Department Embedded Systems Engineering Technology Program Assessment

Oregon Institute of Technology Computer Systems Engineering Technology Department Embedded Systems Engineering Technology Program Assessment Oregon Institute of Technology Computer Systems Engineering Technology Department Embedded Systems Engineering Technology Program Assessment 2014-15 I. Introduction The Embedded Systems Engineering Technology

More information

Application of Virtual Instruments (VIs) for an enhanced learning environment

Application of Virtual Instruments (VIs) for an enhanced learning environment Application of Virtual Instruments (VIs) for an enhanced learning environment Philip Smyth, Dermot Brabazon, Eilish McLoughlin Schools of Mechanical and Physical Sciences Dublin City University Ireland

More information

An OO Framework for building Intelligence and Learning properties in Software Agents

An OO Framework for building Intelligence and Learning properties in Software Agents An OO Framework for building Intelligence and Learning properties in Software Agents José A. R. P. Sardinha, Ruy L. Milidiú, Carlos J. P. Lucena, Patrick Paranhos Abstract Software agents are defined as

More information

System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks

System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks 1 Tzu-Hsuan Yang, 2 Tzu-Hsuan Tseng, and 3 Chia-Ping Chen Department of Computer Science and Engineering

More information

Document number: 2013/ Programs Committee 6/2014 (July) Agenda Item 42.0 Bachelor of Engineering with Honours in Software Engineering

Document number: 2013/ Programs Committee 6/2014 (July) Agenda Item 42.0 Bachelor of Engineering with Honours in Software Engineering Document number: 2013/0006139 Programs Committee 6/2014 (July) Agenda Item 42.0 Bachelor of Engineering with Honours in Software Engineering Program Learning Outcomes Threshold Learning Outcomes for Engineering

More information

Decision Analysis. Decision-Making Problem. Decision Analysis. Part 1 Decision Analysis and Decision Tables. Decision Analysis, Part 1

Decision Analysis. Decision-Making Problem. Decision Analysis. Part 1 Decision Analysis and Decision Tables. Decision Analysis, Part 1 Decision Support: Decision Analysis Jožef Stefan International Postgraduate School, Ljubljana Programme: Information and Communication Technologies [ICT3] Course Web Page: http://kt.ijs.si/markobohanec/ds/ds.html

More information

Generating Test Cases From Use Cases

Generating Test Cases From Use Cases 1 of 13 1/10/2007 10:41 AM Generating Test Cases From Use Cases by Jim Heumann Requirements Management Evangelist Rational Software pdf (155 K) In many organizations, software testing accounts for 30 to

More information

COMPUTATIONAL COMPLEXITY OF LEFT-ASSOCIATIVE GRAMMAR

COMPUTATIONAL COMPLEXITY OF LEFT-ASSOCIATIVE GRAMMAR COMPUTATIONAL COMPLEXITY OF LEFT-ASSOCIATIVE GRAMMAR ROLAND HAUSSER Institut für Deutsche Philologie Ludwig-Maximilians Universität München München, West Germany 1. CHOICE OF A PRIMITIVE OPERATION The

More information

CHAPTER 4: REIMBURSEMENT STRATEGIES 24

CHAPTER 4: REIMBURSEMENT STRATEGIES 24 CHAPTER 4: REIMBURSEMENT STRATEGIES 24 INTRODUCTION Once state level policymakers have decided to implement and pay for CSR, one issue they face is simply how to calculate the reimbursements to districts

More information

Statewide Framework Document for:

Statewide Framework Document for: Statewide Framework Document for: 270301 Standards may be added to this document prior to submission, but may not be removed from the framework to meet state credit equivalency requirements. Performance

More information

ENEE 302h: Digital Electronics, Fall 2005 Prof. Bruce Jacob

ENEE 302h: Digital Electronics, Fall 2005 Prof. Bruce Jacob Course Syllabus ENEE 302h: Digital Electronics, Fall 2005 Prof. Bruce Jacob 1. Basic Information Time & Place Lecture: TuTh 2:00 3:15 pm, CSIC-3118 Discussion Section: Mon 12:00 12:50pm, EGR-1104 Professor

More information

Guru: A Computer Tutor that Models Expert Human Tutors

Guru: A Computer Tutor that Models Expert Human Tutors Guru: A Computer Tutor that Models Expert Human Tutors Andrew Olney 1, Sidney D'Mello 2, Natalie Person 3, Whitney Cade 1, Patrick Hays 1, Claire Williams 1, Blair Lehman 1, and Art Graesser 1 1 University

More information

CAAP. Content Analysis Report. Sample College. Institution Code: 9011 Institution Type: 4-Year Subgroup: none Test Date: Spring 2011

CAAP. Content Analysis Report. Sample College. Institution Code: 9011 Institution Type: 4-Year Subgroup: none Test Date: Spring 2011 CAAP Content Analysis Report Institution Code: 911 Institution Type: 4-Year Normative Group: 4-year Colleges Introduction This report provides information intended to help postsecondary institutions better

More information

Extending Place Value with Whole Numbers to 1,000,000

Extending Place Value with Whole Numbers to 1,000,000 Grade 4 Mathematics, Quarter 1, Unit 1.1 Extending Place Value with Whole Numbers to 1,000,000 Overview Number of Instructional Days: 10 (1 day = 45 minutes) Content to Be Learned Recognize that a digit

More information

Machine Learning and Data Mining. Ensembles of Learners. Prof. Alexander Ihler

Machine Learning and Data Mining. Ensembles of Learners. Prof. Alexander Ihler Machine Learning and Data Mining Ensembles of Learners Prof. Alexander Ihler Ensemble methods Why learn one classifier when you can learn many? Ensemble: combine many predictors (Weighted) combina

More information