Improved Multi-Agent Reinforcement Learning for Minimizing Traffic Waiting Time
|
|
- Dustin Murphy
- 5 years ago
- Views:
Transcription
1 Improved Muli-Agen Reinforcemen Learning for Minimizing Traffic Waiing Time Vijay Kumar M.T.U India B. Kaushik K.E.C., M.T.U., India H. Banka ISM, India ABSTRACT This paper depic using muli-agen reinforcemen learning (MAR algorihm for learning raffic paern o minimize he raveling ime or maximizing safey and opimizing raffic paern (OTP). This model provides a descripion and soluion o opimize raffic paern ha use muli-agen based reinforcemen learning algorihms. MARL uses muli agen srucure where vehicles and raffic signals are working as agens. In his model raffic area divide in differen-differen raffic ZONE. Each zone have own disribued agen and hese agen will pass he informaion one zone o oher hrew he nework. The Opimizaion objecives include he number of vehicle sops, he average waiing ime and maximum queue lengh of he nex (node) inersecion. In addiio This research also inroduce he prioriy conrol of buses and emergen vehicles ino his model. Expeced oucome of he algorihm is comparable o he performance of Q-Learning and Temporal difference learning. The resuls show significan reducion in waiing ime comparable o hose algorihms for he work more efficienly han oher raffic sysem. General Terms Learning Algorihm, Arificial Inelligence, Agen based learning. Keywords Agen Based Sysem, Inelligen Traffic Signal Conrol, Muli Objecive Scheme, Opimizaion Objecives, RL, Muli-Agen Sysem (MAS). 1. INTRODUCTION Manage he raffic in high raffic areas is a big problem. Increasing populaion size requires more efficien ransporaion sysems and hence beer raffic conrol sysem. Even developed counries are suffering high coss because of increasing road congesion levels. In he European Union (EU) alone, congesion coss 0.5% of he member counries Gross Domesic Produc (GDP) [11], [8], and his is expeced o increase o roughly 1% of he EU s GDP by 2009 if he problem is no deal wih properly. In 2002, he number of vehicles per housand persons had reached 460 which is nearly double he number (232) in 1974.In high raffic siuaions and bad driving in he EU (European Union) accouns for up o 50% of fuel consumpion on road neworks resuling in deadly emissions ha could oherwise be diminished. High raffic ranspor conribues 41% of carbon dioxide o give ou from road raffic in he EU hus resuling in serious healh and safey problems. In hese cases o avoid he high coss ha give by hese hreas, UTC has o provide some soluions o he problem of raffic managemen [11], [8]. To achieve he global goal UTC opimizaio increasing global such hreas and vehicles infrasrucure communicaing beween some sysems may be provide some exra deail.these deail may provide help for local view of he raffic condiions. In case medium raffic condiions he Wiering s mehod reduce he overall waiing ime for vehicles. This mehod reduce he waiing ime for vehicles and opimize he goal. In real raffic sysem, his model should consider differendifferen opimizaion objecives in differen raffic siuaio which is called muli-agen conrol scheme in his paper. In he free raffic siuaio presened model ry o minimize he overall number of sops of vehicles in he raffic nework. In case medium raffic siuaion his research ries o minimize he waiing ime on behalf opimal goal. In congesed raffic condiion main focused on queue lengh. So muli-agen conrol scheme can adap o differen raffic condiions and make a more inelligen raffic conrol sysem. Therefore, his model, propose a muli-agen conrol sraegy using MARL. Muli-objecive conrol and paramic simulaion model boh have some problems.firs node raffic siuaion pass o he all nex nodes. If firs node has a free raffic, his condiion will passes all he nex nodes, his is no good way for real raffic so his model will calculae raffic siuaion individually for each node. In congesed raffic siuaio queue spillovers mus be avoided o keep he nework from large-scale congesio hus he queue lengh mus be focused on [6]. In his model cycle is prevened. The value of is no fix (3) i depends on raffic conrol admin in his model.this may be 4, 5 ec. On behalf he value of his model will manage green ligh for emergen vehicles in raffic nework. In his model daa exchange beween vehicles and roadside raffic equipmens is necessary, hus vehicular ad hoc nework is uilized o build a wireless raffic informaion sysem. Therefore disribued nework helpful for uilized o develop a wireless raffic informaion sysem. Differen researchers have chosen varian ypes of arificial inelligence algorihms and mehods for he opimizaion of he raffic flow in real raffic condiions. Geneic algorihm or evoluionary algorihm is one of he mos common mehods inroduced ino he raffic conrol sysem. Rouing of raffic flow using geneic algorihm has shown some improvemen in he raffic conrol. Fuzzy logic conrol is also useful ino he raffic ligh sysems for beer conrol of raffic flow. Increase performance of real raffic ligh sysem is build wih some idea such ha increases green ligh ime period for vehicles. Anoher approach o improve he raffic conrol is using wireless nework communicaions beween vehicles and raffic conrol sysems o ge raffic informaion for raffic flow. This informaion can use for opimizaion in raffic sysem in medium and high raffic condiions. Reinforcemen learning echnique is used in cerain research sudies for he raffic flow conrol and 30
2 opimizaions. So reinforcemen learning echnique can be applied in raffic signal conrol effecively o response o he frequen change of raffic flow and ouperform radiional raffic conrol algorihm ha helpful for opimaliy, reducing raffic delay and build a beer raffic ligh sysem. This model are minimizing ravel ime or maximizing safey, Minimizing vehicle ravel ime, reducing raffic delay, increasing vehicle velociy, and prioriizing emergency raffic Since OTP conrollers by hand is a complex and edious ask,his research sudy how muli-agen reinforcemen learning(mar algorihms can be used for his goal. 2. AGENT BASED MODEL OF TRAFFIC SYSTEM In his model use an agen-based model o describe he pracical raffic sysem. In he roa here are wo ypes of agen one is vehicles and anoher is raffic signal conrollers called as disribued agens. Traffic informaion will be exchange beween hese agens. There are some possibiliy for each raffic conrollers ha preven raffic hreas and accidens. Two raffic lighs from opposing direcions allow vehicle o go sraigh ahead o urn righ, wo raffic lighs a he same direcion of he inersecion allow he vehicle from here o go sraigh ahea urn righ or urn Lef. When new vehicle have been added he raffic ligh decisions are made and each vehicle moves o cell if cell is no occupied.this decision conrol by he raffic sysem according o raffic condiions. There for, each vehicle is a a raffic a direcion a he node (dir), a posiion in he queue (place) and has a paricular desinaion (des). This model use [ place, in sor ([ o denoe he sae of each vehicle [7].The main objec is opimizaion wih reduce waiing ime,number of sops and raffic queue lengh. One name is Reinforcemen Learning ha suppor dynamic environmen using dynamic programming. A more popular approach is o use model-based reinforcemen learning, in which he ransiion and reward funcions are esimaed from experience and hen used o find a policy via planning mehods like dynamic programming. 3.1 Simple model Figure 2 shows he learning process of an agen. A each ime se he agen receives a reinforcemen feedback from he environmen along wih he curren sae. The goal for he agen is o creae an opimal acion selecion policy p o maximize he reward. In many cases, no only he immediae reward bu also he subsequen rewards Delayed rewards? should be considered when acions are aken. Fig 2: Agen wih sae and acion Agen and environmen inerac a discree ime seps: 0,1,2,k Agen observes sae a sep: Produces acion a sep: Ges resuling reward: : s a A r R 1 ( s ) S s And resuling nex sae: 1 Fig 1: Agen Based Model. In his model Q([ acion) o represen he oal expeced value of opimized indices for all raffic lighs for each vehicle. This process will be coninue unil vehicles arrive a he desinaion goal. In Wiering s model, consider firs node raffic siuaion pass o he all nex nodes. If firs node has a free raffic, his condiion passes all he nex nodes bu his model will calculae raffic siuaion individually for each node. This is he mos impor difference beween his model and Wiering s model. 3. REINFORCEMENT LEARNING FOR TRAFFIC CONTROL Previously several mehods for learn raffic have been developed like Sarsa and Q-learning.These all echniques suffered wih same problem in high raffic condiions. In urban or congesed raffic, hese echnique are no scale o muli-agen Reinforcemen Learning. In urban raffic may be possible ha raffic grows dynamically. So need a dynamic mehod for handle urban raffic ha grow dynamically. Q- learning and Sarsa hey are applied only o small nework. Fig 3: A general process model of RL [8] 3.2 This Basic Elemens of Reinforcemen Learning 1. Model of he process 2. Reward funcions. 3. Learning objecive. 4. Conrollers. 5. Exploraion. 3.3 Muli-agen Frame work The muli-agen framework is based on he same idea of Figure 2 bu, his Time, here are several agens deciding on acions over he environmen. The big difference resides in he fac ha all each agen probably has some effec on he environmen an so, acions can have differen oucomes depending on wha he oher agens are doing. Nex Fig. shows he muli-agen model or framework. 31
3 d [ C([ pos, Re d)/ C([ (2) Where C([ vehicle in he sae of C([ Re d) he ligh urns red in such sae. is he number of imes a [ is he number of imes 4.2 Medium raffic condiion In medium raffic condiion main goal of his model is o minimize he overall waiing ime of vehicles. If number of vehicles are larger 100 bu less han 150, i is consider as medium raffic., Fig 4: Muli-Agen Model In addiion o benefis owing o he disribued naure of he muli-agen soluio such as he speedup made possible by parallel compuaio muliple RL agens can harness new benefis from sharing experience, e.g., by communicaio eaching. Conversely, besides challenges inheried from single-agen RL, including he curse of dimensionaliy. 4. MULTI-AGENT CONTROL ALGORITHM BASED ON REINFORCEMENT LEARNING The muli-agen conrol algorihm considers hree ypes of raffic siuaions as follows less raffic (low raffic or free raffic) siuaio medium raffic siuaion and congesed raffic siuaion. 4.1 Free raffic condiion The number of sops will increase when a vehicle moving a a green ligh in curren ime sep mee a red ligh in he nex ime sep. In free raffic condiion he main goal is o minimize he number of sops. So use Q ([ Green) as he expeced cumulaive number of sops. The formulaion of Q ([ Green) is shown as follows. Q([ Green) ( dir', ) d [ ( R([ dir [ Q([ Green)) (1) [ Where means he sae of a vehicle in nex ime sep; d [ gives he probabiliy ha he raffic ligh urns red in nex ime sep; R([ dir, [ is a reward funcion as follows: if a vehicle says a he same raffic ligh, hen R=1, oherwise R=0, (he vehicle ges hrough his inersecion and eners he nex one); is he discoun facor (0 < < 1) which ensure he Q-values are bounded. The probabiliy ha a raffic ligh urns red is calculaed as follows. V ([ P( L [ L LQ ([ (3) Q([ L ( dir' pos) L [ n', ( R[ [ n', des ]) ( n', )) V (4) Where is L he raffic ligh sae (red or green), P ( L [, is calculaed in he same way as equaion 2, ( R [ [ n', is defined as follows as: if a vehicle says a he same raffic ligh, hen R=1, oherwise R=0 and use for force o be green ligh Congesed raffic condiion In his condiio spillovers of queue mus be avoided which will minimize he raffic conrol effec and probably cause large-scale raffic congesion. Q([ Green) ( dir' pos) Gree[ ( R([ [ R' ([,[ node ', V ( des'])) (5) Q([ Re d) ( dir', ) Re [ ' ( R ([ [ V ([ )) (6) Where Q ([ and V ([ have he same meanings as 32
4 under medium raffic condiion. Compared equaion 5 wih equaion 4, anoher reward funcion R '([ [ is added o indicae he influence from raffic condiion a he nex and use for force o be green ligh, 10 R ([ dir, [ Is he reward of vehicles waiing ime while R' ([ [ indicaes he reward from he change of he queue lengh a he nex raffic node. Consider queue lengh when design Q learning procedure, l ' denoe he max queue lengh a nex raffic ligh so l ' can wrien as K. L is he capaciy of he α is he adjusing facor ha lane of nex raffic ligh and deermine queue lengh K l' as follows: 0 IF K l' 0.8L 0 k ) 1.0 IF 0.8L Kl' (7) ( Tl0. 8 L.2 IF Kl L The larges value is se o.2 in his model. 4.4 Prioriy Conrol for Emergen Vehicles In case emergency vehicles like Fire Truck ambulances, Prime Miniser Vehicles ec. so need o manage raffic ligh when hese condiions were arise. For hese siuaions give high prioriy for hese ypes of vehicle. The raffic adminisraor can manage raffic ligh according o raffic condiions. If emergency condiion arises he admin of raffic conrol can reduce ime of he green ligh ha is se prioriy according o ype of vehicles for green ligh. In prioriy condiion he main focus manage green ligh on behalf his, presen model can reduce waiing ime for emergency vehicles. Q ([ pos, Green([ des', ])( R([ [ des' V[ pos ') (8) 5. RESULT In his research 1000 ime seps use for simulaion. For learning process 2000 seps use, and 2000 seps were also used for simulaion resul. The value 0.9 se o facor in his model. is se o be according o emergen Vehicles siuaion ha is for Fire Truck and ambulance he prioriy of green ligh may be differ, no 3(fix). If in a minue number of vehicles are 100 enering in raffic nework, i is consider as free raffic. If number of vehicles are larger 100 bu less han 150, i is consider as medium raffic, and number of vehicles are larger han 150 i is consider as congesed (high raffic) raffic condiion. 5.1 Comparison of average waiing ime Comparison of average waiing ime regard o he increasing of raffic volume rapidly is shown in figure 5.TD means emporal difference, QL means Q-learning algorihm, MARL means Muli-agen reinforcemen learning algorihm he model proposed in his paper. The nex able shows a daa se used in TD, QL, and MARL. Table 1 Visiing Poins wih Q-Capaciy and Q-Lengh visiing Poins q-capaciy q-lengh Lambeh Waford WesDrayon Leaherhead Oford Darford Loughon Aylesford Table 2 Visiers disance Visiors Lambeh Waford WesDrayon Leaherhead Oford Darford Loughon Aylesford Lambeh Waford WesDrayon Leaherhead Oford Darford Loughon Aylesford In Table 2 visiors disance,-1 show here is no any pah beween wo visior nodes. Number of sops under he muli-agen RL conrol will be less han hose under oher conrol sraegies like TD and Q- learning. Reinforcemen learning who minimize number of sops comparable o TD and Q-learning echnique in case medium raffic and congesed raffic condiions. 6. CONCLUSION This paper presened he muli-agen RL conrol algorihm based on reinforcemen learning. The simulaion indicaed ha he MARL go he minimum waiing ime under free raffic, comparable QL, TD. MARL could effecively preven he queue spillovers o avoid large scale raffic jams. There are sill some sysem parameers ha should carefully be deermined by hand. For, example, he adjusing facor α indicaing he influence of he queue a he nex raffic node o 33
5 he waiing ime of vehicles a curren ligh under congesed raffic condiion. This is a very imporan parameer, which we should furher research is deermining way based fuzzy logic approach such as crisp o fuzzy conversion such as Lambda cus for minimizing raffic paern. Neural nework as a ool can also be used for deecing rends in raffic paerns and o predic minimal waiing ime for raffic. Fig 5: Simulaion beween TD, QL and MARL by increasing he opposie raffic lengh. 7. ACKNOWLEDGMENTS Firs and foremos, I would like o express my sincere hanks o my paper advisor Associaive Prof. Baijnah Kaushik for providing me heir precious advices and suggesions. This model wouldn have been a success for me wihou heir cooperaion and valuable commens and suggesions. I also wan o express my graiude o Prof. P. S. Gill (H.O.D.) and Associaive Prof. Sunia Tiwari (M.Tech. Coordinaor) for heir suppor, kind hel coninued ineres and inspiraion during his work. 8. REFERENCES [1] Bowling,M.: Convergence and no-regre in muliagen learning. In: L.K.Saul, Y.Weiss, L. Boou (eds.) Advances in Neural Informaion Processing Sysems 17, pp MIT Press (2005). [2] Bus oniu, L., De Schuer, B., Babuˇska, R.: Muliagen reinforcemen learning wih adapive sae focus. In: Proceedings 17h Belgian-Duch Conference on Arificial Inelligence (BNAIC-05), pp Brussels, Belgium (2005). [3] Chalkiadakis, G.: Muliagen reinforcemen learning: Sochasic games wih muliple learning players. Tech. rep., Dep. of Compuer Science, Universiy of Torono, Canada (2003). [4] Guesri C., Lagoudakis, M.G., Parr, R.: Coordinaed reinforcemen learning. In: Proceedings 19h Inernaional Conference on Machine Learning (ICML- 02), pp Sydney, Ausralia (2002) [5] Hu, J., Wellma M.P.: Nash Q-learning for general-sum sochasic games. Journal of Machine Learning Research 4, (2003) [6] M.Wiering, e al (2004). Inelligen Traffic Ligh Conrol. Technical Repor UU-CS , Universiy Urech. [7] M.Wiering (2000). Muli-Agen Reinforcemen Learning for Traffic Ligh Conrol. Machine Learning: Proceedings of he 17h Inernaional Conference (ICML 2000), [8] Michell, T. M. (1995) he Book of Machine Learning: McGraw-HILL INTERNATIONAL EDITIONS. [9] Nunes L., and Oliveira, E. C. Learning from muliple sources. In Proceedings of he 3rd Inernaional Join Conference on Auonomous Agens and Muli Agen Sysems, AAMAS (New York, USA, July 2004), vol. 3, New York, IEEE Compuer Sociey, pp [10] Oliveira, D., Bazza A. L. C., and Lesser, V. using cooperaive mediaion o coordinae raffic lighs: a case sudy. In Proceedings of he 4h Inernaional Join Conference on Auonomous Agens and Muli Agen Sysems (AAMAS) (July 2005), New York, IEEE Compuer Sociey, pp [11] Price, B., Bouilier, C.: Acceleraing reinforcemen learning hrough implici imiaion Journal of Arificial Inelligence Research 19, (2003). [12] Ta M.: Muli-agen reinforcemen learning: Independen vs. cooperaive agens. In: Proceedings 10h Inernaional Conference on Machine Learning (ICML- 93), pp Amhers, US (1993). IJCA TM : 34
Neural Network Model of the Backpropagation Algorithm
Neural Nework Model of he Backpropagaion Algorihm Rudolf Jakša Deparmen of Cyberneics and Arificial Inelligence Technical Universiy of Košice Lená 9, 4 Košice Slovakia jaksa@neuron.uke.sk Miroslav Karák
More informationAn Effiecient Approach for Resource Auto-Scaling in Cloud Environments
Inernaional Journal of Elecrical and Compuer Engineering (IJECE) Vol. 6, No. 5, Ocober 2016, pp. 2415~2424 ISSN: 2088-8708, DOI: 10.11591/ijece.v6i5.10639 2415 An Effiecien Approach for Resource Auo-Scaling
More informationInformation Propagation for informing Special Population Subgroups about New Ground Transportation Services at Airports
Downloaded from ascelibrary.org by Basil Sephanis on 07/13/16. Copyrigh ASCE. For personal use only; all righs reserved. Informaion Propagaion for informing Special Populaion Subgroups abou New Ground
More informationFast Multi-task Learning for Query Spelling Correction
Fas Muli-ask Learning for Query Spelling Correcion Xu Sun Dep. of Saisical Science Cornell Universiy Ihaca, NY 14853 xusun@cornell.edu Anshumali Shrivasava Dep. of Compuer Science Cornell Universiy Ihaca,
More informationMyLab & Mastering Business
MyLab & Masering Business Efficacy Repor 2013 MyLab & Masering: Business Efficacy Repor 2013 Edied by Michelle D. Speckler 2013 Pearson MyAccouningLab, MyEconLab, MyFinanceLab, MyMarkeingLab, and MyOMLab
More informationChannel Mapping using Bidirectional Long Short-Term Memory for Dereverberation in Hands-Free Voice Controlled Devices
Z. Zhang e al.: Channel Mapping using Bidirecional Long Shor-Term Memory for Dereverberaion in Hands-Free Voice Conrolled Devices 525 Channel Mapping using Bidirecional Long Shor-Term Memory for Dereverberaion
More information1 Language universals
AS LX 500 Topics: Language Uniersals Fall 2010, Sepember 21 4a. Anisymmery 1 Language uniersals Subjec-erb agreemen and order Bach (1971) discusses wh-quesions across SO and SO languages, hypohesizing:...
More informationMore Accurate Question Answering on Freebase
More Accurae Quesion Answering on Freebase Hannah Bas, Elmar Haussmann Deparmen of Compuer Science Universiy of Freiburg 79110 Freiburg, Germany {bas, haussmann}@informaik.uni-freiburg.de ABSTRACT Real-world
More informationReinforcement Learning by Comparing Immediate Reward
Reinforcement Learning by Comparing Immediate Reward Punit Pandey DeepshikhaPandey Dr. Shishir Kumar Abstract This paper introduces an approach to Reinforcement Learning Algorithm by comparing their immediate
More informationAMULTIAGENT system [1] can be defined as a group of
156 IEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS PART C: APPLICATIONS AND REVIEWS, VOL. 38, NO. 2, MARCH 2008 A Comprehensive Survey of Multiagent Reinforcement Learning Lucian Buşoniu, Robert Babuška,
More informationISFA2008U_120 A SCHEDULING REINFORCEMENT LEARNING ALGORITHM
Proceedings of 28 ISFA 28 International Symposium on Flexible Automation Atlanta, GA, USA June 23-26, 28 ISFA28U_12 A SCHEDULING REINFORCEMENT LEARNING ALGORITHM Amit Gil, Helman Stern, Yael Edan, and
More informationTD(λ) and Q-Learning Based Ludo Players
TD(λ) and Q-Learning Based Ludo Players Majed Alhajry, Faisal Alvi, Member, IEEE and Moataz Ahmed Abstract Reinforcement learning is a popular machine learning technique whose inherent self-learning ability
More informationThe Use of Statistical, Computational and Modelling Tools in Higher Learning Institutions: A Case Study of the University of Dodoma
International Journal of Computer Applications (975 8887) The Use of Statistical, Computational and Modelling Tools in Higher Learning Institutions: A Case Study of the University of Dodoma Gilbert M.
More informationLaboratorio di Intelligenza Artificiale e Robotica
Laboratorio di Intelligenza Artificiale e Robotica A.A. 2008-2009 Outline 2 Machine Learning Unsupervised Learning Supervised Learning Reinforcement Learning Genetic Algorithms Genetics-Based Machine Learning
More informationArtificial Neural Networks written examination
1 (8) Institutionen för informationsteknologi Olle Gällmo Universitetsadjunkt Adress: Lägerhyddsvägen 2 Box 337 751 05 Uppsala Artificial Neural Networks written examination Monday, May 15, 2006 9 00-14
More informationOn the Combined Behavior of Autonomous Resource Management Agents
On the Combined Behavior of Autonomous Resource Management Agents Siri Fagernes 1 and Alva L. Couch 2 1 Faculty of Engineering Oslo University College Oslo, Norway siri.fagernes@iu.hio.no 2 Computer Science
More informationReduce the Failure Rate of the Screwing Process with Six Sigma Approach
Proceedings of the 2014 International Conference on Industrial Engineering and Operations Management Bali, Indonesia, January 7 9, 2014 Reduce the Failure Rate of the Screwing Process with Six Sigma Approach
More informationAxiom 2013 Team Description Paper
Axiom 2013 Team Description Paper Mohammad Ghazanfari, S Omid Shirkhorshidi, Farbod Samsamipour, Hossein Rahmatizadeh Zagheli, Mohammad Mahdavi, Payam Mohajeri, S Abbas Alamolhoda Robotics Scientific Association
More informationLecture 10: Reinforcement Learning
Lecture 1: Reinforcement Learning Cognitive Systems II - Machine Learning SS 25 Part III: Learning Programs and Strategies Q Learning, Dynamic Programming Lecture 1: Reinforcement Learning p. Motivation
More informationModule 12. Machine Learning. Version 2 CSE IIT, Kharagpur
Module 12 Machine Learning 12.1 Instructional Objective The students should understand the concept of learning systems Students should learn about different aspects of a learning system Students should
More informationRule-based Expert Systems
Rule-based Expert Systems What is knowledge? is a theoretical or practical understanding of a subject or a domain. is also the sim of what is currently known, and apparently knowledge is power. Those who
More informationACTIVITY: Comparing Combination Locks
5.4 Compound Events outcomes of one or more events? ow can you find the number of possible ACIVIY: Comparing Combination Locks Work with a partner. You are buying a combination lock. You have three choices.
More informationLaboratorio di Intelligenza Artificiale e Robotica
Laboratorio di Intelligenza Artificiale e Robotica A.A. 2008-2009 Outline 2 Machine Learning Unsupervised Learning Supervised Learning Reinforcement Learning Genetic Algorithms Genetics-Based Machine Learning
More informationLearning Methods for Fuzzy Systems
Learning Methods for Fuzzy Systems Rudolf Kruse and Andreas Nürnberger Department of Computer Science, University of Magdeburg Universitätsplatz, D-396 Magdeburg, Germany Phone : +49.39.67.876, Fax : +49.39.67.8
More informationPublic Speaking Rubric
Public Speaking Rubric Speaker s Name or ID: Coder ID: Competency: Uses verbal and nonverbal communication for clear expression of ideas 1. Provides clear central ideas NOTES: 2. Uses organizational patterns
More informationMEE 6501, Advanced Air Quality Control Course Syllabus. Course Description. Course Textbook. Course Learning Outcomes. Credits.
MEE 6501, Advanced Air Quality Control Course Syllabus Course Description An in-depth study of advanced air quality control science and management practices. Addresses health effects, environmental impacts,
More informationA comparative study on cost-sharing in higher education Using the case study approach to contribute to evidence-based policy
A comparative study on cost-sharing in higher education Using the case study approach to contribute to evidence-based policy Tuition fees between sacred cow and cash cow Conference of Vlaams Verbond van
More informationMath 1313 Section 2.1 Example 2: Given the following Linear Program, Determine the vertices of the feasible set. Subject to:
Math 1313 Section 2.1 Example 2: Given the following Linear Program, Determine the vertices of the feasible set Subject to: Min D 3 = 3x + y 10x + 2y 84 8x + 4y 120 x, y 0 3 Math 1313 Section 2.1 Popper
More informationLecture 1: Machine Learning Basics
1/69 Lecture 1: Machine Learning Basics Ali Harakeh University of Waterloo WAVE Lab ali.harakeh@uwaterloo.ca May 1, 2017 2/69 Overview 1 Learning Algorithms 2 Capacity, Overfitting, and Underfitting 3
More informationImproving Fairness in Memory Scheduling
Improving Fairness in Memory Scheduling Using a Team of Learning Automata Aditya Kajwe and Madhu Mutyam Department of Computer Science & Engineering, Indian Institute of Tehcnology - Madras June 14, 2014
More informationAn Introduction to Simio for Beginners
An Introduction to Simio for Beginners C. Dennis Pegden, Ph.D. This white paper is intended to introduce Simio to a user new to simulation. It is intended for the manufacturing engineer, hospital quality
More informationACCOUNTING FOR MANAGERS BU-5190-OL Syllabus
MASTER IN BUSINESS ADMINISTRATION ACCOUNTING FOR MANAGERS BU-5190-OL Syllabus Fall 2011 P LYMOUTH S TATE U NIVERSITY, C OLLEGE OF B USINESS A DMINISTRATION 1 Page 2 PLYMOUTH STATE UNIVERSITY College of
More informationReinForest: Multi-Domain Dialogue Management Using Hierarchical Policies and Knowledge Ontology
ReinForest: Multi-Domain Dialogue Management Using Hierarchical Policies and Knowledge Ontology Tiancheng Zhao CMU-LTI-16-006 Language Technologies Institute School of Computer Science Carnegie Mellon
More informationImpact of Educational Reforms to International Cooperation CASE: Finland
Impact of Educational Reforms to International Cooperation CASE: Finland February 11, 2016 10 th Seminar on Cooperation between Russian and Finnish Institutions of Higher Education Tiina Vihma-Purovaara
More informationWelcome to. ECML/PKDD 2004 Community meeting
Welcome to ECML/PKDD 2004 Community meeting A brief report from the program chairs Jean-Francois Boulicaut, INSA-Lyon, France Floriana Esposito, University of Bari, Italy Fosca Giannotti, ISTI-CNR, Pisa,
More informationA Reinforcement Learning Variant for Control Scheduling
A Reinforcement Learning Variant for Control Scheduling Aloke Guha Honeywell Sensor and System Development Center 3660 Technology Drive Minneapolis MN 55417 Abstract We present an algorithm based on reinforcement
More informationPhilosophy 301L: Early Modern Philosophy, Spring 2012
Philosophy 301L: Early Modern Philosophy, Spring 2012 Topic: Epistemology and metaphysics in the work of five major figures in the History of Modern Philosophy: Descartes, Locke, Berkeley, Leibniz, and
More informationLab 1 - The Scientific Method
Lab 1 - The Scientific Method As Biologists we are interested in learning more about life. Through observations of the living world we often develop questions about various phenomena occurring around us.
More informationVisual CP Representation of Knowledge
Visual CP Representation of Knowledge Heather D. Pfeiffer and Roger T. Hartley Department of Computer Science New Mexico State University Las Cruces, NM 88003-8001, USA email: hdp@cs.nmsu.edu and rth@cs.nmsu.edu
More informationAn OO Framework for building Intelligence and Learning properties in Software Agents
An OO Framework for building Intelligence and Learning properties in Software Agents José A. R. P. Sardinha, Ruy L. Milidiú, Carlos J. P. Lucena, Patrick Paranhos Abstract Software agents are defined as
More informationExclusions Policy. Policy reviewed: May 2016 Policy review date: May OAT Model Policy
Exclusions Policy Policy reviewed: May 2016 Policy review date: May 2018 OAT Model Policy 1 Contents Action to be invoked by Senior Staff in Serious Disciplinary Matters 1. When a serious incident occurs,
More informationSeminar - Organic Computing
Seminar - Organic Computing Self-Organisation of OC-Systems Markus Franke 25.01.2006 Typeset by FoilTEX Timetable 1. Overview 2. Characteristics of SO-Systems 3. Concern with Nature 4. Design-Concepts
More informationData Fusion Models in WSNs: Comparison and Analysis
Proceedings of 2014 Zone 1 Conference of the American Society for Engineering Education (ASEE Zone 1) Data Fusion s in WSNs: Comparison and Analysis Marwah M Almasri, and Khaled M Elleithy, Senior Member,
More informationINPE São José dos Campos
INPE-5479 PRE/1778 MONLINEAR ASPECTS OF DATA INTEGRATION FOR LAND COVER CLASSIFICATION IN A NEDRAL NETWORK ENVIRONNENT Maria Suelena S. Barros Valter Rodrigues INPE São José dos Campos 1993 SECRETARIA
More informationIntelligent Agents. Chapter 2. Chapter 2 1
Intelligent Agents Chapter 2 Chapter 2 1 Outline Agents and environments Rationality PEAS (Performance measure, Environment, Actuators, Sensors) Environment types The structure of agents Chapter 2 2 Agents
More informationACCOUNTING FOR MANAGERS BU-5190-AU7 Syllabus
HEALTH CARE ADMINISTRATION MBA ACCOUNTING FOR MANAGERS BU-5190-AU7 Syllabus Winter 2010 P LYMOUTH S TATE U NIVERSITY, C OLLEGE OF B USINESS A DMINISTRATION 1 Page 2 PLYMOUTH STATE UNIVERSITY College of
More informationEvolutive Neural Net Fuzzy Filtering: Basic Description
Journal of Intelligent Learning Systems and Applications, 2010, 2: 12-18 doi:10.4236/jilsa.2010.21002 Published Online February 2010 (http://www.scirp.org/journal/jilsa) Evolutive Neural Net Fuzzy Filtering:
More informationA student diagnosing and evaluation system for laboratory-based academic exercises
A student diagnosing and evaluation system for laboratory-based academic exercises Maria Samarakou, Emmanouil Fylladitakis and Pantelis Prentakis Technological Educational Institute (T.E.I.) of Athens
More informationSimCity 4 Deluxe Tutorial. Future City Competition
SimCity 4 Deluxe Tutorial Tutorial Outline 1. Getting Started 2. Using SimCity 4 Deluxe Tutorial 3. Building Regions 4. Develop Your Strategy 5. Create Your City Understanding the Toolbars 6. Mayor Mode
More informationFirms and Markets Saturdays Summer I 2014
PRELIMINARY DRAFT VERSION. SUBJECT TO CHANGE. Firms and Markets Saturdays Summer I 2014 Professor Thomas Pugel Office: Room 11-53 KMC E-mail: tpugel@stern.nyu.edu Tel: 212-998-0918 Fax: 212-995-4212 This
More informationAgent-Based Software Engineering
Agent-Based Software Engineering Learning Guide Information for Students 1. Description Grade Module Máster Universitario en Ingeniería de Software - European Master on Software Engineering Advanced Software
More informationTeam Work in International Programs: Why is it so difficult?
Team Work in International Programs: Why is it so difficult? & Henning Madsen Aarhus University Denmark SoTL COMMONS CONFERENCE Karen M. Savannah, Lauridsen GA Centre for Teaching and March Learning 2013
More informationLITERACY ACROSS THE CURRICULUM POLICY
"Pupils should be taught in all subjects to express themselves correctly and appropriately and to read accurately and with understanding." QCA Use of Language across the Curriculum "Thomas Estley Community
More informationInformation Event Master Thesis
Information Event Master Thesis Dr. Michael J. Kendzia Deputy Program Director MSc IB Building Competence. Crossing Borders. Overview Introduction Prior to the master thesis assignment procedure During
More informationPH.D. IN COMPUTER SCIENCE PROGRAM (POST M.S.)
PH.D. IN COMPUTER SCIENCE PROGRAM (POST M.S.) OVERVIEW ADMISSION REQUIREMENTS PROGRAM REQUIREMENTS OVERVIEW FOR THE PH.D. IN COMPUTER SCIENCE Overview The doctoral program is designed for those students
More informationInternational Environmental Policy Spring :374:315:01 Tuesdays, 10:55 am to 1:55 pm, Blake 131
International Environmental Policy Spring 2012-11:374:315:01 Tuesdays, 10:55 am to 1:55 pm, Blake 131 Instructor: Dr. Pamela McElwee Assistant Professor, Department of Human Ecology Cook Office Building,
More informationCOMPUTATIONAL COMPLEXITY OF LEFT-ASSOCIATIVE GRAMMAR
COMPUTATIONAL COMPLEXITY OF LEFT-ASSOCIATIVE GRAMMAR ROLAND HAUSSER Institut für Deutsche Philologie Ludwig-Maximilians Universität München München, West Germany 1. CHOICE OF A PRIMITIVE OPERATION The
More informationMultidisciplinary Engineering Systems 2 nd and 3rd Year College-Wide Courses
Multidisciplinary Engineering Systems 2 nd and 3rd Year College-Wide Courses Kevin Craig College of Engineering Marquette University Milwaukee, WI, USA Mark Nagurka College of Engineering Marquette University
More informationSoft Computing based Learning for Cognitive Radio
Int. J. on Recent Trends in Engineering and Technology, Vol. 10, No. 1, Jan 2014 Soft Computing based Learning for Cognitive Radio Ms.Mithra Venkatesan 1, Dr.A.V.Kulkarni 2 1 Research Scholar, JSPM s RSCOE,Pune,India
More informationMassachusetts Institute of Technology Tel: Massachusetts Avenue Room 32-D558 MA 02139
Hariharan Narayanan Massachusetts Institute of Technology Tel: 773.428.3115 LIDS har@mit.edu 77 Massachusetts Avenue http://www.mit.edu/~har Room 32-D558 MA 02139 EMPLOYMENT Massachusetts Institute of
More informationPh.D in Advance Machine Learning (computer science) PhD submitted, degree to be awarded on convocation, sept B.Tech in Computer science and
Name Qualification Sonia Thomas Ph.D in Advance Machine Learning (computer science) PhD submitted, degree to be awarded on convocation, sept. 2016. M.Tech in Computer science and Engineering. B.Tech in
More informationAssessing System Agreement and Instance Difficulty in the Lexical Sample Tasks of SENSEVAL-2
Assessing System Agreement and Instance Difficulty in the Lexical Sample Tasks of SENSEVAL-2 Ted Pedersen Department of Computer Science University of Minnesota Duluth, MN, 55812 USA tpederse@d.umn.edu
More informationDinesh K. Sharma, Ph.D. Department of Management School of Business and Economics Fayetteville State University
Department of Management School of Business and Economics Fayetteville State University EDUCATION Doctor of Philosophy, Devi Ahilya University, Indore, India (2013) Area of Specialization: Management:
More informationOn-Line Data Analytics
International Journal of Computer Applications in Engineering Sciences [VOL I, ISSUE III, SEPTEMBER 2011] [ISSN: 2231-4946] On-Line Data Analytics Yugandhar Vemulapalli #, Devarapalli Raghu *, Raja Jacob
More informationKeene State College SPECIAL PERMISSION FORM PRACTICUM, INTERNSHIP, EXTERNSHIP, FIELDWORK
Keene State College SPECIAL PERMISSION FORM PRACTICUM, INTERNSHIP, EXTERNSHIP, FIELDWORK DEPARTMENT NUMBER (Official use only) CREDITS COURSE TITLE: STUDENT NAME: (print) TERM: ID#: COURSE OUTLINE: Description
More informationActive Learning. Yingyu Liang Computer Sciences 760 Fall
Active Learning Yingyu Liang Computer Sciences 760 Fall 2017 http://pages.cs.wisc.edu/~yliang/cs760/ Some of the slides in these lectures have been adapted/borrowed from materials developed by Mark Craven,
More informationOPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS
OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS Václav Kocian, Eva Volná, Michal Janošek, Martin Kotyrba University of Ostrava Department of Informatics and Computers Dvořákova 7,
More informationTENNESSEE S ECONOMY: Implications for Economic Development
TENNESSEE S ECONOMY: Implications for Economic Development William F. Fox, Director Center for Business and Economic Research The University of Tennessee, Knoxville August 2005 U.S. ECONOMY W.F. Fox, CBER,
More informationEntrepreneurial Discovery and the Demmert/Klein Experiment: Additional Evidence from Germany
Entrepreneurial Discovery and the Demmert/Klein Experiment: Additional Evidence from Germany Jana Kitzmann and Dirk Schiereck, Endowed Chair for Banking and Finance, EUROPEAN BUSINESS SCHOOL, International
More informationGREAT Britain: Film Brief
GREAT Britain: Film Brief Prepared by Rachel Newton, British Council, 26th April 2012. Overview and aims As part of the UK government s GREAT campaign, Education UK has received funding to promote the
More informationProduct Feature-based Ratings foropinionsummarization of E-Commerce Feedback Comments
Product Feature-based Ratings foropinionsummarization of E-Commerce Feedback Comments Vijayshri Ramkrishna Ingale PG Student, Department of Computer Engineering JSPM s Imperial College of Engineering &
More informationSAM - Sensors, Actuators and Microcontrollers in Mobile Robots
Coordinating unit: Teaching unit: Academic year: Degree: ECTS credits: 2017 230 - ETSETB - Barcelona School of Telecommunications Engineering 710 - EEL - Department of Electronic Engineering BACHELOR'S
More informationA Context-Driven Use Case Creation Process for Specifying Automotive Driver Assistance Systems
A Context-Driven Use Case Creation Process for Specifying Automotive Driver Assistance Systems Hannes Omasreiter, Eduard Metzker DaimlerChrysler AG Research Information and Communication Postfach 23 60
More informationA Game-based Assessment of Children s Choices to Seek Feedback and to Revise
A Game-based Assessment of Children s Choices to Seek Feedback and to Revise Maria Cutumisu, Kristen P. Blair, Daniel L. Schwartz, Doris B. Chin Stanford Graduate School of Education Please address all
More informationExploration. CS : Deep Reinforcement Learning Sergey Levine
Exploration CS 294-112: Deep Reinforcement Learning Sergey Levine Class Notes 1. Homework 4 due on Wednesday 2. Project proposal feedback sent Today s Lecture 1. What is exploration? Why is it a problem?
More informationInteraction Design Considerations for an Aircraft Carrier Deck Agent-based Simulation
Interaction Design Considerations for an Aircraft Carrier Deck Agent-based Simulation Miles Aubert (919) 619-5078 Miles.Aubert@duke. edu Weston Ross (505) 385-5867 Weston.Ross@duke. edu Steven Mazzari
More informationFF+FPG: Guiding a Policy-Gradient Planner
FF+FPG: Guiding a Policy-Gradient Planner Olivier Buffet LAAS-CNRS University of Toulouse Toulouse, France firstname.lastname@laas.fr Douglas Aberdeen National ICT australia & The Australian National University
More informationGCSE. Mathematics A. Mark Scheme for January General Certificate of Secondary Education Unit A503/01: Mathematics C (Foundation Tier)
GCSE Mathematics A General Certificate of Secondary Education Unit A503/0: Mathematics C (Foundation Tier) Mark Scheme for January 203 Oxford Cambridge and RSA Examinations OCR (Oxford Cambridge and RSA)
More informationDegreeWorks Training Guide
DegreeWorks Training Guide A Degree Evaluation and Advising Tool for MERCY COLLEGE Information for Students Last updated 03/2014 What Is DegreeWorks? DegreeWorks is a web-based tool that will provide a
More informationPREDICTING SPEECH RECOGNITION CONFIDENCE USING DEEP LEARNING WITH WORD IDENTITY AND SCORE FEATURES
PREDICTING SPEECH RECOGNITION CONFIDENCE USING DEEP LEARNING WITH WORD IDENTITY AND SCORE FEATURES Po-Sen Huang, Kshitiz Kumar, Chaojun Liu, Yifan Gong, Li Deng Department of Electrical and Computer Engineering,
More informationCircuit Simulators: A Revolutionary E-Learning Platform
Circuit Simulators: A Revolutionary E-Learning Platform Mahi Itagi Padre Conceicao College of Engineering, Verna, Goa, India. itagimahi@gmail.com Akhil Deshpande Gogte Institute of Technology, Udyambag,
More informationLecture 6: Applications
Lecture 6: Applications Michael L. Littman Rutgers University Department of Computer Science Rutgers Laboratory for Real-Life Reinforcement Learning What is RL? Branch of machine learning concerned with
More informationSIE: Speech Enabled Interface for E-Learning
SIE: Speech Enabled Interface for E-Learning Shikha M.Tech Student Lovely Professional University, Phagwara, Punjab INDIA ABSTRACT In today s world, e-learning is very important and popular. E- learning
More informationSelf Study Report Computer Science
Computer Science undergraduate students have access to undergraduate teaching, and general computing facilities in three buildings. Two large classrooms are housed in the Davis Centre, which hold about
More informationFragment Analysis and Test Case Generation using F- Measure for Adaptive Random Testing and Partitioned Block based Adaptive Random Testing
Fragment Analysis and Test Case Generation using F- Measure for Adaptive Random Testing and Partitioned Block based Adaptive Random Testing D. Indhumathi Research Scholar Department of Information Technology
More informationEvolution of Symbolisation in Chimpanzees and Neural Nets
Evolution of Symbolisation in Chimpanzees and Neural Nets Angelo Cangelosi Centre for Neural and Adaptive Systems University of Plymouth (UK) a.cangelosi@plymouth.ac.uk Introduction Animal communication
More informationkey findings Highlights of Results from TIMSS THIRD INTERNATIONAL MATHEMATICS AND SCIENCE STUDY November 1996
TIMSS International Study Center BOSTON COLLEGE Highlights of Results from TIMSS THIRD INTERNATIONAL MATHEMATICS AND SCIENCE STUDY Now Available International comparative results in mathematics and science
More informationKaipaki School. We expect the roll to climb to almost 100 in line with the demographic report from MoE through 2016.
Kaipaki School 687 Kaipaki Rd RD3 Cambridge Kaipaki School Bringing Learning to Life Whakatinanahia te mātauranga Ph: (07) 823 6653 e-mail: principal@kaipaki.school.nz www.kaipaki.school.nz 25 May 2015
More informationAutomatic Discretization of Actions and States in Monte-Carlo Tree Search
Automatic Discretization of Actions and States in Monte-Carlo Tree Search Guy Van den Broeck 1 and Kurt Driessens 2 1 Katholieke Universiteit Leuven, Department of Computer Science, Leuven, Belgium guy.vandenbroeck@cs.kuleuven.be
More informationA Case-Based Approach To Imitation Learning in Robotic Agents
A Case-Based Approach To Imitation Learning in Robotic Agents Tesca Fitzgerald, Ashok Goel School of Interactive Computing Georgia Institute of Technology, Atlanta, GA 30332, USA {tesca.fitzgerald,goel}@cc.gatech.edu
More informationIAT 888: Metacreation Machines endowed with creative behavior. Philippe Pasquier Office 565 (floor 14)
IAT 888: Metacreation Machines endowed with creative behavior Philippe Pasquier Office 565 (floor 14) pasquier@sfu.ca Outline of today's lecture A little bit about me A little bit about you What will that
More informationPair Programming. Spring 2015
CS4 Introduction to Scientific Computing Potter Pair Programming Spring 2015 1 What is Pair Programming? Simply put, pair programming is two people working together at a single computer [1]. The practice
More informationBenjamin Pohl, Yves Richard, Manon Kohler, Justin Emery, Thierry Castel, Benjamin De Lapparent, Denis Thévenin, Thomas Thévenin, Julien Pergaud
Measured and simulated Urban Heat Island in Dijon, France [the Urban Heat Island of a middle-size Franch city as seen by high-resolution numerical experiments and in situ measurements the case of Dijon,
More informationRover Races Grades: 3-5 Prep Time: ~45 Minutes Lesson Time: ~105 minutes
Rover Races Grades: 3-5 Prep Time: ~45 Minutes Lesson Time: ~105 minutes WHAT STUDENTS DO: Establishing Communication Procedures Following Curiosity on Mars often means roving to places with interesting
More informationCONCEPT MAPS AS A DEVICE FOR LEARNING DATABASE CONCEPTS
CONCEPT MAPS AS A DEVICE FOR LEARNING DATABASE CONCEPTS Pirjo Moen Department of Computer Science P.O. Box 68 FI-00014 University of Helsinki pirjo.moen@cs.helsinki.fi http://www.cs.helsinki.fi/pirjo.moen
More informationOn JEE. Milind Sohoni Senate Meeting, IITB 6 th October 2016
On JEE Milind Sohoni Senate Meeting, IITB 6 th October 2016 Terms of Reference A. To recommend structure of a single exam that tests the understanding, conceptual clarity, and innovative thinking of students
More informationModeling function word errors in DNN-HMM based LVCSR systems
Modeling function word errors in DNN-HMM based LVCSR systems Melvin Jose Johnson Premkumar, Ankur Bapna and Sree Avinash Parchuri Department of Computer Science Department of Electrical Engineering Stanford
More informationCLASSIFICATION OF TEXT DOCUMENTS USING INTEGER REPRESENTATION AND REGRESSION: AN INTEGRATED APPROACH
ISSN: 0976-3104 Danti and Bhushan. ARTICLE OPEN ACCESS CLASSIFICATION OF TEXT DOCUMENTS USING INTEGER REPRESENTATION AND REGRESSION: AN INTEGRATED APPROACH Ajit Danti 1 and SN Bharath Bhushan 2* 1 Department
More informationThe development and implementation of a coaching model for project-based learning
The development and implementation of a coaching model for project-based learning W. Van der Hoeven 1 Educational Research Assistant KU Leuven, Faculty of Bioscience Engineering Heverlee, Belgium E-mail:
More informationCommon Core Exemplar for English Language Arts and Social Studies: GRADE 1
The Common Core State Standards and the Social Studies: Preparing Young Students for College, Career, and Citizenship Common Core Exemplar for English Language Arts and Social Studies: Why We Need Rules
More information