Introduction of connectionist models

Size: px
Start display at page:

Download "Introduction of connectionist models"

Transcription

1 Introduction of connectionist models Introduction to ANNs Markus Dambek Uni Bremen 20. Dezember 2010 Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66

2 1 Introduction 2 Information processing in biology 3 The artificial neuron 4 The perceptron 5 The multilayer perceptron 6 Classification 7 Development Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66

3 About what? Connectionism is a set of approaches in the fields of artificial intelligence, cognitive psychology, cognitive science, neuroscience and philosophy of mind, that models mental or behavioral phenomena as the emergent processes of interconnected networks of simple units. There are many forms of connectionism, but the most common forms use artificial neural networks (ANNs) Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66

4 ANNs Artificial neural networks (ANNs) are relatively new computational tools that have found extensive utilization in solving many complex real-world problems. The attractiveness of ANNs comes from their remarkable information processing characteristics pertinent mainly to nonlinearity Although ANNs are drastic abstractions of the biological counterparts, the idea of ANNs is not to replicate the operation of the biological systems. Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66

5 Capabilities nonlinearity noise-insensitivity Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66

6 Capabilities high parallelism learning generalization Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66

7 Information processing in biology Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66

8 The Neuron Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66

9 An impulse, in the form of an electric signal,travels within the dendrites and through the cell body towards the pre-synaptic membrane of the synapse. Upon arrival at the membrane, a neurotransmitter (chemical) is released from the vesicles in quantities proportional to the strength of the incoming signal. The neurotransmitter diffuses within the synaptic gap and eventually into the dendrites of neighboring neurons. Depending on the neighboring neuron s threshold, they force them to generate a new electrical signal. Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66

10 Neural Networks Such neurons are interconnected in networks of billions of neurons, processing incoming information into motor actions and new information. Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66

11 The artificial neuron Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66

12 The artificial neuron Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66

13 Analogy connections = dendrites and axons connection weights = synapses threshold = soma activity Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66

14 Components Input x i w ij net-input-function net j = k x k w kj activation-function a j = f act (net j,θ j ) output-function x j = f out (a j ) (mostly : f out = id) Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66

15 Activation Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66

16 The perceptron Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66

17 Perceptron Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66

18 Perceptron unidirectional two Layers: Input-Layer and Output-Layer Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66

19 Learning Delta-rule Change weights according to the contribution to the error Error denotes the difference between the perceptron s output and the expected output (teaching output) Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66

20 Learning Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66

21 AND-Net Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66

22 AND-Net x 1 x 2 x 1 *1+x 2 *1 > θ = 1,5 output false false false true 1 Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66

23 OR-Net Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66

24 OR-Net x 1 x 2 x 1 *1+x 2 *1 > θ = 0,5 output false true true true 1 Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66

25 XOR-Net Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66

26 Capabilities Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66

27 Capabilities Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66

28 The multilayer perceptron Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66

29 Multilayer-Perceptron Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66

30 Multilayer-Perceptron Adding Hidden-Layers results in the capability to represent even more komplex functions But delta-rule operates on the difference between the perceptron s output and the teaching output We have no clue about the hidden neuron s taching output We need to modify the learning algorithm Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66

31 Many different approaches: Backpropagation Hopfield Adaptive resonance theory... Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66

32 Learning: supervised vs. unsupervised vs. reinforcement Most of them can be divided into three different classes of learning algorithms: supervised learing unsupervised learning reinforcement learning Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66

33 error-correction learning The error-correction learning (ECL) rule is used in supervised learning. The arithmetic difference (error) between the ANN solution and the corresponding correct answer is used to modify the connection weights. Gradual reduction the overall network error. Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66

34 boltzman learning The Boltzmann learning (BL) rule is a stochastic rule. It is similar to ECL, however each neuron generates an output (or state) based on a Boltzmann statistical distribution Renders learning extremely slow. Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66

35 hebbian learning The Hebbian learning (HL) rule, developed based on neurobiological experiments, is the oldest learning rule. It postulates that if neurons on both sides of a synapse are activated synchronously and repeatedly, the synapses strength is selectively increased. Learning is done locally by adjusting the synapse weight based on the activities of the neurons. Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66

36 competitive learning In the competitive learning (CL) rule, all neurons are forced to compete among themselves. Only one neuron will be activated in a given iteration with all the weights attached to it being adjusted. The CL rule is speculated to exist in many biological systems Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66

37 Backpropagation Generalization of delta-rule error-correction learning Idea Represent the net-error as Function of all weights: E(W j ) = E(w j1,w j2,...,w jn ) two weights w 1 and w 2 Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66

38 Backpropagation Objective Find the global minimum of the Error-Function Use Gradient descent Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66

39 Backpropagation Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66

40 BP-XOR-Net Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66

41 BP-XOR-Net x 1 x 2 x 1*1+x 2*1 > θ = 1,5 x 3 x 1*1+x 2*1+x 3*(-2) > θ = 0,5 output false 0 0 false false 0 1 true false 0 1 true true 1 0 false 0 Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66

42 Classification Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66

43 Hopfield networks This network is a symmetric fully connected twolayer recurrent network. When presented with an incomplete or noisy pattern, the network responds by retrieving an internally stored pattern that most closely resembles the presented pattern. Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66

44 Hopfield networks Efficient in solving optimization problems. Learning is done by setting each weight connecting two neurons to the product of the inputs of these two neurons. Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66

45 Adaptive resonance theory These are trained by unsupervised learning. The ART network consists of two fully interconnected layers, an input-layer and an output-layer. The feedforward weights are used to select a winning output neuron (cluster) and serve as the long-term memory for the networks. Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66

46 Adaptive resonance theory The feedback weights serve as the short-term memory for the network. Can be used for pattern recognition, completion, and classification. Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66

47 Kohonen networks These networks are two-layer networks that transform n-dimensional input patterns into lower-ordered data. Patterns project onto points in close proximity to one another. Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66

48 Kohonen networks Kohonen networks are trained in an unsupervised manner to form clusters within the data. pattern recognition, classification and data compression Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66

49 recurrents networks In a recurrent network, the outputs of some neurons are fed back to the same neurons or to neurons in preceding layers. Enables a flow of information in both forward and backward directions. Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66

50 recurrents networks Provides the ANN with a dynamic memory. Special algorithms for training recurrent networks. Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66

51 Counterpropagation networks These networks are trained by hybrid learning to create a self-organizing look-up table. A response is the average for those feature vectors closest to it in the input data space. Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66

52 Counterpropagation networks As input is presented unsupervised learning is carried out to create a Kohonen map of the input data. Meanwhile, supervised learning is used to associate an appropriate output vector with each point on the map. Useful for function approximation and classification. Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66

53 Backpropagation networks (feedforward) A backpropagation (BP) network is an MLP. The term backpropagation refers to the way the error computed at the output side is propagated backward from the output layer, to the hidden layer, and finally to the input layer. Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66

54 Backpropagation networks (feedforward) Uses supervised learning. Can be used for data modeling, classification, forecasting, control, data and image compression, and pattern recognition. Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66

55 General Issues with Backpropagation Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66

56 ANNs or expert Systems The decision as to whether to use ANNs, ESs, or theoretical modeling for an arbitrary problem depends on the availability of the theory and the data. For a problem with abundant data but unclear theory, ANNs can be a perfect tool. When both the data and theory are inadequate, the human experts opinion should be sought followed by coding this knowledge into a set of ES-rules. When the problem is rich in both data and theory, it may be possible to derive a physical model. When both theory and data are abundant but a physical model is hard to formulate, the modeler can also use ANNs. Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66

57 Graduation Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66

58 Phases in ANN development Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66

59 Problem Definition The problem definition and formulation (phase 1) relies heavily on an adequate understanding of the problem. The benefits of ANNs over other techniques should be evaluated before final selection of the modeling technique. Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66

60 System Design Determination of the type of ANN and learning rule that fit the problem. Involves data collection, data preprocessing to fit the type of ANN partitioning the data into three distinct subsets (training, test, and validation subsets). Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66

61 System Realization Training of the network utilizing the training and test subsets Assessing the network performance by analyzing the prediction error. Selection of the various parameters (e.g., network size, learning rate, number of training cycles, acceptable error, etc.) If possible, splitting the problem into smaller sub-problems, if possible, and designing an ensemble of networks Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66

62 System Verification Examination of the best network for its generalization capability using the validation subset. Comparison of the performance of the ANN-based model to those of other approaches. Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66

63 System Implementation Embedding the obtained network in an appropriate working system. Final testing of the integrated system. Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66

64 System Maintenance Updating the developed system as changes in the environment or the system variables occur. Involves a new development cycle. Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66

65 Thank you for your Attention Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66

66 Sources I.A. Basheer, M. Hajmeer: Artificial neural networks: fundamentals, computing, design, and application Uwe Lämmel, Jürgen Cleve: Künstliche Intelligenz 3. Auflage Günter Daniel Rey, Karl F. Wender: Neuronale Netze Eine Einführung in die Grundlagen, Anwendungen und Datenauswertung Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66

Artificial Neural Networks

Artificial Neural Networks Artificial Neural Networks Andres Chavez Math 382/L T/Th 2:00-3:40 April 13, 2010 Chavez2 Abstract The main interest of this paper is Artificial Neural Networks (ANNs). A brief history of the development

More information

Python Machine Learning

Python Machine Learning Python Machine Learning Unlock deeper insights into machine learning with this vital guide to cuttingedge predictive analytics Sebastian Raschka [ PUBLISHING 1 open source I community experience distilled

More information

Artificial Neural Networks written examination

Artificial Neural Networks written examination 1 (8) Institutionen för informationsteknologi Olle Gällmo Universitetsadjunkt Adress: Lägerhyddsvägen 2 Box 337 751 05 Uppsala Artificial Neural Networks written examination Monday, May 15, 2006 9 00-14

More information

Course Outline. Course Grading. Where to go for help. Academic Integrity. EE-589 Introduction to Neural Networks NN 1 EE

Course Outline. Course Grading. Where to go for help. Academic Integrity. EE-589 Introduction to Neural Networks NN 1 EE EE-589 Introduction to Neural Assistant Prof. Dr. Turgay IBRIKCI Room # 305 (322) 338 6868 / 139 Wensdays 9:00-12:00 Course Outline The course is divided in two parts: theory and practice. 1. Theory covers

More information

A Neural Network GUI Tested on Text-To-Phoneme Mapping

A Neural Network GUI Tested on Text-To-Phoneme Mapping A Neural Network GUI Tested on Text-To-Phoneme Mapping MAARTEN TROMPPER Universiteit Utrecht m.f.a.trompper@students.uu.nl Abstract Text-to-phoneme (T2P) mapping is a necessary step in any speech synthesis

More information

Learning Methods for Fuzzy Systems

Learning Methods for Fuzzy Systems Learning Methods for Fuzzy Systems Rudolf Kruse and Andreas Nürnberger Department of Computer Science, University of Magdeburg Universitätsplatz, D-396 Magdeburg, Germany Phone : +49.39.67.876, Fax : +49.39.67.8

More information

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur Module 12 Machine Learning 12.1 Instructional Objective The students should understand the concept of learning systems Students should learn about different aspects of a learning system Students should

More information

SARDNET: A Self-Organizing Feature Map for Sequences

SARDNET: A Self-Organizing Feature Map for Sequences SARDNET: A Self-Organizing Feature Map for Sequences Daniel L. James and Risto Miikkulainen Department of Computer Sciences The University of Texas at Austin Austin, TX 78712 dljames,risto~cs.utexas.edu

More information

Proposal of Pattern Recognition as a necessary and sufficient principle to Cognitive Science

Proposal of Pattern Recognition as a necessary and sufficient principle to Cognitive Science Proposal of Pattern Recognition as a necessary and sufficient principle to Cognitive Science Gilberto de Paiva Sao Paulo Brazil (May 2011) gilbertodpaiva@gmail.com Abstract. Despite the prevalence of the

More information

*** * * * COUNCIL * * CONSEIL OFEUROPE * * * DE L'EUROPE. Proceedings of the 9th Symposium on Legal Data Processing in Europe

*** * * * COUNCIL * * CONSEIL OFEUROPE * * * DE L'EUROPE. Proceedings of the 9th Symposium on Legal Data Processing in Europe *** * * * COUNCIL * * CONSEIL OFEUROPE * * * DE L'EUROPE Proceedings of the 9th Symposium on Legal Data Processing in Europe Bonn, 10-12 October 1989 Systems based on artificial intelligence in the legal

More information

QuickStroke: An Incremental On-line Chinese Handwriting Recognition System

QuickStroke: An Incremental On-line Chinese Handwriting Recognition System QuickStroke: An Incremental On-line Chinese Handwriting Recognition System Nada P. Matić John C. Platt Λ Tony Wang y Synaptics, Inc. 2381 Bering Drive San Jose, CA 95131, USA Abstract This paper presents

More information

Issues in the Mining of Heart Failure Datasets

Issues in the Mining of Heart Failure Datasets International Journal of Automation and Computing 11(2), April 2014, 162-179 DOI: 10.1007/s11633-014-0778-5 Issues in the Mining of Heart Failure Datasets Nongnuch Poolsawad 1 Lisa Moore 1 Chandrasekhar

More information

Lecture 1: Machine Learning Basics

Lecture 1: Machine Learning Basics 1/69 Lecture 1: Machine Learning Basics Ali Harakeh University of Waterloo WAVE Lab ali.harakeh@uwaterloo.ca May 1, 2017 2/69 Overview 1 Learning Algorithms 2 Capacity, Overfitting, and Underfitting 3

More information

Analysis of Hybrid Soft and Hard Computing Techniques for Forex Monitoring Systems

Analysis of Hybrid Soft and Hard Computing Techniques for Forex Monitoring Systems Analysis of Hybrid Soft and Hard Computing Techniques for Forex Monitoring Systems Ajith Abraham School of Business Systems, Monash University, Clayton, Victoria 3800, Australia. Email: ajith.abraham@ieee.org

More information

Unsupervised Learning of Word Semantic Embedding using the Deep Structured Semantic Model

Unsupervised Learning of Word Semantic Embedding using the Deep Structured Semantic Model Unsupervised Learning of Word Semantic Embedding using the Deep Structured Semantic Model Xinying Song, Xiaodong He, Jianfeng Gao, Li Deng Microsoft Research, One Microsoft Way, Redmond, WA 98052, U.S.A.

More information

Analysis of Speech Recognition Models for Real Time Captioning and Post Lecture Transcription

Analysis of Speech Recognition Models for Real Time Captioning and Post Lecture Transcription Analysis of Speech Recognition Models for Real Time Captioning and Post Lecture Transcription Wilny Wilson.P M.Tech Computer Science Student Thejus Engineering College Thrissur, India. Sindhu.S Computer

More information

(Sub)Gradient Descent

(Sub)Gradient Descent (Sub)Gradient Descent CMSC 422 MARINE CARPUAT marine@cs.umd.edu Figures credit: Piyush Rai Logistics Midterm is on Thursday 3/24 during class time closed book/internet/etc, one page of notes. will include

More information

Time series prediction

Time series prediction Chapter 13 Time series prediction Amaury Lendasse, Timo Honkela, Federico Pouzols, Antti Sorjamaa, Yoan Miche, Qi Yu, Eric Severin, Mark van Heeswijk, Erkki Oja, Francesco Corona, Elia Liitiäinen, Zhanxing

More information

Softprop: Softmax Neural Network Backpropagation Learning

Softprop: Softmax Neural Network Backpropagation Learning Softprop: Softmax Neural Networ Bacpropagation Learning Michael Rimer Computer Science Department Brigham Young University Provo, UT 84602, USA E-mail: mrimer@axon.cs.byu.edu Tony Martinez Computer Science

More information

Evolution of Symbolisation in Chimpanzees and Neural Nets

Evolution of Symbolisation in Chimpanzees and Neural Nets Evolution of Symbolisation in Chimpanzees and Neural Nets Angelo Cangelosi Centre for Neural and Adaptive Systems University of Plymouth (UK) a.cangelosi@plymouth.ac.uk Introduction Animal communication

More information

Evolutive Neural Net Fuzzy Filtering: Basic Description

Evolutive Neural Net Fuzzy Filtering: Basic Description Journal of Intelligent Learning Systems and Applications, 2010, 2: 12-18 doi:10.4236/jilsa.2010.21002 Published Online February 2010 (http://www.scirp.org/journal/jilsa) Evolutive Neural Net Fuzzy Filtering:

More information

Neuroscience I. BIOS/PHIL/PSCH 484 MWF 1:00-1:50 Lecture Center F6. Fall credit hours

Neuroscience I. BIOS/PHIL/PSCH 484 MWF 1:00-1:50 Lecture Center F6. Fall credit hours INSTRUCTOR INFORMATION Dr. John Leonard (course coordinator) Neuroscience I BIOS/PHIL/PSCH 484 MWF 1:00-1:50 Lecture Center F6 Fall 2016 3 credit hours leonard@uic.edu Biological Sciences 3055 SEL 312-996-4261

More information

Axiom 2013 Team Description Paper

Axiom 2013 Team Description Paper Axiom 2013 Team Description Paper Mohammad Ghazanfari, S Omid Shirkhorshidi, Farbod Samsamipour, Hossein Rahmatizadeh Zagheli, Mohammad Mahdavi, Payam Mohajeri, S Abbas Alamolhoda Robotics Scientific Association

More information

Framewise Phoneme Classification with Bidirectional LSTM and Other Neural Network Architectures

Framewise Phoneme Classification with Bidirectional LSTM and Other Neural Network Architectures Framewise Phoneme Classification with Bidirectional LSTM and Other Neural Network Architectures Alex Graves and Jürgen Schmidhuber IDSIA, Galleria 2, 6928 Manno-Lugano, Switzerland TU Munich, Boltzmannstr.

More information

Test Effort Estimation Using Neural Network

Test Effort Estimation Using Neural Network J. Software Engineering & Applications, 2010, 3: 331-340 doi:10.4236/jsea.2010.34038 Published Online April 2010 (http://www.scirp.org/journal/jsea) 331 Chintala Abhishek*, Veginati Pavan Kumar, Harish

More information

Knowledge-Based - Systems

Knowledge-Based - Systems Knowledge-Based - Systems ; Rajendra Arvind Akerkar Chairman, Technomathematics Research Foundation and Senior Researcher, Western Norway Research institute Priti Srinivas Sajja Sardar Patel University

More information

OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS

OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS Václav Kocian, Eva Volná, Michal Janošek, Martin Kotyrba University of Ostrava Department of Informatics and Computers Dvořákova 7,

More information

Language Acquisition Fall 2010/Winter Lexical Categories. Afra Alishahi, Heiner Drenhaus

Language Acquisition Fall 2010/Winter Lexical Categories. Afra Alishahi, Heiner Drenhaus Language Acquisition Fall 2010/Winter 2011 Lexical Categories Afra Alishahi, Heiner Drenhaus Computational Linguistics and Phonetics Saarland University Children s Sensitivity to Lexical Categories Look,

More information

CSL465/603 - Machine Learning

CSL465/603 - Machine Learning CSL465/603 - Machine Learning Fall 2016 Narayanan C Krishnan ckn@iitrpr.ac.in Introduction CSL465/603 - Machine Learning 1 Administrative Trivia Course Structure 3-0-2 Lecture Timings Monday 9.55-10.45am

More information

A New Perspective on Combining GMM and DNN Frameworks for Speaker Adaptation

A New Perspective on Combining GMM and DNN Frameworks for Speaker Adaptation A New Perspective on Combining GMM and DNN Frameworks for Speaker Adaptation SLSP-2016 October 11-12 Natalia Tomashenko 1,2,3 natalia.tomashenko@univ-lemans.fr Yuri Khokhlov 3 khokhlov@speechpro.com Yannick

More information

System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks

System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks 1 Tzu-Hsuan Yang, 2 Tzu-Hsuan Tseng, and 3 Chia-Ping Chen Department of Computer Science and Engineering

More information

Deep search. Enhancing a search bar using machine learning. Ilgün Ilgün & Cedric Reichenbach

Deep search. Enhancing a search bar using machine learning. Ilgün Ilgün & Cedric Reichenbach #BaselOne7 Deep search Enhancing a search bar using machine learning Ilgün Ilgün & Cedric Reichenbach We are not researchers Outline I. Periscope: A search tool II. Goals III. Deep learning IV. Applying

More information

Classification Using ANN: A Review

Classification Using ANN: A Review International Journal of Computational Intelligence Research ISSN 0973-1873 Volume 13, Number 7 (2017), pp. 1811-1820 Research India Publications http://www.ripublication.com Classification Using ANN:

More information

INPE São José dos Campos

INPE São José dos Campos INPE-5479 PRE/1778 MONLINEAR ASPECTS OF DATA INTEGRATION FOR LAND COVER CLASSIFICATION IN A NEDRAL NETWORK ENVIRONNENT Maria Suelena S. Barros Valter Rodrigues INPE São José dos Campos 1993 SECRETARIA

More information

Human Emotion Recognition From Speech

Human Emotion Recognition From Speech RESEARCH ARTICLE OPEN ACCESS Human Emotion Recognition From Speech Miss. Aparna P. Wanare*, Prof. Shankar N. Dandare *(Department of Electronics & Telecommunication Engineering, Sant Gadge Baba Amravati

More information

Neuro-Symbolic Approaches for Knowledge Representation in Expert Systems

Neuro-Symbolic Approaches for Knowledge Representation in Expert Systems Published in the International Journal of Hybrid Intelligent Systems 1(3-4) (2004) 111-126 Neuro-Symbolic Approaches for Knowledge Representation in Expert Systems Ioannis Hatzilygeroudis and Jim Prentzas

More information

An empirical study of learning speed in backpropagation

An empirical study of learning speed in backpropagation Carnegie Mellon University Research Showcase @ CMU Computer Science Department School of Computer Science 1988 An empirical study of learning speed in backpropagation networks Scott E. Fahlman Carnegie

More information

Speaker Identification by Comparison of Smart Methods. Abstract

Speaker Identification by Comparison of Smart Methods. Abstract Journal of mathematics and computer science 10 (2014), 61-71 Speaker Identification by Comparison of Smart Methods Ali Mahdavi Meimand Amin Asadi Majid Mohamadi Department of Electrical Department of Computer

More information

I-COMPETERE: Using Applied Intelligence in search of competency gaps in software project managers.

I-COMPETERE: Using Applied Intelligence in search of competency gaps in software project managers. Information Systems Frontiers manuscript No. (will be inserted by the editor) I-COMPETERE: Using Applied Intelligence in search of competency gaps in software project managers. Ricardo Colomo-Palacios

More information

Autoregressive product of multi-frame predictions can improve the accuracy of hybrid models

Autoregressive product of multi-frame predictions can improve the accuracy of hybrid models Autoregressive product of multi-frame predictions can improve the accuracy of hybrid models Navdeep Jaitly 1, Vincent Vanhoucke 2, Geoffrey Hinton 1,2 1 University of Toronto 2 Google Inc. ndjaitly@cs.toronto.edu,

More information

Lecture 1: Basic Concepts of Machine Learning

Lecture 1: Basic Concepts of Machine Learning Lecture 1: Basic Concepts of Machine Learning Cognitive Systems - Machine Learning Ute Schmid (lecture) Johannes Rabold (practice) Based on slides prepared March 2005 by Maximilian Röglinger, updated 2010

More information

Seminar - Organic Computing

Seminar - Organic Computing Seminar - Organic Computing Self-Organisation of OC-Systems Markus Franke 25.01.2006 Typeset by FoilTEX Timetable 1. Overview 2. Characteristics of SO-Systems 3. Concern with Nature 4. Design-Concepts

More information

Knowledge Transfer in Deep Convolutional Neural Nets

Knowledge Transfer in Deep Convolutional Neural Nets Knowledge Transfer in Deep Convolutional Neural Nets Steven Gutstein, Olac Fuentes and Eric Freudenthal Computer Science Department University of Texas at El Paso El Paso, Texas, 79968, U.S.A. Abstract

More information

International Journal of Advanced Networking Applications (IJANA) ISSN No. :

International Journal of Advanced Networking Applications (IJANA) ISSN No. : International Journal of Advanced Networking Applications (IJANA) ISSN No. : 0975-0290 34 A Review on Dysarthric Speech Recognition Megha Rughani Department of Electronics and Communication, Marwadi Educational

More information

On the Formation of Phoneme Categories in DNN Acoustic Models

On the Formation of Phoneme Categories in DNN Acoustic Models On the Formation of Phoneme Categories in DNN Acoustic Models Tasha Nagamine Department of Electrical Engineering, Columbia University T. Nagamine Motivation Large performance gap between humans and state-

More information

Forget catastrophic forgetting: AI that learns after deployment

Forget catastrophic forgetting: AI that learns after deployment Forget catastrophic forgetting: AI that learns after deployment Anatoly Gorshechnikov CTO, Neurala 1 Neurala at a glance Programming neural networks on GPUs since circa 2 B.C. Founded in 2006 expecting

More information

FUZZY EXPERT. Dr. Kasim M. Al-Aubidy. Philadelphia University. Computer Eng. Dept February 2002 University of Damascus-Syria

FUZZY EXPERT. Dr. Kasim M. Al-Aubidy. Philadelphia University. Computer Eng. Dept February 2002 University of Damascus-Syria FUZZY EXPERT SYSTEMS 16-18 18 February 2002 University of Damascus-Syria Dr. Kasim M. Al-Aubidy Computer Eng. Dept. Philadelphia University What is Expert Systems? ES are computer programs that emulate

More information

Syntactic systematicity in sentence processing with a recurrent self-organizing network

Syntactic systematicity in sentence processing with a recurrent self-organizing network Syntactic systematicity in sentence processing with a recurrent self-organizing network Igor Farkaš,1 Department of Applied Informatics, Comenius University Mlynská dolina, 842 48 Bratislava, Slovak Republic

More information

Calibration of Confidence Measures in Speech Recognition

Calibration of Confidence Measures in Speech Recognition Submitted to IEEE Trans on Audio, Speech, and Language, July 2010 1 Calibration of Confidence Measures in Speech Recognition Dong Yu, Senior Member, IEEE, Jinyu Li, Member, IEEE, Li Deng, Fellow, IEEE

More information

Communication and Cybernetics 17

Communication and Cybernetics 17 Communication and Cybernetics 17 Editors: K. S. Fu W. D. Keidel W. J. M. Levelt H. Wolter Communication and Cybernetics Editors: K.S.Fu, W.D.Keidel, W.1.M.Levelt, H.Wolter Vol. Vol. 2 Vol. 3 Vol. 4 Vol.

More information

MASTER OF SCIENCE (M.S.) MAJOR IN COMPUTER SCIENCE

MASTER OF SCIENCE (M.S.) MAJOR IN COMPUTER SCIENCE Master of Science (M.S.) Major in Computer Science 1 MASTER OF SCIENCE (M.S.) MAJOR IN COMPUTER SCIENCE Major Program The programs in computer science are designed to prepare students for doctoral research,

More information

Attributed Social Network Embedding

Attributed Social Network Embedding JOURNAL OF LATEX CLASS FILES, VOL. 14, NO. 8, MAY 2017 1 Attributed Social Network Embedding arxiv:1705.04969v1 [cs.si] 14 May 2017 Lizi Liao, Xiangnan He, Hanwang Zhang, and Tat-Seng Chua Abstract Embedding

More information

Modeling function word errors in DNN-HMM based LVCSR systems

Modeling function word errors in DNN-HMM based LVCSR systems Modeling function word errors in DNN-HMM based LVCSR systems Melvin Jose Johnson Premkumar, Ankur Bapna and Sree Avinash Parchuri Department of Computer Science Department of Electrical Engineering Stanford

More information

Breaking the Habit of Being Yourself Workshop for Quantum University

Breaking the Habit of Being Yourself Workshop for Quantum University Breaking the Habit of Being Yourself Workshop for Quantum University 2 Copyright Dr Joe Dispenza. June 2013. All rights reserved. 3 Copyright Dr Joe Dispenza. June 2013. All rights reserved. 4 Copyright

More information

Lecture 10: Reinforcement Learning

Lecture 10: Reinforcement Learning Lecture 1: Reinforcement Learning Cognitive Systems II - Machine Learning SS 25 Part III: Learning Programs and Strategies Q Learning, Dynamic Programming Lecture 1: Reinforcement Learning p. Motivation

More information

CS Machine Learning

CS Machine Learning CS 478 - Machine Learning Projects Data Representation Basic testing and evaluation schemes CS 478 Data and Testing 1 Programming Issues l Program in any platform you want l Realize that you will be doing

More information

arxiv: v1 [cs.cv] 10 May 2017

arxiv: v1 [cs.cv] 10 May 2017 Inferring and Executing Programs for Visual Reasoning Justin Johnson 1 Bharath Hariharan 2 Laurens van der Maaten 2 Judy Hoffman 1 Li Fei-Fei 1 C. Lawrence Zitnick 2 Ross Girshick 2 1 Stanford University

More information

Spinal Cord. Student Pages. Classroom Ac tivities

Spinal Cord. Student Pages. Classroom Ac tivities Classroom Ac tivities Spinal Cord Student Pages Produced by Regenerative Medicine Partnership in Education Duquesne University Director john A. Pollock (pollock@duq.edu) The spinal column protects the

More information

Accelerated Learning Course Outline

Accelerated Learning Course Outline Accelerated Learning Course Outline Course Description The purpose of this course is to make the advances in the field of brain research more accessible to educators. The techniques and strategies of Accelerated

More information

Modeling function word errors in DNN-HMM based LVCSR systems

Modeling function word errors in DNN-HMM based LVCSR systems Modeling function word errors in DNN-HMM based LVCSR systems Melvin Jose Johnson Premkumar, Ankur Bapna and Sree Avinash Parchuri Department of Computer Science Department of Electrical Engineering Stanford

More information

Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models

Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models Stephan Gouws and GJ van Rooyen MIH Medialab, Stellenbosch University SOUTH AFRICA {stephan,gvrooyen}@ml.sun.ac.za

More information

Accelerated Learning Online. Course Outline

Accelerated Learning Online. Course Outline Accelerated Learning Online Course Outline Course Description The purpose of this course is to make the advances in the field of brain research more accessible to educators. The techniques and strategies

More information

HIERARCHICAL DEEP LEARNING ARCHITECTURE FOR 10K OBJECTS CLASSIFICATION

HIERARCHICAL DEEP LEARNING ARCHITECTURE FOR 10K OBJECTS CLASSIFICATION HIERARCHICAL DEEP LEARNING ARCHITECTURE FOR 10K OBJECTS CLASSIFICATION Atul Laxman Katole 1, Krishna Prasad Yellapragada 1, Amish Kumar Bedi 1, Sehaj Singh Kalra 1 and Mynepalli Siva Chaitanya 1 1 Samsung

More information

Second Exam: Natural Language Parsing with Neural Networks

Second Exam: Natural Language Parsing with Neural Networks Second Exam: Natural Language Parsing with Neural Networks James Cross May 21, 2015 Abstract With the advent of deep learning, there has been a recent resurgence of interest in the use of artificial neural

More information

Lecture 2: Quantifiers and Approximation

Lecture 2: Quantifiers and Approximation Lecture 2: Quantifiers and Approximation Case study: Most vs More than half Jakub Szymanik Outline Number Sense Approximate Number Sense Approximating most Superlative Meaning of most What About Counting?

More information

Phonetic- and Speaker-Discriminant Features for Speaker Recognition. Research Project

Phonetic- and Speaker-Discriminant Features for Speaker Recognition. Research Project Phonetic- and Speaker-Discriminant Features for Speaker Recognition by Lara Stoll Research Project Submitted to the Department of Electrical Engineering and Computer Sciences, University of California

More information

arxiv: v1 [cs.lg] 15 Jun 2015

arxiv: v1 [cs.lg] 15 Jun 2015 Dual Memory Architectures for Fast Deep Learning of Stream Data via an Online-Incremental-Transfer Strategy arxiv:1506.04477v1 [cs.lg] 15 Jun 2015 Sang-Woo Lee Min-Oh Heo School of Computer Science and

More information

The Method of Immersion the Problem of Comparing Technical Objects in an Expert Shell in the Class of Artificial Intelligence Algorithms

The Method of Immersion the Problem of Comparing Technical Objects in an Expert Shell in the Class of Artificial Intelligence Algorithms IOP Conference Series: Materials Science and Engineering PAPER OPEN ACCESS The Method of Immersion the Problem of Comparing Technical Objects in an Expert Shell in the Class of Artificial Intelligence

More information

1 NETWORKS VERSUS SYMBOL SYSTEMS: TWO APPROACHES TO MODELING COGNITION

1 NETWORKS VERSUS SYMBOL SYSTEMS: TWO APPROACHES TO MODELING COGNITION NETWORKS VERSUS SYMBOL SYSTEMS 1 1 NETWORKS VERSUS SYMBOL SYSTEMS: TWO APPROACHES TO MODELING COGNITION 1.1 A Revolution in the Making? The rise of cognitivism in psychology, which, by the 1970s, had successfully

More information

A Case Study: News Classification Based on Term Frequency

A Case Study: News Classification Based on Term Frequency A Case Study: News Classification Based on Term Frequency Petr Kroha Faculty of Computer Science University of Technology 09107 Chemnitz Germany kroha@informatik.tu-chemnitz.de Ricardo Baeza-Yates Center

More information

Networks in Cognitive Science

Networks in Cognitive Science 1 Networks in Cognitive Science Andrea Baronchelli 1,*, Ramon Ferrer-i-Cancho 2, Romualdo Pastor-Satorras 3, Nick Chater 4 and Morten H. Christiansen 5,6 1 Laboratory for the Modeling of Biological and

More information

Software Maintenance

Software Maintenance 1 What is Software Maintenance? Software Maintenance is a very broad activity that includes error corrections, enhancements of capabilities, deletion of obsolete capabilities, and optimization. 2 Categories

More information

Machine Learning and Data Mining. Ensembles of Learners. Prof. Alexander Ihler

Machine Learning and Data Mining. Ensembles of Learners. Prof. Alexander Ihler Machine Learning and Data Mining Ensembles of Learners Prof. Alexander Ihler Ensemble methods Why learn one classifier when you can learn many? Ensemble: combine many predictors (Weighted) combina

More information

IAT 888: Metacreation Machines endowed with creative behavior. Philippe Pasquier Office 565 (floor 14)

IAT 888: Metacreation Machines endowed with creative behavior. Philippe Pasquier Office 565 (floor 14) IAT 888: Metacreation Machines endowed with creative behavior Philippe Pasquier Office 565 (floor 14) pasquier@sfu.ca Outline of today's lecture A little bit about me A little bit about you What will that

More information

Probabilistic Latent Semantic Analysis

Probabilistic Latent Semantic Analysis Probabilistic Latent Semantic Analysis Thomas Hofmann Presentation by Ioannis Pavlopoulos & Andreas Damianou for the course of Data Mining & Exploration 1 Outline Latent Semantic Analysis o Need o Overview

More information

Predicting Early Students with High Risk to Drop Out of University using a Neural Network-Based Approach

Predicting Early Students with High Risk to Drop Out of University using a Neural Network-Based Approach Predicting Early Students with High Risk to Drop Out of University using a Neural Network-Based Approach Miguel Gil, Norma Reyes, María Juárez, Emmanuel Espitia, Julio Mosqueda and Myriam Soria Information

More information

Assignment 1: Predicting Amazon Review Ratings

Assignment 1: Predicting Amazon Review Ratings Assignment 1: Predicting Amazon Review Ratings 1 Dataset Analysis Richard Park r2park@acsmail.ucsd.edu February 23, 2015 The dataset selected for this assignment comes from the set of Amazon reviews for

More information

University of Groningen. Systemen, planning, netwerken Bosman, Aart

University of Groningen. Systemen, planning, netwerken Bosman, Aart University of Groningen Systemen, planning, netwerken Bosman, Aart IMPORTANT NOTE: You are advised to consult the publisher's version (publisher's PDF) if you wish to cite from it. Please check the document

More information

BI408-01: Cellular and Molecular Neurobiology

BI408-01: Cellular and Molecular Neurobiology BI408-01: Cellular and Molecular Neurobiology Spring 2013 Instructor: Jennifer R. Kowalski, Ph.D. Office: Gallahue Hall 271 Phone: 940-8879 Office Hours: 10:00-11:30 a.m. Mon. and Wed. E-mail: jrkowals@butler.edu

More information

A Comparison of Annealing Techniques for Academic Course Scheduling

A Comparison of Annealing Techniques for Academic Course Scheduling A Comparison of Annealing Techniques for Academic Course Scheduling M. A. Saleh Elmohamed 1, Paul Coddington 2, and Geoffrey Fox 1 1 Northeast Parallel Architectures Center Syracuse University, Syracuse,

More information

How People Learn Physics

How People Learn Physics How People Learn Physics Edward F. (Joe) Redish Dept. Of Physics University Of Maryland AAPM, Houston TX, Work supported in part by NSF grants DUE #04-4-0113 and #05-2-4987 Teaching complex subjects 2

More information

Laboratorio di Intelligenza Artificiale e Robotica

Laboratorio di Intelligenza Artificiale e Robotica Laboratorio di Intelligenza Artificiale e Robotica A.A. 2008-2009 Outline 2 Machine Learning Unsupervised Learning Supervised Learning Reinforcement Learning Genetic Algorithms Genetics-Based Machine Learning

More information

Neural Representation and Neural Computation. Philosophical Perspectives, Vol. 4, Action Theory and Philosophy of Mind (1990),

Neural Representation and Neural Computation. Philosophical Perspectives, Vol. 4, Action Theory and Philosophy of Mind (1990), Neural Representation and Neural Computation Patricia Smith Churchland; Terrence J. Sejnowski Philosophical Perspectives, Vol. 4, Action Theory and Philosophy of Mind (1990), 343-382. Stable URL: http://links.jstor.org/sici?sici=1520-8583%28

More information

ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY DOWNLOAD EBOOK : ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY PDF

ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY DOWNLOAD EBOOK : ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY PDF Read Online and Download Ebook ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY DOWNLOAD EBOOK : ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY PDF Click link bellow and free register to download

More information

DIRECT ADAPTATION OF HYBRID DNN/HMM MODEL FOR FAST SPEAKER ADAPTATION IN LVCSR BASED ON SPEAKER CODE

DIRECT ADAPTATION OF HYBRID DNN/HMM MODEL FOR FAST SPEAKER ADAPTATION IN LVCSR BASED ON SPEAKER CODE 2014 IEEE International Conference on Acoustic, Speech and Signal Processing (ICASSP) DIRECT ADAPTATION OF HYBRID DNN/HMM MODEL FOR FAST SPEAKER ADAPTATION IN LVCSR BASED ON SPEAKER CODE Shaofei Xue 1

More information

Machine Learning from Garden Path Sentences: The Application of Computational Linguistics

Machine Learning from Garden Path Sentences: The Application of Computational Linguistics Machine Learning from Garden Path Sentences: The Application of Computational Linguistics http://dx.doi.org/10.3991/ijet.v9i6.4109 J.L. Du 1, P.F. Yu 1 and M.L. Li 2 1 Guangdong University of Foreign Studies,

More information

TD(λ) and Q-Learning Based Ludo Players

TD(λ) and Q-Learning Based Ludo Players TD(λ) and Q-Learning Based Ludo Players Majed Alhajry, Faisal Alvi, Member, IEEE and Moataz Ahmed Abstract Reinforcement learning is a popular machine learning technique whose inherent self-learning ability

More information

BUSINESS INTELLIGENCE FROM WEB USAGE MINING

BUSINESS INTELLIGENCE FROM WEB USAGE MINING BUSINESS INTELLIGENCE FROM WEB USAGE MINING Ajith Abraham Department of Computer Science, Oklahoma State University, 700 N Greenwood Avenue, Tulsa,Oklahoma 74106-0700, USA, ajith.abraham@ieee.org Abstract.

More information

Generative models and adversarial training

Generative models and adversarial training Day 4 Lecture 1 Generative models and adversarial training Kevin McGuinness kevin.mcguinness@dcu.ie Research Fellow Insight Centre for Data Analytics Dublin City University What is a generative model?

More information

WHEN THERE IS A mismatch between the acoustic

WHEN THERE IS A mismatch between the acoustic 808 IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 14, NO. 3, MAY 2006 Optimization of Temporal Filters for Constructing Robust Features in Speech Recognition Jeih-Weih Hung, Member,

More information

Learning Methods in Multilingual Speech Recognition

Learning Methods in Multilingual Speech Recognition Learning Methods in Multilingual Speech Recognition Hui Lin Department of Electrical Engineering University of Washington Seattle, WA 98125 linhui@u.washington.edu Li Deng, Jasha Droppo, Dong Yu, and Alex

More information

Purdue Data Summit Communication of Big Data Analytics. New SAT Predictive Validity Case Study

Purdue Data Summit Communication of Big Data Analytics. New SAT Predictive Validity Case Study Purdue Data Summit 2017 Communication of Big Data Analytics New SAT Predictive Validity Case Study Paul M. Johnson, Ed.D. Associate Vice President for Enrollment Management, Research & Enrollment Information

More information

BUILDING CONTEXT-DEPENDENT DNN ACOUSTIC MODELS USING KULLBACK-LEIBLER DIVERGENCE-BASED STATE TYING

BUILDING CONTEXT-DEPENDENT DNN ACOUSTIC MODELS USING KULLBACK-LEIBLER DIVERGENCE-BASED STATE TYING BUILDING CONTEXT-DEPENDENT DNN ACOUSTIC MODELS USING KULLBACK-LEIBLER DIVERGENCE-BASED STATE TYING Gábor Gosztolya 1, Tamás Grósz 1, László Tóth 1, David Imseng 2 1 MTA-SZTE Research Group on Artificial

More information

Automating the E-learning Personalization

Automating the E-learning Personalization Automating the E-learning Personalization Fathi Essalmi 1, Leila Jemni Ben Ayed 1, Mohamed Jemni 1, Kinshuk 2, and Sabine Graf 2 1 The Research Laboratory of Technologies of Information and Communication

More information

A study of speaker adaptation for DNN-based speech synthesis

A study of speaker adaptation for DNN-based speech synthesis A study of speaker adaptation for DNN-based speech synthesis Zhizheng Wu, Pawel Swietojanski, Christophe Veaux, Steve Renals, Simon King The Centre for Speech Technology Research (CSTR) University of Edinburgh,

More information

Henry Tirri* Petri Myllymgki

Henry Tirri* Petri Myllymgki From: AAAI Technical Report SS-93-04. Compilation copyright 1993, AAAI (www.aaai.org). All rights reserved. Bayesian Case-Based Reasoning with Neural Networks Petri Myllymgki Henry Tirri* email: University

More information

Learning to Schedule Straight-Line Code

Learning to Schedule Straight-Line Code Learning to Schedule Straight-Line Code Eliot Moss, Paul Utgoff, John Cavazos Doina Precup, Darko Stefanović Dept. of Comp. Sci., Univ. of Mass. Amherst, MA 01003 Carla Brodley, David Scheeff Sch. of Elec.

More information

An OO Framework for building Intelligence and Learning properties in Software Agents

An OO Framework for building Intelligence and Learning properties in Software Agents An OO Framework for building Intelligence and Learning properties in Software Agents José A. R. P. Sardinha, Ruy L. Milidiú, Carlos J. P. Lucena, Patrick Paranhos Abstract Software agents are defined as

More information

An Online Handwriting Recognition System For Turkish

An Online Handwriting Recognition System For Turkish An Online Handwriting Recognition System For Turkish Esra Vural, Hakan Erdogan, Kemal Oflazer, Berrin Yanikoglu Sabanci University, Tuzla, Istanbul, Turkey 34956 ABSTRACT Despite recent developments in

More information

AUTOMATED FABRIC DEFECT INSPECTION: A SURVEY OF CLASSIFIERS

AUTOMATED FABRIC DEFECT INSPECTION: A SURVEY OF CLASSIFIERS AUTOMATED FABRIC DEFECT INSPECTION: A SURVEY OF CLASSIFIERS Md. Tarek Habib 1, Rahat Hossain Faisal 2, M. Rokonuzzaman 3, Farruk Ahmed 4 1 Department of Computer Science and Engineering, Prime University,

More information