Introduction of connectionist models
|
|
- Lucas Horn
- 6 years ago
- Views:
Transcription
1 Introduction of connectionist models Introduction to ANNs Markus Dambek Uni Bremen 20. Dezember 2010 Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66
2 1 Introduction 2 Information processing in biology 3 The artificial neuron 4 The perceptron 5 The multilayer perceptron 6 Classification 7 Development Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66
3 About what? Connectionism is a set of approaches in the fields of artificial intelligence, cognitive psychology, cognitive science, neuroscience and philosophy of mind, that models mental or behavioral phenomena as the emergent processes of interconnected networks of simple units. There are many forms of connectionism, but the most common forms use artificial neural networks (ANNs) Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66
4 ANNs Artificial neural networks (ANNs) are relatively new computational tools that have found extensive utilization in solving many complex real-world problems. The attractiveness of ANNs comes from their remarkable information processing characteristics pertinent mainly to nonlinearity Although ANNs are drastic abstractions of the biological counterparts, the idea of ANNs is not to replicate the operation of the biological systems. Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66
5 Capabilities nonlinearity noise-insensitivity Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66
6 Capabilities high parallelism learning generalization Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66
7 Information processing in biology Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66
8 The Neuron Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66
9 An impulse, in the form of an electric signal,travels within the dendrites and through the cell body towards the pre-synaptic membrane of the synapse. Upon arrival at the membrane, a neurotransmitter (chemical) is released from the vesicles in quantities proportional to the strength of the incoming signal. The neurotransmitter diffuses within the synaptic gap and eventually into the dendrites of neighboring neurons. Depending on the neighboring neuron s threshold, they force them to generate a new electrical signal. Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66
10 Neural Networks Such neurons are interconnected in networks of billions of neurons, processing incoming information into motor actions and new information. Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66
11 The artificial neuron Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66
12 The artificial neuron Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66
13 Analogy connections = dendrites and axons connection weights = synapses threshold = soma activity Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66
14 Components Input x i w ij net-input-function net j = k x k w kj activation-function a j = f act (net j,θ j ) output-function x j = f out (a j ) (mostly : f out = id) Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66
15 Activation Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66
16 The perceptron Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66
17 Perceptron Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66
18 Perceptron unidirectional two Layers: Input-Layer and Output-Layer Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66
19 Learning Delta-rule Change weights according to the contribution to the error Error denotes the difference between the perceptron s output and the expected output (teaching output) Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66
20 Learning Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66
21 AND-Net Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66
22 AND-Net x 1 x 2 x 1 *1+x 2 *1 > θ = 1,5 output false false false true 1 Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66
23 OR-Net Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66
24 OR-Net x 1 x 2 x 1 *1+x 2 *1 > θ = 0,5 output false true true true 1 Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66
25 XOR-Net Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66
26 Capabilities Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66
27 Capabilities Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66
28 The multilayer perceptron Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66
29 Multilayer-Perceptron Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66
30 Multilayer-Perceptron Adding Hidden-Layers results in the capability to represent even more komplex functions But delta-rule operates on the difference between the perceptron s output and the teaching output We have no clue about the hidden neuron s taching output We need to modify the learning algorithm Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66
31 Many different approaches: Backpropagation Hopfield Adaptive resonance theory... Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66
32 Learning: supervised vs. unsupervised vs. reinforcement Most of them can be divided into three different classes of learning algorithms: supervised learing unsupervised learning reinforcement learning Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66
33 error-correction learning The error-correction learning (ECL) rule is used in supervised learning. The arithmetic difference (error) between the ANN solution and the corresponding correct answer is used to modify the connection weights. Gradual reduction the overall network error. Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66
34 boltzman learning The Boltzmann learning (BL) rule is a stochastic rule. It is similar to ECL, however each neuron generates an output (or state) based on a Boltzmann statistical distribution Renders learning extremely slow. Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66
35 hebbian learning The Hebbian learning (HL) rule, developed based on neurobiological experiments, is the oldest learning rule. It postulates that if neurons on both sides of a synapse are activated synchronously and repeatedly, the synapses strength is selectively increased. Learning is done locally by adjusting the synapse weight based on the activities of the neurons. Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66
36 competitive learning In the competitive learning (CL) rule, all neurons are forced to compete among themselves. Only one neuron will be activated in a given iteration with all the weights attached to it being adjusted. The CL rule is speculated to exist in many biological systems Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66
37 Backpropagation Generalization of delta-rule error-correction learning Idea Represent the net-error as Function of all weights: E(W j ) = E(w j1,w j2,...,w jn ) two weights w 1 and w 2 Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66
38 Backpropagation Objective Find the global minimum of the Error-Function Use Gradient descent Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66
39 Backpropagation Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66
40 BP-XOR-Net Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66
41 BP-XOR-Net x 1 x 2 x 1*1+x 2*1 > θ = 1,5 x 3 x 1*1+x 2*1+x 3*(-2) > θ = 0,5 output false 0 0 false false 0 1 true false 0 1 true true 1 0 false 0 Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66
42 Classification Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66
43 Hopfield networks This network is a symmetric fully connected twolayer recurrent network. When presented with an incomplete or noisy pattern, the network responds by retrieving an internally stored pattern that most closely resembles the presented pattern. Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66
44 Hopfield networks Efficient in solving optimization problems. Learning is done by setting each weight connecting two neurons to the product of the inputs of these two neurons. Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66
45 Adaptive resonance theory These are trained by unsupervised learning. The ART network consists of two fully interconnected layers, an input-layer and an output-layer. The feedforward weights are used to select a winning output neuron (cluster) and serve as the long-term memory for the networks. Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66
46 Adaptive resonance theory The feedback weights serve as the short-term memory for the network. Can be used for pattern recognition, completion, and classification. Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66
47 Kohonen networks These networks are two-layer networks that transform n-dimensional input patterns into lower-ordered data. Patterns project onto points in close proximity to one another. Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66
48 Kohonen networks Kohonen networks are trained in an unsupervised manner to form clusters within the data. pattern recognition, classification and data compression Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66
49 recurrents networks In a recurrent network, the outputs of some neurons are fed back to the same neurons or to neurons in preceding layers. Enables a flow of information in both forward and backward directions. Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66
50 recurrents networks Provides the ANN with a dynamic memory. Special algorithms for training recurrent networks. Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66
51 Counterpropagation networks These networks are trained by hybrid learning to create a self-organizing look-up table. A response is the average for those feature vectors closest to it in the input data space. Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66
52 Counterpropagation networks As input is presented unsupervised learning is carried out to create a Kohonen map of the input data. Meanwhile, supervised learning is used to associate an appropriate output vector with each point on the map. Useful for function approximation and classification. Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66
53 Backpropagation networks (feedforward) A backpropagation (BP) network is an MLP. The term backpropagation refers to the way the error computed at the output side is propagated backward from the output layer, to the hidden layer, and finally to the input layer. Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66
54 Backpropagation networks (feedforward) Uses supervised learning. Can be used for data modeling, classification, forecasting, control, data and image compression, and pattern recognition. Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66
55 General Issues with Backpropagation Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66
56 ANNs or expert Systems The decision as to whether to use ANNs, ESs, or theoretical modeling for an arbitrary problem depends on the availability of the theory and the data. For a problem with abundant data but unclear theory, ANNs can be a perfect tool. When both the data and theory are inadequate, the human experts opinion should be sought followed by coding this knowledge into a set of ES-rules. When the problem is rich in both data and theory, it may be possible to derive a physical model. When both theory and data are abundant but a physical model is hard to formulate, the modeler can also use ANNs. Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66
57 Graduation Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66
58 Phases in ANN development Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66
59 Problem Definition The problem definition and formulation (phase 1) relies heavily on an adequate understanding of the problem. The benefits of ANNs over other techniques should be evaluated before final selection of the modeling technique. Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66
60 System Design Determination of the type of ANN and learning rule that fit the problem. Involves data collection, data preprocessing to fit the type of ANN partitioning the data into three distinct subsets (training, test, and validation subsets). Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66
61 System Realization Training of the network utilizing the training and test subsets Assessing the network performance by analyzing the prediction error. Selection of the various parameters (e.g., network size, learning rate, number of training cycles, acceptable error, etc.) If possible, splitting the problem into smaller sub-problems, if possible, and designing an ensemble of networks Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66
62 System Verification Examination of the best network for its generalization capability using the validation subset. Comparison of the performance of the ANN-based model to those of other approaches. Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66
63 System Implementation Embedding the obtained network in an appropriate working system. Final testing of the integrated system. Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66
64 System Maintenance Updating the developed system as changes in the environment or the system variables occur. Involves a new development cycle. Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66
65 Thank you for your Attention Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66
66 Sources I.A. Basheer, M. Hajmeer: Artificial neural networks: fundamentals, computing, design, and application Uwe Lämmel, Jürgen Cleve: Künstliche Intelligenz 3. Auflage Günter Daniel Rey, Karl F. Wender: Neuronale Netze Eine Einführung in die Grundlagen, Anwendungen und Datenauswertung Markus Dambek (Uni Bremen) Introduction of connectionist models 20. Dezember / 66
Artificial Neural Networks
Artificial Neural Networks Andres Chavez Math 382/L T/Th 2:00-3:40 April 13, 2010 Chavez2 Abstract The main interest of this paper is Artificial Neural Networks (ANNs). A brief history of the development
More informationPython Machine Learning
Python Machine Learning Unlock deeper insights into machine learning with this vital guide to cuttingedge predictive analytics Sebastian Raschka [ PUBLISHING 1 open source I community experience distilled
More informationArtificial Neural Networks written examination
1 (8) Institutionen för informationsteknologi Olle Gällmo Universitetsadjunkt Adress: Lägerhyddsvägen 2 Box 337 751 05 Uppsala Artificial Neural Networks written examination Monday, May 15, 2006 9 00-14
More informationCourse Outline. Course Grading. Where to go for help. Academic Integrity. EE-589 Introduction to Neural Networks NN 1 EE
EE-589 Introduction to Neural Assistant Prof. Dr. Turgay IBRIKCI Room # 305 (322) 338 6868 / 139 Wensdays 9:00-12:00 Course Outline The course is divided in two parts: theory and practice. 1. Theory covers
More informationA Neural Network GUI Tested on Text-To-Phoneme Mapping
A Neural Network GUI Tested on Text-To-Phoneme Mapping MAARTEN TROMPPER Universiteit Utrecht m.f.a.trompper@students.uu.nl Abstract Text-to-phoneme (T2P) mapping is a necessary step in any speech synthesis
More informationLearning Methods for Fuzzy Systems
Learning Methods for Fuzzy Systems Rudolf Kruse and Andreas Nürnberger Department of Computer Science, University of Magdeburg Universitätsplatz, D-396 Magdeburg, Germany Phone : +49.39.67.876, Fax : +49.39.67.8
More informationModule 12. Machine Learning. Version 2 CSE IIT, Kharagpur
Module 12 Machine Learning 12.1 Instructional Objective The students should understand the concept of learning systems Students should learn about different aspects of a learning system Students should
More informationSARDNET: A Self-Organizing Feature Map for Sequences
SARDNET: A Self-Organizing Feature Map for Sequences Daniel L. James and Risto Miikkulainen Department of Computer Sciences The University of Texas at Austin Austin, TX 78712 dljames,risto~cs.utexas.edu
More informationProposal of Pattern Recognition as a necessary and sufficient principle to Cognitive Science
Proposal of Pattern Recognition as a necessary and sufficient principle to Cognitive Science Gilberto de Paiva Sao Paulo Brazil (May 2011) gilbertodpaiva@gmail.com Abstract. Despite the prevalence of the
More information*** * * * COUNCIL * * CONSEIL OFEUROPE * * * DE L'EUROPE. Proceedings of the 9th Symposium on Legal Data Processing in Europe
*** * * * COUNCIL * * CONSEIL OFEUROPE * * * DE L'EUROPE Proceedings of the 9th Symposium on Legal Data Processing in Europe Bonn, 10-12 October 1989 Systems based on artificial intelligence in the legal
More informationQuickStroke: An Incremental On-line Chinese Handwriting Recognition System
QuickStroke: An Incremental On-line Chinese Handwriting Recognition System Nada P. Matić John C. Platt Λ Tony Wang y Synaptics, Inc. 2381 Bering Drive San Jose, CA 95131, USA Abstract This paper presents
More informationIssues in the Mining of Heart Failure Datasets
International Journal of Automation and Computing 11(2), April 2014, 162-179 DOI: 10.1007/s11633-014-0778-5 Issues in the Mining of Heart Failure Datasets Nongnuch Poolsawad 1 Lisa Moore 1 Chandrasekhar
More informationLecture 1: Machine Learning Basics
1/69 Lecture 1: Machine Learning Basics Ali Harakeh University of Waterloo WAVE Lab ali.harakeh@uwaterloo.ca May 1, 2017 2/69 Overview 1 Learning Algorithms 2 Capacity, Overfitting, and Underfitting 3
More informationAnalysis of Hybrid Soft and Hard Computing Techniques for Forex Monitoring Systems
Analysis of Hybrid Soft and Hard Computing Techniques for Forex Monitoring Systems Ajith Abraham School of Business Systems, Monash University, Clayton, Victoria 3800, Australia. Email: ajith.abraham@ieee.org
More informationUnsupervised Learning of Word Semantic Embedding using the Deep Structured Semantic Model
Unsupervised Learning of Word Semantic Embedding using the Deep Structured Semantic Model Xinying Song, Xiaodong He, Jianfeng Gao, Li Deng Microsoft Research, One Microsoft Way, Redmond, WA 98052, U.S.A.
More informationAnalysis of Speech Recognition Models for Real Time Captioning and Post Lecture Transcription
Analysis of Speech Recognition Models for Real Time Captioning and Post Lecture Transcription Wilny Wilson.P M.Tech Computer Science Student Thejus Engineering College Thrissur, India. Sindhu.S Computer
More information(Sub)Gradient Descent
(Sub)Gradient Descent CMSC 422 MARINE CARPUAT marine@cs.umd.edu Figures credit: Piyush Rai Logistics Midterm is on Thursday 3/24 during class time closed book/internet/etc, one page of notes. will include
More informationTime series prediction
Chapter 13 Time series prediction Amaury Lendasse, Timo Honkela, Federico Pouzols, Antti Sorjamaa, Yoan Miche, Qi Yu, Eric Severin, Mark van Heeswijk, Erkki Oja, Francesco Corona, Elia Liitiäinen, Zhanxing
More informationSoftprop: Softmax Neural Network Backpropagation Learning
Softprop: Softmax Neural Networ Bacpropagation Learning Michael Rimer Computer Science Department Brigham Young University Provo, UT 84602, USA E-mail: mrimer@axon.cs.byu.edu Tony Martinez Computer Science
More informationEvolution of Symbolisation in Chimpanzees and Neural Nets
Evolution of Symbolisation in Chimpanzees and Neural Nets Angelo Cangelosi Centre for Neural and Adaptive Systems University of Plymouth (UK) a.cangelosi@plymouth.ac.uk Introduction Animal communication
More informationEvolutive Neural Net Fuzzy Filtering: Basic Description
Journal of Intelligent Learning Systems and Applications, 2010, 2: 12-18 doi:10.4236/jilsa.2010.21002 Published Online February 2010 (http://www.scirp.org/journal/jilsa) Evolutive Neural Net Fuzzy Filtering:
More informationNeuroscience I. BIOS/PHIL/PSCH 484 MWF 1:00-1:50 Lecture Center F6. Fall credit hours
INSTRUCTOR INFORMATION Dr. John Leonard (course coordinator) Neuroscience I BIOS/PHIL/PSCH 484 MWF 1:00-1:50 Lecture Center F6 Fall 2016 3 credit hours leonard@uic.edu Biological Sciences 3055 SEL 312-996-4261
More informationAxiom 2013 Team Description Paper
Axiom 2013 Team Description Paper Mohammad Ghazanfari, S Omid Shirkhorshidi, Farbod Samsamipour, Hossein Rahmatizadeh Zagheli, Mohammad Mahdavi, Payam Mohajeri, S Abbas Alamolhoda Robotics Scientific Association
More informationFramewise Phoneme Classification with Bidirectional LSTM and Other Neural Network Architectures
Framewise Phoneme Classification with Bidirectional LSTM and Other Neural Network Architectures Alex Graves and Jürgen Schmidhuber IDSIA, Galleria 2, 6928 Manno-Lugano, Switzerland TU Munich, Boltzmannstr.
More informationTest Effort Estimation Using Neural Network
J. Software Engineering & Applications, 2010, 3: 331-340 doi:10.4236/jsea.2010.34038 Published Online April 2010 (http://www.scirp.org/journal/jsea) 331 Chintala Abhishek*, Veginati Pavan Kumar, Harish
More informationKnowledge-Based - Systems
Knowledge-Based - Systems ; Rajendra Arvind Akerkar Chairman, Technomathematics Research Foundation and Senior Researcher, Western Norway Research institute Priti Srinivas Sajja Sardar Patel University
More informationOPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS
OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS Václav Kocian, Eva Volná, Michal Janošek, Martin Kotyrba University of Ostrava Department of Informatics and Computers Dvořákova 7,
More informationLanguage Acquisition Fall 2010/Winter Lexical Categories. Afra Alishahi, Heiner Drenhaus
Language Acquisition Fall 2010/Winter 2011 Lexical Categories Afra Alishahi, Heiner Drenhaus Computational Linguistics and Phonetics Saarland University Children s Sensitivity to Lexical Categories Look,
More informationCSL465/603 - Machine Learning
CSL465/603 - Machine Learning Fall 2016 Narayanan C Krishnan ckn@iitrpr.ac.in Introduction CSL465/603 - Machine Learning 1 Administrative Trivia Course Structure 3-0-2 Lecture Timings Monday 9.55-10.45am
More informationA New Perspective on Combining GMM and DNN Frameworks for Speaker Adaptation
A New Perspective on Combining GMM and DNN Frameworks for Speaker Adaptation SLSP-2016 October 11-12 Natalia Tomashenko 1,2,3 natalia.tomashenko@univ-lemans.fr Yuri Khokhlov 3 khokhlov@speechpro.com Yannick
More informationSystem Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks
System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks 1 Tzu-Hsuan Yang, 2 Tzu-Hsuan Tseng, and 3 Chia-Ping Chen Department of Computer Science and Engineering
More informationDeep search. Enhancing a search bar using machine learning. Ilgün Ilgün & Cedric Reichenbach
#BaselOne7 Deep search Enhancing a search bar using machine learning Ilgün Ilgün & Cedric Reichenbach We are not researchers Outline I. Periscope: A search tool II. Goals III. Deep learning IV. Applying
More informationClassification Using ANN: A Review
International Journal of Computational Intelligence Research ISSN 0973-1873 Volume 13, Number 7 (2017), pp. 1811-1820 Research India Publications http://www.ripublication.com Classification Using ANN:
More informationINPE São José dos Campos
INPE-5479 PRE/1778 MONLINEAR ASPECTS OF DATA INTEGRATION FOR LAND COVER CLASSIFICATION IN A NEDRAL NETWORK ENVIRONNENT Maria Suelena S. Barros Valter Rodrigues INPE São José dos Campos 1993 SECRETARIA
More informationHuman Emotion Recognition From Speech
RESEARCH ARTICLE OPEN ACCESS Human Emotion Recognition From Speech Miss. Aparna P. Wanare*, Prof. Shankar N. Dandare *(Department of Electronics & Telecommunication Engineering, Sant Gadge Baba Amravati
More informationNeuro-Symbolic Approaches for Knowledge Representation in Expert Systems
Published in the International Journal of Hybrid Intelligent Systems 1(3-4) (2004) 111-126 Neuro-Symbolic Approaches for Knowledge Representation in Expert Systems Ioannis Hatzilygeroudis and Jim Prentzas
More informationAn empirical study of learning speed in backpropagation
Carnegie Mellon University Research Showcase @ CMU Computer Science Department School of Computer Science 1988 An empirical study of learning speed in backpropagation networks Scott E. Fahlman Carnegie
More informationSpeaker Identification by Comparison of Smart Methods. Abstract
Journal of mathematics and computer science 10 (2014), 61-71 Speaker Identification by Comparison of Smart Methods Ali Mahdavi Meimand Amin Asadi Majid Mohamadi Department of Electrical Department of Computer
More informationI-COMPETERE: Using Applied Intelligence in search of competency gaps in software project managers.
Information Systems Frontiers manuscript No. (will be inserted by the editor) I-COMPETERE: Using Applied Intelligence in search of competency gaps in software project managers. Ricardo Colomo-Palacios
More informationAutoregressive product of multi-frame predictions can improve the accuracy of hybrid models
Autoregressive product of multi-frame predictions can improve the accuracy of hybrid models Navdeep Jaitly 1, Vincent Vanhoucke 2, Geoffrey Hinton 1,2 1 University of Toronto 2 Google Inc. ndjaitly@cs.toronto.edu,
More informationLecture 1: Basic Concepts of Machine Learning
Lecture 1: Basic Concepts of Machine Learning Cognitive Systems - Machine Learning Ute Schmid (lecture) Johannes Rabold (practice) Based on slides prepared March 2005 by Maximilian Röglinger, updated 2010
More informationSeminar - Organic Computing
Seminar - Organic Computing Self-Organisation of OC-Systems Markus Franke 25.01.2006 Typeset by FoilTEX Timetable 1. Overview 2. Characteristics of SO-Systems 3. Concern with Nature 4. Design-Concepts
More informationKnowledge Transfer in Deep Convolutional Neural Nets
Knowledge Transfer in Deep Convolutional Neural Nets Steven Gutstein, Olac Fuentes and Eric Freudenthal Computer Science Department University of Texas at El Paso El Paso, Texas, 79968, U.S.A. Abstract
More informationInternational Journal of Advanced Networking Applications (IJANA) ISSN No. :
International Journal of Advanced Networking Applications (IJANA) ISSN No. : 0975-0290 34 A Review on Dysarthric Speech Recognition Megha Rughani Department of Electronics and Communication, Marwadi Educational
More informationOn the Formation of Phoneme Categories in DNN Acoustic Models
On the Formation of Phoneme Categories in DNN Acoustic Models Tasha Nagamine Department of Electrical Engineering, Columbia University T. Nagamine Motivation Large performance gap between humans and state-
More informationForget catastrophic forgetting: AI that learns after deployment
Forget catastrophic forgetting: AI that learns after deployment Anatoly Gorshechnikov CTO, Neurala 1 Neurala at a glance Programming neural networks on GPUs since circa 2 B.C. Founded in 2006 expecting
More informationFUZZY EXPERT. Dr. Kasim M. Al-Aubidy. Philadelphia University. Computer Eng. Dept February 2002 University of Damascus-Syria
FUZZY EXPERT SYSTEMS 16-18 18 February 2002 University of Damascus-Syria Dr. Kasim M. Al-Aubidy Computer Eng. Dept. Philadelphia University What is Expert Systems? ES are computer programs that emulate
More informationSyntactic systematicity in sentence processing with a recurrent self-organizing network
Syntactic systematicity in sentence processing with a recurrent self-organizing network Igor Farkaš,1 Department of Applied Informatics, Comenius University Mlynská dolina, 842 48 Bratislava, Slovak Republic
More informationCalibration of Confidence Measures in Speech Recognition
Submitted to IEEE Trans on Audio, Speech, and Language, July 2010 1 Calibration of Confidence Measures in Speech Recognition Dong Yu, Senior Member, IEEE, Jinyu Li, Member, IEEE, Li Deng, Fellow, IEEE
More informationCommunication and Cybernetics 17
Communication and Cybernetics 17 Editors: K. S. Fu W. D. Keidel W. J. M. Levelt H. Wolter Communication and Cybernetics Editors: K.S.Fu, W.D.Keidel, W.1.M.Levelt, H.Wolter Vol. Vol. 2 Vol. 3 Vol. 4 Vol.
More informationMASTER OF SCIENCE (M.S.) MAJOR IN COMPUTER SCIENCE
Master of Science (M.S.) Major in Computer Science 1 MASTER OF SCIENCE (M.S.) MAJOR IN COMPUTER SCIENCE Major Program The programs in computer science are designed to prepare students for doctoral research,
More informationAttributed Social Network Embedding
JOURNAL OF LATEX CLASS FILES, VOL. 14, NO. 8, MAY 2017 1 Attributed Social Network Embedding arxiv:1705.04969v1 [cs.si] 14 May 2017 Lizi Liao, Xiangnan He, Hanwang Zhang, and Tat-Seng Chua Abstract Embedding
More informationModeling function word errors in DNN-HMM based LVCSR systems
Modeling function word errors in DNN-HMM based LVCSR systems Melvin Jose Johnson Premkumar, Ankur Bapna and Sree Avinash Parchuri Department of Computer Science Department of Electrical Engineering Stanford
More informationBreaking the Habit of Being Yourself Workshop for Quantum University
Breaking the Habit of Being Yourself Workshop for Quantum University 2 Copyright Dr Joe Dispenza. June 2013. All rights reserved. 3 Copyright Dr Joe Dispenza. June 2013. All rights reserved. 4 Copyright
More informationLecture 10: Reinforcement Learning
Lecture 1: Reinforcement Learning Cognitive Systems II - Machine Learning SS 25 Part III: Learning Programs and Strategies Q Learning, Dynamic Programming Lecture 1: Reinforcement Learning p. Motivation
More informationCS Machine Learning
CS 478 - Machine Learning Projects Data Representation Basic testing and evaluation schemes CS 478 Data and Testing 1 Programming Issues l Program in any platform you want l Realize that you will be doing
More informationarxiv: v1 [cs.cv] 10 May 2017
Inferring and Executing Programs for Visual Reasoning Justin Johnson 1 Bharath Hariharan 2 Laurens van der Maaten 2 Judy Hoffman 1 Li Fei-Fei 1 C. Lawrence Zitnick 2 Ross Girshick 2 1 Stanford University
More informationSpinal Cord. Student Pages. Classroom Ac tivities
Classroom Ac tivities Spinal Cord Student Pages Produced by Regenerative Medicine Partnership in Education Duquesne University Director john A. Pollock (pollock@duq.edu) The spinal column protects the
More informationAccelerated Learning Course Outline
Accelerated Learning Course Outline Course Description The purpose of this course is to make the advances in the field of brain research more accessible to educators. The techniques and strategies of Accelerated
More informationModeling function word errors in DNN-HMM based LVCSR systems
Modeling function word errors in DNN-HMM based LVCSR systems Melvin Jose Johnson Premkumar, Ankur Bapna and Sree Avinash Parchuri Department of Computer Science Department of Electrical Engineering Stanford
More informationLearning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models
Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models Stephan Gouws and GJ van Rooyen MIH Medialab, Stellenbosch University SOUTH AFRICA {stephan,gvrooyen}@ml.sun.ac.za
More informationAccelerated Learning Online. Course Outline
Accelerated Learning Online Course Outline Course Description The purpose of this course is to make the advances in the field of brain research more accessible to educators. The techniques and strategies
More informationHIERARCHICAL DEEP LEARNING ARCHITECTURE FOR 10K OBJECTS CLASSIFICATION
HIERARCHICAL DEEP LEARNING ARCHITECTURE FOR 10K OBJECTS CLASSIFICATION Atul Laxman Katole 1, Krishna Prasad Yellapragada 1, Amish Kumar Bedi 1, Sehaj Singh Kalra 1 and Mynepalli Siva Chaitanya 1 1 Samsung
More informationSecond Exam: Natural Language Parsing with Neural Networks
Second Exam: Natural Language Parsing with Neural Networks James Cross May 21, 2015 Abstract With the advent of deep learning, there has been a recent resurgence of interest in the use of artificial neural
More informationLecture 2: Quantifiers and Approximation
Lecture 2: Quantifiers and Approximation Case study: Most vs More than half Jakub Szymanik Outline Number Sense Approximate Number Sense Approximating most Superlative Meaning of most What About Counting?
More informationPhonetic- and Speaker-Discriminant Features for Speaker Recognition. Research Project
Phonetic- and Speaker-Discriminant Features for Speaker Recognition by Lara Stoll Research Project Submitted to the Department of Electrical Engineering and Computer Sciences, University of California
More informationarxiv: v1 [cs.lg] 15 Jun 2015
Dual Memory Architectures for Fast Deep Learning of Stream Data via an Online-Incremental-Transfer Strategy arxiv:1506.04477v1 [cs.lg] 15 Jun 2015 Sang-Woo Lee Min-Oh Heo School of Computer Science and
More informationThe Method of Immersion the Problem of Comparing Technical Objects in an Expert Shell in the Class of Artificial Intelligence Algorithms
IOP Conference Series: Materials Science and Engineering PAPER OPEN ACCESS The Method of Immersion the Problem of Comparing Technical Objects in an Expert Shell in the Class of Artificial Intelligence
More information1 NETWORKS VERSUS SYMBOL SYSTEMS: TWO APPROACHES TO MODELING COGNITION
NETWORKS VERSUS SYMBOL SYSTEMS 1 1 NETWORKS VERSUS SYMBOL SYSTEMS: TWO APPROACHES TO MODELING COGNITION 1.1 A Revolution in the Making? The rise of cognitivism in psychology, which, by the 1970s, had successfully
More informationA Case Study: News Classification Based on Term Frequency
A Case Study: News Classification Based on Term Frequency Petr Kroha Faculty of Computer Science University of Technology 09107 Chemnitz Germany kroha@informatik.tu-chemnitz.de Ricardo Baeza-Yates Center
More informationNetworks in Cognitive Science
1 Networks in Cognitive Science Andrea Baronchelli 1,*, Ramon Ferrer-i-Cancho 2, Romualdo Pastor-Satorras 3, Nick Chater 4 and Morten H. Christiansen 5,6 1 Laboratory for the Modeling of Biological and
More informationSoftware Maintenance
1 What is Software Maintenance? Software Maintenance is a very broad activity that includes error corrections, enhancements of capabilities, deletion of obsolete capabilities, and optimization. 2 Categories
More informationMachine Learning and Data Mining. Ensembles of Learners. Prof. Alexander Ihler
Machine Learning and Data Mining Ensembles of Learners Prof. Alexander Ihler Ensemble methods Why learn one classifier when you can learn many? Ensemble: combine many predictors (Weighted) combina
More informationIAT 888: Metacreation Machines endowed with creative behavior. Philippe Pasquier Office 565 (floor 14)
IAT 888: Metacreation Machines endowed with creative behavior Philippe Pasquier Office 565 (floor 14) pasquier@sfu.ca Outline of today's lecture A little bit about me A little bit about you What will that
More informationProbabilistic Latent Semantic Analysis
Probabilistic Latent Semantic Analysis Thomas Hofmann Presentation by Ioannis Pavlopoulos & Andreas Damianou for the course of Data Mining & Exploration 1 Outline Latent Semantic Analysis o Need o Overview
More informationPredicting Early Students with High Risk to Drop Out of University using a Neural Network-Based Approach
Predicting Early Students with High Risk to Drop Out of University using a Neural Network-Based Approach Miguel Gil, Norma Reyes, María Juárez, Emmanuel Espitia, Julio Mosqueda and Myriam Soria Information
More informationAssignment 1: Predicting Amazon Review Ratings
Assignment 1: Predicting Amazon Review Ratings 1 Dataset Analysis Richard Park r2park@acsmail.ucsd.edu February 23, 2015 The dataset selected for this assignment comes from the set of Amazon reviews for
More informationUniversity of Groningen. Systemen, planning, netwerken Bosman, Aart
University of Groningen Systemen, planning, netwerken Bosman, Aart IMPORTANT NOTE: You are advised to consult the publisher's version (publisher's PDF) if you wish to cite from it. Please check the document
More informationBI408-01: Cellular and Molecular Neurobiology
BI408-01: Cellular and Molecular Neurobiology Spring 2013 Instructor: Jennifer R. Kowalski, Ph.D. Office: Gallahue Hall 271 Phone: 940-8879 Office Hours: 10:00-11:30 a.m. Mon. and Wed. E-mail: jrkowals@butler.edu
More informationA Comparison of Annealing Techniques for Academic Course Scheduling
A Comparison of Annealing Techniques for Academic Course Scheduling M. A. Saleh Elmohamed 1, Paul Coddington 2, and Geoffrey Fox 1 1 Northeast Parallel Architectures Center Syracuse University, Syracuse,
More informationHow People Learn Physics
How People Learn Physics Edward F. (Joe) Redish Dept. Of Physics University Of Maryland AAPM, Houston TX, Work supported in part by NSF grants DUE #04-4-0113 and #05-2-4987 Teaching complex subjects 2
More informationLaboratorio di Intelligenza Artificiale e Robotica
Laboratorio di Intelligenza Artificiale e Robotica A.A. 2008-2009 Outline 2 Machine Learning Unsupervised Learning Supervised Learning Reinforcement Learning Genetic Algorithms Genetics-Based Machine Learning
More informationNeural Representation and Neural Computation. Philosophical Perspectives, Vol. 4, Action Theory and Philosophy of Mind (1990),
Neural Representation and Neural Computation Patricia Smith Churchland; Terrence J. Sejnowski Philosophical Perspectives, Vol. 4, Action Theory and Philosophy of Mind (1990), 343-382. Stable URL: http://links.jstor.org/sici?sici=1520-8583%28
More informationADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY DOWNLOAD EBOOK : ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY PDF
Read Online and Download Ebook ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY DOWNLOAD EBOOK : ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY PDF Click link bellow and free register to download
More informationDIRECT ADAPTATION OF HYBRID DNN/HMM MODEL FOR FAST SPEAKER ADAPTATION IN LVCSR BASED ON SPEAKER CODE
2014 IEEE International Conference on Acoustic, Speech and Signal Processing (ICASSP) DIRECT ADAPTATION OF HYBRID DNN/HMM MODEL FOR FAST SPEAKER ADAPTATION IN LVCSR BASED ON SPEAKER CODE Shaofei Xue 1
More informationMachine Learning from Garden Path Sentences: The Application of Computational Linguistics
Machine Learning from Garden Path Sentences: The Application of Computational Linguistics http://dx.doi.org/10.3991/ijet.v9i6.4109 J.L. Du 1, P.F. Yu 1 and M.L. Li 2 1 Guangdong University of Foreign Studies,
More informationTD(λ) and Q-Learning Based Ludo Players
TD(λ) and Q-Learning Based Ludo Players Majed Alhajry, Faisal Alvi, Member, IEEE and Moataz Ahmed Abstract Reinforcement learning is a popular machine learning technique whose inherent self-learning ability
More informationBUSINESS INTELLIGENCE FROM WEB USAGE MINING
BUSINESS INTELLIGENCE FROM WEB USAGE MINING Ajith Abraham Department of Computer Science, Oklahoma State University, 700 N Greenwood Avenue, Tulsa,Oklahoma 74106-0700, USA, ajith.abraham@ieee.org Abstract.
More informationGenerative models and adversarial training
Day 4 Lecture 1 Generative models and adversarial training Kevin McGuinness kevin.mcguinness@dcu.ie Research Fellow Insight Centre for Data Analytics Dublin City University What is a generative model?
More informationWHEN THERE IS A mismatch between the acoustic
808 IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 14, NO. 3, MAY 2006 Optimization of Temporal Filters for Constructing Robust Features in Speech Recognition Jeih-Weih Hung, Member,
More informationLearning Methods in Multilingual Speech Recognition
Learning Methods in Multilingual Speech Recognition Hui Lin Department of Electrical Engineering University of Washington Seattle, WA 98125 linhui@u.washington.edu Li Deng, Jasha Droppo, Dong Yu, and Alex
More informationPurdue Data Summit Communication of Big Data Analytics. New SAT Predictive Validity Case Study
Purdue Data Summit 2017 Communication of Big Data Analytics New SAT Predictive Validity Case Study Paul M. Johnson, Ed.D. Associate Vice President for Enrollment Management, Research & Enrollment Information
More informationBUILDING CONTEXT-DEPENDENT DNN ACOUSTIC MODELS USING KULLBACK-LEIBLER DIVERGENCE-BASED STATE TYING
BUILDING CONTEXT-DEPENDENT DNN ACOUSTIC MODELS USING KULLBACK-LEIBLER DIVERGENCE-BASED STATE TYING Gábor Gosztolya 1, Tamás Grósz 1, László Tóth 1, David Imseng 2 1 MTA-SZTE Research Group on Artificial
More informationAutomating the E-learning Personalization
Automating the E-learning Personalization Fathi Essalmi 1, Leila Jemni Ben Ayed 1, Mohamed Jemni 1, Kinshuk 2, and Sabine Graf 2 1 The Research Laboratory of Technologies of Information and Communication
More informationA study of speaker adaptation for DNN-based speech synthesis
A study of speaker adaptation for DNN-based speech synthesis Zhizheng Wu, Pawel Swietojanski, Christophe Veaux, Steve Renals, Simon King The Centre for Speech Technology Research (CSTR) University of Edinburgh,
More informationHenry Tirri* Petri Myllymgki
From: AAAI Technical Report SS-93-04. Compilation copyright 1993, AAAI (www.aaai.org). All rights reserved. Bayesian Case-Based Reasoning with Neural Networks Petri Myllymgki Henry Tirri* email: University
More informationLearning to Schedule Straight-Line Code
Learning to Schedule Straight-Line Code Eliot Moss, Paul Utgoff, John Cavazos Doina Precup, Darko Stefanović Dept. of Comp. Sci., Univ. of Mass. Amherst, MA 01003 Carla Brodley, David Scheeff Sch. of Elec.
More informationAn OO Framework for building Intelligence and Learning properties in Software Agents
An OO Framework for building Intelligence and Learning properties in Software Agents José A. R. P. Sardinha, Ruy L. Milidiú, Carlos J. P. Lucena, Patrick Paranhos Abstract Software agents are defined as
More informationAn Online Handwriting Recognition System For Turkish
An Online Handwriting Recognition System For Turkish Esra Vural, Hakan Erdogan, Kemal Oflazer, Berrin Yanikoglu Sabanci University, Tuzla, Istanbul, Turkey 34956 ABSTRACT Despite recent developments in
More informationAUTOMATED FABRIC DEFECT INSPECTION: A SURVEY OF CLASSIFIERS
AUTOMATED FABRIC DEFECT INSPECTION: A SURVEY OF CLASSIFIERS Md. Tarek Habib 1, Rahat Hossain Faisal 2, M. Rokonuzzaman 3, Farruk Ahmed 4 1 Department of Computer Science and Engineering, Prime University,
More information