Volgenau School of Engineering. Final Report of Project ECE

Size: px
Start display at page:

Download "Volgenau School of Engineering. Final Report of Project ECE"

Transcription

1 Volgenau School of Engineering Final Report of Project ECE Title: Evaluation of Learning Algorithms on the Data of Self-Organizing Network to Select a Model for Predicting of the Next Call Blocking Probability Authors: HoseinMohammadiMakrani G# Mohammed Rahbini G# Professor: Dr. Monson H. Hayes Fall 2015

2 Table of content 1- Introduction Main Idea Implementation Data Set Bayesian Neural Network Kernel Regression Ensemble Learning Conclusion Reference... 31

3 1- Introduction The increasing availability of large amounts of historical data and the need of performing accurate forecasting of future behavior in several scientific and applied domains demands the efficient techniques able to infer from observations (like time series) the stochastic dependency between past and future [1]. A time series is a set of observations (x), each one being recorded at a specified time (t) [2]. Examples of time series are [3]: Meteorology: weather variables, like temperature, pressure, wind. Economy and finance: economic factors (GNP), financial indexes, exchange rate, spread. Marketing: activity of business, sales. Industry: electric load, power consumption, voltage, sensors. Biomedicine: physiological signals (EEG), heart-rate, patient temperature. Web: clicks, logs. Genomics: time series of gene expression during cell cycle. Time series are studied for several purposes such as the forecasting of the future based on knowledge of the past, the understanding of the phenomenon underlying the measures, or simply a succinct description of the salient features of the series [1]. There are 3 type of series variation. The first one is trend which is a long-term change in the mean level. The second one is seasonal which is periodic and during time the data show same behavior and the last one is irregular. Figure 1 [3] shows the variation. Figure 1

4 2- Main Idea In a Self Organizing Network, it is interesting to predict the probability of call blocking probability at next interval. Based on this prediction, the network can adopt itself with new condition and prepare for likely congestion. Prediction could be done using machine learning methods. For prediction it is essential to study and analyze the patterns which arrive in a certain period. After applying these machinelearning algorithms successfully it is feasible to learn about such time series to predict empty blocks in the spectrum and the probability of congestion in the SON in the future. Therefore, if a model of call blocking through time is available, one can easily predict upcoming delays, blockings, and essentially make the SON much more efficient at allocating its services. In this project we want to evaluate some learning algorithms and make a comparison across them to select the best algorithm to model a time series data, which would allow us to predict the next observation interval from the current time. Actually, our goal is to use machine- learning techniques to learn and thus predict the future congestion in an SON. There are several algorithms available in the technical literature that we will eventually choose from. Our Nominated algorithms are Bayesian neural networks, kernel regression, ensemble learning [4]. These algorithms will apply to a data set (time series), and then the corresponding results will be compared with each other. Eventually, the selected algorithm would be the one with a better prediction and the minimum average prediction error over some period of time (like figure 2 [5]). The criteria for selecting an algorithm are training error, complexity, and validation error.

5 Figure 2 Let s define the problem of this project more clearly. In an SON, we wish to learn properties of the time series formed by the call blocking probabilities measured over a fixed time interval. We then wish to predict the congestion in the network in the short future that follows thereafter. Basically, we have access to the information of a Cell of SON. We need to apply each of the algorithms which we have selected in order to model the call blocking, or congestion rate, probabilities measured over the finite interval. Thus, after we model the data by an algorithm, we then will be able to predict call blocking/congestion probabilities for the next designated interval using that specific algorithm.

6 3- Implementation In this chapter, we describe each technique thoroughly and present the result of prediction for each one. Techniques are Bayesian neural networks, kernel regression, ensemble learning, and Linear discriminate classification. These techniques are implemented by MATLAB at this project. Also we will look at our data set first Data set Our Data Set has call blocking probability of a cell of SON. The Data is for a month (4 weeks). Each interval is half hour, so the total number of interval is equal to This Data set is provided by Reverb Network Company (in a excel file) [6]. Figure 3 shows some part of our data set. Some characteristics of this data set are as below: Average: Standard deviation: Maximum: Minimum: 5e-6 Threshold for considering congestion: 0.2 Figure 3

7 Images 4, 5, and 6 are presentation for one day, one week, and whole data. As it seems, this data has seasonal nature which we described it before. Figure 4 Figure 5

8 Figure 6 We utilize 3 weeks of this data set as a training set and use forth week for testing and prediction. So, all of our reports are based on this division. We must mention here that the testing part (week four) of our data set has 59 congestion spots. Therefore, our desire is that our learning techniques predict these spots as much as possible and still have a minimum prediction error. In our experiment, all errors (training error, prediction error) are reported as MSE (mean squared error) Bayesian Neural network Fortunately, the MATLAB has a data analysis toolbox. One part of this toolbox is Neural Net Time Series (ntstool) which solves nonlinear time series problem using dynamic neural network. Figure 7 shows this toolbox. Figure 7 A Bayesian neural network (BNN) is a neural network designed based on a Bayesian probabilistic formulation. As such BNN s are related to the classical statistics concept of Bayesian parameter estimation, and are also related to the

9 concept of regularization such as in ridge regression. BNN s have enjoyed wide applicability in many areas such as economics/finance and engineering. The idea of BNN is to treat the network parameters or weights as random variables, obeying some a priori distribution. This distribution is designed so as to favor low complexity models, i.e., models producing smooth fits. Once the data are observed, the posterior distribution of the weights is evaluated and the network prediction can be computed. The predictions will then reflect both the smoothness aspect imposed through the prior and the fitness accuracy aspect imposed by the observed data. A closely related concept is the regularization aspect, whereby the following objective function is constructed and minimized this formula: J = a E D + (1 a)e W Where E D is the sum of the square errors in the network outputs, E W is the sum of the squares of the network parameters (i.e., weights), and is the regularization parameter. For the Bayesian approach, the typical choice of the prior is the following normal density that puts more weight onto smaller network parameter values as follow: p(w) = [((1-a)/pi)^(L/2)] * e^ [(1 a)e W ] Where L denotes the number of parameters (weights). The posterior is then given by: p(w D,a ) = p(d w,a) * p(w a) / p(d ) Where D represents the observed data. Assuming normally distributed errors, the probability density of the data given the parameters can be evaluated as: p(d w, a) = [(a/pi)^m/2] * e^( ae D ) Where M is the number of training data points. By substituting the expressions for the densities, we get: p(w D, a) = c exp( J ) Where c is some normalizing constant. The regularization constant is also determined using Bayesian concepts, from: p(a D) = p(d a) * p(a) / p(d) Both two last expressions should be maximized to obtain the optimal weights and alpha (a) parameter, respectively. The term p(d a) in last equation is obtained by a quadratic approximation of J in terms of the weights and then integrating out the weights.

10 We used the MATLAB version trainbr for BNN (applied to a multilayer perceptron architecture). This routine is based on the algorithm proposed by Foresee and Hagan. For prediction, we used Nonlinear Autoregressive model. Figure 8 shows this one. Figure 8 For training part of the network, we set options of test and validation as follow: Figure 9 After setting parameters, then we have to define the configuration of network (number of delays and number of neurons) that figure 10 presents it. Figure 10

11 We made a lot of experiments and tried to play with these two main parameters to find the best configuration for prediction. To find the best configuration we considered following values for delay and neuron: Values for delay: 12, 24, 48, 96 Values for neuron: 5, 10, 15, 20 For the sake of training time, we don t grow our hidden layer s neuron. Actual training time of network for the delay=96 and neuron=20 is about an hour. Table 1 shows all of our experiment and their results. In the table 1, number of correct prediction means that how many congestion spots are predicted correctly by Bayesian neural network. The total number of congestion in testing set is 59. Also, number of wrong prediction indicates that how many times the neural network predicts we will have congestion at next interval but actually there isn t any congestion at next interval. Hidden neuron Table 1 Delay Training Number of correct prediction error error prediction E E E E E E E E E E E E E E E Number of wrong prediction The table 1 reveals that the best configuration for prediction of next call blocking probability and congestion spot (where the probability is bigger than 0.2) is a network with 5 hidden neuron and 96 delay. By this configuration, prediction error

12 (mse) is and also the network is successful to predict 49 congestion spot correctly and only has 7 wrong predictions of congestion. The other information that we can extract from this table is that when we increase the number delays then training error goes down. Figure 11,12, and 13 show this relation for number of neurons 5, 10, and E E E E E E E E E E E Series1 Figure E E E E E E E E E E Series1 Figure 12

13 8.00E E E E E E-04 Series1 2.00E E E Figure 13 Also, there is similar relation between training error and the number of neuron in hidden layer. Figure 14 shows this relation for the number of delay E E E E E E E E E E E Series1 Figure 14 The other relation which can be extracted from this table is that when we increase the complexity of neural network (increase the number of delay), first the prediction error goes to be better but from certain point (where training error is close to zero) this process is inverted and the prediction error is increased. So, we can conclude that by reducing the training error, actually we are decreasing the

14 generalization. Figure 15, 16, and 17 depict this conclusion for the network with the number of neurons 10, 15, and 20. The Y axis is prediction error and X axis represents the step of delays Series Figure Series Figure 16

15 Series Figure 17 Now, we present some picture of training diagram of networks. Following figures show training diagram when neural network is trained by training set. At these pictures, the actual data of training set and the learned data by network are plotted and blow of that is error diagram. Here we just show some nominate pictures. Figure 18, 19, 10, 21, and 22 shows diagrams for network with these configuration (hidden neuron=5, delay=12), (hidden neuron=5, delay=96), (hidden neuron=10, delay=48), (hidden neuron=15, delay=48), and (hidden neuron=20, delay=24). Figure 18

16 Figure 19 Figure 20

17 Figure 21 Figure 22 Here, we present the prediction images of nominated networks. The blue bubbles are the actual data and the red line is prediction of that point. Figure 23, 24, 25, and 26 are prediction result for networks with the configuration as follow: (neuron=5, delay=96), (neuron=10, delay=48), (neuron=15, delay=48), (neuron=20, delay=24).

18 Figure 23 Figure 24

19 Figure 25 Figure 26

20 3-3- Kernel Regression Nadaraya and Watson developed this model. It is commonly called the Nadaraya Watson estimator or the kernel regression estimator. In the machine learning community, the term generalized regression is typically used. The GR model is a nonparametric model where the prediction for a given data point x is given by the average of the target outputs of the training data points in the vicinity of the given point x. The local average is constructed by weighting the points according to their distance from x, using some kernel function. We used the typical Gaussian kernel H(u) = [e^ ((u^2)/2)]/ 2. The parameter H, called the bandwidth, is an important parameter as it determines the smoothness of the fit, since increasing it or decreasing it will control the size of the smoothing region. As it obvious that regression doesn t need to be trained, so we only did regression for week 4 to be able have a fair comparison between Bayesian Neural network and Gaussian Kernel regression. However kernel regression doesn t have the training, it needs some points to predict next interval. In our experiment we regard these initial points as training point and then we calculate training error for them. To show how regression works, we present some pictures which reveal that how bandwidth has effect on the regression and how the next point is predicted by regression. Figures 27, 28, 29, 30, and 31 are related to bandwidth of 0.3, 0.5, 1, 2, and 5. In these pictures, blue bubbles are actual data and the red line is regression. Also, it is visible that the last point of picture (which red line ends) is predicted point. Therefore, here we have 6 actual points and 7 regression points. The last point (7) is prediction.

21 Figure 27 Figure 28

22 Figure 29 Figure 30

23 Figure 31 We can extract some information about the behavior of kernel regression from these images. The first thing is that by increasing the bandwidth (h), the curve goes to be smoother. We can say that when the h is too small, the regression follows the actual data (that we can assume the training error is zero) and the prediction value is equal to the last interval. When we increase the (h), the prediction point starts to change its value. Figure 32 reveal the relation of prediction error and bandwidth Series Figure 32

24 For our implementation, we consider different bandwidth as follow: 0.5, 0.8, 1, 1.2, 1.5, 2, and 5. The table 2 shows the result for kernel regression experiment. Table 2 Bandwidth Training error Prediction error Number of correct Number of prediction wrong prediction E E E E E E E Based on this table, we can say that the best prediction is for regression with bandwidth 0.5. An important result of this regression is that call blocking probabilities are following each other and there is strong relation between next call blocking probability and previous intervals. This kernel regression is successful to do 51 correct predictions of congestion (from 59 congestions in week 4) with the prediction error of Also, the number of wrong predictions of congestion is 8. Figures 32, 33, 34, 35, 36, 37, and 38 depict the regression prediction for week 4 and they represent regression for bandwidth 0.5, 0.8, 1, 1.2, 1.5, 2, 5. It is important to consider that this regression is not just a regression of actual data, of course each point of this regression (red line) is calculated (predicted) based previous interval and finally we draw all predicted data in one figure (red line). The green line is threshold (0.2).

25 Figure 33 Figure 34

26 Figure 35 Figure 36

27 Figure 37 Figure 38

28 Figure Ensemble Learning Ensemble learning is the process by which multiple models, such as classifiers or experts, are strategically generated and combined to solve a particular computational intelligence problem. Ensemble learning is primarily used to improve the (classification, prediction, function approximation, etc.) performance of a model, or reduce the likelihood of an unfortunate selection of a poor one. Other applications of ensemble learning include assigning a confidence to the decision made by the model, selecting optimal (or near optimal) features, data fusion, incremental learning, nonstationary learning and error-correcting. In our project, we used our two previous techniques to build the third technique. So, our first algorithm is Bayesian Neural Network and the second one is Kernel Regression. In our ensemble learning, we first obtain results (prediction points) from each technique and then calculate new prediction point by voting of them. A fair voting technique between two points is averaging. Therefore, we regard the average prediction points of each mentioned algorithm as final prediction point. By this way, we want to increase the prediction of next call blocking probability. The result shows that our expectation is true and the prediction was enhanced

29 significantly. Table 3 shows the value related to ensemble learning and figure 40 depict the prediction line versus actual data. Table 3 Number of correct Number of wrong Prediction error prediction prediction e-4 Figure 40

30 4- Conclusion In this project we evaluate three learning techniques and now we want to make a comparison across them to select the best algorithm to model our time series data, which would allow us to predict the next observation interval from the current time and find congestion spots. Our Nominated algorithms were Bayesian neural networks, kernel regression, ensemble learning. The selected algorithm would be the one with a better prediction and the minimum average prediction error over some period of time. The criterion for selecting this algorithm is prediction error. For the Bayesian Neural Network, the best configuration is 5 neurons with 96 delays. For Kernel regression, the best configuration is bandwidth 0.5. For Ensemble learning we have 2 techniques which we calculate the average of their response as predicted value. Table 4 shows the best result for these 3 techniques. Technique Bayesian neural network Kernel regression Ensemble learning configuration Neurons=5, delays=96 Training error Table 4 Prediction error Number of correct prediction Number of wrong prediction 2.67E H= E NN+KR e

31 5- Reference [1]: Bontempi, Gianluca, Souhaib Ben Taieb, and Yann-Aël Le Borgne. "Machine learning strategies for time series forecasting." In Business Intelligence, pp Springer Berlin Heidelberg, [2]:Brockwell, Peter J., and Richard A. Davis. Time series: theory and methods. Springer Science & Business Media, [3]:Bontempi, Gianluca. "Machine Learning Strategies for Time Series Prediction." [4]: Ahmed, Nesreen K., et al. "An empirical comparison of machine learning models for time series forecasting." Econometric Reviews (2010): [5]: [6]:

Python Machine Learning

Python Machine Learning Python Machine Learning Unlock deeper insights into machine learning with this vital guide to cuttingedge predictive analytics Sebastian Raschka [ PUBLISHING 1 open source I community experience distilled

More information

Lecture 1: Machine Learning Basics

Lecture 1: Machine Learning Basics 1/69 Lecture 1: Machine Learning Basics Ali Harakeh University of Waterloo WAVE Lab ali.harakeh@uwaterloo.ca May 1, 2017 2/69 Overview 1 Learning Algorithms 2 Capacity, Overfitting, and Underfitting 3

More information

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur Module 12 Machine Learning 12.1 Instructional Objective The students should understand the concept of learning systems Students should learn about different aspects of a learning system Students should

More information

Evolutive Neural Net Fuzzy Filtering: Basic Description

Evolutive Neural Net Fuzzy Filtering: Basic Description Journal of Intelligent Learning Systems and Applications, 2010, 2: 12-18 doi:10.4236/jilsa.2010.21002 Published Online February 2010 (http://www.scirp.org/journal/jilsa) Evolutive Neural Net Fuzzy Filtering:

More information

Machine Learning and Data Mining. Ensembles of Learners. Prof. Alexander Ihler

Machine Learning and Data Mining. Ensembles of Learners. Prof. Alexander Ihler Machine Learning and Data Mining Ensembles of Learners Prof. Alexander Ihler Ensemble methods Why learn one classifier when you can learn many? Ensemble: combine many predictors (Weighted) combina

More information

Lecture 10: Reinforcement Learning

Lecture 10: Reinforcement Learning Lecture 1: Reinforcement Learning Cognitive Systems II - Machine Learning SS 25 Part III: Learning Programs and Strategies Q Learning, Dynamic Programming Lecture 1: Reinforcement Learning p. Motivation

More information

Mathematics process categories

Mathematics process categories Mathematics process categories All of the UK curricula define multiple categories of mathematical proficiency that require students to be able to use and apply mathematics, beyond simple recall of facts

More information

Software Maintenance

Software Maintenance 1 What is Software Maintenance? Software Maintenance is a very broad activity that includes error corrections, enhancements of capabilities, deletion of obsolete capabilities, and optimization. 2 Categories

More information

Artificial Neural Networks written examination

Artificial Neural Networks written examination 1 (8) Institutionen för informationsteknologi Olle Gällmo Universitetsadjunkt Adress: Lägerhyddsvägen 2 Box 337 751 05 Uppsala Artificial Neural Networks written examination Monday, May 15, 2006 9 00-14

More information

Test Effort Estimation Using Neural Network

Test Effort Estimation Using Neural Network J. Software Engineering & Applications, 2010, 3: 331-340 doi:10.4236/jsea.2010.34038 Published Online April 2010 (http://www.scirp.org/journal/jsea) 331 Chintala Abhishek*, Veginati Pavan Kumar, Harish

More information

Reinforcement Learning by Comparing Immediate Reward

Reinforcement Learning by Comparing Immediate Reward Reinforcement Learning by Comparing Immediate Reward Punit Pandey DeepshikhaPandey Dr. Shishir Kumar Abstract This paper introduces an approach to Reinforcement Learning Algorithm by comparing their immediate

More information

Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition

Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition Todd Holloway Two Lecture Series for B551 November 20 & 27, 2007 Indiana University Outline Introduction Bias and

More information

Detailed course syllabus

Detailed course syllabus Detailed course syllabus 1. Linear regression model. Ordinary least squares method. This introductory class covers basic definitions of econometrics, econometric model, and economic data. Classification

More information

Time series prediction

Time series prediction Chapter 13 Time series prediction Amaury Lendasse, Timo Honkela, Federico Pouzols, Antti Sorjamaa, Yoan Miche, Qi Yu, Eric Severin, Mark van Heeswijk, Erkki Oja, Francesco Corona, Elia Liitiäinen, Zhanxing

More information

Generative models and adversarial training

Generative models and adversarial training Day 4 Lecture 1 Generative models and adversarial training Kevin McGuinness kevin.mcguinness@dcu.ie Research Fellow Insight Centre for Data Analytics Dublin City University What is a generative model?

More information

Adaptive Learning in Time-Variant Processes With Application to Wind Power Systems

Adaptive Learning in Time-Variant Processes With Application to Wind Power Systems IEEE TRANSACTIONS ON AUTOMATION SCIENCE AND ENGINEERING, VOL 13, NO 2, APRIL 2016 997 Adaptive Learning in Time-Variant Processes With Application to Wind Power Systems Eunshin Byon, Member, IEEE, Youngjun

More information

Genevieve L. Hartman, Ph.D.

Genevieve L. Hartman, Ph.D. Curriculum Development and the Teaching-Learning Process: The Development of Mathematical Thinking for all children Genevieve L. Hartman, Ph.D. Topics for today Part 1: Background and rationale Current

More information

The Good Judgment Project: A large scale test of different methods of combining expert predictions

The Good Judgment Project: A large scale test of different methods of combining expert predictions The Good Judgment Project: A large scale test of different methods of combining expert predictions Lyle Ungar, Barb Mellors, Jon Baron, Phil Tetlock, Jaime Ramos, Sam Swift The University of Pennsylvania

More information

QuickStroke: An Incremental On-line Chinese Handwriting Recognition System

QuickStroke: An Incremental On-line Chinese Handwriting Recognition System QuickStroke: An Incremental On-line Chinese Handwriting Recognition System Nada P. Matić John C. Platt Λ Tony Wang y Synaptics, Inc. 2381 Bering Drive San Jose, CA 95131, USA Abstract This paper presents

More information

OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS

OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS Václav Kocian, Eva Volná, Michal Janošek, Martin Kotyrba University of Ostrava Department of Informatics and Computers Dvořákova 7,

More information

Course Outline. Course Grading. Where to go for help. Academic Integrity. EE-589 Introduction to Neural Networks NN 1 EE

Course Outline. Course Grading. Where to go for help. Academic Integrity. EE-589 Introduction to Neural Networks NN 1 EE EE-589 Introduction to Neural Assistant Prof. Dr. Turgay IBRIKCI Room # 305 (322) 338 6868 / 139 Wensdays 9:00-12:00 Course Outline The course is divided in two parts: theory and practice. 1. Theory covers

More information

The 9 th International Scientific Conference elearning and software for Education Bucharest, April 25-26, / X

The 9 th International Scientific Conference elearning and software for Education Bucharest, April 25-26, / X The 9 th International Scientific Conference elearning and software for Education Bucharest, April 25-26, 2013 10.12753/2066-026X-13-154 DATA MINING SOLUTIONS FOR DETERMINING STUDENT'S PROFILE Adela BÂRA,

More information

School of Innovative Technologies and Engineering

School of Innovative Technologies and Engineering School of Innovative Technologies and Engineering Department of Applied Mathematical Sciences Proficiency Course in MATLAB COURSE DOCUMENT VERSION 1.0 PCMv1.0 July 2012 University of Technology, Mauritius

More information

INPE São José dos Campos

INPE São José dos Campos INPE-5479 PRE/1778 MONLINEAR ASPECTS OF DATA INTEGRATION FOR LAND COVER CLASSIFICATION IN A NEDRAL NETWORK ENVIRONNENT Maria Suelena S. Barros Valter Rodrigues INPE São José dos Campos 1993 SECRETARIA

More information

Probability and Statistics Curriculum Pacing Guide

Probability and Statistics Curriculum Pacing Guide Unit 1 Terms PS.SPMJ.3 PS.SPMJ.5 Plan and conduct a survey to answer a statistical question. Recognize how the plan addresses sampling technique, randomization, measurement of experimental error and methods

More information

System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks

System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks 1 Tzu-Hsuan Yang, 2 Tzu-Hsuan Tseng, and 3 Chia-Ping Chen Department of Computer Science and Engineering

More information

Human Emotion Recognition From Speech

Human Emotion Recognition From Speech RESEARCH ARTICLE OPEN ACCESS Human Emotion Recognition From Speech Miss. Aparna P. Wanare*, Prof. Shankar N. Dandare *(Department of Electronics & Telecommunication Engineering, Sant Gadge Baba Amravati

More information

Introduction to Simulation

Introduction to Simulation Introduction to Simulation Spring 2010 Dr. Louis Luangkesorn University of Pittsburgh January 19, 2010 Dr. Louis Luangkesorn ( University of Pittsburgh ) Introduction to Simulation January 19, 2010 1 /

More information

Axiom 2013 Team Description Paper

Axiom 2013 Team Description Paper Axiom 2013 Team Description Paper Mohammad Ghazanfari, S Omid Shirkhorshidi, Farbod Samsamipour, Hossein Rahmatizadeh Zagheli, Mohammad Mahdavi, Payam Mohajeri, S Abbas Alamolhoda Robotics Scientific Association

More information

Calibration of Confidence Measures in Speech Recognition

Calibration of Confidence Measures in Speech Recognition Submitted to IEEE Trans on Audio, Speech, and Language, July 2010 1 Calibration of Confidence Measures in Speech Recognition Dong Yu, Senior Member, IEEE, Jinyu Li, Member, IEEE, Li Deng, Fellow, IEEE

More information

A Neural Network GUI Tested on Text-To-Phoneme Mapping

A Neural Network GUI Tested on Text-To-Phoneme Mapping A Neural Network GUI Tested on Text-To-Phoneme Mapping MAARTEN TROMPPER Universiteit Utrecht m.f.a.trompper@students.uu.nl Abstract Text-to-phoneme (T2P) mapping is a necessary step in any speech synthesis

More information

Learning Methods for Fuzzy Systems

Learning Methods for Fuzzy Systems Learning Methods for Fuzzy Systems Rudolf Kruse and Andreas Nürnberger Department of Computer Science, University of Magdeburg Universitätsplatz, D-396 Magdeburg, Germany Phone : +49.39.67.876, Fax : +49.39.67.8

More information

Testing A Moving Target: How Do We Test Machine Learning Systems? Peter Varhol Technology Strategy Research, USA

Testing A Moving Target: How Do We Test Machine Learning Systems? Peter Varhol Technology Strategy Research, USA Testing A Moving Target: How Do We Test Machine Learning Systems? Peter Varhol Technology Strategy Research, USA Testing a Moving Target How Do We Test Machine Learning Systems? Peter Varhol, Technology

More information

Lecture 1: Basic Concepts of Machine Learning

Lecture 1: Basic Concepts of Machine Learning Lecture 1: Basic Concepts of Machine Learning Cognitive Systems - Machine Learning Ute Schmid (lecture) Johannes Rabold (practice) Based on slides prepared March 2005 by Maximilian Röglinger, updated 2010

More information

Australian Journal of Basic and Applied Sciences

Australian Journal of Basic and Applied Sciences AENSI Journals Australian Journal of Basic and Applied Sciences ISSN:1991-8178 Journal home page: www.ajbasweb.com Feature Selection Technique Using Principal Component Analysis For Improving Fuzzy C-Mean

More information

Edexcel GCSE. Statistics 1389 Paper 1H. June Mark Scheme. Statistics Edexcel GCSE

Edexcel GCSE. Statistics 1389 Paper 1H. June Mark Scheme. Statistics Edexcel GCSE Edexcel GCSE Statistics 1389 Paper 1H June 2007 Mark Scheme Edexcel GCSE Statistics 1389 NOTES ON MARKING PRINCIPLES 1 Types of mark M marks: method marks A marks: accuracy marks B marks: unconditional

More information

Rule Learning With Negation: Issues Regarding Effectiveness

Rule Learning With Negation: Issues Regarding Effectiveness Rule Learning With Negation: Issues Regarding Effectiveness S. Chua, F. Coenen, G. Malcolm University of Liverpool Department of Computer Science, Ashton Building, Ashton Street, L69 3BX Liverpool, United

More information

A Reinforcement Learning Variant for Control Scheduling

A Reinforcement Learning Variant for Control Scheduling A Reinforcement Learning Variant for Control Scheduling Aloke Guha Honeywell Sensor and System Development Center 3660 Technology Drive Minneapolis MN 55417 Abstract We present an algorithm based on reinforcement

More information

On the Combined Behavior of Autonomous Resource Management Agents

On the Combined Behavior of Autonomous Resource Management Agents On the Combined Behavior of Autonomous Resource Management Agents Siri Fagernes 1 and Alva L. Couch 2 1 Faculty of Engineering Oslo University College Oslo, Norway siri.fagernes@iu.hio.no 2 Computer Science

More information

Lahore University of Management Sciences. FINN 321 Econometrics Fall Semester 2017

Lahore University of Management Sciences. FINN 321 Econometrics Fall Semester 2017 Instructor Syed Zahid Ali Room No. 247 Economics Wing First Floor Office Hours Email szahid@lums.edu.pk Telephone Ext. 8074 Secretary/TA TA Office Hours Course URL (if any) Suraj.lums.edu.pk FINN 321 Econometrics

More information

STA 225: Introductory Statistics (CT)

STA 225: Introductory Statistics (CT) Marshall University College of Science Mathematics Department STA 225: Introductory Statistics (CT) Course catalog description A critical thinking course in applied statistical reasoning covering basic

More information

Physics 270: Experimental Physics

Physics 270: Experimental Physics 2017 edition Lab Manual Physics 270 3 Physics 270: Experimental Physics Lecture: Lab: Instructor: Office: Email: Tuesdays, 2 3:50 PM Thursdays, 2 4:50 PM Dr. Uttam Manna 313C Moulton Hall umanna@ilstu.edu

More information

Task Types. Duration, Work and Units Prepared by

Task Types. Duration, Work and Units Prepared by Task Types Duration, Work and Units Prepared by 1 Introduction Microsoft Project allows tasks with fixed work, fixed duration, or fixed units. Many people ask questions about changes in these values when

More information

Purdue Data Summit Communication of Big Data Analytics. New SAT Predictive Validity Case Study

Purdue Data Summit Communication of Big Data Analytics. New SAT Predictive Validity Case Study Purdue Data Summit 2017 Communication of Big Data Analytics New SAT Predictive Validity Case Study Paul M. Johnson, Ed.D. Associate Vice President for Enrollment Management, Research & Enrollment Information

More information

arxiv: v1 [cs.lg] 15 Jun 2015

arxiv: v1 [cs.lg] 15 Jun 2015 Dual Memory Architectures for Fast Deep Learning of Stream Data via an Online-Incremental-Transfer Strategy arxiv:1506.04477v1 [cs.lg] 15 Jun 2015 Sang-Woo Lee Min-Oh Heo School of Computer Science and

More information

A New Perspective on Combining GMM and DNN Frameworks for Speaker Adaptation

A New Perspective on Combining GMM and DNN Frameworks for Speaker Adaptation A New Perspective on Combining GMM and DNN Frameworks for Speaker Adaptation SLSP-2016 October 11-12 Natalia Tomashenko 1,2,3 natalia.tomashenko@univ-lemans.fr Yuri Khokhlov 3 khokhlov@speechpro.com Yannick

More information

DEVELOPMENT OF AN INTELLIGENT MAINTENANCE SYSTEM FOR ELECTRONIC VALVES

DEVELOPMENT OF AN INTELLIGENT MAINTENANCE SYSTEM FOR ELECTRONIC VALVES DEVELOPMENT OF AN INTELLIGENT MAINTENANCE SYSTEM FOR ELECTRONIC VALVES Luiz Fernando Gonçalves, luizfg@ece.ufrgs.br Marcelo Soares Lubaszewski, luba@ece.ufrgs.br Carlos Eduardo Pereira, cpereira@ece.ufrgs.br

More information

Speech Recognition at ICSI: Broadcast News and beyond

Speech Recognition at ICSI: Broadcast News and beyond Speech Recognition at ICSI: Broadcast News and beyond Dan Ellis International Computer Science Institute, Berkeley CA Outline 1 2 3 The DARPA Broadcast News task Aspects of ICSI

More information

Analysis of Hybrid Soft and Hard Computing Techniques for Forex Monitoring Systems

Analysis of Hybrid Soft and Hard Computing Techniques for Forex Monitoring Systems Analysis of Hybrid Soft and Hard Computing Techniques for Forex Monitoring Systems Ajith Abraham School of Business Systems, Monash University, Clayton, Victoria 3800, Australia. Email: ajith.abraham@ieee.org

More information

Speech Emotion Recognition Using Support Vector Machine

Speech Emotion Recognition Using Support Vector Machine Speech Emotion Recognition Using Support Vector Machine Yixiong Pan, Peipei Shen and Liping Shen Department of Computer Technology Shanghai JiaoTong University, Shanghai, China panyixiong@sjtu.edu.cn,

More information

Speeding Up Reinforcement Learning with Behavior Transfer

Speeding Up Reinforcement Learning with Behavior Transfer Speeding Up Reinforcement Learning with Behavior Transfer Matthew E. Taylor and Peter Stone Department of Computer Sciences The University of Texas at Austin Austin, Texas 78712-1188 {mtaylor, pstone}@cs.utexas.edu

More information

ACTL5103 Stochastic Modelling For Actuaries. Course Outline Semester 2, 2014

ACTL5103 Stochastic Modelling For Actuaries. Course Outline Semester 2, 2014 UNSW Australia Business School School of Risk and Actuarial Studies ACTL5103 Stochastic Modelling For Actuaries Course Outline Semester 2, 2014 Part A: Course-Specific Information Please consult Part B

More information

Visit us at:

Visit us at: White Paper Integrating Six Sigma and Software Testing Process for Removal of Wastage & Optimizing Resource Utilization 24 October 2013 With resources working for extended hours and in a pressurized environment,

More information

An empirical study of learning speed in backpropagation

An empirical study of learning speed in backpropagation Carnegie Mellon University Research Showcase @ CMU Computer Science Department School of Computer Science 1988 An empirical study of learning speed in backpropagation networks Scott E. Fahlman Carnegie

More information

Twitter Sentiment Classification on Sanders Data using Hybrid Approach

Twitter Sentiment Classification on Sanders Data using Hybrid Approach IOSR Journal of Computer Engineering (IOSR-JCE) e-issn: 2278-0661,p-ISSN: 2278-8727, Volume 17, Issue 4, Ver. I (July Aug. 2015), PP 118-123 www.iosrjournals.org Twitter Sentiment Classification on Sanders

More information

MULTIPLE CHOICE. Choose the one alternative that best completes the statement or answers the question.

MULTIPLE CHOICE. Choose the one alternative that best completes the statement or answers the question. Ch 2 Test Remediation Work Name MULTIPLE CHOICE. Choose the one alternative that best completes the statement or answers the question. Provide an appropriate response. 1) High temperatures in a certain

More information

OCR for Arabic using SIFT Descriptors With Online Failure Prediction

OCR for Arabic using SIFT Descriptors With Online Failure Prediction OCR for Arabic using SIFT Descriptors With Online Failure Prediction Andrey Stolyarenko, Nachum Dershowitz The Blavatnik School of Computer Science Tel Aviv University Tel Aviv, Israel Email: stloyare@tau.ac.il,

More information

Seminar - Organic Computing

Seminar - Organic Computing Seminar - Organic Computing Self-Organisation of OC-Systems Markus Franke 25.01.2006 Typeset by FoilTEX Timetable 1. Overview 2. Characteristics of SO-Systems 3. Concern with Nature 4. Design-Concepts

More information

Getting Started with TI-Nspire High School Science

Getting Started with TI-Nspire High School Science Getting Started with TI-Nspire High School Science 2012 Texas Instruments Incorporated Materials for Institute Participant * *This material is for the personal use of T3 instructors in delivering a T3

More information

Learning Methods in Multilingual Speech Recognition

Learning Methods in Multilingual Speech Recognition Learning Methods in Multilingual Speech Recognition Hui Lin Department of Electrical Engineering University of Washington Seattle, WA 98125 linhui@u.washington.edu Li Deng, Jasha Droppo, Dong Yu, and Alex

More information

Word Segmentation of Off-line Handwritten Documents

Word Segmentation of Off-line Handwritten Documents Word Segmentation of Off-line Handwritten Documents Chen Huang and Sargur N. Srihari {chuang5, srihari}@cedar.buffalo.edu Center of Excellence for Document Analysis and Recognition (CEDAR), Department

More information

AGS THE GREAT REVIEW GAME FOR PRE-ALGEBRA (CD) CORRELATED TO CALIFORNIA CONTENT STANDARDS

AGS THE GREAT REVIEW GAME FOR PRE-ALGEBRA (CD) CORRELATED TO CALIFORNIA CONTENT STANDARDS AGS THE GREAT REVIEW GAME FOR PRE-ALGEBRA (CD) CORRELATED TO CALIFORNIA CONTENT STANDARDS 1 CALIFORNIA CONTENT STANDARDS: Chapter 1 ALGEBRA AND WHOLE NUMBERS Algebra and Functions 1.4 Students use algebraic

More information

(Sub)Gradient Descent

(Sub)Gradient Descent (Sub)Gradient Descent CMSC 422 MARINE CARPUAT marine@cs.umd.edu Figures credit: Piyush Rai Logistics Midterm is on Thursday 3/24 during class time closed book/internet/etc, one page of notes. will include

More information

Phonetic- and Speaker-Discriminant Features for Speaker Recognition. Research Project

Phonetic- and Speaker-Discriminant Features for Speaker Recognition. Research Project Phonetic- and Speaker-Discriminant Features for Speaker Recognition by Lara Stoll Research Project Submitted to the Department of Electrical Engineering and Computer Sciences, University of California

More information

Softprop: Softmax Neural Network Backpropagation Learning

Softprop: Softmax Neural Network Backpropagation Learning Softprop: Softmax Neural Networ Bacpropagation Learning Michael Rimer Computer Science Department Brigham Young University Provo, UT 84602, USA E-mail: mrimer@axon.cs.byu.edu Tony Martinez Computer Science

More information

Learning Optimal Dialogue Strategies: A Case Study of a Spoken Dialogue Agent for

Learning Optimal Dialogue Strategies: A Case Study of a Spoken Dialogue Agent for Learning Optimal Dialogue Strategies: A Case Study of a Spoken Dialogue Agent for Email Marilyn A. Walker Jeanne C. Fromer Shrikanth Narayanan walker@research.att.com jeannie@ai.mit.edu shri@research.att.com

More information

Malicious User Suppression for Cooperative Spectrum Sensing in Cognitive Radio Networks using Dixon s Outlier Detection Method

Malicious User Suppression for Cooperative Spectrum Sensing in Cognitive Radio Networks using Dixon s Outlier Detection Method Malicious User Suppression for Cooperative Spectrum Sensing in Cognitive Radio Networks using Dixon s Outlier Detection Method Sanket S. Kalamkar and Adrish Banerjee Department of Electrical Engineering

More information

Bayllocator: A proactive system to predict server utilization and dynamically allocate memory resources using Bayesian networks and ballooning

Bayllocator: A proactive system to predict server utilization and dynamically allocate memory resources using Bayesian networks and ballooning Bayllocator: A proactive system to predict server utilization and dynamically allocate memory resources using Bayesian networks and ballooning Evangelos Tasoulas - University of Oslo Hårek Haugerud - Oslo

More information

Measurement. When Smaller Is Better. Activity:

Measurement. When Smaller Is Better. Activity: Measurement Activity: TEKS: When Smaller Is Better (6.8) Measurement. The student solves application problems involving estimation and measurement of length, area, time, temperature, volume, weight, and

More information

Transfer Learning Action Models by Measuring the Similarity of Different Domains

Transfer Learning Action Models by Measuring the Similarity of Different Domains Transfer Learning Action Models by Measuring the Similarity of Different Domains Hankui Zhuo 1, Qiang Yang 2, and Lei Li 1 1 Software Research Institute, Sun Yat-sen University, Guangzhou, China. zhuohank@gmail.com,lnslilei@mail.sysu.edu.cn

More information

Rule Learning with Negation: Issues Regarding Effectiveness

Rule Learning with Negation: Issues Regarding Effectiveness Rule Learning with Negation: Issues Regarding Effectiveness Stephanie Chua, Frans Coenen, and Grant Malcolm University of Liverpool Department of Computer Science, Ashton Building, Ashton Street, L69 3BX

More information

Knowledge-Based - Systems

Knowledge-Based - Systems Knowledge-Based - Systems ; Rajendra Arvind Akerkar Chairman, Technomathematics Research Foundation and Senior Researcher, Western Norway Research institute Priti Srinivas Sajja Sardar Patel University

More information

AP Calculus AB. Nevada Academic Standards that are assessable at the local level only.

AP Calculus AB. Nevada Academic Standards that are assessable at the local level only. Calculus AB Priority Keys Aligned with Nevada Standards MA I MI L S MA represents a Major content area. Any concept labeled MA is something of central importance to the entire class/curriculum; it is a

More information

Introduction to the Practice of Statistics

Introduction to the Practice of Statistics Chapter 1: Looking at Data Distributions Introduction to the Practice of Statistics Sixth Edition David S. Moore George P. McCabe Bruce A. Craig Statistics is the science of collecting, organizing and

More information

BMBF Project ROBUKOM: Robust Communication Networks

BMBF Project ROBUKOM: Robust Communication Networks BMBF Project ROBUKOM: Robust Communication Networks Arie M.C.A. Koster Christoph Helmberg Andreas Bley Martin Grötschel Thomas Bauschert supported by BMBF grant 03MS616A: ROBUKOM Robust Communication Networks,

More information

Why Did My Detector Do That?!

Why Did My Detector Do That?! Why Did My Detector Do That?! Predicting Keystroke-Dynamics Error Rates Kevin Killourhy and Roy Maxion Dependable Systems Laboratory Computer Science Department Carnegie Mellon University 5000 Forbes Ave,

More information

Exploration. CS : Deep Reinforcement Learning Sergey Levine

Exploration. CS : Deep Reinforcement Learning Sergey Levine Exploration CS 294-112: Deep Reinforcement Learning Sergey Levine Class Notes 1. Homework 4 due on Wednesday 2. Project proposal feedback sent Today s Lecture 1. What is exploration? Why is it a problem?

More information

GCSE Mathematics B (Linear) Mark Scheme for November Component J567/04: Mathematics Paper 4 (Higher) General Certificate of Secondary Education

GCSE Mathematics B (Linear) Mark Scheme for November Component J567/04: Mathematics Paper 4 (Higher) General Certificate of Secondary Education GCSE Mathematics B (Linear) Component J567/04: Mathematics Paper 4 (Higher) General Certificate of Secondary Education Mark Scheme for November 2014 Oxford Cambridge and RSA Examinations OCR (Oxford Cambridge

More information

WHEN THERE IS A mismatch between the acoustic

WHEN THERE IS A mismatch between the acoustic 808 IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 14, NO. 3, MAY 2006 Optimization of Temporal Filters for Constructing Robust Features in Speech Recognition Jeih-Weih Hung, Member,

More information

CSL465/603 - Machine Learning

CSL465/603 - Machine Learning CSL465/603 - Machine Learning Fall 2016 Narayanan C Krishnan ckn@iitrpr.ac.in Introduction CSL465/603 - Machine Learning 1 Administrative Trivia Course Structure 3-0-2 Lecture Timings Monday 9.55-10.45am

More information

Modeling function word errors in DNN-HMM based LVCSR systems

Modeling function word errors in DNN-HMM based LVCSR systems Modeling function word errors in DNN-HMM based LVCSR systems Melvin Jose Johnson Premkumar, Ankur Bapna and Sree Avinash Parchuri Department of Computer Science Department of Electrical Engineering Stanford

More information

A Model to Predict 24-Hour Urinary Creatinine Level Using Repeated Measurements

A Model to Predict 24-Hour Urinary Creatinine Level Using Repeated Measurements Virginia Commonwealth University VCU Scholars Compass Theses and Dissertations Graduate School 2006 A Model to Predict 24-Hour Urinary Creatinine Level Using Repeated Measurements Donna S. Kroos Virginia

More information

CS Machine Learning

CS Machine Learning CS 478 - Machine Learning Projects Data Representation Basic testing and evaluation schemes CS 478 Data and Testing 1 Programming Issues l Program in any platform you want l Realize that you will be doing

More information

Algebra 2- Semester 2 Review

Algebra 2- Semester 2 Review Name Block Date Algebra 2- Semester 2 Review Non-Calculator 5.4 1. Consider the function f x 1 x 2. a) Describe the transformation of the graph of y 1 x. b) Identify the asymptotes. c) What is the domain

More information

Using the Attribute Hierarchy Method to Make Diagnostic Inferences about Examinees Cognitive Skills in Algebra on the SAT

Using the Attribute Hierarchy Method to Make Diagnostic Inferences about Examinees Cognitive Skills in Algebra on the SAT The Journal of Technology, Learning, and Assessment Volume 6, Number 6 February 2008 Using the Attribute Hierarchy Method to Make Diagnostic Inferences about Examinees Cognitive Skills in Algebra on the

More information

IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 17, NO. 3, MARCH

IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 17, NO. 3, MARCH IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 17, NO. 3, MARCH 2009 423 Adaptive Multimodal Fusion by Uncertainty Compensation With Application to Audiovisual Speech Recognition George

More information

Hierarchical Linear Modeling with Maximum Likelihood, Restricted Maximum Likelihood, and Fully Bayesian Estimation

Hierarchical Linear Modeling with Maximum Likelihood, Restricted Maximum Likelihood, and Fully Bayesian Estimation A peer-reviewed electronic journal. Copyright is retained by the first or sole author, who grants right of first publication to Practical Assessment, Research & Evaluation. Permission is granted to distribute

More information

16.1 Lesson: Putting it into practice - isikhnas

16.1 Lesson: Putting it into practice - isikhnas BAB 16 Module: Using QGIS in animal health The purpose of this module is to show how QGIS can be used to assist in animal health scenarios. In order to do this, you will have needed to study, and be familiar

More information

Stacks Teacher notes. Activity description. Suitability. Time. AMP resources. Equipment. Key mathematical language. Key processes

Stacks Teacher notes. Activity description. Suitability. Time. AMP resources. Equipment. Key mathematical language. Key processes Stacks Teacher notes Activity description (Interactive not shown on this sheet.) Pupils start by exploring the patterns generated by moving counters between two stacks according to a fixed rule, doubling

More information

Grade Dropping, Strategic Behavior, and Student Satisficing

Grade Dropping, Strategic Behavior, and Student Satisficing Grade Dropping, Strategic Behavior, and Student Satisficing Lester Hadsell Department of Economics State University of New York, College at Oneonta Oneonta, NY 13820 hadsell@oneonta.edu Raymond MacDermott

More information

COMPUTER-ASSISTED INDEPENDENT STUDY IN MULTIVARIATE CALCULUS

COMPUTER-ASSISTED INDEPENDENT STUDY IN MULTIVARIATE CALCULUS COMPUTER-ASSISTED INDEPENDENT STUDY IN MULTIVARIATE CALCULUS L. Descalço 1, Paula Carvalho 1, J.P. Cruz 1, Paula Oliveira 1, Dina Seabra 2 1 Departamento de Matemática, Universidade de Aveiro (PORTUGAL)

More information

Speech Segmentation Using Probabilistic Phonetic Feature Hierarchy and Support Vector Machines

Speech Segmentation Using Probabilistic Phonetic Feature Hierarchy and Support Vector Machines Speech Segmentation Using Probabilistic Phonetic Feature Hierarchy and Support Vector Machines Amit Juneja and Carol Espy-Wilson Department of Electrical and Computer Engineering University of Maryland,

More information

Multivariate k-nearest Neighbor Regression for Time Series data -

Multivariate k-nearest Neighbor Regression for Time Series data - Multivariate k-nearest Neighbor Regression for Time Series data - a novel Algorithm for Forecasting UK Electricity Demand ISF 2013, Seoul, Korea Fahad H. Al-Qahtani Dr. Sven F. Crone Management Science,

More information

Design Of An Automatic Speaker Recognition System Using MFCC, Vector Quantization And LBG Algorithm

Design Of An Automatic Speaker Recognition System Using MFCC, Vector Quantization And LBG Algorithm Design Of An Automatic Speaker Recognition System Using MFCC, Vector Quantization And LBG Algorithm Prof. Ch.Srinivasa Kumar Prof. and Head of department. Electronics and communication Nalanda Institute

More information

Applications of data mining algorithms to analysis of medical data

Applications of data mining algorithms to analysis of medical data Master Thesis Software Engineering Thesis no: MSE-2007:20 August 2007 Applications of data mining algorithms to analysis of medical data Dariusz Matyja School of Engineering Blekinge Institute of Technology

More information

Paper Reference. Edexcel GCSE Mathematics (Linear) 1380 Paper 1 (Non-Calculator) Foundation Tier. Monday 6 June 2011 Afternoon Time: 1 hour 30 minutes

Paper Reference. Edexcel GCSE Mathematics (Linear) 1380 Paper 1 (Non-Calculator) Foundation Tier. Monday 6 June 2011 Afternoon Time: 1 hour 30 minutes Centre No. Candidate No. Paper Reference 1 3 8 0 1 F Paper Reference(s) 1380/1F Edexcel GCSE Mathematics (Linear) 1380 Paper 1 (Non-Calculator) Foundation Tier Monday 6 June 2011 Afternoon Time: 1 hour

More information

SARDNET: A Self-Organizing Feature Map for Sequences

SARDNET: A Self-Organizing Feature Map for Sequences SARDNET: A Self-Organizing Feature Map for Sequences Daniel L. James and Risto Miikkulainen Department of Computer Sciences The University of Texas at Austin Austin, TX 78712 dljames,risto~cs.utexas.edu

More information

Assignment 1: Predicting Amazon Review Ratings

Assignment 1: Predicting Amazon Review Ratings Assignment 1: Predicting Amazon Review Ratings 1 Dataset Analysis Richard Park r2park@acsmail.ucsd.edu February 23, 2015 The dataset selected for this assignment comes from the set of Amazon reviews for

More information

Spinners at the School Carnival (Unequal Sections)

Spinners at the School Carnival (Unequal Sections) Spinners at the School Carnival (Unequal Sections) Maryann E. Huey Drake University maryann.huey@drake.edu Published: February 2012 Overview of the Lesson Students are asked to predict the outcomes of

More information