to solve real-world problems.

Similar documents
Python Machine Learning

(Sub)Gradient Descent

CS Machine Learning

Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition

Business Analytics and Information Tech COURSE NUMBER: 33:136:494 COURSE TITLE: Data Mining and Business Intelligence

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

Lecture 1: Machine Learning Basics

CSL465/603 - Machine Learning

Lecture 1: Basic Concepts of Machine Learning

Assignment 1: Predicting Amazon Review Ratings

Machine Learning and Data Mining. Ensembles of Learners. Prof. Alexander Ihler

Learning From the Past with Experiment Databases

ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY DOWNLOAD EBOOK : ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY PDF

Large-Scale Web Page Classification. Sathi T Marath. Submitted in partial fulfilment of the requirements. for the degree of Doctor of Philosophy

Welcome to. ECML/PKDD 2004 Community meeting

Artificial Neural Networks written examination

Twitter Sentiment Classification on Sanders Data using Hybrid Approach

Human Emotion Recognition From Speech

arxiv: v2 [cs.cv] 30 Mar 2017

Time series prediction

A survey of multi-view machine learning

Predicting Student Attrition in MOOCs using Sentiment Analysis and Neural Networks

Issues in the Mining of Heart Failure Datasets

Axiom 2013 Team Description Paper

Generative models and adversarial training

CLASSIFICATION OF TEXT DOCUMENTS USING INTEGER REPRESENTATION AND REGRESSION: AN INTEGRATED APPROACH

The 9 th International Scientific Conference elearning and software for Education Bucharest, April 25-26, / X

Active Learning. Yingyu Liang Computer Sciences 760 Fall

Semi-supervised methods of text processing, and an application to medical concept extraction. Yacine Jernite Text-as-Data series September 17.

Modeling function word errors in DNN-HMM based LVCSR systems

Speech Emotion Recognition Using Support Vector Machine

Learning Methods in Multilingual Speech Recognition

Comparison of EM and Two-Step Cluster Method for Mixed Data: An Application

Switchboard Language Model Improvement with Conversational Data from Gigaword

Analysis of Emotion Recognition System through Speech Signal Using KNN & GMM Classifier

A New Perspective on Combining GMM and DNN Frameworks for Speaker Adaptation

CS4491/CS 7265 BIG DATA ANALYTICS INTRODUCTION TO THE COURSE. Mingon Kang, PhD Computer Science, Kennesaw State University

Evaluating Interactive Visualization of Multidimensional Data Projection with Feature Transformation

Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models

Impact of Cluster Validity Measures on Performance of Hybrid Models Based on K-means and Decision Trees

Ensemble Technique Utilization for Indonesian Dependency Parser

Speech Recognition at ICSI: Broadcast News and beyond

Lahore University of Management Sciences. FINN 321 Econometrics Fall Semester 2017

Multivariate k-nearest Neighbor Regression for Time Series data -

Word Segmentation of Off-line Handwritten Documents

OCR for Arabic using SIFT Descriptors With Online Failure Prediction

Australian Journal of Basic and Applied Sciences

Speech Segmentation Using Probabilistic Phonetic Feature Hierarchy and Support Vector Machines

Iterative Cross-Training: An Algorithm for Learning from Unlabeled Web Pages

A Comparison of Two Text Representations for Sentiment Analysis

Laboratorio di Intelligenza Artificiale e Robotica

Course Outline. Course Grading. Where to go for help. Academic Integrity. EE-589 Introduction to Neural Networks NN 1 EE

Model Ensemble for Click Prediction in Bing Search Ads

A Neural Network GUI Tested on Text-To-Phoneme Mapping

Universidade do Minho Escola de Engenharia

A Survey on Unsupervised Machine Learning Algorithms for Automation, Classification and Maintenance

Probabilistic Latent Semantic Analysis

WHEN THERE IS A mismatch between the acoustic

Learning Methods for Fuzzy Systems

CS 446: Machine Learning

Calibration of Confidence Measures in Speech Recognition

Deep search. Enhancing a search bar using machine learning. Ilgün Ilgün & Cedric Reichenbach

PREDICTING SPEECH RECOGNITION CONFIDENCE USING DEEP LEARNING WITH WORD IDENTITY AND SCORE FEATURES

A Case Study: News Classification Based on Term Frequency

Modeling function word errors in DNN-HMM based LVCSR systems

Historical maintenance relevant information roadmap for a self-learning maintenance prediction procedural approach

Rule Learning With Negation: Issues Regarding Effectiveness

SAM - Sensors, Actuators and Microcontrollers in Mobile Robots

Semi-Supervised GMM and DNN Acoustic Model Training with Multi-system Combination and Confidence Re-calibration

June 2, Via . Stephen Metz Bankruptcy Bar Association-District of Maryland 4800 Montgomery Lane Suite 900 Bethesda, MD RE: NEE1609

An OO Framework for building Intelligence and Learning properties in Software Agents

Semi-Supervised Face Detection

Unsupervised Learning of Word Semantic Embedding using the Deep Structured Semantic Model

PELLISSIPPI STATE TECHNICAL COMMUNITY COLLEGE MASTER SYLLABUS APPLIED MECHANICS MET 2025

*Net Perceptions, Inc West 78th Street Suite 300 Minneapolis, MN

Multi-label classification via multi-target regression on data streams

QuickStroke: An Incremental On-line Chinese Handwriting Recognition System

Class-Discriminative Weighted Distortion Measure for VQ-Based Speaker Identification

A Vector Space Approach for Aspect-Based Sentiment Analysis

Developing a Distance Learning Curriculum for Marine Engineering Education

TRANSFER LEARNING IN MIR: SHARING LEARNED LATENT REPRESENTATIONS FOR MUSIC AUDIO CLASSIFICATION AND SIMILARITY

Laboratorio di Intelligenza Artificiale e Robotica

CS 1103 Computer Science I Honors. Fall Instructor Muller. Syllabus

Reducing Features to Improve Bug Prediction

Training a Neural Network to Answer 8th Grade Science Questions Steven Hewitt, An Ju, Katherine Stasaski

OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS

International Journal of Computational Intelligence and Informatics, Vol. 1 : No. 4, January - March 2012

Experiment Databases: Towards an Improved Experimental Methodology in Machine Learning

ADVANCES IN DEEP NEURAL NETWORK APPROACHES TO SPEAKER RECOGNITION

PELLISSIPPI STATE TECHNICAL COMMUNITY COLLEGE MASTER SYLLABUS APPLIED STATICS MET 1040

Rule Learning with Negation: Issues Regarding Effectiveness

POS tagging of Chinese Buddhist texts using Recurrent Neural Networks

Test Effort Estimation Using Neural Network

Phonetic- and Speaker-Discriminant Features for Speaker Recognition. Research Project

HIERARCHICAL DEEP LEARNING ARCHITECTURE FOR 10K OBJECTS CLASSIFICATION

Softprop: Softmax Neural Network Backpropagation Learning

When Student Confidence Clicks

Reinforcement Learning by Comparing Immediate Reward

Knowledge management styles and performance: a knowledge space model from both theoretical and empirical perspectives

Introduction to Causal Inference. Problem Set 1. Required Problems

Transcription:

Subject Code: CSE4020 Indicative Pre-requisite Objective Expected Outcomes Machine Learning L,T,P,J,C 2,0,2,4,4 MAT2001- Statistics for Engineers It introduces theoretical foundations, algorithms, methodologies, and applications of Machine Learning and also provide practical knowledge for handling and analysing data sets covering a variety of real-world applications. After successfully completing the course the student should be able to 1. Recognize the characteristics of machine learning that make it useful to solve real-world problems. 2. Identify real-world applications of machine learning. 3. Identify and apply appropriate machine learning algorithms for analyzing the data for variety of problems. 4. Implement different machine learning algorithms using R and Python for analyzing the data 5. Design test procedures in order to evaluate a model 6. Combine several models in order to gain better results 7. Make choices for a model for new machine learning tasks based on reasoned argument SLO 5. Having design thinking capability 7. Having computational thinking (Ability to translate vast data in to abstract concepts and to understand database reasoning) 9. Having problem solving ability- solving social issues and engineering problems Module Topics L Hrs SLO 1 INTRODUCTION TO MACHINE LEARNING What is Machine Learning, Examples of Various Learning Paradigms, Perspectives and Issues, Version Spaces, Finite and Infinite Hypothesis Spaces, PAC Learning 2 Supervised Learning - I Learning a Class from Examples, Linear, Non-linear, Multi-class and Multi-label classification, Generalization error bounds: VC Dimension, Decision Trees: ID3, Classification and Regression Trees, Regression: Linear Regression, Multiple Linear Regression, Logistic Regression 3 5 4 5,7,9 3 Supervised Learning - II 5 5,7,9

Neural Networks: Introduction, Perceptron, Multilayer Perceptron, Support vector machines: Linear and Non-Linear, Kernel Functions, K- Nearest Neighbors 4 Ensemble Learning Model Combination Schemes, Voting, Error-Correcting Output Codes, Bagging: Random Forest Trees, Boosting: Adaboost, Stacking 5 Unsupervised Learning - I Introduction to clustering, Hierarchical: AGNES, DIANA, Partitional: K-means clustering, K-Mode Clustering, Self-Organizing Map, Expectation Maximization, Gaussian Mixture Models 6 Unsupervised Learning - II Principal components analysis (PCA), Locally Linear Embedding (LLE), Factor Analysis 7 Machine Learning in Practice Design, Analysis and Evaluation of Machine Learning Experiments, Feature selection Mechanisms, Other Issues: Imbalanced data, Missing Values, Outliers 3 5,7,9 7 5,7,9 3 5,7,9 3 5,7 8 Recent Trends in Machine Learning 2 Lab (Indicative List of Experiments (in the areas of ) 30 5,7 1. Implement Decision Tree learning 2. Implement Logistic Regression 3. Implement classification using Multilayer perceptron 4. Implement classification using SVM 5. Implement Adaboost 6. Implement Bagging using Random Forests 7. Implement K-means Clustering to Find Natural Patterns in Data 8. Implement Hierarchical clustering 9. Implement K-mode clustering 10. Implement Principle Component Analysis for Dimensionality Reduction 11. Implement Multiple Correspondence Analysis for Dimensionality Reduction

12. Implement Gaussian Mixture Model Using the Expectation Maximization 13. Evaluating ML algorithm with balanced and unbalanced datasets 14. Comparison of Machine Learning algorithms 15. Implement k-nearest neighbours algorithm Project # Generally a team project [5 to 10 members] # Concepts studied in XXXX should have been used # Down to earth application and innovative idea should have been attempted # Report in Digital format with all drawings using software package to be submitted. # Assessment on a continuous basis with a min of 3 reviews. Projects may be given as group projects The following is the sample project that can be given to students to be implemented: 1. Solving Data Science problems from Kaggle website 2. Applying Machine Learning algorithms in the field of biometrics for reliable and robust identification of humans from their personal traits, mainly for security and authentication purposes 3. Applying Machine Learning for OCR, Video Analytics 4. Applying Machine Learning algorithms in the field of Natural Language Processingfor document clustering and sentiment analysis 5. Applying Machine Learning for Fraud Detection, Customer segmentation etc. Note: Students can down load real time data sets for different Machine Learning Tasks from https://archive.ics.uci.edu/ml/datasets.html and http://sci2s.ugr.es/keel/datasets.php#sub1 and do the projects 60 [Non Contact hrs] 5,7,9 Text Books 1. Ethem Alpaydin,"Introduction to Machine Learning, MIT Press, Prentice Hall of India, Third Edition 2014. Reference Books 2. Mehryar Mohri, Afshin Rostamizadeh, Ameet Talwalkar "Foundations of Machine Learning, MIT Press, 2012. 3. Tom Mitchell, Machine Learning, McGraw Hill, 3rd Edition,1997. 4. Charu C. Aggarwal, Data Classification Algorithms and Applications, CRC Press, 2014. 5. Charu C. Aggarwal, DATA CLUSTERING Algorithms and Applications, CRC Press, 2014. 6. Kevin P. Murphy "Machine Learning: A Probabilistic Perspective", The MIT Press, 2012

Machine Learning Knowledge Areas that contain topics and learning outcomes covered in the course Knowledge Area Total Hours of Coverage CS: IS(Intelligent System) 30 Body of Knowledge coverage [List the Knowledge Units covered in whole or in part in the course. If in part, please indicate which topics and/or learning outcomes are covered. For those not covered, you might want to indicate whether they are covered in another course or not covered in your curriculum at all. This section will likely be the most time-consuming to complete, but is the most valuable for educators planning to adopt the CS2013 guidelines.] KA Knowledge Unit Topics Covered Hours CS: IS IS/Basic Machine Learning Introduction to Machine Learning 3 CS: IS IS/Advanced Machine Learning Supervised Learning - I Unsupervised Learning - I Machine Learning in Practice Supervised Learning - II Unsupervised Learning - II Ensemble Learning Recent Trends 27 Total hours 30

Where does the course fit in the curriculum? [In what year do students commonly take the course? Is it compulsory? Does it have prerequisites, required following courses? How many students take it?] This course is a Elective Course. Suitable from 5 th semester onwards. Knowledge of any one programming language is essential. What is covered in the course? [A short description, and/or a concise list of topics - possibly from your course syllabus.(this is likely to be your longest answer)] Part 1: Introduction to Machine Learning It introduces the concepts learning and various aspects of machine learning such different learning paradigms: Supervised, unsupervised, semi-supervised, reinforced. Part II: Supervised Learning - I It introduces Supervised learning algorithms for classification. Learning a Class from Examples, Linear, Non-linear, Multi-class and Multi-label classification, Generalization error bounds: VC Dimension, Decision Trees: ID3, Classification and Regression Trees, Regression: Linear Regression, Multiple Linear Regression, Logistic Regression Part III: Supervised Learning - II It introduces advanced Supervised learning algorithms for classification. Neural Networks: Introduction, Perceptron, Multilayer Perceptron, Support vector machines: Linear and Non- Linear, Kernel Functions, K-nearest neighbours Part IV: Ensemble Learning - I It introduces ensemble methods combining different models. Model Combination Schemes, Voting, Error-Correcting Output Codes, Bagging: Random Forest Trees, Boosting: Adaboost, Stacking

Part V: Unsupervised Learning - I It introduces various clustering techniques like Hierarchical: AGNES, DIANA, Partitional: K- means clustering, K-means++, K-Mode Clustering, Self-Organizing Map, Expectation Maximization, Gaussian Mixture Models Part VI: Unsupervised Learning - II It introduces dimensionality reduction techniques like Principal components analysis (PCA), Locally Linear Embedding (LLE), Factor Analysis Part VII: Machine Learning in Practice Design, Analysis and Evaluation of Machine Learning Experiments, Feature selection Mechanisms, Other Issues: Imbalanced data, Missing Values, Outliers What is the format of the course? [Is it face to face, online or blended? How many contact hours? Does it have lectures, lab sessions, discussion classes?] This Course is designed with 100 minutes of in-classroom sessions per week, 60 minutes of video/reading instructional material per week, 100 minutes of lab hours per week, as well as 200 minutes of non-contact time spent on implementing course related project. Generally this course should have the combination of lectures, in-class discussion, case studies, guest-lectures, mandatory off-class reading material, quizzes. How are students assessed? [What type, and number, of assignments are students are expected to do? (papers, problem sets, programming projects, etc.). How long do you expect students to spend on completing assessed work?] Students are assessed on a combination group activities, classroom discussion, projects, and continuous, final assessment tests. Additional weightage will be given based on their rank in crowd sourced projects/ Kaggle like competitions. Students can earn additional weightage based on certificate of completion of a related MOOC course.