Government of Russian Federation. Federal State Autonomous Educational Institution of High Professional Education

Similar documents
Generative models and adversarial training

Lecture 1: Machine Learning Basics

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

Python Machine Learning

The 9 th International Scientific Conference elearning and software for Education Bucharest, April 25-26, / X

Probabilistic Latent Semantic Analysis

Business Analytics and Information Tech COURSE NUMBER: 33:136:494 COURSE TITLE: Data Mining and Business Intelligence

Semi-Supervised GMM and DNN Acoustic Model Training with Multi-system Combination and Confidence Re-calibration

CSL465/603 - Machine Learning

(Sub)Gradient Descent

CS4491/CS 7265 BIG DATA ANALYTICS INTRODUCTION TO THE COURSE. Mingon Kang, PhD Computer Science, Kennesaw State University

A New Perspective on Combining GMM and DNN Frameworks for Speaker Adaptation

Calibration of Confidence Measures in Speech Recognition

Rule Learning With Negation: Issues Regarding Effectiveness

Exploration. CS : Deep Reinforcement Learning Sergey Levine

ACTL5103 Stochastic Modelling For Actuaries. Course Outline Semester 2, 2014

Speech Emotion Recognition Using Support Vector Machine

Artificial Neural Networks written examination

Unsupervised Learning of Word Semantic Embedding using the Deep Structured Semantic Model

Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models

Learning Methods in Multilingual Speech Recognition

Lecture 1: Basic Concepts of Machine Learning

Stochastic Calculus for Finance I (46-944) Spring 2008 Syllabus

Introduction to Simulation

Softprop: Softmax Neural Network Backpropagation Learning

PREDICTING SPEECH RECOGNITION CONFIDENCE USING DEEP LEARNING WITH WORD IDENTITY AND SCORE FEATURES

Deep Neural Network Language Models

A Review: Speech Recognition with Deep Learning Methods

Learning From the Past with Experiment Databases

Laboratorio di Intelligenza Artificiale e Robotica

Assignment 1: Predicting Amazon Review Ratings

QuickStroke: An Incremental On-line Chinese Handwriting Recognition System

GRADUATE STUDENT HANDBOOK Master of Science Programs in Biostatistics

CS 1103 Computer Science I Honors. Fall Instructor Muller. Syllabus

Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition

Semi-supervised methods of text processing, and an application to medical concept extraction. Yacine Jernite Text-as-Data series September 17.

Multi-Dimensional, Multi-Level, and Multi-Timepoint Item Response Modeling.

HIERARCHICAL DEEP LEARNING ARCHITECTURE FOR 10K OBJECTS CLASSIFICATION

ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY DOWNLOAD EBOOK : ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY PDF

Chapter 10 APPLYING TOPIC MODELING TO FORENSIC DATA. 1. Introduction. Alta de Waal, Jacobus Venter and Etienne Barnard

Human Emotion Recognition From Speech

Australian Journal of Basic and Applied Sciences

Rule Learning with Negation: Issues Regarding Effectiveness

BUILDING CONTEXT-DEPENDENT DNN ACOUSTIC MODELS USING KULLBACK-LEIBLER DIVERGENCE-BASED STATE TYING

Twitter Sentiment Classification on Sanders Data using Hybrid Approach

Autoregressive product of multi-frame predictions can improve the accuracy of hybrid models

Foothill College Summer 2016

Knowledge-Based - Systems

arxiv: v1 [cs.lg] 15 Jun 2015

Forget catastrophic forgetting: AI that learns after deployment

Evolutive Neural Net Fuzzy Filtering: Basic Description

Chinese Language Parsing with Maximum-Entropy-Inspired Parser

CS 101 Computer Science I Fall Instructor Muller. Syllabus

CS Machine Learning

Universidade do Minho Escola de Engenharia

Math 96: Intermediate Algebra in Context

Laboratorio di Intelligenza Artificiale e Robotica

The Good Judgment Project: A large scale test of different methods of combining expert predictions

Model Ensemble for Click Prediction in Bing Search Ads

Corrective Feedback and Persistent Learning for Information Extraction

Pp. 176{182 in Proceedings of The Second International Conference on Knowledge Discovery and Data Mining. Predictive Data Mining with Finite Mixtures

Mining Association Rules in Student s Assessment Data

Honors Mathematics. Introduction and Definition of Honors Mathematics

Lecture 10: Reinforcement Learning

An OO Framework for building Intelligence and Learning properties in Software Agents

An Introduction to the Minimalist Program

Content-based Image Retrieval Using Image Regions as Query Examples

Proposal of Pattern Recognition as a necessary and sufficient principle to Cognitive Science

COMPUTER-ASSISTED INDEPENDENT STUDY IN MULTIVARIATE CALCULUS

We are strong in research and particularly noted in software engineering, information security and privacy, and humane gaming.

A study of speaker adaptation for DNN-based speech synthesis

Reducing Features to Improve Bug Prediction

Georgetown University at TREC 2017 Dynamic Domain Track

Predicting Student Attrition in MOOCs using Sentiment Analysis and Neural Networks

Speech Recognition at ICSI: Broadcast News and beyond

Rule-based Expert Systems

Experiment Databases: Towards an Improved Experimental Methodology in Machine Learning

Lahore University of Management Sciences. FINN 321 Econometrics Fall Semester 2017

Truth Inference in Crowdsourcing: Is the Problem Solved?

Learning Methods for Fuzzy Systems

Welcome to. ECML/PKDD 2004 Community meeting

Axiom 2013 Team Description Paper

Machine Learning and Data Mining. Ensembles of Learners. Prof. Alexander Ihler

COMPUTER SCIENCE GRADUATE STUDIES Course Descriptions by Methodology

Radius STEM Readiness TM

SARDNET: A Self-Organizing Feature Map for Sequences

THE world surrounding us involves multiple modalities

AUTOMATIC DETECTION OF PROLONGED FRICATIVE PHONEMES WITH THE HIDDEN MARKOV MODELS APPROACH 1. INTRODUCTION

Data Fusion Through Statistical Matching

Deep search. Enhancing a search bar using machine learning. Ilgün Ilgün & Cedric Reichenbach

Dual-Memory Deep Learning Architectures for Lifelong Learning of Everyday Human Behaviors

COMPUTER SCIENCE GRADUATE STUDIES Course Descriptions by Research Area

FUZZY EXPERT. Dr. Kasim M. Al-Aubidy. Philadelphia University. Computer Eng. Dept February 2002 University of Damascus-Syria

Iterative Cross-Training: An Algorithm for Learning from Unlabeled Web Pages

Semi-Supervised Face Detection

Test Effort Estimation Using Neural Network

Development of Multistage Tests based on Teacher Ratings

System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks

DNN ACOUSTIC MODELING WITH MODULAR MULTI-LINGUAL FEATURE EXTRACTION NETWORKS

Math 098 Intermediate Algebra Spring 2018

Transcription:

Government of Russian Federation Federal State Autonomous Educational Institution of High Professional Education National Research University Higher School of Economics Syllabus for the course Advanced topics in Machine Learning for postgraduate program in 09.06.01 Computer Science and Computer Engineering / 05.13.01 Systems Analysis, Control Theory, and Information Processing, 05.13.11 Mathematical Theory and Software for Computing Machinery, Systems, and Networks, 05.13.17 Theoretical Foundations of Computer Science, 05.13.18 Mathematical Modeling, Numerical Methods, and Software Systems Author: Attila Kertesz-Farkas, assistant professor, akerteszfarkas@hse.ru Approved by the Academic Council of the School for Postgraduate Studies in Computer Science on October 26, 2014 Moscow - 2017 This program cannot be used by other departments and other universities without the author s permission. 1

1. Scope of Use This program establishes the minimal requirements to postgraduate students knowledge and skills for 09.06.01 Computer Science and Computer Engineering / 05.13.01 Systems Analysis, Control Theory, and Information Processing, 05.13.11 Mathematical Theory and Software for Computing Machinery, Systems, and Networks, 05.13.17 Theoretical Foundations of Computer Science, 05.13.18 Mathematical Modeling, Numerical Methods, and Software Systems and determines the content of the course and educational techniques used in teaching the course. The present syllabus is aimed at faculty teaching the course and postgraduate students studying 09.06.01 Computer Science and Computer Engineering / 05.13.01 Systems Analysis, Control Theory, and Information Processing, 05.13.11 Mathematical Theory and Software for Computing Machinery, Systems, and Networks, 05.13.17 Theoretical Foundations of Computer Science, 05.13.18 Mathematical Modeling, Numerical Methods, and Software Systems. This syllabus meets the standards required by: Educational standards of National Research University Higher School of Economics; Postgraduate educational program for 09.06.01 Computer Science and Computer Engineering. University curriculum of the postgraduate program for 09.06.01 Computer Science and Computer Engineering / 05.13.01 Systems Analysis, Control Theory, and Information Processing, 05.13.11 Mathematical Theory and Software for Computing Machinery, Systems, and Networks, 05.13.17 Theoretical Foundations of Computer Science, 05.13.18 Mathematical Modeling, Numerical Methods, and approved in 2014. 2. Learning Objectives The learning objective of the course Advanced topics on Machine Learning is to provide students advanced techniques and deeper theoretical and practical knowledge in modern probabilistic learning techniques, such as: Basic principles, Generative Models, Bayesian Network, Random Markov Fields, Boltzmann Machines, Auto Encoders Sampling and Inference, Neural Networks, Deep Learning techniques. 3. Main Competencies Developed after Completing the Study of This Discipline After completing the study of the discipline the PhD student should have: Knowledge about probabilistic models. Knowledge about modern methods such as deep learning techniques. Knowledge about ongoing developments in Machine Learning Hands-on experience with large scale machine learning problems. Knowledge about how to design and develop machine learning programs using a programming language such as R or Python. Think critically with real data. 2

After completing the study of the discipline the student should have developed the following competencies: Competence Code Descriptors (indicators of achievement of the result) the ability to carry out theoretical and experimental research in the field of professional activity the ability to develop new research methods and apply them in research in one s professional field the ability to objectively evaluate the outcomes of research and development carried out by other specialists in other scientific institutions the ability to do research in transformation of information into data and knowledge, models of data and knowledge representation, methods for knowledge processing, machine learning and knowledge discovery methods, principles of building and operating software for automation of these processes ОПК-1 PhD students obtain necessary knowledge in probabilistic generative models ОПК-2 The PhD student is able to choose an appropriate model for real-life problems and to calibrate the hyperparameters. ОПК-4 The PhD student is able to carry out comparative testing of competing models or methods. ПК-4 The PhD student is able to develop and analyze machine learning models, implement them in a programming language in large scale, and select the best model using validation techniques. Educative forms and methods aimed at generation and development of the competence Assignments, additional material/reading provided Examples covered during the lectures and tutorials. Assignments. Examples covered during the lectures and tutorials. Assignments. Lectures, tutorials, and assignments. 4. Place of the Discipline in the Postgraduate Program Structure This is an elective course for 05.13.01 Systems Analysis, Control Theory, and Information Processing, 05.13.11 Mathematical Theory and Software for Computing Machinery, Systems, and Networks, 05.13.17 Theoretical Foundations of Computer Science, 05.13.18 Mathematical Modeling, Numerical Methods, and Software Systems. Postgraduate students are expected to be already familiar with some statistical learning techniques, and have skills in analysis, linear algebra, optimization, computational complexity, and probability theory. The following knowledge and competences are needed to study the discipline: A good command of the English language, both oral and written. A sound knowledge of probability theory, complexity theory, optimization, and linear algebra 3

5. Schedule for one semesters (2 modules) Topic Contact hours Total Lectur Semin Practice hours es ars lessons Self-study 1. Introduction of Machine Learning 9 2 1 6 2. Bayesian Learning, Exponential Families 9 2 1 6 3. Graphical Models 27 6 3 18 4. Sampling and Inference 27 6 3 18 5. Variational Learning 36 8 4 24 6. Generative Learning 113 6 3 104 7. Deep learning techniques 33 6 3 24 8. Optimization and Regularization 18 4 2 12 9. Student Presentation 108 4 2 102 Total 380 44 22 314 6. Requirements and Grading 7. Assessment Mid-Term Exam 1 Mid-semester test. Written exam. Presence 1 Exam 1 Written exam. Preparation time 180 min. Final assessments are based on the mid-exam and the final exam. Students have to demonstrate knowledge of the material covered during the entire course. 8. The grade formula The exam is worth 60% of the final mark. Final course mark is obtained from the following formula: Final=0.2*(Mid-term exam)+ 0.2*(Presence on all lectures and seminars)+0.6*(exam). All grades having a fractional part greater than 0.5 are rounded up. Table of Grade Accordance Ten-point grading Scale 1 - very bad 2 bad 3 no pass 4 pass 5 highly pass 6 good 7 very good 8 almost excellent 9 excellent 10 perfect Five-point grading Scale Unsatisfactory - 2 Satisfactory 3 Good 4 Excellent 5 FAIL PASS 4

9. Course description. National Research University Higher School of Economics Topic 1. Introduction to machine learning, Bayesian Decision Theory, Maximum Likelihood Estimation, and EM. Basic definitions, principles and types of machine learning. Classifiers, Discriminant Functions, and Decision Surfaces, Minimum-Error-Rate Classification, Neyman-Pearson lemma, Distributions, Relation to Logistic Regression, Naïve Bayes classification, basics of MLE, learning parameters of distributions. Gaussian Mixture Models, Latent Variables, Examples, Expectation-Maximization, Latent Dirichlet Allocation. Topic 2. Exponential Family, Sufficient Statistics. Generalized Linear Models, Topic 3. Graphical Models Bayesian Networks, Random Markov Fields, Conditional Random Fields, Boltzmann Machines, Energy-based methods. Hidden Markov Models. Topic 4. Sampling and Inference Exact and Inexact Inference, Gibbs sampling, Bridge Sampling, Simple and Annealed Importance Sampling, Monte-Carlo EM, Junction Tree algorithm Topic 5. Variational Learning Mean-Field, Bethe Approximation, Variational Bayes, Variational Message Passing, Free-Energy, Variational Free Energy. Topic 6. Generative Learning Restricted Boltzmann Machines, Helmoltz Machines and Wake-Sleep algorithms, Energy-based methods. Generative Adversarial Networks, Generative Auto-Encoders, Belief networks, connectionist learning. Topic 7. Deep learning techniques Neural Networks, Shallow networks, Multilayer Neural networks, back-propagation, deep learning, Universal Approximation. Auto Encoders, Stacked Auto-Encoders, Stacked Boltzmann machines, supervised and unsupervised pre-training, Deep Belief Networks. Topic 8. Optimization and Regularization Error Surfaces, Optimization, Regularization. Topic 9. Student Presentation Students to select a topic from this class at their wish and to give a short on this topic in order to improve their communication skills on this topic. 5

10. Educational technologies The following educational technologies are used in the study process: discussion and analysis of the results during the tutorials; regular assignments to test the progress of the PhD student; consultation time on Monday afternoons. 11. Final exam questions The final exam will consist of a selection of problems equally weighted. No material is allowed for the exam. Each question will focus on a particular topic presented during the lectures. The questions consist in exercises on any topic seen during the lectures. To be prepared for the final exam, PhD students must be able to answer questions from the topics covered during the lecture. 12. Reading and Materials Literature: 1. Kevin Murphy, Machine Leaning: A probabilistic Perspective, 2013, MIT press 2. C. Bishop: Pattern Recognition and Machine Learning, 3. I. Goodfellow, Y. Bengio and A. Courville, Deep Learning. 2016 MIT press 4. G. James, D. Witten, T. Hastie, R. Tibshirani. An introduction to Statistical Learning, 2013, Springer 5. Li Deng, Dong Yu: Deep Learning: Methods and Applications, 2014, Now publishers. 6. M. J. Wainwright, M. I. Jordan: Graphical Models, Exponential Families, and Variational Inference, 2008, Now publishers Literature for self-study: 1. Y Bengio: Learning Deep Architectures for AI; Machine Learning, 2009, Vol. 2, No. 1, 13. Equipment. The course requires a computer room, laptop and a projector. 6