Automatic Speech Recognition (CS753)

Similar documents
CSL465/603 - Machine Learning

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

Lecture 1: Machine Learning Basics

CS4491/CS 7265 BIG DATA ANALYTICS INTRODUCTION TO THE COURSE. Mingon Kang, PhD Computer Science, Kennesaw State University

Python Machine Learning

(Sub)Gradient Descent

System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks

CS Machine Learning

The 9 th International Scientific Conference elearning and software for Education Bucharest, April 25-26, / X

Business Analytics and Information Tech COURSE NUMBER: 33:136:494 COURSE TITLE: Data Mining and Business Intelligence

Axiom 2013 Team Description Paper

Lecture 1: Basic Concepts of Machine Learning

Laboratorio di Intelligenza Artificiale e Robotica

Testing A Moving Target: How Do We Test Machine Learning Systems? Peter Varhol Technology Strategy Research, USA

Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition

Generative models and adversarial training

Deep search. Enhancing a search bar using machine learning. Ilgün Ilgün & Cedric Reichenbach

ENME 605 Advanced Control Systems, Fall 2015 Department of Mechanical Engineering

Laboratorio di Intelligenza Artificiale e Robotica

ASTRONOMY 2801A: Stars, Galaxies & Cosmology : Fall term

Firms and Markets Saturdays Summer I 2014

CS224d Deep Learning for Natural Language Processing. Richard Socher, PhD

Rule Learning With Negation: Issues Regarding Effectiveness

Lahore University of Management Sciences. FINN 321 Econometrics Fall Semester 2017

CS 446: Machine Learning

Probabilistic Latent Semantic Analysis

COMPUTER-ASSISTED INDEPENDENT STUDY IN MULTIVARIATE CALCULUS

Learning From the Past with Experiment Databases

Syllabus Foundations of Finance Summer 2014 FINC-UB

Assignment 1: Predicting Amazon Review Ratings

Innovative Methods for Teaching Engineering Courses

MAE Flight Simulation for Aircraft Safety

ACTL5103 Stochastic Modelling For Actuaries. Course Outline Semester 2, 2014

INPE São José dos Campos

EECS 700: Computer Modeling, Simulation, and Visualization Fall 2014

Syllabus ENGR 190 Introductory Calculus (QR)

Training a Neural Network to Answer 8th Grade Science Questions Steven Hewitt, An Ju, Katherine Stasaski

Class Meeting Time and Place: Section 3: MTWF10:00-10:50 TILT 221

Instructor: Matthew Wickes Kilgore Office: ES 310

ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY DOWNLOAD EBOOK : ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY PDF

Spring 2014 SYLLABUS Michigan State University STT 430: Probability and Statistics for Engineering

Computers Change the World

Massachusetts Institute of Technology Tel: Massachusetts Avenue Room 32-D558 MA 02139

We are strong in research and particularly noted in software engineering, information security and privacy, and humane gaming.

Office Hours: Mon & Fri 10:00-12:00. Course Description

arxiv: v1 [cs.cl] 2 Apr 2017

SYLLABUS. EC 322 Intermediate Macroeconomics Fall 2012

Machine Learning from Garden Path Sentences: The Application of Computational Linguistics

Comparison of network inference packages and methods for multiple networks inference

Twitter Sentiment Classification on Sanders Data using Hybrid Approach

Exploration. CS : Deep Reinforcement Learning Sergey Levine

Ph.D in Advance Machine Learning (computer science) PhD submitted, degree to be awarded on convocation, sept B.Tech in Computer science and

Learning Optimal Dialogue Strategies: A Case Study of a Spoken Dialogue Agent for

Human Emotion Recognition From Speech

Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models

arxiv: v1 [cs.lg] 15 Jun 2015

Phys4051: Methods of Experimental Physics I

EECS 571 PRINCIPLES OF REAL-TIME COMPUTING Fall 10. Instructor: Kang G. Shin, 4605 CSE, ;

Iterative Cross-Training: An Algorithm for Learning from Unlabeled Web Pages

CS/SE 3341 Spring 2012

AN EXAMPLE OF THE GOMORY CUTTING PLANE ALGORITHM. max z = 3x 1 + 4x 2. 3x 1 x x x x N 2

An investigation of imitation learning algorithms for structured prediction

MGT/MGP/MGB 261: Investment Analysis

A study of speaker adaptation for DNN-based speech synthesis

BPS Information and Digital Literacy Goals

Australian Journal of Basic and Applied Sciences

Reinforcement Learning by Comparing Immediate Reward

Analysis of Emotion Recognition System through Speech Signal Using KNN & GMM Classifier

Self Study Report Computer Science

EDCI 699 Statistics: Content, Process, Application COURSE SYLLABUS: SPRING 2016

ScienceDirect. A Framework for Clustering Cardiac Patient s Records Using Unsupervised Learning Techniques

Universidade do Minho Escola de Engenharia

Learning Methods for Fuzzy Systems

Time series prediction

Active Learning. Yingyu Liang Computer Sciences 760 Fall

Pre-AP Geometry Course Syllabus Page 1

Model Ensemble for Click Prediction in Bing Search Ads

LOUISIANA HIGH SCHOOL RALLY ASSOCIATION

Word Segmentation of Off-line Handwritten Documents

Evolution of Symbolisation in Chimpanzees and Neural Nets

Speech Recognition at ICSI: Broadcast News and beyond

Spring 2015 IET4451 Systems Simulation Course Syllabus for Traditional, Hybrid, and Online Classes

Artificial Neural Networks written examination

Content-free collaborative learning modeling using data mining

Learning Disability Functional Capacity Evaluation. Dear Doctor,

Unsupervised Learning of Word Semantic Embedding using the Deep Structured Semantic Model

Radius STEM Readiness TM

Physics Experimental Physics II: Electricity and Magnetism Prof. Eno Spring 2017

SY 6200 Behavioral Assessment, Analysis, and Intervention Spring 2016, 3 Credits

ME 443/643 Design Techniques in Mechanical Engineering. Lecture 1: Introduction

ACCOUNTING FOR MANAGERS BU-5190-OL Syllabus

Computational Data Analysis Techniques In Economics And Finance

OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS

Indian Institute of Technology, Kanpur

Probability and Game Theory Course Syllabus

CS 100: Principles of Computing

Semi-Supervised Face Detection

Detailed course syllabus

HIERARCHICAL DEEP LEARNING ARCHITECTURE FOR 10K OBJECTS CLASSIFICATION

ECON492 Senior Capstone Seminar: Cost-Benefit and Local Economic Policy Analysis Fall 2017 Instructor: Dr. Anita Alves Pena

Transcription:

Automatic Speech Recognition (CS753) Introduction to Machine Learning (CS419M) Lecture 1: What and why? Jan 5, 2018

What is Machine Learning? Ability of computers to learn from data or past experience

What is Machine Learning? Ability of computers to learn from data or past experience data: Comes from various sources such as sensors, domain knowledge, experimental runs, etc.

What is Machine Learning? Ability of computers to learn from data or past experience data: Comes from various sources such as sensors, domain knowledge, experimental runs, etc. learn: Make intelligent predictions or decisions based on data

Pigeon Superstition Video link: https://www.youtube.com/watch?v=ttfqlkgwe2u

What is Machine Learning? Ability of computers to learn from data or past experience data: Comes from various sources such as sensors, domain knowledge, experimental runs, etc. learn: Make intelligent predictions or decisions based on data 1. Supervised learning

Supervised Learning Given a labeled set of input-output pairs, D = {(x i,y i )} N i=1 objective is to learn a function mapping the inputs x to outputs y Inputs can be complex objects such as images, sentences, speech signals, etc. Typically take the form of features. Outputs are either categorical (classification tasks) or real-valued (regression tasks). More on these concepts in later classes.

Image recognition Image from ImageNet classification with deep CNNs, Krizhevsky et al.

What is Machine Learning? Ability of computers to learn from data or past experience data: Comes from various sources such as sensors, domain knowledge, experimental runs, etc. learn: Make intelligent predictions or decisions based on data 1. Supervised learning: decision trees, neural networks, etc.

What is Machine Learning? Ability of computers to learn from data or past experience data: Comes from various sources such as sensors, domain knowledge, experimental runs, etc. learn: Make intelligent predictions or decisions based on data 1. Supervised learning: decision trees, neural networks, etc. 2. Unsupervised learning

Unsupervised Learning Given a set of inputs, D = {x i } N i=1, discover some patterns in the data Most common example: Clustering

What is Machine Learning? Ability of computers to learn from data or past experience data: Comes from various sources such as sensors, domain knowledge, experimental runs, etc. learn: Make intelligent predictions or decisions based on data 1. Supervised learning: decision trees, neural networks, etc. 2. Unsupervised learning: k-means clustering, etc.

What is Machine Learning? Ability of computers to learn from data or past experience data: Comes from various sources such as sensors, domain knowledge, experimental runs, etc. learn: Make intelligent predictions or decisions based on data 1. Supervised learning: decision trees, neural networks, etc. 2. Unsupervised learning: k-means clustering, etc. 3. Reinforcement learning

When do we need ML? (I) For tasks that are easily performed by humans but are complex for computer systems to emulate Vision: Identify faces in a photograph, objects in a video or still image, etc. Natural language: Translate a sentence from Hindi to English, question answering, etc. Speech: Recognise spoken words, speaking sentences naturally Game playing: Play games like chess Robotics: Walking, jumping, displaying emotions, etc. Driving a car, flying a plane, navigating a maze, etc.

Relationship between AI, ML, DL Image from: https://blogs.nvidia.com/blog/2016/07/29/whats-difference-artificial-intelligence-machine-learning-deep-learning-ai/

When do we need ML? (II) For tasks that are beyond human capabilities Analysis of large and complex datasets E.g. IBM Watson s Jeopardy-playing machine Image credit: https://i.ytimg.com/vi/p18edakuc1u/maxresdefault.jpg

When do we need ML? (II) For tasks that are beyond human capabilities Analysis of large and complex datasets E.g. Autopilot controls Image credit: https://media.newyorker.com/photos/59095c8a019dfc3494e9f7b9/16:9/w_1200,h_630,c_limit/hazards-of-autopilot.jpg

ML and Statistics? Glossary Machine learning Statistics network, graphs model weights parameters learning fitting generalization test set performance supervised learning regression/classification unsupervised learning density estimation, clustering large grant = $1,000,000 large grant= $50,000 nice place to have a meeting: Snowbird, Utah, French Alps nice place to have a meeting: Las Vegas in August Glossary from: http://statweb.stanford.edu/~tibs/stat315a/glossary.pdf

Course Specifics

Pre-requisites Officially: No prerequisites. But would help if you ve taken Data Structures and Algorithms, Data Analysis and Interpretation, or any equivalent course taught by respective departments When to take this course: Should be comfortable with probability linear algebra multivariable calculus programming

Course webpage https://www.cse.iitb.ac.in/~pjyothi/cs419/

Course logistics Reading: All mandatory reading will be freely available online and posted on the course website. Textbooks (available online): 1. Understanding Machine Learning. Shai Shalev-Shwartz and Shai Ben-David. Cambridge University Press. 2017. 2. The Elements of Statistical Learning. Trevor Hastie, Robert Tibshirani and Jerome Friedman. Second Edition. 2009. Communication: We will use Moodle to communicate about the course. Attendance: Strongly advised to attend all lectures. Lot of material covered in class will not be on the slides. (Also, points for participation.)

Course TAs 1. Himanshu Agarwal, M.Tech. II 2. Aniket Kuiri, M.Tech. II 3. Pooja Palod, M.Tech. II 4. Sunandini Sanyal, M.Tech. I 5. Anupama Vijjapu, M.Tech. I 6. Saurabh Garg, B.Tech. IV 7. Tanmay Parekh, B.Tech. IV

Course Syllabus Provide an overview of machine learning and well-known techniques. We will briefly cover some ML applications as well. Some Topics: Basic foundations of ML, classification/regression, Naive Bayes classifier, linear and logistic regression Supervised learning: Decision trees, support vector machines, neural networks, etc. Unsupervised learning: k-means clustering, etc. Brief introduction to ML applications in computer vision, speech and natural language processing.

Evaluation (subject to change) Assignments: 4-5 assignments contributing to 40% of your grade. Late policy for assignments will vary and will be announced along with the assignment. Assignments will contain programming questions Midsem and final exam: 15% + 25% Participation: 5%. Four in-class quizzes will be randomly handed out. Should have attempted at least two to receive full points for participation.

Academic Integrity Policy Write what you know. Use your own words. If you refer to *any* external material, *always* cite your sources. Follow proper citation guidelines. If you re caught for plagiarism or copying, penalties are much higher than simply omitting that question. In short: Just not worth it. Image credit: https://www.flickr.com/photos/kurok/22196852451

Evaluation Project Grading: Constitutes 15% of the total grade. (Exceptional projects could get extra credit. Details posted on website.) Team: 2-3 members. Individual projects are highly discouraged. Project details: Apply the techniques you studied in class to any interesting problem of your choice Think of a problem early and work on it throughout the course. Project milestones will be posted on Moodle. Examples of project ideas: auto-complete code, generate song lyrics, help irctc predict ticket prices, etc. Feel free to be creative; consult with TAs/me if it s feasible

Datasets abound Kaggle: https://www.kaggle.com/datasets

Datasets abound Kaggle: https://www.kaggle.com/datasets Another good resource: http://deeplearning.net/datasets/ Popular resource for ML beginners: http://archive.ics.uci.edu/ml/index.php Interesting datasets for computational journalists: http://cjlab.stanford.edu/2015/09/30/lab-launch-and-data-sets/ Speech and language resources: www.openslr.org/ and so do ML libraries/toolkits scikit-learn, opencv, Keras, Tensorflow, NLTK, etc.

Some basic concepts

Typical ML approach How do we approach an ML problem? Modeling: Use a model to represent the task

Modeling Word Trans Posn Phn L-Lag T-Lag G-Lag L-Phn Prev- Phn T-Phn G-Phn Lip-op TT-op Glot sur Lip-op sur TT-op sur Glot

Typical ML approach How do we approach an ML problem? Modeling: Use a model to represent the task Decoding/Inference: Given a model, answer questions with respect to the model

Inference Given an observed set of values, how accurately can we predict the identity of a word? Frame # Articulatory Features Word Posn Word list weighted by probabilities 1 G1 LP0 V0 T0 Phn 2 G1 LP1 V0 T0 L-Lag T-Lag G-Lag 3 G1 LP1 V0 T0 4 G1 LP2 V1 T3 L-Phn T-Phn G-Phn 5 G1 LP2 V1 T3 : Lip-Op sur! Lip-Op TT-Op sur! TT-Op Glot sur! Glot Articulatory Features

Typical ML approach How do we approach an ML problem? Modeling: Use a model to represent the task Decoding/Inference: Given a model, answer questions with respect to the model Learning: The model could be parameterized and the parameters are learned using data

How do we know if our model s any good? Generalization: Does the trained model produce good predictions on examples beyond the training set? We should be careful not to overfit the training data Occam s Razor: All other things being equal, pick the simplest solution These concepts will be made more precise in later classes

No free lunch theorem There is no single best model that works optimally for all kinds of problems (Wolpert 1997) No learning is possible without some prior assumptions about the problem at hand Need many different types of models to cover variety of problems in the real world. Each model will have a range of algorithms that can be used to train it.