Security Analytics Review for Final Exam. Purdue University Prof. Ninghui Li

Similar documents
Python Machine Learning

Lecture 1: Machine Learning Basics

(Sub)Gradient Descent

Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition

System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks

A Simple VQA Model with a Few Tricks and Image Features from Bottom-up Attention

Machine Learning and Data Mining. Ensembles of Learners. Prof. Alexander Ihler

Artificial Neural Networks written examination

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

Softprop: Softmax Neural Network Backpropagation Learning

Learning From the Past with Experiment Databases

arxiv: v1 [cs.lg] 15 Jun 2015

Reducing Features to Improve Bug Prediction

Course Outline. Course Grading. Where to go for help. Academic Integrity. EE-589 Introduction to Neural Networks NN 1 EE

QuickStroke: An Incremental On-line Chinese Handwriting Recognition System

Rule Learning With Negation: Issues Regarding Effectiveness

HIERARCHICAL DEEP LEARNING ARCHITECTURE FOR 10K OBJECTS CLASSIFICATION

OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS

CSL465/603 - Machine Learning

Predicting Student Attrition in MOOCs using Sentiment Analysis and Neural Networks

Assignment 1: Predicting Amazon Review Ratings

INPE São José dos Campos

Autoregressive product of multi-frame predictions can improve the accuracy of hybrid models

A Neural Network GUI Tested on Text-To-Phoneme Mapping

CS Machine Learning

arxiv: v1 [cs.cv] 10 May 2017

Rule Learning with Negation: Issues Regarding Effectiveness

Generative models and adversarial training

Unsupervised Learning of Word Semantic Embedding using the Deep Structured Semantic Model

Framewise Phoneme Classification with Bidirectional LSTM and Other Neural Network Architectures

Active Learning. Yingyu Liang Computer Sciences 760 Fall

Twitter Sentiment Classification on Sanders Data using Hybrid Approach

Lecture 1: Basic Concepts of Machine Learning

Model Ensemble for Click Prediction in Bing Search Ads

The 9 th International Scientific Conference elearning and software for Education Bucharest, April 25-26, / X

Deep search. Enhancing a search bar using machine learning. Ilgün Ilgün & Cedric Reichenbach

Iterative Cross-Training: An Algorithm for Learning from Unlabeled Web Pages

Calibration of Confidence Measures in Speech Recognition

Issues in the Mining of Heart Failure Datasets

A New Perspective on Combining GMM and DNN Frameworks for Speaker Adaptation

Analysis of Hybrid Soft and Hard Computing Techniques for Forex Monitoring Systems

Human Emotion Recognition From Speech

Training a Neural Network to Answer 8th Grade Science Questions Steven Hewitt, An Ju, Katherine Stasaski

Switchboard Language Model Improvement with Conversational Data from Gigaword

Speech Emotion Recognition Using Support Vector Machine

Second Exam: Natural Language Parsing with Neural Networks

Machine Learning from Garden Path Sentences: The Application of Computational Linguistics

Using Deep Convolutional Neural Networks in Monte Carlo Tree Search

Test Effort Estimation Using Neural Network

ScienceDirect. A Framework for Clustering Cardiac Patient s Records Using Unsupervised Learning Techniques

SARDNET: A Self-Organizing Feature Map for Sequences

Dropout improves Recurrent Neural Networks for Handwriting Recognition

Knowledge Transfer in Deep Convolutional Neural Nets

Artificial Neural Networks

Dual-Memory Deep Learning Architectures for Lifelong Learning of Everyday Human Behaviors

Multivariate k-nearest Neighbor Regression for Time Series data -

I-COMPETERE: Using Applied Intelligence in search of competency gaps in software project managers.

arxiv: v1 [cs.lg] 7 Apr 2015

POS tagging of Chinese Buddhist texts using Recurrent Neural Networks

CS 1103 Computer Science I Honors. Fall Instructor Muller. Syllabus

Optimizing to Arbitrary NLP Metrics using Ensemble Selection

Applications of data mining algorithms to analysis of medical data

Lecture 10: Reinforcement Learning

Learning to Schedule Straight-Line Code

Глубокие рекуррентные нейронные сети для аспектно-ориентированного анализа тональности отзывов пользователей на различных языках

PREDICTING SPEECH RECOGNITION CONFIDENCE USING DEEP LEARNING WITH WORD IDENTITY AND SCORE FEATURES

Experiment Databases: Towards an Improved Experimental Methodology in Machine Learning

TRANSFER LEARNING OF WEAKLY LABELLED AUDIO. Aleksandr Diment, Tuomas Virtanen

Word learning as Bayesian inference

Learning Methods for Fuzzy Systems

Semantic Segmentation with Histological Image Data: Cancer Cell vs. Stroma

Version Space. Term 2012/2013 LSI - FIB. Javier Béjar cbea (LSI - FIB) Version Space Term 2012/ / 18

Testing A Moving Target: How Do We Test Machine Learning Systems? Peter Varhol Technology Strategy Research, USA

Universidade do Minho Escola de Engenharia

Time series prediction

An Empirical Analysis of the Effects of Mexican American Studies Participation on Student Achievement within Tucson Unified School District

Content-based Image Retrieval Using Image Regions as Query Examples

A Deep Bag-of-Features Model for Music Auto-Tagging

have to be modeled) or isolated words. Output of the system is a grapheme-tophoneme conversion system which takes as its input the spelling of words,

Language Acquisition Fall 2010/Winter Lexical Categories. Afra Alishahi, Heiner Drenhaus

An empirical study of learning speed in backpropagation

Attributed Social Network Embedding

CS 446: Machine Learning

ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY DOWNLOAD EBOOK : ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY PDF

Phonetic- and Speaker-Discriminant Features for Speaker Recognition. Research Project

Mining Association Rules in Student s Assessment Data

Impact of Cluster Validity Measures on Performance of Hybrid Models Based on K-means and Decision Trees

Exploration. CS : Deep Reinforcement Learning Sergey Levine

Multi-Lingual Text Leveling

Linking Task: Identifying authors and book titles in verbose queries

A Case Study: News Classification Based on Term Frequency

FRAMEWORK FOR IDENTIFYING THE MOST LIKELY SUCCESSFUL UNDERPRIVILEGED TERTIARY STUDY BURSARY APPLICANTS

Indian Institute of Technology, Kanpur

Australian Journal of Basic and Applied Sciences

Business Analytics and Information Tech COURSE NUMBER: 33:136:494 COURSE TITLE: Data Mining and Business Intelligence

Using focal point learning to improve human machine tacit coordination

arxiv: v4 [cs.cl] 28 Mar 2016

Algebra 1, Quarter 3, Unit 3.1. Line of Best Fit. Overview

Grade 6: Correlated to AGS Basic Math Skills

On the Formation of Phoneme Categories in DNN Acoustic Models

Transcription:

Security Analytics Review for Final Exam Purdue University Prof. Ninghui Li

Exam Date/Time Monday Dec 10 (8am 10am) LWSN B134

Organization of the Course Basic machine learning algorithms Neural networks Big data analytics Advasarial machine learning

Topic 2 Tasks: Exploratory, Descriptive, Predictive, Pattern Discovery What are the differences between supervised learning and unsupervised learning?

Topic 2 Concepts of Model space Scoring function Search technique Distance metrics Minkowski: Manhattan, Euclidean, L_0, L_\infty Jaccard

Topic 2 Explain the knn algorithm for classification. What is the training process? How to predict a sample x? Does a high k value result in a more complex model or a simpler model (smoother decision boundary)? How should one determine k? Is training fast or slow? How large is the model size?

Topic 4: Probability Review Able to do conditional probability computation Able to judge independent and dependent events Understand the base rate fallacy Under Conditional Independence Able to compute Bernoulli and Binomial

Topic 5: CLassification Accuracy, Precision and recall, F1 score Naïve Bayes on discrete-valued features Smoothing

Topic 6: Logistic Regression and SVM Linear regression Sum-square Error (SSE) Logistic-regression Intuition, Odds-Ratio, Maximum likelihood estimation Intuition behind SVM (margin) Linear versus kernel-based SVM

Topic 7: Decision Trees Inductive Learning Hypothesis IID assumption Understand two sources of inductive bias Language bias Search bias Impossibility of bias-free learning How to build a decision tree Calculating entropy, information gain, Gini impurity Overfitting, prepruning, postpruning (reduced error pruning)

Topic 8: Bagging and Random Forest Bagging: Bootstrap aggregating Bootstrap sampling Limitations of bagging with decision trees (i.d. not i.i.d.) Random forests Need for feature selection Increasing number of trees causes no overfitting

Topic 8: Neural Network (1) Types of neurons Linear, binary threshold, rectified Linear, sigmoid (remember)

Neural Network (2) Architecture of NN Feed-forward, recurrent Percentron classifier Percentron learning rule Training for each instance Multilayered percentron doesn t help without non-linearity The need for hidden layers Without them, limited in the model space Hidden layers learn features

Neural Network (3) Backpropagation Compute gradients (partial derivatives) of error function relative to each weight Online, full batch, and mini-batch

Neural Network (4) Definition of softmax, Definition of cross-entropy

Neural Network (5) Convolutional neural networks Why we need them? What other things we can do if not using CNN? Replicating feature recognizer

Neural Network (6) Ways to speed up mini-batch learning Momentum, separate adaptive learning rate, rprop, rmsprop

Neural Network (9) Ways of dealing of overfitting Weight-decay, Weight-sharing, Early stopping Model averaging, Dropout Creating new training data

Recurrent Neural Networks Types of input-output Understand issue of Vanishing gradients Gated recurrent units LSTM

Map-Reduce Challenges of cluster computing: Node failures, network bottle-neck, programming Meeting the challenges Redundant storage of files, moving jobs to where data is, Map-reduce framework Steps involved in Map-reduce framework. How to combine Map and reduce to solve problems. How the map-reduce framework deal with failures: map worker, reducer, master?

Spark Dataframes Concepts of transformations and actions Why it is faster than map-reduce

PageRank How to compute pagerank for simple examples by power iteration method. Random walk interpretation Dead ends and spider traps How dead ends and spider traps are handled?

Adversarial Machine Learning What are adversarial examples? Not just for Neural Networks Relationship to linearity in input What do the different maps of Adversarial and Random Cross-Sections mean? Concept of transferability