COMP9318 Review. Wei UNSW. June 4, 2018

Similar documents
Python Machine Learning

Mining Association Rules in Student s Assessment Data

(Sub)Gradient Descent

Probabilistic Latent Semantic Analysis

CS Machine Learning

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

WHEN THERE IS A mismatch between the acoustic

CSL465/603 - Machine Learning

Lecture 1: Machine Learning Basics

Australian Journal of Basic and Applied Sciences

Learning From the Past with Experiment Databases

Assignment 1: Predicting Amazon Review Ratings

Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition

The 9 th International Scientific Conference elearning and software for Education Bucharest, April 25-26, / X

Mining Student Evolution Using Associative Classification and Clustering

Unsupervised Learning of Word Semantic Embedding using the Deep Structured Semantic Model

Rule Learning With Negation: Issues Regarding Effectiveness

Rule discovery in Web-based educational systems using Grammar-Based Genetic Programming

Using Web Searches on Important Words to Create Background Sets for LSI Classification

Calibration of Confidence Measures in Speech Recognition

A survey of multi-view machine learning

The Method of Immersion the Problem of Comparing Technical Objects in an Expert Shell in the Class of Artificial Intelligence Algorithms

arxiv: v2 [cs.cv] 30 Mar 2017

Axiom 2013 Team Description Paper

Linking Task: Identifying authors and book titles in verbose queries

A Comparison of Two Text Representations for Sentiment Analysis

Learning Methods for Fuzzy Systems

Rule Learning with Negation: Issues Regarding Effectiveness

Impact of Cluster Validity Measures on Performance of Hybrid Models Based on K-means and Decision Trees

Reducing Features to Improve Bug Prediction

Lecture 1: Basic Concepts of Machine Learning

Version Space. Term 2012/2013 LSI - FIB. Javier Béjar cbea (LSI - FIB) Version Space Term 2012/ / 18

Word Segmentation of Off-line Handwritten Documents

Carnegie Mellon University Department of Computer Science /615 - Database Applications C. Faloutsos & A. Pavlo, Spring 2014.

Applications of data mining algorithms to analysis of medical data

Analysis of Emotion Recognition System through Speech Signal Using KNN & GMM Classifier

Product Feature-based Ratings foropinionsummarization of E-Commerce Feedback Comments

On-Line Data Analytics

Issues in the Mining of Heart Failure Datasets

Speech Recognition at ICSI: Broadcast News and beyond

Attributed Social Network Embedding

Artificial Neural Networks written examination

Large-Scale Web Page Classification. Sathi T Marath. Submitted in partial fulfilment of the requirements. for the degree of Doctor of Philosophy

Automating the E-learning Personalization

Beyond the Pipeline: Discrete Optimization in NLP

AUTOMATIC DETECTION OF PROLONGED FRICATIVE PHONEMES WITH THE HIDDEN MARKOV MODELS APPROACH 1. INTRODUCTION

Twitter Sentiment Classification on Sanders Data using Hybrid Approach

Business Analytics and Information Tech COURSE NUMBER: 33:136:494 COURSE TITLE: Data Mining and Business Intelligence

Conference Presentation

ScienceDirect. A Framework for Clustering Cardiac Patient s Records Using Unsupervised Learning Techniques

A study of speaker adaptation for DNN-based speech synthesis

OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS

Softprop: Softmax Neural Network Backpropagation Learning

Level 6. Higher Education Funding Council for England (HEFCE) Fee for 2017/18 is 9,250*

Data Integration through Clustering and Finding Statistical Relations - Validation of Approach

A New Perspective on Combining GMM and DNN Frameworks for Speaker Adaptation

Experiments with SMS Translation and Stochastic Gradient Descent in Spanish Text Author Profiling

Welcome to. ECML/PKDD 2004 Community meeting

Experiment Databases: Towards an Improved Experimental Methodology in Machine Learning

System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks

Speech Emotion Recognition Using Support Vector Machine

Universidade do Minho Escola de Engenharia

Given a directed graph G =(N A), where N is a set of m nodes and A. destination node, implying a direction for ow to follow. Arcs have limitations

Semi-Supervised Face Detection

Indian Institute of Technology, Kanpur

Chinese Language Parsing with Maximum-Entropy-Inspired Parser

We are strong in research and particularly noted in software engineering, information security and privacy, and humane gaming.

CS 446: Machine Learning

Matching Similarity for Keyword-Based Clustering

Statewide Framework Document for:

Article A Novel, Gradient Boosting Framework for Sentiment Analysis in Languages where NLP Resources Are Not Plentiful: A Case Study for Modern Greek

Comment-based Multi-View Clustering of Web 2.0 Items

Software Maintenance

*Net Perceptions, Inc West 78th Street Suite 300 Minneapolis, MN

Using the Attribute Hierarchy Method to Make Diagnostic Inferences about Examinees Cognitive Skills in Algebra on the SAT

Course Outline. Course Grading. Where to go for help. Academic Integrity. EE-589 Introduction to Neural Networks NN 1 EE

arxiv: v1 [math.at] 10 Jan 2016

Machine Learning from Garden Path Sentences: The Application of Computational Linguistics

Transfer Learning Action Models by Measuring the Similarity of Different Domains

Mathematics. Mathematics

STUDYING ACADEMIC INDICATORS WITHIN VIRTUAL LEARNING ENVIRONMENT USING EDUCATIONAL DATA MINING

Latent Semantic Analysis

Ensemble Technique Utilization for Indonesian Dependency Parser

Semi-supervised methods of text processing, and an application to medical concept extraction. Yacine Jernite Text-as-Data series September 17.

A Case Study: News Classification Based on Term Frequency

IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 17, NO. 3, MARCH

Learning Methods in Multilingual Speech Recognition

GACE Computer Science Assessment Test at a Glance

Detailed course syllabus

AN EXAMPLE OF THE GOMORY CUTTING PLANE ALGORITHM. max z = 3x 1 + 4x 2. 3x 1 x x x x N 2

TIMSS ADVANCED 2015 USER GUIDE FOR THE INTERNATIONAL DATABASE. Pierre Foy

Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models

Iterative Cross-Training: An Algorithm for Learning from Unlabeled Web Pages

Laboratorio di Intelligenza Artificiale e Robotica

Bug triage in open source systems: a review

A Bayesian Learning Approach to Concept-Based Document Classification

Feature Selection based on Sampling and C4.5 Algorithm to Improve the Quality of Text Classification using Naïve Bayes

Machine Learning and Data Mining. Ensembles of Learners. Prof. Alexander Ihler

Team Formation for Generalized Tasks in Expertise Social Networks

Matrices, Compression, Learning Curves: formulation, and the GROUPNTEACH algorithms

Transcription:

COMP9318 Review Wei Wang @ UNSW June 4, 2018

Course Logisitics THE formula: mark = 0.55 exam + 0.15 (ass1 + proj1 + lab) mark = FL, if exam < 40 lab = avg(best of 3(lab1, lab2, lab3, lab4, lab5)) proj1 and ass1 will be marked ASAP; we aim at delivering the result before the exam Pre-exam consultations: 15 Jun: 1500 1700, K17-508 18 Jun: 1200 1400, K17-508 Course feedback: via comments in the course survey or private messages to me on the forum. We are particularly interested in aspects such as coverage, difficulty levels, use of python/jupyter, project, and background required. Note (1) The final exam mark is important and you must achieve at least 40! (2) Supplementary exam is only for those who cannot attend the final exam.

About the Final Exam Time: 1345 1600, 19 Jun 2016 (Tue), 10 minutes reading time + 2 hr closed-book exam. Accessories: UNSW Approved Calculator. Note: watches are prohibited. Designed to test your understanding and familiarity of the core contents of the course. Answer 1 + 6 questions out of 9 questions. Q1: short answer (can use your own words) and compulsory. Choose 6 from Q2 to Q9; thers will requires some calculation (i.e., similar to tute/ass questions)

About the Final Exam /2 Read the instructions carefully. Use your time wisely. Don t spend too much time if stuck on one question or writing excessively long answers on Q1. Tips (1) Write down intermediate steps. (2) Know how to do log 2 (x) on your calculator. (3) Work on easy questions first (but start the answer on a new page on the booklet). Disclaimer We will go through the main contents of each lecture. However, note that it is by no means exhaustive.

Introduction DM vs. KDD Steps of KDD; iterative in nature; results need to be validated. Database (efficiency) vs. Machine learning (effectiveness) vs. Statistics (validity): Able to cast a real problem into a data mining problem.

Data Warehousing and OLAP Understand the four characteristics of DW (DW vs. Data Mart) Differences between OLTP and OLAP Multidimensional data model; data cube; fact, dimension, measure, hierarchies cuboid, cube lattice three types of schemas four typical OLAP operations ROLAP/MOLAP/HOLAP Query processing methods for OLAP servers, including the BUC cubing algorithm. NOT needed: Design good DW schemas and perform ETL from operational data sources to the DW tables.

Linear Algebra Column vectors; Linear combination; Basis vectors; Span Matrix vector multiplication Eigenvalues and eigenvectors SVD: general idea.

Data Preprocessing Understand that real data is dirty (incomplete, noisy, inconsistent) How to handle missing data? How to normalize the data? How to handle noisy data? different binning/histogram method (including V-optimal and MaxDiff) How to discretize data? NOT needed: Feature selection and reduction (e.g., PCA, Random Projection, t-sne)

Classification and Prediction Classification basics: overfitting/underfitting; cross-validation Classification vs prediction; vs clustering (unsupervised learning); eager learning vs. lazy learning (instance-based learning) Decision tree: The ID3 algorithm Decision tree pruning Derive rules from the decision tree The CART algorithm (with gini index) Naive Bayes classifier Smoothing Two ways to apply NB on text data Logistic regression/maxent classifier; Maximum likelihood estimation of the model parameters + regularization; Gradient ascend. SVM: Main idea; the optimization problem in the primal form; the decision function in the dual form; kernel

Cluster Analysis Clustering criteria: minimize intra-cluster distance + maximize inter-cluster distance Distance/similarity how to deal with different types of variables distance functions: Lp metric distance functions

Cluster Analysis /2 Partition-based Clustering: k-means (algorithm, advantages, disadvantages,... ) Hierarchical Clustering: agglomerative, single-link / complete-link / group average hierarchical clustering Graph-based Clustering: Unnormalized graph laplacian and its semantics, overview of spectral clustering algorithm; embedding.

Association Rule Mining Concepts: Input: transaction db Output: (1) frequent itemset (via minsup); (2) association rules (via minconf ) Apriori algorithm: Apriori property (2 versions) The Apriori algorithm How to find frequent itemsets? How to derive the association rules?

Association Rule Mining /2 FP-growth algorithm: How to mine the association rule using FP-trees? Derive association rules from the frequent itemsets.

Thanks You and Good Luck!