Machine Learning Opportunities and Limitations

Similar documents
Lecture 1: Machine Learning Basics

CS Machine Learning

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

Python Machine Learning

(Sub)Gradient Descent

Lecture 1: Basic Concepts of Machine Learning

Artificial Neural Networks written examination

Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition

Laboratorio di Intelligenza Artificiale e Robotica

CSL465/603 - Machine Learning

Predicting Student Attrition in MOOCs using Sentiment Analysis and Neural Networks

ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY DOWNLOAD EBOOK : ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY PDF

System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks

OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS

Laboratorio di Intelligenza Artificiale e Robotica

Machine Learning and Data Mining. Ensembles of Learners. Prof. Alexander Ihler

Australian Journal of Basic and Applied Sciences

Learning From the Past with Experiment Databases

Rule Learning With Negation: Issues Regarding Effectiveness

Probabilistic Latent Semantic Analysis

INPE São José dos Campos

Data Structures and Algorithms

Axiom 2013 Team Description Paper

Assignment 1: Predicting Amazon Review Ratings

Human Emotion Recognition From Speech

California Professional Standards for Education Leaders (CPSELs)

Impact of Cluster Validity Measures on Performance of Hybrid Models Based on K-means and Decision Trees

Lecture 10: Reinforcement Learning

Applications of data mining algorithms to analysis of medical data

MYCIN. The MYCIN Task

Deep search. Enhancing a search bar using machine learning. Ilgün Ilgün & Cedric Reichenbach

Testing A Moving Target: How Do We Test Machine Learning Systems? Peter Varhol Technology Strategy Research, USA

Twitter Sentiment Classification on Sanders Data using Hybrid Approach

Knowledge Transfer in Deep Convolutional Neural Nets

OCR for Arabic using SIFT Descriptors With Online Failure Prediction

Probability estimates in a scenario tree

Welcome to. ECML/PKDD 2004 Community meeting

Experiments with SMS Translation and Stochastic Gradient Descent in Spanish Text Author Profiling

Softprop: Softmax Neural Network Backpropagation Learning

Universidade do Minho Escola de Engenharia

STT 231 Test 1. Fill in the Letter of Your Choice to Each Question in the Scantron. Each question is worth 2 point.

The Good Judgment Project: A large scale test of different methods of combining expert predictions

Seminar - Organic Computing

A Decision Tree Analysis of the Transfer Student Emma Gunu, MS Research Analyst Robert M Roe, PhD Executive Director of Institutional Research and

Generative models and adversarial training

A Neural Network GUI Tested on Text-To-Phoneme Mapping

Machine Learning from Garden Path Sentences: The Application of Computational Linguistics

Purdue Data Summit Communication of Big Data Analytics. New SAT Predictive Validity Case Study

Rule Learning with Negation: Issues Regarding Effectiveness

Issues in the Mining of Heart Failure Datasets

Word Segmentation of Off-line Handwritten Documents

Experiment Databases: Towards an Improved Experimental Methodology in Machine Learning

COMPUTER-ASSISTED INDEPENDENT STUDY IN MULTIVARIATE CALCULUS

Computerized Adaptive Psychological Testing A Personalisation Perspective

Using focal point learning to improve human machine tacit coordination

Historical maintenance relevant information roadmap for a self-learning maintenance prediction procedural approach

Challenges in Deep Reinforcement Learning. Sergey Levine UC Berkeley

Aalya School. Parent Survey Results

Specification and Evaluation of Machine Translation Toy Systems - Criteria for laboratory assignments

Evolutive Neural Net Fuzzy Filtering: Basic Description

The 9 th International Scientific Conference elearning and software for Education Bucharest, April 25-26, / X

Abu Dhabi Indian. Parent Survey Results

Model Ensemble for Click Prediction in Bing Search Ads

Abu Dhabi Grammar School - Canada

Switchboard Language Model Improvement with Conversational Data from Gigaword

Active Learning. Yingyu Liang Computer Sciences 760 Fall

Business Analytics and Information Tech COURSE NUMBER: 33:136:494 COURSE TITLE: Data Mining and Business Intelligence

For Jury Evaluation. The Road to Enlightenment: Generating Insight and Predicting Consumer Actions in Digital Markets

Software Maintenance

CS4491/CS 7265 BIG DATA ANALYTICS INTRODUCTION TO THE COURSE. Mingon Kang, PhD Computer Science, Kennesaw State University

Proposal of Pattern Recognition as a necessary and sufficient principle to Cognitive Science

Chinese Language Parsing with Maximum-Entropy-Inspired Parser

A study of speaker adaptation for DNN-based speech synthesis

Course Outline. Course Grading. Where to go for help. Academic Integrity. EE-589 Introduction to Neural Networks NN 1 EE

TD(λ) and Q-Learning Based Ludo Players

Learning Methods for Fuzzy Systems

Exploration. CS : Deep Reinforcement Learning Sergey Levine

Knowledge based expert systems D H A N A N J A Y K A L B A N D E

Test Effort Estimation Using Neural Network

Using computational modeling in language acquisition research

Calibration of Confidence Measures in Speech Recognition

Linking Task: Identifying authors and book titles in verbose queries

IT Students Workshop within Strategic Partnership of Leibniz University and Peter the Great St. Petersburg Polytechnic University

ScienceDirect. A Framework for Clustering Cardiac Patient s Records Using Unsupervised Learning Techniques

Introduction to Simulation

Reducing Features to Improve Bug Prediction

A New Perspective on Combining GMM and DNN Frameworks for Speaker Adaptation

Comparison of EM and Two-Step Cluster Method for Mixed Data: An Application

arxiv: v2 [cs.cv] 30 Mar 2017

Analysis of Hybrid Soft and Hard Computing Techniques for Forex Monitoring Systems

Learning Optimal Dialogue Strategies: A Case Study of a Spoken Dialogue Agent for

Speech Recognition at ICSI: Broadcast News and beyond

On Human Computer Interaction, HCI. Dr. Saif al Zahir Electrical and Computer Engineering Department UBC

arxiv: v1 [cs.cl] 2 Apr 2017

Notes on The Sciences of the Artificial Adapted from a shorter document written for course (Deciding What to Design) 1

The Internet as a Normative Corpus: Grammar Checking with a Search Engine

AUTOMATED TROUBLESHOOTING OF MOBILE NETWORKS USING BAYESIAN NETWORKS

Web as Corpus. Corpus Linguistics. Web as Corpus 1 / 1. Corpus Linguistics. Web as Corpus. web.pl 3 / 1. Sketch Engine. Corpus Linguistics

Data Fusion Through Statistical Matching

Knowledge-Based - Systems

Transcription:

Machine Learning Opportunities and Limitations Holger H. Hoos LIACS Universiteit Leiden The Netherlands LCDS Conference 2017/11/28

The age of computation Clear, precise instructions flawlessly executed 1

The age of computation Clear, precise instructions flawlessly executed algorithms = recipes for data processing 1

The age of computation Clear, precise instructions flawlessly executed algorithms = recipes for data processing predictable results, behaviour 1

The age of computation Clear, precise instructions flawlessly executed algorithms = recipes for data processing predictable results, behaviour performance guarantees 1

The age of computation Clear, precise instructions flawlessly executed algorithms = recipes for data processing predictable results, behaviour performance guarantees trusted, effective solutions to complex problems 1

The age of advanced computation AI vast amounts of cheap computation 2

The age of advanced computation AI vast amounts of cheap computation automatically designed algorithms 2

The age of advanced computation AI vast amounts of cheap computation automatically designed algorithms effective but complex, heuristic, black-box methods 2

Key idea: explicit programming learning / automatic adaptation to data 3

Key idea: explicit programming learning / automatic adaptation to data Success stories: game playing (e.g., Go, poker) 3

Key idea: explicit programming learning / automatic adaptation to data Success stories: game playing (e.g., Go, poker) medical diagnosis (lung disease) 3

Key idea: explicit programming learning / automatic adaptation to data Success stories: game playing (e.g., Go, poker) medical diagnosis (lung disease) transportation (autonomous driving) 3

Key idea: explicit programming learning / automatic adaptation to data Success stories: game playing (e.g., Go, poker) medical diagnosis (lung disease) transportation (autonomous driving) energy (demand prediction and trading) 3

The Machine Learning Revolution machine learning (ML) = automatic construction of software that works well on given data 4

The Machine Learning Revolution machine learning (ML) = automatic construction of software that works well on given data ideas reach back to 1950s (Alan Turing) 4

The Machine Learning Revolution machine learning (ML) = automatic construction of software that works well on given data ideas reach back to 1950s (Alan Turing) based on statistics, mathematical optimisation 4

The Machine Learning Revolution machine learning (ML) = automatic construction of software that works well on given data ideas reach back to 1950s (Alan Turing) based on statistics, mathematical optimisation and principled experimentation (heuristic mechanisms) 4

The Machine Learning Revolution machine learning (ML) = automatic construction of software that works well on given data ideas reach back to 1950s (Alan Turing) based on statistics, mathematical optimisation and principled experimentation (heuristic mechanisms) key ingredient to artificial intelligence (AI) 4

The Machine Learning Revolution machine learning (ML) = automatic construction of software that works well on given data ideas reach back to 1950s (Alan Turing) based on statistics, mathematical optimisation and principled experimentation (heuristic mechanisms) key ingredient to artificial intelligence (AI) but: AI is more than ML 4

Supervised vs unsupervised ML unsupervised: discover patterns in data 5

Supervised vs unsupervised ML unsupervised: discover patterns in data data mining 5

Supervised vs unsupervised ML unsupervised: discover patterns in data data mining (e.g., clustering) 5

Supervised vs unsupervised ML unsupervised: discover patterns in data data mining (e.g., clustering) supervised: make predictions based on known training examples 5

Supervised vs unsupervised ML unsupervised: discover patterns in data data mining (e.g., clustering) supervised: make predictions based on known training examples statistical modelling 5

Supervised vs unsupervised ML unsupervised: discover patterns in data data mining (e.g., clustering) supervised: make predictions based on known training examples statistical modelling Key assumption: training data is representative Key assumption: of application scenario 5

Supervised vs unsupervised ML unsupervised: discover patterns in data data mining (e.g., clustering) supervised: make predictions based on known training examples statistical modelling Key assumption: training data is representative Key assumption: of application scenario other types of ML exist (e.g., semi-supervised learning, reinforcement learning) 5

Regression Example: predict plant growth for given set Example: of environmental conditions 6

Regression Example: predict plant growth for given set Example: of environmental conditions Given: set of training examples Given: = feature values + numerical outputs 6

Regression Example: predict plant growth for given set Example: of environmental conditions Given: set of training examples Given: = feature values + numerical outputs Objective: predict output for new feature values 6

Classification Example: predict whether someone takes a loan Example: based on demographic + personal financial data 7

Classification Example: predict whether someone takes a loan Example: based on demographic + personal financial data Given: set of training examples Given: = feature values + classes 7

Classification Example: predict whether someone takes a loan Example: based on demographic + personal financial data Given: set of training examples Given: = feature values + classes Objective: predict class for new feature values 7

Classification Example: predict whether someone takes a loan Example: based on demographic + personal financial data Given: set of training examples Given: = feature values + classes Objective: predict class for new feature values Important special case: binary classification Important special case: = 2 classes (e.g., yes/no) 7

Example: Binary classification with decision trees [Source: www.simafore.com] 8

Random forests (state-of-the-art method) [Source: blog.citizennet.com] 9

Key distinction: Classification procedure (classifier; model ): algorithm used for solving a classification problem e.g., decision tree 10

Key distinction: Classification procedure (classifier; model ): algorithm used for solving a classification problem e.g., decision tree Input: feature values Output: class (yes/no) 10

Key distinction: Classification procedure (classifier; model ): algorithm used for solving a classification problem e.g., decision tree Input: feature values Output: class (yes/no) Learning procedure: algorithm used for constructing a classifier e.g., C4.5 (well-known decision tree learning algorithm) 10

Key distinction: Classification procedure (classifier; model ): algorithm used for solving a classification problem e.g., decision tree Input: feature values Output: class (yes/no) Learning procedure: algorithm used for constructing a classifier e.g., C4.5 (well-known decision tree learning algorithm) Input: set of training data Output: classification procedure (decision tree) 10

Evaluation and Bias How to evaluate supervised ML algorithms? Key idea: Assess quality of predictions obtained Key idea: (e.g., from trained binary classifier) 11

Evaluation and Bias How to evaluate supervised ML algorithms? Key idea: Assess quality of predictions obtained Key idea: (e.g., from trained binary classifier) Prediction quality of binary classifiers accuracy: expected rate of misclassifications 11

Evaluation and Bias How to evaluate supervised ML algorithms? Key idea: Assess quality of predictions obtained Key idea: (e.g., from trained binary classifier) Prediction quality of binary classifiers accuracy: expected rate of misclassifications false positive rate: expected rate of incorrect yes predictions 11

Evaluation and Bias How to evaluate supervised ML algorithms? Key idea: Assess quality of predictions obtained Key idea: (e.g., from trained binary classifier) Prediction quality of binary classifiers accuracy: expected rate of misclassifications false positive rate: expected rate of incorrect yes predictions false negative rate: expected rate of incorrect no predictions 11

Evaluation and Bias How to evaluate supervised ML algorithms? Key idea: Assess quality of predictions obtained Key idea: (e.g., from trained binary classifier) Prediction quality of binary classifiers accuracy: expected rate of misclassifications false positive rate: expected rate of incorrect yes predictions false negative rate: expected rate of incorrect no predictions trade-off (weighted average; ROC curve) 11

Caution: Typically, no single correct evaluation metric evaluation metrics can introduce unfairness / bias 12

Caution: Typically, no single correct evaluation metric evaluation metrics can introduce unfairness / bias especially when training sets are unbalanced (many more no than yes cases, prevalence/lack of input feature combinations) 12

Caution: Typically, no single correct evaluation metric evaluation metrics can introduce unfairness / bias especially when training sets are unbalanced (many more no than yes cases, prevalence/lack of input feature combinations) use great care when constructing training sets 12

Caution: Typically, no single correct evaluation metric evaluation metrics can introduce unfairness / bias especially when training sets are unbalanced (many more no than yes cases, prevalence/lack of input feature combinations) use great care when constructing training sets use multiple evaluation metrics 12

Caution: Typically, no single correct evaluation metric evaluation metrics can introduce unfairness / bias especially when training sets are unbalanced (many more no than yes cases, prevalence/lack of input feature combinations) use great care when constructing training sets use multiple evaluation metrics perform detailed evaluations (beyond simple metrics) 12

The problem of overfitting good performance on training data may not generalise to previously unseen data overfitting (well-known problem) 13

The problem of overfitting good performance on training data may not generalise to previously unseen data overfitting (well-known problem) detect overfitting using validation techniques hold-out validation: evaluate on set of test cases hold-out validation: strictly separate from training set 13

The problem of overfitting good performance on training data may not generalise to previously unseen data overfitting (well-known problem) detect overfitting using validation techniques hold-out validation: evaluate on set of test cases hold-out validation: strictly separate from training set cross-validation: like hold-out, but with many different cross-validation: training/test splits 13

The problem of overfitting good performance on training data may not generalise to previously unseen data overfitting (well-known problem) detect overfitting using validation techniques hold-out validation: evaluate on set of test cases hold-out validation: strictly separate from training set cross-validation: like hold-out, but with many different cross-validation: training/test splits prevent overfitting using regularisation techniques (= modification / specific setting of ML method used) 13

The problem of overfitting good performance on training data may not generalise to previously unseen data overfitting (well-known problem) detect overfitting using validation techniques hold-out validation: evaluate on set of test cases hold-out validation: strictly separate from training set cross-validation: like hold-out, but with many different cross-validation: training/test splits prevent overfitting using regularisation techniques (= modification / specific setting of ML method used) Caution: Overfitting can introduce bias! 13

Problematic features certain (input) features can help improve performance, but are inappropriate to use 14

Problematic features certain (input) features can help improve performance, but are inappropriate to use examples: 14

Problematic features certain (input) features can help improve performance, but are inappropriate to use examples: race, gender, sexual orientation 14

Problematic features certain (input) features can help improve performance, but are inappropriate to use examples: race, gender, sexual orientation using problematic features in machine learning can cause (unintentional) discrimination 14

Problematic features certain (input) features can help improve performance, but are inappropriate to use examples: race, gender, sexual orientation using problematic features in machine learning can cause (unintentional) discrimination Easy solution: do not use problematic features 14

Problematic features certain (input) features can help improve performance, but are inappropriate to use examples: race, gender, sexual orientation using problematic features in machine learning can cause (unintentional) discrimination Easy solution: do not use problematic features Wrong!! combinations of other, harmless features can yield equivalent information 14

Problematic features certain (input) features can help improve performance, but are inappropriate to use examples: race, gender, sexual orientation using problematic features in machine learning can cause (unintentional) discrimination Easy solution: do not use problematic features Wrong!! combinations of other, harmless features can yield equivalent information especially problematic for deep learning and other, powerful black-box methods 14

Problematic features certain (input) features can help improve performance, but are inappropriate to use examples: race, gender, sexual orientation using problematic features in machine learning can cause (unintentional) discrimination Easy solution: do not use problematic features Wrong!! combinations of other, harmless features can yield equivalent information especially problematic for deep learning and other, powerful black-box methods Better solution: careful, detailed evaluation 14

Explainability & Transparency Challenge: How can we trust an ML system? 15

Explainability & Transparency Challenge: How can we trust an ML system? carefully evaluate performance; identify strengths and weaknesses (requires detailed evaluation = computational experiments) 15

Explainability & Transparency Challenge: How can we trust an ML system? carefully evaluate performance; identify strengths and weaknesses (requires detailed evaluation = computational experiments) understand how it works 15

Explainability & Transparency Challenge: How can we trust an ML system? carefully evaluate performance; identify strengths and weaknesses (requires detailed evaluation = computational experiments) understand how it works understand its output 15

Key distinction: understanding a classifier (e.g., decision tree) vs understanding the training procedure that produced it 16

Key distinction: understanding a classifier (e.g., decision tree) vs understanding the training procedure that produced it Note: to understand a given classifier (and its output), we do not need to understand how it was built 16

Key distinction: understanding a classifier (e.g., decision tree) vs understanding the training procedure that produced it Note: to understand a given classifier (and its output), we do not need to understand how it was built understanding of what happens at every step does not mean understanding behaviour of an algorithm 16

Key distinction: understanding a classifier (e.g., decision tree) vs understanding the training procedure that produced it Note: to understand a given classifier (and its output), we do not need to understand how it was built understanding of what happens at every step does not mean understanding behaviour of an algorithm some classifiers are easier to understand than others 16

Neural networks [Source: www.texsample.net] 17

Deep learning uses neural networks with many layers 18

Deep learning uses neural networks with many layers AlphaGo Zero: 84 layers 18

Deep learning uses neural networks with many layers AlphaGo Zero: 84 layers idea + research dates back to 1960s/1970s 18

Deep learning uses neural networks with many layers AlphaGo Zero: 84 layers idea + research dates back to 1960s/1970s successful real-world applications since the 1980s 18

Deep learning uses neural networks with many layers AlphaGo Zero: 84 layers idea + research dates back to 1960s/1970s successful real-world applications since the 1980s very popular since 2012 18

Deep learning uses neural networks with many layers AlphaGo Zero: 84 layers idea + research dates back to 1960s/1970s successful real-world applications since the 1980s very popular since 2012 impressive results in increasing number of application areas 18

Deep learning uses neural networks with many layers AlphaGo Zero: 84 layers idea + research dates back to 1960s/1970s successful real-world applications since the 1980s very popular since 2012 impressive results in increasing number of application areas requires large amounts of data, specialised hardware, considerable human expertise + experimentation 18

Deep learning uses neural networks with many layers AlphaGo Zero: 84 layers idea + research dates back to 1960s/1970s successful real-world applications since the 1980s very popular since 2012 impressive results in increasing number of application areas requires large amounts of data, specialised hardware, considerable human expertise + experimentation Caution! Deep learning machine learning AI 18

Deep neural networks are black-box methods easy to understand function of each neuron in the network; very hard / impossible to understand the behaviour of the network 19

Deep neural networks are black-box methods easy to understand function of each neuron in the network; very hard / impossible to understand the behaviour of the network lack of transparency / explainability 19

Deep neural networks are black-box methods easy to understand function of each neuron in the network; very hard / impossible to understand the behaviour of the network lack of transparency / explainability Possible remedies: principled, detailed evaluation of behaviour 19

Deep neural networks are black-box methods easy to understand function of each neuron in the network; very hard / impossible to understand the behaviour of the network lack of transparency / explainability Possible remedies: principled, detailed evaluation of behaviour use alternate methods with similar performance (e.g., random forests) 19

Deep neural networks are black-box methods easy to understand function of each neuron in the network; very hard / impossible to understand the behaviour of the network lack of transparency / explainability Possible remedies: principled, detailed evaluation of behaviour use alternate methods with similar performance (e.g., random forests) trade off performance against explainability 19

Deep neural networks are black-box methods easy to understand function of each neuron in the network; very hard / impossible to understand the behaviour of the network lack of transparency / explainability Possible remedies: principled, detailed evaluation of behaviour use alternate methods with similar performance (e.g., random forests) trade off performance against explainability frugal learning (new research direction) 19

Automated Machine Learning Machine learning is powerful 20

Automated Machine Learning Machine learning is powerful, but successful application is far from trivial. 20

Automated Machine Learning Machine learning is powerful, but successful application is far from trivial. Fundamental problem: Which of many available algorithms (models) applicable to given machine learning problem to use, and with which hyper-parameter settings? 20

Automated Machine Learning Machine learning is powerful, but successful application is far from trivial. Fundamental problem: Which of many available algorithms (models) applicable to given machine learning problem to use, and with which hyper-parameter settings? Example: WEKA contains 39 classification algorithms, Example: 3 8 feature selection methods 20

Automated Machine Learning Machine learning is powerful, but successful application is far from trivial. Fundamental problem: Which of many available algorithms (models) applicable to given machine learning problem to use, and with which hyper-parameter settings? Solution: Automatically select ML methods and hyper-parameter settings 20

Automated Machine Learning Machine learning is powerful, but successful application is far from trivial. Fundamental problem: Which of many available algorithms (models) applicable to given machine learning problem to use, and with which hyper-parameter settings? Solution: Automatically select ML methods and hyper-parameter settings Automated machine learning (AutoML) 20

AutoML... achieves substantial performance improvements over solutions hand-crafted by human experts 21

AutoML... achieves substantial performance improvements over solutions hand-crafted by human experts enables frugal learning (explainable/transparent ML) 21

AutoML... achieves substantial performance improvements over solutions hand-crafted by human experts enables frugal learning (explainable/transparent ML) helps non-experts effectively apply ML techniques 21

AutoML... achieves substantial performance improvements over solutions hand-crafted by human experts enables frugal learning (explainable/transparent ML) helps non-experts effectively apply ML techniques intense international research focus (academia + industry) 21

AutoML... achieves substantial performance improvements over solutions hand-crafted by human experts enables frugal learning (explainable/transparent ML) helps non-experts effectively apply ML techniques intense international research focus (academia + industry) ongoing research focus at LIACS (Leiden Institute of Advanced Computer Science); see ada.liacs.nl/projects, Auto-WEKA. 21

Take-Home Message Machine learning can (help to) solve many proplems 22

Take-Home Message Machine learning can (help to) solve many proplems... but is no panacea. 22

Take-Home Message Machine learning can (help to) solve many proplems... but is no panacea. Methods and results strongly depend on quantity + quality of input data. 22

Take-Home Message Machine learning can (help to) solve many proplems... but is no panacea. Methods and results strongly depend on quantity + quality of input data. Challenges: Risk of overfitting training data, hidden bias 22

Take-Home Message Machine learning can (help to) solve many proplems... but is no panacea. Methods and results strongly depend on quantity + quality of input data. Challenges: Risk of overfitting training data, hidden bias Lack of transparency, explainability 22

Take-Home Message Machine learning can (help to) solve many proplems... but is no panacea. Methods and results strongly depend on quantity + quality of input data. Challenges: Risk of overfitting training data, hidden bias Lack of transparency, explainability Human expertise: crucial for successful, responsible use 22

Take-Home Message Machine learning can (help to) solve many proplems... but is no panacea. Methods and results strongly depend on quantity + quality of input data. Challenges: Risk of overfitting training data, hidden bias Lack of transparency, explainability Human expertise: crucial for successful, responsible use Current + future research (far from solved) 22

Take-Home Message Machine learning can (help to) solve many proplems... but is no panacea. Methods and results strongly depend on quantity + quality of input data. Challenges: Risk of overfitting training data, hidden bias Lack of transparency, explainability Human expertise: crucial for successful, responsible use Current + future research (far from solved) AI should augment, not replace human expertise! (Likewise for machine learning.) 22