GLMs the Good, the Bad, and the Ugly Midwest Actuarial Forum 23 March Christopher Cooksey, FCAS, MAAA EagleEye Analytics

Similar documents
Lecture 1: Machine Learning Basics

Self Study Report Computer Science

Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition

Machine Learning and Data Mining. Ensembles of Learners. Prof. Alexander Ihler

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

(Sub)Gradient Descent

WE GAVE A LAWYER BASIC MATH SKILLS, AND YOU WON T BELIEVE WHAT HAPPENED NEXT

CS Machine Learning

Python Machine Learning

Assignment 1: Predicting Amazon Review Ratings

Instructor: Mario D. Garrett, Ph.D. Phone: Office: Hepner Hall (HH) 100

Machine Learning and Development Policy

Probability and Statistics Curriculum Pacing Guide

Evidence for Reliability, Validity and Learning Effectiveness

Ryerson University Sociology SOC 483: Advanced Research and Statistics

Further, Robert W. Lissitz, University of Maryland Huynh Huynh, University of South Carolina ADEQUATE YEARLY PROGRESS

A Case Study: News Classification Based on Term Frequency

Purdue Data Summit Communication of Big Data Analytics. New SAT Predictive Validity Case Study

Active Learning. Yingyu Liang Computer Sciences 760 Fall

On-Line Data Analytics

Probabilistic Latent Semantic Analysis

Major Milestones, Team Activities, and Individual Deliverables

School Size and the Quality of Teaching and Learning

12- A whirlwind tour of statistics

EDCI 699 Statistics: Content, Process, Application COURSE SYLLABUS: SPRING 2016

Cognitive Thinking Style Sample Report

STA 225: Introductory Statistics (CT)

The Good Judgment Project: A large scale test of different methods of combining expert predictions

A student diagnosing and evaluation system for laboratory-based academic exercises

Analysis of Emotion Recognition System through Speech Signal Using KNN & GMM Classifier

learning collegiate assessment]

Probability estimates in a scenario tree

ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY DOWNLOAD EBOOK : ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY PDF

PM tutor. Estimate Activity Durations Part 2. Presented by Dipo Tepede, PMP, SSBB, MBA. Empowering Excellence. Powered by POeT Solvers Limited

Rule Learning With Negation: Issues Regarding Effectiveness

4.0 CAPACITY AND UTILIZATION

Learning From the Past with Experiment Databases

SACS Reaffirmation of Accreditation: Process and Reports

Hardhatting in a Geo-World

AGS THE GREAT REVIEW GAME FOR PRE-ALGEBRA (CD) CORRELATED TO CALIFORNIA CONTENT STANDARDS

Research Design & Analysis Made Easy! Brainstorming Worksheet

Evaluation of Teach For America:

Model Ensemble for Click Prediction in Bing Search Ads

EGRHS Course Fair. Science & Math AP & IB Courses

Multi-label classification via multi-target regression on data streams

Lecture 1: Basic Concepts of Machine Learning

Generative models and adversarial training

Hierarchical Linear Modeling with Maximum Likelihood, Restricted Maximum Likelihood, and Fully Bayesian Estimation

DRAFT VERSION 2, 02/24/12

Acquiring Competence from Performance Data

COMMUNICATION STRATEGY FOR THE IMPLEMENTATION OF THE SYSTEM OF ENVIRONMENTAL ECONOMIC ACCOUNTING. Version: 14 November 2017

Measurement. When Smaller Is Better. Activity:

State University of New York at Buffalo INTRODUCTION TO STATISTICS PSC 408 Fall 2015 M,W,F 1-1:50 NSC 210

Exploration. CS : Deep Reinforcement Learning Sergey Levine

Quantitative analysis with statistics (and ponies) (Some slides, pony-based examples from Blase Ur)

Alpha provides an overall measure of the internal reliability of the test. The Coefficient Alphas for the STEP are:

Success Factors for Creativity Workshops in RE

Course Outline. Course Grading. Where to go for help. Academic Integrity. EE-589 Introduction to Neural Networks NN 1 EE

*Net Perceptions, Inc West 78th Street Suite 300 Minneapolis, MN

The lab is designed to remind you how to work with scientific data (including dealing with uncertainty) and to review experimental design.

ENME 605 Advanced Control Systems, Fall 2015 Department of Mechanical Engineering

Sociology 521: Social Statistics and Quantitative Methods I Spring Wed. 2 5, Kap 305 Computer Lab. Course Website

Carolina Course Evaluation Item Bank Last Revised Fall 2009

Teacher Quality and Value-added Measurement

Conceptual and Procedural Knowledge of a Mathematics Problem: Their Measurement and Their Causal Interrelations

NCEO Technical Report 27

Word Segmentation of Off-line Handwritten Documents

Practical Research. Planning and Design. Paul D. Leedy. Jeanne Ellis Ormrod. Upper Saddle River, New Jersey Columbus, Ohio

Essentials of Ability Testing. Joni Lakin Assistant Professor Educational Foundations, Leadership, and Technology

Multiple regression as a practical tool for teacher preparation program evaluation

Value Creation Through! Integration Workshop! Value Stream Analysis and Mapping for PD! January 31, 2002!

Knowledge based expert systems D H A N A N J A Y K A L B A N D E

Designing a Rubric to Assess the Modelling Phase of Student Design Projects in Upper Year Engineering Courses

A cognitive perspective on pair programming

AGENDA LEARNING THEORIES LEARNING THEORIES. Advanced Learning Theories 2/22/2016

Consultation skills teaching in primary care TEACHING CONSULTING SKILLS * * * * INTRODUCTION

VOL. 3, NO. 5, May 2012 ISSN Journal of Emerging Trends in Computing and Information Sciences CIS Journal. All rights reserved.

Semantic Segmentation with Histological Image Data: Cancer Cell vs. Stroma

Foothill College Summer 2016

Green Belt Curriculum (This workshop can also be conducted on-site, subject to price change and number of participants)

Rule Learning with Negation: Issues Regarding Effectiveness

The Talent Development High School Model Context, Components, and Initial Impacts on Ninth-Grade Students Engagement and Performance

Learning Disability Functional Capacity Evaluation. Dear Doctor,

Mater Dei Institute of Education A College of Dublin City University

The Effect of Written Corrective Feedback on the Accuracy of English Article Usage in L2 Writing

Chapters 1-5 Cumulative Assessment AP Statistics November 2008 Gillespie, Block 4

SOFTWARE EVALUATION TOOL

ReFresh: Retaining First Year Engineering Students and Retraining for Success

Briefing document CII Continuing Professional Development (CPD) scheme.

PROGRAMME SPECIFICATION UWE UWE. Taught course. JACS code. Ongoing

Certified Six Sigma Professionals International Certification Courses in Six Sigma Green Belt

Reinforcement Learning by Comparing Immediate Reward

1.11 I Know What Do You Know?

WHEN THERE IS A mismatch between the acoustic

MARKETING MANAGEMENT II: MARKETING STRATEGY (MKTG 613) Section 007

Learning or lurking? Tracking the invisible online student

A process by any other name

A Note on Structuring Employability Skills for Accounting Students

Introduction. 1. Evidence-informed teaching Prelude

ACTL5103 Stochastic Modelling For Actuaries. Course Outline Semester 2, 2014

Transcription:

Midwest Actuarial Forum 23 March 2009 Christopher Cooksey, FCAS, MAAA EagleEye Analytics

Agenda 1.A Brief History of GLMs 2.The Good what GLMs do well 3.The Bad what GLMs don t do well 4.The Ugly what GLMs can t do 5.Solutions 2

Section 1 GLM History 3

A Brief History of GLMs Formulated by Nelder and Wedderburn in 1972. First edition of McCullagh/Nelder book on GLMs in 1983. One of the first examples of use in insurance was Statistical Motor Rating: making effective use of your data by Brockman and Wright in 1992. Practitioner s Guide to Generalized Linear Models written in 2007. 4

Section 2 The Good what GLMs do well 5

The Good what GLMs do well There is an established and understood literature. There is increasing DOI acceptance. There are readily available software solutions. GLMs extrapolate over predictor levels with little or no data. GLMs provide easily calculated relativities to use as a rate classification system. GLMs clearly find significant signal in insurance data. 6

The Good what GLMs do well GLMs are parametric and come with all the advantages of parametric approaches. By assuming you know the form of the noise you can do statistical inference to evaluate predictors. You can also provide confidence intervals to communicate the inherent uncertainty in the output. Parametric approaches are very accurate when the assumptions hold. 7

Section 3 The Bad what GLMs don t do well 8

The Bad what GLMs don t do well The assumptions underlying GLMs may not hold. Investigating this issue takes time, as do corrections to the basic assumptions (if necessary). Issues include Appropriateness of the link function Appropriateness of the error function Predictiveness of the model 9

The Bad what GLMs don t do well One assumption is that the log link works well for insurance data. This can be tested with a Box Cox Transformation (an example of this can be found in the Practitioner s Guide ). Use the following link function. g(x) = (x λ -1)/ λ when λ 0 g(x) = ln(x) when λ = 0 10

Taken from A Practitioner s Guide to Generalized Linear Models, Third Edition, page 59. 11

Taken from A Practitioner s Guide to Generalized Linear Models, Third Edition, page 60. 12

The Bad what GLMs don t do well One assumption is that the log link works well for insurance data. There is no fundamental reason that real insurance data must be multiplicative in nature. Usually, a multiplicative model is more appropriate than other model forms. Consequently, multiplicative models are used. This is usually counted as a minor distortion. 13

The Bad what GLMs don t do well A second assumption is that the typical error functions (Poisson and gamma) work well for insurance data. This can be tested by looking at the residuals. Many things can be done to correct for patterns in residuals, but rarely, if ever, do you have perfectly homogeneous residuals. Sometimes you can correct for known distortions (zeroinflated Poisson, for example). These issues are usually counted as minor distortions. 14

15

The Bad what GLMs don t do well The predictiveness of the model is an additional assumption that sometimes isn t considered. Most of our time is spent on significance testing predictors against the training data. This can lead to the erroneous assumption that if a model tests well on training data that is will also validate well on hold-out data. Significance testing alone tends to overfit models. Not a fundamental issue with GLMs, but a common problem instead. 16

The Bad what GLMs don t do well The final category of issues with GLMs revolves around the time and effort involved in doing them well. While all advanced modeling techniques require knowledgeable practitioners, GLMs involve an extensive modeling process. Modelers fluent with the details are required to shepherd the modeling process itself. Also, the relatively numerous and statistically strong assumptions of GLMs require evaluation throughout. 17

Section 4 The Ugly what GLMs can t do 18

The Ugly what GLMs can t do GLM model risk can be mitigated but not removed. There is no theoretical reason that any given error function should fit insurance data precisely. There is no theoretical reason that signal in insurance data must be related to multiplicative predictors. There is always some risk that the imperfections of the model assumptions will substantively impact results. 19

The Ugly what GLMs can t do GLMs simply do not provide a system for finding all of the relevant interactions. It is not practically possible to test through trial and error all possible combinations of three-way interactions, let alone interactions involving four, five or more predictors. The key question is: Do three, four, five-way interactions exist in some way that meaningfully impacts results? Our research says they do. An example will follow. 20

The Ugly what GLMs can t do Another problem with interactions is that GLMs are not formulated to find local interactions. GLMs use global interactions the interaction between all levels of two predictors. Once this interaction is included, it is possible to note relevant portions and to smooth over irrelevant portions, thus creating local interactions between only certain levels of each predictor. This process is only practical for simple interactions. 21

Section 5 Solutions 22

Solutions Keeping in mind a realistic view of GLMs, there are at least three possible responses. 1. Continue to rely solely on GLMs 2. Abandon GLMs for some other alternative 3. Find some supplement to cover for GLMs' weaknesses 23

Solutions If you stick with GLMs, remember the difficulties 1. GLMs are parametric. Model assumptions impact the results. Make sure you test the assumptions and consider alternatives to the typical Poisson/frequency and gamma/severity combinations. 2. GLMs provide no good way to explore the universe of possible interactions. Make sure you set aside time to find these. Use intuition and scan your competitors for options. Also look for where your model is out of balance where observed losses are not close to predicted losses for significant segments of the book of business. 24

Solutions If you abandon GLMs, what else is there? Data mining techniques Minimum bias General Iteration Algorithms (Fu, Wu, 2007) Something else??? 25

Solutions A third approach is to find a supplement to GLMs. Again, consider the difficulties 1. GLMs are parametric. Model assumptions impact the results. 2. GLMs have no good way to explore the universe of possible interactions. All you need to find is a nonparametric, nonlinear approach which quickly finds relevant local interactions. 26

Solutions What possible candidates exist for accomplishing this? There are many nonparametric approaches and other tools to be found in the fields of data mining and machine learning 27

Solutions Some issues in developing a solution include Getting the technical expertise in nonparametric solutions. One-size-fits-all data mining methods have shown moderate performance on insurance-specific data. Better results are found by ensembling multiple methods. Nonparametric methods tend to be greedy significant risk of overfitting. 28

Example Medium size auto portfolio Pure premiums set by consulting actuaries using GLM tools on 2 years of data Used machine learning inductive techniques on the first year of data (training dataset) to search the residuals of the GLMs The second year of data was used as validation Used policy attributes only [Example pulled from Machine Learning vs. Traditional Methods Summary Document by Dr. Paul Beinat. Paper can be found at the EagleEye Analytics website Resources, Papers.] 29

Example Segment Exposure GLM Total Premium Claims Claim Count Training LR Validation LR 1 40,088 9,677,889 7,223,230 5,730 75% 82% 8 26,642 8,770,620 7,454,508 4,717 85% 87% 3 35,946 8,036,238 7,298,945 5,178 91% 84% 4 20,954 6,699,637 6,353,455 3,664 95% 88% 6 26,212 6,754,957 6,534,512 4,127 97% 104% 0 29,558 7,868,872 8,109,686 5,018 103% 98% 9 20,049 5,636,667 5,935,182 3,576 105% 109% 2 33,043 10,830,010 11,614,780 6,287 107% 113% 7 23,203 8,181,896 10,125,938 4,356 124% 117% 5 30,163 7,419,663 9,590,068 5,081 129% 124% 30

Example Correlation between training and validation loss ratios: 0.93 Lift is similar between the two datasets: There is a similar difference between maximum and minimum There is a similar exposure-weighted standard deviation of the loss ratios: Training is 16.4%; Validation is 14.5% Significant improvement in fit on the validation data when machine learning Deviance results are Squared combined error with Chi squared the GLM. error GLM Premiums 34.15388 1463.838 5.47708 Estimated Premiums 15.02064 243.8955 0.946702 31

Section 6 Questions? Contact Info Christopher Cooksey, FCAS, MAAA EagleEye Analytics ccooksey@eeanalytics.com www.eeanalytics.com 32