Understanding and Interpreting the NRC s Data-Based Assessment of Research-Doctorate Programs in the United States (2010)

Similar documents
Chapters 1-5 Cumulative Assessment AP Statistics November 2008 Gillespie, Block 4

On-the-Fly Customization of Automated Essay Scoring

Psychometric Research Brief Office of Shared Accountability

Algebra 1, Quarter 3, Unit 3.1. Line of Best Fit. Overview

Probability and Statistics Curriculum Pacing Guide

CONSTRUCTION OF AN ACHIEVEMENT TEST Introduction One of the important duties of a teacher is to observe the student in the classroom, laboratory and

NCEO Technical Report 27

Centre for Evaluation & Monitoring SOSCA. Feedback Information

Conceptual and Procedural Knowledge of a Mathematics Problem: Their Measurement and Their Causal Interrelations

Session 2B From understanding perspectives to informing public policy the potential and challenges for Q findings to inform survey design

Evidence for Reliability, Validity and Learning Effectiveness

AGS THE GREAT REVIEW GAME FOR PRE-ALGEBRA (CD) CORRELATED TO CALIFORNIA CONTENT STANDARDS

STA 225: Introductory Statistics (CT)

Race, Class, and the Selective College Experience

PROFESSIONAL TREATMENT OF TEACHERS AND STUDENT ACADEMIC ACHIEVEMENT. James B. Chapman. Dissertation submitted to the Faculty of the Virginia

Corpus Linguistics (L615)

VOL. 3, NO. 5, May 2012 ISSN Journal of Emerging Trends in Computing and Information Sciences CIS Journal. All rights reserved.

Redirected Inbound Call Sampling An Example of Fit for Purpose Non-probability Sample Design

ACTL5103 Stochastic Modelling For Actuaries. Course Outline Semester 2, 2014

Biological Sciences, BS and BA

PIRLS. International Achievement in the Processes of Reading Comprehension Results from PIRLS 2001 in 35 Countries

TIMSS ADVANCED 2015 USER GUIDE FOR THE INTERNATIONAL DATABASE. Pierre Foy

LANGUAGE DIVERSITY AND ECONOMIC DEVELOPMENT. Paul De Grauwe. University of Leuven

Lecture 1: Machine Learning Basics

A Comparison of Charter Schools and Traditional Public Schools in Idaho

learning collegiate assessment]

Accessing Higher Education in Developing Countries: panel data analysis from India, Peru and Vietnam

An Empirical Analysis of the Effects of Mexican American Studies Participation on Student Achievement within Tucson Unified School District

A Retrospective Study

ECON 365 fall papers GEOS 330Z fall papers HUMN 300Z fall papers PHIL 370 fall papers

Using Blackboard.com Software to Reach Beyond the Classroom: Intermediate

The Efficacy of PCI s Reading Program - Level One: A Report of a Randomized Experiment in Brevard Public Schools and Miami-Dade County Public Schools

Graduate Division Annual Report Key Findings

Introduction. Chem 110: Chemical Principles 1 Sections 40-52

AP Statistics Summer Assignment 17-18

PEER EFFECTS IN THE CLASSROOM: LEARNING FROM GENDER AND RACE VARIATION *

Access Center Assessment Report

Linking the Ohio State Assessments to NWEA MAP Growth Tests *

THE PENNSYLVANIA STATE UNIVERSITY SCHREYER HONORS COLLEGE DEPARTMENT OF MATHEMATICS ASSESSING THE EFFECTIVENESS OF MULTIPLE CHOICE MATH TESTS

Office of Planning and Budgets. Provost Market for Fiscal Year Resource Guide

STT 231 Test 1. Fill in the Letter of Your Choice to Each Question in the Scantron. Each question is worth 2 point.

Algebra 2- Semester 2 Review

MAT 122 Intermediate Algebra Syllabus Summer 2016

Analysis of Enzyme Kinetic Data

w o r k i n g p a p e r s

Instructor: Mario D. Garrett, Ph.D. Phone: Office: Hepner Hall (HH) 100

Empowering Students Learning Achievement Through Project-Based Learning As Perceived By Electrical Instructors And Students

Accountability in the Netherlands

CHAPTER 4: REIMBURSEMENT STRATEGIES 24

American Journal of Business Education October 2009 Volume 2, Number 7

PROMOTION and TENURE GUIDELINES. DEPARTMENT OF ECONOMICS Gordon Ford College of Business Western Kentucky University

Background Information. Instructions. Problem Statement. HOMEWORK INSTRUCTIONS Homework #3 Higher Education Salary Problem

Manipulative Mathematics Using Manipulatives to Promote Understanding of Math Concepts

How and Why Has Teacher Quality Changed in Australia?

Spring 2014 SYLLABUS Michigan State University STT 430: Probability and Statistics for Engineering

Peer Influence on Academic Achievement: Mean, Variance, and Network Effects under School Choice

MINUTE TO WIN IT: NAMING THE PRESIDENTS OF THE UNITED STATES

Purdue Data Summit Communication of Big Data Analytics. New SAT Predictive Validity Case Study

(I couldn t find a Smartie Book) NEW Grade 5/6 Mathematics: (Number, Statistics and Probability) Title Smartie Mathematics

Third Misconceptions Seminar Proceedings (1993)

GradinG SyStem IE-SMU MBA

BLACKBOARD TRAINING PHASE 2 CREATE ASSESSMENT. Essential Tool Part 1 Rubrics, page 3-4. Assignment Tool Part 2 Assignments, page 5-10

System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks

Principal vacancies and appointments

EDUCATIONAL ATTAINMENT

Evidence-based Practice: A Workshop for Training Adult Basic Education, TANF and One Stop Practitioners and Program Administrators

1 3-5 = Subtraction - a binary operation

Interpreting ACER Test Results

BENCHMARK TREND COMPARISON REPORT:

(Includes a Detailed Analysis of Responses to Overall Satisfaction and Quality of Academic Advising Items) By Steve Chatman

ABILITY SORTING AND THE IMPORTANCE OF COLLEGE QUALITY TO STUDENT ACHIEVEMENT: EVIDENCE FROM COMMUNITY COLLEGES

Educational Attainment

Grade 6: Correlated to AGS Basic Math Skills

End-of-Module Assessment Task

PROJECT MANAGEMENT AND COMMUNICATION SKILLS DEVELOPMENT STUDENTS PERCEPTION ON THEIR LEARNING

Stimulating Techniques in Micro Teaching. Puan Ng Swee Teng Ketua Program Kursus Lanjutan U48 Kolej Sains Kesihatan Bersekutu, SAS, Ulu Kinta

Capturing and Organizing Prior Student Learning with the OCW Backpack

National Collegiate Retention and. Persistence-to-Degree Rates

CENTRAL MAINE COMMUNITY COLLEGE Introduction to Computer Applications BCA ; FALL 2011

Introduction. Educational policymakers in most schools and districts face considerable pressure to

Class-Discriminative Weighted Distortion Measure for VQ-Based Speaker Identification

Introduction to Simulation

Mathematics. Mathematics

UNIT ONE Tools of Algebra

Level 1 Mathematics and Statistics, 2015

Longitudinal Analysis of the Effectiveness of DCPS Teachers

Quantitative Research Questionnaire

IS FINANCIAL LITERACY IMPROVED BY PARTICIPATING IN A STOCK MARKET GAME?

PREDISPOSING FACTORS TOWARDS EXAMINATION MALPRACTICE AMONG STUDENTS IN LAGOS UNIVERSITIES: IMPLICATIONS FOR COUNSELLING

Association Between Categorical Variables

CS177 Python Programming

Best Colleges Main Survey

PSYCHOLOGY 353: SOCIAL AND PERSONALITY DEVELOPMENT IN CHILDREN SPRING 2006

School of Innovative Technologies and Engineering

Teaching a Laboratory Section

AN ANALYSIS OF GRAMMTICAL ERRORS MADE BY THE SECOND YEAR STUDENTS OF SMAN 5 PADANG IN WRITING PAST EXPERIENCES

CHMB16H3 TECHNIQUES IN ANALYTICAL CHEMISTRY

Role Models, the Formation of Beliefs, and Girls Math. Ability: Evidence from Random Assignment of Students. in Chinese Middle Schools

Perceptions of value and value beyond perceptions: measuring the quality and value of journal article readings

Transcription:

Understanding and Interpreting the NRC s Data-Based Assessment of Research-Doctorate Programs in the United States (2010) Jaxk Reeves, SCC Director Kim Love-Myers, SCC Associate Director Presented at UGA Graduate School Workshop 10/19/2010 About the SCC The Statistical Consulting Center (SCC) is part of the Department of Statistics at UGA We offer statistical advice to on- and off-campus clients Web page: www.stat.uga.edu/consulting 2 1

Topical Outline Brief Overview NRC Excel File Explanation and Demonstration Interpreting Individual Program Results 3 Overview The NRC study covers 5,004 doctoral programs in 62 fields at 212 universities (50 programs at UGA) Provides ranking information in 59 fields (48 ranked programs at UGA) To be ranked, a field must have produced at least 500 Ph.D.s in the five academic years from 1999-2004, and a program must have produced at least 5 Ph.D.s in the five academic years from 2000-2005 4 2

Overview NRC surveys conducted about once every 15 years (1983, 1995, 2010) One goal is to evaluate (rank) PhD-granting programs in various fields Current survey has more quantifiable aspects to counter previous criticisms of being a popularity contest that unduly favored established programs 5 Overview Rankings provided as a range of possible rankings for each program overall (two methods, R and S) and in each of three dimensions (Research Activity, Student Support and Outcomes, and Diversity of the Academic Environment) According to the NRC, Users are encouraged to look at groups of programs that are in the same range as their own programs, as well as programs whose ranges are above or below, in trying to answer the question Where do we stand? 6 3

This is a summary of the NRC s methodology For a complete SCC presentation and handout on the NRC s methodology prepared in 2009, visit http://www.grad.uga.edu/nrc/index.html The NRC s published methodology guide can be found at http://www.nap.edu/rdp/, along with the complete Assessment; the methodology guide is the basis for this presentation 7 The NRC developed two methods of assigning overall ranks to programs: a regression ( R ) ranking, and a survey ( S ) ranking The survey ranking has also been modified to create the three dimensional rankings mentioned Program data collected from universities and opinions of tenure-track faculty members contributed to both ranking methods 8 4

Program data (used in both ranking methods) included information about students, faculty, and characteristics of the program Data used in the final rankings were obtained on 20 key variables from three groups: Faculty Characteristics, Student Characteristics, and Program Characteristics 9 Faculty Characteristic Variables Publications per faculty member per year [2000-2006] R Avg. citations per publication [2000-2006] R % Faculty with Grant [2005-06 AY] R % Interdisciplinary Faculty [2005-06 AY] % non-asian Minority Faculty [2005-06 AY] D % Female Faculty [2005-06 AY] D Awards/Faculty Member [2001-2006] R 10 5

Student Characteristic Variables Median GRE V or Q Score of Entering Students [2004-2006] % First-year Receiving Full Support [Fall 2005] S % First-year With External Support [Fall 2005] Average # of student pubs/presentations [not collected] % non-asian Minority Students [Fall 2005] D % Female Students [Fall 2005] D % International Students [Fall 2005] D 11 Program Characteristic Variables Average # completing PhD/year % PhDs graduating within 6 years % Median Time to PhD Degree % PhD s with Academic Positions [2004-2006] S [F96 to F01 cohorts] S [F96 to F01 cohorts] S [2001-2005] S PhD Student Work Space [-1, +1] [Fall 2005] PhD Student Health Insurance [-1, +1] [Fall 2005] Average Number of Student Support Mechanisms [Fall 2005] 12 6

The 20 variables measured for each program are in different units and scales, making it hard to compare them, since we desire to combine them into one overall score Variables were standardized according to program field so eventual importance/ weight of variables could be easily modeled and compared 13 Standardize using where x i, j z i j i j,, Avg SD = Value recorded for variable j by program i in field Avg j = Average for variable j over all programs in field SD j = Standard deviation for variable j over all programs in field x j j 14 7

A positive z-score indicates the value for a program is above average, negative indicates below average If the overall distribution of scores for a variable (over all programs in a field) is roughly mound-shaped, then about 2/3 of z-scores are in the range [-1,+1], about 95% are within [-2,+2], and values outside of [-3,+3] are very rare 15 For the S rankings, all faculty (86% response rate) were asked to identify up to 4 important variables (score of 1), and of those, up to 2 most important variables (score of 2) in each of three categories: Faculty Characteristics, Student Characteristics, and Program Characteristics They were also asked to assign weights of importance totaling 100 to each of the 3 categories 16 8

The final weight each faculty member assigned to a variable was the variable s score times the weight of importance of its category, divided by the sum of that faculty member s 20 weights (so weights would sum to 1) Within each field, the weight for a variable was the average final weight taken across all faculty members who responded x 17 For each program, the survey score could then simply be calculated as Survey score = x 1 z1 x2z2... x20z20 If no randomization is incorporated, ordering survey scores for all programs provides rankings (larger survey score implies lower-numbered rank) The NRC incorporated randomness into both the z-scores and the x weights to yield a range of possible ranks 18 9

For the R rankings, a stratified random sample of faculty members were given program information and asked to rate 15 programs in their field (30 or 50 programs included over all raters, depending on size of field, also a stratified random sample) For programs that did not receive ratings from all faculty raters (58% response rate), additional faculty members were chosen until almost every program in the field had been rated by 40 or more faculty members 19 A regression equation was then calculated in each field to relate the program variables for these programs to their ratings (principal components accounted for correlation between standardized variables z, and weights r were adjusted so absolute values of weights summed to 1): Adjusted predicted rating = 1 z1 r2 z2... r20z20 Ordering adjusted predicted ratings for all programs provides ranks if no randomness is incorporated r 20 10

Stopping here would provide one rank for each program within its field by both R and S methods To account for variability in program data (i.e., year-toyear fluctuation) and in R and S weights (since results would vary according to who answers the surveys involved), the NRC decided to incorporate random variation 21 To account for variation in R and S weights, half of the faculty raters or survey respondents were randomly selected for the weight creation process, and this was repeated 500 times To account for variation in program data, recorded values (before standardization) for each of the 20 variables were perturbed by adding a random error term from a normal distribution with a zero mean and a standard error that depended on the variable; this was also repeated 500 times 22 11

The 500 sets of R and S weights were used with the 500 sets of perturbed program data to produce 500 simulated ranking orders in each field with each method For each program, the lowest and highest 5% of rankings (i.e., 25 lowest and 25 highest) achieved in this process were removed, and the lowest and highest of the remaining 450 rankings were reported as the 5 th and 95 th percentile rankings for both R and S methods 23 To create the dimensional rankings, the variables from the survey method were analyzed separately for each dimension in each of the 500 simulations The dimensional rankings process for each dimension was otherwise similar to the overall survey ranking process 24 12

Philosophical Observations Incorporation of randomness R vs. S rankings Effects of Program Size Politically Incorrect Results 25 NRC Excel File The Excel workbook, downloadable as ResDocTableWin.xls (if using a Windows computer), contains 15 sheets 26 13

NRC Excel File The first sheet, Start, has basic instructions for using the Excel workbook Make sure to enable active content, or the workbook will not be fully functional! 27 NRC Excel File The second sheet, Guide, contains more detailed information about using the spreadsheet, as well as basic information about the data and the methodology 28 14

NRC Excel File The third sheet, Master, contains information about all programs in the study 29 NRC Excel File The Master sheet is the link to all remaining sheets Each row is associated with one program 30 15

NRC Excel File Clicking on that row will link all remaining sheets to information specific to that program (if active content is enabled!) 31 NRC Excel File Variables contains the actual program data, as well as the standardized values and descriptive ranges of fieldspecific R and S weights 32 16

NRC Excel File The next 10 sheets give 5 th and 95 th percentile information for both the R and S methods of overall ranking, as well as for the three dimensional rankings Each sheet gives the actual program data, along with the perturbed standardized data and random R or S weights that achieved that percentile ranking and how they combined 33 NRC Excel File The final sheet, Emerging Programs, provides brief data on 14 fields in 4 broad fields These programs are unranked 34 17

Interpreting Results Regression (R) rankings This school did well in its field with its.05 R ranking Large positive numbers in (Col 6) show why program received a high (low-numbered) ranking 35 Interpreting Results Regression (R) rankings For this same school,.95 R still ranked highly Same variables counted less, however, due to variation 36 18

Interpreting Results Survey (S) rankings This school did not compete as well as the previous one Large negative numbers in (Col 6) demonstrate why the school received a low (high-numbered) ranking in its.05 S ranking 37 Interpreting Results Survey (S) rankings For this same school, the.95 S ranking was affected by similar areas as the.05 S ranking, but random variation resulted in a highernumbered ranking 38 19

Interpreting Results To find the variables in R or S rankings that positively influenced your program s rankings, look for large (in absolute value) standardized program values and large coefficients of the same sign (+/-) To find the variables that negatively influenced your program s rankings, look for large (in absolute value) standardized program values and large coefficients of the opposite sign Important areas are those with large coefficients Compare your program to others in your range of rankings, and others that received ranges outside of your own (both lower and higher) 39 Interpreting Results Dimensional Rankings Same basic idea as S rankings, but more specific concentration of variables High (positive) values in (Col 6) show where a program did well 40 20

Interpreting Results Of the dimensional rankings, Research Activity had the most impact on overall rankings, and Diversity of the Academic Environment had the least However, programs shouldn t disregard these areas while the NRC s R and S rankings may not have depended highly on certain variables, students may value them highly when deciding on a program 41 Questions? 42 21