Utilizing Predictive Modeling to Improve Policy through Improved - - PowerPoint PPT Presentation

utilizing predictive modeling to improve policy through
SMART_READER_LITE
LIVE PREVIEW

Utilizing Predictive Modeling to Improve Policy through Improved - - PowerPoint PPT Presentation

Utilizing Predictive Modeling to Improve Policy through Improved Targeting of Agency Resources: A Case Study on Placement Instability among Foster Children Dallas J. Elgin, Ph.D. IMPAQ International Randi Walters, Ph.D. Casey Family Programs


slide-1
SLIDE 1

Utilizing Predictive Modeling to Improve Policy through Improved Targeting of Agency Resources: A Case Study on Placement Instability among Foster Children

Dallas J. Elgin, Ph.D. IMPAQ International Randi Walters, Ph.D. Casey Family Programs

2016 APPAM Fall Research Conference

Image Credit: The Strengths Initiative

slide-2
SLIDE 2
  • Ch

Challenge: Government agencies operate in an environment that increasingly requires using limited resources to meet nearly limitless demands.

  • Op

Opportu tunity ty: Advances in computing technology & administrative data can be leveraged via predictive modeling to predict the likelihood of future events

  • Go

Goal: : To provide an improved understanding of the methodology & identify associated best practices

The Utility of Predictive Modeling for Government Agencies

slide-3
SLIDE 3
  • Process of selecting a model that best predicts the probability of an
  • utcome (Geisser, 1993), or generating an accurate prediction (Kuhn

& Johnson, 2013).

  • Over the past several decades, predictive modeling has been utilized

in a variety of fields to predict diverse outcomes

  • Within child welfare, predictive models have been used to inform

decision-making:

– Risk assessment instruments – Maltreatment recurrence, future involvement, child fatalities

What is Predictive Modeling?

slide-4
SLIDE 4
  • Data:

– 2013 Adoption and Foster Care Analysis and Reporting System (AFCARS)

  • Publicly-available dataset resembling administrative data

– Sample: 15,000 foster care children that were in care throughout 2013

  • Operationalization: 3 or more moves, or a total of 4 placements

(Hartnett, Falconnier, Leathers & Testa, 1999; Webster, Barth & Needell, 2000)

– 11,649 children with 3 or fewer placements – 3,351 children with 4 or more placements

Case: Placement Instability

slide-5
SLIDE 5

Methodological Approach: Data Partition Strategy

  • The entire dataset of 15,000 children was split into 2 groups:

– A training set used to train the models (75% of dataset= 11,250 children) – A test set used to validate the models (25% of dataset= 3,750 children)

slide-6
SLIDE 6

Methodological Approach: Data Training Strategy

  • Train a collection of 10 models using the training set
  • Utilize ROC Curves to evaluate how well the models calculate:
  • 1. The true-positive rate (sensitivity)
  • 2. The false-positive rate (specificity)

Mo Model Type Mo Model In Interpretability Co Computation Time Linear Discriminant Analysis Logistic Regression High Low Partial Least Squares Discriminant Analysis High Low Elastic Net/Lasso High Low Non-Linear Classification Models K-Nearest Neighbors Low High Neural Networks Low High Support Vector Machines Low High Multivariate Adaptive Regression Splines Moderate Moderate Classification Trees & Rule-Based Models Classification Tree High High Boosted Trees Low High Random Forest Low High

slide-7
SLIDE 7

Model Performance on the Test Set

  • 3 models with highest ROC scores were applied to the test set

(3,750 observations)

  • Overall Accuracy= 87.8% - 87.9%

– Less than 3 Moves= 90.1% - 90.2% – 4 or More Moves= 77.4% - 77.8%

Ne Neural Ne Network Model 4 or More Moves Less than 3 Moves 4 or More Moves 535 153 Less than 3 Moves 302 2,759 Ra Random Forest Model 4 or More Moves Less than 3 Moves 4 or More Moves 537 157 Less than 3 Moves 300 2,755 Bo Boosted Tree Model 4 or More Moves Less than 3 Moves 4 or More Moves 540 158 Less than 3 Moves 297 2,754

slide-8
SLIDE 8

Improving Model Accuracy

  • Iterative process involving: transforming variables, ‘fine-tuning’

model parameters, or combination of both

  • ‘Fine-Tuning’ parameters of the neural network model
  • Improved Overall Accuracy= 88.2%

Un Un-tu tuned Neural Netw twork Model 4 or More Moves Less than 3 Moves 4 or More Moves 535 153 Less than 3 Moves 302 2,759 Tu Tuned Neural Network Model 4 or More Moves Less than 3 Moves 4 or More Moves 569 176 Less than 3 Moves 268 2,736

slide-9
SLIDE 9

Improving Model Accuracy: Cost-Sensitive Tuning

  • Considerable improvements in reducing false negatives, but at

expense of notable increases in the number of false positives.

Cl Classification Tree with No Co Cost Penalty 4 or More Moves Less than 3 Moves Sensitivity Specificity 4 or More Moves 515 181 0.615 0.938 Less than 3 Moves 322 2,731 Cl Classification Tree with Co Cost Penalty of 2 4 or More Moves Less than 3 Moves Sensitivity Specificity 4 or More Moves 620 354 0.741 0.878 Less than 3 Moves 217 2,558 Cl Classification Tree with Co Cost Penalty of 5 4 or More Moves Less than 3 Moves Sensitivity Specificity 4 or More Moves 756 758 0.903 0.740 Less than 3 Moves 81 2,154 Cl Classification Tree with Co Cost Penalty of 10 4 or More Moves Less than 3 Moves Sensitivity Specificity 4 or More Moves 790 970 0.944 0.667 Less than 3 Moves 47 1,942 Cl Classification Tree with Co Cost Penalty of 20 4 or More Moves Less than 3 Moves Sensitivity Specificity 4 or More Moves 803 1,161 0.959 0.601 Less than 3 Moves 34 1,751

slide-10
SLIDE 10

1. Predictive Models Can Improve Upon, but Not Replace, Traditional Decision-Making Processes within Government Agencies. 2. Government Agencies Should Clearly Articulate the Methodological Approach and the Predictive Accuracy of their Models. 3. Consider Opportunities for Incorporating Community Engagement into the Predictive Modeling Process.

Best Practices for Designing & Implementing Predictive Models

slide-11
SLIDE 11

~Questions & Feedback~

Dallas Elgin, Research Associate IMPAQ International delgin@impaqint.com Randi Walters, Senior Director of Knowledge Management Casey Family Programs

slide-12
SLIDE 12
  • What is it?

Occurs when a child in the care of a child welfare system experiences multiple moves to different settings

  • Why does it matter?

Placement instability can have significant consequences on children:

– Greater risk for impaired development & psychosocial well-being – Greater uncertainty surrounding a child’s future – Greater likelihood of re-entry and/or emancipation

  • Is it a big issue?

25% of foster care children experience three or moves while in care (Doyle, 2007)

Placement Instability

slide-13
SLIDE 13

Improving Model Accuracy: Cost-Sensitive Tuning

  • False-negative predictions may be unacceptable as a failure to

correctly identify placement instability could result in unnecessary exposure to adverse events

  • Cost-sensitive models impose cost penalties to minimize the

likelihood of false predictions

slide-14
SLIDE 14

Data

  • 2013 Adoption and Foster Care Analysis and Reporting System

(AFCARS): Federal data provided by the states on all children in foster care

  • Sample: 15,000 foster care children that were in care throughout

2013

77.66% of Children in the Sample have 3 or fewer moves 22.34% of Children in the Sample have 4 or more moves

slide-15
SLIDE 15

3 Highest Performing Models on the Training Set

  • Boosted Trees: build upon traditional classification tree models

– Fit a series of independent decision trees and then aggregate the trees to form a single predictive model

  • Random Forests: build upon traditional classification tree models by

utilizing bootstrapping methods to build a collection of decision trees

– Consideration of smaller subset of predictors minimizes the likelihood of a high degree of correlation among multiple trees

  • Neural Networks: resemble the physiological structure of the

human brain or nervous system

– Use multiple layers (or algorithms) for processing pieces of information

slide-16
SLIDE 16

Linear Discriminant Analysis Models

  • Utilize linear functions to categorize observations into groups based
  • n predictor characteristics
  • Examples: logistic regressions, partial least squares discriminant

analysis, and Elastic Net/Lasso models

  • These models commonly have:

– High degree of interpretability – Low amount of computational time

slide-17
SLIDE 17

Non-Linear Classification Models

  • Utilize non-linear functions to categorize observations
  • Examples: k-nearest neighbors, neural networks, support

vector machines, and multivariate adaptive regression splines

  • These models commonly have:

– Low to moderate interpretability – Moderate to high computational time

slide-18
SLIDE 18

Classification Trees and Rule-Based Models

  • Utilize rules to partition observations into smaller homogenous

groups

  • Examples: classification trees, boosted trees, and random

forests

  • These models commonly have:

– Low to high interpretability – High degree of computational time

slide-19
SLIDE 19

Model Performance on the Test Set

  • 3 models with the highest ROC values were applied to the test set
  • f 3,750 children
slide-20
SLIDE 20

Identifying Prominent Predictors

  • The caret package’s variable importance feature provides one option

for characterizing the general effects of predictors with predictive models

  • The feature was ran on the neural network, random forest, and

boosted tree models to identify the most important variables

Va Variable Name Ne Neural Ne Network Ra Ranki king Ra Random Forest Ra Ranki king Bo Boosted Trees Ra Ranki king Av Average Ra Ranki king Date of Latest Removal 3 1 1 1.7 Beginning Date for Current Placement Setting 2 3 4 3.0 Date of First Removal 1 5 5 3.7 Child's Date of Birth 4 6 6 5.3 Emotionally Disturbed Diagnosis 8 11 8 9.0 Discharge Date of Child's Previous Removal 5 12 11 9.3 Currently Placed in Non-Relative Foster Home 9 14 13 12.0 Currently Placed in an Institution 6 20 10 12.0 Number of Days in Current Placement Setting 36 4 3 14.3 Female Child 16 17 18 17.0