Fast Item Response Theory (IRT) Analysis by using GPUs
Lei Chen lei.chen@liulishuo.com Liulishuo Silicon Valley AI Lab
1
Fast Item Response Theory (IRT) Analysis by using GPUs Lei Chen - - PowerPoint PPT Presentation
Fast Item Response Theory (IRT) Analysis by using GPUs Lei Chen lei.chen@liulishuo.com Liulishuo Silicon Valley AI Lab 1 Outline A brief introduction of Item Response Theory (IRT) Edward, a new probabilistic programming (PP) toolkit
Lei Chen lei.chen@liulishuo.com Liulishuo Silicon Valley AI Lab
1
estimation on both CPU and GPU computing platforms
2
3
grown threefold between 2013 and 2016, according to a new analysis. EdWeek market brief 7/14/2017
4
ability levels, which are latent
a correct answer p(X=1) depends on
a lucky guess, carelessness …
5
these factors and has been widely used to build up modern assessment industry
6
7
8
8
8
8
standard tests be possible
8
standard tests be possible
items on the same scale
8
testing” presentation made for a hands-on workshop by Rust, Cek, Sun, and Kosinski from University of Cambridge The Psychometrics Center
score, and CAT.
9
Binary items
Parameters: Measured concept (theta) Probability of getting item right 1 Models:
10
Binary items
Parameters:
Measured concept (theta) Probability of getting item right 1 Models:
Difficulty
10
Binary items
Parameters:
Measured concept (theta) Probability of getting item right 1 D i s c r i m i n a t i
( s l
e ) Models:
Difficulty
10
Binary items
Parameters:
Measured concept (theta) Probability of getting item right 1 D i s c r i m i n a t i
( s l
e ) Models:
Difficulty Guessing
10
Binary items
Parameters:
Measured concept (theta) Probability of getting item right 1 D i s c r i m i n a t i
( s l
e ) Models:
Difficulty Guessing Inattention
10
Binary items
Parameters:
Measured concept (theta) Probability of getting item right 1 D i s c r i m i n a t i
( s l
e ) Models:
Difficulty Guessing Inattention
10
Test:
Probability
0.0 0.2 0.4 0.6 0.8 1.0
Theta
0.0 1.0 2.0 3.0 11
Test:
Probability
0.0 0.2 0.4 0.6 0.8 1.0
Theta
0.0 1.0 2.0 3.0 11
Test:
Probability
0.0 0.2 0.4 0.6 0.8 1.0
Theta
0.0 1.0 2.0 3.0 11
Most likely score
Test:
Probability
0.0 0.2 0.4 0.6 0.8 1.0
Theta
0.0 1.0 2.0 3.0 11
Test:
Probability
0.0 0.2 0.4 0.6 0.8 1.0
Theta
0.0 1.0 2.0 3.0 11
Test:
Most likely score
Probability
0.0 0.2 0.4 0.6 0.8 1.0
Theta
0.0 1.0 2.0 3.0 11
Test:
Probability
0.0 0.2 0.4 0.6 0.8 1.0
Theta
0.0 1.0 2.0 3.0 11
Test:
Probability
0.0 0.2 0.4 0.6 0.8 1.0
Theta
0.0 1.0 2.0 3.0
Most likely score
11
Test:
Probability
0.0 0.2 0.4 0.6 0.8 1.0
Theta
0.0 1.0 2.0 3.0
Most likely score
11
Test:
Probability
0.0 0.2 0.4 0.6 0.8 1.0
Theta
0.0 1.0 2.0 3.0
Most likely score
11
test-taker
12
Probability
0.0 0.2 0.4 0.6 0.8 1.0
Theta
0.0 1.0 2.0 3.0
Start the test:
13
Probability
0.0 0.2 0.4 0.6 0.8 1.0
Theta
0.0 1.0 2.0 3.0
Start the test:
medium difficulty Correct response Incorrect response
13
Probability
0.0 0.2 0.4 0.6 0.8 1.0
Theta
0.0 1.0 2.0 3.0
Start the test:
medium difficulty
13
Probability
0.0 0.2 0.4 0.6 0.8 1.0
Theta
0.0 1.0 2.0 3.0
Start the test:
medium difficulty
Most likely score Normal distribution
13
Probability
0.0 0.2 0.4 0.6 0.8 1.0
Theta
0.0 1.0 2.0 3.0
Start the test:
medium difficulty
Most likely score
13
Probability
0.0 0.2 0.4 0.6 0.8 1.0
Theta
0.0 1.0 2.0 3.0
Start the test:
medium difficulty
difficulty around the most likely score (or with the max
information)
Most likely score Difficulty
13
Probability
0.0 0.2 0.4 0.6 0.8 1.0
Theta
0.0 1.0 2.0 3.0
Start the test:
medium difficulty
difficulty around the most likely score (or with the max
information)
stopping rule is reached Most likely score Difficulty
13
integrating over theta
14
sized
15
estimation
sampling many data points from the posterior probability
parameter spaces. HMC utilizes the geometry of the important regions of the posterior for making better proposals.
16
can minimizes divergence to the true posterior
leading to faster estimation
measure two distributions’ closeness
17
based Gibbs sampler for a unidimensional IRT
18
criticism
University
Pelham Box.
probabilistic programming
19
interested in trying PP but don’t want to be swamped by many math and statistics details
20
direction)
21
to Bayesian Machine Learning with PyMC3 and Edward” at PyCon 2017
prob(H)
22
23
questions; need jointly estimate 2,000 + 250 ability and item parameters for a 1-PL model
from two normal distributions
answer vector
24
1+exp(logit), which is 1-PL IRT model where logit = trait
25
HMC
26
variational loc and scale
27
estimated parameters
28
CPU running
Inference/Platform Running time (Sec) MSE HMC/CPU 893 HMC/GPU 222 0.900 VB/CPU 116 VB/GPU 29 0.023
29
educational applications
ability levels and items) can be time consuming
convenient way for doing IRT parameter estimation using Bayesian methods, and it enables fast GPU computations
30
documents/Concerto/irtandcat.pdf
(2016). Bayesian Prior Choice in IRT Estimation Using MCMC and Variational Bayes. Frontiers in Psychology, 7, 1422. http:// doi.org/10.3389/fpsyg.2016.01422
31