7) Functional Methods for Testing Data

Download Report

Transcript 7) Functional Methods for Testing Data

Functional Methods for Testing
Data
2.5
1.5
1
0.5
1
Item 29
0.8
0
0.6
-0.5
0.4
-1
-1.5
0.2
-2
0
1
-2.5
0.8
1
0.6
0.8
0.6
0.4
0.4
0.2
Item 4
0.2
0
0
Item 3
The data




Each person a = 1,…,N
responds to each item i = 1,…,n
and makes a binary response = uai , where
0 indicates “wrong” and 1 indicates
“right”.
We want to estimate Pai = the probability
of person a getting item i right.
The model



The response space is an n-dimensional
unit hypercube. The data vectors
{ua1,…,uan} are on the corners.
The vectors of correct response
probabilities {Pa1,…,Pan} fall along a
smooth curve in response space, the
response manifold.
This manifold is, in principle, identifiable
from the data, is therefore not a latent
trait.
Item response functions



We can define a smooth charting function that
maps each point on the response manifold to a
corresponding real number θ. E.g.: arc length.
In this way we establish a metric defining
positions on this manifold.
Pi(θ) is the success probability for item i of all
those at position θ. This is a smooth function of
θ : The item response function for item i.


The response
manifold for 3
test items: 3, 4,
and 29.
The curve
indicates the
possible values
of Pai = Pi (θ) .
The circles
correspond to 11
fixed values of
θ.
2.5
1.5
1
0.5
1
0.8
Item 29

Three items from an
Intro Psych test
0
0.6
-0.5
0.4
-1
-1.5
0.2
-2
0
1
-2.5
0.8
1
0.6
0.8
0.6
0.4
0.4
0.2
Item 4
0.2
0
0
Item 3
What does “smooth” mean?
If θ has a standard normal distribution,
then experience indicates that usually:
 The function Pi (θ) is monotonic.
 It has slopes near 0 for extreme θ values.
 The lower asymptote is positive, and the
upper asymptote is one.
 There is only one inflection point.
The three-parameter logistic item
response function is smooth
1
0.9
0.8
0.7
P()
0.6
0.5
0.4
0.3
0.2
0.1
0
-3
-2
-1
0

1
2
3
The challenges




Let the item response functions take
whatever shapes are supported by the
data.
But control their smoothness in this sense.
Constrain the function values to lie within
[0,1].
We want a smooth derivative, too, for the
item information function.
The log-odds transformation deals
with the [0,1] constraint
We actually estimate
Wi(θ) = log [Pi (θ) /Qi (θ)], where
Qi (θ) = 1 – Pi (θ).
“Smooth” in terms of Wi (θ) means linear
behavior for extreme θ, with small slope
on the left and larger positive slope on the
right.
The log-odds transformation of a
3PL item response function
4
3
W()
2
1
0
-1
-2
-3
-2
-1
0

1
2
3
A three-dimensional model for a
smooth log-odds function

W ( )  c1  c2  c3 log(e  1)
The function log(e θ+1) has the desired behavior
at the extremes.
The other two terms add vertical shift and tilt as
required.
Comparing the 3D model and the
3PL log-odds functions
4
3
3D
3PL
W()
2
1
0
-1
-2
-3
-2
-1
0

1
2
3
B-spline expansions for W(θ)
But, in fact, we will actually estimate our
log-odds functions by
• expanding W(θ) in terms of a set of K
B-spline basis functions, while
• smoothing these expansions towards
these simpler three-dimensional models.
Fitting the data
We use maximum marginal likelihood estimation,
using the EM algorithm to maximize


log ML   log    Pai   Qai   g   d 
a
 i

where g(θ) is a prior density on θ, often taken to
be the standard normal.
Maximization is with respect to the nK coefficients
defining the B-spline expansions of the log-odds
functions.
What about smoothness?


We have defined smoothness here in a less
orthodox fashion; It isn’t defined only in terms
of the second derivative.
Instead, we define smooth in terms of the size
of

1 e 2
3
LW    
D W    D W  

1 e
How did you come up with this?


If W(θ) conforms exactly to the threedimensional smooth model, then LW(θ) = 0.
In other words, if
then

1 e 2
D W   
D W  

1 e
3




Our strategy is to define a lowdimensional family of prototype functions
that capture what we mean by “smooth.”
Then we represent this family by a linear
differential equation.
This differential equation defines a
measure of “roughness”, which we
penalize.
The more we penalize this kind of
roughness, the more we force the fitted
functions to be smooth.

In general, if we begin with a linear model of
dimension m, we can find a linear differential
equation of order m such that all versions of this
model will satisfy:
DmW(θ) = b0 (θ) W(θ)+ b1 (θ) DW(θ)+ …
+ bm-1 (θ) Dm-1W(θ)
for some choice of coefficient functions bj(θ).

We change the equation to a roughness penalty
by converting it to operator form:
LW(θ) =
b0 (θ) W(θ)+ b1 (θ) DW(θ)+…+ bm-1 (θ) Dm-1W (θ)
+ DmW(θ) = 0.
The roughness-penalty
R(W )   [ LW  ] d
2
measures the departure of W(θ) from this
smooth model.
Roughness-penalized log marginal
likelihood
Consequently, we actually maximize
log ML    R(Wi )
i
Smoothing parameter λ controls the amount of
smoothness in the W(θ) ‘s; the larger it is, the
more these will look like the three-dimensional
versions.
Some examples



Here are three estimates of the item
response functions for items 3, 4, 29, and
96 for an introductory psychology test
The test had 100 items, and was given to
379 students.
Each function W(θ) is defined by an
expansion in terms of 13 B-spline basis
functions.
λ=0
Item 3
Item 4
1
1
0.8
0.8
0.6
0.6
0.4
0.4
0.2
0.2
0
-2
-1
0

1
2
0
-2
-1
Item 29
1
0.8
0.8
0.6
0.6
0.4
0.4
0.2
0.2
-2
-1
0


1
2
1
2
Item 96
1
0
0
1
2
0
-2
-1
0

λ=0.01
Item 3
Item 4
1
1
0.8
0.8
0.6
0.6
0.4
0.4
0.2
0.2
0
-2
-1
0

1
2
0
-2
-1
Item 29
1
0.8
0.8
0.6
0.6
0.4
0.4
0.2
0.2
-2
-1
0


1
2
1
2
Item 96
1
0
0
1
2
0
-2
-1
0

λ=50
Item 3
Item 4
1
1
0.8
0.8
0.6
0.6
0.4
0.4
0.2
0.2
0
-2
-1
0

1
2
0
-2
-1
Item 29
1
0.8
0.8
0.6
0.6
0.4
0.4
0.2
0.2
-2
-1
0


1
2
1
2
Item 96
1
0
0
1
2
0
-2
-1
0

What does θ mean?




We have fallen into the habit of calling θ a
“latent trait score”.
Actually, it is the value of a function that is
chosen more or less arbitrarily to map
position along the response manifold.
The assumption of a standard normal
distribution is pure convention.
We can choose otherwise.
What charting functions would be
more useful?


Three choices of charting functions are
especially interesting, and none are
“latent” in any sense.
Each leads to interesting diagnostic
statistics and graphics.
The arc length charting

Arc length s measures the Euclidean distance
traveled along the manifold from its origin at
θ0 to a given position θ:
s( )  

0
  DP  
i
i
2
d
Item discrimination in arc length
metric
One useful property of arc length is
  DP  s 
i
i
2
1
Each squared item discrimination is a proportion
total test discrimination, and therefore has a
familiar frame of reference.
Expected score charting
Assuming that expected score is monotonically
related to θ, (there aren’t too many items like
96), then
     Pi  
i
Provides a metric that is familiar to users and easy
for them to interpret.
Expected score is already used extensively as a
basis for assessing differential item functioning (DIF).
ACT Math test for males and
females


Three items from a 60
item math test.
Around 2000
examinees.
The male and female
response manifolds
differ.
F
M
M
F
M
1
F
M
0.8
Item 19

F
0.6
M
0.4
0.2
M
0
1
F
M
0.8
F
0.6
F
0.4
0.8
0.6
M
0.4
0.2
Item 17
1
F
M
0.2
0
0
Item 14
Differential item functioning for an
ACT Math test item
Item 17
1
0.9
Probability of Success
0.8
0.7
0.6
0.5
0.4
Male
Female
0.3
0.2
0.1
0
10
20
30
40
Expected Score 
50
60
Total change charting
The following total change in probability of
success measure is closely related to arc
length:

c     DPi   d
0
Some general lessons



Fitting functional models to non-functional
data is relatively straight-forward.
But we do need to transform constrained
functions into unconstrained versions.
We can define smoothness or roughness
in customized ways that capture the
default or baseline behavior of our
estimated functions.




“Latent trait models” aren’t really latent at
all.
They express the idea of a one-dimensional
subspace for modeling the data.
Differential geometry gives us the
appropriate mathematical tools.
There is room for creativity in choosing
charting functions.
Looking ahead


There is an intimate connection between
designer roughness penalties and the
estimation of differential equations from
data.
We will use discrete data to estimate a
differential equation that describes the
data.
References
More technical details on fitting test data
with functional models are in
Rossi, N., Wang, X. and Ramsay, J. O.
(2002) Nonparametric item response
function estimates with the EM algorithm.
Journal of Educational and Behavioral
Statistics, 27, 291-317.