Home  /  Products  /  Features  /  IRT (item response theory)

<-  See Stata's other features

Highlights

  • Binary response models

    • One-parameter logistic (1PL)

    • Two-parameter logistic (2PL)

    • Three-parameter logistic (3PL)

  • Ordinal response models

    • Graded response

    • Partial credit

    • Rating scale

  • Categorical response models

    • Nominal response

  • Hybrid models with differing response types

  • IRT models for multiple groups

  • Graphs

    • Item characteristic curve

    • Test characteristic curve

    • Item information function

    • Test information function

  • Differential item functioning (DIF)

Many researchers study cognitive abilities, personality traits, attitudes, quality of life, patient satisfaction, and other attributes that cannot be measured directly. To quantify these types of latent traits, a researcher often develops an instrument—a questionnaire or test consisting of binary, ordinal, or categorical items—to determine individuals' levels of the trait.

Item response theory (IRT) models can be used to evaluate the relationships between the latent trait of interest and the items intended to measure the trait. With IRT, we can also determine how the instrument as a whole relates to the latent trait.

IRT is used when new instruments are developed, when analyzing and scoring data collected from these instruments, when comparing instruments that measure the same trait, and more.

For instance, when we develop a new instrument, we have a set of items that we believe to be good measurements of our latent trait. We can use IRT models to determine whether these items, or a subset, can be combined to form a good measurement tool. With IRT, we evaluate the amount of information each item provides. If some items do not provide much information, we may eliminate them. IRT models estimate the difficulty of each item. This tells us the level of the trait that is assessed by the item. We want items that provide information across the full continuum of the latent trait scale. We can also ask how much information an instrument, as a whole, provides for each level of the latent trait. If there are ranges of the latent trait for which little information is provided, we may add items to the test.

Let's see it work

Suppose we have a test designed to assess mathematical ability based on eight questions that are scored 0 (incorrect) or 1 (correct). We fit a one-parameter logistic model, a model that estimates only the difficulty of each of our eight items, by typing

irt 1pl q1-q8

Even better, we can fit our model from the IRT Control Panel.

Either way, here are the results.

. irt 1pl q1-q8

Fitting fixed-effects model:

Iteration 0:  Log likelihood =   -3790.87
Iteration 1:  Log likelihood =  -3786.285
Iteration 2:  Log likelihood = -3786.2825
Iteration 3:  Log likelihood = -3786.2825

Fitting full model:

Iteration 0:  Log likelihood = -3681.3257
Iteration 1:  Log likelihood = -3669.4837
Iteration 2:  Log likelihood = -3669.4715
Iteration 3:  Log likelihood = -3669.4715

One-parameter logistic model                               Number of obs = 800
Log likelihood = -3669.4715
Coefficient Std. err. z P>|z| [95% conf. interval]
Discrim .8926881 .0497697 17.94 0.000 .7951412 .990235
q1
Diff -.6826695 .1000551 -6.82 0.000 -.8787739 -.486565
q2
Diff -.1177849 .0930488 -1.27 0.206 -.3001572 .0645875
q3
Diff -1.754276 .1356766 -12.93 0.000 -2.020197 -1.488355
q4
Diff .3101872 .0943281 3.29 0.001 .1253074 .4950669
q5
Diff 1.595213 .1288328 12.38 0.000 1.342705 1.847721
q6
Diff .6694488 .0997334 6.71 0.000 .4739748 .8649227
q7
Diff 1.279229 .1167531 10.96 0.000 1.050397 1.508061
q8
Diff -2.328184 .1640633 -14.19 0.000 -2.649742 -2.006625

Coefficients labeled "Diff" report difficulty. Based on this model, question 8 is the easiest with a coefficient of −2.328. Question 5 is the most difficult with a coefficient of 1.595.

We have only eight questions in our example. If we had 50 questions, it would not be as easy to spot those that correspond to a particular difficulty level. We can use estat report to sort the questions by difficulty.

. estat report, sort(b) byparm

One-parameter logistic model                               Number of obs = 800
Log likelihood = -3669.4715
Coefficient Std. err. z P>|z| [95% conf. interval]
Discrim .8926881 .0497697 17.94 0.000 .7951412 .990235
Diff
q8 -2.328184 .1640633 -14.19 0.000 -2.649742 -2.006625
q3 -1.754276 .1356766 -12.93 0.000 -2.020197 -1.488355
q1 -.6826695 .1000551 -6.82 0.000 -.8787739 -.486565
q2 -.1177849 .0930488 -1.27 0.206 -.3001572 .0645875
q4 .3101872 .0943281 3.29 0.001 .1253074 .4950669
q6 .6694488 .0997334 6.71 0.000 .4739748 .8649227
q7 1.279229 .1167531 10.96 0.000 1.050397 1.508061
q5 1.595213 .1288328 12.38 0.000 1.342705 1.847721

We can visualize the relationship between questions and mathematical ability—between items and latent trait—by graphing the item characteristic curves (ICCs) using irtgraph icc.

icc1_stcolor.svg

We made the easiest question blue and the hardest, red. The probability of succeeding on the easiest question is higher than the probability of succeeding on all other questions. Because we fit a 1PL model, this is true at every level of ability.

irtgraph tif graphs the test information function.

tif1_stcolor.svg

The hump in the middle shows that this test provides the most information for average mathematical ability levels.

When we have binary items, we can fit a 1PL, 2PL, or 3PL model. The irt 2pl command fits a 2PL model and allows items to have different difficulties and different abilities to discriminate between high and low levels of the latent trait. Visually, differing discriminations means that the slopes of our ICC curves differ across items. The irt 3pl command extends the 2PL model to allow for the possibility of guessing correct answers.

Let's see it work with ordinal items

IRT models can be fit to ordinal and categorical items, too. Here we have a new test, also with eight questions. Individuals are expected to show their work as they solve each problem. Responses are scored as 0 (incorrect), 1 (partially correct), or 2 (correct).

With ordinal data, we could fit a graded response model, a partial credit model, or a rating scale model. These models make different assumptions about how the ordered scores relate to the latent trait. Here we fit a graded response model by typing

irt grm q1-q8

The results are

. irt grm q1-q8

Fitting fixed-effects model:

Iteration 0:  Log likelihood = -23241.067
Iteration 1:  Log likelihood =  -20961.66
Iteration 2:  Log likelihood = -20874.212
Iteration 3:  Log likelihood = -20869.975
Iteration 4:  Log likelihood = -20869.947
Iteration 5:  Log likelihood = -20869.947

Fitting full model:

Iteration 0:  Log likelihood = -19894.671
Iteration 1:  Log likelihood = -19645.066
Iteration 2:  Log likelihood = -19634.227
Iteration 3:  Log likelihood = -19634.173
Iteration 4:  Log likelihood = -19634.173

Graded response model                                    Number of obs = 2,941
Log likelihood = -19634.173
Coefficient Std. err. z P>|z| [95% conf. interval]
q1
Discrim 1.75666 .1081947 16.24 0.000 1.544603 1.968718
Diff
>=1 -2.138871 .0879781 -2.311305 -1.966437
=2 -1.238469 .0530588 -1.342462 -1.134476
q2
Discrim 1.575855 .0888881 17.73 0.000 1.401637 1.750072
Diff
>=1 -1.345774 .0596825 -1.46275 -1.228798
=2 -.5571402 .0383507 -.6323062 -.4819741
q3
Discrim 1.100984 .0626802 17.57 0.000 .9781335 1.223835
Diff
>=1 -1.604283 .0847658 -1.770421 -1.438145
=2 .1580019 .041744 .0761851 .2398186
q4
Discrim .9245333 .0549971 16.81 0.000 .8167411 1.032326
Diff
>=1 -.7752653 .0616916 -.8961786 -.6543519
=2 1.227147 .076343 1.077518 1.376777
q5
Discrim 1.528995 .0896173 17.06 0.000 1.353349 1.704642
Diff
>=1 .0365143 .0342353 -.0305857 .1036143
=2 .3670961 .0366753 .2952138 .4389784
q6
Discrim .6986686 .0516767 13.52 0.000 .5973842 .799953
Diff
>=1 -.3764114 .0636946 -.5012505 -.2515723
=2 4.226257 .2991581 3.639917 4.812596
q7
Discrim 1.430949 .0874188 16.37 0.000 1.259611 1.602286
Diff
>=1 .6332313 .0418799 .5511482 .7153143
=2 1.242491 .0612475 1.122448 1.362534
q8
Discrim .9449605 .0647013 14.60 0.000 .8181483 1.071773
Diff
>=1 1.041434 .0704925 .9032708 1.179596
=2 2.72978 .1638681 2.408605 3.050956

One way to evaluate how an individual item, say, q3, relates to mathematical ability is to look at the category characteristic curves produced by irtgraph icc.

icc2_stcolor.svg

Respondents with mathematical ability levels below −1.3 are most likely to answer q3 with a completely incorrect answer, those with levels between −1.3 and −0.2 are most likely to give a partially correct answer, and those with ability levels above −0.15 are most likely to give a completely correct answer.

From the test characteristic curve produced by irtgraph tcc, we see how the expected total test score relates to mathematical ability levels.

tcc1_stcolor.svg

Out of a possible 16 points on the test, a person with above-average mathematical ability (above 0) is expected to score above 7.94 or, because all scores are integers, above 7.

Not interested in standardized testing?

IRT models can be used to measure many types of latent traits. For example,

  • attitudes

  • personality traits

  • health outcomes

  • quality of life

Use IRT for analyzing any unobservable characteristic for which binary or categorical measurements are observed.

Tell me more

Stata's IRT features are documented in their own manual. You can read more about IRT and more about Stata's IRT features and see several worked examples in the Item Response Theory Reference Manual.