Computational Statistics and Data Analysis 59 (2013) 144–160
Contents lists available at SciVerse ScienceDirect
Computational Statistics and Data Analysis
journal homepage: www.elsevier.com/locate/csda
Simultaneous multifactor DIF analysis and detection in
Item Response Theory
F.B. Gonçalves
a,∗
, D. Gamerman
b
, T.M. Soares
c,d
a
Departamento de Estatística, Universidade Federal de Minas Gerais, Brazil
b
Departamento de Métodos Estatísticos, Universidade Federal do Rio de Janeiro, Brazil
c
Departamento de Estatística, Universidade Federal de Juiz de Fora, Brazil
d
CAEd, Universidade Federal de Juiz de Fora, Brazil
article info
Article history:
Received 12 July 2011
Received in revised form 7 October 2012
Accepted 10 October 2012
Available online 23 October 2012
Keywords:
Bayesian analysis
Educational assessment
Differential Item Functioning
Item Response Theory
MCMC
Multifactor
abstract
Item Response Theory (IRT) is a psychometric theory widely used in educational
assessment and cognitive psychology to analyse data emerged from answers given to
items contained in exams, questionnaires, etc. Standard IRT, however, is based on models
which assume that items behave equally to all individuals. This may not be a reasonable
assumption, especially when the individuals taking the test have different social and/or
cultural backgrounds. Differential Item Functioning (DIF) is an area of IRT which allows
an item to be perceived differently by distinct groups, respecting its usual characteristics.
DIF hypothesis avoids neglecting items that may behave differently among groups and
may also be used to provide important information about differences in the populations
involved in the study. In this paper, two integrated Bayesian models for DIF analysis in
IRT are proposed and compared. Both models are based on a two component mixture
with one component describing DIF and the other accounting for the absence of DIF.
Another contribution of this paper is the approach of the simultaneous presence of
multiple factors causing DIF. Ideas from ANOVA models are used to characterize different
possibilities associated with these factors. The models are also extended to account for
explanation and detection in each factor. A simulation study was conducted to assess the
model’s capabilities and to compare it against existing alternatives. Special attention has
been directed to the conditions required to ensure model identification. An analysis of
a Mathematics exam applied nationally to Brazilian elementary school students is made
considering two DIF factors: geographical region and type of school. The results highlight
the relevance of the proposed methodology to address important issues in educational
studying and testing.
© 2012 Elsevier B.V. All rights reserved.
1. Introduction
Item Response Theory (IRT) is a psychometric theory widely used in educational assessment and cognitive psychology. As
its name suggests, it analyses data emerged from answers given by individuals to items contained in exams, questionnaires,
etc.
The main educational assessment tests in the world (like PISA, GRE, SAT, TOEFL) make use of this theory to obtain their
results. However, the applicability of IRT is not restricted to this kind of test, being applicable to any psychological test to
estimate latent characteristics of the test taker that are related to the answers given in the test.
∗
Correspondence to: Av. Antônio Carlos, 6627 - ICEx - DEST, 31270-901, Belo Horizonte, MG, Brazil. Tel.: +55 31 34095939.
E-mail address: fbgoncalves@est.ufmg.br (F.B. Gonçalves).
0167-9473/$ – see front matter © 2012 Elsevier B.V. All rights reserved.
doi:10.1016/j.csda.2012.10.011