Computational Statistics and Data Analysis 59 (2013) 144–160 Contents lists available at SciVerse ScienceDirect Computational Statistics and Data Analysis journal homepage: www.elsevier.com/locate/csda Simultaneous multifactor DIF analysis and detection in Item Response Theory F.B. Gonçalves a, , D. Gamerman b , T.M. Soares c,d a Departamento de Estatística, Universidade Federal de Minas Gerais, Brazil b Departamento de Métodos Estatísticos, Universidade Federal do Rio de Janeiro, Brazil c Departamento de Estatística, Universidade Federal de Juiz de Fora, Brazil d CAEd, Universidade Federal de Juiz de Fora, Brazil article info Article history: Received 12 July 2011 Received in revised form 7 October 2012 Accepted 10 October 2012 Available online 23 October 2012 Keywords: Bayesian analysis Educational assessment Differential Item Functioning Item Response Theory MCMC Multifactor abstract Item Response Theory (IRT) is a psychometric theory widely used in educational assessment and cognitive psychology to analyse data emerged from answers given to items contained in exams, questionnaires, etc. Standard IRT, however, is based on models which assume that items behave equally to all individuals. This may not be a reasonable assumption, especially when the individuals taking the test have different social and/or cultural backgrounds. Differential Item Functioning (DIF) is an area of IRT which allows an item to be perceived differently by distinct groups, respecting its usual characteristics. DIF hypothesis avoids neglecting items that may behave differently among groups and may also be used to provide important information about differences in the populations involved in the study. In this paper, two integrated Bayesian models for DIF analysis in IRT are proposed and compared. Both models are based on a two component mixture with one component describing DIF and the other accounting for the absence of DIF. Another contribution of this paper is the approach of the simultaneous presence of multiple factors causing DIF. Ideas from ANOVA models are used to characterize different possibilities associated with these factors. The models are also extended to account for explanation and detection in each factor. A simulation study was conducted to assess the model’s capabilities and to compare it against existing alternatives. Special attention has been directed to the conditions required to ensure model identification. An analysis of a Mathematics exam applied nationally to Brazilian elementary school students is made considering two DIF factors: geographical region and type of school. The results highlight the relevance of the proposed methodology to address important issues in educational studying and testing. © 2012 Elsevier B.V. All rights reserved. 1. Introduction Item Response Theory (IRT) is a psychometric theory widely used in educational assessment and cognitive psychology. As its name suggests, it analyses data emerged from answers given by individuals to items contained in exams, questionnaires, etc. The main educational assessment tests in the world (like PISA, GRE, SAT, TOEFL) make use of this theory to obtain their results. However, the applicability of IRT is not restricted to this kind of test, being applicable to any psychological test to estimate latent characteristics of the test taker that are related to the answers given in the test. Correspondence to: Av. Antônio Carlos, 6627 - ICEx - DEST, 31270-901, Belo Horizonte, MG, Brazil. Tel.: +55 31 34095939. E-mail address: fbgoncalves@est.ufmg.br (F.B. Gonçalves). 0167-9473/$ – see front matter © 2012 Elsevier B.V. All rights reserved. doi:10.1016/j.csda.2012.10.011