Computational Statistics and Data Analysis 52 (2008) 4608–4624 Contents lists available at ScienceDirect Computational Statistics and Data Analysis journal homepage: www.elsevier.com/locate/csda Marginal likelihoods for non-Gaussian models using auxiliary mixture sampling Sylvia Frühwirth-Schnatter , Helga Wagner Johannes Kepler Universität Linz, Department of Applied Statistics and Econometrics, Altenbergerstraße 69, A-4040 Linz, Austria article info Article history: Received 13 June 2007 Received in revised form 13 March 2008 Accepted 14 March 2008 Available online 3 April 2008 abstract Several new estimators of the marginal likelihood for complex non-Gaussian models are developed. These estimators make use of the output of auxiliary mixture sampling for count data and for binary and multinomial data. One of these estimators is based on combining Chib’s estimator with data augmentation as in auxiliary mixture sampling, while the other estimators are importance sampling and bridge sampling based on constructing an unsupervised importance density from the output of auxiliary mixture sampling. These estimators are applied to a logit regression model, to a Poisson regression model, to a binomial model with random intercept, as well as to state space modeling of count data. © 2008 Elsevier B.V. All rights reserved. 1. Introduction For an applied statistician, choosing an appropriate model from a class of candidate models is a fundamental data analytical task. In a classical framework, model selection problems are addressed either via hypothesis testing for nested models, or by using information criteria such as the Akaike information criterion (Akaike, 1974) or Schwarz’s criterion or BIC (Schwarz, 1978). In a Bayesian setting, model selection relies on the posterior probabilities of a model given the data, see Bernardo and Smith (1994) as well as the recent reviews by Godsill (2001), Green (2003) and Kadane and Lazar (2004). More formally, suppose there are K different models M 1 ,..., M K , which are candidates for having generated the data y. Each of these models is assigned a prior probability p(M k ) and the goal is to derive the posterior model probabilities p(M k |y) for each model M k , k = 1,..., K. There are basically two strategies to implement Bayesian model selection. Model space MCMC methods directly sample from the discrete model space (M 1 ,..., M K ) by drawing jointly model indicators and parameters, using e.g. the reversible jump MCMC algorithm (Green, 1995) or the stochastic variable selection approach (George and McCulloch, 1993, 1997). A more classical strategy which dates back to Jeffreys (1948) and Zellner (1971) determines the posterior model probabilities p(M k |y) separately for each model by using Bayes’ rule: p(M k |y) p(y|M k )p(M k ), where p(y|M k ) is the marginal likelihood of model M k . Let ϑ k denote the vector containing the unknown parameters of model M k . Then the marginal likelihood is given as p(y|M k ) = Θ k p(y|ϑ k )p(ϑ k )dϑ k , (1) with p(ϑ k ) being the prior distribution of model parameter ϑ k with support Θ k . Corresponding author. Tel.: +43 732 2468 8294; fax: +43 732 2468 9846. E-mail address: sylvia.fruehwirth-schnatter@jku.at (S. Frühwirth-Schnatter). 0167-9473/$ – see front matter © 2008 Elsevier B.V. All rights reserved. doi:10.1016/j.csda.2008.03.028