Prior elicitation, variable selection and Bayesian computation for logistic regression models

Ming Hui Chen, Joseph G. Ibrahim, Constantin Yiannoutsos

Research output: Contribution to journalArticle

74 Scopus citations

Abstract

Bayesian selection of variables is often difficult to carry out because of the challenge in specifying prior distributions for the regression parameters for all possible models, specifying a prior distribution on the model space and computations. We address these three issues for the logistic regression model. For the first, we propose an informative prior distribution for variable selection. Several theoretical and computational properties of the prior are derived and illustrated with several examples. For the second, we propose a method for specifying an informative prior on the model space, and for the third we propose novel methods for computing the marginal distribution of the data. The new computational algorithms only require Gibbs samples from the full model to facilitate the computation of the prior and posterior model probabilities for all possible models. Several properties of the algorithms are also derived. The prior specification for the first challenge focuses on the observables in that the elicitation is based on a prior prediction y0 for the response vector and a quantity a0 quantifying the uncertainty in y0. Then, y0 and a0 are used to specify a prior for the regression coefficients semi-automatically. Examples using real data are given to demonstrate the methodology.

Original languageEnglish (US)
Pages (from-to)223-242
Number of pages20
JournalJournal of the Royal Statistical Society. Series B: Statistical Methodology
Volume61
Issue number1
DOIs
StatePublished - Jan 1 1999

    Fingerprint

Keywords

  • Gibbs sampling
  • Logistic regression
  • Normal prior
  • Posterior distribution
  • Prior distribution
  • Selection of variables

ASJC Scopus subject areas

  • Statistics and Probability
  • Statistics, Probability and Uncertainty

Cite this