Local model influence

Research output: Contribution to journalArticlepeer-review

144 Scopus citations

Abstract

This article develops a general method for assessing the influence of model assumptions in a Bayesian analysis. We assume that model choices are indexed by a hyperparameter with some given initial choice. We use the term “model” to encompass both the sampling model and the prior distribution. We wish to assess the effect of changing the hyperparameter away from the initial choice. We are performing a sensitivity analysis, with the hyperparameter defining our perturbations. We use the Kullback—Leibler divergence to measure the difference between posteriors corresponding to different choices of the hyperparameter. We also measure the change in priors. If small changes in the priors lead to large changes in posteriors, the choice of hyperparameter is influential. The second-order difference in the Kullback—Leibler divergence is expressed by Fisher information matrices. The relative change in posteriors compared with priors may be summarized by the relative eigenvalue of the posterior and prior Fisher information matrices. The corresponding eigenvector indicates which aspects of the perturbation hyperparameter are most influential. Examples considered are the choice of conjugate prior in regression, case weights in regression, and the choice of Dirichlet prior for multinomials.

Original languageEnglish (US)
Pages (from-to)473-478
Number of pages6
JournalJournal of the American Statistical Association
Volume84
Issue number406
DOIs
StatePublished - Jun 1989
Externally publishedYes

Keywords

  • Diagnostics
  • Fisher information
  • Kullback—Leibler divergence
  • Posterior distribution
  • Predictive distribution

ASJC Scopus subject areas

  • Statistics and Probability
  • Statistics, Probability and Uncertainty

Fingerprint

Dive into the research topics of 'Local model influence'. Together they form a unique fingerprint.

Cite this