Bayesian sparse graphical models and their mixtures. Academic Article uri icon

abstract

  • We propose Bayesian methods for Gaussian graphical models that lead to sparse and adaptively shrunk estimators of the precision (inverse covariance) matrix. Our methods are based on lasso-type regularization priors leading to parsimonious parameterization of the precision matrix, which is essential in several applications involving learning relationships among the variables. In this context, we introduce a novel type of selection prior that develops a sparse structure on the precision matrix by making most of the elements exactly zero, in addition to ensuring positive definiteness - thus conducting model selection and estimation simultaneously. More importantly, we extend these methods to analyze clustered data using finite mixtures of Gaussian graphical model and infinite mixtures of Gaussian graphical models. We discuss appropriate posterior simulation schemes to implement posterior inference in the proposed models, including the evaluation of normalizing constants that are functions of parameters of interest, which result from the restriction of positive definiteness on the correlation matrix. We evaluate the operating characteristics of our method via several simulations and demonstrate the application to real data examples in genomics.

published proceedings

  • Stat

altmetric score

  • 0.5

author list (cited authors)

  • Talluri, R., Baladandayuthapani, V., & Mallick, B. K.

citation count

  • 4

complete list of authors

  • Talluri, Rajesh||Baladandayuthapani, Veerabhadran||Mallick, Bani K

publication date

  • January 2014

publisher

published in