z-logo
Premium
A convergence diagnostic for Bayesian clustering
Author(s) -
Lysy Martin,
Asgharian Masoud,
Partovi Nia Vahid
Publication year - 2020
Publication title -
wiley interdisciplinary reviews: computational statistics
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 0.693
H-Index - 38
eISSN - 1939-0068
pISSN - 1939-5108
DOI - 10.1002/wics.1536
Subject(s) - markov chain monte carlo , cluster analysis , computer science , gibbs sampling , posterior probability , bayesian probability , data mining , markov chain , machine learning , artificial intelligence
In many applications of Bayesian clustering, posterior sampling on the discrete state space of cluster allocations is achieved via Markov chain Monte Carlo (MCMC) techniques. As it is typically challenging to design transition kernels to explore this state space efficiently, MCMC convergence diagnostics for clustering applications are especially important. Here we propose a diagnostic tool for discrete‐space MCMC, focusing on Bayesian clustering applications where the model parameters have been integrated out. We construct a Hotelling‐type statistic on the highest probability states, and use regenerative sampling theory to derive its equilibrium distribution. By leveraging information from the unnormalized posterior, our diagnostic offers added protection against seemingly convergent chains in which the relative frequency of visited states is incorrect. The methodology is illustrated with a Bayesian clustering analysis of genetic mutants of the flowering plant Arabidopsis thaliana . This article is categorized under: Statistical Learning and Exploratory Methods of the Data Sciences > Clustering and Classification Statistical Learning and Exploratory Methods of the Data Sciences > Knowledge Discovery Statistical and Graphical Methods of Data Analysis > Markov Chain Monte Carlo

This content is not available in your region!

Continue researching here.

Having issues? You can contact us here