Clustering approach and MCMC practicalities of stochastic block models

Clement Lee

2019-06-13 (Thu)

1. Introduction

Background

A Bigger Picture

Three main types of models for networks

  1. Generative models
    • Preferential attachment model (Barabasi and Albert, 1999)
    • Small world model (Watts and Strogatz, 1998)
  2. Exponential random graph models
  3. Latent models
    • Stochastic block models (SBMs)
    • Latent feature models (Miller, Griffiths and Jordan, 2009; Morup, Schmidt and Hansen, 2011)
    • Latent class analysis models (Ng and Murphy, 2018)
    • Latent space models (Hoff, Raftery and Handcock, 2002; Handcock, Raftery and Tantrum, 2007)

Stochastic Block Models

Notation

Structural Equivalence

Bayesian Inference

MCMC Algorithms

2. Soft Clustering

Comparing Models

Mixed Membership SBM

Airoldi, Blei, Fienberg and Xing (2008)

Likelihood & Inference

Comparing with Clustering Non-relational Data

Latent Dirichlet Allocation (for example)

Mixed membership SBM

Going Back to the Basics?

Some Modifications

To the model

To the MCMC algorithm

3. Back to Hard Clustering

Poisson Approximation

Previously (Bernoulli SBM)

Karrer and Newman(2011); Peixoto (2018)

Comparing the Priors

Bernoulli SBM

Poisson SBM

Integrating out \(\boldsymbol{C}\)

Inference

Prior of \(\boldsymbol{Z}\)

Summary of SBMs

Bernoulli Mixed membership Poisson
Clustering Hard Soft Hard
Quantity of interest \(\boldsymbol{Z}\) \(\boldsymbol{\Theta}\) \(\boldsymbol{Z}\)
Marginalisation? \(\boldsymbol{C}\) \(\boldsymbol{C}\) & \(\boldsymbol{D}\), or \(\boldsymbol{Z}\) \(\boldsymbol{C}\) & \(\mu\)
Remarks Neither marginalisation particularly useful Exponential priors for \(\boldsymbol{C}\) with dep. on \(\boldsymbol{Z}\)
Quadratic computational cost Can extend to nested version

4. Practicalities

Component-wise Moves

Metropolis

Component-wise Moves

Gibbs

Bigger Moves

Gibbs + M3 (McDaid, Murphy, Friel and Hurley, 2013)

M3 & Label Switching

Propose to randomise nodes in two groups

  1. All nodes are placed in a randomly reordered list
  2. Each node is placed in one group according to some assignment probability
  3. Nobile and Fearnside (2007): Choose the ratio of the assignment probabilities as the ratio of the two posterior probabilities resulting from the assignments of the previous nodes
    • Heuristic, should lead to “good” choices
  4. The same list traversed to calculate reverse proposal probability
  5. Check agreement between current and proposed \(\boldsymbol{Z}\); switch labels if agreement < 50%

Informed Proposals

Zanella (2019)

Multiple Local Modes

Metropolis-coupled MCMC [(MC)\(^3\)] / Parallel Tempering

Model Selection / K

Modelled

Estimated

Criterion

Model Selection / K

Criterion: marginal likelihood

\[ f(\boldsymbol{Y})=\frac{f(\boldsymbol{Y}|\boldsymbol{Z})\pi(\boldsymbol{Z})}{\pi(\boldsymbol{Z}|\boldsymbol{Y})}\] \[ \log{}f(\boldsymbol{Y})=\underbrace{\log\left[f(\boldsymbol{Y}|\boldsymbol{Z})\pi(\boldsymbol{Z})\right]}_{\text{log-ICL}} - \log\pi(\boldsymbol{Z}|\boldsymbol{Y}) \]

Summary

Next steps

Questions? Comments?