scispace - formally typeset
Open AccessJournal ArticleDOI

Mixed Membership Stochastic Blockmodels

TLDR
In this article, the authors introduce a class of variance allocation models for pairwise measurements, called mixed membership stochastic blockmodels, which combine global parameters that instantiate dense patches of connectivity (blockmodel) with local parameters (mixed membership), and develop a general variational inference algorithm for fast approximate posterior inference.
Abstract
Consider data consisting of pairwise measurements, such as presence or absence of links between pairs of objects. These data arise, for instance, in the analysis of protein interactions and gene regulatory networks, collections of author-recipient email, and social networks. Analyzing pairwise measurements with probabilistic models requires special assumptions, since the usual independence or exchangeability assumptions no longer hold. Here we introduce a class of variance allocation models for pairwise measurements: mixed membership stochastic blockmodels. These models combine global parameters that instantiate dense patches of connectivity (blockmodel) with local parameters that instantiate node-specific variability in the connections (mixed membership). We develop a general variational inference algorithm for fast approximate posterior inference. We demonstrate the advantages of mixed membership stochastic blockmodels with applications to social networks and protein interaction networks.

read more

Content maybe subject to copyright    Report

Citations
More filters
Journal ArticleDOI

Optimizing Variational Graph Autoencoder for Community Detection with Dual Optimization.

TL;DR: A dual optimization procedure is introduced on Variational Graph Autoencoder for Community Detection (VGAECD) that aims to guide the optimization process and encourage learning of the primary objective, and linearize the encoder to reduce the number of learning parameters.
Proceedings Article

Clustering and Ranking in Heterogeneous Information Networks via Gamma-Poisson Model

TL;DR: A probabilistic generative model is proposed that simultaneously achieves clustering and ranking on a heterogeneous network that can follow arbitrary schema, where the edges from different types are sampled from a Poisson distribution with the parameters determined by the ranking scores of the nodes in each cluster.
Posted Content

Smoothed Gradients for Stochastic Variational Inference

TL;DR: This paper replaces the natural gradient with a similarly constructed vector that uses a fixed-window moving average of some of its previous terms that enjoys significant variance reduction over the unbiased estimates, smaller bias than averaged gradients, and leads to smaller mean-squared error against the full gradient.
Journal ArticleDOI

Mining Overlapping Communities and Inner Role Assignments through Bayesian Mixed-Membership Models of Networks with Context-Dependent Interactions

TL;DR: Two new model-based machine-learning approaches are presented, wherein community discovery and role assignment are seamlessly integrated and simultaneously performed through approximate posterior inference in Bayesian mixed-membership models of directed networks.
Journal ArticleDOI

Features and heterogeneities in growing network models.

TL;DR: A generalization of growing network models with preferential attachment that includes the effect of heterogeneous features of the nodes, and concludes that both small clustering coefficients and disassortative mixing are outcomes of the preferential attachment mechanism in general growing networks.
References
More filters
Journal ArticleDOI

Gene Ontology: tool for the unification of biology

TL;DR: The goal of the Gene Ontology Consortium is to produce a dynamic, controlled vocabulary that can be applied to all eukaryotes even as knowledge of gene and protein roles in cells is accumulating and changing.
Journal ArticleDOI

Latent dirichlet allocation

TL;DR: This work proposes a generative model for text and other collections of discrete data that generalizes or improves on several previous models including naive Bayes/unigram, mixture of unigrams, and Hofmann's aspect model.
Journal ArticleDOI

Finding scientific topics

TL;DR: A generative model for documents is described, introduced by Blei, Ng, and Jordan, and a Markov chain Monte Carlo algorithm is presented for inference in this model, which is used to analyze abstracts from PNAS by using Bayesian model selection to establish the number of topics.
Related Papers (5)