Mixed Membership Stochastic Blockmodels
TLDR
In this article, the authors introduce a class of variance allocation models for pairwise measurements, called mixed membership stochastic blockmodels, which combine global parameters that instantiate dense patches of connectivity (blockmodel) with local parameters (mixed membership), and develop a general variational inference algorithm for fast approximate posterior inference.Abstract:
Consider data consisting of pairwise measurements, such as presence or absence of links between pairs of objects. These data arise, for instance, in the analysis of protein interactions and gene regulatory networks, collections of author-recipient email, and social networks. Analyzing pairwise measurements with probabilistic models requires special assumptions, since the usual independence or exchangeability assumptions no longer hold. Here we introduce a class of variance allocation models for pairwise measurements: mixed membership stochastic blockmodels. These models combine global parameters that instantiate dense patches of connectivity (blockmodel) with local parameters that instantiate node-specific variability in the connections (mixed membership). We develop a general variational inference algorithm for fast approximate posterior inference. We demonstrate the advantages of mixed membership stochastic blockmodels with applications to social networks and protein interaction networks.read more
Citations
More filters
Journal ArticleDOI
Variational Inference: A Review for Statisticians
TL;DR: For instance, mean-field variational inference as discussed by the authors approximates probability densities through optimization, which is used in many applications and tends to be faster than classical methods, such as Markov chain Monte Carlo sampling.
Journal ArticleDOI
Link prediction in complex networks: A survey
TL;DR: Recent progress about link prediction algorithms is summarized, emphasizing on the contributions from physical perspectives and approaches, such as the random-walk-based methods and the maximum likelihood methods.
Proceedings Article
Learning to Discover Social Circles in Ego Networks
Jure Leskovec,Julian McAuley +1 more
TL;DR: A novel machine learning task of identifying users' social circles is defined as a node clustering problem on a user's ego-network, a network of connections between her friends, and a model for detecting circles is developed that combines network structure as well as user profile information.
Journal ArticleDOI
Feature Selection: A Data Perspective
TL;DR: This survey revisits feature selection research from a data perspective and reviews representative feature selection algorithms for conventional data, structured data, heterogeneous data and streaming data, and categorizes them into four main groups: similarity- based, information-theoretical-based, sparse-learning-based and statistical-based.
Journal ArticleDOI
Probabilistic Topic Models
TL;DR: In this paper, a review of probabilistic topic models can be found, which can be used to summarize a large collection of documents with a smaller number of distributions over words.
References
More filters
ReportDOI
A Collapsed Variational Bayesian Inference Algorithm for Latent Dirichlet Allocation
TL;DR: This paper proposes the collapsed variational Bayesian inference algorithm for LDA, and shows that it is computationally efficient, easy to implement and significantly more accurate than standard variationalBayesian inference for L DA.
Journal ArticleDOI
Stochastic Blockmodels for Directed Graphs
Yuchung J. Wang,George Y. Wong +1 more
TL;DR: An iterative scaling algorithm is presented for fitting the model parameters by maximum likelihood and blockmodels that are simple extensions of the p 1 model are proposed specifically for such data.
The variational Bayesian EM algorithm for incomplete data: with application to scoring graphical model structures
TL;DR: This method constructs and optimises a lower bound on the marginal likelihood using variational calculus, resulting in an iterative algorithm which generalises the EM algorithm by maintaining posterior distributions over both latent variables and parameters.
Proceedings Article
Link Prediction in Relational Data
TL;DR: It is shown that the collective classification approach of RMNs, and the introduction of subgraph patterns over link labels, provide significant improvements in accuracy over flat classification, which attempts to predict each link in isolation.
Proceedings Article
Expectation-propagation for the generative aspect model
Tom Minka,John Lafferty +1 more
TL;DR: This paper showed that the simple variational methods of Blei et al. (2001) can lead to inaccurate inferences and biased learning for the generative aspect model and developed an alternative approach that leads to higher accuracy at comparable cost.