scispace - formally typeset
Open AccessPosted Content

Variational Autoencoders for Collaborative Filtering

TLDR
In this paper, a variational autoencoder (VAE) was extended to collaborative filtering for implicit feedback, and a generative model with multinomial likelihood and Bayesian inference for parameter estimation was proposed.
Abstract
We extend variational autoencoders (VAEs) to collaborative filtering for implicit feedback. This non-linear probabilistic model enables us to go beyond the limited modeling capacity of linear factor models which still largely dominate collaborative filtering research.We introduce a generative model with multinomial likelihood and use Bayesian inference for parameter estimation. Despite widespread use in language modeling and economics, the multinomial likelihood receives less attention in the recommender systems literature. We introduce a different regularization parameter for the learning objective, which proves to be crucial for achieving competitive performance. Remarkably, there is an efficient way to tune the parameter using annealing. The resulting model and learning algorithm has information-theoretic connections to maximum entropy discrimination and the information bottleneck principle. Empirically, we show that the proposed approach significantly outperforms several state-of-the-art baselines, including two recently-proposed neural network approaches, on several real-world datasets. We also provide extended experiments comparing the multinomial likelihood with other commonly used likelihood functions in the latent factor collaborative filtering literature and show favorable results. Finally, we identify the pros and cons of employing a principled Bayesian inference approach and characterize settings where it provides the most significant improvements.

read more

Citations
More filters
Posted Content

NSML: Meet the MLaaS platform with a real-world case study

TL;DR: This work proposed NSML, a machine learning as a service (MLaaS) platform, which helps machine learning work be easily launched on a NSML cluster and provides a collaborative environment which can afford development at enterprise scale.
Proceedings ArticleDOI

Improving Graph Collaborative Filtering with Neighborhood-enriched Contrastive Learning

TL;DR: A novel contrastive learning approach, named Neighborhood-enriched Contrastive Learning, named NCL, which explicitly incorporates the potential neighbors into contrastive pairs, and can be optimized with EM algorithm and generalized to apply to graph collaborative filtering methods.
Posted Content

Avoiding Latent Variable Collapse With Generative Skip Models

TL;DR: Compared to existing VAE architectures, it is shown that generative skip models maintain similar predictive performance but lead to less collapse and provide more meaningful representations of the data.
Proceedings ArticleDOI

Knowledge Graph Contrastive Learning for Recommendation

TL;DR: This work proposes a knowledge graph augmentation schema to suppress KG noise in information aggregation, and derive more robust knowledge-aware representations for items, and exploits additional supervision signals from the KG augmentation process to guide a cross-view contrastive learning paradigm.
Journal ArticleDOI

Deep Learning for Recommender Systems: A Netflix Case Study

TL;DR: The challenges encountered and lessons learned in using deep learning for recommender systems at Netflix are outlined and some of the take-aways that may generalize to other applications beyond Netflix are summarized.
References
More filters
Journal ArticleDOI

Latent dirichlet allocation

TL;DR: This work proposes a generative model for text and other collections of discrete data that generalizes or improves on several previous models including naive Bayes/unigram, mixture of unigrams, and Hofmann's aspect model.
Posted Content

Adam: A Method for Stochastic Optimization

TL;DR: In this article, the adaptive estimates of lower-order moments are used for first-order gradient-based optimization of stochastic objective functions, based on adaptive estimate of lowerorder moments.
Posted Content

Distributed Representations of Words and Phrases and their Compositionality

TL;DR: In this paper, the Skip-gram model is used to learn high-quality distributed vector representations that capture a large number of precise syntactic and semantic word relationships and improve both the quality of the vectors and the training speed.
Posted Content

Auto-Encoding Variational Bayes

TL;DR: In this paper, a stochastic variational inference and learning algorithm was proposed for directed probabilistic models with intractable posterior distributions and large datasets, which scales to large datasets.
Related Papers (5)