Deep Neural Networks for YouTube Recommendations
Paul Covington,Jay Adams,Emre Sargin +2 more
- pp 191-198
Reads0
Chats0
TLDR
This paper details a deep candidate generation model and then describes a separate deep ranking model and provides practical lessons and insights derived from designing, iterating and maintaining a massive recommendation system with enormous user-facing impact.Abstract:
YouTube represents one of the largest scale and most sophisticated industrial recommendation systems in existence. In this paper, we describe the system at a high level and focus on the dramatic performance improvements brought by deep learning. The paper is split according to the classic two-stage information retrieval dichotomy: first, we detail a deep candidate generation model and then describe a separate deep ranking model. We also provide practical lessons and insights derived from designing, iterating and maintaining a massive recommendation system with enormous user-facing impact.read more
Citations
More filters
Journal Article
Algorithmic Distortion of Informational Landscapes
TL;DR: In this paper, a double dichotomy of algorithmic biases on prior information rearrangement and posterior information arrangement is presented, where algorithms empirically appear to expand the cognitive and social horizon of users, from those where they seem to limit that horizon.
Posted Content
Quantifying Long Range Dependence in Language and User Behavior to improve RNNs
TL;DR: The authors proposed a principled estimation procedure of Long Range Dependence (LRD) in sequential datasets based on established LRD theory for real-valued time series and apply it to sequences of symbols with million-item-scale dictionaries.
Proceedings ArticleDOI
Learning to Build User-tag Profile in Recommendation System
TL;DR: This paper proposes a user tag profiling model (UTPM) to study user-tag profiling as a multi-label classification task using deep neural networks and introduces the improved FM-based cross feature layer, which outperforms many state-of-the-art cross feature methods and further enhances model performance.
Proceedings ArticleDOI
Hercules: Heterogeneity-Aware Inference Serving for At-Scale Personalized Recommendation
TL;DR: Hercules is proposed, an optimized framework for personalized recommendation inference serving that targets diverse industry-representative models and cloud-scale heterogeneous systems and reduces the provisioned power by 23.7% over a state-of-the-art greedy scheduler.
Proceedings ArticleDOI
Learning Audio Embeddings with User Listening Data for Content-Based Music Recommendation
TL;DR: Wang et al. as mentioned in this paper explored user listening history and demographics to construct a user embedding representing the user's music preference, which can be obtained for each track using metric learning with Siamese networks.
References
More filters
Proceedings Article
Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift
Sergey Ioffe,Christian Szegedy +1 more
TL;DR: Applied to a state-of-the-art image classification model, Batch Normalization achieves the same accuracy with 14 times fewer training steps, and beats the original model by a significant margin.
Proceedings Article
Distributed Representations of Words and Phrases and their Compositionality
TL;DR: This paper presents a simple method for finding phrases in text, and shows that learning good vector representations for millions of phrases is possible and describes a simple alternative to the hierarchical softmax called negative sampling.
Posted Content
Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift
Sergey Ioffe,Christian Szegedy +1 more
TL;DR: Batch Normalization as mentioned in this paper normalizes layer inputs for each training mini-batch to reduce the internal covariate shift in deep neural networks, and achieves state-of-the-art performance on ImageNet.
Posted Content
Distributed Representations of Words and Phrases and their Compositionality
TL;DR: In this paper, the Skip-gram model is used to learn high-quality distributed vector representations that capture a large number of precise syntactic and semantic word relationships and improve both the quality of the vectors and the training speed.