scispace - formally typeset
Open AccessProceedings ArticleDOI

Deep Neural Networks for YouTube Recommendations

Paul Covington, +2 more
- pp 191-198
Reads0
Chats0
TLDR
This paper details a deep candidate generation model and then describes a separate deep ranking model and provides practical lessons and insights derived from designing, iterating and maintaining a massive recommendation system with enormous user-facing impact.
Abstract
YouTube represents one of the largest scale and most sophisticated industrial recommendation systems in existence. In this paper, we describe the system at a high level and focus on the dramatic performance improvements brought by deep learning. The paper is split according to the classic two-stage information retrieval dichotomy: first, we detail a deep candidate generation model and then describe a separate deep ranking model. We also provide practical lessons and insights derived from designing, iterating and maintaining a massive recommendation system with enormous user-facing impact.

read more

Citations
More filters
Journal Article

Algorithmic Distortion of Informational Landscapes

Camille Roth
- 21 Jul 2019 - 
TL;DR: In this paper, a double dichotomy of algorithmic biases on prior information rearrangement and posterior information arrangement is presented, where algorithms empirically appear to expand the cognitive and social horizon of users, from those where they seem to limit that horizon.
Posted Content

Quantifying Long Range Dependence in Language and User Behavior to improve RNNs

TL;DR: The authors proposed a principled estimation procedure of Long Range Dependence (LRD) in sequential datasets based on established LRD theory for real-valued time series and apply it to sequences of symbols with million-item-scale dictionaries.
Proceedings ArticleDOI

Learning to Build User-tag Profile in Recommendation System

TL;DR: This paper proposes a user tag profiling model (UTPM) to study user-tag profiling as a multi-label classification task using deep neural networks and introduces the improved FM-based cross feature layer, which outperforms many state-of-the-art cross feature methods and further enhances model performance.
Proceedings ArticleDOI

Hercules: Heterogeneity-Aware Inference Serving for At-Scale Personalized Recommendation

TL;DR: Hercules is proposed, an optimized framework for personalized recommendation inference serving that targets diverse industry-representative models and cloud-scale heterogeneous systems and reduces the provisioned power by 23.7% over a state-of-the-art greedy scheduler.
Proceedings ArticleDOI

Learning Audio Embeddings with User Listening Data for Content-Based Music Recommendation

TL;DR: Wang et al. as mentioned in this paper explored user listening history and demographics to construct a user embedding representing the user's music preference, which can be obtained for each track using metric learning with Siamese networks.
References
More filters
Book ChapterDOI

I and J

Proceedings Article

Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift

TL;DR: Applied to a state-of-the-art image classification model, Batch Normalization achieves the same accuracy with 14 times fewer training steps, and beats the original model by a significant margin.
Proceedings Article

Distributed Representations of Words and Phrases and their Compositionality

TL;DR: This paper presents a simple method for finding phrases in text, and shows that learning good vector representations for millions of phrases is possible and describes a simple alternative to the hierarchical softmax called negative sampling.
Posted Content

Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift

TL;DR: Batch Normalization as mentioned in this paper normalizes layer inputs for each training mini-batch to reduce the internal covariate shift in deep neural networks, and achieves state-of-the-art performance on ImageNet.
Posted Content

Distributed Representations of Words and Phrases and their Compositionality

TL;DR: In this paper, the Skip-gram model is used to learn high-quality distributed vector representations that capture a large number of precise syntactic and semantic word relationships and improve both the quality of the vectors and the training speed.