Deep Neural Networks for YouTube Recommendations
Paul Covington,Jay Adams,Emre Sargin +2 more
- pp 191-198
Reads0
Chats0
TLDR
This paper details a deep candidate generation model and then describes a separate deep ranking model and provides practical lessons and insights derived from designing, iterating and maintaining a massive recommendation system with enormous user-facing impact.Abstract:
YouTube represents one of the largest scale and most sophisticated industrial recommendation systems in existence. In this paper, we describe the system at a high level and focus on the dramatic performance improvements brought by deep learning. The paper is split according to the classic two-stage information retrieval dichotomy: first, we detail a deep candidate generation model and then describe a separate deep ranking model. We also provide practical lessons and insights derived from designing, iterating and maintaining a massive recommendation system with enormous user-facing impact.read more
Citations
More filters
Book ChapterDOI
From fake to junk news : The data politics of online virality
TL;DR: In this paper, the authors argue that fake news is not related to computational analytics or political intentions, but it does mean that this relation is not straightforward and propose a different definition of this phenomenon based on its circulation rather than of its contents.
Proceedings ArticleDOI
User-Video Co-Attention Network for Personalized Micro-video Recommendation
TL;DR: A novel framework User-Video Co-Attention Network (UVCAN) is proposed, which can learn multi-modal information from both user and microvideo side using attention mechanism and reasons about the attention in a stacked attention network fashion for both users and micro-video.
Proceedings ArticleDOI
DGCN: Diversified Recommendation with Graph Convolutional Networks
TL;DR: Zhang et al. as mentioned in this paper proposed to perform rebalanced neighbor discovering, category-boosted negative sampling and adversarial learning on top of GCN to push the diversification to the upstream candidate generation stage, with the help of graph convolutional networks.
Proceedings ArticleDOI
XDL: an industrial deep learning framework for high-dimensional sparse data
Biye Jiang,Chao Deng,Huimin Yi,Zelin Hu,Guorui Zhou,Yang Zheng,Sui Huang,Xinyang Guo,Dongyue Wang,Yue Song,Liqin Zhao,Zhi Wang,Peng Sun,Yu Zhang,Di Zhang,Jinhui Li,Jian Xu,Xiaoqiang Zhu,Kun Gai +18 more
TL;DR: A high-performance, large-scale and distributed DL framework named XDL which provides an elegant solution to fill the gap between general design of existing DL frameworks and industrial requirements arising from high-dimensional sparse data.
Proceedings ArticleDOI
Interact and Decide: Medley of Sub-Attention Networks for Effective Group Recommendation
TL;DR: This paper proposes Medley of Sub-Attention Networks (MoSAN), a new novel neural architecture for the group recommendation task that not only achieves state-of-the-art performance but also improves standard baselines by a considerable margin.
References
More filters
Proceedings Article
Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift
Sergey Ioffe,Christian Szegedy +1 more
TL;DR: Applied to a state-of-the-art image classification model, Batch Normalization achieves the same accuracy with 14 times fewer training steps, and beats the original model by a significant margin.
Proceedings Article
Distributed Representations of Words and Phrases and their Compositionality
TL;DR: This paper presents a simple method for finding phrases in text, and shows that learning good vector representations for millions of phrases is possible and describes a simple alternative to the hierarchical softmax called negative sampling.
Posted Content
Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift
Sergey Ioffe,Christian Szegedy +1 more
TL;DR: Batch Normalization as mentioned in this paper normalizes layer inputs for each training mini-batch to reduce the internal covariate shift in deep neural networks, and achieves state-of-the-art performance on ImageNet.
Posted Content
Distributed Representations of Words and Phrases and their Compositionality
TL;DR: In this paper, the Skip-gram model is used to learn high-quality distributed vector representations that capture a large number of precise syntactic and semantic word relationships and improve both the quality of the vectors and the training speed.