scispace - formally typeset
Open AccessProceedings ArticleDOI

Deep Neural Networks for YouTube Recommendations

Paul Covington, +2 more
- pp 191-198
Reads0
Chats0
TLDR
This paper details a deep candidate generation model and then describes a separate deep ranking model and provides practical lessons and insights derived from designing, iterating and maintaining a massive recommendation system with enormous user-facing impact.
Abstract
YouTube represents one of the largest scale and most sophisticated industrial recommendation systems in existence. In this paper, we describe the system at a high level and focus on the dramatic performance improvements brought by deep learning. The paper is split according to the classic two-stage information retrieval dichotomy: first, we detail a deep candidate generation model and then describe a separate deep ranking model. We also provide practical lessons and insights derived from designing, iterating and maintaining a massive recommendation system with enormous user-facing impact.

read more

Citations
More filters
Book ChapterDOI

From fake to junk news : The data politics of online virality

TL;DR: In this paper, the authors argue that fake news is not related to computational analytics or political intentions, but it does mean that this relation is not straightforward and propose a different definition of this phenomenon based on its circulation rather than of its contents.
Proceedings ArticleDOI

User-Video Co-Attention Network for Personalized Micro-video Recommendation

TL;DR: A novel framework User-Video Co-Attention Network (UVCAN) is proposed, which can learn multi-modal information from both user and microvideo side using attention mechanism and reasons about the attention in a stacked attention network fashion for both users and micro-video.
Proceedings ArticleDOI

DGCN: Diversified Recommendation with Graph Convolutional Networks

TL;DR: Zhang et al. as mentioned in this paper proposed to perform rebalanced neighbor discovering, category-boosted negative sampling and adversarial learning on top of GCN to push the diversification to the upstream candidate generation stage, with the help of graph convolutional networks.
Proceedings ArticleDOI

XDL: an industrial deep learning framework for high-dimensional sparse data

TL;DR: A high-performance, large-scale and distributed DL framework named XDL which provides an elegant solution to fill the gap between general design of existing DL frameworks and industrial requirements arising from high-dimensional sparse data.
Proceedings ArticleDOI

Interact and Decide: Medley of Sub-Attention Networks for Effective Group Recommendation

TL;DR: This paper proposes Medley of Sub-Attention Networks (MoSAN), a new novel neural architecture for the group recommendation task that not only achieves state-of-the-art performance but also improves standard baselines by a considerable margin.
References
More filters
Book ChapterDOI

I and J

Proceedings Article

Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift

TL;DR: Applied to a state-of-the-art image classification model, Batch Normalization achieves the same accuracy with 14 times fewer training steps, and beats the original model by a significant margin.
Proceedings Article

Distributed Representations of Words and Phrases and their Compositionality

TL;DR: This paper presents a simple method for finding phrases in text, and shows that learning good vector representations for millions of phrases is possible and describes a simple alternative to the hierarchical softmax called negative sampling.
Posted Content

Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift

TL;DR: Batch Normalization as mentioned in this paper normalizes layer inputs for each training mini-batch to reduce the internal covariate shift in deep neural networks, and achieves state-of-the-art performance on ImageNet.
Posted Content

Distributed Representations of Words and Phrases and their Compositionality

TL;DR: In this paper, the Skip-gram model is used to learn high-quality distributed vector representations that capture a large number of precise syntactic and semantic word relationships and improve both the quality of the vectors and the training speed.