scispace - formally typeset
Open AccessProceedings ArticleDOI

Deep Neural Networks for YouTube Recommendations

Paul Covington, +2 more
- pp 191-198
TLDR
This paper details a deep candidate generation model and then describes a separate deep ranking model and provides practical lessons and insights derived from designing, iterating and maintaining a massive recommendation system with enormous user-facing impact.
Abstract
YouTube represents one of the largest scale and most sophisticated industrial recommendation systems in existence. In this paper, we describe the system at a high level and focus on the dramatic performance improvements brought by deep learning. The paper is split according to the classic two-stage information retrieval dichotomy: first, we detail a deep candidate generation model and then describe a separate deep ranking model. We also provide practical lessons and insights derived from designing, iterating and maintaining a massive recommendation system with enormous user-facing impact.

read more

Citations
More filters
Posted Content

Improving Native Ads CTR Prediction by Large Scale Event Embedding and Recurrent Networks

TL;DR: This paper proposes a large-scale event embedding scheme to encode the each user browsing event by training a Siamese network with weak supervision on the users' consecutive events to solve the click through rate prediction problem.
Proceedings ArticleDOI

iMARS: an in-memory-computing architecture for recommendation systems

TL;DR: An in-memory-computing (IMC) architecture (iMARS) for accelerating the filtering and ranking stages of deep neural network-based RecSys, which leverages IMC-friendly embedding tables implemented inside a ferroelectric FET based IMC fabric.
Posted Content

RecSim NG: Toward Principled Uncertainty Modeling for Recommender Ecosystems.

TL;DR: RecSim NG as discussed by the authors is a probabilistic platform for the simulation of multi-agent recommender systems, which can be used to create transparent, configurable, end-to-end models of a recommender ecosystem.
Proceedings ArticleDOI

Candidate Selection for Large Scale Personalized Search and Recommender Systems

TL;DR: This tutorial surveys various candidate selection techniques and deep dives into case studies on a large scale social media platform and provides hands-on tutorial where the author explores building these candidate selection models on a real world dataset and sees how to balance the tradeoff between relevance and latency.
Proceedings ArticleDOI

A Deep Prediction Network for Understanding Advertiser Intent and Satisfaction

TL;DR: Wang et al. as mentioned in this paper proposed a Deep Satisfaction Prediction Network (DSPN), which jointly learns advertiser intent vector and satisfaction by considering the features of advertiser's action information and advertising performance indicators.
References
More filters
Book ChapterDOI

I and J

Proceedings Article

Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift

TL;DR: Applied to a state-of-the-art image classification model, Batch Normalization achieves the same accuracy with 14 times fewer training steps, and beats the original model by a significant margin.
Proceedings Article

Distributed Representations of Words and Phrases and their Compositionality

TL;DR: This paper presents a simple method for finding phrases in text, and shows that learning good vector representations for millions of phrases is possible and describes a simple alternative to the hierarchical softmax called negative sampling.
Posted Content

Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift

TL;DR: Batch Normalization as mentioned in this paper normalizes layer inputs for each training mini-batch to reduce the internal covariate shift in deep neural networks, and achieves state-of-the-art performance on ImageNet.
Posted Content

Distributed Representations of Words and Phrases and their Compositionality

TL;DR: In this paper, the Skip-gram model is used to learn high-quality distributed vector representations that capture a large number of precise syntactic and semantic word relationships and improve both the quality of the vectors and the training speed.