scispace - formally typeset
Open AccessProceedings ArticleDOI

Deep Interest Network for Click-Through Rate Prediction

Reads0
Chats0
TLDR
A novel model: Deep Interest Network (DIN) is proposed which tackles this challenge by designing a local activation unit to adaptively learn the representation of user interests from historical behaviors with respect to a certain ad.
Abstract
Click-through rate prediction is an essential task in industrial applications, such as online advertising. Recently deep learning based models have been proposed, which follow a similar Embedding&MLP paradigm. In these methods large scale sparse input features are first mapped into low dimensional embedding vectors, and then transformed into fixed-length vectors in a group-wise manner, finally concatenated together to fed into a multilayer perceptron (MLP) to learn the nonlinear relations among features. In this way, user features are compressed into a fixed-length representation vector, in regardless of what candidate ads are. The use of fixed-length vector will be a bottleneck, which brings difficulty for Embedding&MLP methods to capture user's diverse interests effectively from rich historical behaviors. In this paper, we propose a novel model: Deep Interest Network (DIN) which tackles this challenge by designing a local activation unit to adaptively learn the representation of user interests from historical behaviors with respect to a certain ad. This representation vector varies over different ads, improving the expressive ability of model greatly. Besides, we develop two techniques: mini-batch aware regularization and data adaptive activation function which can help training industrial deep networks with hundreds of millions of parameters. Experiments on two public datasets as well as an Alibaba real production dataset with over 2 billion samples demonstrate the effectiveness of proposed approaches, which achieve superior performance compared with state-of-the-art methods. DIN now has been successfully deployed in the online display advertising system in Alibaba, serving the main traffic.

read more

Citations
More filters
Proceedings ArticleDOI

Product Knowledge Graph Embedding for E-commerce

TL;DR: The key entities are defined and summarize the pivotal product relations that are critical for general e-commerce applications including marketing, advertisement, search ranking and recommendation.
Proceedings ArticleDOI

One Model to Serve All: Star Topology Adaptive Recommender for Multi-Domain CTR Prediction

TL;DR: Zhang et al. as discussed by the authors proposed the Star Topology Adaptive Recommender (STAR) model to train a single model to serve all domains by leveraging data from all domains simultaneously, capturing the characteristics of each domain, and modeling the commonalities between different domains.
Proceedings ArticleDOI

User Behavior Retrieval for Click-Through Rate Prediction

TL;DR: In UBR4CTR, the most relevant and appropriate user behaviors will be firstly retrieved from the entire user history sequence using a learnable search method and fed into a deep model to make the final prediction instead of simply using the most recent ones.
Proceedings ArticleDOI

Improving Ad Click Prediction by Considering Non-displayed Events

TL;DR: This paper proposes a novel framework for counterfactual CTR prediction by considering not only displayed events but also non-displayed events and compares this framework against state-of-the-art conventional CTR models and existingcounterfactual learning approaches.
Proceedings ArticleDOI

MixGCF: An Improved Training Method for Graph Neural Network-based Recommender Systems

TL;DR: MixGCF as discussed by the authors is a general negative sampling plugin that can be directly used to train GNN-based recommender systems, rather than sampling raw negatives from data, the hop mixing technique to synthesize hard negatives.
References
More filters
Proceedings Article

Adam: A Method for Stochastic Optimization

TL;DR: This work introduces Adam, an algorithm for first-order gradient-based optimization of stochastic objective functions, based on adaptive estimates of lower-order moments, and provides a regret bound on the convergence rate that is comparable to the best known results under the online convex optimization framework.
Journal ArticleDOI

Regression Shrinkage and Selection via the Lasso

TL;DR: A new method for estimation in linear models called the lasso, which minimizes the residual sum of squares subject to the sum of the absolute value of the coefficients being less than a constant, is proposed.
Journal Article

Dropout: a simple way to prevent neural networks from overfitting

TL;DR: It is shown that dropout improves the performance of neural networks on supervised learning tasks in vision, speech recognition, document classification and computational biology, obtaining state-of-the-art results on many benchmark data sets.
Journal Article

Visualizing Data using t-SNE

TL;DR: A new technique called t-SNE that visualizes high-dimensional data by giving each datapoint a location in a two or three-dimensional map, a variation of Stochastic Neighbor Embedding that is much easier to optimize, and produces significantly better visualizations by reducing the tendency to crowd points together in the center of the map.
Proceedings ArticleDOI

Densely Connected Convolutional Networks

TL;DR: DenseNet as mentioned in this paper proposes to connect each layer to every other layer in a feed-forward fashion, which can alleviate the vanishing gradient problem, strengthen feature propagation, encourage feature reuse, and substantially reduce the number of parameters.
Related Papers (5)
Trending Questions (1)
How does Google Ads use historical data and machine learning algorithms to predict keyword performance?

The provided paper does not mention Google Ads or how historical data and machine learning algorithms are used to predict keyword performance. The paper focuses on click-through rate prediction in online advertising, specifically in the Alibaba display advertising system.