Deep Interest Network for Click-Through Rate Prediction
Guorui Zhou,Xiaoqiang Zhu,Chenru Song,Ying Fan,Han Zhu,Xiao Ma,Yan Yanghui,Junqi Jin,Han Li,Kun Gai +9 more
- pp 1059-1068
Reads0
Chats0
TLDR
A novel model: Deep Interest Network (DIN) is proposed which tackles this challenge by designing a local activation unit to adaptively learn the representation of user interests from historical behaviors with respect to a certain ad.Abstract:
Click-through rate prediction is an essential task in industrial applications, such as online advertising. Recently deep learning based models have been proposed, which follow a similar Embedding&MLP paradigm. In these methods large scale sparse input features are first mapped into low dimensional embedding vectors, and then transformed into fixed-length vectors in a group-wise manner, finally concatenated together to fed into a multilayer perceptron (MLP) to learn the nonlinear relations among features. In this way, user features are compressed into a fixed-length representation vector, in regardless of what candidate ads are. The use of fixed-length vector will be a bottleneck, which brings difficulty for Embedding&MLP methods to capture user's diverse interests effectively from rich historical behaviors. In this paper, we propose a novel model: Deep Interest Network (DIN) which tackles this challenge by designing a local activation unit to adaptively learn the representation of user interests from historical behaviors with respect to a certain ad. This representation vector varies over different ads, improving the expressive ability of model greatly. Besides, we develop two techniques: mini-batch aware regularization and data adaptive activation function which can help training industrial deep networks with hundreds of millions of parameters. Experiments on two public datasets as well as an Alibaba real production dataset with over 2 billion samples demonstrate the effectiveness of proposed approaches, which achieve superior performance compared with state-of-the-art methods. DIN now has been successfully deployed in the online display advertising system in Alibaba, serving the main traffic.read more
Citations
More filters
Proceedings ArticleDOI
AliMeKG: Domain Knowledge Graph Construction and Application in E-commerce
TL;DR: AliMe KG as mentioned in this paper is a domain knowledge graph in E-commerce that captures user problems, points of inter-est (POI), item information and relations thereof, which helps to under-stand user needs, answer pre-sales questions and generate explana-tion texts.
Proceedings ArticleDOI
Neural Auction: End-to-End Learning of Auction Mechanisms for E-Commerce Advertising
Xiangyu Liu,Chuan Yu,Zhilin Zhang,Zhenzhe Zheng,Yu Rong,Hongtao Lv,Da Huo,Yiqing Wang,Dagui Chen,Jian Xu,Fan Wu,Guihai Chen,Xiaoqiang Zhu +12 more
TL;DR: In this article, Deep Neural Auctions (DNAs) are proposed to enable end-to-end auction learning by proposing a differentiable model to relax the discrete sorting operation, a key component in auctions.
Proceedings ArticleDOI
Personalized Transfer of User Preferences for Cross-domain Recommendation
TL;DR: Wang et al. as discussed by the authors proposed a novel framework named Personalized Transfer of User Preferences for Cross-domain Recommendation (PTUPCDR), which uses a meta network fed with users' characteristic embeddings to generate personalized bridge functions to achieve personalized transfer of preferences for each user.
Proceedings ArticleDOI
Deep Time-Aware Item Evolution Network for Click-Through Rate Prediction
TL;DR: This work proposes a CTR prediction model TIEN based on the time-aware item behavior and shows that the TIEN model consistently achieves remarkable improvements to the state-of-the-art methods.
Proceedings ArticleDOI
FleetRec: Large-Scale Recommendation Inference on Hybrid GPU-FPGA Clusters
Wenqi Jiang,Zhenhao He,Shuai Zhang,Kai Zeng,Liang Feng,Jiansong Zhang,Tongxuan Liu,Yong Li,Jingren Zhou,Ce Zhang,Gustavo Alonso +10 more
TL;DR: Experiments on three production models show that FleetRec outperforms optimized CPU baseline by more than one order of magnitude in terms of throughput while achieving significantly lower latency.
References
More filters
Proceedings Article
Adam: A Method for Stochastic Optimization
Diederik P. Kingma,Jimmy Ba +1 more
TL;DR: This work introduces Adam, an algorithm for first-order gradient-based optimization of stochastic objective functions, based on adaptive estimates of lower-order moments, and provides a regret bound on the convergence rate that is comparable to the best known results under the online convex optimization framework.
Journal ArticleDOI
Regression Shrinkage and Selection via the Lasso
TL;DR: A new method for estimation in linear models called the lasso, which minimizes the residual sum of squares subject to the sum of the absolute value of the coefficients being less than a constant, is proposed.
Journal Article
Dropout: a simple way to prevent neural networks from overfitting
TL;DR: It is shown that dropout improves the performance of neural networks on supervised learning tasks in vision, speech recognition, document classification and computational biology, obtaining state-of-the-art results on many benchmark data sets.
Journal Article
Visualizing Data using t-SNE
TL;DR: A new technique called t-SNE that visualizes high-dimensional data by giving each datapoint a location in a two or three-dimensional map, a variation of Stochastic Neighbor Embedding that is much easier to optimize, and produces significantly better visualizations by reducing the tendency to crowd points together in the center of the map.
Proceedings ArticleDOI
Densely Connected Convolutional Networks
TL;DR: DenseNet as mentioned in this paper proposes to connect each layer to every other layer in a feed-forward fashion, which can alleviate the vanishing gradient problem, strengthen feature propagation, encourage feature reuse, and substantially reduce the number of parameters.