scispace - formally typeset
Proceedings ArticleDOI

When Recurrent Neural Networks meet the Neighborhood for Session-Based Recommendation

TLDR
This work shows based on a comprehensive empirical evaluation that a heuristics-based nearest neighbor (kNN) scheme for sessions outperforms GRU4REC in the large majority of the tested configurations and datasets and ensures the scalability of the kNN method.
Abstract
Deep learning methods have led to substantial progress in various application fields of AI, and in recent years a number of proposals were made to improve recommender systems with artificial neural networks. For the problem of making session-based recommendations, i.e., for recommending the next item in an anonymous session, Hidasi et al.~recently investigated the application of recurrent neural networks with Gated Recurrent Units (GRU4REC). Assessing the true effectiveness of such novel approaches based only on what is reported in the literature is however difficult when no standard evaluation protocols are applied and when the strength of the baselines used in the performance comparison is not clear. In this work we show based on a comprehensive empirical evaluation that a heuristics-based nearest neighbor (kNN) scheme for sessions outperforms GRU4REC in the large majority of the tested configurations and datasets. Neighborhood sampling and efficient in-memory data structures ensure the scalability of the kNN method. The best results in the end were often achieved when we combine the kNN approach with GRU4REC, which shows that RNNs can leverage sequential signals in the data that cannot be detected by the co-occurrence-based kNN method.

read more

Citations
More filters
Journal ArticleDOI

Deep Learning Based Recommender System: A Survey and New Perspectives

TL;DR: A comprehensive review of recent research efforts on deep learning-based recommender systems is provided in this paper, along with a comprehensive summary of the state-of-the-art.
Proceedings ArticleDOI

Personalized Top-N Sequential Recommendation via Convolutional Sequence Embedding

TL;DR: A Convolutional Sequence Embedding Recommendation Model »Caser» is proposed, which is to embed a sequence of recent items into an image in the time and latent spaces and learn sequential patterns as local features of the image using convolutional filters.
Journal ArticleDOI

Session-Based Recommendation with Graph Neural Networks

TL;DR: Wang et al. as discussed by the authors proposed Session-based Recommendation with Graph Neural Networks (SR-GNN) to capture complex transitions of items, which are difficult to be revealed by previous conventional sequential methods.
Journal ArticleDOI

Deep Learning based Recommender System: A Survey and New Perspectives.

TL;DR: A taxonomy of deep learning-based recommendation models is provided and a comprehensive summary of the state of the art is provided, along with new perspectives pertaining to this new and exciting development of the field.
Proceedings ArticleDOI

Are we really making much progress? A worrying analysis of recent neural recommendation approaches

TL;DR: A systematic analysis of algorithmic proposals for top-n recommendation tasks that were presented at top-level research conferences in the last years sheds light on a number of potential problems in today's machine learning scholarship and calls for improved scientific practices in this area.
References
More filters
Journal ArticleDOI

Long short-term memory

TL;DR: A novel, efficient, gradient based method called long short-term memory (LSTM) is introduced, which can learn to bridge minimal time lags in excess of 1000 discrete-time steps by enforcing constant error flow through constant error carousels within special units.
Journal ArticleDOI

Finding Structure in Time

TL;DR: A proposal along these lines first described by Jordan (1986) which involves the use of recurrent links in order to provide networks with a dynamic memory and suggests a method for representing lexical categories and the type/token distinction is developed.
Posted Content

Empirical evaluation of gated recurrent neural networks on sequence modeling

TL;DR: These advanced recurrent units that implement a gating mechanism, such as a long short-term memory (LSTM) unit and a recently proposed gated recurrent unit (GRU), are found to be comparable to LSTM.
Posted Content

Generating Sequences With Recurrent Neural Networks

TL;DR: This paper shows how Long Short-term Memory recurrent neural networks can be used to generate complex sequences with long-range structure, simply by predicting one data point at a time.
Proceedings ArticleDOI

Deep Neural Networks for YouTube Recommendations

TL;DR: This paper details a deep candidate generation model and then describes a separate deep ranking model and provides practical lessons and insights derived from designing, iterating and maintaining a massive recommendation system with enormous user-facing impact.
Related Papers (5)