scispace - formally typeset
Book ChapterDOI

Evaluating Recommender Systems

Reads0
Chats0
TLDR
This paper discusses how to compare recommenders based on a set of properties that are relevant for the application, and focuses on comparative studies, where a few algorithms are compared using some evaluation metric, rather than absolute benchmarking of algorithms.
Abstract
Recommender systems are now popular both commercially and in the research community, where many approaches have been suggested for providing recommendations. In many cases a system designer that wishes to employ a recommendater system must choose between a set of candidate approaches. A first step towards selecting an appropriate algorithm is to decide which properties of the application to focus upon when making this choice. Indeed, recommender systems have a variety of properties that may affect user experience, such as accuracy, robustness, scalability, and so forth. In this paper we discuss how to compare recommenders based on a set of properties that are relevant for the application. We focus on comparative studies, where a few algorithms are compared using some evaluation metric, rather than absolute benchmarking of algorithms. We describe experimental settings appropriate for making choices between algorithms. We review three types of experiments, starting with an offline setting, where recommendation approaches are compared without user interaction, then reviewing user studies, where a small group of subjects experiment with the system and report on the experience, and finally describe large scale online experiments, where real user populations interact with the system. In each of these cases we describe types of questions that can be answered, and suggest protocols for experimentation. We also discuss how to draw trustworthy conclusions from the conducted experiments. We then review a large set of properties, and explain how to evaluate systems given relevant properties. We also survey a large set of evaluation metrics in the context of the property that they evaluate.

read more

Content maybe subject to copyright    Report

Citations
More filters
Journal ArticleDOI

The use of machine learning algorithms in recommender systems: A systematic review

TL;DR: The study concludes that Bayesian and decision tree algorithms are widely used in recommender systems because of their relative simplicity, and that requirement and design phases of recommender system development appear to offer opportunities for further research.
Posted Content

The Use of Machine Learning Algorithms in Recommender Systems: A Systematic Review

TL;DR: In this paper, the authors present a systematic review of the literature that analyzes the use of machine learning algorithms in recommender systems and identifies research opportunities for software engineering research, and conclude that Bayesian and decision tree algorithms are widely used in recommendation systems because of their relative simplicity and that requirement and design phases of recommender system development appear to offer opportunities for further research.
Journal ArticleDOI

Current challenges and visions in music recommender systems research

TL;DR: In this article, the authors identify and shed light on what they believe are the most pressing challenges in recommender systems from both academic and industry perspectives, and detail possible future directions and visions for the further evolution of the field.
Journal ArticleDOI

Facebook single and cross domain data for recommendation systems

TL;DR: This paper investigates the feasibility and effectiveness of utilizing existing available data from social networks for the recommendation process, specifically from Facebook, and highlights the benefits of utilizing cross domain Facebook data to achieve improvement in recommendation performance.
Proceedings ArticleDOI

Let Me Explain: Impact of Personal and Impersonal Explanations on Trust in Recommender Systems

TL;DR: It is suggested that RS should provide richer explanations in order to increase their perceived recommendation quality and trustworthiness.
References
More filters
Journal ArticleDOI

Controlling the false discovery rate: a practical and powerful approach to multiple testing

TL;DR: In this paper, a different approach to problems of multiple significance testing is presented, which calls for controlling the expected proportion of falsely rejected hypotheses -the false discovery rate, which is equivalent to the FWER when all hypotheses are true but is smaller otherwise.
Journal Article

Statistical Comparisons of Classifiers over Multiple Data Sets

TL;DR: A set of simple, yet safe and robust non-parametric tests for statistical comparisons of classifiers is recommended: the Wilcoxon signed ranks test for comparison of two classifiers and the Friedman test with the corresponding post-hoc tests for comparisons of more classifiers over multiple data sets.
Proceedings ArticleDOI

Item-based collaborative filtering recommendation algorithms

TL;DR: This paper analyzes item-based collaborative ltering techniques and suggests that item- based algorithms provide dramatically better performance than user-based algorithms, while at the same time providing better quality than the best available userbased algorithms.
Journal ArticleDOI

A new measure of rank correlation

Maurice G. Kendall
- 01 Jun 1938 - 
TL;DR: Rank correlation as mentioned in this paper is a measure of similarity between two rankings of the same set of individuals, and it has been used in psychological work to compare two different rankings of individuals in order to indicate similarity of taste.
Journal ArticleDOI

Evaluating collaborative filtering recommender systems

TL;DR: The key decisions in evaluating collaborative filtering recommender systems are reviewed: the user tasks being evaluated, the types of analysis and datasets being used, the ways in which prediction quality is measured, the evaluation of prediction attributes other than quality, and the user-based evaluation of the system as a whole.
Related Papers (5)