4.5 Article

AFramework for Evaluating Personalized Ranking Systems by Fusing Different Evaluation Measures

Journal

BIG DATA RESEARCH
Volume 25, Issue -, Pages -

Publisher

ELSEVIER
DOI: 10.1016/j.bdr.2021.100211

Keywords

Recommender systems; Evaluation; Benchmarking

Funding

  1. Slovenian Research Agency [P2-0098, P2-0103, Z2-1867, J2-9230]

Ask authors/readers for more resources

Personalized ranking systems, also known as recommender systems, employ various big data methods, but existing performance measures do not effectively assist end-users in selecting suitable algorithms. To address this issue, we introduce a novel benchmarking framework that combines different evaluation measures to rank recommender systems on individual benchmark datasets.
Personalized ranking systems - also known as recommender systems - use different big data methods, including collaborative filtering, graph random-walks, matrix factorization, and latent-factor models. With their wide use in various social-network, e-commerce, and content platforms, online platforms and developers are in need of better ways to choose the systems that are most suitable for their usecases. At the same time, the research literature on recommender systems describes a multitude of performance measures to evaluate the performance of different algorithms. For the end-user however, the large number of available measures do not provide much help in deciding which algorithm to deploy. Some of the measures are correlated, while others deal with different aspects of recommendation performance like accuracy and diversity. To address this problem, we propose a novel benchmarking framework that mixes different evaluation measures in order to rank the recommender systems on each benchmark dataset, separately. Additionally, our approach discovers sets of correlated measures as well as sets of evaluation measures that are least correlated. We investigate the robustness of the proposed methodology using published results from an experimental study involving multiple big datasets and evaluation measures. Our work provides a general framework that can handle an arbitrary number of evaluation measures and help end-users rank the systems available to them. (C) 2021 The Author(s). Published by Elsevier Inc.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.5
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available