Tri Kurniawan Wijaya, Edoardo D'Amico, Gabor Fodor, Manuel V. Loureiro
{"title":"RBoard: A Unified Platform for Reproducible and Reusable Recommender System Benchmarks","authors":"Tri Kurniawan Wijaya, Edoardo D'Amico, Gabor Fodor, Manuel V. Loureiro","doi":"arxiv-2409.05526","DOIUrl":null,"url":null,"abstract":"Recommender systems research lacks standardized benchmarks for\nreproducibility and algorithm comparisons. We introduce RBoard, a novel\nframework addressing these challenges by providing a comprehensive platform for\nbenchmarking diverse recommendation tasks, including CTR prediction, Top-N\nrecommendation, and others. RBoard's primary objective is to enable fully\nreproducible and reusable experiments across these scenarios. The framework\nevaluates algorithms across multiple datasets within each task, aggregating\nresults for a holistic performance assessment. It implements standardized\nevaluation protocols, ensuring consistency and comparability. To facilitate\nreproducibility, all user-provided code can be easily downloaded and executed,\nallowing researchers to reliably replicate studies and build upon previous\nwork. By offering a unified platform for rigorous, reproducible evaluation\nacross various recommendation scenarios, RBoard aims to accelerate progress in\nthe field and establish a new standard for recommender systems benchmarking in\nboth academia and industry. The platform is available at https://rboard.org and\nthe demo video can be found at https://bit.ly/rboard-demo.","PeriodicalId":501281,"journal":{"name":"arXiv - CS - Information Retrieval","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2024-09-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Information Retrieval","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.05526","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Recommender systems research lacks standardized benchmarks for
reproducibility and algorithm comparisons. We introduce RBoard, a novel
framework addressing these challenges by providing a comprehensive platform for
benchmarking diverse recommendation tasks, including CTR prediction, Top-N
recommendation, and others. RBoard's primary objective is to enable fully
reproducible and reusable experiments across these scenarios. The framework
evaluates algorithms across multiple datasets within each task, aggregating
results for a holistic performance assessment. It implements standardized
evaluation protocols, ensuring consistency and comparability. To facilitate
reproducibility, all user-provided code can be easily downloaded and executed,
allowing researchers to reliably replicate studies and build upon previous
work. By offering a unified platform for rigorous, reproducible evaluation
across various recommendation scenarios, RBoard aims to accelerate progress in
the field and establish a new standard for recommender systems benchmarking in
both academia and industry. The platform is available at https://rboard.org and
the demo video can be found at https://bit.ly/rboard-demo.