Tri Kurniawan Wijaya, Edoardo D'Amico, Gabor Fodor, Manuel V. Loureiro
{"title":"RBoard:可重复使用的推荐系统基准统一平台","authors":"Tri Kurniawan Wijaya, Edoardo D'Amico, Gabor Fodor, Manuel V. Loureiro","doi":"arxiv-2409.05526","DOIUrl":null,"url":null,"abstract":"Recommender systems research lacks standardized benchmarks for\nreproducibility and algorithm comparisons. We introduce RBoard, a novel\nframework addressing these challenges by providing a comprehensive platform for\nbenchmarking diverse recommendation tasks, including CTR prediction, Top-N\nrecommendation, and others. RBoard's primary objective is to enable fully\nreproducible and reusable experiments across these scenarios. The framework\nevaluates algorithms across multiple datasets within each task, aggregating\nresults for a holistic performance assessment. It implements standardized\nevaluation protocols, ensuring consistency and comparability. To facilitate\nreproducibility, all user-provided code can be easily downloaded and executed,\nallowing researchers to reliably replicate studies and build upon previous\nwork. By offering a unified platform for rigorous, reproducible evaluation\nacross various recommendation scenarios, RBoard aims to accelerate progress in\nthe field and establish a new standard for recommender systems benchmarking in\nboth academia and industry. The platform is available at https://rboard.org and\nthe demo video can be found at https://bit.ly/rboard-demo.","PeriodicalId":501281,"journal":{"name":"arXiv - CS - Information Retrieval","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2024-09-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"RBoard: A Unified Platform for Reproducible and Reusable Recommender System Benchmarks\",\"authors\":\"Tri Kurniawan Wijaya, Edoardo D'Amico, Gabor Fodor, Manuel V. Loureiro\",\"doi\":\"arxiv-2409.05526\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Recommender systems research lacks standardized benchmarks for\\nreproducibility and algorithm comparisons. We introduce RBoard, a novel\\nframework addressing these challenges by providing a comprehensive platform for\\nbenchmarking diverse recommendation tasks, including CTR prediction, Top-N\\nrecommendation, and others. RBoard's primary objective is to enable fully\\nreproducible and reusable experiments across these scenarios. The framework\\nevaluates algorithms across multiple datasets within each task, aggregating\\nresults for a holistic performance assessment. It implements standardized\\nevaluation protocols, ensuring consistency and comparability. To facilitate\\nreproducibility, all user-provided code can be easily downloaded and executed,\\nallowing researchers to reliably replicate studies and build upon previous\\nwork. By offering a unified platform for rigorous, reproducible evaluation\\nacross various recommendation scenarios, RBoard aims to accelerate progress in\\nthe field and establish a new standard for recommender systems benchmarking in\\nboth academia and industry. The platform is available at https://rboard.org and\\nthe demo video can be found at https://bit.ly/rboard-demo.\",\"PeriodicalId\":501281,\"journal\":{\"name\":\"arXiv - CS - Information Retrieval\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-09-09\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - CS - Information Retrieval\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2409.05526\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Information Retrieval","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.05526","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
RBoard: A Unified Platform for Reproducible and Reusable Recommender System Benchmarks
Recommender systems research lacks standardized benchmarks for
reproducibility and algorithm comparisons. We introduce RBoard, a novel
framework addressing these challenges by providing a comprehensive platform for
benchmarking diverse recommendation tasks, including CTR prediction, Top-N
recommendation, and others. RBoard's primary objective is to enable fully
reproducible and reusable experiments across these scenarios. The framework
evaluates algorithms across multiple datasets within each task, aggregating
results for a holistic performance assessment. It implements standardized
evaluation protocols, ensuring consistency and comparability. To facilitate
reproducibility, all user-provided code can be easily downloaded and executed,
allowing researchers to reliably replicate studies and build upon previous
work. By offering a unified platform for rigorous, reproducible evaluation
across various recommendation scenarios, RBoard aims to accelerate progress in
the field and establish a new standard for recommender systems benchmarking in
both academia and industry. The platform is available at https://rboard.org and
the demo video can be found at https://bit.ly/rboard-demo.