{"title":"Multilingual Prompts in LLM-Based Recommenders: Performance Across Languages","authors":"Makbule Gulcin Ozsoy","doi":"arxiv-2409.07604","DOIUrl":null,"url":null,"abstract":"Large language models (LLMs) are increasingly used in natural language\nprocessing tasks. Recommender systems traditionally use methods such as\ncollaborative filtering and matrix factorization, as well as advanced\ntechniques like deep learning and reinforcement learning. Although language\nmodels have been applied in recommendation, the recent trend have focused on\nleveraging the generative capabilities of LLMs for more personalized\nsuggestions. While current research focuses on English due to its resource\nrichness, this work explores the impact of non-English prompts on\nrecommendation performance. Using OpenP5, a platform for developing and\nevaluating LLM-based recommendations, we expanded its English prompt templates\nto include Spanish and Turkish. Evaluation on three real-world datasets, namely\nML1M, LastFM, and Amazon-Beauty, showed that usage of non-English prompts\ngenerally reduce performance, especially in less-resourced languages like\nTurkish. We also retrained an LLM-based recommender model with multilingual\nprompts to analyze performance variations. Retraining with multilingual prompts\nresulted in more balanced performance across languages, but slightly reduced\nEnglish performance. This work highlights the need for diverse language support\nin LLM-based recommenders and suggests future research on creating evaluation\ndatasets, using newer models and additional languages.","PeriodicalId":501281,"journal":{"name":"arXiv - CS - Information Retrieval","volume":"4 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-09-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Information Retrieval","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.07604","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Large language models (LLMs) are increasingly used in natural language
processing tasks. Recommender systems traditionally use methods such as
collaborative filtering and matrix factorization, as well as advanced
techniques like deep learning and reinforcement learning. Although language
models have been applied in recommendation, the recent trend have focused on
leveraging the generative capabilities of LLMs for more personalized
suggestions. While current research focuses on English due to its resource
richness, this work explores the impact of non-English prompts on
recommendation performance. Using OpenP5, a platform for developing and
evaluating LLM-based recommendations, we expanded its English prompt templates
to include Spanish and Turkish. Evaluation on three real-world datasets, namely
ML1M, LastFM, and Amazon-Beauty, showed that usage of non-English prompts
generally reduce performance, especially in less-resourced languages like
Turkish. We also retrained an LLM-based recommender model with multilingual
prompts to analyze performance variations. Retraining with multilingual prompts
resulted in more balanced performance across languages, but slightly reduced
English performance. This work highlights the need for diverse language support
in LLM-based recommenders and suggests future research on creating evaluation
datasets, using newer models and additional languages.