{"title":"Logic-Enhanced Language Model Agents for Trustworthy Social Simulations","authors":"Agnieszka Mensfelt, Kostas Stathis, Vince Trencsenyi","doi":"arxiv-2408.16081","DOIUrl":null,"url":null,"abstract":"We introduce the Logic-Enhanced Language Model Agents (LELMA) framework, a\nnovel approach to enhance the trustworthiness of social simulations that\nutilize large language models (LLMs). While LLMs have gained attention as\nagents for simulating human behaviour, their applicability in this role is\nlimited by issues such as inherent hallucinations and logical inconsistencies.\nLELMA addresses these challenges by integrating LLMs with symbolic AI, enabling\nlogical verification of the reasoning generated by LLMs. This verification\nprocess provides corrective feedback, refining the reasoning output. The\nframework consists of three main components: an LLM-Reasoner for producing\nstrategic reasoning, an LLM-Translator for mapping natural language reasoning\nto logic queries, and a Solver for evaluating these queries. This study focuses\non decision-making in game-theoretic scenarios as a model of human interaction.\nExperiments involving the Hawk-Dove game, Prisoner's Dilemma, and Stag Hunt\nhighlight the limitations of state-of-the-art LLMs, GPT-4 Omni and Gemini 1.0\nPro, in producing correct reasoning in these contexts. LELMA demonstrates high\naccuracy in error detection and improves the reasoning correctness of LLMs via\nself-refinement, particularly in GPT-4 Omni.","PeriodicalId":501208,"journal":{"name":"arXiv - CS - Logic in Computer Science","volume":"24 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-08-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Logic in Computer Science","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2408.16081","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
We introduce the Logic-Enhanced Language Model Agents (LELMA) framework, a
novel approach to enhance the trustworthiness of social simulations that
utilize large language models (LLMs). While LLMs have gained attention as
agents for simulating human behaviour, their applicability in this role is
limited by issues such as inherent hallucinations and logical inconsistencies.
LELMA addresses these challenges by integrating LLMs with symbolic AI, enabling
logical verification of the reasoning generated by LLMs. This verification
process provides corrective feedback, refining the reasoning output. The
framework consists of three main components: an LLM-Reasoner for producing
strategic reasoning, an LLM-Translator for mapping natural language reasoning
to logic queries, and a Solver for evaluating these queries. This study focuses
on decision-making in game-theoretic scenarios as a model of human interaction.
Experiments involving the Hawk-Dove game, Prisoner's Dilemma, and Stag Hunt
highlight the limitations of state-of-the-art LLMs, GPT-4 Omni and Gemini 1.0
Pro, in producing correct reasoning in these contexts. LELMA demonstrates high
accuracy in error detection and improves the reasoning correctness of LLMs via
self-refinement, particularly in GPT-4 Omni.