CauseJudger:用 LLMs 找出原因,进行归纳逻辑推理

Jinwei He, Feng Lu
{"title":"CauseJudger:用 LLMs 找出原因,进行归纳逻辑推理","authors":"Jinwei He, Feng Lu","doi":"arxiv-2409.05559","DOIUrl":null,"url":null,"abstract":"Large language models (LLMs) have been utilized in solving diverse reasoning\ntasks, encompassing common sense, arithmetic and deduction tasks. However, with\ndifficulties of reversing thinking patterns and irrelevant premises, how to\ndetermine the authenticity of the cause in abductive logical reasoning remains\nunderexplored. Inspired by hypothesis and verification method and\nidentification of irrelevant information in human thinking process, we propose\na new framework for LLMs abductive logical reasoning called CauseJudger (CJ),\nwhich identifies the authenticity of possible cause by transforming thinking\nfrom reverse to forward and removing irrelevant information. In addition, we\nconstruct an abductive logical reasoning dataset for decision task called\nCauseLogics, which contains 200,000 tasks of varying reasoning lengths. Our\nexperiments show the efficiency of CJ with overall experiments and ablation\nexperiments as well as case studies on our dataset and reconstructed public\ndataset. Notably, CJ's implementation is efficient, requiring only two calls to\nLLM. Its impact is profound: when using gpt-3.5, CJ achieves a maximum\ncorrectness improvement of 41% compared to Zero-Shot-CoT. Moreover, with gpt-4,\nCJ attains an accuracy exceeding 90% across all datasets.","PeriodicalId":501479,"journal":{"name":"arXiv - CS - Artificial Intelligence","volume":"62 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-09-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"CauseJudger: Identifying the Cause with LLMs for Abductive Logical Reasoning\",\"authors\":\"Jinwei He, Feng Lu\",\"doi\":\"arxiv-2409.05559\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Large language models (LLMs) have been utilized in solving diverse reasoning\\ntasks, encompassing common sense, arithmetic and deduction tasks. However, with\\ndifficulties of reversing thinking patterns and irrelevant premises, how to\\ndetermine the authenticity of the cause in abductive logical reasoning remains\\nunderexplored. Inspired by hypothesis and verification method and\\nidentification of irrelevant information in human thinking process, we propose\\na new framework for LLMs abductive logical reasoning called CauseJudger (CJ),\\nwhich identifies the authenticity of possible cause by transforming thinking\\nfrom reverse to forward and removing irrelevant information. In addition, we\\nconstruct an abductive logical reasoning dataset for decision task called\\nCauseLogics, which contains 200,000 tasks of varying reasoning lengths. Our\\nexperiments show the efficiency of CJ with overall experiments and ablation\\nexperiments as well as case studies on our dataset and reconstructed public\\ndataset. Notably, CJ's implementation is efficient, requiring only two calls to\\nLLM. Its impact is profound: when using gpt-3.5, CJ achieves a maximum\\ncorrectness improvement of 41% compared to Zero-Shot-CoT. Moreover, with gpt-4,\\nCJ attains an accuracy exceeding 90% across all datasets.\",\"PeriodicalId\":501479,\"journal\":{\"name\":\"arXiv - CS - Artificial Intelligence\",\"volume\":\"62 1\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-09-09\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - CS - Artificial Intelligence\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2409.05559\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Artificial Intelligence","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.05559","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

大语言模型(LLM)已被用于解决各种推理任务,包括常识、算术和演绎任务。然而,在归纳逻辑推理中,由于存在思维模式逆转和前提不相关等困难,如何确定原因的真伪仍未得到探索。受假设和验证方法以及人类思维过程中无关信息识别方法的启发,我们提出了一种新的LLMs归纳逻辑推理框架,称为 "原因评判器(CJ)",它通过将思维从逆向转化为正向并去除无关信息来识别可能原因的真实性。此外,我们还构建了一个名为 "原因逻辑"(CauseLogics)的决策任务归纳逻辑推理数据集,其中包含 200,000 个推理长度各不相同的任务。我们通过整体实验、消融实验以及在我们的数据集和重建的公共数据集上进行的案例研究,展示了 CJ 的效率。值得注意的是,CJ 的实现非常高效,只需要调用两次LLM。它的影响是深远的:在使用 gpt-3.5 时,CJ 与 Zero-Shot-CoT 相比,最大正确率提高了 41%。此外,在使用 gpt-4 时,CJ 在所有数据集上的正确率都超过了 90%。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
CauseJudger: Identifying the Cause with LLMs for Abductive Logical Reasoning
Large language models (LLMs) have been utilized in solving diverse reasoning tasks, encompassing common sense, arithmetic and deduction tasks. However, with difficulties of reversing thinking patterns and irrelevant premises, how to determine the authenticity of the cause in abductive logical reasoning remains underexplored. Inspired by hypothesis and verification method and identification of irrelevant information in human thinking process, we propose a new framework for LLMs abductive logical reasoning called CauseJudger (CJ), which identifies the authenticity of possible cause by transforming thinking from reverse to forward and removing irrelevant information. In addition, we construct an abductive logical reasoning dataset for decision task called CauseLogics, which contains 200,000 tasks of varying reasoning lengths. Our experiments show the efficiency of CJ with overall experiments and ablation experiments as well as case studies on our dataset and reconstructed public dataset. Notably, CJ's implementation is efficient, requiring only two calls to LLM. Its impact is profound: when using gpt-3.5, CJ achieves a maximum correctness improvement of 41% compared to Zero-Shot-CoT. Moreover, with gpt-4, CJ attains an accuracy exceeding 90% across all datasets.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Abductive explanations of classifiers under constraints: Complexity and properties Explaining Non-monotonic Normative Reasoning using Argumentation Theory with Deontic Logic Towards Explainable Goal Recognition Using Weight of Evidence (WoE): A Human-Centered Approach A Metric Hybrid Planning Approach to Solving Pandemic Planning Problems with Simple SIR Models Neural Networks for Vehicle Routing Problem
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1