Abdolmahdi Bagheri, Matin Alinejad, Kevin Bello, Alireza Akhondi-Asl
{"title":"C2P:具有因果推理功能的大型语言模型","authors":"Abdolmahdi Bagheri, Matin Alinejad, Kevin Bello, Alireza Akhondi-Asl","doi":"arxiv-2407.18069","DOIUrl":null,"url":null,"abstract":"Causal reasoning is the primary bottleneck that Large Language Models (LLMs)\nmust overcome to attain human-level intelligence. To address this, we introduce\nthe Causal Chain of Prompting (C2P) as the first reasoning framework that\nequips current LLMs with causal reasoning capabilities. C2P operates\nautonomously, avoiding reliance on external tools or modules during both the\ncausal learning and reasoning phases, and can be seamlessly implemented during\nthe training or fine-tuning of LLMs. Experimental results across various\nbenchmark datasets demonstrate a significant improvement in causal learning and\nsubsequent reasoning accuracy of LLMs. We illustrate how C2P enhances LLMs'\nability to causally reason in real-world scenarios, addressing complex problems\nin fields such as healthcare, medicine, economics, education, social sciences,\nenvironmental science, and marketing. With few-shot learning, GPT-4 Turbo using\nC2P with as few as six examples achieves significant performance improvements,\nboasting over a 33% increase in reasoning accuracy over the most\nstate-of-the-art LLMs, which perform nearly randomly in similar circumstances.\nThis demonstrates the transformative potential of integrating C2P into LLM\ntraining or fine-tuning processes, thereby empowering these models with\nadvanced causal reasoning capabilities.","PeriodicalId":501208,"journal":{"name":"arXiv - CS - Logic in Computer Science","volume":"94 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-07-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"C2P: Featuring Large Language Models with Causal Reasoning\",\"authors\":\"Abdolmahdi Bagheri, Matin Alinejad, Kevin Bello, Alireza Akhondi-Asl\",\"doi\":\"arxiv-2407.18069\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Causal reasoning is the primary bottleneck that Large Language Models (LLMs)\\nmust overcome to attain human-level intelligence. To address this, we introduce\\nthe Causal Chain of Prompting (C2P) as the first reasoning framework that\\nequips current LLMs with causal reasoning capabilities. C2P operates\\nautonomously, avoiding reliance on external tools or modules during both the\\ncausal learning and reasoning phases, and can be seamlessly implemented during\\nthe training or fine-tuning of LLMs. Experimental results across various\\nbenchmark datasets demonstrate a significant improvement in causal learning and\\nsubsequent reasoning accuracy of LLMs. We illustrate how C2P enhances LLMs'\\nability to causally reason in real-world scenarios, addressing complex problems\\nin fields such as healthcare, medicine, economics, education, social sciences,\\nenvironmental science, and marketing. With few-shot learning, GPT-4 Turbo using\\nC2P with as few as six examples achieves significant performance improvements,\\nboasting over a 33% increase in reasoning accuracy over the most\\nstate-of-the-art LLMs, which perform nearly randomly in similar circumstances.\\nThis demonstrates the transformative potential of integrating C2P into LLM\\ntraining or fine-tuning processes, thereby empowering these models with\\nadvanced causal reasoning capabilities.\",\"PeriodicalId\":501208,\"journal\":{\"name\":\"arXiv - CS - Logic in Computer Science\",\"volume\":\"94 1\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-07-25\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - CS - Logic in Computer Science\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2407.18069\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Logic in Computer Science","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2407.18069","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
C2P: Featuring Large Language Models with Causal Reasoning
Causal reasoning is the primary bottleneck that Large Language Models (LLMs)
must overcome to attain human-level intelligence. To address this, we introduce
the Causal Chain of Prompting (C2P) as the first reasoning framework that
equips current LLMs with causal reasoning capabilities. C2P operates
autonomously, avoiding reliance on external tools or modules during both the
causal learning and reasoning phases, and can be seamlessly implemented during
the training or fine-tuning of LLMs. Experimental results across various
benchmark datasets demonstrate a significant improvement in causal learning and
subsequent reasoning accuracy of LLMs. We illustrate how C2P enhances LLMs'
ability to causally reason in real-world scenarios, addressing complex problems
in fields such as healthcare, medicine, economics, education, social sciences,
environmental science, and marketing. With few-shot learning, GPT-4 Turbo using
C2P with as few as six examples achieves significant performance improvements,
boasting over a 33% increase in reasoning accuracy over the most
state-of-the-art LLMs, which perform nearly randomly in similar circumstances.
This demonstrates the transformative potential of integrating C2P into LLM
training or fine-tuning processes, thereby empowering these models with
advanced causal reasoning capabilities.