Siyu Wu, Alessandro Oltramari, Jonathan Francis, C. Lee Giles, Frank E. Ritter
{"title":"认知 LLMs:为制造决策整合认知架构和大型语言模型","authors":"Siyu Wu, Alessandro Oltramari, Jonathan Francis, C. Lee Giles, Frank E. Ritter","doi":"arxiv-2408.09176","DOIUrl":null,"url":null,"abstract":"Resolving the dichotomy between the human-like yet constrained reasoning\nprocesses of Cognitive Architectures and the broad but often noisy inference\nbehavior of Large Language Models (LLMs) remains a challenging but exciting\npursuit, for enabling reliable machine reasoning capabilities in production\nsystems. Because Cognitive Architectures are famously developed for the purpose\nof modeling the internal mechanisms of human cognitive decision-making at a\ncomputational level, new investigations consider the goal of informing LLMs\nwith the knowledge necessary for replicating such processes, e.g., guided\nperception, memory, goal-setting, and action. Previous approaches that use LLMs\nfor grounded decision-making struggle with complex reasoning tasks that require\nslower, deliberate cognition over fast and intuitive inference -- reporting\nissues related to the lack of sufficient grounding, as in hallucination. To\nresolve these challenges, we introduce LLM-ACTR, a novel neuro-symbolic\narchitecture that provides human-aligned and versatile decision-making by\nintegrating the ACT-R Cognitive Architecture with LLMs. Our framework extracts\nand embeds knowledge of ACT-R's internal decision-making process as latent\nneural representations, injects this information into trainable LLM adapter\nlayers, and fine-tunes the LLMs for downstream prediction. Our experiments on\nnovel Design for Manufacturing tasks show both improved task performance as\nwell as improved grounded decision-making capability of our approach, compared\nto LLM-only baselines that leverage chain-of-thought reasoning strategies.","PeriodicalId":501033,"journal":{"name":"arXiv - CS - Symbolic Computation","volume":"322 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-08-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Cognitive LLMs: Towards Integrating Cognitive Architectures and Large Language Models for Manufacturing Decision-making\",\"authors\":\"Siyu Wu, Alessandro Oltramari, Jonathan Francis, C. Lee Giles, Frank E. Ritter\",\"doi\":\"arxiv-2408.09176\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Resolving the dichotomy between the human-like yet constrained reasoning\\nprocesses of Cognitive Architectures and the broad but often noisy inference\\nbehavior of Large Language Models (LLMs) remains a challenging but exciting\\npursuit, for enabling reliable machine reasoning capabilities in production\\nsystems. Because Cognitive Architectures are famously developed for the purpose\\nof modeling the internal mechanisms of human cognitive decision-making at a\\ncomputational level, new investigations consider the goal of informing LLMs\\nwith the knowledge necessary for replicating such processes, e.g., guided\\nperception, memory, goal-setting, and action. Previous approaches that use LLMs\\nfor grounded decision-making struggle with complex reasoning tasks that require\\nslower, deliberate cognition over fast and intuitive inference -- reporting\\nissues related to the lack of sufficient grounding, as in hallucination. To\\nresolve these challenges, we introduce LLM-ACTR, a novel neuro-symbolic\\narchitecture that provides human-aligned and versatile decision-making by\\nintegrating the ACT-R Cognitive Architecture with LLMs. Our framework extracts\\nand embeds knowledge of ACT-R's internal decision-making process as latent\\nneural representations, injects this information into trainable LLM adapter\\nlayers, and fine-tunes the LLMs for downstream prediction. Our experiments on\\nnovel Design for Manufacturing tasks show both improved task performance as\\nwell as improved grounded decision-making capability of our approach, compared\\nto LLM-only baselines that leverage chain-of-thought reasoning strategies.\",\"PeriodicalId\":501033,\"journal\":{\"name\":\"arXiv - CS - Symbolic Computation\",\"volume\":\"322 1\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-08-17\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - CS - Symbolic Computation\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2408.09176\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Symbolic Computation","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2408.09176","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Cognitive LLMs: Towards Integrating Cognitive Architectures and Large Language Models for Manufacturing Decision-making
Resolving the dichotomy between the human-like yet constrained reasoning
processes of Cognitive Architectures and the broad but often noisy inference
behavior of Large Language Models (LLMs) remains a challenging but exciting
pursuit, for enabling reliable machine reasoning capabilities in production
systems. Because Cognitive Architectures are famously developed for the purpose
of modeling the internal mechanisms of human cognitive decision-making at a
computational level, new investigations consider the goal of informing LLMs
with the knowledge necessary for replicating such processes, e.g., guided
perception, memory, goal-setting, and action. Previous approaches that use LLMs
for grounded decision-making struggle with complex reasoning tasks that require
slower, deliberate cognition over fast and intuitive inference -- reporting
issues related to the lack of sufficient grounding, as in hallucination. To
resolve these challenges, we introduce LLM-ACTR, a novel neuro-symbolic
architecture that provides human-aligned and versatile decision-making by
integrating the ACT-R Cognitive Architecture with LLMs. Our framework extracts
and embeds knowledge of ACT-R's internal decision-making process as latent
neural representations, injects this information into trainable LLM adapter
layers, and fine-tunes the LLMs for downstream prediction. Our experiments on
novel Design for Manufacturing tasks show both improved task performance as
well as improved grounded decision-making capability of our approach, compared
to LLM-only baselines that leverage chain-of-thought reasoning strategies.