首页 > 最新文献

arXiv - CS - Symbolic Computation最新文献

英文 中文
Synthesizing Evolving Symbolic Representations for Autonomous Systems 为自主系统合成不断演化的符号表征
Pub Date : 2024-09-18 DOI: arxiv-2409.11756
Gabriele Sartor, Angelo Oddi, Riccardo Rasconi, Vieri Giuliano Santucci, Rosa Meo
Recently, AI systems have made remarkable progress in various tasks. DeepReinforcement Learning(DRL) is an effective tool for agents to learn policiesin low-level state spaces to solve highly complex tasks. Researchers haveintroduced Intrinsic Motivation(IM) to the RL mechanism, which simulates theagent's curiosity, encouraging agents to explore interesting areas of theenvironment. This new feature has proved vital in enabling agents to learnpolicies without being given specific goals. However, even though DRLintelligence emerges through a sub-symbolic model, there is still a need for asort of abstraction to understand the knowledge collected by the agent. To thisend, the classical planning formalism has been used in recent research toexplicitly represent the knowledge an autonomous agent acquires and effectivelyreach extrinsic goals. Despite classical planning usually presents limitedexpressive capabilities, PPDDL demonstrated usefulness in reviewing theknowledge gathered by an autonomous system, making explicit causalcorrelations, and can be exploited to find a plan to reach any state the agentfaces during its experience. This work presents a new architecture implementingan open-ended learning system able to synthesize from scratch its experienceinto a PPDDL representation and update it over time. Without a predefined setof goals and tasks, the system integrates intrinsic motivations to explore theenvironment in a self-directed way, exploiting the high-level knowledgeacquired during its experience. The system explores the environment anditeratively: (a) discover options, (b) explore the environment using options,(c) abstract the knowledge collected and (d) plan. This paper proposes analternative approach to implementing open-ended learning architecturesexploiting low-level and high-level representations to extend its knowledge ina virtuous loop.
最近,人工智能系统在各种任务中取得了显著进展。深度强化学习(DRL)是代理在低级状态空间中学习策略以解决高度复杂任务的有效工具。研究人员在强化学习机制中引入了内在动机(IM),它可以模拟代理的好奇心,鼓励代理探索环境中有趣的领域。事实证明,这一新功能非常重要,它能让代理在没有特定目标的情况下学习策略。不过,尽管 DRL 智能是通过子符号模型产生的,但仍需要进行一定的抽象才能理解代理收集的知识。为此,最近的研究中使用了经典规划形式来明确表示自主代理获取的知识,并有效地实现外在目标。尽管经典规划的表达能力通常有限,但 PPDDL 在回顾自主系统收集的知识、明确因果关系方面表现出了实用性,并可用于寻找计划,以达到代理在其经历过程中所面临的任何状态。这项工作提出了一种新的架构,实现了一种开放式学习系统,该系统能够从头开始将其经验合成为 PPDDL 表征,并随着时间的推移不断更新。在没有预定目标和任务的情况下,该系统利用在体验过程中获得的高级知识,整合内在动机,以自我导向的方式探索环境。该系统探索环境的过程包括:(a)发现选项;(b)利用选项探索环境;(c)对收集到的知识进行抽象;(d)制定计划。本文提出了实现开放式学习架构的替代方法,即利用低级和高级表征来扩展其知识的良性循环。
{"title":"Synthesizing Evolving Symbolic Representations for Autonomous Systems","authors":"Gabriele Sartor, Angelo Oddi, Riccardo Rasconi, Vieri Giuliano Santucci, Rosa Meo","doi":"arxiv-2409.11756","DOIUrl":"https://doi.org/arxiv-2409.11756","url":null,"abstract":"Recently, AI systems have made remarkable progress in various tasks. Deep\u0000Reinforcement Learning(DRL) is an effective tool for agents to learn policies\u0000in low-level state spaces to solve highly complex tasks. Researchers have\u0000introduced Intrinsic Motivation(IM) to the RL mechanism, which simulates the\u0000agent's curiosity, encouraging agents to explore interesting areas of the\u0000environment. This new feature has proved vital in enabling agents to learn\u0000policies without being given specific goals. However, even though DRL\u0000intelligence emerges through a sub-symbolic model, there is still a need for a\u0000sort of abstraction to understand the knowledge collected by the agent. To this\u0000end, the classical planning formalism has been used in recent research to\u0000explicitly represent the knowledge an autonomous agent acquires and effectively\u0000reach extrinsic goals. Despite classical planning usually presents limited\u0000expressive capabilities, PPDDL demonstrated usefulness in reviewing the\u0000knowledge gathered by an autonomous system, making explicit causal\u0000correlations, and can be exploited to find a plan to reach any state the agent\u0000faces during its experience. This work presents a new architecture implementing\u0000an open-ended learning system able to synthesize from scratch its experience\u0000into a PPDDL representation and update it over time. Without a predefined set\u0000of goals and tasks, the system integrates intrinsic motivations to explore the\u0000environment in a self-directed way, exploiting the high-level knowledge\u0000acquired during its experience. The system explores the environment and\u0000iteratively: (a) discover options, (b) explore the environment using options,\u0000(c) abstract the knowledge collected and (d) plan. This paper proposes an\u0000alternative approach to implementing open-ended learning architectures\u0000exploiting low-level and high-level representations to extend its knowledge in\u0000a virtuous loop.","PeriodicalId":501033,"journal":{"name":"arXiv - CS - Symbolic Computation","volume":"18 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142253478","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Introducing Quantification into a Hierarchical Graph Rewriting Language 将量化引入层次图重写语言
Pub Date : 2024-09-17 DOI: arxiv-2409.11015
Haruto Mishina, Kazunori Ueda
LMNtal is a programming and modeling language based on hierarchical graphrewriting that uses logical variables to represent connectivity and membranesto represent hierarchy. On the theoretical side, it allows logicalinterpretation based on intuitionistic linear logic; on the practical side, itsfull-fledged implementation supports a graph-based parallel model checker andhas been used to model diverse applications including various computationalmodels. This paper discuss how we extend LMNtal to QLMNtal (LMNtal withQuantification) to further enhance the usefulness of hierarchical graphrewriting for high-level modeling by introducing quantifiers into rewriting aswell as matching. Those quantifiers allows us to express universalquantification, cardinality and non-existence in an integrated manner. Unlikeother attempts to introduce quantifiers into graph rewriting, QLMNtal hasterm-based syntax, whose semantics is smoothly integrated into the small-stepsemantics of the base language LMNtal. The proposed constructs allow combinedand nested use of quantifiers within individual rewrite rules.
LMNtal 是一种基于分层图重写的编程和建模语言,它使用逻辑变量来表示连接性,使用膜来表示层次结构。在理论方面,它允许基于直觉线性逻辑的逻辑解释;在实践方面,它的完整实现支持基于图的并行模型检查器,并已被用于包括各种计算模型在内的多种应用建模。本文讨论了我们如何将 LMNtal 扩展为 QLMNtal(带量词的 LMNtal),通过在重写和匹配中引入量词,进一步增强分层图重写在高层建模中的实用性。这些量词使我们能够以一种综合的方式表达通用量词、卡片性和不存在性。与其他在图重写中引入量词的尝试不同,QLMNtal 具有基于term的语法,其语义可以平滑地集成到基础语言 LMNtal 的小步语义中。所提出的构造允许在单个重写规则中组合和嵌套使用量词。
{"title":"Introducing Quantification into a Hierarchical Graph Rewriting Language","authors":"Haruto Mishina, Kazunori Ueda","doi":"arxiv-2409.11015","DOIUrl":"https://doi.org/arxiv-2409.11015","url":null,"abstract":"LMNtal is a programming and modeling language based on hierarchical graph\u0000rewriting that uses logical variables to represent connectivity and membranes\u0000to represent hierarchy. On the theoretical side, it allows logical\u0000interpretation based on intuitionistic linear logic; on the practical side, its\u0000full-fledged implementation supports a graph-based parallel model checker and\u0000has been used to model diverse applications including various computational\u0000models. This paper discuss how we extend LMNtal to QLMNtal (LMNtal with\u0000Quantification) to further enhance the usefulness of hierarchical graph\u0000rewriting for high-level modeling by introducing quantifiers into rewriting as\u0000well as matching. Those quantifiers allows us to express universal\u0000quantification, cardinality and non-existence in an integrated manner. Unlike\u0000other attempts to introduce quantifiers into graph rewriting, QLMNtal has\u0000term-based syntax, whose semantics is smoothly integrated into the small-step\u0000semantics of the base language LMNtal. The proposed constructs allow combined\u0000and nested use of quantifiers within individual rewrite rules.","PeriodicalId":501033,"journal":{"name":"arXiv - CS - Symbolic Computation","volume":"65 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142253476","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Symbolic Regression with a Learned Concept Library 利用学习概念库进行符号回归
Pub Date : 2024-09-14 DOI: arxiv-2409.09359
Arya Grayeli, Atharva Sehgal, Omar Costilla-Reyes, Miles Cranmer, Swarat Chaudhuri
We present a novel method for symbolic regression (SR), the task of searchingfor compact programmatic hypotheses that best explain a dataset. The problem iscommonly solved using genetic algorithms; we show that we can enhance suchmethods by inducing a library of abstract textual concepts. Our algorithm,called LaSR, uses zero-shot queries to a large language model (LLM) to discoverand evolve concepts occurring in known high-performing hypotheses. We discovernew hypotheses using a mix of standard evolutionary steps and LLM-guided steps(obtained through zero-shot LLM queries) conditioned on discovered concepts.Once discovered, hypotheses are used in a new round of concept abstraction andevolution. We validate LaSR on the Feynman equations, a popular SR benchmark,as well as a set of synthetic tasks. On these benchmarks, LaSR substantiallyoutperforms a variety of state-of-the-art SR approaches based on deep learningand evolutionary algorithms. Moreover, we show that LaSR can be used todiscover a novel and powerful scaling law for LLMs.
我们为符号回归(SR)提出了一种新方法,符号回归的任务是搜索最能解释数据集的简洁程序假设。这个问题通常使用遗传算法来解决;我们的研究表明,我们可以通过诱导抽象文本概念库来增强这种方法。我们的算法称为 LaSR,它使用对大型语言模型 (LLM) 的零点查询来发现和演化出现在已知高效假设中的概念。我们使用标准进化步骤和 LLM 引导步骤(通过零次 LLM 查询获得)的组合,以发现的概念为条件,发现新的假设。我们在费曼方程(一种流行的 SR 基准)和一组合成任务上验证了 LaSR。在这些基准测试中,LaSR 的性能大大优于各种基于深度学习和进化算法的先进 SR 方法。此外,我们还展示了 LaSR 可用于发现 LLMs 的新颖而强大的缩放规律。
{"title":"Symbolic Regression with a Learned Concept Library","authors":"Arya Grayeli, Atharva Sehgal, Omar Costilla-Reyes, Miles Cranmer, Swarat Chaudhuri","doi":"arxiv-2409.09359","DOIUrl":"https://doi.org/arxiv-2409.09359","url":null,"abstract":"We present a novel method for symbolic regression (SR), the task of searching\u0000for compact programmatic hypotheses that best explain a dataset. The problem is\u0000commonly solved using genetic algorithms; we show that we can enhance such\u0000methods by inducing a library of abstract textual concepts. Our algorithm,\u0000called LaSR, uses zero-shot queries to a large language model (LLM) to discover\u0000and evolve concepts occurring in known high-performing hypotheses. We discover\u0000new hypotheses using a mix of standard evolutionary steps and LLM-guided steps\u0000(obtained through zero-shot LLM queries) conditioned on discovered concepts.\u0000Once discovered, hypotheses are used in a new round of concept abstraction and\u0000evolution. We validate LaSR on the Feynman equations, a popular SR benchmark,\u0000as well as a set of synthetic tasks. On these benchmarks, LaSR substantially\u0000outperforms a variety of state-of-the-art SR approaches based on deep learning\u0000and evolutionary algorithms. Moreover, we show that LaSR can be used to\u0000discover a novel and powerful scaling law for LLMs.","PeriodicalId":501033,"journal":{"name":"arXiv - CS - Symbolic Computation","volume":"100 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142253480","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Towards Verified Polynomial Factorisation 实现验证多项式因式分解
Pub Date : 2024-09-14 DOI: arxiv-2409.09533
James H. Davenport
Computer algebra systems are really good at factoring polynomials, i.e.writing f as a product of irreducible factors. It is relatively easy to verifythat we have a factorisation, but verifying that these factors are irreducibleis a much harder problem. This paper reports work-in-progress to do suchverification in Lean.
计算机代数系统非常擅长对多项式进行因式分解,即把 f 写成不可还原因式的乘积。验证我们是否有因式分解相对容易,但验证这些因式是否不可还原却是一个难得多的问题。本文报告了在 Lean 中进行这种验证的工作进展。
{"title":"Towards Verified Polynomial Factorisation","authors":"James H. Davenport","doi":"arxiv-2409.09533","DOIUrl":"https://doi.org/arxiv-2409.09533","url":null,"abstract":"Computer algebra systems are really good at factoring polynomials, i.e.\u0000writing f as a product of irreducible factors. It is relatively easy to verify\u0000that we have a factorisation, but verifying that these factors are irreducible\u0000is a much harder problem. This paper reports work-in-progress to do such\u0000verification in Lean.","PeriodicalId":501033,"journal":{"name":"arXiv - CS - Symbolic Computation","volume":"66 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142253477","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Active Symbolic Discovery of Ordinary Differential Equations via Phase Portrait Sketching 通过相位肖像素描主动符号化发现常微分方程
Pub Date : 2024-09-02 DOI: arxiv-2409.01416
Nan Jiang, Md Nasim, Yexiang Xue
Discovering Ordinary Differential Equations (ODEs) from trajectory data is acrucial task in AI-driven scientific discovery. Recent methods for symbolicdiscovery of ODEs primarily rely on fixed training datasets collected a-priori,often leading to suboptimal performance, as observed in our experiments inFigure 1. Inspired by active learning, we explore methods for queryinginformative trajectory data to evaluate predicted ODEs, where data are obtainedby the specified initial conditions of the trajectory. Chaos theory indicatesthat small changes in the initial conditions of a dynamical system can resultin vastly different trajectories, necessitating the maintenance of a large setof initial conditions of the trajectory. To address this challenge, weintroduce Active Symbolic Discovery of Ordinary Differential Equations viaPhase Portrait Sketching (APPS). Instead of directly selecting individualinitial conditions, APPS first identifies an informative region and samples abatch of initial conditions within that region. Compared to traditional activelearning methods, APPS eliminates the need for maintaining a large amount ofdata. Extensive experiments demonstrate that APPS consistently discovers moreaccurate ODE expressions than baseline methods using passively collecteddatasets.
从轨迹数据中发现常微分方程(ODE)是人工智能驱动的科学发现中的一项重要任务。最近用于符号发现 ODE 的方法主要依赖于事先收集的固定训练数据集,这往往会导致性能不理想,正如我们在图 1 中的实验所观察到的那样。受主动学习的启发,我们探索了查询信息轨迹数据以评估预测的 ODE 的方法,其中数据是通过指定的轨迹初始条件获得的。混沌理论表明,动态系统初始条件的微小变化都可能导致轨迹的巨大差异,因此有必要维护大量的轨迹初始条件集。为了应对这一挑战,我们引入了通过相位肖像草图(APPS)主动发现常微分方程的符号方法(Active Symbolic Discovery of Ordinary Differential Equations viaPhase Portrait Sketching)。APPS 不直接选择单个初始条件,而是首先确定一个信息区域,然后在该区域内对一批初始条件进行采样。与传统的主动学习方法相比,APPS 无需维护大量数据。大量实验证明,与使用被动收集数据集的基线方法相比,APPS 始终能发现更精确的 ODE 表达式。
{"title":"Active Symbolic Discovery of Ordinary Differential Equations via Phase Portrait Sketching","authors":"Nan Jiang, Md Nasim, Yexiang Xue","doi":"arxiv-2409.01416","DOIUrl":"https://doi.org/arxiv-2409.01416","url":null,"abstract":"Discovering Ordinary Differential Equations (ODEs) from trajectory data is a\u0000crucial task in AI-driven scientific discovery. Recent methods for symbolic\u0000discovery of ODEs primarily rely on fixed training datasets collected a-priori,\u0000often leading to suboptimal performance, as observed in our experiments in\u0000Figure 1. Inspired by active learning, we explore methods for querying\u0000informative trajectory data to evaluate predicted ODEs, where data are obtained\u0000by the specified initial conditions of the trajectory. Chaos theory indicates\u0000that small changes in the initial conditions of a dynamical system can result\u0000in vastly different trajectories, necessitating the maintenance of a large set\u0000of initial conditions of the trajectory. To address this challenge, we\u0000introduce Active Symbolic Discovery of Ordinary Differential Equations via\u0000Phase Portrait Sketching (APPS). Instead of directly selecting individual\u0000initial conditions, APPS first identifies an informative region and samples a\u0000batch of initial conditions within that region. Compared to traditional active\u0000learning methods, APPS eliminates the need for maintaining a large amount of\u0000data. Extensive experiments demonstrate that APPS consistently discovers more\u0000accurate ODE expressions than baseline methods using passively collected\u0000datasets.","PeriodicalId":501033,"journal":{"name":"arXiv - CS - Symbolic Computation","volume":"96 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142211406","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Boolean Matrix Logic Programming 布尔矩阵逻辑编程
Pub Date : 2024-08-19 DOI: arxiv-2408.10369
Lun Ai, Stephen H. Muggleton
We describe a datalog query evaluation approach based on efficient andcomposable boolean matrix manipulation modules. We first define an overarchingproblem, Boolean Matrix Logic Programming (BMLP), which uses boolean matricesas an alternative computation to evaluate datalog programs. We develop twonovel BMLP modules for bottom-up inferences on linear dyadic recursive datalogprograms, and show how additional modules can extend this capability to computeboth linear and non-linear recursive datalog programs of arity two. Ourempirical results demonstrate that these modules outperform general-purpose andspecialised systems by factors of 30x and 9x, respectively, when evaluatinglarge programs with millions of facts. This boolean matrix approachsignificantly enhances the efficiency of datalog querying to support logicprogramming techniques.
我们描述了一种基于高效和可组合布尔矩阵操作模块的数据模型查询评估方法。我们首先定义了一个总体问题,即布尔矩阵逻辑编程(BMLP),它使用布尔矩阵作为评估数据模型程序的另一种计算方法。我们开发了两级 BMLP 模块,用于对线性二元递归数据模型程序进行自下而上的推理,并展示了附加模块如何将这一能力扩展到计算线性和非线性递归数据模型程序。实证结果表明,在评估具有数百万事实的大型程序时,这些模块的性能分别是通用系统和专用系统的 30 倍和 9 倍。这种布尔矩阵方法显著提高了数据模型查询的效率,从而支持逻辑编程技术。
{"title":"Boolean Matrix Logic Programming","authors":"Lun Ai, Stephen H. Muggleton","doi":"arxiv-2408.10369","DOIUrl":"https://doi.org/arxiv-2408.10369","url":null,"abstract":"We describe a datalog query evaluation approach based on efficient and\u0000composable boolean matrix manipulation modules. We first define an overarching\u0000problem, Boolean Matrix Logic Programming (BMLP), which uses boolean matrices\u0000as an alternative computation to evaluate datalog programs. We develop two\u0000novel BMLP modules for bottom-up inferences on linear dyadic recursive datalog\u0000programs, and show how additional modules can extend this capability to compute\u0000both linear and non-linear recursive datalog programs of arity two. Our\u0000empirical results demonstrate that these modules outperform general-purpose and\u0000specialised systems by factors of 30x and 9x, respectively, when evaluating\u0000large programs with millions of facts. This boolean matrix approach\u0000significantly enhances the efficiency of datalog querying to support logic\u0000programming techniques.","PeriodicalId":501033,"journal":{"name":"arXiv - CS - Symbolic Computation","volume":"7 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-08-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142211409","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Cognitive LLMs: Towards Integrating Cognitive Architectures and Large Language Models for Manufacturing Decision-making 认知 LLMs:为制造决策整合认知架构和大型语言模型
Pub Date : 2024-08-17 DOI: arxiv-2408.09176
Siyu Wu, Alessandro Oltramari, Jonathan Francis, C. Lee Giles, Frank E. Ritter
Resolving the dichotomy between the human-like yet constrained reasoningprocesses of Cognitive Architectures and the broad but often noisy inferencebehavior of Large Language Models (LLMs) remains a challenging but excitingpursuit, for enabling reliable machine reasoning capabilities in productionsystems. Because Cognitive Architectures are famously developed for the purposeof modeling the internal mechanisms of human cognitive decision-making at acomputational level, new investigations consider the goal of informing LLMswith the knowledge necessary for replicating such processes, e.g., guidedperception, memory, goal-setting, and action. Previous approaches that use LLMsfor grounded decision-making struggle with complex reasoning tasks that requireslower, deliberate cognition over fast and intuitive inference -- reportingissues related to the lack of sufficient grounding, as in hallucination. Toresolve these challenges, we introduce LLM-ACTR, a novel neuro-symbolicarchitecture that provides human-aligned and versatile decision-making byintegrating the ACT-R Cognitive Architecture with LLMs. Our framework extractsand embeds knowledge of ACT-R's internal decision-making process as latentneural representations, injects this information into trainable LLM adapterlayers, and fine-tunes the LLMs for downstream prediction. Our experiments onnovel Design for Manufacturing tasks show both improved task performance aswell as improved grounded decision-making capability of our approach, comparedto LLM-only baselines that leverage chain-of-thought reasoning strategies.
要在产品系统中实现可靠的机器推理能力,解决认知架构的类人但受限的推理过程与大型语言模型(LLM)的广泛但往往嘈杂的推理行为之间的对立仍然是一项具有挑战性但令人兴奋的探索。由于认知架构的开发目的是在计算层面上模拟人类认知决策的内部机制,因此新的研究考虑的目标是为 LLM 提供复制此类过程(如引导感知、记忆、目标设定和行动)所需的知识。以往使用 LLM 进行基础决策的方法在复杂的推理任务中举步维艰,因为这些任务需要较慢的、深思熟虑的认知,而不是快速的、直观的推理--报告的问题与缺乏足够的基础有关,就像在幻觉中一样。为了解决这些难题,我们引入了 LLM-ACTR,这是一种新颖的神经符号架构,通过将 ACT-R 认知架构与 LLM 相结合,提供与人类相一致的多功能决策。我们的框架提取并嵌入 ACT-R 内部决策过程的知识作为潜在神经表征,将这些信息注入可训练的 LLM 适配器,并微调 LLM 以进行下游预测。我们在新颖的 "制造设计 "任务上进行的实验表明,与利用思维链推理策略的纯 LLM 基线相比,我们的方法既提高了任务性能,也提高了基础决策能力。
{"title":"Cognitive LLMs: Towards Integrating Cognitive Architectures and Large Language Models for Manufacturing Decision-making","authors":"Siyu Wu, Alessandro Oltramari, Jonathan Francis, C. Lee Giles, Frank E. Ritter","doi":"arxiv-2408.09176","DOIUrl":"https://doi.org/arxiv-2408.09176","url":null,"abstract":"Resolving the dichotomy between the human-like yet constrained reasoning\u0000processes of Cognitive Architectures and the broad but often noisy inference\u0000behavior of Large Language Models (LLMs) remains a challenging but exciting\u0000pursuit, for enabling reliable machine reasoning capabilities in production\u0000systems. Because Cognitive Architectures are famously developed for the purpose\u0000of modeling the internal mechanisms of human cognitive decision-making at a\u0000computational level, new investigations consider the goal of informing LLMs\u0000with the knowledge necessary for replicating such processes, e.g., guided\u0000perception, memory, goal-setting, and action. Previous approaches that use LLMs\u0000for grounded decision-making struggle with complex reasoning tasks that require\u0000slower, deliberate cognition over fast and intuitive inference -- reporting\u0000issues related to the lack of sufficient grounding, as in hallucination. To\u0000resolve these challenges, we introduce LLM-ACTR, a novel neuro-symbolic\u0000architecture that provides human-aligned and versatile decision-making by\u0000integrating the ACT-R Cognitive Architecture with LLMs. Our framework extracts\u0000and embeds knowledge of ACT-R's internal decision-making process as latent\u0000neural representations, injects this information into trainable LLM adapter\u0000layers, and fine-tunes the LLMs for downstream prediction. Our experiments on\u0000novel Design for Manufacturing tasks show both improved task performance as\u0000well as improved grounded decision-making capability of our approach, compared\u0000to LLM-only baselines that leverage chain-of-thought reasoning strategies.","PeriodicalId":501033,"journal":{"name":"arXiv - CS - Symbolic Computation","volume":"322 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-08-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142227915","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
ONSEP: A Novel Online Neural-Symbolic Framework for Event Prediction Based on Large Language Model ONSEP:基于大型语言模型的事件预测在线神经符号框架
Pub Date : 2024-08-14 DOI: arxiv-2408.07840
Xuanqing Yu, Wangtao Sun, Jingwei Li, Kang Liu, Chengbao Liu, Jie Tan
In the realm of event prediction, temporal knowledge graph forecasting (TKGF)stands as a pivotal technique. Previous approaches face the challenges of notutilizing experience during testing and relying on a single short-term history,which limits adaptation to evolving data. In this paper, we introduce theOnline Neural-Symbolic Event Prediction (ONSEP) framework, which innovates byintegrating dynamic causal rule mining (DCRM) and dual history augmentedgeneration (DHAG). DCRM dynamically constructs causal rules from real-timedata, allowing for swift adaptation to new causal relationships. In parallel,DHAG merges short-term and long-term historical contexts, leveraging abi-branch approach to enrich event prediction. Our framework demonstratesnotable performance enhancements across diverse datasets, with significantHit@k (k=1,3,10) improvements, showcasing its ability to augment large languagemodels (LLMs) for event prediction without necessitating extensive retraining.The ONSEP framework not only advances the field of TKGF but also underscoresthe potential of neural-symbolic approaches in adapting to dynamic dataenvironments.
在事件预测领域,时间知识图谱预测(TKGF)是一项关键技术。以往的方法面临着在测试过程中无法利用经验和依赖单一短期历史的挑战,这限制了对不断变化的数据的适应。本文介绍了在线神经符号事件预测(ONSEP)框架,该框架通过整合动态因果规则挖掘(DCRM)和双历史增强生成(DHAG)进行了创新。DCRM 从实时数据中动态构建因果规则,从而快速适应新的因果关系。与此同时,DHAG 融合了短期和长期历史背景,利用非分支方法来丰富事件预测。我们的框架在不同的数据集上实现了显著的性能提升,Hit@k(k=1,3,10)有了明显改善,展示了它在无需大量重新训练的情况下增强大型语言模型(LLM)进行事件预测的能力。
{"title":"ONSEP: A Novel Online Neural-Symbolic Framework for Event Prediction Based on Large Language Model","authors":"Xuanqing Yu, Wangtao Sun, Jingwei Li, Kang Liu, Chengbao Liu, Jie Tan","doi":"arxiv-2408.07840","DOIUrl":"https://doi.org/arxiv-2408.07840","url":null,"abstract":"In the realm of event prediction, temporal knowledge graph forecasting (TKGF)\u0000stands as a pivotal technique. Previous approaches face the challenges of not\u0000utilizing experience during testing and relying on a single short-term history,\u0000which limits adaptation to evolving data. In this paper, we introduce the\u0000Online Neural-Symbolic Event Prediction (ONSEP) framework, which innovates by\u0000integrating dynamic causal rule mining (DCRM) and dual history augmented\u0000generation (DHAG). DCRM dynamically constructs causal rules from real-time\u0000data, allowing for swift adaptation to new causal relationships. In parallel,\u0000DHAG merges short-term and long-term historical contexts, leveraging a\u0000bi-branch approach to enrich event prediction. Our framework demonstrates\u0000notable performance enhancements across diverse datasets, with significant\u0000Hit@k (k=1,3,10) improvements, showcasing its ability to augment large language\u0000models (LLMs) for event prediction without necessitating extensive retraining.\u0000The ONSEP framework not only advances the field of TKGF but also underscores\u0000the potential of neural-symbolic approaches in adapting to dynamic data\u0000environments.","PeriodicalId":501033,"journal":{"name":"arXiv - CS - Symbolic Computation","volume":"29 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-08-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142211410","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Algebraic Representations for Faster Predictions in Convolutional Neural Networks 用代数表示法加快卷积神经网络的预测速度
Pub Date : 2024-08-14 DOI: arxiv-2408.07815
Johnny Joyce, Jan Verschelde
Convolutional neural networks (CNNs) are a popular choice of model for tasksin computer vision. When CNNs are made with many layers, resulting in a deepneural network, skip connections may be added to create an easier gradientoptimization problem while retaining model expressiveness. In this paper, weshow that arbitrarily complex, trained, linear CNNs with skip connections canbe simplified into a single-layer model, resulting in greatly reducedcomputational requirements during prediction time. We also present a method fortraining nonlinear models with skip connections that are gradually removedthroughout training, giving the benefits of skip connections without requiringcomputational overhead during during prediction time. These results aredemonstrated with practical examples on Residual Networks (ResNet)architecture.
卷积神经网络(CNN)是计算机视觉任务中常用的模型选择。当卷积神经网络有很多层时,就会形成深度神经网络,这时可以添加跳转连接,从而在保持模型表现力的同时,更容易解决梯度优化问题。在本文中,我们展示了可以将任意复杂、训练有素、带有跳转连接的线性 CNN 简化为单层模型,从而大大降低预测时的计算要求。我们还提出了一种训练带有跳接的非线性模型的方法,这些跳接在整个训练过程中被逐渐移除,从而在预测过程中无需计算开销即可获得跳接的优势。这些结果通过残差网络(ResNet)架构上的实际例子进行了演示。
{"title":"Algebraic Representations for Faster Predictions in Convolutional Neural Networks","authors":"Johnny Joyce, Jan Verschelde","doi":"arxiv-2408.07815","DOIUrl":"https://doi.org/arxiv-2408.07815","url":null,"abstract":"Convolutional neural networks (CNNs) are a popular choice of model for tasks\u0000in computer vision. When CNNs are made with many layers, resulting in a deep\u0000neural network, skip connections may be added to create an easier gradient\u0000optimization problem while retaining model expressiveness. In this paper, we\u0000show that arbitrarily complex, trained, linear CNNs with skip connections can\u0000be simplified into a single-layer model, resulting in greatly reduced\u0000computational requirements during prediction time. We also present a method for\u0000training nonlinear models with skip connections that are gradually removed\u0000throughout training, giving the benefits of skip connections without requiring\u0000computational overhead during during prediction time. These results are\u0000demonstrated with practical examples on Residual Networks (ResNet)\u0000architecture.","PeriodicalId":501033,"journal":{"name":"arXiv - CS - Symbolic Computation","volume":"7 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-08-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142211408","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MathPartner: An Artificial Intelligence Cloud Service MathPartner:人工智能云服务
Pub Date : 2024-08-09 DOI: arxiv-2408.04999
Gennadi Malaschonok, Alexandr Seliverstov
In a broad sense, artificial intelligence is a service to find a solution tocomplex intellectual problems. In this sense, the MathPartner service providesartificial intelligence that allows us to formulate questions and receiveanswers to questions formulated in a mathematical language. For mathematiciansand physicists today, such a language is LaTeX. The MathPartner service uses adialect of LaTeX, which is called Mathpar. The service is a cloud-basedcomputer algebra system and provides users with the opportunity to solve manymathematical problems. In this publication, we focus only on a small class ofextremum problems, which are widely applied in economics, management,logistics, and in many engineering fields. In particular, we consider theshortest path problem and discuss an algorithm that is based on the tropicalmathematics. The ability to work with many types of classical and tropicalalgebras, which are freely available to users, is an important distinguishingfeature of this intelligent tool for symbolic-numerical calculations. We alsoconsider the use of the simplex algorithm for solving optimization problems.
从广义上讲,人工智能是一种为复杂的智力问题寻找解决方案的服务。从这个意义上说,MathPartner 服务提供的人工智能可以让我们提出问题,并获得用数学语言提出的问题的答案。对于今天的数学家和物理学家来说,这种语言就是 LaTeX。MathPartner 服务使用的就是 Mathpar。该服务是一个基于云的计算机代数系统,为用户提供了解决许多数学问题的机会。在本出版物中,我们只关注一小类极端问题,这些问题广泛应用于经济、管理、物流和许多工程领域。特别是,我们考虑了最短路径问题,并讨论了一种基于热带数学的算法。这种用于符号-数值计算的智能工具的一个重要特点是,它能够处理用户免费获得的多种类型的经典和热带卷积。我们还考虑使用单纯形算法来解决优化问题。
{"title":"MathPartner: An Artificial Intelligence Cloud Service","authors":"Gennadi Malaschonok, Alexandr Seliverstov","doi":"arxiv-2408.04999","DOIUrl":"https://doi.org/arxiv-2408.04999","url":null,"abstract":"In a broad sense, artificial intelligence is a service to find a solution to\u0000complex intellectual problems. In this sense, the MathPartner service provides\u0000artificial intelligence that allows us to formulate questions and receive\u0000answers to questions formulated in a mathematical language. For mathematicians\u0000and physicists today, such a language is LaTeX. The MathPartner service uses a\u0000dialect of LaTeX, which is called Mathpar. The service is a cloud-based\u0000computer algebra system and provides users with the opportunity to solve many\u0000mathematical problems. In this publication, we focus only on a small class of\u0000extremum problems, which are widely applied in economics, management,\u0000logistics, and in many engineering fields. In particular, we consider the\u0000shortest path problem and discuss an algorithm that is based on the tropical\u0000mathematics. The ability to work with many types of classical and tropical\u0000algebras, which are freely available to users, is an important distinguishing\u0000feature of this intelligent tool for symbolic-numerical calculations. We also\u0000consider the use of the simplex algorithm for solving optimization problems.","PeriodicalId":501033,"journal":{"name":"arXiv - CS - Symbolic Computation","volume":"32 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-08-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141938851","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
arXiv - CS - Symbolic Computation
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1