首页 > 最新文献

Proceedings of the AAAI Symposium Series最新文献

英文 中文
Analogy as the Swiss Army Knife of Human-like Learning 类比是类人学习的瑞士军刀
Pub Date : 2024-05-20 DOI: 10.1609/aaaiss.v3i1.31272
Kenneth D. Forbus
There is ample psychological evidence that analogy is ubiquitous in human learning, suggesting that computational models of analogy can play important roles in AI systems that learn in human-like ways. This talk will provide evidence for this, focusing mostly on recent advances in hierarchical analogical learning and working-memory analogical generalizations.
大量心理学证据表明,类比在人类学习中无处不在,这表明类比的计算模型可以在以类似人类的方式学习的人工智能系统中发挥重要作用。本讲座将提供这方面的证据,主要侧重于分层类比学习和工作记忆类比泛化方面的最新进展。
{"title":"Analogy as the Swiss Army Knife of Human-like Learning","authors":"Kenneth D. Forbus","doi":"10.1609/aaaiss.v3i1.31272","DOIUrl":"https://doi.org/10.1609/aaaiss.v3i1.31272","url":null,"abstract":"There is ample psychological evidence that analogy is ubiquitous in human learning, suggesting that computational models of analogy can play important roles in AI systems that learn in human-like ways. This talk will provide evidence for this, focusing mostly on recent advances in hierarchical analogical learning and working-memory analogical generalizations.","PeriodicalId":516827,"journal":{"name":"Proceedings of the AAAI Symposium Series","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-05-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141123214","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Perception-Dominant Control Types for Human/Machine Systems 人机系统的感知主导控制类型
Pub Date : 2024-05-20 DOI: 10.1609/aaaiss.v3i1.31177
Ted Goranson
We explore a novel approach to complex domain modelling by emphasising primitives based on perception. The usual approach either focuses on actors or cognition associated with tokens that convey information. In related research, we have examined using effects and/or outcomes as primitives, and influences as the generator of those outcomes via categoric functors. That approach (influences, effects) has advantages: it leverages what is known and supports the expanded logics we use, where we want to anticipate and engineer possible futures. But it has weaknesses when placed in a dynamic human-machine system where what is perceived or assumed matters more than what is known. The work reported here builds on previous advances in type specification and reasoning to ‘move the primitives forward’ more toward situation encounter and away from situation understanding. The goal is in the context of shared human-machine systems where:• reaction times are shorter than the traditional ingestion/comprehension/response loop can support;• situations that are too complex or dynamic for current comprehension by any means;• there simply is insufficient knowledge about governing situations for the comprehension model to support action; and/or,• the many machine/human and system/system interfaces that are incapable of conveying the needed insights; that is, the communication channels choke the information or influence flows. While the approach is motivated by the above unfriendly conditions, we expect significant benefits. We will explore these but engineer toward a federated decision paradigm where decisions by local human, machine or synthesis are not whole-situation-aware, but that collectively ‘swarm’ locally across the larger system to be more effective, ‘wiser’ than a convention paradigm may produce. The supposed implementation strategy will be through extending an existing ‘playbooks as code’ project whose goals are to advise on local action by modelling and gaming complex system dynamics. A sponsoring context is ‘grey zone’ competition that avoids armed conflict, but that can segue to a mixed system course of action advisory. The general context is a costly ‘blue swan’ risk in large commercial and government enterprises. The method will focus on patterns and relationships in synthetic categories used to model type transitions within topological models of system influence. One may say this is applied intuitionistic type theory, following mechanisms generally described by synthetic differential geometry. In this context, the motivating supposition of this study is that information-carrying influence channels are best modelled in our challenging domain as perceived types rather than understood types.
通过强调基于感知的基元,我们探索了一种复杂领域建模的新方法。通常的方法要么侧重于行动者,要么侧重于与传递信息的标记相关的认知。在相关研究中,我们探讨了使用效果和/或结果作为基元,并通过分类函数使用影响作为这些结果的生成器。 这种方法(影响、结果)有其优点:它充分利用了已知的信息,并支持我们使用的扩展逻辑,在这种逻辑中,我们希望预测和设计可能的未来。但在动态人机系统中,感知或假设的东西比已知的东西更重要,因此这种方法也有弱点。本文所报告的工作建立在之前在类型规范和推理方面所取得的进展基础之上,目的是 "向前推进基元",使其更多地与情境相遇,而不是理解情境。 我们的目标是在人机共用系统中:- 反应时间短于传统的摄取/理解/响应循环所能支持的时间;- 情境过于复杂或动态,目前的理解手段无法应对;- 对于理解模型而言,有关管理情境的知识根本不足以支持行动;和/或- 许多机器/人和系统/系统界面无法传递所需的洞察力;也就是说,通信渠道阻塞了信息流或影响流。虽然这种方法是基于上述不友好的条件,但我们期望它能带来显著的益处。我们将探索这些益处,但会朝着联合决策范式的方向努力。在这种范式中,本地人、机器或合成物的决策并不具有整体情境意识,但它们会在更大的系统中集体 "蜂拥",从而比传统范式更有效、更 "明智"。所谓的实施策略是通过扩展现有的 "代码游戏 "项目,其目标是通过对复杂的系统动态进行建模和博弈,为本地行动提供建议。项目的发起背景是避免武装冲突的 "灰色地带 "竞争,但也可以过渡到混合系统行动方案咨询。一般背景是大型商业和政府企业中代价高昂的 "蓝天鹅 "风险。该方法将侧重于合成类别中的模式和关系,用于在系统影响的拓扑模型中模拟类型转换。可以说,这是应用直观类型理论,遵循合成微分几何学所描述的一般机制。在这种情况下,本研究的动机假设是,在我们所面临的挑战领域中,最好将携带信息的影响渠道建模为感知类型,而不是理解类型。
{"title":"Perception-Dominant Control Types for Human/Machine Systems","authors":"Ted Goranson","doi":"10.1609/aaaiss.v3i1.31177","DOIUrl":"https://doi.org/10.1609/aaaiss.v3i1.31177","url":null,"abstract":"We explore a novel approach to complex domain modelling by emphasising primitives based on perception. The usual approach either focuses on actors or cognition associated with tokens that convey information. In related research, we have examined using effects and/or outcomes as primitives, and influences as the generator of those outcomes via categoric functors. \u0000 That approach (influences, effects) has advantages: it leverages what is known and supports the expanded logics we use, where we want to anticipate and engineer possible futures. But it has weaknesses when placed in a dynamic human-machine system where what is perceived or assumed matters more than what is known. The work reported here builds on previous advances in type specification and reasoning to ‘move the primitives forward’ more toward situation encounter and away from situation understanding. \u0000 The goal is in the context of shared human-machine systems where:\u0000• reaction times are shorter than the traditional ingestion/comprehension/response loop can support;\u0000• situations that are too complex or dynamic for current comprehension by any means;\u0000• there simply is insufficient knowledge about governing situations for the comprehension model to support action; and/or,\u0000• the many machine/human and system/system interfaces that are incapable of conveying the needed insights; that is, the communication channels choke the information or influence flows.\u0000 While the approach is motivated by the above unfriendly conditions, we expect significant benefits. We will explore these but engineer toward a federated decision paradigm where decisions by local human, machine or synthesis are not whole-situation-aware, but that collectively ‘swarm’ locally across the larger system to be more effective, ‘wiser’ than a convention paradigm may produce.\u0000 The supposed implementation strategy will be through extending an existing ‘playbooks as code’ project whose goals are to advise on local action by modelling and gaming complex system dynamics. A sponsoring context is ‘grey zone’ competition that avoids armed conflict, but that can segue to a mixed system course of action advisory. The general context is a costly ‘blue swan’ risk in large commercial and government enterprises.\u0000 The method will focus on patterns and relationships in synthetic categories used to model type transitions within topological models of system influence. One may say this is applied intuitionistic type theory, following mechanisms generally described by synthetic differential geometry. In this context, the motivating supposition of this study is that information-carrying influence channels are best modelled in our challenging domain as perceived types rather than understood types.","PeriodicalId":516827,"journal":{"name":"Proceedings of the AAAI Symposium Series","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-05-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141122434","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Ethical Considerations of Generative AI: A Survey Exploring the Role of Decision Makers in the Loop 生成式人工智能的伦理考量:探索环路中决策者角色的调查
Pub Date : 2024-05-20 DOI: 10.1609/aaaiss.v3i1.31243
Yohn Jairo Parra Bautista, Carlos Theran, Richard A. Aló
We explore the foresighted concerns that Norbert Wiener voiced in 1960 about the potential of machines to learn and create strategies that could not be anticipated, drawing parallels to the fable "The Sorcerer's Apprentice" by Goethe. The progress in artificial intelligence (AI) has brought these worries back to the forefront, as shown by a survey AI Impacts conducted in 2022 with more than 700 machine learning researchers. This survey found a five percentage probability that advanced AI might cause "extremely adverse" outcomes, including the possibility of human extinction. Importantly, the introduction of OpenAI's ChatGPT, powered by GPT-4, has led to a surge in entrepreneurial activities, highlighting the ease of use of large language models (LLMs).AI's potential for adverse outcomes, such as military control and unregulated AI races, is explored alongside concerns about AI's role in governance, healthcare, media portrayal, and surpassing human intelligence. Given their transformative impact on content creation, the prominence of generative AI tools such as ChatGPT is noted. The societal assessment of Artificial Intelligence (AI) has grown increasingly intricate and pressing in tandem with the rapid evolution of this technology. As AI continues to advance at a swift pace, the need to comprehensively evaluate its societal implications has become more complex and urgent, necessitating a thorough examination of its potential impact on various domains such as governance, healthcare, media portrayal, and surpassing human intelligence. This assessment is crucial in addressing ethical concerns related to bias, data misuse, technical limitations, and transparency gaps, and in integrating ethical and legal principles throughout AI algorithm lifecycles to ensure alignment with societal well-being. Furthermore, the urgency of addressing the societal implications of AI is underscored by the need for healthcare workforce upskilling and ethical considerations in the era of AI-assisted medicine, emphasizing the critical importance of integrating societal well-being into the development and deployment of AI technologies. Our study entails an examination of the ethical quandaries and obstacles presented when developing methods to evaluate and predict the broader societal impacts of AI on decision-making processes involving the generating of images, videos, and textual content.
我们探讨了诺伯特-维纳(Norbert Wiener)在1960年对机器学习潜力和创造无法预料的战略所表达的前瞻性担忧,并将其与歌德的寓言故事《巫师的学徒》相提并论。2022 年,AI Impacts 对 700 多名机器学习研究人员进行了一项调查。这项调查发现,先进的人工智能可能造成 "极其不利 "结果的概率为 5%,其中包括人类灭绝的可能性。重要的是,由GPT-4驱动的OpenAI的ChatGPT的推出导致了创业活动的激增,凸显了大型语言模型(LLMs)的易用性。在探讨人工智能可能导致的不利结果(如军事控制和不受监管的人工智能竞赛)的同时,还探讨了人们对人工智能在治理、医疗保健、媒体描绘以及超越人类智能方面的作用的担忧。鉴于其对内容创作的变革性影响,ChatGPT 等生成式人工智能工具的重要性也得到了关注。随着人工智能(AI)技术的快速发展,社会对该技术的评估也变得越来越复杂和紧迫。随着人工智能的持续快速发展,全面评估其社会影响的需求也变得更加复杂和迫切,这就要求对其在治理、医疗保健、媒体形象以及超越人类智能等各个领域的潜在影响进行彻底审查。这种评估对于解决与偏见、数据滥用、技术限制和透明度差距有关的伦理问题,以及在整个人工智能算法生命周期中整合伦理和法律原则以确保与社会福祉保持一致至关重要。此外,在人工智能辅助医疗时代,医疗保健人员需要提高技能和考虑伦理问题,这也凸显了解决人工智能社会影响的紧迫性,强调了将社会福祉纳入人工智能技术开发和部署的极端重要性。我们的研究需要审查在开发评估和预测人工智能对涉及生成图像、视频和文本内容的决策过程的广泛社会影响的方法时所遇到的伦理窘境和障碍。
{"title":"Ethical Considerations of Generative AI: A Survey Exploring the Role of Decision Makers in the Loop","authors":"Yohn Jairo Parra Bautista, Carlos Theran, Richard A. Aló","doi":"10.1609/aaaiss.v3i1.31243","DOIUrl":"https://doi.org/10.1609/aaaiss.v3i1.31243","url":null,"abstract":"We explore the foresighted concerns that Norbert Wiener voiced in 1960 about the potential of machines to learn and create strategies that could not be anticipated, drawing parallels to the fable \"The Sorcerer's Apprentice\" by Goethe. The progress in artificial intelligence (AI) has brought these worries back to the forefront, as shown by a survey AI Impacts conducted in 2022 with more than 700 machine learning researchers. This survey found a five percentage probability that advanced AI might cause \"extremely adverse\" outcomes, including the possibility of human extinction. Importantly, the introduction of OpenAI's ChatGPT, powered by GPT-4, has led to a surge in entrepreneurial activities, highlighting the ease of use of large language models (LLMs).AI's potential for adverse outcomes, such as military control and unregulated AI races, is explored alongside concerns about AI's role in governance, healthcare, media portrayal, and surpassing human intelligence. Given their transformative impact on content creation, the prominence of generative AI tools such as ChatGPT is noted. The societal assessment of Artificial Intelligence (AI) has grown increasingly intricate and pressing in tandem with the rapid evolution of this technology. As AI continues to advance at a swift pace, the need to comprehensively evaluate its societal implications has become more complex and urgent, necessitating a thorough examination of its potential impact on various domains such as governance, healthcare, media portrayal, and surpassing human intelligence. This assessment is crucial in addressing ethical concerns related to bias, data misuse, technical limitations, and transparency gaps, and in integrating ethical and legal principles throughout AI algorithm lifecycles to ensure alignment with societal well-being. Furthermore, the urgency of addressing the societal implications of AI is underscored by the need for healthcare workforce upskilling and ethical considerations in the era of AI-assisted medicine, emphasizing the critical importance of integrating societal well-being into the development and deployment of AI technologies. Our study entails an examination of the ethical quandaries and obstacles presented when developing methods to evaluate and predict the broader societal impacts of AI on decision-making processes involving the generating of images, videos, and textual content.","PeriodicalId":516827,"journal":{"name":"Proceedings of the AAAI Symposium Series","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-05-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141122995","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Faithful Reasoning over Scientific Claims 对科学主张的忠实推理
Pub Date : 2024-05-20 DOI: 10.1609/aaaiss.v3i1.31209
N. Tan, Niket Tandon, David Wadden, Oyvind Tafjord, M. Gahegan, Michael Witbrock
Claim verification in scientific domains requires models that faithfully incorporate relevant knowledge from the ever-growing, vast existing literature. Unfaithful claim verifications can lead to misinformation such as those observed during the COVID-19 pandemic. Fact-checking systems often fail to capture the complex relationship between claims and evidence, especially with ambiguous claims and implicit assumptions. Relying only on current LLMs poses challenges due to hallucinations and information traceability issues. To address these challenges, our approach considers multiple viewpoints onto the scientific literature, enabling the assessment of contradictory arguments and implicit assumptions. Our proposed inference method adds faithful reasoning to large language models by distilling information from diverse, relevant scientific abstracts. This method provides a verdict label that can be weighted by the reputation of the scientific articles and an explanation that can be traced back to sources. Our findings demonstrate that humans not only perceive our explanation to be significantly superior to the off-the-shelf model, but they also evaluate it as faithfully enabling the tracing of evidence back to its original sources.
科学领域的索赔验证要求模型能够忠实地纳入不断增长的大量现有文献中的相关知识。不忠实的声称验证可能会导致错误信息,例如在 COVID-19 大流行期间观察到的错误信息。事实核查系统往往无法捕捉到主张与证据之间的复杂关系,尤其是在主张模棱两可和隐含假设的情况下。由于幻觉和信息可追溯性问题,仅依靠当前的法律知识会带来挑战。为了应对这些挑战,我们的方法考虑了科学文献的多种观点,从而能够评估相互矛盾的论点和隐含假设。我们提出的推理方法通过从不同的相关科学摘要中提炼信息,为大型语言模型添加了忠实推理。该方法提供了一个可根据科学文章的声誉加权的判决标签,以及一个可追溯来源的解释。我们的研究结果表明,人类不仅认为我们的解释明显优于现成的模型,而且还认为我们的解释能够忠实地将证据追溯到其原始来源。
{"title":"Faithful Reasoning over Scientific Claims","authors":"N. Tan, Niket Tandon, David Wadden, Oyvind Tafjord, M. Gahegan, Michael Witbrock","doi":"10.1609/aaaiss.v3i1.31209","DOIUrl":"https://doi.org/10.1609/aaaiss.v3i1.31209","url":null,"abstract":"Claim verification in scientific domains requires models that faithfully incorporate relevant knowledge from the ever-growing, vast existing literature. \u0000Unfaithful claim verifications can lead to misinformation such as those observed during the COVID-19 pandemic. Fact-checking systems often fail to capture the complex relationship between claims and evidence, especially with ambiguous claims and implicit assumptions. Relying only on current LLMs poses challenges due to hallucinations and information traceability issues. To address these challenges, our approach considers multiple viewpoints onto the scientific literature, enabling the assessment of contradictory arguments and implicit assumptions. Our proposed inference method adds faithful reasoning to large language models by distilling information from diverse, relevant scientific abstracts. This method provides a verdict label that can be weighted by the reputation of the scientific articles and an explanation that can be traced back to sources. Our findings demonstrate that humans not only perceive our explanation to be significantly superior to the off-the-shelf model, but they also evaluate it as faithfully enabling the tracing of evidence back to its original sources.","PeriodicalId":516827,"journal":{"name":"Proceedings of the AAAI Symposium Series","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-05-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141120971","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Toward Autonomy: Metacognitive Learning for Enhanced AI Performance 迈向自主:元认知学习提升人工智能性能
Pub Date : 2024-05-20 DOI: 10.1609/aaaiss.v3i1.31270
Brendan Conway-Smith, Robert L. West
Large Language Models (LLMs) lack robust metacognitive learning abilities and depend on human-provided algorithms and prompts for learning and output generation. Metacognition involves processes that monitor and enhance cognition. Learning how to learn - metacognitive learning - is crucial for adapting and optimizing learning strategies over time. Although LLMs possess limited metacognitive abilities, they cannot autonomously refine or optimize these strategies. Humans possess innate mechanisms for metacognitive learning that enable at least two unique abilities: discerning which metacognitive strategies are best and automatizing learning strategies. These processes have been effectively modeled in the ACT-R cognitive architecture, providing insights on a path toward greater learning autonomy in AI. Incorporating human-like metacognitive learning abilities into AI could potentially lead to the development of more autonomous and versatile learning mechanisms, as well as improved problem-solving capabilities and performance across diverse tasks.
大型语言模型(LLM)缺乏强大的元认知学习能力,其学习和输出生成依赖于人类提供的算法和提示。元认知涉及监测和增强认知的过程。学会如何学习--元认知学习--对于随着时间的推移调整和优化学习策略至关重要。虽然低等语言学习者拥有有限的元认知能力,但他们无法自主完善或优化这些策略。人类拥有与生俱来的元认知学习机制,至少可以实现两种独特的能力:辨别哪种元认知策略是最好的,以及将学习策略自动化。ACT-R 认知架构对这些过程进行了有效建模,为人工智能实现更高的学习自主性提供了启示。将类似人类的元认知学习能力融入人工智能,有可能开发出更加自主和多用途的学习机制,并提高解决问题的能力和完成各种任务的性能。
{"title":"Toward Autonomy: Metacognitive Learning for Enhanced AI Performance","authors":"Brendan Conway-Smith, Robert L. West","doi":"10.1609/aaaiss.v3i1.31270","DOIUrl":"https://doi.org/10.1609/aaaiss.v3i1.31270","url":null,"abstract":"Large Language Models (LLMs) lack robust metacognitive learning abilities and depend on human-provided algorithms and prompts for learning and output generation. Metacognition involves processes that monitor and enhance cognition. Learning how to learn - metacognitive learning - is crucial for adapting and optimizing learning strategies over time. Although LLMs possess limited metacognitive abilities, they cannot autonomously refine or optimize these strategies. Humans possess innate mechanisms for metacognitive learning that enable at least two unique abilities: discerning which metacognitive strategies are best and automatizing learning strategies. These processes have been effectively modeled in the ACT-R cognitive architecture, providing insights on a path toward greater learning autonomy in AI. Incorporating human-like metacognitive learning abilities into AI could potentially lead to the development of more autonomous and versatile learning mechanisms, as well as improved problem-solving capabilities and performance across diverse tasks.","PeriodicalId":516827,"journal":{"name":"Proceedings of the AAAI Symposium Series","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-05-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141120698","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Toward Human-Like Representation Learning for Cognitive Architectures 面向认知架构的类人表征学习
Pub Date : 2024-05-20 DOI: 10.1609/aaaiss.v3i1.31274
Steven Jones, Peter Lindes
Human-like learning includes an ability to learn concepts from a stream of embodiment sensor data. Echoing previous thoughts such as those from Barsalou that cognition and perception share a common representation system, we suggest an addendum to the common model of cognition. This addendum poses a simultaneous semantic memory and perception learning that bypasses working memory, and that uses parallel processing to learn concepts apart from deliberate reasoning. The goal is to provide a general outline for how to extend a class of cognitive architectures to implement a more human-like interface between cognition and embodiment of an agent, where a critical aspect of that interface is that it is dynamic because of learning.
类人学习包括从体现传感器数据流中学习概念的能力。与巴萨罗等人之前关于认知和感知共享一个共同表征系统的观点相呼应,我们建议对认知的共同模型进行增补。该附录提出了一种同时学习语义记忆和感知的方法,它绕过了工作记忆,利用并行处理来学习刻意推理之外的概念。我们的目标是为如何扩展一类认知架构提供一个总纲,以便在认知和代理的体现之间实现一个更像人类的界面,而该界面的一个关键方面是,由于学习,它是动态的。
{"title":"Toward Human-Like Representation Learning for Cognitive Architectures","authors":"Steven Jones, Peter Lindes","doi":"10.1609/aaaiss.v3i1.31274","DOIUrl":"https://doi.org/10.1609/aaaiss.v3i1.31274","url":null,"abstract":"Human-like learning includes an ability to learn concepts from a stream of embodiment sensor data. Echoing previous thoughts such as those from Barsalou that cognition and perception share a common representation system, we suggest an addendum to the common model of cognition. This addendum poses a simultaneous semantic memory and perception learning that bypasses working memory, and that uses parallel processing to learn concepts apart from deliberate reasoning. The goal is to provide a general outline for how to extend a class of cognitive architectures to implement a more human-like interface between cognition and embodiment of an agent, where a critical aspect of that interface is that it is dynamic because of learning.","PeriodicalId":516827,"journal":{"name":"Proceedings of the AAAI Symposium Series","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-05-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141122270","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
On Replacing Humans with Large Language Models in Voice-Based Human-in-the-Loop Systems 在基于语音的 "人在回路 "系统中用大型语言模型取代人类
Pub Date : 2024-05-20 DOI: 10.1609/aaaiss.v3i1.31178
Shih-Hong Huang, Ting-Hao 'Kenneth' Huang
It is easy to assume that Large Language Models (LLMs) will seamlessly take over applications, especially those that are largely automated. In the case of conversational voice assistants, commercial systems have been widely deployed and used over the past decade. However, are we indeed on the cusp of the future we envisioned? There exists a social-technical gap between what people want to accomplish and the actual capability of technology. In this paper, we present a case study comparing two voice assistants built on Amazon Alexa: one employing a human-in-the-loop workflow, the other utilizes LLM to engage in conversations with users. In our comparison, we discovered that the issues arising in current human-in-the-loop and LLM systems are not identical. However, the presence of a set of similar issues in both systems leads us to believe that focusing on the interaction between users and systems is crucial, perhaps even more so than focusing solely on the underlying technology itself. Merely enhancing the performance of the workers or the models may not adequately address these issues. This observation prompts our research question: What are the overlooked contributing factors in the effort to improve the capabilities of voice assistants, which might not have been emphasized in prior research?
人们很容易假定,大型语言模型(LLM)将无缝接管各种应用,尤其是那些基本自动化的应用。就会话语音助手而言,商业系统在过去十年中得到了广泛部署和使用。然而,我们是否真的站在了我们所设想的未来的风口浪尖上?人们想要实现的目标与技术的实际能力之间存在着社会技术差距。在本文中,我们介绍了一项案例研究,比较了基于亚马逊 Alexa 的两个语音助手:一个采用了人在回路中的工作流程,另一个则利用 LLM 与用户进行对话。在比较过程中,我们发现当前人工智能系统和 LLM 系统中出现的问题并不相同。不过,两个系统中存在的一系列类似问题让我们相信,关注用户与系统之间的互动至关重要,这或许比只关注底层技术本身更为重要。仅仅提高工作人员或模型的性能可能无法充分解决这些问题。这一观察结果引发了我们的研究问题:在努力提高语音助手能力的过程中,有哪些因素被忽视了,而这些因素在以前的研究中可能并没有得到重视?
{"title":"On Replacing Humans with Large Language Models in Voice-Based Human-in-the-Loop Systems","authors":"Shih-Hong Huang, Ting-Hao 'Kenneth' Huang","doi":"10.1609/aaaiss.v3i1.31178","DOIUrl":"https://doi.org/10.1609/aaaiss.v3i1.31178","url":null,"abstract":"It is easy to assume that Large Language Models (LLMs) will seamlessly take over applications, especially those that are largely automated. In the case of conversational voice assistants, commercial systems have been widely deployed and used over the past decade. However, are we indeed on the cusp of the future we envisioned? There exists a social-technical gap between what people want to accomplish and the actual capability of technology. In this paper, we present a case study comparing two voice assistants built on Amazon Alexa: one employing a human-in-the-loop workflow, the other utilizes LLM to engage in conversations with users. In our comparison, we discovered that the issues arising in current human-in-the-loop and LLM systems are not identical. However, the presence of a set of similar issues in both systems leads us to believe that focusing on the interaction between users and systems is crucial, perhaps even more so than focusing solely on the underlying technology itself. Merely enhancing the performance of the workers or the models may not adequately address these issues. This observation prompts our research question: What are the overlooked contributing factors in the effort to improve the capabilities of voice assistants, which might not have been emphasized in prior research?","PeriodicalId":516827,"journal":{"name":"Proceedings of the AAAI Symposium Series","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-05-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141118958","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Human-Like Learning of Social Reasoning via Analogy 通过类比进行类人社会推理学习
Pub Date : 2024-05-20 DOI: 10.1609/aaaiss.v3i1.31284
Irina Rabkina
Neurotypical adult humans are impeccably good social reasoners. Despite the occasional faux pas, we know how to interact in most social settings and how to consider others' points of view. Young children, on the other hand, do not. Social reasoning, like many of our most important skills, is learned. Much like human children, AI agents are not good social reasoners. While some algorithms can perform some aspects of social reasoning, we are a ways off from AI that can interact naturally and appropriately in the broad range of settings that people can. In this talk, I will argue that learning social reasoning via the same processes used by people will help AI agents reason--and interact--more like people do. Specifically, I will argue that children learn social reasoning via analogy, and that AI agents should, too. I will present evidence from cognitive modeling experiments demonstrating the former and AI experiments demonstrating the latter. I will also propose future directions for social reasoning research that both demonstrate the need for robust, human-like social reasoning in AI and test the utility of common approaches.
神经正常的成年人是无可挑剔的社交推理高手。尽管偶尔也会犯错,但我们知道如何在大多数社交场合进行互动,知道如何考虑他人的观点。而幼儿则不然。社交推理和我们许多最重要的技能一样,都是后天学习的。与人类儿童一样,人工智能代理也不擅长社交推理。虽然有些算法可以进行某些方面的社会推理,但我们距离人工智能能够像人类一样在广泛的环境中自然、恰当地进行互动还有一段距离。在本讲座中,我将论证,通过与人类相同的过程学习社会推理,将有助于人工智能代理更像人类那样进行推理和互动。具体来说,我将论证儿童通过类比学习社会推理,人工智能代理也应该如此。我将从认知建模实验和人工智能实验中提出证据,证明前者和后者。我还将提出社会推理研究的未来方向,既证明人工智能需要强大的、类似人类的社会推理,又测试常用方法的实用性。
{"title":"Human-Like Learning of Social Reasoning via Analogy","authors":"Irina Rabkina","doi":"10.1609/aaaiss.v3i1.31284","DOIUrl":"https://doi.org/10.1609/aaaiss.v3i1.31284","url":null,"abstract":"Neurotypical adult humans are impeccably good social reasoners. Despite the occasional faux pas, we know how to interact in most social settings and how to consider others' points of view. Young children, on the other hand, do not. Social reasoning, like many of our most important skills, is learned. \u0000\u0000Much like human children, AI agents are not good social reasoners. While some algorithms can perform some aspects of social reasoning, we are a ways off from AI that can interact naturally and appropriately in the broad range of settings that people can. In this talk, I will argue that learning social reasoning via the same processes used by people will help AI agents reason--and interact--more like people do. Specifically, I will argue that children learn social reasoning via analogy, and that AI agents should, too. I will present evidence from cognitive modeling experiments demonstrating the former and AI experiments demonstrating the latter. I will also propose future directions for social reasoning research that both demonstrate the need for robust, human-like social reasoning in AI and test the utility of common approaches.","PeriodicalId":516827,"journal":{"name":"Proceedings of the AAAI Symposium Series","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-05-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141119950","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Inclusion Ethics in AI: Use Cases in African Fashion 人工智能中的包容伦理:非洲时尚界的使用案例
Pub Date : 2024-05-20 DOI: 10.1609/aaaiss.v3i1.31266
Christelle Scharff, James Brusseau, K. Bathula, Kaleemunnisa Fnu, Samyak Rakesh Meshram, Om Gaikhe
This paper addresses the ethics of inclusion in artificial in-telligence in the context of African fashion. Despite the proliferation of fashion-related AI applications and da-tasets global diversity remains limited, and African fash-ion is significantly underrepresented. This paper docu-ments two use-cases that enhance AI's inclusivity by in-corporating sub-Saharan fashion elements. The first case details the creation of a Senegalese fashion dataset and a model for classifying traditional apparel using transfer learning. The second case investigates African wax textile patterns generated through generative adversarial net-works (GANs), specifically StyleGAN architectures, and machine learning diffusion models. Alongside the practi-cal, technological advances, theoretical ethical progress is made in two directions. First, the cases are used to elabo-rate and define the ethics of inclusion, while also contrib-uting to current debates about how inclusion differs from ethical fairness. Second, the cases engage with the ethical debate on whether AI innovation should be slowed to prevent ethical imbalances or accelerated to solve them.
本文探讨了非洲时尚背景下人工智能的包容性伦理问题。尽管与时尚相关的人工智能应用和工具集不断涌现,但全球多样性仍然有限,非洲时尚的代表性严重不足。本文记录了两个通过融入撒哈拉以南地区的时尚元素来增强人工智能包容性的案例。第一个案例详细介绍了塞内加尔时尚数据集的创建,以及利用迁移学习对传统服装进行分类的模型。第二个案例研究了通过生成式对抗网络工程(GAN)(特别是 StyleGAN 架构)和机器学习扩散模型生成的非洲蜡纺织品图案。在实践和技术进步的同时,理论伦理也在两个方向上取得了进展。首先,这些案例被用来确定和定义全纳伦理,同时也有助于当前关于全纳与伦理公平有何不同的辩论。其次,案例参与了关于人工智能创新应该放缓以防止伦理失衡,还是加快以解决伦理失衡的伦理辩论。
{"title":"Inclusion Ethics in AI: Use Cases in African Fashion","authors":"Christelle Scharff, James Brusseau, K. Bathula, Kaleemunnisa Fnu, Samyak Rakesh Meshram, Om Gaikhe","doi":"10.1609/aaaiss.v3i1.31266","DOIUrl":"https://doi.org/10.1609/aaaiss.v3i1.31266","url":null,"abstract":"This paper addresses the ethics of inclusion in artificial in-telligence in the context of African fashion. Despite the proliferation of fashion-related AI applications and da-tasets global diversity remains limited, and African fash-ion is significantly underrepresented. This paper docu-ments two use-cases that enhance AI's inclusivity by in-corporating sub-Saharan fashion elements. The first case details the creation of a Senegalese fashion dataset and a model for classifying traditional apparel using transfer learning. The second case investigates African wax textile patterns generated through generative adversarial net-works (GANs), specifically StyleGAN architectures, and machine learning diffusion models. Alongside the practi-cal, technological advances, theoretical ethical progress is made in two directions. First, the cases are used to elabo-rate and define the ethics of inclusion, while also contrib-uting to current debates about how inclusion differs from ethical fairness. Second, the cases engage with the ethical debate on whether AI innovation should be slowed to prevent ethical imbalances or accelerated to solve them.","PeriodicalId":516827,"journal":{"name":"Proceedings of the AAAI Symposium Series","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-05-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141120817","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multi-Criterion Client Selection for Efficient Federated Learning 高效联盟学习的多标准客户端选择
Pub Date : 2024-05-20 DOI: 10.1609/aaaiss.v3i1.31227
Mehreen Tahir, Muhammad Intizar Ali
Federated Learning (FL) has received tremendous attention as a decentralized machine learning (ML) framework that allows distributed data owners to collaboratively train a global model without sharing raw data. Since FL trains the model directly on edge devices, the heterogeneity of participating clients in terms of data distribution, hardware capabilities and network connectivity can significantly impact the overall performance of FL systems. Optimizing for model accuracy could extend the training time due to the diverse and resource-constrained nature of edge devices while minimizing training time could compromise the model's accuracy. Effective client selection thus becomes crucial to ensure that the training process is not only efficient but also capitalizes on the diverse data and computational capabilities of different devices. To this end, we propose FedPROM, a novel framework that tackles client selection in FL as a multi-criteria optimization problem. By leveraging the PROMETHEE method, FedPROM ranks clients based on their suitability for a given FL task, considering multiple criteria such as system resources, network conditions, and data quality. This approach allows FedPROM to dynamically select the most appropriate set of clients for each learning round, optimizing both model accuracy and training efficiency. Our evaluations on diverse datasets demonstrate that FedPROM outperforms several state-of-the-art FL client selection protocols in terms of convergence speed, and accuracy, highlighting the framework's effectiveness and the importance of multi-criteria client selection in FL.
联邦学习(FL)作为一种去中心化的机器学习(ML)框架受到了极大的关注,它允许分布式数据所有者在不共享原始数据的情况下协作训练一个全局模型。由于联邦学习直接在边缘设备上训练模型,参与的客户端在数据分布、硬件能力和网络连接方面的异质性会极大地影响联邦学习系统的整体性能。由于边缘设备的多样性和资源有限性,优化模型准确性可能会延长训练时间,而尽量缩短训练时间则可能会影响模型的准确性。因此,有效的客户端选择对于确保训练过程不仅高效,而且充分利用不同设备的数据和计算能力至关重要。为此,我们提出了 FedPROM,这是一个新颖的框架,将 FL 中的客户端选择作为一个多标准优化问题来处理。通过利用 PROMETHEE 方法,FedPROM 在考虑系统资源、网络条件和数据质量等多重标准的基础上,根据客户端对特定 FL 任务的适用性对其进行排序。通过这种方法,FedPROM 可以为每一轮学习动态选择最合适的客户端集,从而优化模型准确性和训练效率。我们在不同数据集上进行的评估表明,FedPROM 在收敛速度和准确性方面都优于几种最先进的 FL 客户端选择协议,这凸显了该框架的有效性以及多标准客户端选择在 FL 中的重要性。
{"title":"Multi-Criterion Client Selection for Efficient Federated Learning","authors":"Mehreen Tahir, Muhammad Intizar Ali","doi":"10.1609/aaaiss.v3i1.31227","DOIUrl":"https://doi.org/10.1609/aaaiss.v3i1.31227","url":null,"abstract":"Federated Learning (FL) has received tremendous attention as a decentralized machine learning (ML) framework that allows distributed data owners to collaboratively train a global model without sharing raw data. Since FL trains the model directly on edge devices, the heterogeneity of participating clients in terms of data distribution, hardware capabilities and network connectivity can significantly impact the overall performance of FL systems. Optimizing for model accuracy could extend the training time due to the diverse and resource-constrained nature of edge devices while minimizing training time could compromise the model's accuracy. Effective client selection thus becomes crucial to ensure that the training process is not only efficient but also capitalizes on the diverse data and computational capabilities of different devices. To this end, we propose FedPROM, a novel framework that tackles client selection in FL as a multi-criteria optimization problem. By leveraging the PROMETHEE method, FedPROM ranks clients based on their suitability for a given FL task, considering multiple criteria such as system resources, network conditions, and data quality. This approach allows FedPROM to dynamically select the most appropriate set of clients for each learning round, optimizing both model accuracy and training efficiency. Our evaluations on diverse datasets demonstrate that FedPROM outperforms several state-of-the-art FL client selection protocols in terms of convergence speed, and accuracy, highlighting the framework's effectiveness and the importance of multi-criteria client selection in FL.","PeriodicalId":516827,"journal":{"name":"Proceedings of the AAAI Symposium Series","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-05-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141119064","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Proceedings of the AAAI Symposium Series
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1