首页 > 最新文献

Proceedings of the AAAI Symposium Series最新文献

英文 中文
Modeling Human-Like Acquisition of Language and Concepts 模拟人类学习语言和概念的过程
Pub Date : 2024-05-20 DOI: 10.1609/aaaiss.v3i1.31275
Peter Lindes, Steven Jones
Humans acquire language and related concepts in a trajectory over a lifetime. Concepts for simple interaction with the world are learned before language. Later, words are learned to name these concepts along with structures needed to represent larger meanings. Eventually, language advances to where it can drive the learning of new concepts. Throughout this trajectory a language processing capability uses architectural mechanisms to process language using the knowledge already acquired. We assume that this growing body of knowledge is made up of small units of form-meaning mapping that can be composed in many ways, suggesting that these units are learned incrementally from experience. In prior work we have built a system to comprehend human language within an autonomous robot using knowledge in such units developed by hand. Here we propose a research program to develop the ability of an artificial agent to acquire this knowledge incrementally and autonomously from its experience in a similar trajectory. We then propose a strategy for evaluating this human-like learning system using a large benchmark created as a tool for training deep learning systems. We expect that our human-like learning system will produce better task performance from training on only a small subset of this benchmark.
人的一生都在循序渐进地学习语言和相关概念。在学习语言之前,人们先学会了与世界进行简单互动的概念。后来,人们学会了命名这些概念的词汇,以及表达更大含义所需的结构。最终,语言可以推动新概念的学习。在这一发展过程中,语言处理能力将利用已掌握的知识架构机制来处理语言。我们假定,这个不断增长的知识体系是由形式-意义映射的小单元组成的,它们可以以多种方式组成,这表明这些单元是从经验中逐步学习的。在之前的工作中,我们利用手工开发的此类单元知识,在自主机器人中建立了一个理解人类语言的系统。在此,我们提出了一项研究计划,以开发人工智能代理从类似轨迹的经验中逐步自主获取这些知识的能力。然后,我们提出一种策略,利用作为深度学习系统训练工具而创建的大型基准来评估这种类人学习系统。我们预计,我们的类人学习系统只需在该基准的一小部分上进行训练,就能产生更好的任务性能。
{"title":"Modeling Human-Like Acquisition of Language and Concepts","authors":"Peter Lindes, Steven Jones","doi":"10.1609/aaaiss.v3i1.31275","DOIUrl":"https://doi.org/10.1609/aaaiss.v3i1.31275","url":null,"abstract":"Humans acquire language and related concepts in a trajectory over a lifetime. Concepts for simple interaction with the world are learned before language. Later, words are learned to name these concepts along with structures needed to represent larger meanings. Eventually, language advances to where it can drive the learning of new concepts. Throughout this trajectory a language processing capability uses architectural mechanisms to process language using the knowledge already acquired. We assume that this growing body of knowledge is made up of small units of form-meaning mapping that can be composed in many ways, suggesting that these units are learned incrementally from experience. In prior work we have built a system to comprehend human language within an autonomous robot using knowledge in such units developed by hand. Here we propose a research program to develop the ability of an artificial agent to acquire this knowledge incrementally and autonomously from its experience in a similar trajectory. We then propose a strategy for evaluating this human-like learning system using a large benchmark created as a tool for training deep learning systems. We expect that our human-like learning system will produce better task performance from training on only a small subset of this benchmark.","PeriodicalId":516827,"journal":{"name":"Proceedings of the AAAI Symposium Series","volume":"81 10","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-05-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141121330","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Analogy as the Swiss Army Knife of Human-like Learning 类比是类人学习的瑞士军刀
Pub Date : 2024-05-20 DOI: 10.1609/aaaiss.v3i1.31272
Kenneth D. Forbus
There is ample psychological evidence that analogy is ubiquitous in human learning, suggesting that computational models of analogy can play important roles in AI systems that learn in human-like ways. This talk will provide evidence for this, focusing mostly on recent advances in hierarchical analogical learning and working-memory analogical generalizations.
大量心理学证据表明,类比在人类学习中无处不在,这表明类比的计算模型可以在以类似人类的方式学习的人工智能系统中发挥重要作用。本讲座将提供这方面的证据,主要侧重于分层类比学习和工作记忆类比泛化方面的最新进展。
{"title":"Analogy as the Swiss Army Knife of Human-like Learning","authors":"Kenneth D. Forbus","doi":"10.1609/aaaiss.v3i1.31272","DOIUrl":"https://doi.org/10.1609/aaaiss.v3i1.31272","url":null,"abstract":"There is ample psychological evidence that analogy is ubiquitous in human learning, suggesting that computational models of analogy can play important roles in AI systems that learn in human-like ways. This talk will provide evidence for this, focusing mostly on recent advances in hierarchical analogical learning and working-memory analogical generalizations.","PeriodicalId":516827,"journal":{"name":"Proceedings of the AAAI Symposium Series","volume":"73 11","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-05-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141123214","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Exploiting Machine Learning Bias: Predicting Medical Denials 利用机器学习的偏差:预测医疗拒绝率
Pub Date : 2024-05-20 DOI: 10.1609/aaaiss.v3i1.31181
Stephen Russell, Fabio Montes Suros, Ashwin Kumar
For a large healthcare system, ignoring costs associated with managing the patient encounter denial process (staffing, contracts, etc.), total denial-related amounts can be more than $1B annually in gross charges. Being able to predict a denial before it occurs has the potential for tremendous savings. Using machine learning to predict denial has the potential to allow denial-preventing interventions. However, challenges of data imbalance make creating a single generalized model difficult. We employ two biased models in a hybrid voting scheme to achieve results that exceed the state-of-the art and allow for incremental predictions as the encounter progresses. The model had the added benefit of monitoring the human-driven denial process that affect the underlying distribution, on which the models’ bias is based.
对于大型医疗保健系统而言,如果不考虑与管理病人就诊拒绝流程相关的成本(人员配备、合同等),每年与拒绝相关的总费用可能超过 10 亿美元。如果能在拒付发生之前预测到拒付,就有可能节省大量费用。利用机器学习预测拒付有可能实现防止拒付的干预措施。然而,数据不平衡的挑战使得创建一个单一的通用模型变得困难。我们在一个混合投票方案中采用了两个有偏差的模型,从而获得了超越最先进技术的结果,并允许随着遭遇的进展进行增量预测。该模型的另一个好处是可以监控影响基本分布的人为驱动的拒绝过程,而模型的偏差正是建立在这一基础之上的。
{"title":"Exploiting Machine Learning Bias: Predicting Medical Denials","authors":"Stephen Russell, Fabio Montes Suros, Ashwin Kumar","doi":"10.1609/aaaiss.v3i1.31181","DOIUrl":"https://doi.org/10.1609/aaaiss.v3i1.31181","url":null,"abstract":"For a large healthcare system, ignoring costs associated with managing the patient encounter denial process (staffing, contracts, etc.), total denial-related amounts can be more than $1B annually in gross charges. Being able to predict a denial before it occurs has the potential for tremendous savings. Using machine learning to predict denial has the potential to allow denial-preventing interventions. However, challenges of data imbalance make creating a single generalized model difficult. We employ two biased models in a hybrid voting scheme to achieve results that exceed the state-of-the art and allow for incremental predictions as the encounter progresses. The model had the added benefit of monitoring the human-driven denial process that affect the underlying distribution, on which the models’ bias is based.","PeriodicalId":516827,"journal":{"name":"Proceedings of the AAAI Symposium Series","volume":"60 2","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-05-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141123298","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Embodying Human-Like Modes of Balance Control Through Human-In-the-Loop Dyadic Learning 通过 "人在回路中 "的双向学习实现类似人类的平衡控制模式
Pub Date : 2024-05-20 DOI: 10.1609/aaaiss.v3i1.31278
Sheikh Mannan, V. Vimal, Paul DiZio, Nikhil Krishnaswamy
In this paper, we explore how humans and AIs trained to perform a virtual inverted pendulum (VIP) balancing task converge and differ in their learning and performance strategies. We create a visual analogue of disoriented IP balancing, as may be experienced by pilots suffering from spatial disorientation, and train AI models on data from human subjects performing a real-world disoriented balancing task. We then place the trained AI models in a dyadic human-in-the-loop (HITL) training setting. Episodes in which human subjects disagreed with AI actions were logged and used to fine-tune the AI model. Human subjects then performed the task while being given guidance from pretrained and dyadically fine-tuned versions of an AI model. We examine the effects of HITL training on AI performance, AI guidance on human performance, and the behavior patterns of human subjects and AI models during task performance. We find that in many cases, HITL training improves AI performance, AI guidance improves human performance, and after dyadic training the two converge on similar behavior patterns.
在本文中,我们探讨了人类和人工智能在执行虚拟倒立摆(VIP)平衡任务时,在学习和执行策略上是如何趋同和差异的。我们创建了一个迷失方向的 IP 平衡视觉模拟(飞行员可能会经历空间迷失),并根据执行真实世界迷失方向平衡任务的人类受试者的数据训练人工智能模型。然后,我们将训练好的人工智能模型置于双人环内(HITL)训练环境中。我们记录了人类受试者与人工智能操作不一致的情况,并利用这些情况对人工智能模型进行微调。然后,人类受试者在人工智能模型的预训练和双向微调版本的指导下执行任务。我们研究了 HITL 训练对人工智能性能的影响、人工智能对人类性能的指导,以及人类受试者和人工智能模型在执行任务过程中的行为模式。我们发现,在许多情况下,HITL 训练提高了人工智能的性能,人工智能指导提高了人类的性能,而且在经过双向训练后,两者的行为模式趋于相似。
{"title":"Embodying Human-Like Modes of Balance Control Through Human-In-the-Loop Dyadic Learning","authors":"Sheikh Mannan, V. Vimal, Paul DiZio, Nikhil Krishnaswamy","doi":"10.1609/aaaiss.v3i1.31278","DOIUrl":"https://doi.org/10.1609/aaaiss.v3i1.31278","url":null,"abstract":"In this paper, we explore how humans and AIs trained to perform a virtual inverted pendulum (VIP) balancing task converge and differ in their learning and performance strategies. We create a visual analogue of disoriented IP balancing, as may be experienced by pilots suffering from spatial disorientation, and train AI models on data from human subjects performing a real-world disoriented balancing task. We then place the trained AI models in a dyadic human-in-the-loop (HITL) training setting. Episodes in which human subjects disagreed with AI actions were logged and used to fine-tune the AI model. Human subjects then performed the task while being given guidance from pretrained and dyadically fine-tuned versions of an AI model. We examine the effects of HITL training on AI performance, AI guidance on human performance, and the behavior patterns of human subjects and AI models during task performance. We find that in many cases, HITL training improves AI performance, AI guidance improves human performance, and after dyadic training the two converge on similar behavior patterns.","PeriodicalId":516827,"journal":{"name":"Proceedings of the AAAI Symposium Series","volume":"48 4","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-05-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141118648","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
On Replacing Humans with Large Language Models in Voice-Based Human-in-the-Loop Systems 在基于语音的 "人在回路 "系统中用大型语言模型取代人类
Pub Date : 2024-05-20 DOI: 10.1609/aaaiss.v3i1.31178
Shih-Hong Huang, Ting-Hao 'Kenneth' Huang
It is easy to assume that Large Language Models (LLMs) will seamlessly take over applications, especially those that are largely automated. In the case of conversational voice assistants, commercial systems have been widely deployed and used over the past decade. However, are we indeed on the cusp of the future we envisioned? There exists a social-technical gap between what people want to accomplish and the actual capability of technology. In this paper, we present a case study comparing two voice assistants built on Amazon Alexa: one employing a human-in-the-loop workflow, the other utilizes LLM to engage in conversations with users. In our comparison, we discovered that the issues arising in current human-in-the-loop and LLM systems are not identical. However, the presence of a set of similar issues in both systems leads us to believe that focusing on the interaction between users and systems is crucial, perhaps even more so than focusing solely on the underlying technology itself. Merely enhancing the performance of the workers or the models may not adequately address these issues. This observation prompts our research question: What are the overlooked contributing factors in the effort to improve the capabilities of voice assistants, which might not have been emphasized in prior research?
人们很容易假定,大型语言模型(LLM)将无缝接管各种应用,尤其是那些基本自动化的应用。就会话语音助手而言,商业系统在过去十年中得到了广泛部署和使用。然而,我们是否真的站在了我们所设想的未来的风口浪尖上?人们想要实现的目标与技术的实际能力之间存在着社会技术差距。在本文中,我们介绍了一项案例研究,比较了基于亚马逊 Alexa 的两个语音助手:一个采用了人在回路中的工作流程,另一个则利用 LLM 与用户进行对话。在比较过程中,我们发现当前人工智能系统和 LLM 系统中出现的问题并不相同。不过,两个系统中存在的一系列类似问题让我们相信,关注用户与系统之间的互动至关重要,这或许比只关注底层技术本身更为重要。仅仅提高工作人员或模型的性能可能无法充分解决这些问题。这一观察结果引发了我们的研究问题:在努力提高语音助手能力的过程中,有哪些因素被忽视了,而这些因素在以前的研究中可能并没有得到重视?
{"title":"On Replacing Humans with Large Language Models in Voice-Based Human-in-the-Loop Systems","authors":"Shih-Hong Huang, Ting-Hao 'Kenneth' Huang","doi":"10.1609/aaaiss.v3i1.31178","DOIUrl":"https://doi.org/10.1609/aaaiss.v3i1.31178","url":null,"abstract":"It is easy to assume that Large Language Models (LLMs) will seamlessly take over applications, especially those that are largely automated. In the case of conversational voice assistants, commercial systems have been widely deployed and used over the past decade. However, are we indeed on the cusp of the future we envisioned? There exists a social-technical gap between what people want to accomplish and the actual capability of technology. In this paper, we present a case study comparing two voice assistants built on Amazon Alexa: one employing a human-in-the-loop workflow, the other utilizes LLM to engage in conversations with users. In our comparison, we discovered that the issues arising in current human-in-the-loop and LLM systems are not identical. However, the presence of a set of similar issues in both systems leads us to believe that focusing on the interaction between users and systems is crucial, perhaps even more so than focusing solely on the underlying technology itself. Merely enhancing the performance of the workers or the models may not adequately address these issues. This observation prompts our research question: What are the overlooked contributing factors in the effort to improve the capabilities of voice assistants, which might not have been emphasized in prior research?","PeriodicalId":516827,"journal":{"name":"Proceedings of the AAAI Symposium Series","volume":"45 10","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-05-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141118958","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Human-Like Learning of Social Reasoning via Analogy 通过类比进行类人社会推理学习
Pub Date : 2024-05-20 DOI: 10.1609/aaaiss.v3i1.31284
Irina Rabkina
Neurotypical adult humans are impeccably good social reasoners. Despite the occasional faux pas, we know how to interact in most social settings and how to consider others' points of view. Young children, on the other hand, do not. Social reasoning, like many of our most important skills, is learned. Much like human children, AI agents are not good social reasoners. While some algorithms can perform some aspects of social reasoning, we are a ways off from AI that can interact naturally and appropriately in the broad range of settings that people can. In this talk, I will argue that learning social reasoning via the same processes used by people will help AI agents reason--and interact--more like people do. Specifically, I will argue that children learn social reasoning via analogy, and that AI agents should, too. I will present evidence from cognitive modeling experiments demonstrating the former and AI experiments demonstrating the latter. I will also propose future directions for social reasoning research that both demonstrate the need for robust, human-like social reasoning in AI and test the utility of common approaches.
神经正常的成年人是无可挑剔的社交推理高手。尽管偶尔也会犯错,但我们知道如何在大多数社交场合进行互动,知道如何考虑他人的观点。而幼儿则不然。社交推理和我们许多最重要的技能一样,都是后天学习的。与人类儿童一样,人工智能代理也不擅长社交推理。虽然有些算法可以进行某些方面的社会推理,但我们距离人工智能能够像人类一样在广泛的环境中自然、恰当地进行互动还有一段距离。在本讲座中,我将论证,通过与人类相同的过程学习社会推理,将有助于人工智能代理更像人类那样进行推理和互动。具体来说,我将论证儿童通过类比学习社会推理,人工智能代理也应该如此。我将从认知建模实验和人工智能实验中提出证据,证明前者和后者。我还将提出社会推理研究的未来方向,既证明人工智能需要强大的、类似人类的社会推理,又测试常用方法的实用性。
{"title":"Human-Like Learning of Social Reasoning via Analogy","authors":"Irina Rabkina","doi":"10.1609/aaaiss.v3i1.31284","DOIUrl":"https://doi.org/10.1609/aaaiss.v3i1.31284","url":null,"abstract":"Neurotypical adult humans are impeccably good social reasoners. Despite the occasional faux pas, we know how to interact in most social settings and how to consider others' points of view. Young children, on the other hand, do not. Social reasoning, like many of our most important skills, is learned. \u0000\u0000Much like human children, AI agents are not good social reasoners. While some algorithms can perform some aspects of social reasoning, we are a ways off from AI that can interact naturally and appropriately in the broad range of settings that people can. In this talk, I will argue that learning social reasoning via the same processes used by people will help AI agents reason--and interact--more like people do. Specifically, I will argue that children learn social reasoning via analogy, and that AI agents should, too. I will present evidence from cognitive modeling experiments demonstrating the former and AI experiments demonstrating the latter. I will also propose future directions for social reasoning research that both demonstrate the need for robust, human-like social reasoning in AI and test the utility of common approaches.","PeriodicalId":516827,"journal":{"name":"Proceedings of the AAAI Symposium Series","volume":"10 15","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-05-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141119950","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Toward Autonomy: Metacognitive Learning for Enhanced AI Performance 迈向自主:元认知学习提升人工智能性能
Pub Date : 2024-05-20 DOI: 10.1609/aaaiss.v3i1.31270
Brendan Conway-Smith, Robert L. West
Large Language Models (LLMs) lack robust metacognitive learning abilities and depend on human-provided algorithms and prompts for learning and output generation. Metacognition involves processes that monitor and enhance cognition. Learning how to learn - metacognitive learning - is crucial for adapting and optimizing learning strategies over time. Although LLMs possess limited metacognitive abilities, they cannot autonomously refine or optimize these strategies. Humans possess innate mechanisms for metacognitive learning that enable at least two unique abilities: discerning which metacognitive strategies are best and automatizing learning strategies. These processes have been effectively modeled in the ACT-R cognitive architecture, providing insights on a path toward greater learning autonomy in AI. Incorporating human-like metacognitive learning abilities into AI could potentially lead to the development of more autonomous and versatile learning mechanisms, as well as improved problem-solving capabilities and performance across diverse tasks.
大型语言模型(LLM)缺乏强大的元认知学习能力,其学习和输出生成依赖于人类提供的算法和提示。元认知涉及监测和增强认知的过程。学会如何学习--元认知学习--对于随着时间的推移调整和优化学习策略至关重要。虽然低等语言学习者拥有有限的元认知能力,但他们无法自主完善或优化这些策略。人类拥有与生俱来的元认知学习机制,至少可以实现两种独特的能力:辨别哪种元认知策略是最好的,以及将学习策略自动化。ACT-R 认知架构对这些过程进行了有效建模,为人工智能实现更高的学习自主性提供了启示。将类似人类的元认知学习能力融入人工智能,有可能开发出更加自主和多用途的学习机制,并提高解决问题的能力和完成各种任务的性能。
{"title":"Toward Autonomy: Metacognitive Learning for Enhanced AI Performance","authors":"Brendan Conway-Smith, Robert L. West","doi":"10.1609/aaaiss.v3i1.31270","DOIUrl":"https://doi.org/10.1609/aaaiss.v3i1.31270","url":null,"abstract":"Large Language Models (LLMs) lack robust metacognitive learning abilities and depend on human-provided algorithms and prompts for learning and output generation. Metacognition involves processes that monitor and enhance cognition. Learning how to learn - metacognitive learning - is crucial for adapting and optimizing learning strategies over time. Although LLMs possess limited metacognitive abilities, they cannot autonomously refine or optimize these strategies. Humans possess innate mechanisms for metacognitive learning that enable at least two unique abilities: discerning which metacognitive strategies are best and automatizing learning strategies. These processes have been effectively modeled in the ACT-R cognitive architecture, providing insights on a path toward greater learning autonomy in AI. Incorporating human-like metacognitive learning abilities into AI could potentially lead to the development of more autonomous and versatile learning mechanisms, as well as improved problem-solving capabilities and performance across diverse tasks.","PeriodicalId":516827,"journal":{"name":"Proceedings of the AAAI Symposium Series","volume":"35 13","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-05-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141120698","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Leveraging Generative Artificial Intelligence to Broaden Participation in Computer Science 利用生成式人工智能扩大计算机科学的参与范围
Pub Date : 2024-05-20 DOI: 10.1609/aaaiss.v3i1.31262
Devang Jayachandran, P. Maldikar, Tyler S. Love, Jeremy Blum
Generative Artificial Intelligence (AI) was incorporated into a competitive programming event that targeted undergraduate students, including those with little programming experience. The competition incorporated a range of challenge design approaches that promoted meaningful interaction with generative AI system, even while keeping the challenge difficulty level to an appropriate level. An analysis of survey responses and competition data showed that this format lowered barriers to participation, successfully engaged students throughout the competition, and increased the likelihood that they would participate in a similar event. In an extension of this work, a professional development workshop for high school teachers is being developed, along with a contest for high school students. Participant surveys and logs of interaction with the contest and generative AI systems will be analyzed to measure the effect of generative AI on student self-efficacy and suggest ways to integrate generative AI instruction into computer science curriculum.
生成式人工智能(AI)被纳入了一项编程竞技活动,其目标群体是本科生,包括那些编程经验不足的学生。比赛采用了一系列挑战设计方法,促进与生成式人工智能系统进行有意义的互动,同时将挑战难度保持在适当水平。对调查反馈和比赛数据的分析表明,这种形式降低了参赛门槛,成功吸引了学生参与整个比赛,并提高了他们参加类似活动的可能性。作为这项工作的延伸,目前正在为高中教师举办专业发展研讨会,同时为高中学生举办竞赛。我们将分析参赛者调查以及与竞赛和生成式人工智能系统的互动记录,以衡量生成式人工智能对学生自我效能的影响,并就如何将生成式人工智能教学纳入计算机科学课程提出建议。
{"title":"Leveraging Generative Artificial Intelligence to Broaden Participation in Computer Science","authors":"Devang Jayachandran, P. Maldikar, Tyler S. Love, Jeremy Blum","doi":"10.1609/aaaiss.v3i1.31262","DOIUrl":"https://doi.org/10.1609/aaaiss.v3i1.31262","url":null,"abstract":"Generative Artificial Intelligence (AI) was incorporated into a competitive programming event that targeted undergraduate students, including those with little programming experience. The competition incorporated a range of challenge design approaches that promoted meaningful interaction with generative AI system, even while keeping the challenge difficulty level to an appropriate level. An analysis of survey responses and competition data showed that this format lowered barriers to participation, successfully engaged students throughout the competition, and increased the likelihood that they would participate in a similar event. In an extension of this work, a professional development workshop for high school teachers is being developed, along with a contest for high school students. Participant surveys and logs of interaction with the contest and generative AI systems will be analyzed to measure the effect of generative AI on student self-efficacy and suggest ways to integrate generative AI instruction into computer science curriculum.","PeriodicalId":516827,"journal":{"name":"Proceedings of the AAAI Symposium Series","volume":"21 17","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-05-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141120792","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Faithful Reasoning over Scientific Claims 对科学主张的忠实推理
Pub Date : 2024-05-20 DOI: 10.1609/aaaiss.v3i1.31209
N. Tan, Niket Tandon, David Wadden, Oyvind Tafjord, M. Gahegan, Michael Witbrock
Claim verification in scientific domains requires models that faithfully incorporate relevant knowledge from the ever-growing, vast existing literature. Unfaithful claim verifications can lead to misinformation such as those observed during the COVID-19 pandemic. Fact-checking systems often fail to capture the complex relationship between claims and evidence, especially with ambiguous claims and implicit assumptions. Relying only on current LLMs poses challenges due to hallucinations and information traceability issues. To address these challenges, our approach considers multiple viewpoints onto the scientific literature, enabling the assessment of contradictory arguments and implicit assumptions. Our proposed inference method adds faithful reasoning to large language models by distilling information from diverse, relevant scientific abstracts. This method provides a verdict label that can be weighted by the reputation of the scientific articles and an explanation that can be traced back to sources. Our findings demonstrate that humans not only perceive our explanation to be significantly superior to the off-the-shelf model, but they also evaluate it as faithfully enabling the tracing of evidence back to its original sources.
科学领域的索赔验证要求模型能够忠实地纳入不断增长的大量现有文献中的相关知识。不忠实的声称验证可能会导致错误信息,例如在 COVID-19 大流行期间观察到的错误信息。事实核查系统往往无法捕捉到主张与证据之间的复杂关系,尤其是在主张模棱两可和隐含假设的情况下。由于幻觉和信息可追溯性问题,仅依靠当前的法律知识会带来挑战。为了应对这些挑战,我们的方法考虑了科学文献的多种观点,从而能够评估相互矛盾的论点和隐含假设。我们提出的推理方法通过从不同的相关科学摘要中提炼信息,为大型语言模型添加了忠实推理。该方法提供了一个可根据科学文章的声誉加权的判决标签,以及一个可追溯来源的解释。我们的研究结果表明,人类不仅认为我们的解释明显优于现成的模型,而且还认为我们的解释能够忠实地将证据追溯到其原始来源。
{"title":"Faithful Reasoning over Scientific Claims","authors":"N. Tan, Niket Tandon, David Wadden, Oyvind Tafjord, M. Gahegan, Michael Witbrock","doi":"10.1609/aaaiss.v3i1.31209","DOIUrl":"https://doi.org/10.1609/aaaiss.v3i1.31209","url":null,"abstract":"Claim verification in scientific domains requires models that faithfully incorporate relevant knowledge from the ever-growing, vast existing literature. \u0000Unfaithful claim verifications can lead to misinformation such as those observed during the COVID-19 pandemic. Fact-checking systems often fail to capture the complex relationship between claims and evidence, especially with ambiguous claims and implicit assumptions. Relying only on current LLMs poses challenges due to hallucinations and information traceability issues. To address these challenges, our approach considers multiple viewpoints onto the scientific literature, enabling the assessment of contradictory arguments and implicit assumptions. Our proposed inference method adds faithful reasoning to large language models by distilling information from diverse, relevant scientific abstracts. This method provides a verdict label that can be weighted by the reputation of the scientific articles and an explanation that can be traced back to sources. Our findings demonstrate that humans not only perceive our explanation to be significantly superior to the off-the-shelf model, but they also evaluate it as faithfully enabling the tracing of evidence back to its original sources.","PeriodicalId":516827,"journal":{"name":"Proceedings of the AAAI Symposium Series","volume":"77 4","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-05-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141120971","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multi-Criterion Client Selection for Efficient Federated Learning 高效联盟学习的多标准客户端选择
Pub Date : 2024-05-20 DOI: 10.1609/aaaiss.v3i1.31227
Mehreen Tahir, Muhammad Intizar Ali
Federated Learning (FL) has received tremendous attention as a decentralized machine learning (ML) framework that allows distributed data owners to collaboratively train a global model without sharing raw data. Since FL trains the model directly on edge devices, the heterogeneity of participating clients in terms of data distribution, hardware capabilities and network connectivity can significantly impact the overall performance of FL systems. Optimizing for model accuracy could extend the training time due to the diverse and resource-constrained nature of edge devices while minimizing training time could compromise the model's accuracy. Effective client selection thus becomes crucial to ensure that the training process is not only efficient but also capitalizes on the diverse data and computational capabilities of different devices. To this end, we propose FedPROM, a novel framework that tackles client selection in FL as a multi-criteria optimization problem. By leveraging the PROMETHEE method, FedPROM ranks clients based on their suitability for a given FL task, considering multiple criteria such as system resources, network conditions, and data quality. This approach allows FedPROM to dynamically select the most appropriate set of clients for each learning round, optimizing both model accuracy and training efficiency. Our evaluations on diverse datasets demonstrate that FedPROM outperforms several state-of-the-art FL client selection protocols in terms of convergence speed, and accuracy, highlighting the framework's effectiveness and the importance of multi-criteria client selection in FL.
联邦学习(FL)作为一种去中心化的机器学习(ML)框架受到了极大的关注,它允许分布式数据所有者在不共享原始数据的情况下协作训练一个全局模型。由于联邦学习直接在边缘设备上训练模型,参与的客户端在数据分布、硬件能力和网络连接方面的异质性会极大地影响联邦学习系统的整体性能。由于边缘设备的多样性和资源有限性,优化模型准确性可能会延长训练时间,而尽量缩短训练时间则可能会影响模型的准确性。因此,有效的客户端选择对于确保训练过程不仅高效,而且充分利用不同设备的数据和计算能力至关重要。为此,我们提出了 FedPROM,这是一个新颖的框架,将 FL 中的客户端选择作为一个多标准优化问题来处理。通过利用 PROMETHEE 方法,FedPROM 在考虑系统资源、网络条件和数据质量等多重标准的基础上,根据客户端对特定 FL 任务的适用性对其进行排序。通过这种方法,FedPROM 可以为每一轮学习动态选择最合适的客户端集,从而优化模型准确性和训练效率。我们在不同数据集上进行的评估表明,FedPROM 在收敛速度和准确性方面都优于几种最先进的 FL 客户端选择协议,这凸显了该框架的有效性以及多标准客户端选择在 FL 中的重要性。
{"title":"Multi-Criterion Client Selection for Efficient Federated Learning","authors":"Mehreen Tahir, Muhammad Intizar Ali","doi":"10.1609/aaaiss.v3i1.31227","DOIUrl":"https://doi.org/10.1609/aaaiss.v3i1.31227","url":null,"abstract":"Federated Learning (FL) has received tremendous attention as a decentralized machine learning (ML) framework that allows distributed data owners to collaboratively train a global model without sharing raw data. Since FL trains the model directly on edge devices, the heterogeneity of participating clients in terms of data distribution, hardware capabilities and network connectivity can significantly impact the overall performance of FL systems. Optimizing for model accuracy could extend the training time due to the diverse and resource-constrained nature of edge devices while minimizing training time could compromise the model's accuracy. Effective client selection thus becomes crucial to ensure that the training process is not only efficient but also capitalizes on the diverse data and computational capabilities of different devices. To this end, we propose FedPROM, a novel framework that tackles client selection in FL as a multi-criteria optimization problem. By leveraging the PROMETHEE method, FedPROM ranks clients based on their suitability for a given FL task, considering multiple criteria such as system resources, network conditions, and data quality. This approach allows FedPROM to dynamically select the most appropriate set of clients for each learning round, optimizing both model accuracy and training efficiency. Our evaluations on diverse datasets demonstrate that FedPROM outperforms several state-of-the-art FL client selection protocols in terms of convergence speed, and accuracy, highlighting the framework's effectiveness and the importance of multi-criteria client selection in FL.","PeriodicalId":516827,"journal":{"name":"Proceedings of the AAAI Symposium Series","volume":"30 10","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-05-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141119064","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Proceedings of the AAAI Symposium Series
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1