首页 > 最新文献

Ai Magazine最新文献

英文 中文
Training robots with natural and lightweight human feedback 用自然和轻量级的人类反馈来训练机器人
IF 3.2 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-01-09 DOI: 10.1002/aaai.70037
Erdem Bıyık

Generalist robot models promise broad applicability across domains but currently require extensive expert demonstrations for task specialization, which is a costly and impractical barrier for real-world deployment. In this article, which summarizes the author's presentation in the New Faculty Highlights Track of the 39th annual AAAI Conference on Artificial Intelligence, we present algorithms that enable non-expert users to adapt and continually improve robot policies through natural and lightweight feedback modalities, such as preference comparisons, rankings, ratings, natural language, and users' own demonstrations, combining them with active learning strategies to maximize data-efficiency. We further introduce methods for leveraging real-time human interventions as rich training signals, modeling both their timing and absence to refine policies continually. Our approaches achieve substantial gains in sample-efficiency, adaptability, and user-friendliness, demonstrated across simulated and real-world robotic tasks. By aligning robot learning with how humans naturally teach, we hope to move toward autonomous systems that are more personalized, capable, and deployable in everyday environments.

通才机器人模型承诺广泛的跨领域适用性,但目前需要大量的专家演示来实现任务专业化,这是一个昂贵且不切实际的现实部署障碍。在这篇文章中,总结了作者在第39届年度AAAI人工智能会议新学院亮点轨道上的演讲,我们提出了一些算法,使非专业用户能够通过自然和轻量级的反馈方式(如偏好比较、排名、评级、自然语言和用户自己的演示)适应并不断改进机器人策略,并将它们与主动学习策略相结合,以最大限度地提高数据效率。我们进一步介绍了利用实时人为干预作为丰富的训练信号的方法,对其时间和缺席进行建模,以不断完善政策。我们的方法在样本效率、适应性和用户友好性方面取得了实质性的进展,并在模拟和现实世界的机器人任务中得到了证明。通过将机器人的学习与人类的自然教学方式结合起来,我们希望朝着更加个性化、更有能力、更能在日常环境中部署的自主系统迈进。
{"title":"Training robots with natural and lightweight human feedback","authors":"Erdem Bıyık","doi":"10.1002/aaai.70037","DOIUrl":"https://doi.org/10.1002/aaai.70037","url":null,"abstract":"<p>Generalist robot models promise broad applicability across domains but currently require extensive expert demonstrations for task specialization, which is a costly and impractical barrier for real-world deployment. In this article, which summarizes the author's presentation in the New Faculty Highlights Track of the 39<sup>th</sup> annual AAAI Conference on Artificial Intelligence, we present algorithms that enable non-expert users to adapt and continually improve robot policies through natural and lightweight feedback modalities, such as preference comparisons, rankings, ratings, natural language, and users' own demonstrations, combining them with active learning strategies to maximize data-efficiency. We further introduce methods for leveraging real-time human interventions as rich training signals, modeling both their timing and absence to refine policies continually. Our approaches achieve substantial gains in sample-efficiency, adaptability, and user-friendliness, demonstrated across simulated and real-world robotic tasks. By aligning robot learning with how humans naturally teach, we hope to move toward autonomous systems that are more personalized, capable, and deployable in everyday environments.</p>","PeriodicalId":7854,"journal":{"name":"Ai Magazine","volume":"47 1","pages":""},"PeriodicalIF":3.2,"publicationDate":"2026-01-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/aaai.70037","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145964133","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Knowledge Engineering for Open Science: Building and Deploying Knowledge Bases for Metadata Standards 面向开放科学的知识工程:为元数据标准构建和部署知识库
IF 3.2 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-12-31 DOI: 10.1002/aaai.70048
Mark A. Musen, Martin J. O'Connor, Josef Hardi, Marcos Martínez-Romero

For more than a decade, scientists have been striving to make their datasets available in open repositories, with the goal that they be findable, accessible, interoperable, and reusable (FAIR). Although it is hard for most investigators to remember all the “guiding principles” associated with FAIR data, there is one overarching requirement: The data need to be annotated with “rich,” discipline-specific, standardized metadata that can enable third parties to understand who performed the experiment, who or what the subjects were, what the experimental conditions were, and what the results appear to show. Most areas of science lack standards for such metadata and, when such standards exist, it can be difficult for investigators or data curators to apply them. The Center for Expanded Data Annotation and Retrieval (CEDAR) builds technology that enables scientists to encode descriptive metadata standards as templates that enumerate the attributes of different kinds of experiments and that link those attributes to ontologies or value sets that may supply controlled values for those attributes. These metadata templates capture the preferences of groups of investigators regarding how their data should be described and what a third party needs to know to make sense of their datasets. CEDAR templates describing community metadata preferences have been used to standardize metadata for a variety of scientific consortia. They have been used as the basis for data-annotation systems that acquire metadata through Web forms or through spreadsheets, and they can help correct metadata to ensure adherence to standards. Like the declarative knowledge bases that underpinned intelligent systems decades ago, CEDAR templates capture the knowledge of a community of practice in symbolic form, and they allow that knowledge to be applied in a variety of settings. They provide a mechanism for scientific communities to create shared metadata standards and to encode their preferences for the application of those standards, and for deploying those standards in a range of intelligent systems to promote open science.

十多年来,科学家们一直在努力使他们的数据集在开放存储库中可用,目标是可查找、可访问、可互操作和可重用(FAIR)。尽管大多数研究者很难记住与FAIR数据相关的所有“指导原则”,但有一个首要要求:数据需要用“丰富的”、特定学科的、标准化的元数据进行注释,这可以使第三方了解谁进行了实验,谁或什么受试者,实验条件是什么,以及结果显示了什么。大多数科学领域缺乏此类元数据的标准,即使存在此类标准,调查人员或数据管理员也很难应用它们。扩展数据注释和检索中心(CEDAR)建立了一种技术,使科学家能够将描述性元数据标准编码为模板,这些模板列举了不同类型实验的属性,并将这些属性链接到本体或值集,这些本体或值集可能为这些属性提供受控值。这些元数据模板捕获了调查小组关于如何描述他们的数据以及第三方需要知道什么才能理解他们的数据集的偏好。描述社区元数据偏好的CEDAR模板已被用于标准化各种科学联盟的元数据。它们被用作通过Web表单或电子表格获取元数据的数据注释系统的基础,并且它们可以帮助纠正元数据以确保符合标准。就像几十年前支撑智能系统的声明性知识库一样,CEDAR模板以符号形式捕获实践社区的知识,并允许将这些知识应用于各种设置中。它们为科学界提供了一种机制,以创建共享的元数据标准,并为这些标准的应用编码他们的偏好,并在一系列智能系统中部署这些标准,以促进开放科学。
{"title":"Knowledge Engineering for Open Science: Building and Deploying Knowledge Bases for Metadata Standards","authors":"Mark A. Musen,&nbsp;Martin J. O'Connor,&nbsp;Josef Hardi,&nbsp;Marcos Martínez-Romero","doi":"10.1002/aaai.70048","DOIUrl":"https://doi.org/10.1002/aaai.70048","url":null,"abstract":"<p>For more than a decade, scientists have been striving to make their datasets available in open repositories, with the goal that they be findable, accessible, interoperable, and reusable (FAIR). Although it is hard for most investigators to remember all the “guiding principles” associated with FAIR data, there is one overarching requirement: The data need to be annotated with “rich,” discipline-specific, standardized metadata that can enable third parties to understand who performed the experiment, who or what the subjects were, what the experimental conditions were, and what the results appear to show. Most areas of science lack standards for such metadata and, when such standards exist, it can be difficult for investigators or data curators to apply them. The Center for Expanded Data Annotation and Retrieval (CEDAR) builds technology that enables scientists to encode descriptive metadata standards as <i>templates</i> that enumerate the attributes of different kinds of experiments and that link those attributes to ontologies or value sets that may supply controlled values for those attributes. These metadata templates capture the preferences of groups of investigators regarding how their data should be described and what a third party needs to know to make sense of their datasets. CEDAR templates describing community metadata preferences have been used to standardize metadata for a variety of scientific consortia. They have been used as the basis for data-annotation systems that acquire metadata through Web forms or through spreadsheets, and they can help correct metadata to ensure adherence to standards. Like the declarative knowledge bases that underpinned intelligent systems decades ago, CEDAR templates capture the knowledge of a community of practice in symbolic form, and they allow that knowledge to be applied in a variety of settings. They provide a mechanism for scientific communities to create shared metadata standards and to encode their preferences for the application of those standards, and for deploying those standards in a range of intelligent systems to promote open science.</p>","PeriodicalId":7854,"journal":{"name":"Ai Magazine","volume":"47 1","pages":""},"PeriodicalIF":3.2,"publicationDate":"2025-12-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/aaai.70048","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145887954","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
AI for social science: A sociology PhD candidate's autoethnography on how LLMs are changing research work 社会科学的人工智能:一位社会学博士候选人关于法学硕士如何改变研究工作的自我民族志
IF 3.2 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-12-16 DOI: 10.1002/aaai.70046
Shuo Wang

Will AI replace social scientists? The real issue concerns reshaping rather than replacement. Confronting the integration of large language models (LLMs) into academic training establishes “prompt engineering” as the core interface for collaboration, defining it as a method to translate sociological thinking into precise instructions. LLMs are becoming essential partners across the research spectrum. They transform qualitative analysis from a solitary craft into a dialogical coding process and assist in theoretical localization and the construction of localized measurement scales. Beyond text analysis, they provide a low-cost virtual testbed for experimental design through “silicon samples” and enable the deduction of complex social interactions via “generative agents.” In the quantitative realm, they act as translators connecting research intentions with statistical code. Ultimately, the core challenge facing researchers is not technical. It lies in proactively cultivating a critical “literacy for human-AI collaboration” to master this paradigm shift.

人工智能会取代社会科学家吗?真正的问题在于重塑,而不是取代。面对将大型语言模型(llm)集成到学术培训中的问题,建立了“提示工程”作为协作的核心接口,将其定义为将社会学思维转化为精确指令的方法。法学硕士正在成为整个研究领域的重要合作伙伴。它们将定性分析从一种孤立的工艺转变为一种对话编码过程,并有助于理论定位和本地化测量尺度的构建。除了文本分析,他们还通过“硅样品”为实验设计提供了一个低成本的虚拟测试平台,并通过“生成代理”推导出复杂的社会互动。在定量领域,他们充当翻译,将研究意图与统计代码联系起来。最终,研究人员面临的核心挑战不是技术上的。关键在于积极培养一种关键的“人类与人工智能协作的素养”,以掌握这种范式转变。
{"title":"AI for social science: A sociology PhD candidate's autoethnography on how LLMs are changing research work","authors":"Shuo Wang","doi":"10.1002/aaai.70046","DOIUrl":"https://doi.org/10.1002/aaai.70046","url":null,"abstract":"<p>Will AI replace social scientists? The real issue concerns reshaping rather than replacement. Confronting the integration of large language models (LLMs) into academic training establishes “prompt engineering” as the core interface for collaboration, defining it as a method to translate sociological thinking into precise instructions. LLMs are becoming essential partners across the research spectrum. They transform qualitative analysis from a solitary craft into a dialogical coding process and assist in theoretical localization and the construction of localized measurement scales. Beyond text analysis, they provide a low-cost virtual testbed for experimental design through “silicon samples” and enable the deduction of complex social interactions via “generative agents.” In the quantitative realm, they act as translators connecting research intentions with statistical code. Ultimately, the core challenge facing researchers is not technical. It lies in proactively cultivating a critical “literacy for human-AI collaboration” to master this paradigm shift.</p>","PeriodicalId":7854,"journal":{"name":"Ai Magazine","volume":"46 4","pages":""},"PeriodicalIF":3.2,"publicationDate":"2025-12-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/aaai.70046","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145824391","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
The ETHICAL Protocol for Responsible Use of Generative AI for Research Purposes in Higher Education 在高等教育中为研究目的负责任地使用生成人工智能的伦理协议
IF 3.2 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-12-12 DOI: 10.1002/aaai.70047
Ahmed Alduais, Saba Qadhi, Youmen Chaaban, Majeda Khraisheh

Generative AI's growing use in higher education research requires strong protocols for responsible use. This need arises from the potential for misuse and the current uncertainty around ethical concerns and intellectual property. The lack of clear rules about openness in AI use, along with the “black box” nature of many AI systems, raises worries about reproducibility and the possibility of biased or fake results. This paper focuses specifically on generative AI tools (e.g., LLMs like ChatGPT, research-specific platforms like Elicit/SciSpace). The paper presents the ETHICAL protocol (i.e., Establish your purpose, Thoroughly explore options, Harness the appropriate tool, Inspect and verify output, Cite and reference accurately, Acknowledge AI usage transparently, and Look over publisher's guidelines), a detailed guide designed to direct researchers in the ethical and responsible inclusion of generative AI in their work. The protocol was created through a multi-step process, including a scientometric review of current trends, a systematic review of researcher experiences, and a policy analysis of 74 documents from various stakeholders (authorities, universities, publishers, and publication manuals). This analysis shaped the creation of a seven-heading, nine-item checklist covering key aspects of responsible AI use, from setting clear research goals to checking outputs and openly acknowledging AI help. The ETHICAL protocol gives practical examples and detailed explanations for each item, highlighting the importance of AI literacy and careful choice of suitable tools. It also stresses the vital need for checking AI-generated content to lessen the risk of errors and made-up information (“hallucinations”). The resulting protocol offers a practical and easy-to-use guide for researchers, encouraging responsible AI practices and upholding academic integrity. The ETHICAL protocol offers a helpful tool for managing the complex area of AI in research, ultimately leading to more open, reliable, and ethically sound scholarly work. Its broad acceptance could greatly improve the responsible use of AI in higher education, building trust and furthering knowledge growth.

生成式人工智能在高等教育研究中的越来越多的应用需要强有力的协议来负责任的使用。这种需求源于滥用的可能性以及目前围绕伦理问题和知识产权的不确定性。缺乏关于人工智能使用开放性的明确规则,加上许多人工智能系统的“黑箱”性质,引发了人们对可重复性以及有偏见或虚假结果的可能性的担忧。本文特别关注生成式人工智能工具(例如,法学硕士,如ChatGPT,研究专用平台,如Elicit/SciSpace)。本文介绍了伦理协议(即,建立您的目的,彻底探索选项,利用适当的工具,检查和验证输出,准确引用和参考,透明地承认人工智能的使用,并查看出版商的指导方针),这是一个详细的指南,旨在指导研究人员在其工作中道德和负责任地包括生成人工智能。该协议是通过一个多步骤过程创建的,包括对当前趋势进行科学计量学审查,对研究人员经验进行系统审查,以及对来自各利益攸关方(当局、大学、出版商和出版手册)的74份文件进行政策分析。这一分析形成了一个七标题、九项的清单,涵盖了负责任的人工智能使用的关键方面,从设定明确的研究目标到检查产出和公开承认人工智能的帮助。伦理协议为每个项目提供了实际示例和详细解释,强调了人工智能素养和谨慎选择合适工具的重要性。它还强调了检查人工智能生成内容的重要性,以减少错误和虚构信息(“幻觉”)的风险。由此产生的协议为研究人员提供了实用且易于使用的指南,鼓励负责任的人工智能实践并维护学术诚信。伦理协议为管理人工智能研究的复杂领域提供了一个有用的工具,最终导致更加开放、可靠和合乎道德的学术工作。它的广泛接受可以极大地改善人工智能在高等教育中的负责任地使用,建立信任并促进知识增长。
{"title":"The ETHICAL Protocol for Responsible Use of Generative AI for Research Purposes in Higher Education","authors":"Ahmed Alduais,&nbsp;Saba Qadhi,&nbsp;Youmen Chaaban,&nbsp;Majeda Khraisheh","doi":"10.1002/aaai.70047","DOIUrl":"https://doi.org/10.1002/aaai.70047","url":null,"abstract":"<p>Generative AI's growing use in higher education research requires strong protocols for responsible use. This need arises from the potential for misuse and the current uncertainty around ethical concerns and intellectual property. The lack of clear rules about openness in AI use, along with the “black box” nature of many AI systems, raises worries about reproducibility and the possibility of biased or fake results. This paper focuses specifically on generative AI tools (e.g., LLMs like ChatGPT, research-specific platforms like Elicit/SciSpace). The paper presents the ETHICAL protocol (i.e., <b>E</b>stablish your purpose, <b>T</b>horoughly explore options, <b>H</b>arness the appropriate tool, <b>I</b>nspect and verify output, <b>C</b>ite and reference accurately, <b>A</b>cknowledge AI usage transparently, and <b>L</b>ook over publisher's guidelines), a detailed guide designed to direct researchers in the ethical and responsible inclusion of generative AI in their work. The protocol was created through a multi-step process, including a scientometric review of current trends, a systematic review of researcher experiences, and a policy analysis of 74 documents from various stakeholders (authorities, universities, publishers, and publication manuals). This analysis shaped the creation of a seven-heading, nine-item checklist covering key aspects of responsible AI use, from setting clear research goals to checking outputs and openly acknowledging AI help. The ETHICAL protocol gives practical examples and detailed explanations for each item, highlighting the importance of AI literacy and careful choice of suitable tools. It also stresses the vital need for checking AI-generated content to lessen the risk of errors and made-up information (“hallucinations”). The resulting protocol offers a practical and easy-to-use guide for researchers, encouraging responsible AI practices and upholding academic integrity. The ETHICAL protocol offers a helpful tool for managing the complex area of AI in research, ultimately leading to more open, reliable, and ethically sound scholarly work. Its broad acceptance could greatly improve the responsible use of AI in higher education, building trust and furthering knowledge growth.</p>","PeriodicalId":7854,"journal":{"name":"Ai Magazine","volume":"46 4","pages":""},"PeriodicalIF":3.2,"publicationDate":"2025-12-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/aaai.70047","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145739767","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
How Proportional Representation Can Shape Artificial Intelligence 比例代表制如何影响人工智能
IF 3.2 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-12-03 DOI: 10.1002/aaai.70044
Evi Micha

Proportional representation is a foundational principle in social choice theory, ensuring that groups influence collective decisions in proportion to their size. While it has traditionally been studied in the context of political elections, recent work in computational social choice has broadened its scope to a variety of voting frameworks. This article showcases how proportional representation can be formalized and applied beyond these frameworks, spotlighting AI domains where it naturally takes shape. In particular, we focus on two such domains: clustering and AI alignment. In clustering, proportionality ensures that sufficiently large and cohesive groups of data points or agents are adequately represented in the selection of cluster centers or group assignments, to both centroid-based and noncentroid-based paradigms. In AI alignment, particularly in reinforcement learning from human feedback (RLHF), proportionality provides a principled framework for aggregating heterogeneous preferences by designing committees of reward functions that reflect annotators' viewpoints in proportion to their prevalence. We also discuss additional promising applications, including client selection in federated learning and forming committees of pre-trained models in meta-learning, and argue that incorporating proportional representation into AI systems provides a mathematically rigorous foundation for aligning algorithmic outcomes with the breadth of human viewpoints.

比例代表制是社会选择理论中的一个基本原则,它确保群体对集体决策的影响与其规模成正比。虽然它传统上是在政治选举的背景下研究的,但最近在计算社会选择方面的工作已将其范围扩大到各种投票框架。本文展示了比例代表如何形式化并应用于这些框架之外,突出了它自然形成的人工智能领域。我们特别关注两个这样的领域:聚类和AI对齐。在聚类中,比例性确保足够大且有凝聚力的数据点或代理在聚类中心的选择或组分配中得到充分的代表,无论是基于质心的还是基于非质心的范式。在人工智能校准中,特别是在从人类反馈中强化学习(RLHF)中,比例性提供了一个原则性框架,通过设计奖励函数委员会来聚合异质偏好,奖励函数委员会反映了注释者的观点与他们的流行程度成比例。我们还讨论了其他有前途的应用,包括联邦学习中的客户选择和元学习中预先训练模型的组成委员会,并认为将比例表示纳入人工智能系统为将算法结果与人类观点的广度相一致提供了数学上严格的基础。
{"title":"How Proportional Representation Can Shape Artificial Intelligence","authors":"Evi Micha","doi":"10.1002/aaai.70044","DOIUrl":"https://doi.org/10.1002/aaai.70044","url":null,"abstract":"<p>Proportional representation is a foundational principle in social choice theory, ensuring that groups influence collective decisions in proportion to their size. While it has traditionally been studied in the context of political elections, recent work in computational social choice has broadened its scope to a variety of voting frameworks. This article showcases how proportional representation can be formalized and applied beyond these frameworks, spotlighting AI domains where it naturally takes shape. In particular, we focus on two such domains: clustering and AI alignment. In clustering, proportionality ensures that sufficiently large and cohesive groups of data points or agents are adequately represented in the selection of cluster centers or group assignments, to both centroid-based and noncentroid-based paradigms. In AI alignment, particularly in reinforcement learning from human feedback (RLHF), proportionality provides a principled framework for aggregating heterogeneous preferences by designing committees of reward functions that reflect annotators' viewpoints in proportion to their prevalence. We also discuss additional promising applications, including client selection in federated learning and forming committees of pre-trained models in meta-learning, and argue that incorporating proportional representation into AI systems provides a mathematically rigorous foundation for aligning algorithmic outcomes with the breadth of human viewpoints.</p>","PeriodicalId":7854,"journal":{"name":"Ai Magazine","volume":"46 4","pages":""},"PeriodicalIF":3.2,"publicationDate":"2025-12-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/aaai.70044","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145686395","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Human at the Center: A Framework for Human-Driven AI Development 以人为中心:人类驱动的人工智能发展框架
IF 3.2 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-12-03 DOI: 10.1002/aaai.70043
Danniell Hu, Diana Acosta Navas, Susanne Gaube, Hussein Mozannar, Matthew E. Taylor, Krishnamurthy Dvijotham, Elizabeth Bondi-Kelly

Artificial Intelligence (AI) systems increasingly shape many aspects of daily life, influencing our jobs, finances, healthcare, and online content. This expansion has led to the rise of human–AI systems, where humans communicate, collaborate, or otherwise interact with AI, such as using AI outputs to make decisions. While these systems have shown potential to enhance human capabilities and improve performance on benchmarks, evidence suggests that they often underperform compared to AI-only or human-only approaches in experiments and real-world applications. Here, we argue that human–AI systems should be developed with a greater emphasis on human-centered factors—such as usability, fairness, trust, and user autonomy—within the algorithmic design and evaluation process. We advocate for integrating human-centered principles into AI development through human-centered algorithmic design and contextual evaluation with real users. Drawing on interdisciplinary research and our tutorial at two major AI conferences, we highlight examples and strategies for AI researchers and practitioners to embed these principles effectively. This work offers a systematic synthesis that integrates technical, practical, and ethical insights into a unified framework. Additionally, we highlight critical ethical considerations, including fairness, labor, privacy, and human agency to ensure that systems meet performance goals while serving broader societal interests. Through this work, we aim to inspire the field to embrace a truly human-centered approach to algorithmic design and deployment.

人工智能(AI)系统越来越多地影响着日常生活的方方面面,影响着我们的工作、财务、医疗保健和在线内容。这种扩张导致了人类-人工智能系统的兴起,在这种系统中,人类与人工智能进行交流、协作或以其他方式互动,例如使用人工智能的输出来做出决策。虽然这些系统已经显示出增强人类能力和提高基准性能的潜力,但有证据表明,在实验和实际应用中,与纯人工智能或纯人类方法相比,它们的表现往往不佳。在这里,我们认为人类-人工智能系统应该在算法设计和评估过程中更加强调以人为中心的因素,如可用性、公平性、信任和用户自主性。我们倡导通过以人为本的算法设计和真实用户的情境评估,将以人为本的原则融入人工智能开发。通过跨学科研究和我们在两个主要人工智能会议上的指导,我们为人工智能研究人员和从业者提供了有效嵌入这些原则的例子和策略。这项工作提供了一个系统的综合,将技术,实践和伦理见解集成到一个统一的框架中。此外,我们还强调了关键的道德考虑因素,包括公平、劳动、隐私和人类代理,以确保系统在满足性能目标的同时服务于更广泛的社会利益。通过这项工作,我们的目标是激励该领域采用真正以人为中心的算法设计和部署方法。
{"title":"Human at the Center: A Framework for Human-Driven AI Development","authors":"Danniell Hu,&nbsp;Diana Acosta Navas,&nbsp;Susanne Gaube,&nbsp;Hussein Mozannar,&nbsp;Matthew E. Taylor,&nbsp;Krishnamurthy Dvijotham,&nbsp;Elizabeth Bondi-Kelly","doi":"10.1002/aaai.70043","DOIUrl":"https://doi.org/10.1002/aaai.70043","url":null,"abstract":"<p>Artificial Intelligence (AI) systems increasingly shape many aspects of daily life, influencing our jobs, finances, healthcare, and online content. This expansion has led to the rise of human–AI systems, where humans communicate, collaborate, or otherwise interact with AI, such as using AI outputs to make decisions. While these systems have shown potential to enhance human capabilities and improve performance on benchmarks, evidence suggests that they often underperform compared to AI-only or human-only approaches in experiments and real-world applications. Here, we argue that human–AI systems should be developed with a greater emphasis on human-centered factors—such as usability, fairness, trust, and user autonomy—within the algorithmic design and evaluation process. We advocate for integrating human-centered principles into AI development through human-centered algorithmic design and contextual evaluation with real users. Drawing on interdisciplinary research and our tutorial at two major AI conferences, we highlight examples and strategies for AI researchers and practitioners to embed these principles effectively. This work offers a systematic synthesis that integrates technical, practical, and ethical insights into a unified framework. Additionally, we highlight critical ethical considerations, including fairness, labor, privacy, and human agency to ensure that systems meet performance goals while serving broader societal interests. Through this work, we aim to inspire the field to embrace a truly human-centered approach to algorithmic design and deployment.</p>","PeriodicalId":7854,"journal":{"name":"Ai Magazine","volume":"46 4","pages":""},"PeriodicalIF":3.2,"publicationDate":"2025-12-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/aaai.70043","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145686394","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Foundation Models for Anomaly Detection: Vision and Challenges 异常检测的基础模型:愿景与挑战
IF 3.2 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-12-02 DOI: 10.1002/aaai.70045
Jing Ren, Tao Tang, Hong Jia, Ziqi Xu, Haytham Fayek, Xiaodong Li, Suyu Ma, Xiwei Xu, Feng Xia

As data continues to grow in volume and complexity across domains such as finance, manufacturing, and healthcare, effective anomaly detection is essential for identifying irregular patterns that may signal critical issues. Recently, foundation models (FMs) have emerged as a powerful tool for advancing anomaly detection. They have demonstrated unprecedented capabilities in enhancing anomaly identification, generating detailed data descriptions, and providing visual explanations. This survey presents the first comprehensive review of recent advancements in FM-based anomaly detection. We propose a novel taxonomy for classifying FMs in anomaly detection tasks, based on the roles they play within the detection pipeline. Specifically, we categorize FMs as encoders, detectors, or interpreters, reflecting whether they are used for feature extraction, anomaly detection, or result interpretation, respectively. We provide a systematic analysis of state-of-the-art methods and discuss key challenges in leveraging FMs for improved anomaly detection. We also outline future research directions in this rapidly evolving field.

随着数据在金融、制造业和医疗保健等领域的数量和复杂性不断增长,有效的异常检测对于识别可能预示关键问题的不规则模式至关重要。近年来,基础模型(FMs)已成为推进异常检测的有力工具。它们在增强异常识别、生成详细数据描述和提供可视化解释方面展示了前所未有的能力。本调查首次全面回顾了基于fms的异常检测的最新进展。我们提出了一种新的分类方法来对异常检测任务中的FMs进行分类,基于它们在检测管道中所扮演的角色。具体来说,我们将fm分类为编码器、检测器或解释器,分别反映它们是否用于特征提取、异常检测或结果解释。我们对最先进的方法进行了系统的分析,并讨论了利用FMs改进异常检测的关键挑战。我们还概述了这一快速发展领域的未来研究方向。
{"title":"Foundation Models for Anomaly Detection: Vision and Challenges","authors":"Jing Ren,&nbsp;Tao Tang,&nbsp;Hong Jia,&nbsp;Ziqi Xu,&nbsp;Haytham Fayek,&nbsp;Xiaodong Li,&nbsp;Suyu Ma,&nbsp;Xiwei Xu,&nbsp;Feng Xia","doi":"10.1002/aaai.70045","DOIUrl":"https://doi.org/10.1002/aaai.70045","url":null,"abstract":"<p>As data continues to grow in volume and complexity across domains such as finance, manufacturing, and healthcare, effective anomaly detection is essential for identifying irregular patterns that may signal critical issues. Recently, foundation models (FMs) have emerged as a powerful tool for advancing anomaly detection. They have demonstrated unprecedented capabilities in enhancing anomaly identification, generating detailed data descriptions, and providing visual explanations. This survey presents the first comprehensive review of recent advancements in FM-based anomaly detection. We propose a novel taxonomy for classifying FMs in anomaly detection tasks, based on the roles they play within the detection pipeline. Specifically, we categorize FMs as encoders, detectors, or interpreters, reflecting whether they are used for feature extraction, anomaly detection, or result interpretation, respectively. We provide a systematic analysis of state-of-the-art methods and discuss key challenges in leveraging FMs for improved anomaly detection. We also outline future research directions in this rapidly evolving field.</p>","PeriodicalId":7854,"journal":{"name":"Ai Magazine","volume":"46 4","pages":""},"PeriodicalIF":3.2,"publicationDate":"2025-12-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/aaai.70045","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145695122","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Imitation learning is probably existentially safe 模仿学习可能是存在安全的
IF 3.2 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-11-21 DOI: 10.1002/aaai.70040
Michael K. Cohen, Marcus Hutter

Concerns about extinction risk from AI vary among experts in the field. However, AI encompasses a very broad category of algorithms. Perhaps some algorithms would pose an extinction risk, and others would not. Such an observation might be of great interest to both regulators and innovators. This paper argues that advanced imitation learners would likely not cause human extinction. We first present a simple argument to that effect, and then we rebut six different arguments that have been made to the contrary. A common theme of most of these arguments is a story for how a subroutine within an advanced imitation learner could hijack the imitation learner's behavior toward its own ends. However, we argue that each argument is flawed and each story implausible.

该领域的专家对人工智能带来的灭绝风险的担忧各不相同。然而,人工智能包含了非常广泛的算法类别。也许有些算法会带来灭绝的风险,而另一些则不会。这样的观察结果可能会引起监管者和创新者的极大兴趣。本文认为,高级模仿学习者可能不会导致人类灭绝。我们首先提出一个简单的论点,然后反驳六个不同的观点。这些争论的一个共同主题是关于高级模仿学习器中的子程序如何劫持模仿学习器的行为以达到自己的目的。然而,我们认为每个论点都有缺陷,每个故事都不可信。
{"title":"Imitation learning is probably existentially safe","authors":"Michael K. Cohen,&nbsp;Marcus Hutter","doi":"10.1002/aaai.70040","DOIUrl":"https://doi.org/10.1002/aaai.70040","url":null,"abstract":"<p>Concerns about extinction risk from AI vary among experts in the field. However, AI encompasses a very broad category of algorithms. Perhaps some algorithms would pose an extinction risk, and others would not. Such an observation might be of great interest to both regulators and innovators. This paper argues that advanced imitation learners would likely <i>not</i> cause human extinction. We first present a simple argument to that effect, and then we rebut six different arguments that have been made to the contrary. A common theme of most of these arguments is a story for how a subroutine within an advanced imitation learner could hijack the imitation learner's behavior toward its own ends. However, we argue that each argument is flawed and each story implausible.</p>","PeriodicalId":7854,"journal":{"name":"Ai Magazine","volume":"46 4","pages":""},"PeriodicalIF":3.2,"publicationDate":"2025-11-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/aaai.70040","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145581240","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
What do we want from AI? 我们想从人工智能中得到什么?
IF 3.2 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-11-08 DOI: 10.1002/aaai.70042
Robert B. Fisher

Most recent writing about the directions for AI has focused on the potential risks of widespread use of AI and what we DO NOT want from AI. This has led to many, largely ignored, calls for a pause in research and deployment. This essay takes the view that there are too many factors in play to slow the deployment much and for long. Hence, instead, this paper looks at what we DO want from AI (18 principles or goals) and how to get there.

最近关于人工智能方向的文章主要集中在人工智能广泛使用的潜在风险以及我们不希望从人工智能中得到什么。这导致许多人呼吁暂停研究和部署,但在很大程度上被忽视了。本文的观点是,有太多的因素在起作用,以减缓部署太多和长时间。因此,本文将着眼于我们真正想从AI中得到什么(18个原则或目标)以及如何实现这一目标。
{"title":"What do we want from AI?","authors":"Robert B. Fisher","doi":"10.1002/aaai.70042","DOIUrl":"https://doi.org/10.1002/aaai.70042","url":null,"abstract":"<p>Most recent writing about the directions for AI has focused on the potential risks of widespread use of AI and what we DO NOT want from AI. This has led to many, largely ignored, calls for a pause in research and deployment. This essay takes the view that there are too many factors in play to slow the deployment much and for long. Hence, instead, this paper looks at what we DO want from AI (18 principles or goals) and how to get there.</p>","PeriodicalId":7854,"journal":{"name":"Ai Magazine","volume":"46 4","pages":""},"PeriodicalIF":3.2,"publicationDate":"2025-11-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/aaai.70042","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145521580","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
RevOrder: A novel equation format for arithmetic operations in language models RevOrder:语言模型中用于算术运算的新型方程格式
IF 3.2 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-11-06 DOI: 10.1002/aaai.70038
Si Shen, Peijun Shen, Danhao Zhu

This paper proposes to understand arithmetic operations in Language Models (LM) by framing them as digit-based reasoning challenges. Our research focuses on arithmetic optimization challenges specific to LLMs, not on solving mathematical word problems. We introduce a metric called the Count of Sequential Intermediate Digits (CSID), which measures the complexity of arithmetic equations by counting the missing steps in digit reasoning. Our empirical findings suggest that increasing the model size does little to improve the handling of equations with high CSID values.

We propose RevOrder, a method that incorporates techniques such as reversing the output order, step-by-step decomposition, and rollback mechanisms to maintain a low CSID, thereby enhancing the solvability of arithmetic equations in LMs. RevOrder also introduces a more compact reasoning process, which reduces the token requirements without affecting the CSID, significantly enhancing token efficiency.

Comprehensive testing shows that RevOrder achieves perfect accuracy in operations such as addition, subtraction, and multiplication, and substantially improves performance in division tasks, especially with large numbers where traditional models falter.

本文提出通过将语言模型(LM)中的算术运算框架化为基于数字的推理挑战来理解它们。我们的研究重点是法学硕士特有的算法优化挑战,而不是解决数学单词问题。我们引入了一个称为顺序中间数字计数(CSID)的度量,它通过计算数字推理中缺失的步骤来度量算术方程的复杂性。我们的实证研究结果表明,增加模型大小对提高高CSID值方程的处理效果不大。我们提出RevOrder,这是一种结合了诸如反转输出顺序、逐步分解和回滚机制等技术的方法,以保持较低的CSID,从而增强了LMs中算术方程的可解性。RevOrder还引入了一个更紧凑的推理过程,在不影响CSID的情况下减少了令牌需求,显著提高了令牌效率。综合测试表明,RevOrder在加法、减法、乘法等运算中达到了完美的准确性,并且在除法任务中的性能得到了大幅提高,特别是在传统模型出现问题的大数字情况下。
{"title":"RevOrder: A novel equation format for arithmetic operations in language models","authors":"Si Shen,&nbsp;Peijun Shen,&nbsp;Danhao Zhu","doi":"10.1002/aaai.70038","DOIUrl":"https://doi.org/10.1002/aaai.70038","url":null,"abstract":"<p>This paper proposes to understand arithmetic operations in Language Models (LM) by framing them as digit-based reasoning challenges. Our research focuses on arithmetic optimization challenges specific to LLMs, not on solving mathematical word problems. We introduce a metric called the Count of Sequential Intermediate Digits (CSID), which measures the complexity of arithmetic equations by counting the missing steps in digit reasoning. Our empirical findings suggest that increasing the model size does little to improve the handling of equations with high CSID values.</p><p>We propose RevOrder, a method that incorporates techniques such as reversing the output order, step-by-step decomposition, and rollback mechanisms to maintain a low CSID, thereby enhancing the solvability of arithmetic equations in LMs. RevOrder also introduces a more compact reasoning process, which reduces the token requirements without affecting the CSID, significantly enhancing token efficiency.</p><p>Comprehensive testing shows that RevOrder achieves perfect accuracy in operations such as addition, subtraction, and multiplication, and substantially improves performance in division tasks, especially with large numbers where traditional models falter.</p>","PeriodicalId":7854,"journal":{"name":"Ai Magazine","volume":"46 4","pages":""},"PeriodicalIF":3.2,"publicationDate":"2025-11-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/aaai.70038","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145469655","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Ai Magazine
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1