首页 > 最新文献

Ai Magazine最新文献

英文 中文
When science fiction collides with reality: The future of learning and the one after that 当科幻小说与现实碰撞:学习的未来和未来
IF 2.5 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-07-04 DOI: 10.1002/aaai.70009
Steve Joordens

This article provides a somewhat whimsical discussion of the impact that AI, or “robots”, will have on the future of education. Interwoven with numerous references to science fiction, and at least one to Alice Cooper, is a very serious consideration of the manner in which AI may completely redefine the way we learn and grow as humans. From Holistic Assessment of Learning (HAL) to a Yoda on Yer Shoulda, a future of life-embedding learning is described.

本文对人工智能或“机器人”将对未来教育产生的影响进行了一些异想天开的讨论。与科幻小说交织在一起,至少有一个与爱丽丝·库珀(Alice Cooper)有关,这是对人工智能可能完全重新定义我们作为人类学习和成长方式的一种非常严肃的思考。从学习的整体评估(HAL)到yyshoulda上的尤达,描述了嵌入生命的学习的未来。
{"title":"When science fiction collides with reality: The future of learning and the one after that","authors":"Steve Joordens","doi":"10.1002/aaai.70009","DOIUrl":"https://doi.org/10.1002/aaai.70009","url":null,"abstract":"<p>This article provides a somewhat whimsical discussion of the impact that AI, or “robots”, will have on the future of education. Interwoven with numerous references to science fiction, and at least one to Alice Cooper, is a very serious consideration of the manner in which AI may completely redefine the way we learn and grow as humans. From Holistic Assessment of Learning (HAL) to a Yoda on Yer Shoulda, a future of life-embedding learning is described.</p>","PeriodicalId":7854,"journal":{"name":"Ai Magazine","volume":"46 2","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-07-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/aaai.70009","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144551058","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Governance in the age of artificial intelligence: A comparative analysis of policy framework in BRICS nations 人工智能时代的治理:金砖国家政策框架的比较分析
IF 2.5 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-07-04 DOI: 10.1002/aaai.70010
Animesh Kumar Sharma, Rahul Sharma

This study investigates the dynamic landscape of governance frameworks for emerging technologies, particularly artificial intelligence (AI), within the context of public policy in expanded BRICS nations (Brazil, Russia, India, China, South Africa, Egypt, Ethiopia, Iran, and the United Arab Emirates). Understanding the ethical implications and crafting policy tools to guide the development and deployment of AI is crucial. Analyzing findings from AI policy initiatives, this research delves into managing new technologies, emphasizing the evolving discourse on AI ethics. It stresses the importance of embedding ethical considerations into governance frameworks to address societal concerns and foster responsible AI advancement. Additionally, strong legal frameworks are essential, striking a balance between fostering innovation and ensuring accountability, thereby enhancing confidence and transparency in AI systems. This study underscores the significance of public policy in shaping AI governance, advocating for inclusive, participatory approaches involving stakeholders from diverse sectors. Adaptive governance frameworks capable of navigating the evolving AI landscape and its societal ramifications are emphasized. A holistic governance strategy based on insights from AI policy is recommended, aiming to reconcile innovation with ethical, legal, and societal considerations. Policymakers are urged to foster stakeholder engagement, ensuring that AI advancements benefit society while upholding ethical, just, and accountable standards.

本研究在扩大后的金砖国家(巴西、俄罗斯、印度、中国、南非、埃及、埃塞俄比亚、伊朗和阿拉伯联合酋长国)的公共政策背景下,调查了新兴技术治理框架的动态格局,尤其是人工智能(AI)。理解伦理影响并制定政策工具来指导人工智能的开发和部署至关重要。本研究分析了人工智能政策倡议的结果,深入研究了新技术的管理,强调了人工智能伦理的不断发展。它强调了将伦理考虑纳入治理框架的重要性,以解决社会关切并促进负责任的人工智能发展。此外,强有力的法律框架至关重要,在促进创新和确保问责制之间取得平衡,从而增强对人工智能系统的信心和透明度。本研究强调了公共政策在塑造人工智能治理方面的重要性,倡导采用涉及不同部门利益相关者的包容性、参与性方法。强调了能够驾驭不断发展的人工智能景观及其社会后果的自适应治理框架。建议采用基于人工智能政策见解的整体治理策略,旨在将创新与道德、法律和社会考虑相协调。敦促政策制定者促进利益相关者的参与,确保人工智能的进步造福社会,同时坚持道德、公正和负责任的标准。
{"title":"Governance in the age of artificial intelligence: A comparative analysis of policy framework in BRICS nations","authors":"Animesh Kumar Sharma,&nbsp;Rahul Sharma","doi":"10.1002/aaai.70010","DOIUrl":"https://doi.org/10.1002/aaai.70010","url":null,"abstract":"<p>This study investigates the dynamic landscape of governance frameworks for emerging technologies, particularly artificial intelligence (AI), within the context of public policy in expanded BRICS nations (Brazil, Russia, India, China, South Africa, Egypt, Ethiopia, Iran, and the United Arab Emirates). Understanding the ethical implications and crafting policy tools to guide the development and deployment of AI is crucial. Analyzing findings from AI policy initiatives, this research delves into managing new technologies, emphasizing the evolving discourse on AI ethics. It stresses the importance of embedding ethical considerations into governance frameworks to address societal concerns and foster responsible AI advancement. Additionally, strong legal frameworks are essential, striking a balance between fostering innovation and ensuring accountability, thereby enhancing confidence and transparency in AI systems. This study underscores the significance of public policy in shaping AI governance, advocating for inclusive, participatory approaches involving stakeholders from diverse sectors. Adaptive governance frameworks capable of navigating the evolving AI landscape and its societal ramifications are emphasized. A holistic governance strategy based on insights from AI policy is recommended, aiming to reconcile innovation with ethical, legal, and societal considerations. Policymakers are urged to foster stakeholder engagement, ensuring that AI advancements benefit society while upholding ethical, just, and accountable standards.</p>","PeriodicalId":7854,"journal":{"name":"Ai Magazine","volume":"46 2","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-07-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/aaai.70010","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144551018","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Building trust: Foundations of security, safety, and transparency in AI 建立信任:人工智能安全、安全和透明的基础
IF 2.5 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-05-20 DOI: 10.1002/aaai.70005
Huzaifa Sidhpurwala, Garth Mollett, Emily Fox, Mark Bestavros, Huamin Chen

This paper explores the rapidly evolving ecosystem of publicly available AI models and their potential implications on the security and safety landscape. Understanding their potential risks and vulnerabilities is crucial as AI models become increasingly prevalent. We review the current security and safety scenarios while highlighting challenges such as tracking issues, remediation, and the absence of AI model lifecycle and ownership processes. Comprehensive strategies to enhance security and safety for both model developers and end-users are proposed. This paper provides several foundational pieces for more standardized security, safety, and transparency in developing and operating generative AI models and the larger open ecosystems and communities forming around them.

本文探讨了公开可用的人工智能模型的快速发展生态系统及其对安全和安全领域的潜在影响。随着人工智能模型变得越来越普遍,了解它们的潜在风险和漏洞至关重要。我们回顾了当前的安全和安全场景,同时强调了诸如跟踪问题、补救以及缺乏人工智能模型生命周期和所有权流程等挑战。提出了增强模型开发人员和最终用户的安全性的综合策略。本文为开发和操作生成式人工智能模型以及围绕它们形成的更大的开放生态系统和社区提供了更标准化的安全性、安全性和透明度的几个基础部分。
{"title":"Building trust: Foundations of security, safety, and transparency in AI","authors":"Huzaifa Sidhpurwala,&nbsp;Garth Mollett,&nbsp;Emily Fox,&nbsp;Mark Bestavros,&nbsp;Huamin Chen","doi":"10.1002/aaai.70005","DOIUrl":"https://doi.org/10.1002/aaai.70005","url":null,"abstract":"<p>This paper explores the rapidly evolving ecosystem of publicly available AI models and their potential implications on the security and safety landscape. Understanding their potential risks and vulnerabilities is crucial as AI models become increasingly prevalent. We review the current security and safety scenarios while highlighting challenges such as tracking issues, remediation, and the absence of AI model lifecycle and ownership processes. Comprehensive strategies to enhance security and safety for both model developers and end-users are proposed. This paper provides several foundational pieces for more standardized security, safety, and transparency in developing and operating generative AI models and the larger open ecosystems and communities forming around them.</p>","PeriodicalId":7854,"journal":{"name":"Ai Magazine","volume":"46 2","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-05-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/aaai.70005","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144100705","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
What is reproducibility in artificial intelligence and machine learning research? 人工智能和机器学习研究中的可重复性是什么?
IF 2.5 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-04-18 DOI: 10.1002/aaai.70004
Abhyuday Desai, Mohamed Abdelhamid, Nakul R. Padalkar

In the rapidly evolving fields of artificial intelligence (AI) and machine learning (ML), the reproducibility crisis underscores the urgent need for clear validation methodologies to maintain scientific integrity and encourage advancement. The crisis is compounded by the prevalent confusion over validation terminology. In response to this challenge, we introduce a framework that clarifies the roles and definitions of key validation efforts: repeatability, dependent and independent reproducibility, and direct and conceptual replicability. This structured framework aims to provide AI/ML researchers with the necessary clarity on these essential concepts, facilitating the appropriate design, conduct, and interpretation of validation studies. By articulating the nuances and specific roles of each type of validation study, we aim to enhance the reliability and trustworthiness of research findings and support the community's efforts to address reproducibility challenges effectively.

在快速发展的人工智能(AI)和机器学习(ML)领域,可重复性危机凸显出迫切需要明确的验证方法来维护科学的完整性并鼓励进步。验证术语的普遍混淆加剧了这一危机。为了应对这一挑战,我们提出了一个框架,明确了关键验证工作的作用和定义:可重复性、从属和独立可重复性以及直接和概念可重复性。这一结构化框架旨在为人工智能/移动语言研究人员提供有关这些基本概念的必要清晰度,从而促进验证研究的适当设计、实施和解释。通过阐明每类验证研究的细微差别和具体作用,我们旨在提高研究成果的可靠性和可信度,并支持社会各界有效应对可重复性挑战的努力。
{"title":"What is reproducibility in artificial intelligence and machine learning research?","authors":"Abhyuday Desai,&nbsp;Mohamed Abdelhamid,&nbsp;Nakul R. Padalkar","doi":"10.1002/aaai.70004","DOIUrl":"https://doi.org/10.1002/aaai.70004","url":null,"abstract":"<p>In the rapidly evolving fields of artificial intelligence (AI) and machine learning (ML), the reproducibility crisis underscores the urgent need for clear validation methodologies to maintain scientific integrity and encourage advancement. The crisis is compounded by the prevalent confusion over validation terminology. In response to this challenge, we introduce a framework that clarifies the roles and definitions of key validation efforts: repeatability, dependent and independent reproducibility, and direct and conceptual replicability. This structured framework aims to provide AI/ML researchers with the necessary clarity on these essential concepts, facilitating the appropriate design, conduct, and interpretation of validation studies. By articulating the nuances and specific roles of each type of validation study, we aim to enhance the reliability and trustworthiness of research findings and support the community's efforts to address reproducibility challenges effectively.</p>","PeriodicalId":7854,"journal":{"name":"Ai Magazine","volume":"46 2","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-04-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/aaai.70004","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143849301","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Open issues in open world learning 开放世界学习中的开放问题
IF 2.5 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-04-14 DOI: 10.1002/aaai.70001
Steve Cruz, Katarina Doctor, Christopher Funk, Walter Scheirer

Meaningful progress has been made in open world learning (OWL), enhancing the ability of agents to detect, characterize, and incrementally learn novelty in dynamic environments. However, novelty remains a persistent challenge for agents relying on state-of-the-art learning algorithms. This article considers the current state of OWL, drawing on insights from a recent DARPA research program on this topic. We identify open issues that impede further advancements spanning theory, design, and evaluation. In particular, we emphasize the challenges posed by dynamic scenarios that are crucial to understand for ensuring the viability of agents designed for real-world environments. The article provides suggestions for setting a new research agenda that effectively addresses these open issues.

在开放世界学习(OWL)方面取得了有意义的进展,增强了智能体在动态环境中检测、表征和增量学习新颖性的能力。然而,对于依赖最先进的学习算法的代理来说,新颖性仍然是一个持续的挑战。本文考虑了OWL的当前状态,借鉴了DARPA最近关于该主题的一个研究项目的见解。我们发现了阻碍理论、设计和评估进一步发展的开放性问题。我们特别强调了动态场景所带来的挑战,这些挑战对于理解为现实世界环境设计的智能体的可行性至关重要。本文为制定新的研究议程提供了建议,以有效地解决这些悬而未决的问题。
{"title":"Open issues in open world learning","authors":"Steve Cruz,&nbsp;Katarina Doctor,&nbsp;Christopher Funk,&nbsp;Walter Scheirer","doi":"10.1002/aaai.70001","DOIUrl":"https://doi.org/10.1002/aaai.70001","url":null,"abstract":"<p>Meaningful progress has been made in open world learning (OWL), enhancing the ability of agents to detect, characterize, and incrementally learn novelty in dynamic environments. However, novelty remains a persistent challenge for agents relying on state-of-the-art learning algorithms. This article considers the current state of OWL, drawing on insights from a recent DARPA research program on this topic. We identify open issues that impede further advancements spanning theory, design, and evaluation. In particular, we emphasize the challenges posed by dynamic scenarios that are crucial to understand for ensuring the viability of agents designed for real-world environments. The article provides suggestions for setting a new research agenda that effectively addresses these open issues.</p>","PeriodicalId":7854,"journal":{"name":"Ai Magazine","volume":"46 2","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-04-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/aaai.70001","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143831393","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Reproducibility in machine-learning-based research: Overview, barriers, and drivers 基于机器学习的研究的可重复性:概述、障碍和驱动因素
IF 2.5 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-04-14 DOI: 10.1002/aaai.70002
Harald Semmelrock, Tony Ross-Hellauer, Simone Kopeinik, Dieter Theiler, Armin Haberl, Stefan Thalmann, Dominik Kowald

Many research fields are currently reckoning with issues of poor levels of reproducibility. Some label it a “crisis,” and research employing or building machine learning (ML) models is no exception. Issues including lack of transparency, data or code, poor adherence to standards, and the sensitivity of ML training conditions mean that many papers are not even reproducible in principle. Where they are, though, reproducibility experiments have found worryingly low degrees of similarity with original results. Despite previous appeals from ML researchers on this topic and various initiatives from conference reproducibility tracks to the ACM's new Emerging Interest Group on Reproducibility and Replicability, we contend that the general community continues to take this issue too lightly. Poor reproducibility threatens trust in and integrity of research results. Therefore, in this article, we lay out a new perspective on the key barriers and drivers (both procedural and technical) to increased reproducibility at various levels (methods, code, data, and experiments). We then map the drivers to the barriers to give concrete advice for strategies for researchers to mitigate reproducibility issues in their own work, to lay out key areas where further research is needed in specific areas, and to further ignite discussion on the threat presented by these urgent issues.

许多研究领域目前都在考虑可重复性差的问题。一些人将其称为“危机”,使用或构建机器学习(ML)模型的研究也不例外。缺乏透明度、数据或代码、对标准的依从性差以及ML训练条件的敏感性等问题意味着许多论文在原则上甚至是不可复制的。然而,可重复性实验发现,与原始结果的相似度低得令人担忧。尽管之前机器学习研究人员对这个主题提出了呼吁,并且从会议可重复性跟踪到ACM新成立的可重复性和可复制性兴趣小组,我们认为一般社区仍然对这个问题过于轻视。低可重复性威胁到对研究结果的信任和完整性。因此,在本文中,我们将从一个新的角度来看待在不同层次(方法、代码、数据和实验)上提高再现性的关键障碍和驱动因素(程序和技术)。然后,我们将驱动因素映射到障碍,为研究人员提供具体的策略建议,以减轻他们自己工作中的可重复性问题,列出在特定领域需要进一步研究的关键领域,并进一步引发对这些紧迫问题所带来的威胁的讨论。
{"title":"Reproducibility in machine-learning-based research: Overview, barriers, and drivers","authors":"Harald Semmelrock,&nbsp;Tony Ross-Hellauer,&nbsp;Simone Kopeinik,&nbsp;Dieter Theiler,&nbsp;Armin Haberl,&nbsp;Stefan Thalmann,&nbsp;Dominik Kowald","doi":"10.1002/aaai.70002","DOIUrl":"https://doi.org/10.1002/aaai.70002","url":null,"abstract":"<p>Many research fields are currently reckoning with issues of poor levels of reproducibility. Some label it a “crisis,” and research employing or building machine learning (ML) models is no exception. Issues including lack of transparency, data or code, poor adherence to standards, and the sensitivity of ML training conditions mean that many papers are not even reproducible in principle. Where they are, though, reproducibility experiments have found worryingly low degrees of similarity with original results. Despite previous appeals from ML researchers on this topic and various initiatives from conference reproducibility tracks to the ACM's new Emerging Interest Group on Reproducibility and Replicability, we contend that the general community continues to take this issue too lightly. Poor reproducibility threatens trust in and integrity of research results. Therefore, in this article, we lay out a new perspective on the key barriers and drivers (both procedural and technical) to increased reproducibility at various levels (methods, code, data, and experiments). We then map the drivers to the barriers to give concrete advice for strategies for researchers to mitigate reproducibility issues in their own work, to lay out key areas where further research is needed in specific areas, and to further ignite discussion on the threat presented by these urgent issues.</p>","PeriodicalId":7854,"journal":{"name":"Ai Magazine","volume":"46 2","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-04-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/aaai.70002","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143831231","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
PADTHAI-MM: Principles-based approach for designing trustworthy, human-centered AI using the MAST methodology PADTHAI-MM:基于原则的方法,使用MAST方法设计可信赖的、以人为本的人工智能
IF 2.5 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-03-18 DOI: 10.1002/aaai.70000
Myke C. Cohen, Nayoung Kim, Yang Ba, Anna Pan, Shawaiz Bhatti, Pouria Salehi, James Sung, Erik Blasch, Mickey V. Mancenido, Erin K. Chiou

Despite an extensive body of literature on trust in technology, designing trustworthy AI systems for high-stakes decision domains remains a significant challenge. Widely used system design guidelines and tools are rarely attuned to domain-specific trustworthiness principles. In this study, we introduce a design framework to address this gap within intelligence analytic tasks, called the Principles-based Approach for Designing Trustworthy, Human-centered AI using the MAST Methodology (PADTHAI-MM). PADTHAI-MM builds on the Multisource AI Scorecard Table (MAST), an AI decision support system evaluation tool designed in accordance to the U.S. Intelligence Community's standards for system trustworthiness. We demonstrate PADTHAI-MM in our development of the Reporting Assistant for Defense and Intelligence Tasks (READIT), a research platform that leverages data visualizations and natural language processing-based text analysis to emulate AI-enabled intelligence reporting aids. To empirically assess the efficacy of PADTHAI-MM, we developed two versions of READIT for comparison: a “High-MAST” version, which incorporates AI contextual information and explanations, and a “Low-MAST” version, designed to be akin to inscrutable “black box” AI systems. Through an iterative design process guided by stakeholder feedback, our multidisciplinary design team developed prototypes that were evaluated by experienced intelligence analysts. Results substantially supported the viability of PADTHAI-MM in designing for system trustworthiness in this task domain. We also explored the relationship between analysts' MAST ratings and three theoretical categories of information known to impact trust: process, purpose, and performance. Overall, our study supports the practical and theoretical viability of PADTHAI-MM as an approach to designing trustable AI systems.

尽管有大量关于技术信任的文献,但为高风险决策领域设计值得信赖的人工智能系统仍然是一个重大挑战。广泛使用的系统设计指南和工具很少与领域特定的可信度原则相协调。在本研究中,我们引入了一个设计框架来解决智能分析任务中的这一差距,称为基于原则的方法,用于使用MAST方法论(PADTHAI-MM)设计可信赖的、以人为中心的人工智能。PADTHAI-MM基于多源AI记分卡表(MAST),这是一种根据美国情报界系统可信度标准设计的AI决策支持系统评估工具。我们在国防和情报任务报告助理(READIT)的开发中展示了PADTHAI-MM,这是一个利用数据可视化和基于自然语言处理的文本分析来模拟人工智能智能报告辅助的研究平台。为了从经验上评估PADTHAI-MM的有效性,我们开发了两个版本的READIT进行比较:一个是“高桅杆”版本,其中包含人工智能上下文信息和解释,另一个是“低桅杆”版本,旨在类似于不可思议的“黑匣子”人工智能系统。通过由利益相关者反馈指导的迭代设计过程,我们的多学科设计团队开发了由经验丰富的情报分析师评估的原型。结果充分支持了PADTHAI-MM在该任务域中设计系统可信度的可行性。我们还探讨了分析师的MAST评级与已知影响信任的三个理论信息类别之间的关系:过程、目的和绩效。总的来说,我们的研究支持PADTHAI-MM作为设计可信赖的人工智能系统的方法的实践和理论可行性。
{"title":"PADTHAI-MM: Principles-based approach for designing trustworthy, human-centered AI using the MAST methodology","authors":"Myke C. Cohen,&nbsp;Nayoung Kim,&nbsp;Yang Ba,&nbsp;Anna Pan,&nbsp;Shawaiz Bhatti,&nbsp;Pouria Salehi,&nbsp;James Sung,&nbsp;Erik Blasch,&nbsp;Mickey V. Mancenido,&nbsp;Erin K. Chiou","doi":"10.1002/aaai.70000","DOIUrl":"https://doi.org/10.1002/aaai.70000","url":null,"abstract":"<p>Despite an extensive body of literature on trust in technology, designing trustworthy AI systems for high-stakes decision domains remains a significant challenge. Widely used system design guidelines and tools are rarely attuned to domain-specific trustworthiness principles. In this study, we introduce a design framework to address this gap within intelligence analytic tasks, called the Principles-based Approach for Designing Trustworthy, Human-centered AI using the MAST Methodology (PADTHAI-MM). PADTHAI-MM builds on the Multisource AI Scorecard Table (MAST), an AI decision support system evaluation tool designed in accordance to the U.S. Intelligence Community's standards for system trustworthiness. We demonstrate PADTHAI-MM in our development of the Reporting Assistant for Defense and Intelligence Tasks (READIT), a research platform that leverages data visualizations and natural language processing-based text analysis to emulate AI-enabled intelligence reporting aids. To empirically assess the efficacy of PADTHAI-MM, we developed two versions of READIT for comparison: a “High-MAST” version, which incorporates AI contextual information and explanations, and a “Low-MAST” version, designed to be akin to inscrutable “black box” AI systems. Through an iterative design process guided by stakeholder feedback, our multidisciplinary design team developed prototypes that were evaluated by experienced intelligence analysts. Results substantially supported the viability of PADTHAI-MM in designing for system trustworthiness in this task domain. We also explored the relationship between analysts' MAST ratings and three theoretical categories of information known to impact trust: <i>process</i>, <i>purpose</i>, and <i>performance</i>. Overall, our study supports the practical and theoretical viability of PADTHAI-MM as an approach to designing trustable AI systems.</p>","PeriodicalId":7854,"journal":{"name":"Ai Magazine","volume":"46 1","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-03-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/aaai.70000","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143638790","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
What AIs are not learning (and why) 人工智能没有学到什么(以及为什么)
IF 2.5 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-03-03 DOI: 10.1002/aaai.12213
Mark Stefik

Today's robots do not yet learn the general skills that are necessary to provide home care, to be nursing assistants, to interact with people, or do household chores nearly as well as people do. Addressing the aspirational goal of creating service robots requires improving how they are created. Today's mainstream AIs are not created by agents learning from experiences doing tasks in real-world contexts and interacting with people. Today's robots do not learn by sensing, acting, doing experiments, and collaborating. Future robots will need to learn from such experiences in order to be ready for robust deployment in human service applications. This paper investigates what aspirational future autonomous human-compatible service robots will need to know. It recommends developing experiential (robotic) foundation models (FMs) for bootstrapping them.

今天的机器人还没有学会提供家庭护理、成为护理助理、与人互动或像人一样做家务所必需的一般技能。要实现创造服务型机器人的理想目标,就需要改进它们的创造方式。今天的主流人工智能并不是由智能体从现实世界中完成任务和与人互动的经验中学习出来的。今天的机器人不是通过感知、行动、实验和合作来学习的。未来的机器人将需要从这些经验中学习,以便为人类服务应用的强大部署做好准备。本文研究了未来自主的人类兼容服务机器人需要知道什么。它建议开发经验(机器人)基础模型(FMs)来引导他们。
{"title":"What AIs are not learning (and why)","authors":"Mark Stefik","doi":"10.1002/aaai.12213","DOIUrl":"https://doi.org/10.1002/aaai.12213","url":null,"abstract":"<p>Today's robots do not yet learn the general skills that are necessary to provide home care, to be nursing assistants, to interact with people, or do household chores nearly as well as people do. Addressing the aspirational goal of creating service robots requires improving how they are created. Today's mainstream AIs are not created by agents learning from experiences doing tasks in real-world contexts and interacting with people. Today's robots do not learn by sensing, acting, doing experiments, and collaborating. Future robots will need to learn from such experiences in order to be ready for robust deployment in human service applications. This paper investigates what aspirational future autonomous human-compatible service robots will need to know. It recommends developing <i>experiential</i> (robotic) foundation models (FMs) for bootstrapping them.</p>","PeriodicalId":7854,"journal":{"name":"Ai Magazine","volume":"46 1","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-03-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/aaai.12213","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143535974","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Fairness amidst non-IID graph data: A literature review 非iid图形数据中的公平性:文献综述
IF 2.5 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-01-28 DOI: 10.1002/aaai.12212
Wenbin Zhang, Shuigeng Zhou, Toby Walsh, Jeremy C. Weiss

The growing importance of understanding and addressing algorithmic bias in artificial intelligence (AI) has led to a surge in research on AI fairness, which often assumes that the underlying data are independent and identically distributed (IID). However, real-world data frequently exist in non-IID graph structures that capture connections among individual units. To effectively mitigate bias in AI systems, it is essential to bridge the gap between traditional fairness literature, designed for IID data, and the prevalence of non-IID graph data. This survey reviews recent advancements in fairness amidst non-IID graph data, including the newly introduced fair graph generation and the commonly studied fair graph classification. In addition, available datasets and evaluation metrics for future research are identified, the limitations of existing work are highlighted, and promising future directions are proposed.

理解和解决人工智能(AI)中算法偏见的重要性日益增加,这导致了人工智能公平性研究的激增,这些研究通常假设底层数据是独立和同分布的(IID)。然而,真实世界的数据经常存在于非iid图结构中,这些结构捕获了各个单元之间的连接。为了有效减轻人工智能系统中的偏见,有必要弥合为IID数据设计的传统公平文献与非IID图形数据的流行之间的差距。本调查回顾了在非iid图数据中公平性的最新进展,包括新引入的公平图生成和通常研究的公平图分类。此外,确定了未来研究的可用数据集和评估指标,强调了现有工作的局限性,并提出了有希望的未来方向。
{"title":"Fairness amidst non-IID graph data: A literature review","authors":"Wenbin Zhang,&nbsp;Shuigeng Zhou,&nbsp;Toby Walsh,&nbsp;Jeremy C. Weiss","doi":"10.1002/aaai.12212","DOIUrl":"https://doi.org/10.1002/aaai.12212","url":null,"abstract":"<p>The growing importance of understanding and addressing algorithmic bias in artificial intelligence (AI) has led to a surge in research on AI fairness, which often assumes that the underlying data are independent and identically distributed (IID). However, real-world data frequently exist in non-IID graph structures that capture connections among individual units. To effectively mitigate bias in AI systems, it is essential to bridge the gap between traditional fairness literature, designed for IID data, and the prevalence of non-IID graph data. This survey reviews recent advancements in fairness amidst non-IID graph data, including the newly introduced fair graph generation and the commonly studied fair graph classification. In addition, available datasets and evaluation metrics for future research are identified, the limitations of existing work are highlighted, and promising future directions are proposed.</p>","PeriodicalId":7854,"journal":{"name":"Ai Magazine","volume":"46 1","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-01-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/aaai.12212","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143119936","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Beyond scaleup: Knowledge-aware parsimony learning from deep networks 超越规模化:从深度网络中进行知识感知的节俭学习
IF 2.5 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-01-28 DOI: 10.1002/aaai.12211
Quanming Yao, Yongqi Zhang, Yaqing Wang, Nan Yin, James Kwok, Qiang Yang

The brute-force scaleup of training datasets, learnable parameters and computation power, has become a prevalent strategy for developing more robust learning models. However, due to bottlenecks in data, computation, and trust, the sustainability of this strategy is a serious concern. In this paper, we attempt to address this issue in a parsimonious manner (i.e., achieving greater potential with simpler models). The key is to drive models using domain-specific knowledge, such as symbols, logic, and formulas, instead of purely relying on scaleup. This approach allows us to build a framework that uses this knowledge as “building blocks” to achieve parsimony in model design, training, and interpretation. Empirical results show that our methods surpass those that typically follow the scaling law. We also demonstrate our framework in AI for science, specifically in the problem of drug-drug interaction prediction. We hope our research can foster more diverse technical roadmaps in the era of foundation models.

训练数据集、可学习参数和计算能力的蛮力放大,已经成为开发更健壮的学习模型的普遍策略。然而,由于数据、计算和信任方面的瓶颈,这种策略的可持续性是一个严重的问题。在本文中,我们试图以一种简约的方式解决这个问题(即,用更简单的模型实现更大的潜力)。关键是使用特定于领域的知识来驱动模型,例如符号、逻辑和公式,而不是纯粹依赖于缩放。这种方法允许我们构建一个框架,使用这些知识作为“构建块”,在模型设计、训练和解释中实现简约。实证结果表明,我们的方法优于通常遵循标度定律的方法。我们还展示了我们在科学领域的人工智能框架,特别是在药物-药物相互作用预测问题上。我们希望我们的研究能够在基础模型时代培育出更多样化的技术路线图。
{"title":"Beyond scaleup: Knowledge-aware parsimony learning from deep networks","authors":"Quanming Yao,&nbsp;Yongqi Zhang,&nbsp;Yaqing Wang,&nbsp;Nan Yin,&nbsp;James Kwok,&nbsp;Qiang Yang","doi":"10.1002/aaai.12211","DOIUrl":"https://doi.org/10.1002/aaai.12211","url":null,"abstract":"<p>The brute-force scaleup of training datasets, learnable parameters and computation power, has become a prevalent strategy for developing more robust learning models. However, due to bottlenecks in data, computation, and trust, the sustainability of this strategy is a serious concern. In this paper, we attempt to address this issue in a parsimonious manner (i.e., achieving greater potential with simpler models). The key is to drive models using domain-specific knowledge, such as symbols, logic, and formulas, instead of purely relying on scaleup. This approach allows us to build a framework that uses this knowledge as “building blocks” to achieve parsimony in model design, training, and interpretation. Empirical results show that our methods surpass those that typically follow the scaling law. We also demonstrate our framework in AI for science, specifically in the problem of drug-drug interaction prediction. We hope our research can foster more diverse technical roadmaps in the era of foundation models.</p>","PeriodicalId":7854,"journal":{"name":"Ai Magazine","volume":"46 1","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-01-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/aaai.12211","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143119937","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Ai Magazine
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1