首页 > 最新文献

Ai Magazine最新文献

英文 中文
The virtual driving instructor: Multi-agent system collaborating via knowledge graph for scalable driver education 虚拟驾驶教练:基于知识图谱的多智能体系统协作,实现可扩展的驾驶员教育
IF 2.5 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-10-18 DOI: 10.1002/aaai.12201
Johannes Rehm, Irina Reshodko, Stian Zimmermann Børresen, Odd Erik Gundersen

This work introduces the design, development, and deployment of a virtual driving instructor (VDI) for enhanced driver education. The VDI provides personalized, real-time feedback to students in a driving simulator, addressing some of the limitations of traditional driver instruction. Employing a hybrid AI system, the VDI combines rule-based agents, learning-based agents, knowledge graphs, and Bayesian networks to assess and monitor student performance in a comprehensive manner. Implemented in multiple simulators at a driving school in Norway, the system aims to leverage AI and driving simulation to improve both the learning experience and the efficiency of instruction. Initial feedback from students has been largely positive, highlighting the effectiveness of this integration while also pointing to areas for further improvement. This marks a significant stride in infusing technology into driver education, offering a scalable and efficient approach to instruction.

本工作介绍了虚拟驾驶教练(VDI)的设计、开发和部署,以增强驾驶员教育。VDI在驾驶模拟器中为学生提供个性化的实时反馈,解决了传统驾驶教学的一些局限性。VDI采用混合人工智能系统,将基于规则的智能体、基于学习的智能体、知识图和贝叶斯网络相结合,以全面的方式评估和监控学生的表现。该系统在挪威一所驾校的多个模拟器中实施,旨在利用人工智能和驾驶模拟来改善学习体验和教学效率。来自学生的初步反馈基本上是积极的,突出了这种整合的有效性,同时也指出了进一步改进的领域。这标志着在将技术注入驾驶员教育方面迈出了重要的一步,提供了一种可扩展和有效的教学方法。
{"title":"The virtual driving instructor: Multi-agent system collaborating via knowledge graph for scalable driver education","authors":"Johannes Rehm,&nbsp;Irina Reshodko,&nbsp;Stian Zimmermann Børresen,&nbsp;Odd Erik Gundersen","doi":"10.1002/aaai.12201","DOIUrl":"https://doi.org/10.1002/aaai.12201","url":null,"abstract":"<p>This work introduces the design, development, and deployment of a virtual driving instructor (VDI) for enhanced driver education. The VDI provides personalized, real-time feedback to students in a driving simulator, addressing some of the limitations of traditional driver instruction. Employing a hybrid AI system, the VDI combines rule-based agents, learning-based agents, knowledge graphs, and Bayesian networks to assess and monitor student performance in a comprehensive manner. Implemented in multiple simulators at a driving school in Norway, the system aims to leverage AI and driving simulation to improve both the learning experience and the efficiency of instruction. Initial feedback from students has been largely positive, highlighting the effectiveness of this integration while also pointing to areas for further improvement. This marks a significant stride in infusing technology into driver education, offering a scalable and efficient approach to instruction.</p>","PeriodicalId":7854,"journal":{"name":"Ai Magazine","volume":"45 4","pages":"514-525"},"PeriodicalIF":2.5,"publicationDate":"2024-10-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/aaai.12201","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142851523","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Framework to enable and test conversational assistant for APIs and RPAs 框架来启用和测试api和rpa的会话助手
IF 2.5 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-10-18 DOI: 10.1002/aaai.12198
Jayachandu Bandlamudi, Kushal Mukherjee, Prerna Agarwal, Ritwik Chaudhuri, Rakesh Pimplikar, Sampath Dechu, Alex Straley, Anbumunee Ponniah, Renuka Sindhgatta

In the realm of business automation, conversational assistants are emerging as the primary method for making automation software accessible to users in various business sectors. Access to automation primarily occurs through application programming interface (APIs) and robotic process automation (RPAs). To effectively convert APIs and RPAs into chatbots on a larger scale, it is crucial to establish an automated process for generating data and training models that can recognize user intentions, identify questions for conversational slot filling, and provide recommendations for subsequent actions. In this paper, we present a technique for enhancing and generating natural language conversational artifacts from API specifications using large language models (LLMs). The goal is to utilize LLMs in the “build” phase to assist humans in creating skills for digital assistants. As a result, the system does not need to rely on LLMs during conversations with business users, leading to efficient deployment. Along with enabling digital assistants, our system employs LLMs as proxies to simulate human interaction and automatically evaluate the digital assistant's performance. Experimental results highlight the effectiveness of our proposed approach. Our system is deployed in the IBM Watson Orchestrate product for general availability.

在业务自动化领域,会话助手正在成为各种业务部门的用户访问自动化软件的主要方法。对自动化的访问主要是通过应用程序编程接口(api)和机器人过程自动化(rpa)实现的。为了有效地将api和rpa转换为更大规模的聊天机器人,建立一个自动化的过程来生成数据和训练模型是至关重要的,这些模型可以识别用户意图,识别会话槽填充的问题,并为后续行动提供建议。在本文中,我们提出了一种使用大型语言模型(llm)从API规范增强和生成自然语言会话工件的技术。目标是在“构建”阶段利用法学硕士来帮助人类为数字助理创造技能。因此,系统在与业务用户对话时不需要依赖llm,从而实现了高效的部署。随着数字助理的启用,我们的系统采用法学硕士作为代理来模拟人类互动并自动评估数字助理的表现。实验结果表明了该方法的有效性。我们的系统部署在IBM Watson Orchestrate产品中,以提供一般可用性。
{"title":"Framework to enable and test conversational assistant for APIs and RPAs","authors":"Jayachandu Bandlamudi,&nbsp;Kushal Mukherjee,&nbsp;Prerna Agarwal,&nbsp;Ritwik Chaudhuri,&nbsp;Rakesh Pimplikar,&nbsp;Sampath Dechu,&nbsp;Alex Straley,&nbsp;Anbumunee Ponniah,&nbsp;Renuka Sindhgatta","doi":"10.1002/aaai.12198","DOIUrl":"https://doi.org/10.1002/aaai.12198","url":null,"abstract":"<p>In the realm of business automation, conversational assistants are emerging as the primary method for making automation software accessible to users in various business sectors. Access to automation primarily occurs through application programming interface (APIs) and robotic process automation (RPAs). To effectively convert APIs and RPAs into chatbots on a larger scale, it is crucial to establish an automated process for generating data and training models that can recognize user intentions, identify questions for conversational slot filling, and provide recommendations for subsequent actions. In this paper, we present a technique for enhancing and generating natural language conversational artifacts from API specifications using large language models (LLMs). The goal is to utilize LLMs in the “build” phase to assist humans in creating skills for digital assistants. As a result, the system does not need to rely on LLMs during conversations with business users, leading to efficient deployment. Along with enabling digital assistants, our system employs LLMs as proxies to simulate human interaction and automatically evaluate the digital assistant's performance. Experimental results highlight the effectiveness of our proposed approach. Our system is deployed in the IBM Watson Orchestrate product for general availability.</p>","PeriodicalId":7854,"journal":{"name":"Ai Magazine","volume":"45 4","pages":"443-456"},"PeriodicalIF":2.5,"publicationDate":"2024-10-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/aaai.12198","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142851522","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Data-efficient graph learning: Problems, progress, and prospects 数据高效图学习:问题、进展和前景
IF 2.5 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-10-18 DOI: 10.1002/aaai.12200
Kaize Ding, Yixin Liu, Chuxu Zhang, Jianling Wang

Graph-structured data, ranging from social networks to financial transaction networks, from citation networks to gene regulatory networks, have been widely used for modeling a myriad of real-world systems. As a prevailing model architecture to model graph-structured data, graph neural networks (GNNs) have drawn much attention in both academic and industrial communities in the past decades. Despite their success in different graph learning tasks, existing methods usually rely on learning from “big” data, requiring a large amount of labeled data for model training. However, it is common that real-world graphs are associated with “small” labeled data as data annotation and labeling on graphs is always time and resource-consuming. Therefore, it is imperative to investigate graph machine learning (graph ML) with low-cost human supervision for low-resource settings where limited or even no labeled data is available. This paper investigates a new research field—data-efficient graph learning, which aims to push forward the performance boundary of graph ML models with different kinds of low-cost supervision signals. Specifically, we outline the fundamental research problems, review the current progress, and discuss the future prospects of data-efficient graph learning, aiming to illuminate the path for subsequent research in this field.

从社交网络到金融交易网络,从引文网络到基因调控网络,图结构数据已被广泛用于模拟现实世界中的各种系统。作为图结构数据建模的主流模型架构,图神经网络(GNN)在过去几十年中引起了学术界和工业界的广泛关注。尽管它们在不同的图学习任务中取得了成功,但现有方法通常依赖于从 "大 "数据中学习,需要大量标注数据来进行模型训练。然而,现实世界中的图通常与 "小 "标注数据相关联,因为对图进行数据注释和标注总是耗费时间和资源。因此,在资源有限甚至没有标注数据的情况下,研究具有低成本人工监督的图机器学习(graph ML)势在必行。本文探讨了一个新的研究领域--数据高效图学习,旨在通过不同类型的低成本监督信号来推动图 ML 模型的性能边界。具体而言,我们概述了数据高效图学习的基础研究问题,回顾了当前的研究进展,并讨论了其未来前景,旨在为该领域的后续研究指明方向。
{"title":"Data-efficient graph learning: Problems, progress, and prospects","authors":"Kaize Ding,&nbsp;Yixin Liu,&nbsp;Chuxu Zhang,&nbsp;Jianling Wang","doi":"10.1002/aaai.12200","DOIUrl":"https://doi.org/10.1002/aaai.12200","url":null,"abstract":"<p>Graph-structured data, ranging from social networks to financial transaction networks, from citation networks to gene regulatory networks, have been widely used for modeling a myriad of real-world systems. As a prevailing model architecture to model graph-structured data, graph neural networks (GNNs) have drawn much attention in both academic and industrial communities in the past decades. Despite their success in different graph learning tasks, existing methods usually rely on learning from “big” data, requiring a large amount of labeled data for model training. However, it is common that real-world graphs are associated with “small” labeled data as data annotation and labeling on graphs is always time and resource-consuming. Therefore, it is imperative to investigate graph machine learning (graph ML) with low-cost human supervision for low-resource settings where limited or even no labeled data is available. This paper investigates a new research field—data-efficient graph learning, which aims to push forward the performance boundary of graph ML models with different kinds of low-cost supervision signals. Specifically, we outline the fundamental research problems, review the current progress, and discuss the future prospects of data-efficient graph learning, aiming to illuminate the path for subsequent research in this field.</p>","PeriodicalId":7854,"journal":{"name":"Ai Magazine","volume":"45 4","pages":"549-560"},"PeriodicalIF":2.5,"publicationDate":"2024-10-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/aaai.12200","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142861732","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
DCV 2 I $text{DCV}^2text{I}$ : Leveraging deep vision models to support geographers' visual interpretation in dune segmentation DCV 2 I $text{DCV}^2text{I}$:利用深度视觉模型支持地理学家在沙丘分割中的视觉解释
IF 2.5 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-10-18 DOI: 10.1002/aaai.12199
Anqi Lu, Zifeng Wu, Zheng Jiang, Wei Wang, Eerdun Hasi, Yi Wang

Visual interpretation is extremely important in human geography as the primary technique for geographers to use photograph data in identifying, classifying, and quantifying geographic and topological objects or regions. However, it is also time-consuming and requires overwhelming manual effort from professional geographers. This paper describes our interdisciplinary team's efforts in integrating computer vision models with geographers' visual image interpretation process to reduce their workload in interpreting images. Focusing on the dune segmentation task, we proposed an approach called DCV2I${bf DCV}^2{bf I}$ featuring a deep dune segmentation model to identify dunes and label their ranges in an automated way. By developing a tool to connect our model with ArcGIS—one of the most popular workbenches for visual interpretation, geographers can further refine the automatically generated dune segmentation on images without learning any CV or deep learning techniques. Our approach thus realized a noninvasive change to geographers' visual interpretation routines, reducing their manual efforts while incurring minimal interruptions to their work routines and tools they are familiar with. Deployment with a leading Chinese geography research institution demonstrated the potential of DCV2I${bf DCV}^2{bf I}$ in supporting geographers in researching and solving drylands desertification.

目视判读在人文地理学中极为重要,是地理学家利用照片数据识别、分类和量化地理和地形对象或区域的主要技术。然而,这也非常耗时,需要专业地理学家付出大量的人工努力。本文介绍了我们的跨学科团队如何将计算机视觉模型与地理学家的视觉图像判读过程相结合,以减轻他们判读图像的工作量。针对沙丘分割任务,我们提出了一种名为 DCV 2 I ${bf DCV}^2{bf I}$ 的方法,其特点是采用深度沙丘分割模型来自动识别沙丘并标注其范围。通过开发一种工具,将我们的模型与 ArcGIS(最流行的可视化解释工作台之一)连接起来,地理学家无需学习任何 CV 或深度学习技术,就能进一步完善自动生成的沙丘分割图像。因此,我们的方法实现了对地理学家可视化判读例程的非侵入式改变,减少了他们的手工操作,同时对他们的工作例程和熟悉的工具产生了最小的干扰。在中国领先的地理研究机构的部署表明,DCV 2 I ${bf DCV}^2{bf I}$在支持地理学家研究和解决旱地荒漠化问题方面具有潜力。
{"title":"DCV\u0000 2\u0000 \u0000 I\u0000 \u0000 $text{DCV}^2text{I}$\u0000 : Leveraging deep vision models to support geographers' visual interpretation in dune segmentation","authors":"Anqi Lu,&nbsp;Zifeng Wu,&nbsp;Zheng Jiang,&nbsp;Wei Wang,&nbsp;Eerdun Hasi,&nbsp;Yi Wang","doi":"10.1002/aaai.12199","DOIUrl":"https://doi.org/10.1002/aaai.12199","url":null,"abstract":"<p>Visual interpretation is extremely important in human geography as the primary technique for geographers to use photograph data in identifying, classifying, and quantifying geographic and topological objects or regions. However, it is also time-consuming and requires overwhelming manual effort from professional geographers. This paper describes our interdisciplinary team's efforts in integrating computer vision models with geographers' visual image interpretation process to reduce their workload in interpreting images. Focusing on the dune segmentation task, we proposed an approach called <span></span><math>\u0000 <semantics>\u0000 <mrow>\u0000 <msup>\u0000 <mi>DCV</mi>\u0000 <mn>2</mn>\u0000 </msup>\u0000 <mi>I</mi>\u0000 </mrow>\u0000 <annotation>${bf DCV}^2{bf I}$</annotation>\u0000 </semantics></math> featuring a deep dune segmentation model to identify dunes and label their ranges in an automated way. By developing a tool to connect our model with ArcGIS—one of the most popular workbenches for visual interpretation, geographers can further refine the automatically generated dune segmentation on images without learning any CV or deep learning techniques. Our approach thus realized a noninvasive change to geographers' visual interpretation routines, reducing their manual efforts while incurring minimal interruptions to their work routines and tools they are familiar with. Deployment with a leading Chinese geography research institution demonstrated the potential of <span></span><math>\u0000 <semantics>\u0000 <mrow>\u0000 <msup>\u0000 <mi>DCV</mi>\u0000 <mn>2</mn>\u0000 </msup>\u0000 <mi>I</mi>\u0000 </mrow>\u0000 <annotation>${bf DCV}^2{bf I}$</annotation>\u0000 </semantics></math> in supporting geographers in researching and solving drylands desertification.</p>","PeriodicalId":7854,"journal":{"name":"Ai Magazine","volume":"45 4","pages":"472-485"},"PeriodicalIF":2.5,"publicationDate":"2024-10-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/aaai.12199","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142861675","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A submodular optimization approach to trustworthy loan approval automation 可信赖贷款审批自动化的子模块优化方法
IF 2.5 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-10-01 DOI: 10.1002/aaai.12195
Kyungsik Lee, Hana Yoo, Sumin Shin, Wooyoung Kim, Yeonung Baek, Hyunjin Kang, Jaehyun Kim, Kee-Eung Kim

In the field of finance, the underwriting process is an essential step in evaluating every loan application. During this stage, the borrowers' creditworthiness and ability to repay the loan are assessed to ultimately decide whether to approve the loan application. One of the core components of underwriting is credit scoring, in which the probability of default is estimated. As such, there has been significant progress in enhancing the predictive accuracy of credit scoring models through the use of machine learning, but there still exists a need to ultimately construct an approval rule that takes into consideration additional criteria beyond the score itself. This construction process is traditionally done manually to ensure that the approval rule remains interpretable to humans. In this paper, we outline an automated system for optimizing a rule-based system for approving loan applications, which has been deployed at Hyundai Capital Services (HCS). The main challenge lays in creating a high-quality rule base that is simultaneously simple enough to be interpretable by risk analysts as well as customers, since the approval decision should be easily understandable. We addressed this challenge through principled submodular optimization. The deployment of our system has led to a 14% annual growth in the volume of loan services at HCS, while maintaining the target bad rate, and has resulted in the approval of customers who might have otherwise been rejected.

在金融领域,承销过程是评估每一笔贷款申请的重要步骤。在此阶段,评估借款人的信誉和偿还贷款的能力,以最终决定是否批准贷款申请。承保的核心组成部分之一是信用评分,其中估计了违约的可能性。因此,通过使用机器学习,在提高信用评分模型的预测准确性方面已经取得了重大进展,但仍然需要最终构建一个考虑分数本身之外的其他标准的审批规则。这个构造过程传统上是手动完成的,以确保审批规则仍然对人类是可解释的。在本文中,我们概述了一个自动化系统,用于优化基于规则的系统,以批准贷款申请,该系统已部署在现代资本服务公司(HCS)。主要的挑战在于创建一个高质量的规则基础,它同时足够简单,可以被风险分析师和客户解释,因为批准决定应该很容易理解。我们通过原则性的子模块优化解决了这一挑战。我们系统的部署使HCS的贷款业务量每年增长14%,同时保持了目标不良率,并获得了可能被拒绝的客户的认可。
{"title":"A submodular optimization approach to trustworthy loan approval automation","authors":"Kyungsik Lee,&nbsp;Hana Yoo,&nbsp;Sumin Shin,&nbsp;Wooyoung Kim,&nbsp;Yeonung Baek,&nbsp;Hyunjin Kang,&nbsp;Jaehyun Kim,&nbsp;Kee-Eung Kim","doi":"10.1002/aaai.12195","DOIUrl":"https://doi.org/10.1002/aaai.12195","url":null,"abstract":"<p>In the field of finance, the underwriting process is an essential step in evaluating every loan application. During this stage, the borrowers' creditworthiness and ability to repay the loan are assessed to ultimately decide whether to approve the loan application. One of the core components of underwriting is credit scoring, in which the probability of default is estimated. As such, there has been significant progress in enhancing the predictive accuracy of credit scoring models through the use of machine learning, but there still exists a need to ultimately construct an approval rule that takes into consideration additional criteria beyond the score itself. This construction process is traditionally done manually to ensure that the approval rule remains interpretable to humans. In this paper, we outline an automated system for optimizing a rule-based system for approving loan applications, which has been deployed at Hyundai Capital Services (HCS). The main challenge lays in creating a high-quality rule base that is simultaneously simple enough to be interpretable by risk analysts as well as customers, since the approval decision should be easily understandable. We addressed this challenge through principled submodular optimization. The deployment of our system has led to a 14% annual growth in the volume of loan services at HCS, while maintaining the target bad rate, and has resulted in the approval of customers who might have otherwise been rejected.</p>","PeriodicalId":7854,"journal":{"name":"Ai Magazine","volume":"45 4","pages":"502-513"},"PeriodicalIF":2.5,"publicationDate":"2024-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/aaai.12195","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142859946","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Effective knowledge representation and utilization for sustainable collaborative learning across heterogeneous systems 有效的知识表示和利用,促进跨异构系统的可持续协作学习
IF 2.5 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-09-22 DOI: 10.1002/aaai.12193
Trong Nghia Hoang

The increasingly decentralized and private nature of data in our digital society has motivated the development of collaborative intelligent systems that enable knowledge aggregation among data owners. However, collaborative learning has only been investigated in simple settings. For example, clients are often assumed to train solution models de novo, disregarding all prior expertise. The learned model is typically represented in task-specific forms that are not generalizable to unseen, emerging scenarios. Finally, a universal model representation is enforced among collaborators, ignoring their local compute constraints or input representations. These limitations hampers the practicality of prior collaborative systems in learning scenarios with limited task data that demand constant knowledge adaptation and transfer across information silos, tasks, and learning models, as well as the utilization of prior solution expertise. Furthermore, prior collaborative learning frameworks are not sustainable on a macro scale where participants desire fairness allocation of benefits (e.g., access to the combined model) based on their costs of participation (e.g., overhead of model sharing and training synchronization, risk of information breaches, etc.). This necessitates a new perspective of collaborative learning where the server not only aggregates but also conducts valuation of the participant's contribution, and distribute aggregated information to individuals in commensurate to their contribution. To substantiate the above vision, we propose a new research agenda on developing effective and sustainable collaborative learning frameworks across heterogeneous systems, featuring three novel computational capabilities on knowledge organization: model expression, comprehension, and valuation.

在我们的数字社会中,数据的分散性和私密性越来越强,这促使人们开发能够在数据所有者之间聚合知识的协作智能系统。然而,人们只在简单的环境中研究过协作学习。例如,通常假设客户从头开始训练解决方案模型,而不考虑所有先前的专业知识。学习到的模型通常以特定任务的形式表示,无法推广到未见过的新兴场景中。最后,合作者之间强制使用通用模型表示法,忽略了他们的本地计算约束或输入表示法。这些局限性妨碍了先前的协作系统在任务数据有限的学习场景中的实用性,因为这种场景需要在信息孤岛、任务和学习模型之间不断进行知识调整和转移,并需要利用先前的解决方案专长。此外,先前的协作学习框架在宏观上是不可持续的,因为参与者希望根据他们的参与成本(如模型共享和培训同步的开销、信息泄露的风险等)公平分配利益(如访问组合模型)。这就需要一种新的协作学习视角,即服务器不仅要汇总信息,还要对参与者的贡献进行评估,并将汇总信息按贡献分配给个人。为了证实上述愿景,我们提出了一个新的研究议程,即在异构系统中开发有效、可持续的协作学习框架,其中包括三种新的知识组织计算能力:模型表达、理解和评估。
{"title":"Effective knowledge representation and utilization for sustainable collaborative learning across heterogeneous systems","authors":"Trong Nghia Hoang","doi":"10.1002/aaai.12193","DOIUrl":"https://doi.org/10.1002/aaai.12193","url":null,"abstract":"<p>The increasingly decentralized and private nature of data in our digital society has motivated the development of collaborative intelligent systems that enable knowledge aggregation among data owners. However, collaborative learning has only been investigated in simple settings. For example, clients are often assumed to train solution models <i>de novo</i>, disregarding all prior expertise. The learned model is typically represented in task-specific forms that are not generalizable to unseen, emerging scenarios. Finally, a universal model representation is enforced among collaborators, ignoring their local compute constraints or input representations. These limitations hampers the practicality of prior collaborative systems in learning scenarios with limited task data that demand constant knowledge adaptation and transfer across information silos, tasks, and learning models, as well as the utilization of prior solution expertise. Furthermore, prior collaborative learning frameworks are not sustainable on a macro scale where participants desire fairness allocation of benefits (e.g., access to the combined model) based on their costs of participation (e.g., overhead of model sharing and training synchronization, risk of information breaches, etc.). This necessitates a new perspective of collaborative learning where the server not only aggregates but also conducts valuation of the participant's contribution, and distribute aggregated information to individuals in commensurate to their contribution. To substantiate the above vision, we propose a new research agenda on developing effective and sustainable collaborative learning frameworks across heterogeneous systems, featuring three novel computational capabilities on knowledge organization: model expression, comprehension, and valuation.</p>","PeriodicalId":7854,"journal":{"name":"Ai Magazine","volume":"45 3","pages":"404-410"},"PeriodicalIF":2.5,"publicationDate":"2024-09-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/aaai.12193","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142324480","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Fair and optimal prediction via post-processing 通过后处理实现公平和优化预测
IF 2.5 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-09-22 DOI: 10.1002/aaai.12191
Han Zhao

With the development of machine learning algorithms and the increasing computational resources available, artificial intelligence has achieved great success in many application domains. However, the success of machine learning has also raised concerns about the fairness of the learned models. For instance, the learned models can perpetuate and even exacerbate the potential bias and discrimination in the training data. This issue has become a major obstacle to the deployment of machine learning systems in high-stakes domains, for example, criminal judgment, medical testing, online advertising, hiring process, and so forth. To mitigate the potential bias exhibited by machine learning models, fairness criteria can be integrated into the training process to ensure fair treatment across all demographics, but it often comes at the expense of model performance. Understanding such tradeoffs, therefore, is crucial to the design of optimal and fair algorithms. My research focuses on characterizing the inherent tradeoff between fairness and accuracy in machine learning, and developing algorithms that can achieve both fairness and optimality. In this article, I will discuss our recent work on designing post-processing algorithms for fair classification, which can be applied to a wide range of fairness criteria, including statistical parity, equal opportunity, and equalized odds, under both attribute-aware and attribute-blind settings, and is particularly suited to large-scale foundation models where retraining is expensive or even infeasible. I will also discuss the connections between our work and other related research on trustworthy machine learning, including the connections between algorithmic fairness and differential privacy as well as adversarial robustness.

随着机器学习算法的发展和可用计算资源的不断增加,人工智能在许多应用领域取得了巨大成功。然而,机器学习的成功也引发了人们对所学模型公平性的担忧。例如,学习到的模型可能会延续甚至加剧训练数据中潜在的偏见和歧视。这一问题已成为机器学习系统在刑事判决、医疗测试、在线广告、招聘流程等高风险领域部署的主要障碍。为了减轻机器学习模型可能表现出的偏差,可以将公平性标准集成到训练过程中,以确保公平对待所有人群,但这往往以牺牲模型性能为代价。因此,了解这种权衡对于设计最佳公平算法至关重要。我的研究重点是描述机器学习中公平性和准确性之间固有的权衡,并开发能同时实现公平性和最优性的算法。在本文中,我将讨论我们最近在设计公平分类的后处理算法方面所做的工作,该算法可在属性感知和属性盲设置下应用于广泛的公平标准,包括统计均等、机会均等和赔率均等,尤其适用于重新训练成本高昂甚至不可行的大规模基础模型。我还将讨论我们的工作与其他可信机器学习相关研究之间的联系,包括算法公平性与差异隐私以及对抗鲁棒性之间的联系。
{"title":"Fair and optimal prediction via post-processing","authors":"Han Zhao","doi":"10.1002/aaai.12191","DOIUrl":"https://doi.org/10.1002/aaai.12191","url":null,"abstract":"<p>With the development of machine learning algorithms and the increasing computational resources available, artificial intelligence has achieved great success in many application domains. However, the success of machine learning has also raised concerns about the <i>fairness</i> of the learned models. For instance, the learned models can perpetuate and even exacerbate the potential bias and discrimination in the training data. This issue has become a major obstacle to the deployment of machine learning systems in high-stakes domains, for example, criminal judgment, medical testing, online advertising, hiring process, and so forth. To mitigate the potential bias exhibited by machine learning models, fairness criteria can be integrated into the training process to ensure fair treatment across all demographics, but it often comes at the expense of model performance. Understanding such tradeoffs, therefore, is crucial to the design of optimal and fair algorithms. My research focuses on characterizing the inherent tradeoff between fairness and accuracy in machine learning, and developing algorithms that can achieve both fairness and optimality. In this article, I will discuss our recent work on designing post-processing algorithms for fair classification, which can be applied to a wide range of fairness criteria, including statistical parity, equal opportunity, and equalized odds, under both attribute-aware and attribute-blind settings, and is particularly suited to large-scale foundation models where retraining is expensive or even infeasible. I will also discuss the connections between our work and other related research on trustworthy machine learning, including the connections between algorithmic fairness and differential privacy as well as adversarial robustness.</p>","PeriodicalId":7854,"journal":{"name":"Ai Magazine","volume":"45 3","pages":"411-418"},"PeriodicalIF":2.5,"publicationDate":"2024-09-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/aaai.12191","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142324481","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Efficient and robust sequential decision making algorithms 高效稳健的顺序决策算法
IF 2.5 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-09-22 DOI: 10.1002/aaai.12186
Pan Xu

Sequential decision-making involves making informed decisions based on continuous interactions with a complex environment. This process is ubiquitous in various applications, including recommendation systems and clinical treatment design. My research has concentrated on addressing two pivotal challenges in sequential decision-making: (1) How can we design algorithms that efficiently learn the optimal decision strategy with minimal interactions and limited sample data? (2) How can we ensure robustness in decision-making algorithms when faced with distributional shifts due to environmental changes and the sim-to-real gap? This paper summarizes and expands upon the talk I presented at the AAAI 2024 New Faculty Highlights program, detailing how my research aims to tackle these challenges.

顺序决策涉及根据与复杂环境的持续互动做出明智的决定。这一过程在各种应用中无处不在,包括推荐系统和临床治疗设计。我的研究集中于解决顺序决策中的两个关键挑战:(1) 我们如何设计算法,在最小的交互和有限的样本数据中高效地学习最优决策策略?(2) 面对环境变化和模拟与实际差距造成的分布变化,我们如何确保决策算法的稳健性?本文总结并扩展了我在 AAAI 2024 新教师亮点计划中的演讲,详细介绍了我的研究如何应对这些挑战。
{"title":"Efficient and robust sequential decision making algorithms","authors":"Pan Xu","doi":"10.1002/aaai.12186","DOIUrl":"https://doi.org/10.1002/aaai.12186","url":null,"abstract":"<p>Sequential decision-making involves making informed decisions based on continuous interactions with a complex environment. This process is ubiquitous in various applications, including recommendation systems and clinical treatment design. My research has concentrated on addressing two pivotal challenges in sequential decision-making: (1) How can we design algorithms that efficiently learn the optimal decision strategy with minimal interactions and limited sample data? (2) How can we ensure robustness in decision-making algorithms when faced with distributional shifts due to environmental changes and the sim-to-real gap? This paper summarizes and expands upon the talk I presented at the AAAI 2024 New Faculty Highlights program, detailing how my research aims to tackle these challenges.</p>","PeriodicalId":7854,"journal":{"name":"Ai Magazine","volume":"45 3","pages":"376-385"},"PeriodicalIF":2.5,"publicationDate":"2024-09-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/aaai.12186","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142324470","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
AI fairness in practice: Paradigm, challenges, and prospects 实践中的人工智能公平性:范式、挑战和前景
IF 2.5 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-09-22 DOI: 10.1002/aaai.12189
Wenbin Zhang

Understanding and correcting algorithmic bias in artificial intelligence (AI) has become increasingly important, leading to a surge in research on AI fairness within both the AI community and broader society. Traditionally, this research operates within the constrained supervised learning paradigm, assuming the presence of class labels, independent and identically distributed (IID) data, and batch-based learning necessitating the simultaneous availability of all training data. However, in practice, class labels may be absent due to censoring, data is often represented using non-IID graph structures that capture connections among individual units, and data can arrive and evolve over time. These prevalent real-world data representations limit the applicability of existing fairness literature, which typically addresses fairness in static and tabular supervised learning settings. This paper reviews recent advances in AI fairness aimed at bridging these gaps for practical deployment in real-world scenarios. Additionally, opportunities are envisioned by highlighting the limitations and significant potential for real applications.

了解和纠正人工智能(AI)中的算法偏差已变得越来越重要,这导致人工智能界和更广泛的社会对人工智能公平性的研究激增。传统上,这项研究是在受限的监督学习范式下进行的,假定存在类标签、独立且同分布(IID)的数据,以及基于批量的学习(必须同时提供所有训练数据)。然而,在实践中,由于删减的原因,类标签可能不存在,数据通常使用非独立且同分布(IID)的图结构来表示,以捕捉单个单元之间的联系,而且数据可能随着时间的推移而到达和演变。这些现实世界中普遍存在的数据表示方式限制了现有公平性文献的适用性,因为这些文献通常涉及静态和表格监督学习环境中的公平性问题。本文回顾了人工智能公平性方面的最新进展,旨在缩小这些差距,以便在现实世界场景中进行实际部署。此外,本文还通过强调实际应用的局限性和巨大潜力,展望了各种机遇。
{"title":"AI fairness in practice: Paradigm, challenges, and prospects","authors":"Wenbin Zhang","doi":"10.1002/aaai.12189","DOIUrl":"https://doi.org/10.1002/aaai.12189","url":null,"abstract":"<p>Understanding and correcting algorithmic bias in artificial intelligence (AI) has become increasingly important, leading to a surge in research on AI fairness within both the AI community and broader society. Traditionally, this research operates within the constrained supervised learning paradigm, assuming the presence of class labels, independent and identically distributed (IID) data, and batch-based learning necessitating the simultaneous availability of all training data. However, in practice, class labels may be absent due to censoring, data is often represented using non-IID graph structures that capture connections among individual units, and data can arrive and evolve over time. These prevalent real-world data representations limit the applicability of existing fairness literature, which typically addresses fairness in static and tabular supervised learning settings. This paper reviews recent advances in AI fairness aimed at bridging these gaps for practical deployment in real-world scenarios. Additionally, opportunities are envisioned by highlighting the limitations and significant potential for real applications.</p>","PeriodicalId":7854,"journal":{"name":"Ai Magazine","volume":"45 3","pages":"386-395"},"PeriodicalIF":2.5,"publicationDate":"2024-09-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/aaai.12189","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142324460","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Towards smooth mobile robot deployments in dynamic human environments 在动态人类环境中顺利部署移动机器人
IF 2.5 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-09-22 DOI: 10.1002/aaai.12192
Christoforos Mavrogiannis

Recently, there has been great interest in deploying autonomous mobile robots in airports, malls, and hospitals to complete a range of tasks such as delivery, cleaning, and patrolling. The rich context of these environments gives rise to highly unstructured motion that is challenging for robots to anticipate and adapt to. This results in uncomfortable and unsafe human–robot encounters, poor robot performance, and even catastrophic failures that hinder robot acceptance. Such observations have motivated my work on social robot navigation, the problem of enabling robots to navigate in human environments while accounting for human safety and comfort. In this article, I highlight prior work on expanding the classical autonomy stack with mathematical models and algorithms designed to contribute towards smoother mobile robot deployments in complex environments.

最近,人们对在机场、商场和医院部署自主移动机器人以完成一系列任务(如送货、清洁和巡逻)产生了浓厚的兴趣。这些环境的丰富背景导致了高度非结构化的运动,对机器人的预测和适应能力提出了挑战。这就导致了不舒适和不安全的人机交互、糟糕的机器人性能,甚至是阻碍机器人被接受的灾难性故障。这些观察结果激发了我在社交机器人导航方面的研究,即如何让机器人在人类环境中导航,同时考虑到人类的安全和舒适。在这篇文章中,我将重点介绍之前的工作,即利用数学模型和算法扩展经典的自主性堆栈,以促进移动机器人在复杂环境中更顺利地部署。
{"title":"Towards smooth mobile robot deployments in dynamic human environments","authors":"Christoforos Mavrogiannis","doi":"10.1002/aaai.12192","DOIUrl":"https://doi.org/10.1002/aaai.12192","url":null,"abstract":"<p>Recently, there has been great interest in deploying autonomous mobile robots in airports, malls, and hospitals to complete a range of tasks such as delivery, cleaning, and patrolling. The rich context of these environments gives rise to highly unstructured motion that is challenging for robots to anticipate and adapt to. This results in uncomfortable and unsafe human–robot encounters, poor robot performance, and even catastrophic failures that hinder robot acceptance. Such observations have motivated my work on social robot navigation, the problem of enabling robots to navigate in human environments while accounting for human safety and comfort. In this article, I highlight prior work on expanding the classical autonomy stack with mathematical models and algorithms designed to contribute towards smoother mobile robot deployments in complex environments.</p>","PeriodicalId":7854,"journal":{"name":"Ai Magazine","volume":"45 3","pages":"419-428"},"PeriodicalIF":2.5,"publicationDate":"2024-09-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/aaai.12192","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142324482","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Ai Magazine
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1