首页 > 最新文献

Ai Magazine最新文献

英文 中文
Learning representations for robust human–robot interaction 鲁棒人机交互的学习表征
IF 2.5 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-10-21 DOI: 10.1002/aaai.12197
Yen-Ling Kuo

This article summarizes the author's presentation in the New Faculty Highlight at the Thirty-Eighth AAAI Conference on Artificial Intelligence. It discusses the desired properties of representations for enabling robust human–robot interaction. Examples from the author's work are presented to show how to build these properties into models for performing tasks with natural language guidance and engaging in social interactions with other agents.

本文总结了作者在第38届美国人工智能协会(AAAI)人工智能会议“新教员亮点”上的演讲。它讨论了实现鲁棒人机交互所需的表征属性。本文给出了作者工作中的示例,以展示如何将这些属性构建到模型中,以便在自然语言指导下执行任务,并与其他代理进行社会互动。
{"title":"Learning representations for robust human–robot interaction","authors":"Yen-Ling Kuo","doi":"10.1002/aaai.12197","DOIUrl":"https://doi.org/10.1002/aaai.12197","url":null,"abstract":"<p>This article summarizes the author's presentation in the New Faculty Highlight at the Thirty-Eighth AAAI Conference on Artificial Intelligence. It discusses the desired properties of representations for enabling robust human–robot interaction. Examples from the author's work are presented to show how to build these properties into models for performing tasks with natural language guidance and engaging in social interactions with other agents.</p>","PeriodicalId":7854,"journal":{"name":"Ai Magazine","volume":"45 4","pages":"561-568"},"PeriodicalIF":2.5,"publicationDate":"2024-10-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/aaai.12197","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142851503","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Fusing remote and social sensing data for flood impact mapping 融合遥感和社会遥感数据进行洪水影响制图
IF 2.5 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-10-18 DOI: 10.1002/aaai.12196
Zainab Akhtar, Umair Qazi, Aya El-Sakka, Rizwan Sadiq, Ferda Ofli, Muhammad Imran

The absence of comprehensive situational awareness information poses a significant challenge for humanitarian organizations during their response efforts. We present Flood Insights, an end-to-end system, that ingests data from multiple nontraditional data sources such as remote sensing, social sensing, and geospatial data. We employ state-of-the-art natural language processing and computer vision models to identify flood exposure, ground-level damage and flood reports, and most importantly, urgent needs of affected people. We deploy and test the system during a recent real-world catastrophe, the 2022 Pakistan floods, to surface critical situational and damage information at the district level. We validated the system's effectiveness through various statistical analyses using official ground-truth data, showcasing its strong performance and explanatory power of integrating multiple data sources. Moreover, the system was commended by the United Nations Development Programme stationed in Pakistan, as well as local authorities, for pinpointing hard-hit districts and enhancing disaster response.

缺乏全面的态势感知信息给人道主义组织的救灾工作带来了巨大挑战。我们介绍的 "洪水洞察 "是一个端到端系统,可从遥感、社会感应和地理空间数据等多个非传统数据源获取数据。我们采用最先进的自然语言处理和计算机视觉模型来识别洪水风险、地面损失和洪水报告,最重要的是识别受灾人口的迫切需求。我们在 2022 年巴基斯坦洪灾这一最近发生的实际灾难中部署并测试了该系统,以显示地区一级的重要情况和损失信息。我们通过使用官方地面实况数据进行各种统计分析,验证了该系统的有效性,展示了其整合多种数据源的强大性能和解释能力。此外,该系统还受到了联合国开发计划署驻巴基斯坦办事处和地方当局的赞扬,因为它准确定位了重灾区,增强了救灾能力。
{"title":"Fusing remote and social sensing data for flood impact mapping","authors":"Zainab Akhtar,&nbsp;Umair Qazi,&nbsp;Aya El-Sakka,&nbsp;Rizwan Sadiq,&nbsp;Ferda Ofli,&nbsp;Muhammad Imran","doi":"10.1002/aaai.12196","DOIUrl":"https://doi.org/10.1002/aaai.12196","url":null,"abstract":"<p>The absence of comprehensive situational awareness information poses a significant challenge for humanitarian organizations during their response efforts. We present Flood Insights, an end-to-end system, that ingests data from multiple nontraditional data sources such as remote sensing, social sensing, and geospatial data. We employ state-of-the-art natural language processing and computer vision models to identify flood exposure, ground-level damage and flood reports, and most importantly, urgent needs of affected people. We deploy and test the system during a recent real-world catastrophe, the 2022 Pakistan floods, to surface critical situational and damage information at the district level. We validated the system's effectiveness through various statistical analyses using official ground-truth data, showcasing its strong performance and explanatory power of integrating multiple data sources. Moreover, the system was commended by the United Nations Development Programme stationed in Pakistan, as well as local authorities, for pinpointing hard-hit districts and enhancing disaster response.</p>","PeriodicalId":7854,"journal":{"name":"Ai Magazine","volume":"45 4","pages":"486-501"},"PeriodicalIF":2.5,"publicationDate":"2024-10-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/aaai.12196","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142861676","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A survey of out-of-distribution generalization for graph machine learning from a causal view 基于因果关系的图机器学习分布外泛化研究
IF 2.5 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-10-18 DOI: 10.1002/aaai.12202
Jing Ma

Graph machine learning (GML) has been successfully applied across a wide range of tasks. Nonetheless, GML faces significant challenges in generalizing over out-of-distribution (OOD) data, which raises concerns about its wider applicability. Recent advancements have underscored the crucial role of causality-driven approaches in overcoming these generalization challenges. Distinct from traditional GML methods that primarily rely on statistical dependencies, causality-focused strategies delve into the underlying causal mechanisms of data generation and model prediction, thus significantly improving the generalization of GML across different environments. This paper offers a thorough review of recent progress in causality-involved GML generalization. We elucidate the fundamental concepts of employing causality to enhance graph model generalization and categorize the various approaches, providing detailed descriptions of their methodologies and the connections among them. Furthermore, we explore the incorporation of causality in other related important areas of trustworthy GML, such as explanation, fairness, and robustness. Concluding with a discussion on potential future research directions, this review seeks to articulate the continuing development and future potential of causality in enhancing the trustworthiness of GML.

图机器学习(GML)已经成功地应用于广泛的任务中。尽管如此,GML在泛化分布外(out- distribution, OOD)数据方面面临着重大挑战,这引发了对其更广泛适用性的担忧。最近的进展强调了因果关系驱动的方法在克服这些泛化挑战中的关键作用。与主要依赖统计依赖的传统GML方法不同,以因果关系为中心的策略深入研究了数据生成和模型预测的潜在因果机制,从而显著提高了GML在不同环境中的泛化能力。本文提供了一个全面的综述最近进展的因果关系涉及GML泛化。我们阐明了利用因果关系来增强图模型泛化的基本概念,并对各种方法进行了分类,详细描述了它们的方法和它们之间的联系。此外,我们探讨了可信赖GML的其他相关重要领域中因果关系的结合,如解释、公平性和鲁棒性。最后讨论了未来潜在的研究方向,本文旨在阐明因果关系在提高GML可信度方面的持续发展和未来潜力。
{"title":"A survey of out-of-distribution generalization for graph machine learning from a causal view","authors":"Jing Ma","doi":"10.1002/aaai.12202","DOIUrl":"https://doi.org/10.1002/aaai.12202","url":null,"abstract":"<p>Graph machine learning (GML) has been successfully applied across a wide range of tasks. Nonetheless, GML faces significant challenges in generalizing over out-of-distribution (OOD) data, which raises concerns about its wider applicability. Recent advancements have underscored the crucial role of causality-driven approaches in overcoming these generalization challenges. Distinct from traditional GML methods that primarily rely on statistical dependencies, causality-focused strategies delve into the underlying causal mechanisms of data generation and model prediction, thus significantly improving the generalization of GML across different environments. This paper offers a thorough review of recent progress in causality-involved GML generalization. We elucidate the fundamental concepts of employing causality to enhance graph model generalization and categorize the various approaches, providing detailed descriptions of their methodologies and the connections among them. Furthermore, we explore the incorporation of causality in other related important areas of trustworthy GML, such as explanation, fairness, and robustness. Concluding with a discussion on potential future research directions, this review seeks to articulate the continuing development and future potential of causality in enhancing the trustworthiness of GML.</p>","PeriodicalId":7854,"journal":{"name":"Ai Magazine","volume":"45 4","pages":"537-548"},"PeriodicalIF":2.5,"publicationDate":"2024-10-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/aaai.12202","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142851524","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
The virtual driving instructor: Multi-agent system collaborating via knowledge graph for scalable driver education 虚拟驾驶教练:基于知识图谱的多智能体系统协作,实现可扩展的驾驶员教育
IF 2.5 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-10-18 DOI: 10.1002/aaai.12201
Johannes Rehm, Irina Reshodko, Stian Zimmermann Børresen, Odd Erik Gundersen

This work introduces the design, development, and deployment of a virtual driving instructor (VDI) for enhanced driver education. The VDI provides personalized, real-time feedback to students in a driving simulator, addressing some of the limitations of traditional driver instruction. Employing a hybrid AI system, the VDI combines rule-based agents, learning-based agents, knowledge graphs, and Bayesian networks to assess and monitor student performance in a comprehensive manner. Implemented in multiple simulators at a driving school in Norway, the system aims to leverage AI and driving simulation to improve both the learning experience and the efficiency of instruction. Initial feedback from students has been largely positive, highlighting the effectiveness of this integration while also pointing to areas for further improvement. This marks a significant stride in infusing technology into driver education, offering a scalable and efficient approach to instruction.

本工作介绍了虚拟驾驶教练(VDI)的设计、开发和部署,以增强驾驶员教育。VDI在驾驶模拟器中为学生提供个性化的实时反馈,解决了传统驾驶教学的一些局限性。VDI采用混合人工智能系统,将基于规则的智能体、基于学习的智能体、知识图和贝叶斯网络相结合,以全面的方式评估和监控学生的表现。该系统在挪威一所驾校的多个模拟器中实施,旨在利用人工智能和驾驶模拟来改善学习体验和教学效率。来自学生的初步反馈基本上是积极的,突出了这种整合的有效性,同时也指出了进一步改进的领域。这标志着在将技术注入驾驶员教育方面迈出了重要的一步,提供了一种可扩展和有效的教学方法。
{"title":"The virtual driving instructor: Multi-agent system collaborating via knowledge graph for scalable driver education","authors":"Johannes Rehm,&nbsp;Irina Reshodko,&nbsp;Stian Zimmermann Børresen,&nbsp;Odd Erik Gundersen","doi":"10.1002/aaai.12201","DOIUrl":"https://doi.org/10.1002/aaai.12201","url":null,"abstract":"<p>This work introduces the design, development, and deployment of a virtual driving instructor (VDI) for enhanced driver education. The VDI provides personalized, real-time feedback to students in a driving simulator, addressing some of the limitations of traditional driver instruction. Employing a hybrid AI system, the VDI combines rule-based agents, learning-based agents, knowledge graphs, and Bayesian networks to assess and monitor student performance in a comprehensive manner. Implemented in multiple simulators at a driving school in Norway, the system aims to leverage AI and driving simulation to improve both the learning experience and the efficiency of instruction. Initial feedback from students has been largely positive, highlighting the effectiveness of this integration while also pointing to areas for further improvement. This marks a significant stride in infusing technology into driver education, offering a scalable and efficient approach to instruction.</p>","PeriodicalId":7854,"journal":{"name":"Ai Magazine","volume":"45 4","pages":"514-525"},"PeriodicalIF":2.5,"publicationDate":"2024-10-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/aaai.12201","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142851523","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Framework to enable and test conversational assistant for APIs and RPAs 框架来启用和测试api和rpa的会话助手
IF 2.5 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-10-18 DOI: 10.1002/aaai.12198
Jayachandu Bandlamudi, Kushal Mukherjee, Prerna Agarwal, Ritwik Chaudhuri, Rakesh Pimplikar, Sampath Dechu, Alex Straley, Anbumunee Ponniah, Renuka Sindhgatta

In the realm of business automation, conversational assistants are emerging as the primary method for making automation software accessible to users in various business sectors. Access to automation primarily occurs through application programming interface (APIs) and robotic process automation (RPAs). To effectively convert APIs and RPAs into chatbots on a larger scale, it is crucial to establish an automated process for generating data and training models that can recognize user intentions, identify questions for conversational slot filling, and provide recommendations for subsequent actions. In this paper, we present a technique for enhancing and generating natural language conversational artifacts from API specifications using large language models (LLMs). The goal is to utilize LLMs in the “build” phase to assist humans in creating skills for digital assistants. As a result, the system does not need to rely on LLMs during conversations with business users, leading to efficient deployment. Along with enabling digital assistants, our system employs LLMs as proxies to simulate human interaction and automatically evaluate the digital assistant's performance. Experimental results highlight the effectiveness of our proposed approach. Our system is deployed in the IBM Watson Orchestrate product for general availability.

在业务自动化领域,会话助手正在成为各种业务部门的用户访问自动化软件的主要方法。对自动化的访问主要是通过应用程序编程接口(api)和机器人过程自动化(rpa)实现的。为了有效地将api和rpa转换为更大规模的聊天机器人,建立一个自动化的过程来生成数据和训练模型是至关重要的,这些模型可以识别用户意图,识别会话槽填充的问题,并为后续行动提供建议。在本文中,我们提出了一种使用大型语言模型(llm)从API规范增强和生成自然语言会话工件的技术。目标是在“构建”阶段利用法学硕士来帮助人类为数字助理创造技能。因此,系统在与业务用户对话时不需要依赖llm,从而实现了高效的部署。随着数字助理的启用,我们的系统采用法学硕士作为代理来模拟人类互动并自动评估数字助理的表现。实验结果表明了该方法的有效性。我们的系统部署在IBM Watson Orchestrate产品中,以提供一般可用性。
{"title":"Framework to enable and test conversational assistant for APIs and RPAs","authors":"Jayachandu Bandlamudi,&nbsp;Kushal Mukherjee,&nbsp;Prerna Agarwal,&nbsp;Ritwik Chaudhuri,&nbsp;Rakesh Pimplikar,&nbsp;Sampath Dechu,&nbsp;Alex Straley,&nbsp;Anbumunee Ponniah,&nbsp;Renuka Sindhgatta","doi":"10.1002/aaai.12198","DOIUrl":"https://doi.org/10.1002/aaai.12198","url":null,"abstract":"<p>In the realm of business automation, conversational assistants are emerging as the primary method for making automation software accessible to users in various business sectors. Access to automation primarily occurs through application programming interface (APIs) and robotic process automation (RPAs). To effectively convert APIs and RPAs into chatbots on a larger scale, it is crucial to establish an automated process for generating data and training models that can recognize user intentions, identify questions for conversational slot filling, and provide recommendations for subsequent actions. In this paper, we present a technique for enhancing and generating natural language conversational artifacts from API specifications using large language models (LLMs). The goal is to utilize LLMs in the “build” phase to assist humans in creating skills for digital assistants. As a result, the system does not need to rely on LLMs during conversations with business users, leading to efficient deployment. Along with enabling digital assistants, our system employs LLMs as proxies to simulate human interaction and automatically evaluate the digital assistant's performance. Experimental results highlight the effectiveness of our proposed approach. Our system is deployed in the IBM Watson Orchestrate product for general availability.</p>","PeriodicalId":7854,"journal":{"name":"Ai Magazine","volume":"45 4","pages":"443-456"},"PeriodicalIF":2.5,"publicationDate":"2024-10-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/aaai.12198","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142851522","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Data-efficient graph learning: Problems, progress, and prospects 数据高效图学习:问题、进展和前景
IF 2.5 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-10-18 DOI: 10.1002/aaai.12200
Kaize Ding, Yixin Liu, Chuxu Zhang, Jianling Wang

Graph-structured data, ranging from social networks to financial transaction networks, from citation networks to gene regulatory networks, have been widely used for modeling a myriad of real-world systems. As a prevailing model architecture to model graph-structured data, graph neural networks (GNNs) have drawn much attention in both academic and industrial communities in the past decades. Despite their success in different graph learning tasks, existing methods usually rely on learning from “big” data, requiring a large amount of labeled data for model training. However, it is common that real-world graphs are associated with “small” labeled data as data annotation and labeling on graphs is always time and resource-consuming. Therefore, it is imperative to investigate graph machine learning (graph ML) with low-cost human supervision for low-resource settings where limited or even no labeled data is available. This paper investigates a new research field—data-efficient graph learning, which aims to push forward the performance boundary of graph ML models with different kinds of low-cost supervision signals. Specifically, we outline the fundamental research problems, review the current progress, and discuss the future prospects of data-efficient graph learning, aiming to illuminate the path for subsequent research in this field.

从社交网络到金融交易网络,从引文网络到基因调控网络,图结构数据已被广泛用于模拟现实世界中的各种系统。作为图结构数据建模的主流模型架构,图神经网络(GNN)在过去几十年中引起了学术界和工业界的广泛关注。尽管它们在不同的图学习任务中取得了成功,但现有方法通常依赖于从 "大 "数据中学习,需要大量标注数据来进行模型训练。然而,现实世界中的图通常与 "小 "标注数据相关联,因为对图进行数据注释和标注总是耗费时间和资源。因此,在资源有限甚至没有标注数据的情况下,研究具有低成本人工监督的图机器学习(graph ML)势在必行。本文探讨了一个新的研究领域--数据高效图学习,旨在通过不同类型的低成本监督信号来推动图 ML 模型的性能边界。具体而言,我们概述了数据高效图学习的基础研究问题,回顾了当前的研究进展,并讨论了其未来前景,旨在为该领域的后续研究指明方向。
{"title":"Data-efficient graph learning: Problems, progress, and prospects","authors":"Kaize Ding,&nbsp;Yixin Liu,&nbsp;Chuxu Zhang,&nbsp;Jianling Wang","doi":"10.1002/aaai.12200","DOIUrl":"https://doi.org/10.1002/aaai.12200","url":null,"abstract":"<p>Graph-structured data, ranging from social networks to financial transaction networks, from citation networks to gene regulatory networks, have been widely used for modeling a myriad of real-world systems. As a prevailing model architecture to model graph-structured data, graph neural networks (GNNs) have drawn much attention in both academic and industrial communities in the past decades. Despite their success in different graph learning tasks, existing methods usually rely on learning from “big” data, requiring a large amount of labeled data for model training. However, it is common that real-world graphs are associated with “small” labeled data as data annotation and labeling on graphs is always time and resource-consuming. Therefore, it is imperative to investigate graph machine learning (graph ML) with low-cost human supervision for low-resource settings where limited or even no labeled data is available. This paper investigates a new research field—data-efficient graph learning, which aims to push forward the performance boundary of graph ML models with different kinds of low-cost supervision signals. Specifically, we outline the fundamental research problems, review the current progress, and discuss the future prospects of data-efficient graph learning, aiming to illuminate the path for subsequent research in this field.</p>","PeriodicalId":7854,"journal":{"name":"Ai Magazine","volume":"45 4","pages":"549-560"},"PeriodicalIF":2.5,"publicationDate":"2024-10-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/aaai.12200","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142861732","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
DCV 2 I $text{DCV}^2text{I}$ : Leveraging deep vision models to support geographers' visual interpretation in dune segmentation DCV 2 I $text{DCV}^2text{I}$:利用深度视觉模型支持地理学家在沙丘分割中的视觉解释
IF 2.5 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-10-18 DOI: 10.1002/aaai.12199
Anqi Lu, Zifeng Wu, Zheng Jiang, Wei Wang, Eerdun Hasi, Yi Wang

Visual interpretation is extremely important in human geography as the primary technique for geographers to use photograph data in identifying, classifying, and quantifying geographic and topological objects or regions. However, it is also time-consuming and requires overwhelming manual effort from professional geographers. This paper describes our interdisciplinary team's efforts in integrating computer vision models with geographers' visual image interpretation process to reduce their workload in interpreting images. Focusing on the dune segmentation task, we proposed an approach called DCV2I${bf DCV}^2{bf I}$ featuring a deep dune segmentation model to identify dunes and label their ranges in an automated way. By developing a tool to connect our model with ArcGIS—one of the most popular workbenches for visual interpretation, geographers can further refine the automatically generated dune segmentation on images without learning any CV or deep learning techniques. Our approach thus realized a noninvasive change to geographers' visual interpretation routines, reducing their manual efforts while incurring minimal interruptions to their work routines and tools they are familiar with. Deployment with a leading Chinese geography research institution demonstrated the potential of DCV2I${bf DCV}^2{bf I}$ in supporting geographers in researching and solving drylands desertification.

目视判读在人文地理学中极为重要,是地理学家利用照片数据识别、分类和量化地理和地形对象或区域的主要技术。然而,这也非常耗时,需要专业地理学家付出大量的人工努力。本文介绍了我们的跨学科团队如何将计算机视觉模型与地理学家的视觉图像判读过程相结合,以减轻他们判读图像的工作量。针对沙丘分割任务,我们提出了一种名为 DCV 2 I ${bf DCV}^2{bf I}$ 的方法,其特点是采用深度沙丘分割模型来自动识别沙丘并标注其范围。通过开发一种工具,将我们的模型与 ArcGIS(最流行的可视化解释工作台之一)连接起来,地理学家无需学习任何 CV 或深度学习技术,就能进一步完善自动生成的沙丘分割图像。因此,我们的方法实现了对地理学家可视化判读例程的非侵入式改变,减少了他们的手工操作,同时对他们的工作例程和熟悉的工具产生了最小的干扰。在中国领先的地理研究机构的部署表明,DCV 2 I ${bf DCV}^2{bf I}$在支持地理学家研究和解决旱地荒漠化问题方面具有潜力。
{"title":"DCV\u0000 2\u0000 \u0000 I\u0000 \u0000 $text{DCV}^2text{I}$\u0000 : Leveraging deep vision models to support geographers' visual interpretation in dune segmentation","authors":"Anqi Lu,&nbsp;Zifeng Wu,&nbsp;Zheng Jiang,&nbsp;Wei Wang,&nbsp;Eerdun Hasi,&nbsp;Yi Wang","doi":"10.1002/aaai.12199","DOIUrl":"https://doi.org/10.1002/aaai.12199","url":null,"abstract":"<p>Visual interpretation is extremely important in human geography as the primary technique for geographers to use photograph data in identifying, classifying, and quantifying geographic and topological objects or regions. However, it is also time-consuming and requires overwhelming manual effort from professional geographers. This paper describes our interdisciplinary team's efforts in integrating computer vision models with geographers' visual image interpretation process to reduce their workload in interpreting images. Focusing on the dune segmentation task, we proposed an approach called <span></span><math>\u0000 <semantics>\u0000 <mrow>\u0000 <msup>\u0000 <mi>DCV</mi>\u0000 <mn>2</mn>\u0000 </msup>\u0000 <mi>I</mi>\u0000 </mrow>\u0000 <annotation>${bf DCV}^2{bf I}$</annotation>\u0000 </semantics></math> featuring a deep dune segmentation model to identify dunes and label their ranges in an automated way. By developing a tool to connect our model with ArcGIS—one of the most popular workbenches for visual interpretation, geographers can further refine the automatically generated dune segmentation on images without learning any CV or deep learning techniques. Our approach thus realized a noninvasive change to geographers' visual interpretation routines, reducing their manual efforts while incurring minimal interruptions to their work routines and tools they are familiar with. Deployment with a leading Chinese geography research institution demonstrated the potential of <span></span><math>\u0000 <semantics>\u0000 <mrow>\u0000 <msup>\u0000 <mi>DCV</mi>\u0000 <mn>2</mn>\u0000 </msup>\u0000 <mi>I</mi>\u0000 </mrow>\u0000 <annotation>${bf DCV}^2{bf I}$</annotation>\u0000 </semantics></math> in supporting geographers in researching and solving drylands desertification.</p>","PeriodicalId":7854,"journal":{"name":"Ai Magazine","volume":"45 4","pages":"472-485"},"PeriodicalIF":2.5,"publicationDate":"2024-10-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/aaai.12199","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142861675","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A submodular optimization approach to trustworthy loan approval automation 可信赖贷款审批自动化的子模块优化方法
IF 2.5 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-10-01 DOI: 10.1002/aaai.12195
Kyungsik Lee, Hana Yoo, Sumin Shin, Wooyoung Kim, Yeonung Baek, Hyunjin Kang, Jaehyun Kim, Kee-Eung Kim

In the field of finance, the underwriting process is an essential step in evaluating every loan application. During this stage, the borrowers' creditworthiness and ability to repay the loan are assessed to ultimately decide whether to approve the loan application. One of the core components of underwriting is credit scoring, in which the probability of default is estimated. As such, there has been significant progress in enhancing the predictive accuracy of credit scoring models through the use of machine learning, but there still exists a need to ultimately construct an approval rule that takes into consideration additional criteria beyond the score itself. This construction process is traditionally done manually to ensure that the approval rule remains interpretable to humans. In this paper, we outline an automated system for optimizing a rule-based system for approving loan applications, which has been deployed at Hyundai Capital Services (HCS). The main challenge lays in creating a high-quality rule base that is simultaneously simple enough to be interpretable by risk analysts as well as customers, since the approval decision should be easily understandable. We addressed this challenge through principled submodular optimization. The deployment of our system has led to a 14% annual growth in the volume of loan services at HCS, while maintaining the target bad rate, and has resulted in the approval of customers who might have otherwise been rejected.

在金融领域,承销过程是评估每一笔贷款申请的重要步骤。在此阶段,评估借款人的信誉和偿还贷款的能力,以最终决定是否批准贷款申请。承保的核心组成部分之一是信用评分,其中估计了违约的可能性。因此,通过使用机器学习,在提高信用评分模型的预测准确性方面已经取得了重大进展,但仍然需要最终构建一个考虑分数本身之外的其他标准的审批规则。这个构造过程传统上是手动完成的,以确保审批规则仍然对人类是可解释的。在本文中,我们概述了一个自动化系统,用于优化基于规则的系统,以批准贷款申请,该系统已部署在现代资本服务公司(HCS)。主要的挑战在于创建一个高质量的规则基础,它同时足够简单,可以被风险分析师和客户解释,因为批准决定应该很容易理解。我们通过原则性的子模块优化解决了这一挑战。我们系统的部署使HCS的贷款业务量每年增长14%,同时保持了目标不良率,并获得了可能被拒绝的客户的认可。
{"title":"A submodular optimization approach to trustworthy loan approval automation","authors":"Kyungsik Lee,&nbsp;Hana Yoo,&nbsp;Sumin Shin,&nbsp;Wooyoung Kim,&nbsp;Yeonung Baek,&nbsp;Hyunjin Kang,&nbsp;Jaehyun Kim,&nbsp;Kee-Eung Kim","doi":"10.1002/aaai.12195","DOIUrl":"https://doi.org/10.1002/aaai.12195","url":null,"abstract":"<p>In the field of finance, the underwriting process is an essential step in evaluating every loan application. During this stage, the borrowers' creditworthiness and ability to repay the loan are assessed to ultimately decide whether to approve the loan application. One of the core components of underwriting is credit scoring, in which the probability of default is estimated. As such, there has been significant progress in enhancing the predictive accuracy of credit scoring models through the use of machine learning, but there still exists a need to ultimately construct an approval rule that takes into consideration additional criteria beyond the score itself. This construction process is traditionally done manually to ensure that the approval rule remains interpretable to humans. In this paper, we outline an automated system for optimizing a rule-based system for approving loan applications, which has been deployed at Hyundai Capital Services (HCS). The main challenge lays in creating a high-quality rule base that is simultaneously simple enough to be interpretable by risk analysts as well as customers, since the approval decision should be easily understandable. We addressed this challenge through principled submodular optimization. The deployment of our system has led to a 14% annual growth in the volume of loan services at HCS, while maintaining the target bad rate, and has resulted in the approval of customers who might have otherwise been rejected.</p>","PeriodicalId":7854,"journal":{"name":"Ai Magazine","volume":"45 4","pages":"502-513"},"PeriodicalIF":2.5,"publicationDate":"2024-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/aaai.12195","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142859946","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Effective knowledge representation and utilization for sustainable collaborative learning across heterogeneous systems 有效的知识表示和利用,促进跨异构系统的可持续协作学习
IF 2.5 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-09-22 DOI: 10.1002/aaai.12193
Trong Nghia Hoang

The increasingly decentralized and private nature of data in our digital society has motivated the development of collaborative intelligent systems that enable knowledge aggregation among data owners. However, collaborative learning has only been investigated in simple settings. For example, clients are often assumed to train solution models de novo, disregarding all prior expertise. The learned model is typically represented in task-specific forms that are not generalizable to unseen, emerging scenarios. Finally, a universal model representation is enforced among collaborators, ignoring their local compute constraints or input representations. These limitations hampers the practicality of prior collaborative systems in learning scenarios with limited task data that demand constant knowledge adaptation and transfer across information silos, tasks, and learning models, as well as the utilization of prior solution expertise. Furthermore, prior collaborative learning frameworks are not sustainable on a macro scale where participants desire fairness allocation of benefits (e.g., access to the combined model) based on their costs of participation (e.g., overhead of model sharing and training synchronization, risk of information breaches, etc.). This necessitates a new perspective of collaborative learning where the server not only aggregates but also conducts valuation of the participant's contribution, and distribute aggregated information to individuals in commensurate to their contribution. To substantiate the above vision, we propose a new research agenda on developing effective and sustainable collaborative learning frameworks across heterogeneous systems, featuring three novel computational capabilities on knowledge organization: model expression, comprehension, and valuation.

在我们的数字社会中,数据的分散性和私密性越来越强,这促使人们开发能够在数据所有者之间聚合知识的协作智能系统。然而,人们只在简单的环境中研究过协作学习。例如,通常假设客户从头开始训练解决方案模型,而不考虑所有先前的专业知识。学习到的模型通常以特定任务的形式表示,无法推广到未见过的新兴场景中。最后,合作者之间强制使用通用模型表示法,忽略了他们的本地计算约束或输入表示法。这些局限性妨碍了先前的协作系统在任务数据有限的学习场景中的实用性,因为这种场景需要在信息孤岛、任务和学习模型之间不断进行知识调整和转移,并需要利用先前的解决方案专长。此外,先前的协作学习框架在宏观上是不可持续的,因为参与者希望根据他们的参与成本(如模型共享和培训同步的开销、信息泄露的风险等)公平分配利益(如访问组合模型)。这就需要一种新的协作学习视角,即服务器不仅要汇总信息,还要对参与者的贡献进行评估,并将汇总信息按贡献分配给个人。为了证实上述愿景,我们提出了一个新的研究议程,即在异构系统中开发有效、可持续的协作学习框架,其中包括三种新的知识组织计算能力:模型表达、理解和评估。
{"title":"Effective knowledge representation and utilization for sustainable collaborative learning across heterogeneous systems","authors":"Trong Nghia Hoang","doi":"10.1002/aaai.12193","DOIUrl":"https://doi.org/10.1002/aaai.12193","url":null,"abstract":"<p>The increasingly decentralized and private nature of data in our digital society has motivated the development of collaborative intelligent systems that enable knowledge aggregation among data owners. However, collaborative learning has only been investigated in simple settings. For example, clients are often assumed to train solution models <i>de novo</i>, disregarding all prior expertise. The learned model is typically represented in task-specific forms that are not generalizable to unseen, emerging scenarios. Finally, a universal model representation is enforced among collaborators, ignoring their local compute constraints or input representations. These limitations hampers the practicality of prior collaborative systems in learning scenarios with limited task data that demand constant knowledge adaptation and transfer across information silos, tasks, and learning models, as well as the utilization of prior solution expertise. Furthermore, prior collaborative learning frameworks are not sustainable on a macro scale where participants desire fairness allocation of benefits (e.g., access to the combined model) based on their costs of participation (e.g., overhead of model sharing and training synchronization, risk of information breaches, etc.). This necessitates a new perspective of collaborative learning where the server not only aggregates but also conducts valuation of the participant's contribution, and distribute aggregated information to individuals in commensurate to their contribution. To substantiate the above vision, we propose a new research agenda on developing effective and sustainable collaborative learning frameworks across heterogeneous systems, featuring three novel computational capabilities on knowledge organization: model expression, comprehension, and valuation.</p>","PeriodicalId":7854,"journal":{"name":"Ai Magazine","volume":"45 3","pages":"404-410"},"PeriodicalIF":2.5,"publicationDate":"2024-09-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/aaai.12193","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142324480","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Fair and optimal prediction via post-processing 通过后处理实现公平和优化预测
IF 2.5 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-09-22 DOI: 10.1002/aaai.12191
Han Zhao

With the development of machine learning algorithms and the increasing computational resources available, artificial intelligence has achieved great success in many application domains. However, the success of machine learning has also raised concerns about the fairness of the learned models. For instance, the learned models can perpetuate and even exacerbate the potential bias and discrimination in the training data. This issue has become a major obstacle to the deployment of machine learning systems in high-stakes domains, for example, criminal judgment, medical testing, online advertising, hiring process, and so forth. To mitigate the potential bias exhibited by machine learning models, fairness criteria can be integrated into the training process to ensure fair treatment across all demographics, but it often comes at the expense of model performance. Understanding such tradeoffs, therefore, is crucial to the design of optimal and fair algorithms. My research focuses on characterizing the inherent tradeoff between fairness and accuracy in machine learning, and developing algorithms that can achieve both fairness and optimality. In this article, I will discuss our recent work on designing post-processing algorithms for fair classification, which can be applied to a wide range of fairness criteria, including statistical parity, equal opportunity, and equalized odds, under both attribute-aware and attribute-blind settings, and is particularly suited to large-scale foundation models where retraining is expensive or even infeasible. I will also discuss the connections between our work and other related research on trustworthy machine learning, including the connections between algorithmic fairness and differential privacy as well as adversarial robustness.

随着机器学习算法的发展和可用计算资源的不断增加,人工智能在许多应用领域取得了巨大成功。然而,机器学习的成功也引发了人们对所学模型公平性的担忧。例如,学习到的模型可能会延续甚至加剧训练数据中潜在的偏见和歧视。这一问题已成为机器学习系统在刑事判决、医疗测试、在线广告、招聘流程等高风险领域部署的主要障碍。为了减轻机器学习模型可能表现出的偏差,可以将公平性标准集成到训练过程中,以确保公平对待所有人群,但这往往以牺牲模型性能为代价。因此,了解这种权衡对于设计最佳公平算法至关重要。我的研究重点是描述机器学习中公平性和准确性之间固有的权衡,并开发能同时实现公平性和最优性的算法。在本文中,我将讨论我们最近在设计公平分类的后处理算法方面所做的工作,该算法可在属性感知和属性盲设置下应用于广泛的公平标准,包括统计均等、机会均等和赔率均等,尤其适用于重新训练成本高昂甚至不可行的大规模基础模型。我还将讨论我们的工作与其他可信机器学习相关研究之间的联系,包括算法公平性与差异隐私以及对抗鲁棒性之间的联系。
{"title":"Fair and optimal prediction via post-processing","authors":"Han Zhao","doi":"10.1002/aaai.12191","DOIUrl":"https://doi.org/10.1002/aaai.12191","url":null,"abstract":"<p>With the development of machine learning algorithms and the increasing computational resources available, artificial intelligence has achieved great success in many application domains. However, the success of machine learning has also raised concerns about the <i>fairness</i> of the learned models. For instance, the learned models can perpetuate and even exacerbate the potential bias and discrimination in the training data. This issue has become a major obstacle to the deployment of machine learning systems in high-stakes domains, for example, criminal judgment, medical testing, online advertising, hiring process, and so forth. To mitigate the potential bias exhibited by machine learning models, fairness criteria can be integrated into the training process to ensure fair treatment across all demographics, but it often comes at the expense of model performance. Understanding such tradeoffs, therefore, is crucial to the design of optimal and fair algorithms. My research focuses on characterizing the inherent tradeoff between fairness and accuracy in machine learning, and developing algorithms that can achieve both fairness and optimality. In this article, I will discuss our recent work on designing post-processing algorithms for fair classification, which can be applied to a wide range of fairness criteria, including statistical parity, equal opportunity, and equalized odds, under both attribute-aware and attribute-blind settings, and is particularly suited to large-scale foundation models where retraining is expensive or even infeasible. I will also discuss the connections between our work and other related research on trustworthy machine learning, including the connections between algorithmic fairness and differential privacy as well as adversarial robustness.</p>","PeriodicalId":7854,"journal":{"name":"Ai Magazine","volume":"45 3","pages":"411-418"},"PeriodicalIF":2.5,"publicationDate":"2024-09-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/aaai.12191","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142324481","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Ai Magazine
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1