首页 > 最新文献

Artificial Intelligence Review最新文献

英文 中文
Reinforcement learning for single-agent to multi-agent systems: from basic theory to industrial application progress, a survey 单智能体到多智能体系统的强化学习:从基础理论到工业应用进展综述
IF 13.9 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-11-27 DOI: 10.1007/s10462-025-11439-9
Dehua Zhang, Qingsong Yuan, Lei Meng, Ruixue Xia, Wei Liu, Chunbin Qin

Reinforcement learning (RL), as an emerging interdisciplinary field formed by the integration of artificial intelligence and control science, is currently demonstrating a cross-disciplinary development trend led by artificial intelligence and has become a research hotspot in the field of optimal control. This paper systematically reviews the development context of RL, focusing on the intrinsic connection between single-agent reinforcement learning (SARL) and multi-agent reinforcement learning (MARL). Firstly, starting from the formation and development of RL, it elaborates on the similarities and differences between RL and other learning paradigms in machine learning, and briefly introduces the main branches of current RL. Then, with the basic knowledge and core ideas of SARL as the basic framework, and expanding to multi-agent system (MAS) collaborative control, it explores the coherence characteristics of the two in theoretical frameworks and algorithm design. On this basis, this paper reconfigures SARL algorithms into dynamic programming, value function decomposition and policy gradient (PG) type, and abstracts MARL algorithms into four paradigms: behavior analysis, centralized learning, communication learning and collaborative learning, thus establishing an algorithm mapping relationship from single-agent to multi-agent scenarios. This innovative framework provides a new perspective for understanding the evolutionary correlation of the two methods, and also discusses the challenges and solution ideas of MARL in solving large-scale MAS problems. This paper aims to provide a reference for researchers in this field, and to promote the development of cooperative control and optimization methods for MAS as well as the advancement of related application research.

强化学习(Reinforcement learning, RL)作为人工智能与控制科学融合形成的新兴跨学科领域,目前呈现出以人工智能为主导的跨学科发展趋势,已成为最优控制领域的研究热点。本文系统回顾了强化学习的发展背景,重点讨论了单智能体强化学习(SARL)和多智能体强化学习(MARL)之间的内在联系。首先,从强化学习的形成和发展出发,阐述了强化学习与机器学习中其他学习范式的异同,并简要介绍了当前强化学习的主要分支。然后,以SARL的基本知识和核心思想为基本框架,扩展到多智能体系统(MAS)协同控制,探讨两者在理论框架和算法设计上的一致性特征。在此基础上,本文将SARL算法重新配置为动态规划、价值函数分解和策略梯度(PG)型,并将MARL算法抽象为行为分析、集中学习、沟通学习和协作学习四种范式,从而建立了单智能体到多智能体场景的算法映射关系。这一创新框架为理解两种方法的演化相关性提供了新的视角,并讨论了MARL在解决大规模MAS问题时面临的挑战和解决思路。本文旨在为该领域的研究人员提供参考,并促进MAS协同控制和优化方法的发展以及相关应用研究的推进。
{"title":"Reinforcement learning for single-agent to multi-agent systems: from basic theory to industrial application progress, a survey","authors":"Dehua Zhang,&nbsp;Qingsong Yuan,&nbsp;Lei Meng,&nbsp;Ruixue Xia,&nbsp;Wei Liu,&nbsp;Chunbin Qin","doi":"10.1007/s10462-025-11439-9","DOIUrl":"10.1007/s10462-025-11439-9","url":null,"abstract":"<div><p>Reinforcement learning (RL), as an emerging interdisciplinary field formed by the integration of artificial intelligence and control science, is currently demonstrating a cross-disciplinary development trend led by artificial intelligence and has become a research hotspot in the field of optimal control. This paper systematically reviews the development context of RL, focusing on the intrinsic connection between single-agent reinforcement learning (SARL) and multi-agent reinforcement learning (MARL). Firstly, starting from the formation and development of RL, it elaborates on the similarities and differences between RL and other learning paradigms in machine learning, and briefly introduces the main branches of current RL. Then, with the basic knowledge and core ideas of SARL as the basic framework, and expanding to multi-agent system (MAS) collaborative control, it explores the coherence characteristics of the two in theoretical frameworks and algorithm design. On this basis, this paper reconfigures SARL algorithms into dynamic programming, value function decomposition and policy gradient (PG) type, and abstracts MARL algorithms into four paradigms: behavior analysis, centralized learning, communication learning and collaborative learning, thus establishing an algorithm mapping relationship from single-agent to multi-agent scenarios. This innovative framework provides a new perspective for understanding the evolutionary correlation of the two methods, and also discusses the challenges and solution ideas of MARL in solving large-scale MAS problems. This paper aims to provide a reference for researchers in this field, and to promote the development of cooperative control and optimization methods for MAS as well as the advancement of related application research.</p></div>","PeriodicalId":8449,"journal":{"name":"Artificial Intelligence Review","volume":"59 2","pages":""},"PeriodicalIF":13.9,"publicationDate":"2025-11-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10462-025-11439-9.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145909005","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Advances in machine learning for wetland classification: a comprehensive survey of methods and applications 机器学习在湿地分类中的进展:方法和应用的综合调查
IF 13.9 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-11-25 DOI: 10.1007/s10462-025-11413-5
Derrick Effah, Ali Zia, Mohammad Awrangjeb, Yongsheng Gao, Kwabena Sarpong

Wetlands are critical ecosystems supporting biodiversity and providing essential environmental services. With increasing threats to wetlands, efficient classification techniques are essential for effective conservation. Several researchers have contributed to wetland classification across reputable journals. However, some challenges (data scarcity, noisy labels, and model generalisability) still exist. Therefore, this paper presents a comprehensive survey of recent advancements in machine learning (ML) and deep learning (DL) for wetland classification, focusing on developments from 2018 to 2025. Key methodologies, including convolutional neural networks, transformers, and generative adversarial networks, are critically reviewed, highlighting their strengths, limitations, and applications in remote sensing. Unlike previous reviews, this work emphasises underexplored techniques such as few-shot learning and Mamba networks, offering practical recommendations for handling limited training data and improving model generalisability. The study also identifies promising research directions, such as test-time training and hybrid loss functions, to address challenges in wetland classification. This survey aims to guide researchers and practitioners in advancing state-of-the-art wetland classification through ML and DL technologies.

湿地是支持生物多样性和提供基本环境服务的重要生态系统。随着湿地面临的威胁日益增加,有效的分类技术是有效保护湿地的必要条件。一些研究人员在知名期刊上对湿地分类做出了贡献。然而,仍然存在一些挑战(数据稀缺性、噪声标签和模型通用性)。因此,本文对用于湿地分类的机器学习(ML)和深度学习(DL)的最新进展进行了全面调查,重点关注2018年至2025年的发展。关键的方法,包括卷积神经网络,变压器和生成对抗网络,严格审查,突出其优势,局限性,并在遥感应用。与以前的评论不同,这项工作强调了未被开发的技术,如少镜头学习和曼巴网络,为处理有限的训练数据和提高模型的通用性提供了实用的建议。该研究还确定了有前途的研究方向,如测试时间训练和混合损失函数,以解决湿地分类中的挑战。本调查旨在指导研究人员和实践者通过ML和DL技术推进最先进的湿地分类。
{"title":"Advances in machine learning for wetland classification: a comprehensive survey of methods and applications","authors":"Derrick Effah,&nbsp;Ali Zia,&nbsp;Mohammad Awrangjeb,&nbsp;Yongsheng Gao,&nbsp;Kwabena Sarpong","doi":"10.1007/s10462-025-11413-5","DOIUrl":"10.1007/s10462-025-11413-5","url":null,"abstract":"<div><p>Wetlands are critical ecosystems supporting biodiversity and providing essential environmental services. With increasing threats to wetlands, efficient classification techniques are essential for effective conservation. Several researchers have contributed to wetland classification across reputable journals. However, some challenges (data scarcity, noisy labels, and model generalisability) still exist. Therefore, this paper presents a comprehensive survey of recent advancements in machine learning (ML) and deep learning (DL) for wetland classification, focusing on developments from 2018 to 2025. Key methodologies, including convolutional neural networks, transformers, and generative adversarial networks, are critically reviewed, highlighting their strengths, limitations, and applications in remote sensing. Unlike previous reviews, this work emphasises underexplored techniques such as few-shot learning and Mamba networks, offering practical recommendations for handling limited training data and improving model generalisability. The study also identifies promising research directions, such as test-time training and hybrid loss functions, to address challenges in wetland classification. This survey aims to guide researchers and practitioners in advancing state-of-the-art wetland classification through ML and DL technologies.</p></div>","PeriodicalId":8449,"journal":{"name":"Artificial Intelligence Review","volume":"59 1","pages":""},"PeriodicalIF":13.9,"publicationDate":"2025-11-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10462-025-11413-5.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145612496","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A computer graphics-based model to generate dynamic 3D animations for corresponding Bangla sign language gestures using HamNoSys to SiGML conversion 一个基于计算机图形的模型,使用HamNoSys到SiGML的转换,为相应的孟加拉语手语手势生成动态3D动画
IF 13.9 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-11-25 DOI: 10.1007/s10462-025-11370-z
Ahsanul Karim, Muhammad Aminur Rahaman, Md. Ariful Islam, Md. Ariful Islam, Anichur Rahman, Tanoy Debnath, Utpol Kanti Das

Effective communication is essential for human touch, and it’s especially crucial for those who rely on sign language because they have speech or hearing difficulties. Communication between deaf and normal people is restricted by the inability of current technology to automatically create flexible Bangla Sign Language (BdSL) animations from Bangla text or voice. This work introduces a revolutionary computer graphics-based system that takes voice or text input in Bangla and uses it to create BdSL animations. The system uses a HamNoSys to Gesture Markup Language (SiGML) conversion to dynamically translate input into gestures. With the use of 94 classes in the dataset that cover Bangla numbers, alphabets, and word motions, the model can generate any word or phrase in Bangla. Additionally, the system creates gestures for inputs that aren’t in the dataset by spelling out the letters. The system converts Bangla text into three-dimensional animated BdSL movements by parsing it based on Linguistic principles. According to performance evaluations, each input has a processing cost of 79.57 ms, and the average accuracy for text and voice input is 97.50% and 94.75%, respectively. The method ensures fluency and naturalness by taking into consideration crucial elements of sign language, such as hand shape, palm orientation, and non-manual messages. By reducing communication barriers between the sign and non-sign populations, this study significantly advances accessibility. Visit this GitHub repository link for further details on the implementation of the suggested system and the SiGML dataset: https://gitlab.com/devarifkhan/bdsl-3d-animation

有效的沟通对于人与人之间的接触至关重要,对于那些依赖手语的人来说尤其重要,因为他们有语言或听力障碍。聋哑人与正常人之间的交流受到当前技术的限制,无法从孟加拉语文本或语音自动创建灵活的孟加拉语手语动画。这项工作介绍了一个革命性的基于计算机图形的系统,该系统采用孟加拉语的语音或文本输入并使用它来创建BdSL动画。该系统使用HamNoSys到手势标记语言(SiGML)的转换来动态地将输入转换为手势。通过使用数据集中94个类,涵盖孟加拉语数字、字母和单词运动,该模型可以生成孟加拉语中的任何单词或短语。此外,系统通过拼写字母为数据集中没有的输入创建手势。该系统基于语言学原理对孟加拉语文本进行解析,将其转换为三维动画BdSL运动。根据性能评估,每次输入的处理成本为79.57 ms,文本和语音输入的平均准确率分别为97.50%和94.75%。该方法通过考虑手语的关键要素,如手的形状、手掌的方向和非手动信息,确保流利和自然。通过减少手语和非手语人群之间的交流障碍,本研究显著提高了可达性。请访问此GitHub存储库链接,了解有关建议系统和SiGML数据集实现的更多详细信息:https://gitlab.com/devarifkhan/bdsl-3d-animation
{"title":"A computer graphics-based model to generate dynamic 3D animations for corresponding Bangla sign language gestures using HamNoSys to SiGML conversion","authors":"Ahsanul Karim,&nbsp;Muhammad Aminur Rahaman,&nbsp;Md. Ariful Islam,&nbsp;Md. Ariful Islam,&nbsp;Anichur Rahman,&nbsp;Tanoy Debnath,&nbsp;Utpol Kanti Das","doi":"10.1007/s10462-025-11370-z","DOIUrl":"10.1007/s10462-025-11370-z","url":null,"abstract":"<div><p>Effective communication is essential for human touch, and it’s especially crucial for those who rely on sign language because they have speech or hearing difficulties. Communication between deaf and normal people is restricted by the inability of current technology to automatically create flexible Bangla Sign Language (BdSL) animations from Bangla text or voice. This work introduces a revolutionary computer graphics-based system that takes voice or text input in Bangla and uses it to create BdSL animations. The system uses a HamNoSys to Gesture Markup Language (SiGML) conversion to dynamically translate input into gestures. With the use of 94 classes in the dataset that cover Bangla numbers, alphabets, and word motions, the model can generate any word or phrase in Bangla. Additionally, the system creates gestures for inputs that aren’t in the dataset by spelling out the letters. The system converts Bangla text into three-dimensional animated BdSL movements by parsing it based on Linguistic principles. According to performance evaluations, each input has a processing cost of 79.57 ms, and the average accuracy for text and voice input is 97.50% and 94.75%, respectively. The method ensures fluency and naturalness by taking into consideration crucial elements of sign language, such as hand shape, palm orientation, and non-manual messages. By reducing communication barriers between the sign and non-sign populations, this study significantly advances accessibility. Visit this GitHub repository link for further details on the implementation of the suggested system and the SiGML dataset: https://gitlab.com/devarifkhan/bdsl-3d-animation</p></div>","PeriodicalId":8449,"journal":{"name":"Artificial Intelligence Review","volume":"59 1","pages":""},"PeriodicalIF":13.9,"publicationDate":"2025-11-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10462-025-11370-z.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145612495","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A comprehensive review of current robot-based pollinators for crop pollination 目前农作物传粉机器人传粉器的综合综述
IF 13.9 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-11-25 DOI: 10.1007/s10462-025-11409-1
Rajmeet Singh, Lakmal Seneviratne, Irfan Hussain

The decline of bee and wind-based pollination systems in greenhouses due to controlled environments and limited access has boosted the importance of finding alternative pollination methods. Robot-based pollination systems have emerged as a promising solution, ensuring adequate crop yield even in challenging pollination scenarios. This paper presents a comprehensive review of the current robotic-based pollinators employed in agriculture. The review categorizes pollinator technologies into major categories such as air-jet, water-jet, linear actuator, ultrasonic wave, and air-liquid spray, each suitable for specific crop pollination requirements. However, these technologies are often tailored to particular crops, limiting their versatility. The advancement of science and technology has led to the integration of automated pollination technology, encompassing information technology, automatic perception, detection, control, and operation. This integration not only reduces labor shortage problem but also fosters the ongoing progress of modern agriculture by refining technology, enhancing automation, and promoting intelligence in agricultural practices. Finally, the challenges encountered in the design of robot-based pollinators are addressed, and a forward-looking perspective is taken towards future developments, aiming to contribute to the sustainable advancement of this technology.

由于环境受控制和通道受限,温室中蜜蜂和风媒传粉系统的减少提高了寻找替代传粉方法的重要性。基于机器人的授粉系统已经成为一种有前途的解决方案,即使在具有挑战性的授粉情况下也能确保足够的作物产量。本文介绍了目前在农业中应用的基于机器人的传粉器的全面综述。本文将传粉技术分为空气喷射、水喷射、线性执行器、超声波和气液喷雾等大类,每种传粉技术都适合特定的作物传粉要求。然而,这些技术通常是针对特定作物定制的,限制了它们的通用性。科学技术的进步导致了自动化授粉技术的融合,包括信息技术、自动感知、检测、控制和操作。这种整合不仅减少了劳动力短缺问题,而且通过改进技术,提高自动化程度,促进农业实践的智能化,促进现代农业的不断进步。最后,讨论了机器人传粉器设计中遇到的挑战,并对未来的发展采取了前瞻性的观点,旨在为该技术的可持续发展做出贡献。
{"title":"A comprehensive review of current robot-based pollinators for crop pollination","authors":"Rajmeet Singh,&nbsp;Lakmal Seneviratne,&nbsp;Irfan Hussain","doi":"10.1007/s10462-025-11409-1","DOIUrl":"10.1007/s10462-025-11409-1","url":null,"abstract":"<div><p>The decline of bee and wind-based pollination systems in greenhouses due to controlled environments and limited access has boosted the importance of finding alternative pollination methods. Robot-based pollination systems have emerged as a promising solution, ensuring adequate crop yield even in challenging pollination scenarios. This paper presents a comprehensive review of the current robotic-based pollinators employed in agriculture. The review categorizes pollinator technologies into major categories such as air-jet, water-jet, linear actuator, ultrasonic wave, and air-liquid spray, each suitable for specific crop pollination requirements. However, these technologies are often tailored to particular crops, limiting their versatility. The advancement of science and technology has led to the integration of automated pollination technology, encompassing information technology, automatic perception, detection, control, and operation. This integration not only reduces labor shortage problem but also fosters the ongoing progress of modern agriculture by refining technology, enhancing automation, and promoting intelligence in agricultural practices. Finally, the challenges encountered in the design of robot-based pollinators are addressed, and a forward-looking perspective is taken towards future developments, aiming to contribute to the sustainable advancement of this technology.</p></div>","PeriodicalId":8449,"journal":{"name":"Artificial Intelligence Review","volume":"59 1","pages":""},"PeriodicalIF":13.9,"publicationDate":"2025-11-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10462-025-11409-1.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145612537","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
On the use of transfer learning in nature-inspired algorithms: a systematic review 关于迁移学习在自然启发算法中的应用:系统回顾
IF 13.9 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-11-25 DOI: 10.1007/s10462-025-11404-6
Rita Xavier, Leandro Nunes de Castro

Transfer Learning (TL) has gained significant traction in machine learning, especially in deep learning contexts. However, its integration with Nature-Inspired Algorithms (NIAs) remains fragmented, with limited understanding of strategies, challenges, and outcomes. This paper presents the first systematic review focused exclusively on the use of TL in NIAs, excluding deep learning approaches. Major challenges include dealing with domain/task similarity, avoiding negative transfer, selecting what and when to transfer, and adapting TL mechanisms to population-based search paradigms. To address these issues, we conducted a structured analysis of 47 primary studies, categorizing them by TL strategies, learning paradigms, and algorithmic goals. Our findings reveal recurring patterns, highlight open research gaps, and propose future directions for developing robust TL-based NIAs. This review provides a foundation for researchers interested in designing adaptive, efficient, and knowledge-guided metaheuristics for complex optimization tasks.

迁移学习(TL)在机器学习,特别是在深度学习环境中获得了显著的吸引力。然而,它与自然启发算法(nia)的集成仍然是碎片化的,对策略、挑战和结果的理解有限。本文提出了第一个系统综述,专门关注在nia中使用TL,不包括深度学习方法。主要的挑战包括处理领域/任务相似性,避免负迁移,选择迁移的内容和时间,以及使TL机制适应基于人群的搜索范式。为了解决这些问题,我们对47项主要研究进行了结构化分析,并根据学习策略、学习范式和算法目标对它们进行了分类。我们的发现揭示了反复出现的模式,突出了开放的研究空白,并提出了开发强大的基于tl的nia的未来方向。这一综述为研究人员在复杂优化任务中设计自适应、高效和知识引导的元启发式算法提供了基础。
{"title":"On the use of transfer learning in nature-inspired algorithms: a systematic review","authors":"Rita Xavier,&nbsp;Leandro Nunes de Castro","doi":"10.1007/s10462-025-11404-6","DOIUrl":"10.1007/s10462-025-11404-6","url":null,"abstract":"<div><p>Transfer Learning (TL) has gained significant traction in machine learning, especially in deep learning contexts. However, its integration with Nature-Inspired Algorithms (NIAs) remains fragmented, with limited understanding of strategies, challenges, and outcomes. This paper presents the first systematic review focused exclusively on the use of TL in NIAs, excluding deep learning approaches. Major challenges include dealing with domain/task similarity, avoiding negative transfer, selecting what and when to transfer, and adapting TL mechanisms to population-based search paradigms. To address these issues, we conducted a structured analysis of 47 primary studies, categorizing them by TL strategies, learning paradigms, and algorithmic goals. Our findings reveal recurring patterns, highlight open research gaps, and propose future directions for developing robust TL-based NIAs. This review provides a foundation for researchers interested in designing adaptive, efficient, and knowledge-guided metaheuristics for complex optimization tasks.</p></div>","PeriodicalId":8449,"journal":{"name":"Artificial Intelligence Review","volume":"59 1","pages":""},"PeriodicalIF":13.9,"publicationDate":"2025-11-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10462-025-11404-6.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145612493","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Exploring unanswerability in machine reading comprehension: approaches, benchmarks, and open challenges 探索机器阅读理解中的不可回答性:方法、基准和开放挑战
IF 13.9 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-11-25 DOI: 10.1007/s10462-025-11421-5
Hadiseh Moradisani, Fattane Zarrinkalam, Zeinab Noorian, Faezeh Ensan

The challenge of unanswerable questions in Machine Reading Comprehension (MRC) has drawn considerable attention, as current MRC systems are typically designed under the assumption that every question has a valid answer within the provided context. However, these systems often encounter real-world situations where no valid answer is available. This paper provides a comprehensive review of existing methods for addressing unanswerable questions in MRC systems, categorizing them into model-agnostic and model-specific approaches. It explores key strategies, examines relevant datasets, and evaluates commonly used metrics. This work aims to provide a comprehensive understanding of current techniques and identify critical gaps in the field, offering insights and key challenges to direct future research toward developing more robust MRC systems capable of handling unanswerable questions.

机器阅读理解(MRC)中不可回答问题的挑战引起了相当大的关注,因为当前的MRC系统通常是在假设每个问题在给定的上下文中都有一个有效的答案的情况下设计的。然而,这些系统经常遇到没有有效答案的实际情况。本文提供了解决MRC系统中无法回答的问题的现有方法的全面回顾,将它们分为模型不可知和模型特定的方法。它探讨了关键策略,检查了相关数据集,并评估了常用的指标。这项工作旨在提供对当前技术的全面理解,并确定该领域的关键差距,为指导未来研究开发更强大的MRC系统提供见解和关键挑战,这些系统能够处理无法回答的问题。
{"title":"Exploring unanswerability in machine reading comprehension: approaches, benchmarks, and open challenges","authors":"Hadiseh Moradisani,&nbsp;Fattane Zarrinkalam,&nbsp;Zeinab Noorian,&nbsp;Faezeh Ensan","doi":"10.1007/s10462-025-11421-5","DOIUrl":"10.1007/s10462-025-11421-5","url":null,"abstract":"<div><p>The challenge of unanswerable questions in Machine Reading Comprehension (MRC) has drawn considerable attention, as current MRC systems are typically designed under the assumption that every question has a valid answer within the provided context. However, these systems often encounter real-world situations where no valid answer is available. This paper provides a comprehensive review of existing methods for addressing unanswerable questions in MRC systems, categorizing them into model-agnostic and model-specific approaches. It explores key strategies, examines relevant datasets, and evaluates commonly used metrics. This work aims to provide a comprehensive understanding of current techniques and identify critical gaps in the field, offering insights and key challenges to direct future research toward developing more robust MRC systems capable of handling unanswerable questions.</p></div>","PeriodicalId":8449,"journal":{"name":"Artificial Intelligence Review","volume":"59 1","pages":""},"PeriodicalIF":13.9,"publicationDate":"2025-11-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10462-025-11421-5.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145612492","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Tactical decision making for autonomous trucks by deep reinforcement learning with total cost of operation based reward 基于总运营成本奖励的深度强化学习自动驾驶卡车战术决策
IF 13.9 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-11-25 DOI: 10.1007/s10462-025-11448-8
Deepthi Pathare, Leo Laine, Morteza Haghir Chehreghani

We develop a deep reinforcement learning framework for tactical decision making in an autonomous truck, specifically for Adaptive Cruise Control (ACC) and lane change maneuvers in a highway scenario. Our results demonstrate that it is beneficial to separate high-level decision-making processes and low-level control actions between the reinforcement learning agent and the low-level controllers based on physical models. In the following, we study optimizing the performance with a realistic and multi-objective reward function based on Total Cost of Operation (TCOP) of the truck using different approaches; by adding weights to reward components, by normalizing the reward components and by using curriculum learning techniques.

我们开发了一个深度强化学习框架,用于自动驾驶卡车的战术决策,特别是自适应巡航控制(ACC)和高速公路场景中的变道机动。我们的研究结果表明,基于物理模型的强化学习代理和低级控制器之间将高层决策过程和低级控制动作分开是有益的。接下来,我们研究了基于总运营成本(TCOP)的现实多目标奖励函数下卡车性能优化的不同方法;通过增加奖励成分的权重,通过规范化奖励成分和使用课程学习技术。
{"title":"Tactical decision making for autonomous trucks by deep reinforcement learning with total cost of operation based reward","authors":"Deepthi Pathare,&nbsp;Leo Laine,&nbsp;Morteza Haghir Chehreghani","doi":"10.1007/s10462-025-11448-8","DOIUrl":"10.1007/s10462-025-11448-8","url":null,"abstract":"<div><p>We develop a <i>deep reinforcement learning</i> framework for tactical decision making in an autonomous truck, specifically for Adaptive Cruise Control (ACC) and lane change maneuvers in a highway scenario. Our results demonstrate that it is beneficial to separate high-level decision-making processes and low-level control actions between the reinforcement learning agent and the low-level controllers based on physical models. In the following, we study optimizing the performance with a realistic and multi-objective reward function based on Total Cost of Operation (TCOP) of the truck using different approaches; by adding weights to reward components, by normalizing the reward components and by using curriculum learning techniques.</p></div>","PeriodicalId":8449,"journal":{"name":"Artificial Intelligence Review","volume":"59 1","pages":""},"PeriodicalIF":13.9,"publicationDate":"2025-11-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10462-025-11448-8.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145612494","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Revisiting U-Net: a foundational backbone for modern generative AI 重新审视U-Net:现代生成人工智能的基础支柱
IF 13.9 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-11-24 DOI: 10.1007/s10462-025-11450-0
Marvin John Ignacio, Sangyun Shin, Hulin Jin, Seong Joon Yoo, Dongil Han, Yong-Guk Kim

This survey explores the evolution and application of U-Net in generative AI, highlighting its success across various modalities, including image, text, audio, video, 3D, and pose/action generation. Initially designed for biomedical segmentation, U-Net has been adapted and enhanced with architectural innovations such as normalization techniques, self and cross-attention mechanisms, and residual connections. These advancements have made U-Net a powerful backbone for modern generative models in diffusion-based frameworks, GANs, and autoregressive architectures. The survey comprehensively reviews U-Net’s modality-specific applications, from high-resolution image synthesis and text-to-image generation to speech enhancement, video generation, 3D reconstruction, and pose/action generation. Despite its widespread success, U-Net faces challenges in computational efficiency, contextual understanding, and scalability for multimodal tasks. Future directions focus on optimizing U-Net for lightweight and real-time applications, enhancing its contextual awareness, and improving its integration with emerging architectures like transformers and diffusion models.

本调查探讨了U-Net在生成式人工智能中的演变和应用,突出了其在各种模式下的成功,包括图像、文本、音频、视频、3D和姿势/动作生成。U-Net最初是为生物医学分割而设计的,经过架构创新,如规范化技术、自我和交叉注意机制以及剩余连接,U-Net进行了调整和增强。这些进步使U-Net成为基于扩散的框架、gan和自回归架构中现代生成模型的强大支柱。该调查全面回顾了U-Net的模式特定应用,从高分辨率图像合成、文本到图像生成、语音增强、视频生成、3D重建和姿态/动作生成。尽管U-Net取得了广泛的成功,但它在计算效率、上下文理解和多模式任务的可扩展性方面仍面临挑战。未来的方向将集中在优化U-Net的轻量级和实时应用,增强其上下文感知,并改善其与新兴架构(如变压器和扩散模型)的集成。
{"title":"Revisiting U-Net: a foundational backbone for modern generative AI","authors":"Marvin John Ignacio,&nbsp;Sangyun Shin,&nbsp;Hulin Jin,&nbsp;Seong Joon Yoo,&nbsp;Dongil Han,&nbsp;Yong-Guk Kim","doi":"10.1007/s10462-025-11450-0","DOIUrl":"10.1007/s10462-025-11450-0","url":null,"abstract":"<div><p>This survey explores the evolution and application of U-Net in generative AI, highlighting its success across various modalities, including image, text, audio, video, 3D, and pose/action generation. Initially designed for biomedical segmentation, U-Net has been adapted and enhanced with architectural innovations such as normalization techniques, self and cross-attention mechanisms, and residual connections. These advancements have made U-Net a powerful backbone for modern generative models in diffusion-based frameworks, GANs, and autoregressive architectures. The survey comprehensively reviews U-Net’s modality-specific applications, from high-resolution image synthesis and text-to-image generation to speech enhancement, video generation, 3D reconstruction, and pose/action generation. Despite its widespread success, U-Net faces challenges in computational efficiency, contextual understanding, and scalability for multimodal tasks. Future directions focus on optimizing U-Net for lightweight and real-time applications, enhancing its contextual awareness, and improving its integration with emerging architectures like transformers and diffusion models.</p></div>","PeriodicalId":8449,"journal":{"name":"Artificial Intelligence Review","volume":"59 2","pages":""},"PeriodicalIF":13.9,"publicationDate":"2025-11-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10462-025-11450-0.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145909006","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Legal lay summarization: exploring methods and data generation with large language models 法律法规总结:用大语言模型探索方法和数据生成
IF 13.9 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-11-21 DOI: 10.1007/s10462-025-11392-7
Gianluca Moro, Leonardo David Matteo Magnani, Luca Ragazzi

This paper explores advancements in Natural Language Processing (NLP) for legal lay summarization by systematically analyzing existing methodologies, datasets, and research findings. We review current literature, highlighting key challenges such as data scarcity and the complexity of legal language. A primary contribution of this study is the development of LegalEase, a specialized dataset designed to improve model training for summarizing legal documents in layman’s terms. Our findings demonstrate that subdomain-specific datasets within the legal domain outperform general legal datasets in enhancing NLP model performance for generating accurate and comprehensible legal summaries. The insights and methodologies presented provide a foundation for future research in legal lay summarization.

本文通过系统分析现有的方法、数据集和研究成果,探讨了自然语言处理(NLP)在法律法规摘要方面的进展。我们回顾了当前的文献,突出了数据稀缺和法律语言复杂性等关键挑战。本研究的一个主要贡献是开发了LegalEase,这是一个专门的数据集,旨在改进以外行术语总结法律文件的模型训练。我们的研究结果表明,法律领域的子领域特定数据集在提高NLP模型性能以生成准确和可理解的法律摘要方面优于一般法律数据集。所提出的见解和方法为今后的法律实务总结研究奠定了基础。
{"title":"Legal lay summarization: exploring methods and data generation with large language models","authors":"Gianluca Moro,&nbsp;Leonardo David Matteo Magnani,&nbsp;Luca Ragazzi","doi":"10.1007/s10462-025-11392-7","DOIUrl":"10.1007/s10462-025-11392-7","url":null,"abstract":"<div><p>This paper explores advancements in Natural Language Processing (NLP) for legal lay summarization by systematically analyzing existing methodologies, datasets, and research findings. We review current literature, highlighting key challenges such as data scarcity and the complexity of legal language. A primary contribution of this study is the development of LegalEase, a specialized dataset designed to improve model training for summarizing legal documents in layman’s terms. Our findings demonstrate that subdomain-specific datasets within the legal domain outperform general legal datasets in enhancing NLP model performance for generating accurate and comprehensible legal summaries. The insights and methodologies presented provide a foundation for future research in legal lay summarization.</p></div>","PeriodicalId":8449,"journal":{"name":"Artificial Intelligence Review","volume":"59 1","pages":""},"PeriodicalIF":13.9,"publicationDate":"2025-11-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10462-025-11392-7.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145561584","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Neural network-based prediction of SMTP errors and bounces in cold emailing: a comparative study of GRU, CNN, and TCN 基于神经网络的冷邮件SMTP错误和反弹预测:GRU、CNN和TCN的比较研究
IF 13.9 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-11-21 DOI: 10.1007/s10462-025-11371-y
Anna Jach

The cold email industry faces significant challenges in ensuring successful message delivery, with SMTP errors and bounces being common occurrences. Predicting these errors can help optimize email delivery and improve senders’ reputation. In this study, the abilities of Gated Recurrent Units (GRU), Convolutional Neural Network (CNN), and Temporal Convolutional Network (TCN) were examined to predict SMTP errors and bounces in the context of a cold email dataset. The study reveals that both the GRU networks and the CNNs have achieved over 70% accuracy in predicting SMTP errors and bounces. Exploiting all three architectures, GRU, MLP, and TCN, can substantially enhance the management and optimization of the cold email sending process, leading to improved email delivery rates and better sender reputation.

冷电子邮件行业在确保成功的消息传递方面面临着重大挑战,SMTP错误和反弹经常发生。预测这些错误可以帮助优化电子邮件的发送,提高发件人的声誉。在本研究中,研究了门控循环单元(GRU)、卷积神经网络(CNN)和时间卷积网络(TCN)在冷电子邮件数据集背景下预测SMTP错误和反弹的能力。研究表明,GRU网络和cnn在预测SMTP错误和反弹方面的准确率都达到了70%以上。利用GRU、MLP和TCN这三种架构,可以大大增强对冷电子邮件发送过程的管理和优化,从而提高电子邮件的投递率,提高发件人的声誉。
{"title":"Neural network-based prediction of SMTP errors and bounces in cold emailing: a comparative study of GRU, CNN, and TCN","authors":"Anna Jach","doi":"10.1007/s10462-025-11371-y","DOIUrl":"10.1007/s10462-025-11371-y","url":null,"abstract":"<div><p>The cold email industry faces significant challenges in ensuring successful message delivery, with SMTP errors and bounces being common occurrences. Predicting these errors can help optimize email delivery and improve senders’ reputation. In this study, the abilities of Gated Recurrent Units (GRU), Convolutional Neural Network (CNN), and Temporal Convolutional Network (TCN) were examined to predict SMTP errors and bounces in the context of a cold email dataset. The study reveals that both the GRU networks and the CNNs have achieved over 70% accuracy in predicting SMTP errors and bounces. Exploiting all three architectures, GRU, MLP, and TCN, can substantially enhance the management and optimization of the cold email sending process, leading to improved email delivery rates and better sender reputation.</p></div>","PeriodicalId":8449,"journal":{"name":"Artificial Intelligence Review","volume":"59 1","pages":""},"PeriodicalIF":13.9,"publicationDate":"2025-11-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10462-025-11371-y.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145561521","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Artificial Intelligence Review
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1