首页 > 最新文献

Artificial Intelligence Review最新文献

英文 中文
LegalAI Research in LLM Era: Data, Modeling and Evaluation 法学硕士时代的法学研究:数据、建模与评估
IF 13.9 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-02-15 DOI: 10.1007/s10462-026-11514-9
Xiao Chi, Wei Wang, Ziyao Zhang, Ang Li, Yuting Huang, Yiquan Wu, Kun Kuang, Changlong Sun, Xiaozhong Liu, Fei Wu, Minghui Xiong

Legal artificial intelligence (LegalAI) refers to the use of artificial intelligence technologies to automate various legal tasks. Recent advances in large-scale language models have significantly enhanced the capabilities of LegalAI, marking a new stage in its development. In this paper, we present a comprehensive survey of how large language models (LLMs) are reshaping the research paradigm of LegalAI. Beyond improving task performance, LLMs now serve as integral components across the perspectives of data, modeling, and evaluation. We propose a role-based schema that categorizes the involvement of LLMs along these perspectives and use it to systematically review existing studies in three major legal tasks, including legal classification, legal retrieval, and legal generation. Besides, we conduct a detailed quantitative comparison of LLM effectiveness across roles and tasks, and our findings reveal that the impact of LLMs is shaped by both their assigned roles and the nature of the legal tasks.

法律人工智能(LegalAI)是指利用人工智能技术将各种法律任务自动化。近年来在大规模语言模型方面的进展大大增强了LegalAI的能力,标志着它的发展进入了一个新的阶段。在本文中,我们对大型语言模型(llm)如何重塑LegalAI的研究范式进行了全面的调查。除了提高任务性能之外,法学硕士现在在数据、建模和评估方面都是不可或缺的组成部分。我们提出了一个基于角色的模式,根据这些观点对法学硕士的参与进行分类,并用它来系统地回顾现有的三个主要法律任务的研究,包括法律分类、法律检索和法律生成。此外,我们对不同角色和任务的法学硕士有效性进行了详细的定量比较,我们的研究结果表明,法学硕士的影响是由他们所分配的角色和法律任务的性质共同决定的。
{"title":"LegalAI Research in LLM Era: Data, Modeling and Evaluation","authors":"Xiao Chi,&nbsp;Wei Wang,&nbsp;Ziyao Zhang,&nbsp;Ang Li,&nbsp;Yuting Huang,&nbsp;Yiquan Wu,&nbsp;Kun Kuang,&nbsp;Changlong Sun,&nbsp;Xiaozhong Liu,&nbsp;Fei Wu,&nbsp;Minghui Xiong","doi":"10.1007/s10462-026-11514-9","DOIUrl":"10.1007/s10462-026-11514-9","url":null,"abstract":"<div><p>Legal artificial intelligence (LegalAI) refers to the use of artificial intelligence technologies to automate various legal tasks. Recent advances in large-scale language models have significantly enhanced the capabilities of LegalAI, marking a new stage in its development. In this paper, we present a comprehensive survey of how large language models (LLMs) are reshaping the research paradigm of LegalAI. Beyond improving task performance, LLMs now serve as integral components across the perspectives of data, modeling, and evaluation. We propose a role-based schema that categorizes the involvement of LLMs along these perspectives and use it to systematically review existing studies in three major legal tasks, including legal classification, legal retrieval, and legal generation. Besides, we conduct a detailed quantitative comparison of LLM effectiveness across roles and tasks, and our findings reveal that the impact of LLMs is shaped by both their assigned roles and the nature of the legal tasks.</p></div>","PeriodicalId":8449,"journal":{"name":"Artificial Intelligence Review","volume":"59 4","pages":""},"PeriodicalIF":13.9,"publicationDate":"2026-02-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10462-026-11514-9.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147441723","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Deep learning based visually rich document content understanding: a survey 基于深度学习的视觉丰富文档内容理解:综述
IF 13.9 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-02-13 DOI: 10.1007/s10462-025-11477-3
Yihao Ding, Soyeon Caren Han, Jean Lee, Eduard Hovy

Visually Rich Documents (VRDs) play a vital role in domains such as academia, finance, healthcare, and marketing, as they convey information through a combination of text, layout, and visual elements. Traditional approaches to extracting information from VRDs rely heavily on expert knowledge and manual annotation, making them labor-intensive and inefficient. Recent advances in deep learning have transformed this landscape by enabling multimodal models that integrate vision, language, and layout features through pretraining, significantly improving information extraction performance. This survey presents a comprehensive overview of deep learning-based frameworks for VRD Content Understanding. We categorize existing methods based on their modeling strategies and downstream tasks, and provide a comparative analysis of key components, including feature representation, fusion techniques, model architectures, and pretraining objectives. Additionally, we highlight the strengths and limitations of each approach and discuss their suitability for different applications. The paper concludes with a discussion of current challenges and emerging trends, offering guidance for future research and practical deployment in real-world scenarios.

视觉丰富文档(vrd)在学术、金融、医疗保健和营销等领域发挥着至关重要的作用,因为它们通过文本、布局和视觉元素的组合来传达信息。从vrd中提取信息的传统方法严重依赖于专家知识和人工注释,这使得它们劳动密集型且效率低下。深度学习的最新进展已经改变了这一格局,通过预训练实现了集成视觉、语言和布局特征的多模态模型,显著提高了信息提取性能。本调查对基于深度学习的VRD内容理解框架进行了全面概述。我们根据建模策略和下游任务对现有方法进行了分类,并对关键组件进行了比较分析,包括特征表示、融合技术、模型架构和预训练目标。此外,我们强调了每种方法的优点和局限性,并讨论了它们对不同应用程序的适用性。论文最后讨论了当前的挑战和新兴趋势,为未来的研究和实际部署提供了指导。
{"title":"Deep learning based visually rich document content understanding: a survey","authors":"Yihao Ding,&nbsp;Soyeon Caren Han,&nbsp;Jean Lee,&nbsp;Eduard Hovy","doi":"10.1007/s10462-025-11477-3","DOIUrl":"10.1007/s10462-025-11477-3","url":null,"abstract":"<div><p>Visually Rich Documents (VRDs) play a vital role in domains such as academia, finance, healthcare, and marketing, as they convey information through a combination of text, layout, and visual elements. Traditional approaches to extracting information from VRDs rely heavily on expert knowledge and manual annotation, making them labor-intensive and inefficient. Recent advances in deep learning have transformed this landscape by enabling multimodal models that integrate vision, language, and layout features through pretraining, significantly improving information extraction performance. This survey presents a comprehensive overview of deep learning-based frameworks for VRD Content Understanding. We categorize existing methods based on their modeling strategies and downstream tasks, and provide a comparative analysis of key components, including feature representation, fusion techniques, model architectures, and pretraining objectives. Additionally, we highlight the strengths and limitations of each approach and discuss their suitability for different applications. The paper concludes with a discussion of current challenges and emerging trends, offering guidance for future research and practical deployment in real-world scenarios.</p></div>","PeriodicalId":8449,"journal":{"name":"Artificial Intelligence Review","volume":"59 4","pages":""},"PeriodicalIF":13.9,"publicationDate":"2026-02-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10462-025-11477-3.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147441603","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Geospatial reasoning and awareness in large language models: a systematic review 大语言模型中的地理空间推理和意识:系统综述
IF 13.9 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-02-11 DOI: 10.1007/s10462-026-11512-x
Gabriel Ionut Dorobantu, Ana Cornelia Badea

This research examines the evolution and integration of large language models within the geospatial domain, exploring both theoretical aspects and practical applications. Using a systematic review guided by the PRISMA methodology, the study investigates GeoAI developments and assesses the impact of foundation models in geospatial contexts. The findings highlight that commercial models demonstrate notable capabilities in interpreting geospatial concepts and generating functional code, although they face limitations concerning accessibility, transparency and reliance on external infrastructures. Smaller, open-source models, adapted through approaches such as fine-tuning and Retrieval-Augmented Generation, are identified as feasible alternatives, providing a balanced solution in terms of accuracy, efficiency and customization. The study emphasizes a need for large-scale, standardized datasets for effective training and evaluation of geospatial models, pointing toward a clear direction for future research. Despite significant advancements, achieving full autonomy of geospatial agents in complex task-solving scenarios remains an unresolved challenge. The future progression of GeoAI will rely heavily on interdisciplinary collaboration and the development of robust, transparent and ethical models capable of supporting real-time decision-making and promoting digital transformation in public administration and related fields.

本研究探讨了地理空间领域中大型语言模型的演变和集成,从理论和实际应用两个方面进行了探索。采用PRISMA方法指导的系统回顾,本研究调查了GeoAI的发展,并评估了基础模型在地理空间背景下的影响。研究结果强调,商业模型在解释地理空间概念和生成功能代码方面表现出显著的能力,尽管它们在可访问性、透明度和对外部基础设施的依赖方面面临限制。通过微调和检索增强生成等方法进行调整的较小的开源模型被确定为可行的替代方案,在准确性、效率和定制方面提供了平衡的解决方案。该研究强调需要大规模、标准化的数据集来有效地训练和评估地理空间模型,为未来的研究指明了明确的方向。尽管取得了重大进展,但在复杂任务解决方案中实现地理空间代理的完全自治仍然是一个未解决的挑战。GeoAI的未来发展将在很大程度上依赖于跨学科合作,以及能够支持实时决策和促进公共管理及相关领域数字化转型的强大、透明和道德模型的发展。
{"title":"Geospatial reasoning and awareness in large language models: a systematic review","authors":"Gabriel Ionut Dorobantu,&nbsp;Ana Cornelia Badea","doi":"10.1007/s10462-026-11512-x","DOIUrl":"10.1007/s10462-026-11512-x","url":null,"abstract":"<div><p>This research examines the evolution and integration of large language models within the geospatial domain, exploring both theoretical aspects and practical applications. Using a systematic review guided by the PRISMA methodology, the study investigates GeoAI developments and assesses the impact of foundation models in geospatial contexts. The findings highlight that commercial models demonstrate notable capabilities in interpreting geospatial concepts and generating functional code, although they face limitations concerning accessibility, transparency and reliance on external infrastructures. Smaller, open-source models, adapted through approaches such as fine-tuning and Retrieval-Augmented Generation, are identified as feasible alternatives, providing a balanced solution in terms of accuracy, efficiency and customization. The study emphasizes a need for large-scale, standardized datasets for effective training and evaluation of geospatial models, pointing toward a clear direction for future research. Despite significant advancements, achieving full autonomy of geospatial agents in complex task-solving scenarios remains an unresolved challenge. The future progression of GeoAI will rely heavily on interdisciplinary collaboration and the development of robust, transparent and ethical models capable of supporting real-time decision-making and promoting digital transformation in public administration and related fields.</p></div>","PeriodicalId":8449,"journal":{"name":"Artificial Intelligence Review","volume":"59 4","pages":""},"PeriodicalIF":13.9,"publicationDate":"2026-02-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10462-026-11512-x.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147338422","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Semantic and temporal-aware hybrid embedding for transformer-based sequential recommendation 基于变压器顺序推荐的语义和时间感知混合嵌入
IF 13.9 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-02-11 DOI: 10.1007/s10462-025-11483-5
Smriti Vipin Madangarli, G. Suganeshwari, S. P. Syed Ibrahim, Muthu Subash Kavitha, Norio Setozaki

Sequential recommendation has achieved remarkable progress with self-attentive architectures such as SASRec (Self-Attentive Sequential Recommendation), yet existing approaches often underutilize semantic relationships among items and fail to model temporal dynamics effectively. These limitations reduce the ability of existing systems to capture nuanced user preferences in real-world settings. To address these issues, this work introduces two novel frameworks: one for effective semantic integration and another for a semantic and temporal-aware hybrid embedding generation that enhances the representational capacity of SASRec. These frameworks construct a semantically rich hybrid matrix utilizing Markov State Transition probabilities, Cosine similarity, Personalized PageRank(PPR) based normalization and additionally Time-weighted decay information for the temporal variant. The constructed hybrid matrix is smoothed using a graph convolutional network (GCN) to generate item embeddings, which are then passed to the transformer-based sequential recommendation model SASRec for next-item prediction. Evaluated on three real-world benchmark datasets (MovieLens, Yelp and Amazon Beauty), our proposed temporal variant achieves up to 10.4% improvement in HR@10 and 8.2% in NDCG@10 over SASRec, while maintaining competitive efficiency. Our studies confirm the effectiveness of each component, including semantic graph construction, temporal weighting, and contrastive alignment. These results demonstrate that incorporating semantic and temporal signals into sequential recommenders substantially enhances both accuracy and relevancy of recommendations.

序列推荐在自关注架构(如SASRec(自关注序列推荐))方面取得了显著进展,但现有的方法往往没有充分利用项目之间的语义关系,无法有效地建模时间动态。这些限制降低了现有系统在现实环境中捕捉细微的用户偏好的能力。为了解决这些问题,本工作引入了两个新的框架:一个用于有效的语义集成,另一个用于语义和时间感知的混合嵌入生成,以增强SASRec的表示能力。这些框架利用马尔可夫状态转移概率、余弦相似度、基于个性化PageRank(PPR)的归一化和时间变量的额外时间加权衰减信息构建了一个语义丰富的混合矩阵。构建的混合矩阵使用图卷积网络(GCN)进行平滑以生成项目嵌入,然后将其传递给基于变压器的顺序推荐模型SASRec进行下一个项目预测。在三个真实世界的基准数据集(MovieLens, Yelp和Amazon Beauty)上进行评估,我们提出的时间变量在保持竞争效率的同时,在HR@10和NDCG@10上分别实现了10.4%和8.2%的改进。我们的研究证实了每个组成部分的有效性,包括语义图构建、时间加权和对比对齐。这些结果表明,将语义和时间信号结合到顺序推荐中可以大大提高推荐的准确性和相关性。
{"title":"Semantic and temporal-aware hybrid embedding for transformer-based sequential recommendation","authors":"Smriti Vipin Madangarli,&nbsp;G. Suganeshwari,&nbsp;S. P. Syed Ibrahim,&nbsp;Muthu Subash Kavitha,&nbsp;Norio Setozaki","doi":"10.1007/s10462-025-11483-5","DOIUrl":"10.1007/s10462-025-11483-5","url":null,"abstract":"<div><p>Sequential recommendation has achieved remarkable progress with self-attentive architectures such as SASRec (Self-Attentive Sequential Recommendation), yet existing approaches often underutilize semantic relationships among items and fail to model temporal dynamics effectively. These limitations reduce the ability of existing systems to capture nuanced user preferences in real-world settings. To address these issues, this work introduces two novel frameworks: one for effective semantic integration and another for a semantic and temporal-aware hybrid embedding generation that enhances the representational capacity of SASRec. These frameworks construct a semantically rich hybrid matrix utilizing <b>M</b>arkov State Transition probabilities, <b>C</b>osine similarity, <b>P</b>ersonalized PageRank(PPR) based normalization and additionally <b>T</b>ime-weighted decay information for the temporal variant. The constructed hybrid matrix is smoothed using a graph convolutional network (GCN) to generate item embeddings, which are then passed to the transformer-based sequential recommendation model SASRec for next-item prediction. Evaluated on three real-world benchmark datasets (MovieLens, Yelp and Amazon Beauty), our proposed temporal variant achieves up to 10.4% improvement in HR@10 and 8.2% in NDCG@10 over SASRec, while maintaining competitive efficiency. Our studies confirm the effectiveness of each component, including semantic graph construction, temporal weighting, and contrastive alignment. These results demonstrate that incorporating semantic and temporal signals into sequential recommenders substantially enhances both accuracy and relevancy of recommendations.</p></div>","PeriodicalId":8449,"journal":{"name":"Artificial Intelligence Review","volume":"59 4","pages":""},"PeriodicalIF":13.9,"publicationDate":"2026-02-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10462-025-11483-5.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147338421","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Glider snake optimizer (GSO): a nature-inspired metaheuristic algorithm for global and engineering optimization problems 滑翔机蛇形优化器(GSO):一种受自然启发的全局和工程优化问题的元启发式算法
IF 13.9 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-02-10 DOI: 10.1007/s10462-026-11504-x
El-Sayed M. El-kenawy, Nima Khodadadi, Seyedali Mirjalili, Ahmed Mohamed Zaki, Abdelhameed Ibrahim, Amel Ali Alhussan, Doaa Sami Khafaga, Marwa M. Eid

The rapid expansion of complex engineering and real-world optimization problems necessitates the development of efficient, adaptable, and computationally lightweight metaheuristic algorithms. In this study, a novel nature-inspired algorithm called glider snake optimization (GSO) is proposed, which draws behavioral inspiration from the gliding and serpentine locomotion patterns of arboreal snakes to enhance solution exploration and convergence control. The GSO algorithm incorporates a multi-segment movement mechanism, a flexible gliding path generator, and an elite guidance model to ensure effective balance between exploration and exploitation. Extensive experimental validation is conducted using a comprehensive set of 23 classical benchmark functions, high-dimensional test cases (100D, and 500D), the CEC 2019 benchmark suite, and several constrained engineering design problems. The results demonstrate that GSO outperforms or matches 13 state-of-the-art algorithms, including particle swarm optimization (PSO), grey wolf optimizer (GWO), whale optimization algorithm (WOA), and differential evolution (DE) in terms of accuracy, convergence speed, computational cost, and robustness. The algorithm also exhibits exceptional stability across parameter variations, as confirmed through sensitivity analysis and statistical significance testing. These findings highlight the potential of GSO as a powerful and efficient tool for solving complex optimization problems in both theoretical and practical domains. Additionally, GSO achieves leading performance on most benchmark functions, with error reductions of up to 90% compared with competing algorithms. GSO also demonstrates faster convergence and lower variance across repeated trials, confirming its robustness. These quantitative outcomes further reinforce the effectiveness of the proposed algorithm. The MATLAB and Python implementations of GSO are available at https://nimakhodadadi.com.

复杂工程和现实世界优化问题的迅速发展要求开发高效、适应性强、计算量轻的元启发式算法。本文从树栖蛇的滑翔和蛇形运动模式中汲取行为灵感,提出了一种新颖的自然启发算法滑翔蛇优化(glider snake optimization, GSO),以增强解的探索和收敛控制。该算法结合多段运动机制、柔性滑动路径生成器和精英制导模型,实现了勘探与开发的有效平衡。利用23个经典基准函数、高维测试用例(100D和500D)、CEC 2019基准测试套件和几个约束工程设计问题进行了广泛的实验验证。结果表明,GSO在精度、收敛速度、计算成本和鲁棒性等方面优于或匹配13种最先进的算法,包括粒子群优化算法(PSO)、灰狼优化算法(GWO)、鲸鱼优化算法(WOA)和差分进化算法(DE)。通过灵敏度分析和统计显著性检验证实,该算法在参数变化中也表现出卓越的稳定性。这些发现突出了GSO在理论和实践领域作为解决复杂优化问题的强大有效工具的潜力。此外,GSO在大多数基准函数上实现了领先的性能,与竞争算法相比,误差减少高达90%。GSO在重复试验中也表现出更快的收敛性和更低的方差,证实了它的稳健性。这些量化结果进一步增强了算法的有效性。GSO的MATLAB和Python实现可在https://nimakhodadadi.com上获得。
{"title":"Glider snake optimizer (GSO): a nature-inspired metaheuristic algorithm for global and engineering optimization problems","authors":"El-Sayed M. El-kenawy,&nbsp;Nima Khodadadi,&nbsp;Seyedali Mirjalili,&nbsp;Ahmed Mohamed Zaki,&nbsp;Abdelhameed Ibrahim,&nbsp;Amel Ali Alhussan,&nbsp;Doaa Sami Khafaga,&nbsp;Marwa M. Eid","doi":"10.1007/s10462-026-11504-x","DOIUrl":"10.1007/s10462-026-11504-x","url":null,"abstract":"<div><p>The rapid expansion of complex engineering and real-world optimization problems necessitates the development of efficient, adaptable, and computationally lightweight metaheuristic algorithms. In this study, a novel nature-inspired algorithm called glider snake optimization (GSO) is proposed, which draws behavioral inspiration from the gliding and serpentine locomotion patterns of arboreal snakes to enhance solution exploration and convergence control. The GSO algorithm incorporates a multi-segment movement mechanism, a flexible gliding path generator, and an elite guidance model to ensure effective balance between exploration and exploitation. Extensive experimental validation is conducted using a comprehensive set of 23 classical benchmark functions, high-dimensional test cases (100D, and 500D), the CEC 2019 benchmark suite, and several constrained engineering design problems. The results demonstrate that GSO outperforms or matches 13 state-of-the-art algorithms, including particle swarm optimization (PSO), grey wolf optimizer (GWO), whale optimization algorithm (WOA), and differential evolution (DE) in terms of accuracy, convergence speed, computational cost, and robustness. The algorithm also exhibits exceptional stability across parameter variations, as confirmed through sensitivity analysis and statistical significance testing. These findings highlight the potential of GSO as a powerful and efficient tool for solving complex optimization problems in both theoretical and practical domains. Additionally, GSO achieves leading performance on most benchmark functions, with error reductions of up to 90% compared with competing algorithms. GSO also demonstrates faster convergence and lower variance across repeated trials, confirming its robustness. These quantitative outcomes further reinforce the effectiveness of the proposed algorithm. The MATLAB and Python implementations of GSO are available at https://nimakhodadadi.com.</p></div>","PeriodicalId":8449,"journal":{"name":"Artificial Intelligence Review","volume":"59 3","pages":""},"PeriodicalIF":13.9,"publicationDate":"2026-02-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10462-026-11504-x.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147338271","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Quality over quantity: a data-centric survey of annotation errors in object detection datasets 质量重于数量:以数据为中心的对象检测数据集中标注错误的调查
IF 13.9 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-02-07 DOI: 10.1007/s10462-026-11502-z
Adnan Hussain, Kaleem Ullah, Muhammad Afaq, Muhammad Munsif, Altaf Hussain, Sung Wook Baik

In recent years, object detection has become a cornerstone of many computer vision applications, relying heavily on the availability of high-quality annotated datasets. However, even widely used benchmarks often suffer from annotation issues such as inaccurate bounding boxes, misclassified objects, and missing labels. These annotation errors, especially localization errors, can greatly affect the training and evaluation of detection models. In this survey, we provide a data-centric and comprehensive review of existing methods for identifying and analyzing errors in object detection datasets. We examine the main components of error detection workflows, including annotation error taxonomies and model-agnostic detection techniques. In addition, we develop a standardized categorization of annotation error types specific to object detection, providing a foundation for consistent analysis and comparison across studies. We also perform manual inspections of selected benchmark datasets to observe and quantify common annotation errors in practice. Moreover, the survey highlights the datasets used for evaluating error detection methods and compares their scope and inherent challenges. Finally, we summarize the types of annotation errors found in existing benchmarks and provide recommendations for future research to enhance dataset quality and reliability in object detection.

近年来,目标检测已成为许多计算机视觉应用的基石,这在很大程度上依赖于高质量注释数据集的可用性。然而,即使是广泛使用的基准测试也经常受到标注问题的困扰,例如不准确的边界框、错误分类的对象和缺少标签。这些标注错误,尤其是定位错误,会极大地影响检测模型的训练和评价。在本调查中,我们以数据为中心,全面回顾了识别和分析目标检测数据集中错误的现有方法。我们研究了错误检测工作流的主要组成部分,包括注释错误分类和与模型无关的检测技术。此外,我们开发了特定于对象检测的注释错误类型的标准化分类,为跨研究的一致分析和比较提供了基础。我们还对选定的基准数据集进行手动检查,以观察和量化实践中常见的注释错误。此外,该调查强调了用于评估错误检测方法的数据集,并比较了它们的范围和固有挑战。最后,我们总结了在现有基准测试中发现的标注错误类型,并为未来的研究提供了建议,以提高数据集的质量和目标检测的可靠性。
{"title":"Quality over quantity: a data-centric survey of annotation errors in object detection datasets","authors":"Adnan Hussain,&nbsp;Kaleem Ullah,&nbsp;Muhammad Afaq,&nbsp;Muhammad Munsif,&nbsp;Altaf Hussain,&nbsp;Sung Wook Baik","doi":"10.1007/s10462-026-11502-z","DOIUrl":"10.1007/s10462-026-11502-z","url":null,"abstract":"<div><p>In recent years, object detection has become a cornerstone of many computer vision applications, relying heavily on the availability of high-quality annotated datasets. However, even widely used benchmarks often suffer from annotation issues such as inaccurate bounding boxes, misclassified objects, and missing labels. These annotation errors, especially localization errors, can greatly affect the training and evaluation of detection models. In this survey, we provide a data-centric and comprehensive review of existing methods for identifying and analyzing errors in object detection datasets. We examine the main components of error detection workflows, including annotation error taxonomies and model-agnostic detection techniques. In addition, we develop a standardized categorization of annotation error types specific to object detection, providing a foundation for consistent analysis and comparison across studies. We also perform manual inspections of selected benchmark datasets to observe and quantify common annotation errors in practice. Moreover, the survey highlights the datasets used for evaluating error detection methods and compares their scope and inherent challenges. Finally, we summarize the types of annotation errors found in existing benchmarks and provide recommendations for future research to enhance dataset quality and reliability in object detection.</p></div>","PeriodicalId":8449,"journal":{"name":"Artificial Intelligence Review","volume":"59 3","pages":""},"PeriodicalIF":13.9,"publicationDate":"2026-02-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10462-026-11502-z.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147337682","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Harnessing deep learning for air pollution forecasting: trends, techniques, and future prospects 利用深度学习进行空气污染预测:趋势、技术和未来前景
IF 13.9 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-02-06 DOI: 10.1007/s10462-026-11496-8
Salvator Lawrence, Srimuruganandam Bhathmanabhan

Air pollution is a serious global public health threat arising from exposure to toxic ambient pollutants, including particulate matter (PM), sulphur oxides (SOx), nitrogen oxides (NOx), ozone (O₃), carbon monoxide (CO), and ammonia (NH₃). Traditional statistical and deterministic forecasting models often fail to adequately represent nonlinear interactions among multiple pollutants, meteorological drivers, and anthropogenic influences, motivating the growing adoption of deep learning (DL) approaches. This systematic review synthesizes evidence from more than 150 peer-reviewed studies conducted across diverse geographical regions and employing a wide range of DL architectures, including standalone, hybrid, and advanced spatiotemporal models. Using structured quantitative summaries, rank-based performance comparisons, and methodological assessments, the review identifies leading model families, analyzes pollutant- and horizon-specific performance trends, and evaluates robustness and generalizability across spatial and temporal contexts. Overall, DL models generally outperform traditional approaches, particularly when multi-source inputs and spatiotemporal dependencies are explicitly modeled. Nevertheless, the literature remains fragmented, with a strong concentration of studies in data-rich urban regions of Asia, heterogeneous datasets, inconsistent evaluation protocols, limited transparency, and weak external validity. Addressing these limitations requires standardized preprocessing and benchmarking practices, improved explainability and uncertainty quantification, and the development of globally representative datasets. Emerging directions, including hybrid, physics-informed, and generative DL architectures, offer promising pathways to enhance reliability and operational deployment. Collectively, this review provides a comprehensive and critical synthesis of DL-based air quality forecasting, offering actionable insights for researchers, practitioners, and policymakers seeking transparent, generalizable, and policy-relevant prediction systems for environmental management and public health protection.

空气污染是一种严重的全球公共健康威胁,由暴露于有毒的环境污染物引起,包括颗粒物(PM)、硫氧化物(SOx)、氮氧化物(NOx)、臭氧(O₃)、一氧化碳(CO)和氨(NH₃)。传统的统计和确定性预测模型往往不能充分代表多种污染物、气象驱动因素和人为影响之间的非线性相互作用,这促使人们越来越多地采用深度学习(DL)方法。本系统综述综合了150多项同行评议研究的证据,这些研究跨越不同的地理区域,采用了广泛的深度学习架构,包括独立的、混合的和先进的时空模型。通过结构化的定量总结、基于排名的绩效比较和方法评估,该报告确定了领先的模型族,分析了污染物和特定水平的绩效趋势,并评估了在时空背景下的稳健性和普遍性。总体而言,深度学习模型通常优于传统方法,特别是在明确建模多源输入和时空依赖关系时。然而,文献仍然是碎片化的,研究集中在数据丰富的亚洲城市地区,数据集异构,评估方案不一致,透明度有限,外部有效性弱。解决这些限制需要标准化的预处理和基准实践,改进的可解释性和不确定性量化,以及开发具有全球代表性的数据集。新兴方向,包括混合、物理信息和生成式深度学习架构,为提高可靠性和操作部署提供了有希望的途径。总的来说,这篇综述提供了基于dl的空气质量预测的全面和关键的综合,为研究人员、从业者和政策制定者寻求透明的、通用的、与政策相关的环境管理和公共卫生保护预测系统提供了可操作的见解。
{"title":"Harnessing deep learning for air pollution forecasting: trends, techniques, and future prospects","authors":"Salvator Lawrence,&nbsp;Srimuruganandam Bhathmanabhan","doi":"10.1007/s10462-026-11496-8","DOIUrl":"10.1007/s10462-026-11496-8","url":null,"abstract":"<div><p>Air pollution is a serious global public health threat arising from exposure to toxic ambient pollutants, including particulate matter (PM), sulphur oxides (SOx), nitrogen oxides (NOx), ozone (O₃), carbon monoxide (CO), and ammonia (NH₃). Traditional statistical and deterministic forecasting models often fail to adequately represent nonlinear interactions among multiple pollutants, meteorological drivers, and anthropogenic influences, motivating the growing adoption of deep learning (DL) approaches. This systematic review synthesizes evidence from more than 150 peer-reviewed studies conducted across diverse geographical regions and employing a wide range of DL architectures, including standalone, hybrid, and advanced spatiotemporal models. Using structured quantitative summaries, rank-based performance comparisons, and methodological assessments, the review identifies leading model families, analyzes pollutant- and horizon-specific performance trends, and evaluates robustness and generalizability across spatial and temporal contexts. Overall, DL models generally outperform traditional approaches, particularly when multi-source inputs and spatiotemporal dependencies are explicitly modeled. Nevertheless, the literature remains fragmented, with a strong concentration of studies in data-rich urban regions of Asia, heterogeneous datasets, inconsistent evaluation protocols, limited transparency, and weak external validity. Addressing these limitations requires standardized preprocessing and benchmarking practices, improved explainability and uncertainty quantification, and the development of globally representative datasets. Emerging directions, including hybrid, physics-informed, and generative DL architectures, offer promising pathways to enhance reliability and operational deployment. Collectively, this review provides a comprehensive and critical synthesis of DL-based air quality forecasting, offering actionable insights for researchers, practitioners, and policymakers seeking transparent, generalizable, and policy-relevant prediction systems for environmental management and public health protection.</p></div>","PeriodicalId":8449,"journal":{"name":"Artificial Intelligence Review","volume":"59 3","pages":""},"PeriodicalIF":13.9,"publicationDate":"2026-02-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10462-026-11496-8.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147337305","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Dynamic causal modeling and efficient fuzzy temporal causality reasoning in industrial fault diagnosis: a C-DUCG approach 工业故障诊断中的动态因果建模和高效模糊时间因果推理:C-DUCG方法
IF 13.9 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-02-05 DOI: 10.1007/s10462-026-11509-6
Chunling Dong, Jianxiang Cao, Wenjing Liu

Online fault diagnosis is crucial for improving the reliability of safety–critical industrial systems. Fault diagnosis becomes increasingly complex if the time-variations, fuzziness, and uncertain causal relationships related to the various internal factors in the system are present. In view of the time-varying working conditions of industrial systems, as well as the fuzzy uncertainty of fault data and knowledge, a C-DUCG (Cubic dynamic uncertain causality graph) approach is proposed for fault diagnosis. Such a solution is intended to help develop a generative cubic causality graph modeling scheme and a FUTURE (fuzzy temporal causality reasoning) algorithm, both of which facilitate the representation and reasoning about complex fault situations (involving temporal causalities and uncertain evidence). In C-DUCG, the strategies of causality simplification and EELA (event-oriented early logical absorption) are proposed to mitigate the complexities of modeling and reasoning. Comparative experiments and a sequence of fault diagnosis tests on a nuclear power plant (NPP) simulator validate the efficiency, recall, accuracy, and interpretability of C-DUCG in large-scale dynamic systems. Experiments reveal that the proposed algorithm achieves 0.62 and 0.999 in terms of recall rate AC@1 and AC@k, respectively, with the average reasoning time being 10 ms, and the average time spent in NPP fault diagnosis tests is 9.42 ms.

在线故障诊断对于提高安全关键型工业系统的可靠性至关重要。如果存在与系统内部各种因素相关的时变、模糊性和不确定因果关系,则故障诊断将变得越来越复杂。针对工业系统的时变工况,以及故障数据和知识的模糊不确定性,提出了一种用于故障诊断的C-DUCG(三次动态不确定因果图)方法。这种解决方案旨在帮助开发生成三次因果图建模方案和FUTURE(模糊时间因果推理)算法,这两种算法都有助于对复杂故障情况(涉及时间因果关系和不确定证据)的表示和推理。在C-DUCG中,提出了因果关系简化和面向事件的早期逻辑吸收(EELA)策略来降低建模和推理的复杂性。对比实验和一系列核电厂仿真器故障诊断试验验证了C-DUCG在大型动态系统中的效率、召回率、准确性和可解释性。实验表明,该算法的召回率AC@1和AC@k分别达到0.62和0.999,平均推理时间为10 ms, NPP故障诊断测试平均耗时为9.42 ms。
{"title":"Dynamic causal modeling and efficient fuzzy temporal causality reasoning in industrial fault diagnosis: a C-DUCG approach","authors":"Chunling Dong,&nbsp;Jianxiang Cao,&nbsp;Wenjing Liu","doi":"10.1007/s10462-026-11509-6","DOIUrl":"10.1007/s10462-026-11509-6","url":null,"abstract":"<div>\u0000 \u0000 <p>Online fault diagnosis is crucial for improving the reliability of safety–critical industrial systems. Fault diagnosis becomes increasingly complex if the time-variations, fuzziness, and uncertain causal relationships related to the various internal factors in the system are present. In view of the time-varying working conditions of industrial systems, as well as the fuzzy uncertainty of fault data and knowledge, a C-DUCG (Cubic dynamic uncertain causality graph) approach is proposed for fault diagnosis. Such a solution is intended to help develop a generative cubic causality graph modeling scheme and a FUTURE (fuzzy temporal causality reasoning) algorithm, both of which facilitate the representation and reasoning about complex fault situations (involving temporal causalities and uncertain evidence). In C-DUCG, the strategies of causality simplification and EELA (event-oriented early logical absorption) are proposed to mitigate the complexities of modeling and reasoning. Comparative experiments and a sequence of fault diagnosis tests on a nuclear power plant (NPP) simulator validate the efficiency, recall, accuracy, and interpretability of C-DUCG in large-scale dynamic systems. Experiments reveal that the proposed algorithm achieves 0.62 and 0.999 in terms of recall rate AC@1 and AC@k, respectively, with the average reasoning time being 10 ms, and the average time spent in NPP fault diagnosis tests is 9.42 ms.</p>\u0000 </div>","PeriodicalId":8449,"journal":{"name":"Artificial Intelligence Review","volume":"59 3","pages":""},"PeriodicalIF":13.9,"publicationDate":"2026-02-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10462-026-11509-6.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147337077","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Advances in Alzheimer’s disease diagnosis with machine learning and deep learning techniques: a comprehensive review 机器学习和深度学习技术在阿尔茨海默病诊断中的进展:综述
IF 13.9 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-02-04 DOI: 10.1007/s10462-025-11476-4
Tazin Alam, Jahanara Suchi, Ankita Saha, Fahad Pathan, Aritra Das, Md. Mohsin Kabir, M. F. Mridha

Alzheimer’s disease (AD) is a chronic neurological disease and one of the main causes of dementia in around the world. Traditional diagnostic techniques have limits in terms of subjectivity and resource availability, despite the fact that early and accurate identification of AD is essential for efficient supervision. Deep learning (DL) and Machine Learning (ML) have emerged as powerful tools in medical imaging and have shown promising results in AD detection. This review provides a comprehensive analysis of the latest developments in ML and DL for AD diagnosis, covering essential data sets such as Alzheimer’s Disease Neuroimaging Initiative (ADNI), Open Access Series of Imaging Studies (OASIS), and Australian Imaging, Biomarkers and Lifestyle Study of Ageing (AIBL), as well as preprocessing techniques that enhance data quality. Here, we review some of the significant AD studies and investigate how ML and DL might assist researchers in making an early diagnosis more accurate.

阿尔茨海默病(AD)是一种慢性神经系统疾病,是世界范围内痴呆症的主要病因之一。传统的诊断技术在主观性和资源可用性方面存在局限性,尽管早期和准确识别阿尔茨海默病对于有效的监督至关重要。深度学习(DL)和机器学习(ML)已成为医学成像领域的强大工具,并在AD检测中显示出令人鼓舞的结果。本文综述了机器学习和深度学习在阿尔茨海默病诊断中的最新进展,涵盖了基本数据集,如阿尔茨海默病神经影像学倡议(ADNI)、开放获取系列影像学研究(OASIS)和澳大利亚成像、生物标志物和衰老生活方式研究(AIBL),以及提高数据质量的预处理技术。在这里,我们回顾了一些重要的阿尔茨海默病研究,并探讨ML和DL如何帮助研究人员做出更准确的早期诊断。
{"title":"Advances in Alzheimer’s disease diagnosis with machine learning and deep learning techniques: a comprehensive review","authors":"Tazin Alam,&nbsp;Jahanara Suchi,&nbsp;Ankita Saha,&nbsp;Fahad Pathan,&nbsp;Aritra Das,&nbsp;Md. Mohsin Kabir,&nbsp;M. F. Mridha","doi":"10.1007/s10462-025-11476-4","DOIUrl":"10.1007/s10462-025-11476-4","url":null,"abstract":"<div><p>Alzheimer’s disease (AD) is a chronic neurological disease and one of the main causes of dementia in around the world. Traditional diagnostic techniques have limits in terms of subjectivity and resource availability, despite the fact that early and accurate identification of AD is essential for efficient supervision. Deep learning (DL) and Machine Learning (ML) have emerged as powerful tools in medical imaging and have shown promising results in AD detection. This review provides a comprehensive analysis of the latest developments in ML and DL for AD diagnosis, covering essential data sets such as Alzheimer’s Disease Neuroimaging Initiative (ADNI), Open Access Series of Imaging Studies (OASIS), and Australian Imaging, Biomarkers and Lifestyle Study of Ageing (AIBL), as well as preprocessing techniques that enhance data quality. Here, we review some of the significant AD studies and investigate how ML and DL might assist researchers in making an early diagnosis more accurate.</p></div>","PeriodicalId":8449,"journal":{"name":"Artificial Intelligence Review","volume":"59 3","pages":""},"PeriodicalIF":13.9,"publicationDate":"2026-02-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10462-025-11476-4.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147336412","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
From data to diagnosis: a systematic review on AI-driven approaches to diabetes prediction 从数据到诊断:人工智能驱动的糖尿病预测方法的系统综述
IF 13.9 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-02-03 DOI: 10.1007/s10462-025-11485-3
Saleha Masood, Mousa Ahmad Al-Bashrawi, Muhammad Attique Khan, Yogesh K. Dwivedi

Diabetes mellitus is one of the most pressing global health challenges, and early prediction is critical to reducing complications, mortality, and healthcare costs. Conventional diagnostic tools remain limited, as they often rely on a small set of biomarkers and fail to capture lifestyle, genetic, or environmental risk factors. This review systematically evaluates how artificial intelligence (AI), including machine learning (ML) and deep learning (DL), enhances diabetes prediction by integrating multimodal data and improving clinical interpretability. Following PRISMA 2020 guidelines, a systematic search was conducted across PubMed, Scopus, IEEE Xplore, Web of Science, and ACM Digital Library for studies published between 2010 and 2024. Inclusion criteria required AI-based diabetes prediction models with reported performance metrics. From 2134 records, 155 studies met the criteria and were synthesized. AI models consistently outperformed conventional approaches (60–75% accuracy), Ensemble methods such as Random Forests and XGBoost achieved accuracy of 85–90% and AUC-ROC values > 0.90 (Abnoosian in BMC Bioinf 24:373, 2023. https://doi.org/10.1186/s12859-023-05465-z; Chen and Guestrin in Proceedings of the 22nd ACM SIGKDD international conference on knowledge discovery and data mining, Association for Computing Machinery, 2016. https://doi.org/10.1145/2939672.2939785). DL architectures showed notable strengths in unstructured data: CNNs achieved AUC > 0.95 in retinal image analysis (Gulshan et al. in JAMA 316(22):2402–2410, 2016. https://doi.org/10.1001/jama.2016.17216; Kermany et al. in Cell 172(5):1122–1131, 2018. https://doi.org/10.1016/j.cell.2018.02.010), while LSTMs reached 85–90% accuracy in continuous glucose monitoring predictions (De Bois et al. in Prediction-coherent LSTM-based recurrent neural network for safer glucose predictions in diabetic people, arXiv:2009.03722, 2020; Bian et al. in PLoS ONE 19(9):e0310084, 2024. https://doi.org/10.1371/journal.pone.0310084). Emerging methods such as federated learning enable privacy-preserving cross-institutional collaboration with comparable accuracy (~ 88%), while explainable AI techniques (e.g., SHAP, LIME, attention mechanisms) enhance transparency and clinical trust. Real-world case studies demonstrate improvements in early detection, reduced hospitalizations, and increased patient engagement when AI models are integrated into EHRs or mobile health apps. This review contributes a novel synthesis by combining methodological insights with clinical applications. The central takeaway is that AI-driven diabetes prediction offers significant advantages over traditional methods, but challenges in data quality, generalizability, fairness, and regulatory compliance remain. Addressing these will be essential to ensure safe, equitable, and clinically meaningful adoption in healthcare practice.

糖尿病是全球最紧迫的健康挑战之一,早期预测对于减少并发症、死亡率和医疗保健费用至关重要。传统的诊断工具仍然有限,因为它们通常依赖于一小部分生物标志物,无法捕捉到生活方式、遗传或环境风险因素。本文系统地评估了人工智能(AI),包括机器学习(ML)和深度学习(DL)如何通过整合多模态数据和提高临床可解释性来增强糖尿病预测。遵循PRISMA 2020指南,系统检索PubMed、Scopus、IEEE explore、Web of Science和ACM数字图书馆,检索2010年至2024年间发表的研究。纳入标准需要基于人工智能的糖尿病预测模型和报告的性能指标。从2134项记录中,155项研究符合标准并进行了综合。人工智能模型始终优于传统方法(准确率为60-75%),随机森林和XGBoost等集成方法的准确率为85-90%,AUC-ROC值为0.90 (Abnoosian, BMC bioinf24:373, 2023)。https://doi.org/10.1186/s12859 - 023 - 05465 - z;Chen和Guestrin,第22届ACM SIGKDD知识发现和数据挖掘国际会议论文集,计算机协会,2016。https://doi.org/10.1145/2939672.2939785)。深度学习架构在非结构化数据中表现出显著的优势:cnn在视网膜图像分析中达到了AUC >; 0.95 (Gulshan等人,JAMA 316(22): 2402-2410, 2016)。https://doi.org/10.1001/jama.2016.17216;Kermany et al.中国生物医学工程学报(英文版),2018(5):1122-1131。https://doi.org/10.1016/j.cell.2018.02.010),而LSTMs在连续血糖监测预测中准确率达到85-90% (De Bois等人在预测-coherent基于lstm的递归神经网络用于更安全的糖尿病人血糖预测,arXiv:2009.03722, 2020; Bian等人在PLoS ONE 19(9):e0310084, 2024)。https://doi.org/10.1371/journal.pone.0310084)。联邦学习等新兴方法使保护隐私的跨机构合作具有相当的准确性(约88%),而可解释的人工智能技术(例如,SHAP, LIME,注意力机制)提高了透明度和临床信任。实际案例研究表明,将人工智能模型集成到电子病历或移动健康应用程序中,可以改善早期发现,减少住院治疗,并提高患者参与度。这篇综述通过将方法学见解与临床应用相结合,提出了一种新的综合方法。研究的核心结论是,人工智能驱动的糖尿病预测比传统方法具有显著优势,但在数据质量、普遍性、公平性和法规遵从性方面仍然存在挑战。解决这些问题对于确保在医疗保健实践中安全、公平和临床有意义的采用至关重要。
{"title":"From data to diagnosis: a systematic review on AI-driven approaches to diabetes prediction","authors":"Saleha Masood,&nbsp;Mousa Ahmad Al-Bashrawi,&nbsp;Muhammad Attique Khan,&nbsp;Yogesh K. Dwivedi","doi":"10.1007/s10462-025-11485-3","DOIUrl":"10.1007/s10462-025-11485-3","url":null,"abstract":"<div>\u0000 \u0000 <p>Diabetes mellitus is one of the most pressing global health challenges, and early prediction is critical to reducing complications, mortality, and healthcare costs. Conventional diagnostic tools remain limited, as they often rely on a small set of biomarkers and fail to capture lifestyle, genetic, or environmental risk factors. This review systematically evaluates how artificial intelligence (AI), including machine learning (ML) and deep learning (DL), enhances diabetes prediction by integrating multimodal data and improving clinical interpretability. Following PRISMA 2020 guidelines, a systematic search was conducted across PubMed, Scopus, IEEE Xplore, Web of Science, and ACM Digital Library for studies published between 2010 and 2024. Inclusion criteria required AI-based diabetes prediction models with reported performance metrics. From 2134 records, 155 studies met the criteria and were synthesized. AI models consistently outperformed conventional approaches (60–75% accuracy), Ensemble methods such as Random Forests and XGBoost achieved accuracy of 85–90% and AUC-ROC values &gt; 0.90 (Abnoosian in BMC Bioinf 24:373, 2023. https://doi.org/10.1186/s12859-023-05465-z; Chen and Guestrin in Proceedings of the 22nd ACM SIGKDD international conference on knowledge discovery and data mining, Association for Computing Machinery, 2016. https://doi.org/10.1145/2939672.2939785). DL architectures showed notable strengths in unstructured data: CNNs achieved AUC &gt; 0.95 in retinal image analysis (Gulshan et al. in JAMA 316(22):2402–2410, 2016. https://doi.org/10.1001/jama.2016.17216; Kermany et al. in Cell 172(5):1122–1131, 2018. https://doi.org/10.1016/j.cell.2018.02.010), while LSTMs reached 85–90% accuracy in continuous glucose monitoring predictions (De Bois et al. in Prediction-coherent LSTM-based recurrent neural network for safer glucose predictions in diabetic people, arXiv:2009.03722, 2020; Bian et al. in PLoS ONE 19(9):e0310084, 2024. https://doi.org/10.1371/journal.pone.0310084). Emerging methods such as federated learning enable privacy-preserving cross-institutional collaboration with comparable accuracy (~ 88%), while explainable AI techniques (e.g., SHAP, LIME, attention mechanisms) enhance transparency and clinical trust. Real-world case studies demonstrate improvements in early detection, reduced hospitalizations, and increased patient engagement when AI models are integrated into EHRs or mobile health apps. This review contributes a novel synthesis by combining methodological insights with clinical applications. The central takeaway is that AI-driven diabetes prediction offers significant advantages over traditional methods, but challenges in data quality, generalizability, fairness, and regulatory compliance remain. Addressing these will be essential to ensure safe, equitable, and clinically meaningful adoption in healthcare practice.</p>\u0000 </div>","PeriodicalId":8449,"journal":{"name":"Artificial Intelligence Review","volume":"59 3","pages":""},"PeriodicalIF":13.9,"publicationDate":"2026-02-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10462-025-11485-3.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147336425","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Artificial Intelligence Review
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1