首页 > 最新文献

Artificial Intelligence Review最新文献

英文 中文
Deep dive into RNA: a systematic literature review on RNA structure prediction using machine learning methods 深入研究 RNA:利用机器学习方法预测 RNA 结构的系统文献综述
IF 10.7 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-08-15 DOI: 10.1007/s10462-024-10910-3
Michał Budnik, Jakub Wawrzyniak, Łukasz Grala, Miłosz Kadziński, Natalia Szóstak

The discovery of non-coding RNAs (ncRNAs) has expanded our comprehension of RNAs’ inherent nature and capabilities. The intricate three-dimensional structures assumed by RNAs dictate their specific functions and molecular interactions. However, the limited number of mapped structures, partly due to experimental constraints of methods such as nuclear magnetic resonance (NMR), highlights the importance of in silico prediction solutions. This is particularly crucial in potential applications in therapeutic drug discovery. In this context, machine learning (ML) methods have emerged as prominent candidates, having previously demonstrated prowess in solving complex challenges across various domains. This review focuses on analyzing the development of ML-based solutions for RNA structure prediction, specifically oriented toward recent advancements in the deep learning (DL) domain. A systematic analysis of 33 works reveals insights into the representation of RNA structures, secondary structure motifs, and tertiary interactions. The review highlights current trends in ML methods used for RNA structure prediction, demonstrates the growing research involvement in this field, and summarizes the most valuable findings.

非编码 RNA(ncRNA)的发现拓展了我们对 RNA 固有性质和功能的理解。RNA 复杂的三维结构决定了它们的特定功能和分子相互作用。然而,部分由于核磁共振(NMR)等方法的实验限制,绘制的结构图数量有限,这凸显了硅预测解决方案的重要性。这对于治疗药物发现的潜在应用尤为重要。在这种情况下,机器学习(ML)方法已成为重要的候选方法,它们以前曾在解决各个领域的复杂挑战方面表现出卓越的能力。本综述重点分析基于 ML 的 RNA 结构预测解决方案的发展情况,特别是深度学习(DL)领域的最新进展。通过对 33 篇论文的系统分析,我们可以深入了解 RNA 结构、二级结构主题和三级相互作用的表征。综述重点介绍了用于 RNA 结构预测的 ML 方法的当前趋势,展示了该领域日益增长的研究参与,并总结了最有价值的发现。
{"title":"Deep dive into RNA: a systematic literature review on RNA structure prediction using machine learning methods","authors":"Michał Budnik,&nbsp;Jakub Wawrzyniak,&nbsp;Łukasz Grala,&nbsp;Miłosz Kadziński,&nbsp;Natalia Szóstak","doi":"10.1007/s10462-024-10910-3","DOIUrl":"10.1007/s10462-024-10910-3","url":null,"abstract":"<div><p>The discovery of non-coding RNAs (ncRNAs) has expanded our comprehension of RNAs’ inherent nature and capabilities. The intricate three-dimensional structures assumed by RNAs dictate their specific functions and molecular interactions. However, the limited number of mapped structures, partly due to experimental constraints of methods such as nuclear magnetic resonance (NMR), highlights the importance of in silico prediction solutions. This is particularly crucial in potential applications in therapeutic drug discovery. In this context, machine learning (ML) methods have emerged as prominent candidates, having previously demonstrated prowess in solving complex challenges across various domains. This review focuses on analyzing the development of ML-based solutions for RNA structure prediction, specifically oriented toward recent advancements in the deep learning (DL) domain. A systematic analysis of 33 works reveals insights into the representation of RNA structures, secondary structure motifs, and tertiary interactions. The review highlights current trends in ML methods used for RNA structure prediction, demonstrates the growing research involvement in this field, and summarizes the most valuable findings.</p></div>","PeriodicalId":8449,"journal":{"name":"Artificial Intelligence Review","volume":null,"pages":null},"PeriodicalIF":10.7,"publicationDate":"2024-08-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10462-024-10910-3.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142222104","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Forensic research of satellite images forgery: a comprehensive survey 卫星图像伪造的法医研究:全面调查
IF 10.7 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-08-14 DOI: 10.1007/s10462-024-10909-w
Xiangling Ding, Yuchen Nie, Jizhou Yao, Jia Tang, Yubo Lang

Satellite imagery is a significant and attractive area in remote sensing applications, widely applied in monitoring, managing, and tracking natural disasters. Due to the proliferation of commercial satellites, there is an increasing availability of high-resolution satellite images. However, the ubiquity of image editing tools and the advancement of image processing technologies have made satellite image forgery relatively easy, allowing for the arbitrary addition, removal, or modification of target objects. In recent years, satellite image forgery has caused significant negative effects and potential threats to the nation, society, and individuals, drawing the attention of many scholars. Although forensics of satellite image tampering is an emerging research field that offers new insights into the field of information security, there has been a scarcity of comprehensive surveys in this area. This paper aims to fill this gap and investigates recent advances in satellite image forensics, focusing on tampering strategies and forensic methodologies. First, we discuss the concept of satellite images, the definition of satellite image forgery from global and local perspectives, and the datasets commonly used for satellite image forensics. We then detail each tampering detection and localization method, including their characteristics, advantages, disadvantages, and performance in detection or localization across various notable datasets. We also compare some representative forensic networks using evaluation metrics and public datasets. Finally, the anticipated future directions for satellite image forgery forensics are discussed.

卫星图像是遥感应用中一个重要而有吸引力的领域,广泛应用于监测、管理和跟踪自然灾害。由于商业卫星的普及,高分辨率卫星图像的可用性越来越高。然而,图像编辑工具的普及和图像处理技术的进步使得卫星图像的伪造变得相对容易,可以任意添加、删除或修改目标对象。近年来,卫星图像伪造给国家、社会和个人带来了巨大的负面影响和潜在威胁,引起了众多学者的关注。虽然卫星图像篡改取证是一个新兴的研究领域,为信息安全领域提供了新的见解,但该领域的全面调查一直很少。本文旨在填补这一空白,研究卫星图像取证的最新进展,重点关注篡改策略和取证方法。首先,我们讨论了卫星图像的概念、从全球和本地角度对卫星图像伪造的定义以及卫星图像取证常用的数据集。然后,我们详细介绍了每种篡改检测和定位方法,包括其特点、优缺点以及在各种著名数据集上的检测或定位性能。我们还使用评估指标和公共数据集对一些具有代表性的取证网络进行了比较。最后,讨论了卫星图像伪造取证的预期未来方向。
{"title":"Forensic research of satellite images forgery: a comprehensive survey","authors":"Xiangling Ding,&nbsp;Yuchen Nie,&nbsp;Jizhou Yao,&nbsp;Jia Tang,&nbsp;Yubo Lang","doi":"10.1007/s10462-024-10909-w","DOIUrl":"10.1007/s10462-024-10909-w","url":null,"abstract":"<div><p>Satellite imagery is a significant and attractive area in remote sensing applications, widely applied in monitoring, managing, and tracking natural disasters. Due to the proliferation of commercial satellites, there is an increasing availability of high-resolution satellite images. However, the ubiquity of image editing tools and the advancement of image processing technologies have made satellite image forgery relatively easy, allowing for the arbitrary addition, removal, or modification of target objects. In recent years, satellite image forgery has caused significant negative effects and potential threats to the nation, society, and individuals, drawing the attention of many scholars. Although forensics of satellite image tampering is an emerging research field that offers new insights into the field of information security, there has been a scarcity of comprehensive surveys in this area. This paper aims to fill this gap and investigates recent advances in satellite image forensics, focusing on tampering strategies and forensic methodologies. First, we discuss the concept of satellite images, the definition of satellite image forgery from global and local perspectives, and the datasets commonly used for satellite image forensics. We then detail each tampering detection and localization method, including their characteristics, advantages, disadvantages, and performance in detection or localization across various notable datasets. We also compare some representative forensic networks using evaluation metrics and public datasets. Finally, the anticipated future directions for satellite image forgery forensics are discussed.</p></div>","PeriodicalId":8449,"journal":{"name":"Artificial Intelligence Review","volume":null,"pages":null},"PeriodicalIF":10.7,"publicationDate":"2024-08-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10462-024-10909-w.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142222107","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A review of deep learning models and online healthcare databases for electronic health records and their use for health prediction 深度学习模型和电子健康记录在线医疗数据库及其在健康预测中的应用综述
IF 10.7 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-08-13 DOI: 10.1007/s10462-024-10876-2
Nurul Athirah Nasarudin, Fatma Al Jasmi, Richard O. Sinnott, Nazar Zaki, Hany Al Ashwal, Elfadil A. Mohamed, Mohd Saberi Mohamad

A fundamental obstacle to healthcare transformation continues to be the acquisition of knowledge and insightful data from complex, high dimensional, and heterogeneous biological data. As technology has improved, a wide variety of data sources, including omics data, imaging data, and health records, have been available for use in healthcare research contexts. Electronic health records (EHRs), which are digitalized versions of medical records, have given researchers a significant chance to create computational methods for analyzing healthcare data. EHR systems typically keep track of all the data relating to a patient’s medical history, including clinical notes, demographic background, and diagnosis details. EHR data can offer valuable insights and support doctors in making better decisions related to disease and diagnostic forecasts. As a result, several academics use deep learning to forecast diseases and track health trajectories in EHR. Recent advances in deep learning technology have produced innovative and practical paradigms for building end-to-end learning models. However, scholars have limited access to online HER databases, and there is an inherent need to address this issue. This research examines deep learning models, their architectures, and readily accessible EHR online databases. The goal of this paper is to examine how various architectures, models, and databases differ in terms of features and usability. It is anticipated that the outcomes of this review will lead to the development of more robust deep learning models that facilitate medical decision-making processes based on EHR data and inform efforts to support the selection of architectures, models, and databases for specific research purposes.

从复杂、高维和异构的生物数据中获取知识和有洞察力的数据,仍然是医疗保健转型的根本障碍。随着技术的进步,各种数据源,包括全息数据、成像数据和健康记录,都可用于医疗保健研究。电子健康记录(EHR)是医疗记录的数字化版本,它为研究人员提供了创建医疗数据分析计算方法的重要机会。电子病历系统通常会记录与病人病史有关的所有数据,包括临床笔记、人口统计背景和诊断细节。电子病历数据可以提供有价值的见解,帮助医生做出更好的疾病决策和诊断预测。因此,一些学者利用深度学习来预测疾病并跟踪电子病历中的健康轨迹。深度学习技术的最新进展为建立端到端学习模型提供了创新而实用的范例。然而,学者们对在线 HER 数据库的访问有限,因此有必要解决这一问题。本研究探讨了深度学习模型、其架构以及易于访问的电子病历在线数据库。本文的目的是研究各种架构、模型和数据库在功能和可用性方面有何不同。预计本综述的结果将有助于开发更强大的深度学习模型,以促进基于电子病历数据的医疗决策过程,并为支持为特定研究目的选择架构、模型和数据库的工作提供信息。
{"title":"A review of deep learning models and online healthcare databases for electronic health records and their use for health prediction","authors":"Nurul Athirah Nasarudin,&nbsp;Fatma Al Jasmi,&nbsp;Richard O. Sinnott,&nbsp;Nazar Zaki,&nbsp;Hany Al Ashwal,&nbsp;Elfadil A. Mohamed,&nbsp;Mohd Saberi Mohamad","doi":"10.1007/s10462-024-10876-2","DOIUrl":"10.1007/s10462-024-10876-2","url":null,"abstract":"<div><p>A fundamental obstacle to healthcare transformation continues to be the acquisition of knowledge and insightful data from complex, high dimensional, and heterogeneous biological data. As technology has improved, a wide variety of data sources, including omics data, imaging data, and health records, have been available for use in healthcare research contexts. Electronic health records (EHRs), which are digitalized versions of medical records, have given researchers a significant chance to create computational methods for analyzing healthcare data. EHR systems typically keep track of all the data relating to a patient’s medical history, including clinical notes, demographic background, and diagnosis details. EHR data can offer valuable insights and support doctors in making better decisions related to disease and diagnostic forecasts. As a result, several academics use deep learning to forecast diseases and track health trajectories in EHR. Recent advances in deep learning technology have produced innovative and practical paradigms for building end-to-end learning models. However, scholars have limited access to online HER databases, and there is an inherent need to address this issue. This research examines deep learning models, their architectures, and readily accessible EHR online databases. The goal of this paper is to examine how various architectures, models, and databases differ in terms of features and usability. It is anticipated that the outcomes of this review will lead to the development of more robust deep learning models that facilitate medical decision-making processes based on EHR data and inform efforts to support the selection of architectures, models, and databases for specific research purposes.</p></div>","PeriodicalId":8449,"journal":{"name":"Artificial Intelligence Review","volume":null,"pages":null},"PeriodicalIF":10.7,"publicationDate":"2024-08-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10462-024-10876-2.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142222084","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Deciphering Arabic question: a dedicated survey on Arabic question analysis methods, challenges, limitations and future pathways 解密阿拉伯语问题:关于阿拉伯语问题分析方法、挑战、局限性和未来途径的专项调查
IF 10.7 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-08-13 DOI: 10.1007/s10462-024-10880-6
Mariam Essam, Mohanad A. Deif, Rania Elgohary

This survey reviews different research on question analysis, including other comparative studies of question analysis approaches and an evaluation of the questions by different NLP techniques that are used in question interpretation and categorization. Among these key findings noted includes the assessment of deep learning models such as M-BiGRU-CNN and M-TF-IDF, which come with high precision and accuracy when applied with the effectiveness of use in dealing with the complexities involved in a language. Some of the most mature machine learning algorithms, for example, SVM or logistic regression, remain powerful models, especially on the classification task, meaning that the latter continues to be relevant. This study further underlines the applicability of rule-based or hybrid methodologies in certain linguistic situations, and it must be said that custom design solutions are required. We could recommend, on this basis, directing future work towards the integration of these hybrid systems and towards the definition of more general methodologies of evaluation that are in line with the constant evolution of NLP technologies. It revealed that the underlying challenges and barriers in the domain are very complex syntactic and dialectic variations, unavailability of software tools, very critical standardization in Arabic datasets, benchmark creation, handling of translated data, and the integration of Large Language Models (LLMs). The paper discusses the lack of identity and processing of such structures through online systems for comparison. This comprehensive review highlights not only the diversified potential for the capabilities of NLP techniques in refining question analysis but also the potential way of great promises for further enhancements and improvements in this progressive domain.

本调查回顾了有关问题分析的各种研究,包括对问题分析方法的其他比较研究,以及通过用于问题解释和分类的不同 NLP 技术对问题进行的评估。在这些主要研究成果中,包括对 M-BiGRU-CNN 和 M-TF-IDF 等深度学习模型的评估,这些模型在应用时具有较高的精确度和准确性,并能有效处理语言中涉及的复杂问题。一些最成熟的机器学习算法,如 SVM 或逻辑回归,仍然是功能强大的模型,尤其是在分类任务中,这意味着后者仍然具有相关性。本研究进一步强调了基于规则的方法或混合方法在某些语言情况下的适用性,必须指出的是,需要定制设计解决方案。在此基础上,我们建议将未来的工作导向这些混合系统的整合,以及与 NLP 技术的不断发展相适应的更通用的评估方法的定义。论文揭示了该领域的基本挑战和障碍,包括非常复杂的句法和方言变化、软件工具的不可用性、阿拉伯语数据集的标准化、基准创建、翻译数据的处理以及大型语言模型(LLM)的整合。本文讨论了缺乏通过在线系统对此类结构进行识别和处理以进行比较的问题。这篇全面综述不仅强调了 NLP 技术在改进问题分析方面的多样化潜力,而且还强调了在这一进步领域进一步提高和改进的潜在途径。
{"title":"Deciphering Arabic question: a dedicated survey on Arabic question analysis methods, challenges, limitations and future pathways","authors":"Mariam Essam,&nbsp;Mohanad A. Deif,&nbsp;Rania Elgohary","doi":"10.1007/s10462-024-10880-6","DOIUrl":"10.1007/s10462-024-10880-6","url":null,"abstract":"<div><p>This survey reviews different research on question analysis, including other comparative studies of question analysis approaches and an evaluation of the questions by different NLP techniques that are used in question interpretation and categorization. Among these key findings noted includes the assessment of deep learning models such as M-BiGRU-CNN and M-TF-IDF, which come with high precision and accuracy when applied with the effectiveness of use in dealing with the complexities involved in a language. Some of the most mature machine learning algorithms, for example, SVM or logistic regression, remain powerful models, especially on the classification task, meaning that the latter continues to be relevant. This study further underlines the applicability of rule-based or hybrid methodologies in certain linguistic situations, and it must be said that custom design solutions are required. We could recommend, on this basis, directing future work towards the integration of these hybrid systems and towards the definition of more general methodologies of evaluation that are in line with the constant evolution of NLP technologies. It revealed that the underlying challenges and barriers in the domain are very complex syntactic and dialectic variations, unavailability of software tools, very critical standardization in Arabic datasets, benchmark creation, handling of translated data, and the integration of Large Language Models (LLMs). The paper discusses the lack of identity and processing of such structures through online systems for comparison. This comprehensive review highlights not only the diversified potential for the capabilities of NLP techniques in refining question analysis but also the potential way of great promises for further enhancements and improvements in this progressive domain.</p></div>","PeriodicalId":8449,"journal":{"name":"Artificial Intelligence Review","volume":null,"pages":null},"PeriodicalIF":10.7,"publicationDate":"2024-08-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10462-024-10880-6.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142222106","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A brief review of hypernetworks in deep learning 深度学习中的超网络简评
IF 10.7 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-08-13 DOI: 10.1007/s10462-024-10862-8
Vinod Kumar Chauhan, Jiandong Zhou, Ping Lu, Soheila Molaei, David A. Clifton

Hypernetworks, or hypernets for short, are neural networks that generate weights for another neural network, known as the target network. They have emerged as a powerful deep learning technique that allows for greater flexibility, adaptability, dynamism, faster training, information sharing, and model compression. Hypernets have shown promising results in a variety of deep learning problems, including continual learning, causal inference, transfer learning, weight pruning, uncertainty quantification, zero-shot learning, natural language processing, and reinforcement learning. Despite their success across different problem settings, there is currently no comprehensive review available to inform researchers about the latest developments and to assist in utilizing hypernets. To fill this gap, we review the progress in hypernets. We present an illustrative example of training deep neural networks using hypernets and propose categorizing hypernets based on five design criteria: inputs, outputs, variability of inputs and outputs, and the architecture of hypernets. We also review applications of hypernets across different deep learning problem settings, followed by a discussion of general scenarios where hypernets can be effectively employed. Finally, we discuss the challenges and future directions that remain underexplored in the field of hypernets. We believe that hypernetworks have the potential to revolutionize the field of deep learning. They offer a new way to design and train neural networks, and they have the potential to improve the performance of deep learning models on a variety of tasks. Through this review, we aim to inspire further advancements in deep learning through hypernetworks.

超网络,简称超网,是一种为另一个神经网络(称为目标网络)生成权重的神经网络。超网络已成为一种强大的深度学习技术,可实现更高的灵活性、适应性、动态性、更快的训练、信息共享和模型压缩。超网络在持续学习、因果推理、迁移学习、权重剪枝、不确定性量化、零点学习、自然语言处理和强化学习等各种深度学习问题上都取得了可喜的成果。尽管它们在不同的问题设置中都取得了成功,但目前还没有全面的综述可供研究人员了解最新进展,并帮助他们利用超网络。为了填补这一空白,我们回顾了超网络的进展。我们介绍了一个使用超网络训练深度神经网络的示例,并建议根据五个设计标准对超网络进行分类:输入、输出、输入和输出的可变性以及超网络的架构。我们还回顾了超网络在不同深度学习问题设置中的应用,随后讨论了可以有效使用超网络的一般场景。最后,我们讨论了超网络领域仍未充分探索的挑战和未来方向。我们相信,超网络有可能彻底改变深度学习领域。它们提供了一种设计和训练神经网络的新方法,并有可能提高深度学习模型在各种任务中的性能。通过这篇综述,我们希望通过超网络进一步推动深度学习的发展。
{"title":"A brief review of hypernetworks in deep learning","authors":"Vinod Kumar Chauhan,&nbsp;Jiandong Zhou,&nbsp;Ping Lu,&nbsp;Soheila Molaei,&nbsp;David A. Clifton","doi":"10.1007/s10462-024-10862-8","DOIUrl":"10.1007/s10462-024-10862-8","url":null,"abstract":"<div><p>Hypernetworks, or hypernets for short, are neural networks that generate weights for another neural network, known as the target network. They have emerged as a powerful deep learning technique that allows for greater flexibility, adaptability, dynamism, faster training, information sharing, and model compression. Hypernets have shown promising results in a variety of deep learning problems, including continual learning, causal inference, transfer learning, weight pruning, uncertainty quantification, zero-shot learning, natural language processing, and reinforcement learning. Despite their success across different problem settings, there is currently no comprehensive review available to inform researchers about the latest developments and to assist in utilizing hypernets. To fill this gap, we review the progress in hypernets. We present an illustrative example of training deep neural networks using hypernets and propose categorizing hypernets based on five design criteria: inputs, outputs, variability of inputs and outputs, and the architecture of hypernets. We also review applications of hypernets across different deep learning problem settings, followed by a discussion of general scenarios where hypernets can be effectively employed. Finally, we discuss the challenges and future directions that remain underexplored in the field of hypernets. We believe that hypernetworks have the potential to revolutionize the field of deep learning. They offer a new way to design and train neural networks, and they have the potential to improve the performance of deep learning models on a variety of tasks. Through this review, we aim to inspire further advancements in deep learning through hypernetworks.</p></div>","PeriodicalId":8449,"journal":{"name":"Artificial Intelligence Review","volume":null,"pages":null},"PeriodicalIF":10.7,"publicationDate":"2024-08-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10462-024-10862-8.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142222103","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Comparison of the impact of dimensionality reduction and data splitting on classification performance in credit risk assessment 比较降维和数据拆分对信用风险评估分类性能的影响
IF 10.7 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-08-13 DOI: 10.1007/s10462-024-10904-1
Cem Bulut, Emel Arslan

Credit risk assessment (CRA) plays an important role in credit decision-making process of financial institutions. Today, developing big data analysis and machine learning methods have marked a new era in credit risk estimation. In recent years, using machine learning methods in credit risk estimation has emerged as an alternative method for financial institutions. The past demographic and financial data of the person whose CRA will be performed is important for creating an automatic artificial intelligence credit score prediction model based on machine learning. It is also necessary to use features correctly to create accurate machine learning models. This article aims to investigate the effects of dimensionality reduction and data splitting steps on the performance of classification algorithms widely used in the literature. In our study, dimensionality reduction was performed using Principal Component Analysis (PCA). Random Forest (RF), Logistic Regression (LR), Decision Tree (DT), Naive Bayes (NB) algorithms were chosen for classification. Percentage splitting (PER, 66–34%) and k-fold (k = 10) cross-validation techniques were used when dividing the data set into training and test data. The results obtained were evaluated with accuracy, recall, F1 score, precision, and AUC metrics. German data set was used in this study. The effect of data splitting and dimension reduction on the classification of CRA systems was examined. The highest ACC in PER and CV data splitting was obtained with the RF algorithm. Using data splitting methods and PCA, the highest accuracy was observed with RF and the highest AUC with NB, with 13 PCs in which 80% of the variance was obtained. As a result, the data set consisting of a total of 20 features, expressed by 13 PCs, achieved similar or higher success than the results obtained from the original data set.

信用风险评估(CRA)在金融机构的信贷决策过程中发挥着重要作用。如今,大数据分析和机器学习方法的发展标志着信用风险评估进入了一个新时代。近年来,使用机器学习方法进行信用风险评估已成为金融机构的一种替代方法。要创建一个基于机器学习的人工智能信用评分自动预测模型,将被执行 CRA 的人过去的人口和财务数据非常重要。要创建准确的机器学习模型,还必须正确使用特征。本文旨在研究降维和数据拆分步骤对文献中广泛使用的分类算法性能的影响。在我们的研究中,采用主成分分析法(PCA)进行降维。分类算法选择了随机森林(RF)、逻辑回归(LR)、决策树(DT)和奈夫贝叶斯(NB)算法。在将数据集分为训练数据和测试数据时,使用了百分比分割(PER,66-34%)和 k 倍(k = 10)交叉验证技术。获得的结果通过准确率、召回率、F1 分数、精确度和 AUC 指标进行评估。本研究使用的是德国数据集。研究考察了数据分割和降维对 CRA 系统分类的影响。RF 算法在 PER 和 CV 数据拆分中获得了最高的 ACC。在使用数据拆分方法和 PCA 时,RF 算法的准确率最高,NB 算法的 AUC 最高,有 13 个 PC,其中 80% 的方差是通过 PC 获得的。因此,由 13 个 PC 表示的共 20 个特征组成的数据集取得了与原始数据集相似或更高的成功率。
{"title":"Comparison of the impact of dimensionality reduction and data splitting on classification performance in credit risk assessment","authors":"Cem Bulut,&nbsp;Emel Arslan","doi":"10.1007/s10462-024-10904-1","DOIUrl":"10.1007/s10462-024-10904-1","url":null,"abstract":"<div><p>Credit risk assessment (CRA) plays an important role in credit decision-making process of financial institutions. Today, developing big data analysis and machine learning methods have marked a new era in credit risk estimation. In recent years, using machine learning methods in credit risk estimation has emerged as an alternative method for financial institutions. The past demographic and financial data of the person whose CRA will be performed is important for creating an automatic artificial intelligence credit score prediction model based on machine learning. It is also necessary to use features correctly to create accurate machine learning models. This article aims to investigate the effects of dimensionality reduction and data splitting steps on the performance of classification algorithms widely used in the literature. In our study, dimensionality reduction was performed using Principal Component Analysis (PCA). Random Forest (RF), Logistic Regression (LR), Decision Tree (DT), Naive Bayes (NB) algorithms were chosen for classification. Percentage splitting (PER, 66–34%) and k-fold (k = 10) cross-validation techniques were used when dividing the data set into training and test data. The results obtained were evaluated with accuracy, recall, F1 score, precision, and AUC metrics. German data set was used in this study. The effect of data splitting and dimension reduction on the classification of CRA systems was examined. The highest ACC in PER and CV data splitting was obtained with the RF algorithm. Using data splitting methods and PCA, the highest accuracy was observed with RF and the highest AUC with NB, with 13 PCs in which 80% of the variance was obtained. As a result, the data set consisting of a total of 20 features, expressed by 13 PCs, achieved similar or higher success than the results obtained from the original data set.</p></div>","PeriodicalId":8449,"journal":{"name":"Artificial Intelligence Review","volume":null,"pages":null},"PeriodicalIF":10.7,"publicationDate":"2024-08-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10462-024-10904-1.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142222108","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Machine learning-based bee recognition and tracking for advancing insect behavior research 基于机器学习的蜜蜂识别和跟踪技术,促进昆虫行为研究
IF 10.7 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-08-12 DOI: 10.1007/s10462-024-10879-z
Erez Rozenbaum, Tammar Shrot, Hadassa Daltrophe, Yehuda Kunya, Sharoni Shafir

The study of insect behavior, particularly that of honey bees, has a broad scope and significance. Tracking bee flying patterns grants much helpful information about bee behavior. However, tracking a small yet fast-moving object, such as a bee, is difficult. Hence, we present artificial intelligence, machine-learning-based bee recognition, and tracking systems to assist the researcher in studying the bee’s behavior. To develop a machine learning system, a labeled database is required for model training. To address this, we implemented an automated system for analyzing and labeling bee videos. This labeled database served as the foundation for two distinct bee-tracking solutions. The first solution (planar bee tracking system) tracked individual bees in closed mazes using a neural network. The second solution (spatial bee tracking system) utilized a neural network and a tracking algorithm to identify and track flying bees in open environments. Both systems tackle the challenge of tracking small-bodied creatures with rapid and diverse movement patterns. Although we applied these systems to entomological cognition research in this paper, their relevance extends to general insect research and developing tracking solutions for small organisms with swift movements. We present the complete architecture and detailed methodologies to facilitate the utilization of these models in future research endeavors. Our approach is a simple and inexpensive method that contributes to the growing number of image-analysis tools used for tracking animal movement, with future potential applications under less sterile field conditions. The tools presented in this paper could assist the study of movement ecology, specifically in insects, by providing accurate movement specifications. Following the movement of pollinators or natural enemies, for example, greatly contributes to the study of pollination or biological control, respectively, in natural and agro-ecosystems.

昆虫行为研究,尤其是蜜蜂行为研究,范围广泛,意义重大。跟踪蜜蜂的飞行模式可以获得许多有关蜜蜂行为的有用信息。然而,跟踪蜜蜂这样体积小但移动速度快的物体却很困难。因此,我们提出了基于人工智能、机器学习的蜜蜂识别和跟踪系统,以帮助研究人员研究蜜蜂的行为。要开发机器学习系统,需要一个标注数据库来进行模型训练。为此,我们实施了一套自动系统,用于分析和标注蜜蜂视频。该标记数据库是两种不同蜜蜂跟踪解决方案的基础。第一种方案(平面蜜蜂跟踪系统)使用神经网络在封闭迷宫中跟踪单个蜜蜂。第二种解决方案(空间蜜蜂跟踪系统)利用神经网络和跟踪算法来识别和跟踪开放环境中的飞行蜜蜂。这两个系统都解决了追踪具有快速和多样化运动模式的小型生物的难题。虽然我们在本文中将这些系统应用于昆虫学认知研究,但它们的相关性可扩展到一般昆虫研究以及为具有快速运动的小型生物开发追踪解决方案。我们介绍了完整的结构和详细的方法,以方便在未来的研究工作中使用这些模型。我们的方法是一种简单而廉价的方法,为越来越多的用于追踪动物运动的图像分析工具做出了贡献,未来有可能应用于无菌条件较差的野外环境。本文介绍的工具可以通过提供准确的运动规格来帮助研究运动生态学,特别是昆虫的运动生态学。例如,跟踪传粉昆虫或天敌的移动,可极大地促进对自然和农业生态系统中传粉或生物控制的研究。
{"title":"Machine learning-based bee recognition and tracking for advancing insect behavior research","authors":"Erez Rozenbaum,&nbsp;Tammar Shrot,&nbsp;Hadassa Daltrophe,&nbsp;Yehuda Kunya,&nbsp;Sharoni Shafir","doi":"10.1007/s10462-024-10879-z","DOIUrl":"10.1007/s10462-024-10879-z","url":null,"abstract":"<div><p>The study of insect behavior, particularly that of honey bees, has a broad scope and significance. Tracking bee flying patterns grants much helpful information about bee behavior. However, tracking a small yet fast-moving object, such as a bee, is difficult. Hence, we present artificial intelligence, machine-learning-based bee recognition, and tracking systems to assist the researcher in studying the bee’s behavior. To develop a machine learning system, a labeled database is required for model training. To address this, we implemented an automated system for analyzing and labeling bee videos. This labeled database served as the foundation for two distinct bee-tracking solutions. The first solution (planar bee tracking system) tracked individual bees in closed mazes using a neural network. The second solution (spatial bee tracking system) utilized a neural network and a tracking algorithm to identify and track flying bees in open environments. Both systems tackle the challenge of tracking small-bodied creatures with rapid and diverse movement patterns. Although we applied these systems to entomological cognition research in this paper, their relevance extends to general insect research and developing tracking solutions for small organisms with swift movements. We present the complete architecture and detailed methodologies to facilitate the utilization of these models in future research endeavors. Our approach is a simple and inexpensive method that contributes to the growing number of image-analysis tools used for tracking animal movement, with future potential applications under less sterile field conditions. The tools presented in this paper could assist the study of movement ecology, specifically in insects, by providing accurate movement specifications. Following the movement of pollinators or natural enemies, for example, greatly contributes to the study of pollination or biological control, respectively, in natural and agro-ecosystems.</p></div>","PeriodicalId":8449,"journal":{"name":"Artificial Intelligence Review","volume":null,"pages":null},"PeriodicalIF":10.7,"publicationDate":"2024-08-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10462-024-10879-z.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141933377","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An in-depth survey of the artificial gorilla troops optimizer: outcomes, variations, and applications 人工猩猩部队优化器的深入调查:成果、变化和应用
IF 10.7 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-08-12 DOI: 10.1007/s10462-024-10838-8
Abdelazim G. Hussien, Anas Bouaouda, Abdullah Alzaqebah, Sumit Kumar, Gang Hu, Heming Jia

A recently developed algorithm inspired by natural processes, known as the Artificial Gorilla Troops Optimizer (GTO), boasts a straightforward structure, unique stabilizing features, and notably high effectiveness. Its primary objective is to efficiently find solutions for a wide array of challenges, whether they involve constraints or not. The GTO takes its inspiration from the behavior of Gorilla Troops in the natural world. To emulate the impact of gorillas at each stage of the search process, the GTO employs a flexible weighting mechanism rooted in its concept. Its exceptional qualities, including its independence from derivatives, lack of parameters, user-friendliness, adaptability, and simplicity, have resulted in its rapid adoption for addressing various optimization challenges. This review is dedicated to the examination and discussion of the foundational research that forms the basis of the GTO. It delves into the evolution of this algorithm, drawing insights from 112 research studies that highlight its effectiveness. Additionally, it explores proposed enhancements to the GTO’s behavior, with a specific focus on aligning the geometry of the search area with real-world optimization problems. The review also introduces the GTO solver, providing details about its identification and organization, and demonstrates its application in various optimization scenarios. Furthermore, it provides a critical assessment of the convergence behavior while addressing the primary limitation of the GTO. In conclusion, this review summarizes the key findings of the study and suggests potential avenues for future advancements and adaptations related to the GTO.

最近,一种受自然过程启发而开发的算法被称为 "人工大猩猩部队优化器"(GTO),它具有简单明了的结构、独特的稳定功能和显著的高效性。它的主要目标是为各种挑战高效地找到解决方案,无论这些挑战是否涉及约束条件。GTO 的灵感来源于自然界中大猩猩部队的行为。为了模仿大猩猩在搜索过程中每个阶段的影响,GTO 采用了植根于其概念的灵活加权机制。它具有独立于导数、无需参数、用户友好、适应性强和简单易行等卓越品质,因此在应对各种优化挑战时被迅速采用。本综述专门研究和讨论构成 GTO 基础的基础研究。它深入探讨了该算法的演变过程,从 112 项研究中汲取了深刻的见解,突出了该算法的有效性。此外,它还探讨了对 GTO 行为的改进建议,重点是使搜索区域的几何形状与现实世界的优化问题相一致。综述还介绍了 GTO 求解器,提供了有关其识别和组织的详细信息,并演示了其在各种优化场景中的应用。此外,它还针对 GTO 的主要局限性,对其收敛行为进行了批判性评估。最后,本综述总结了研究的主要发现,并提出了未来与 GTO 相关的改进和调整的潜在途径。
{"title":"An in-depth survey of the artificial gorilla troops optimizer: outcomes, variations, and applications","authors":"Abdelazim G. Hussien,&nbsp;Anas Bouaouda,&nbsp;Abdullah Alzaqebah,&nbsp;Sumit Kumar,&nbsp;Gang Hu,&nbsp;Heming Jia","doi":"10.1007/s10462-024-10838-8","DOIUrl":"10.1007/s10462-024-10838-8","url":null,"abstract":"<div><p>A recently developed algorithm inspired by natural processes, known as the Artificial Gorilla Troops Optimizer (GTO), boasts a straightforward structure, unique stabilizing features, and notably high effectiveness. Its primary objective is to efficiently find solutions for a wide array of challenges, whether they involve constraints or not. The GTO takes its inspiration from the behavior of Gorilla Troops in the natural world. To emulate the impact of gorillas at each stage of the search process, the GTO employs a flexible weighting mechanism rooted in its concept. Its exceptional qualities, including its independence from derivatives, lack of parameters, user-friendliness, adaptability, and simplicity, have resulted in its rapid adoption for addressing various optimization challenges. This review is dedicated to the examination and discussion of the foundational research that forms the basis of the GTO. It delves into the evolution of this algorithm, drawing insights from 112 research studies that highlight its effectiveness. Additionally, it explores proposed enhancements to the GTO’s behavior, with a specific focus on aligning the geometry of the search area with real-world optimization problems. The review also introduces the GTO solver, providing details about its identification and organization, and demonstrates its application in various optimization scenarios. Furthermore, it provides a critical assessment of the convergence behavior while addressing the primary limitation of the GTO. In conclusion, this review summarizes the key findings of the study and suggests potential avenues for future advancements and adaptations related to the GTO.</p></div>","PeriodicalId":8449,"journal":{"name":"Artificial Intelligence Review","volume":null,"pages":null},"PeriodicalIF":10.7,"publicationDate":"2024-08-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10462-024-10838-8.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141933379","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Link prediction for hypothesis generation: an active curriculum learning infused temporal graph-based approach 生成假设的链接预测:基于时态图的主动课程学习注入法
IF 10.7 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-08-12 DOI: 10.1007/s10462-024-10885-1
Uchenna Akujuobi, Priyadarshini Kumari, Jihun Choi, Samy Badreddine, Kana Maruyama, Sucheendra K. Palaniappan, Tarek R. Besold

Over the last few years Literature-based Discovery (LBD) has regained popularity as a means to enhance the scientific research process. The resurgent interest has spurred the development of supervised and semi-supervised machine learning models aimed at making previously implicit connections between scientific concepts/entities within often extensive bodies of literature explicit—i.e., suggesting novel scientific hypotheses. In doing so, understanding the temporally evolving interactions between these entities can provide valuable information for predicting the future development of entity relationships. However, existing methods often underutilize the latent information embedded in the temporal aspects of the interaction data. Motivated by applications in the food domain—where we aim to connect nutritional information with health-related benefits—we address the hypothesis-generation problem using a temporal graph-based approach. Given that hypothesis generation involves predicting future (i.e., still to be discovered) entity connections, in our view the ability to capture the dynamic evolution of connections over time is pivotal for a robust model. To address this, we introduce THiGER, a novel batch contrastive temporal node-pair embedding method. THiGER excels in providing a more expressive node-pair encoding by effectively harnessing node-pair relationships. Furthermore, we present THiGER-A, an incremental training approach that incorporates an active curriculum learning strategy to mitigate label bias arising from unobserved connections. By progressively training on increasingly challenging and high-utility samples, our approach significantly enhances the performance of the embedding model. Empirical validation of our proposed method demonstrates its effectiveness on established temporal-graph benchmark datasets, as well as on real-world datasets within the food domain.

在过去几年里,基于文献的发现(LBD)作为加强科学研究过程的一种手段重新受到人们的欢迎。这种兴趣的复苏刺激了有监督和半监督机器学习模型的发展,这些模型旨在将通常大量文献中的科学概念/实体之间以前隐含的联系显性化,即提出新的科学假设。在此过程中,了解这些实体之间随时间演变的互动关系可以为预测实体关系的未来发展提供有价值的信息。然而,现有的方法往往没有充分利用交互数据的时间方面所蕴含的潜在信息。受食品领域应用的启发--我们的目标是将营养信息与健康相关的益处联系起来--我们采用基于时序图的方法来解决假设生成问题。鉴于假设生成涉及预测未来(即仍有待发现的)实体连接,我们认为捕捉连接随时间的动态演变的能力对于建立一个稳健的模型至关重要。为了解决这个问题,我们引入了 THiGER,这是一种新颖的批量对比性时间节点对嵌入方法。通过有效利用节点对关系,THiGER 能够提供更具表现力的节点对编码。此外,我们还介绍了 THiGER-A,这是一种渐进式训练方法,它采用了主动课程学习策略,以减轻因未观察到的连接而产生的标签偏差。我们的方法通过在越来越具有挑战性和高实用性的样本上进行渐进式训练,显著提高了嵌入模型的性能。我们提出的方法在已建立的时间图基准数据集以及食品领域的实际数据集上进行了经验验证,证明了它的有效性。
{"title":"Link prediction for hypothesis generation: an active curriculum learning infused temporal graph-based approach","authors":"Uchenna Akujuobi,&nbsp;Priyadarshini Kumari,&nbsp;Jihun Choi,&nbsp;Samy Badreddine,&nbsp;Kana Maruyama,&nbsp;Sucheendra K. Palaniappan,&nbsp;Tarek R. Besold","doi":"10.1007/s10462-024-10885-1","DOIUrl":"10.1007/s10462-024-10885-1","url":null,"abstract":"<div><p>Over the last few years Literature-based Discovery (LBD) has regained popularity as a means to enhance the scientific research process. The resurgent interest has spurred the development of supervised and semi-supervised machine learning models aimed at making previously implicit connections between scientific concepts/entities within often extensive bodies of literature explicit—i.e., suggesting novel scientific hypotheses. In doing so, understanding the temporally evolving interactions between these entities can provide valuable information for predicting the future development of entity relationships. However, existing methods often underutilize the latent information embedded in the temporal aspects of the interaction data. Motivated by applications in the food domain—where we aim to connect nutritional information with health-related benefits—we address the hypothesis-generation problem using a temporal graph-based approach. Given that hypothesis generation involves predicting future (i.e., still to be discovered) entity connections, in our view the ability to capture the dynamic evolution of connections over time is pivotal for a robust model. To address this, we introduce <i>THiGER</i>, a novel batch contrastive temporal node-pair embedding method. <i>THiGER</i> excels in providing a more expressive node-pair encoding by effectively harnessing node-pair relationships. Furthermore, we present <i>THiGER-A</i>, an incremental training approach that incorporates an active curriculum learning strategy to mitigate label bias arising from unobserved connections. By progressively training on increasingly challenging and high-utility samples, our approach significantly enhances the performance of the embedding model. Empirical validation of our proposed method demonstrates its effectiveness on established temporal-graph benchmark datasets, as well as on real-world datasets within the food domain.</p></div>","PeriodicalId":8449,"journal":{"name":"Artificial Intelligence Review","volume":null,"pages":null},"PeriodicalIF":10.7,"publicationDate":"2024-08-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10462-024-10885-1.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141933382","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Domain generalization for semantic segmentation: a survey 语义分割的领域泛化:一项调查
IF 10.7 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-08-12 DOI: 10.1007/s10462-024-10817-z
Taki Hasan Rafi, Ratul Mahjabin, Emon Ghosh, Young-Woong Ko, Jeong-Gun Lee

Deep neural networks (DNNs) have proven explicit contributions in making autonomous driving cars and related tasks such as semantic segmentation, motion tracking, object detection, sensor fusion, and planning. However, in challenging situations, DNNs are not generalizable because of the inherent domain shift due to the nature of training under the i.i.d. assumption. The goal of semantic segmentation is to preserve information from a given image into multiple meaningful categories for visual understanding. Particularly for semantic segmentation, pixel-wise annotation is extremely costly and not always feasible. Domain generalization for semantic segmentation aims to learn pixel-level semantic labels from multiple source domains and generalize to predict pixel-level semantic labels on multiple unseen target domains. In this survey, for the first time, we present a comprehensive review of DG for semantic segmentation. we present a comprehensive summary of recent works related to domain generalization in semantic segmentation, which establishes the importance of generalizing to new environments of segmentation models. Although domain adaptation has gained more attention in segmentation tasks than domain generalization, it is still worth unveiling new trends that are adopted from domain generalization methods in semantic segmentation. We cover most of the recent and dominant DG methods in the context of semantic segmentation and also provide some other related applications. We conclude this survey by highlighting the future directions in this area.

事实证明,深度神经网络(DNN)在自动驾驶汽车以及语义分割、运动跟踪、物体检测、传感器融合和规划等相关任务中做出了明确的贡献。然而,在具有挑战性的情况下,由于在 i.i.d. 假设下进行训练的性质所导致的固有领域偏移,DNN 并不具有通用性。语义分割的目标是将给定图像中的信息保存为多个有意义的类别,以便于视觉理解。特别是对于语义分割来说,像素标注成本极高,而且并不总是可行。语义分割的领域泛化旨在从多个源领域学习像素级语义标签,并泛化到预测多个未见目标领域的像素级语义标签。在本调查报告中,我们首次对用于语义分割的领域泛化进行了全面回顾。我们对近期与语义分割领域泛化相关的工作进行了全面总结,从而确定了分割模型泛化到新环境的重要性。虽然在分割任务中,领域适应比领域泛化更受关注,但在语义分割中采用领域泛化方法的新趋势仍然值得揭示。我们介绍了语义分割领域中最近出现的大多数主流 DG 方法,并提供了一些其他相关应用。最后,我们强调了这一领域的未来发展方向。
{"title":"Domain generalization for semantic segmentation: a survey","authors":"Taki Hasan Rafi,&nbsp;Ratul Mahjabin,&nbsp;Emon Ghosh,&nbsp;Young-Woong Ko,&nbsp;Jeong-Gun Lee","doi":"10.1007/s10462-024-10817-z","DOIUrl":"10.1007/s10462-024-10817-z","url":null,"abstract":"<div><p>Deep neural networks (DNNs) have proven explicit contributions in making autonomous driving cars and related tasks such as semantic segmentation, motion tracking, object detection, sensor fusion, and planning. However, in challenging situations, DNNs are not generalizable because of the inherent domain shift due to the nature of training under the i.i.d. assumption. The goal of semantic segmentation is to preserve information from a given image into multiple meaningful categories for visual understanding. Particularly for semantic segmentation, pixel-wise annotation is extremely costly and not always feasible. Domain generalization for semantic segmentation aims to learn pixel-level semantic labels from multiple source domains and generalize to predict pixel-level semantic labels on multiple unseen target domains. In this survey, for the first time, we present a comprehensive review of DG for semantic segmentation. we present a comprehensive summary of recent works related to domain generalization in semantic segmentation, which establishes the importance of generalizing to new environments of segmentation models. Although domain adaptation has gained more attention in segmentation tasks than domain generalization, it is still worth unveiling new trends that are adopted from domain generalization methods in semantic segmentation. We cover most of the recent and dominant DG methods in the context of semantic segmentation and also provide some other related applications. We conclude this survey by highlighting the future directions in this area.</p></div>","PeriodicalId":8449,"journal":{"name":"Artificial Intelligence Review","volume":null,"pages":null},"PeriodicalIF":10.7,"publicationDate":"2024-08-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10462-024-10817-z.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141933381","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Artificial Intelligence Review
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1