首页 > 最新文献

Artificial Intelligence Review最新文献

英文 中文
A comprehensive investigation of multimodal deep learning fusion strategies for breast cancer classification 乳腺癌分类的多模态深度学习融合策略综合研究
IF 10.7 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-10-12 DOI: 10.1007/s10462-024-10984-z
Fatima-Zahrae Nakach, Ali Idri, Evgin Goceri

In breast cancer research, diverse data types and formats, such as radiological images, clinical records, histological data, and expression analysis, are employed. Given the intricate nature of natural phenomena, relying on the features of a single modality is seldom sufficient for comprehensive analysis. Therefore, it is possible to guarantee medical relevance and achieve improved clinical outcomes by combining several modalities. The presen study carefully maps and reviews 47 primary articles from six well-known digital libraries that were published between 2018 and 2023 for breast cancer classification based on multimodal deep learning fusion (MDLF) techniques. This systematic literature review encompasses various aspects, including the medical modalities combined, the datasets utilized in these studies, the techniques, models, and architectures used in MDLF and it also discusses the advantages and limitations of each approach. The analysis of selected papers has revealed a compelling trend: the emergence of new modalities and combinations that were previously unexplored in the context of breast cancer classification. This exploration has not only expanded the scope of predictive models but also introduced fresh perspectives for addressing diverse targets, ranging from screening to diagnosis and prognosis. The practical advantages of MDLF are evident in its ability to enhance the predictive capabilities of machine learning models, resulting in improved accuracy across diverse applications. The prevalence of deep learning models underscores their success in autonomously discerning complex patterns, offering a substantial departure from traditional machine learning approaches. Furthermore, the paper explores the challenges and future directions in this field, including the need for larger datasets, the use of ensemble learning methods, and the interpretation of multimodal models.

在乳腺癌研究中,需要使用多种数据类型和格式,如放射图像、临床记录、组织学数据和表达分析。由于自然现象错综复杂,仅靠单一模式的特征很少能进行全面分析。因此,将几种模式结合起来,才有可能保证医学相关性并改善临床效果。本研究对2018年至2023年期间发表的6个知名数字图书馆中的47篇主要文章进行了仔细的映射和回顾,以研究基于多模态深度学习融合(MDLF)技术的乳腺癌分类。这篇系统性文献综述涵盖了各个方面,包括结合的医疗模式、这些研究中使用的数据集、MDLF 中使用的技术、模型和架构,它还讨论了每种方法的优势和局限性。对所选论文的分析揭示了一个引人注目的趋势:在乳腺癌分类方面出现了以前从未探索过的新模式和新组合。这种探索不仅扩大了预测模型的范围,还为解决从筛查到诊断和预后等不同目标引入了新的视角。MDLF 的实际优势体现在它能够增强机器学习模型的预测能力,从而提高各种应用的准确性。深度学习模型的盛行凸显了它们在自主辨别复杂模式方面的成功,与传统的机器学习方法大相径庭。此外,论文还探讨了这一领域的挑战和未来方向,包括对更大数据集的需求、集合学习方法的使用以及多模态模型的解释。
{"title":"A comprehensive investigation of multimodal deep learning fusion strategies for breast cancer classification","authors":"Fatima-Zahrae Nakach,&nbsp;Ali Idri,&nbsp;Evgin Goceri","doi":"10.1007/s10462-024-10984-z","DOIUrl":"10.1007/s10462-024-10984-z","url":null,"abstract":"<div><p>In breast cancer research, diverse data types and formats, such as radiological images, clinical records, histological data, and expression analysis, are employed. Given the intricate nature of natural phenomena, relying on the features of a single modality is seldom sufficient for comprehensive analysis. Therefore, it is possible to guarantee medical relevance and achieve improved clinical outcomes by combining several modalities. The presen study carefully maps and reviews 47 primary articles from six well-known digital libraries that were published between 2018 and 2023 for breast cancer classification based on multimodal deep learning fusion (MDLF) techniques. This systematic literature review encompasses various aspects, including the medical modalities combined, the datasets utilized in these studies, the techniques, models, and architectures used in MDLF and it also discusses the advantages and limitations of each approach. The analysis of selected papers has revealed a compelling trend: the emergence of new modalities and combinations that were previously unexplored in the context of breast cancer classification. This exploration has not only expanded the scope of predictive models but also introduced fresh perspectives for addressing diverse targets, ranging from screening to diagnosis and prognosis. The practical advantages of MDLF are evident in its ability to enhance the predictive capabilities of machine learning models, resulting in improved accuracy across diverse applications. The prevalence of deep learning models underscores their success in autonomously discerning complex patterns, offering a substantial departure from traditional machine learning approaches. Furthermore, the paper explores the challenges and future directions in this field, including the need for larger datasets, the use of ensemble learning methods, and the interpretation of multimodal models.</p></div>","PeriodicalId":8449,"journal":{"name":"Artificial Intelligence Review","volume":"57 12","pages":""},"PeriodicalIF":10.7,"publicationDate":"2024-10-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10462-024-10984-z.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142411525","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A systematic review of computer vision-based personal protective equipment compliance in industry practice: advancements, challenges and future directions 系统回顾基于计算机视觉的个人防护设备在工业实践中的合规性:进步、挑战和未来方向
IF 10.7 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-10-10 DOI: 10.1007/s10462-024-10978-x
Arso M. Vukicevic, Milos Petrovic, Pavle Milosevic, Aleksandar Peulic, Kosta Jovanovic, Aleksandar Novakovic

Computerized compliance of Personal Protective Equipment (PPE) is an emerging topic in academic literature that aims to enhance workplace safety through the automation of compliance and prevention of PPE misuse (which currently relies on manual employee supervision and reporting). Although trends in the scientific literature indicate a high potential for solving the compliance problem by employing computer vision (CV) techniques, the practice has revealed a series of barriers that limit their wider applications. This article aims to contribute to the advancement of CV-based PPE compliance by providing a comparative review of high-level approaches, algorithms, datasets, and technologies used in the literature. The systematic review highlights industry-specific challenges, environmental variations, and computational costs related to the real-time management of PPE compliance. The issues of employee identification and identity management are also discussed, along with ethical and cybersecurity concerns. Through the concept of CV-based PPE Compliance 4.0, which encapsulates PPE, human, and company spatio-temporal variabilities, this study provides guidelines for future research directions for addressing the identified barriers. The further advancements and adoption of CV-based solutions for PPE compliance will require simultaneously addressing human identification, pose estimation, object recognition and tracking, necessitating the development of corresponding public datasets.

个人防护设备(PPE)的计算机合规性是学术文献中的一个新兴课题,其目的是通过自动化合规性和防止个人防护设备的滥用(目前依赖于员工的人工监督和报告)来提高工作场所的安全性。尽管科学文献中的趋势表明,采用计算机视觉(CV)技术解决合规性问题的潜力很大,但实践中发现的一系列障碍限制了其更广泛的应用。本文旨在通过对文献中使用的高级方法、算法、数据集和技术进行比较综述,推动基于 CV 的个人防护设备合规性的发展。系统综述强调了与个人防护设备合规性实时管理相关的特定行业挑战、环境变化和计算成本。此外,还讨论了员工识别和身份管理问题,以及道德和网络安全问题。基于 CV 的个人防护设备合规性 4.0 概念囊括了个人防护设备、人类和公司的时空变化,通过这一概念,本研究为解决已识别障碍的未来研究方向提供了指导。要进一步推进和采用基于 CV 的个人防护设备合规性解决方案,就必须同时解决人体识别、姿势估计、物体识别和跟踪等问题,这就需要开发相应的公共数据集。
{"title":"A systematic review of computer vision-based personal protective equipment compliance in industry practice: advancements, challenges and future directions","authors":"Arso M. Vukicevic,&nbsp;Milos Petrovic,&nbsp;Pavle Milosevic,&nbsp;Aleksandar Peulic,&nbsp;Kosta Jovanovic,&nbsp;Aleksandar Novakovic","doi":"10.1007/s10462-024-10978-x","DOIUrl":"10.1007/s10462-024-10978-x","url":null,"abstract":"<div><p>Computerized compliance of Personal Protective Equipment (PPE) is an emerging topic in academic literature that aims to enhance workplace safety through the automation of compliance and prevention of PPE misuse (which currently relies on manual employee supervision and reporting). Although trends in the scientific literature indicate a high potential for solving the compliance problem by employing computer vision (CV) techniques, the practice has revealed a series of barriers that limit their wider applications. This article aims to contribute to the advancement of CV-based PPE compliance by providing a comparative review of high-level approaches, algorithms, datasets, and technologies used in the literature. The systematic review highlights industry-specific challenges, environmental variations, and computational costs related to the real-time management of PPE compliance. The issues of employee identification and identity management are also discussed, along with ethical and cybersecurity concerns. Through the concept of CV-based PPE Compliance 4.0, which encapsulates PPE, human, and company spatio-temporal variabilities, this study provides guidelines for future research directions for addressing the identified barriers. The further advancements and adoption of CV-based solutions for PPE compliance will require simultaneously addressing human identification, pose estimation, object recognition and tracking, necessitating the development of corresponding public datasets.</p></div>","PeriodicalId":8449,"journal":{"name":"Artificial Intelligence Review","volume":"57 12","pages":""},"PeriodicalIF":10.7,"publicationDate":"2024-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10462-024-10978-x.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142411190","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A systematic literature review on pancreas segmentation from traditional to non-supervised techniques in abdominal medical images 关于腹部医学图像中胰腺分割(从传统技术到非监督技术)的系统性文献综述
IF 10.7 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-10-10 DOI: 10.1007/s10462-024-10966-1
Suchi Jain, Geeta Sikka, Renu Dhir

Abdominal organs play a significant role in regulating various functional systems. Any impairment in its functioning can lead to cancerous diseases. Diagnosing these diseases mainly relies on radiologists’ subjective assessment, which varies according to professional abilities and clinical experience. Computer-Aided Diagnosis (CAD) system is designed to assist clinicians in identifying various pathological changes. Hence, automatic pancreas segmentation is a vital input to the CAD system in the diagnosis of cancer at its early stages. Automatic segmentation is achieved through traditional methods like atlas-based and statistical models, and nowadays, it is achieved through artificial intelligence approaches like machine learning and deep learning using various imaging modalities. This study investigates and analyses the various state-of-the-art multi-organ and pancreas segmentation approaches to identify the research gaps and future perspectives for the research community. The objective is achieved by framing the research questions using the PICOC framework and then selecting 140 research articles using a systematic process through the Covidence tool to conclude the answers to the respective questions. The literature search has been conducted on five databases of original studies published from 2003 to 2023. Initially, the literature analysis is presented in terms of publication, and the comparative analysis of the current study is presented with existing review studies. Then, existing studies are analyzed, focusing on semi-automatic and automatic multi-organ segmentation and pancreas segmentation, using various learning methods. Finally, the various critical issues, the research gaps and the future perspectives of segmentation methods based on published evidence are summarized.

腹部器官在调节各种功能系统方面发挥着重要作用。任何功能障碍都可能导致癌症疾病。诊断这些疾病主要依靠放射科医生的主观评估,而主观评估因专业能力和临床经验而异。计算机辅助诊断(CAD)系统旨在协助临床医生识别各种病理变化。因此,自动胰腺分割是计算机辅助诊断系统在癌症早期诊断中的重要输入。自动分割是通过基于图集和统计模型等传统方法实现的,如今则是通过机器学习和深度学习等人工智能方法,利用各种成像模式实现的。本研究调查并分析了各种最先进的多器官和胰腺分割方法,以确定研究界的研究空白和未来展望。为了实现这一目标,研究人员使用 PICOC 框架提出了研究问题,然后通过 Covidence 工具以系统化的流程筛选出 140 篇研究文章,从而总结出相应问题的答案。文献检索在五个数据库中进行,这些数据库收录了 2003 年至 2023 年间发表的原创研究。首先,从发表的角度对文献进行分析,并将当前研究与现有的综述研究进行对比分析。然后,分析了现有研究,重点是使用各种学习方法进行半自动和自动多器官分割以及胰腺分割。最后,根据已发表的证据总结了分割方法的各种关键问题、研究空白和未来展望。
{"title":"A systematic literature review on pancreas segmentation from traditional to non-supervised techniques in abdominal medical images","authors":"Suchi Jain,&nbsp;Geeta Sikka,&nbsp;Renu Dhir","doi":"10.1007/s10462-024-10966-1","DOIUrl":"10.1007/s10462-024-10966-1","url":null,"abstract":"<div><p>Abdominal organs play a significant role in regulating various functional systems. Any impairment in its functioning can lead to cancerous diseases. Diagnosing these diseases mainly relies on radiologists’ subjective assessment, which varies according to professional abilities and clinical experience. Computer-Aided Diagnosis (CAD) system is designed to assist clinicians in identifying various pathological changes. Hence, automatic pancreas segmentation is a vital input to the CAD system in the diagnosis of cancer at its early stages. Automatic segmentation is achieved through traditional methods like atlas-based and statistical models, and nowadays, it is achieved through artificial intelligence approaches like machine learning and deep learning using various imaging modalities. This study investigates and analyses the various state-of-the-art multi-organ and pancreas segmentation approaches to identify the research gaps and future perspectives for the research community. The objective is achieved by framing the research questions using the PICOC framework and then selecting 140 research articles using a systematic process through the Covidence tool to conclude the answers to the respective questions. The literature search has been conducted on five databases of original studies published from 2003 to 2023. Initially, the literature analysis is presented in terms of publication, and the comparative analysis of the current study is presented with existing review studies. Then, existing studies are analyzed, focusing on semi-automatic and automatic multi-organ segmentation and pancreas segmentation, using various learning methods. Finally, the various critical issues, the research gaps and the future perspectives of segmentation methods based on published evidence are summarized.</p></div>","PeriodicalId":8449,"journal":{"name":"Artificial Intelligence Review","volume":"57 12","pages":""},"PeriodicalIF":10.7,"publicationDate":"2024-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10462-024-10966-1.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142411200","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Neuromorphic computing for modeling neurological and psychiatric disorders: implications for drug development 用于神经和精神疾病建模的神经形态计算:对药物开发的影响
IF 10.7 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-10-10 DOI: 10.1007/s10462-024-10948-3
Amisha S. Raikar, J Andrew, Pranjali Prabhu Dessai, Sweta M. Prabhu, Shounak Jathar, Aishwarya Prabhu, Mayuri B. Naik, Gokuldas Vedant S. Raikar

The emergence of neuromorphic computing, inspired by the structure and function of the human brain, presents a transformative framework for modelling neurological disorders in drug development. This article investigates the implications of applying neuromorphic computing to simulate and comprehend complex neural systems affected by conditions like Alzheimer’s, Parkinson’s, and epilepsy, drawing from extensive literature. It explores the intersection of neuromorphic computing with neurology and pharmaceutical development, emphasizing the significance of understanding neural processes and integrating deep learning techniques. Technical considerations, such as integrating neural circuits into CMOS technology and employing memristive devices for synaptic emulation, are discussed. The review evaluates how neuromorphic computing optimizes drug discovery and improves clinical trials by precisely simulating biological systems. It also examines the role of neuromorphic models in comprehending and simulating neurological disorders, facilitating targeted treatment development. Recent progress in neuromorphic drug discovery is highlighted, indicating the potential for transformative therapeutic interventions. As technology advances, the synergy between neuromorphic computing and neuroscience holds promise for revolutionizing the study of the human brain’s complexities and addressing neurological challenges.

受人脑结构和功能的启发,神经形态计算的出现为药物开发中的神经系统疾病建模提供了一个变革性框架。本文通过大量文献,探讨了应用神经形态计算模拟和理解受阿尔茨海默氏症、帕金森氏症和癫痫等疾病影响的复杂神经系统的意义。文章探讨了神经形态计算与神经学和药物开发的交叉点,强调了理解神经过程和整合深度学习技术的重要性。文中还讨论了一些技术考虑因素,如将神经电路集成到 CMOS 技术中,以及采用记忆器件进行突触仿真。综述评估了神经形态计算如何通过精确模拟生物系统来优化药物发现和改进临床试验。它还探讨了神经形态模型在理解和模拟神经系统疾病方面的作用,从而促进有针对性的治疗开发。报告重点介绍了神经形态药物发现的最新进展,指出了变革性治疗干预的潜力。随着技术的进步,神经形态计算与神经科学之间的协同作用有望彻底改变对人类大脑复杂性的研究,并解决神经学方面的难题。
{"title":"Neuromorphic computing for modeling neurological and psychiatric disorders: implications for drug development","authors":"Amisha S. Raikar,&nbsp;J Andrew,&nbsp;Pranjali Prabhu Dessai,&nbsp;Sweta M. Prabhu,&nbsp;Shounak Jathar,&nbsp;Aishwarya Prabhu,&nbsp;Mayuri B. Naik,&nbsp;Gokuldas Vedant S. Raikar","doi":"10.1007/s10462-024-10948-3","DOIUrl":"10.1007/s10462-024-10948-3","url":null,"abstract":"<div><p>The emergence of neuromorphic computing, inspired by the structure and function of the human brain, presents a transformative framework for modelling neurological disorders in drug development. This article investigates the implications of applying neuromorphic computing to simulate and comprehend complex neural systems affected by conditions like Alzheimer’s, Parkinson’s, and epilepsy, drawing from extensive literature. It explores the intersection of neuromorphic computing with neurology and pharmaceutical development, emphasizing the significance of understanding neural processes and integrating deep learning techniques. Technical considerations, such as integrating neural circuits into CMOS technology and employing memristive devices for synaptic emulation, are discussed. The review evaluates how neuromorphic computing optimizes drug discovery and improves clinical trials by precisely simulating biological systems. It also examines the role of neuromorphic models in comprehending and simulating neurological disorders, facilitating targeted treatment development. Recent progress in neuromorphic drug discovery is highlighted, indicating the potential for transformative therapeutic interventions. As technology advances, the synergy between neuromorphic computing and neuroscience holds promise for revolutionizing the study of the human brain’s complexities and addressing neurological challenges.</p></div>","PeriodicalId":8449,"journal":{"name":"Artificial Intelligence Review","volume":"57 12","pages":""},"PeriodicalIF":10.7,"publicationDate":"2024-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10462-024-10948-3.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142411182","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A hybrid approach for Bengali sentence validation 孟加拉语句子验证的混合方法
IF 10.7 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-10-07 DOI: 10.1007/s10462-024-10795-2
Juel Sikder, Prosenjit Chakraborty, Utpol Kanti Das, Krity Dhar

Bengali is the official language of Bangladesh and is widely used in Bangladesh and West Bengal in India. Due to the growing accessibility of the internet and smart devices, the use of digital text material and documents in Bengali is growing with time. An automated Bengali Sentence Validation System is proposed in this study to effectively determine the correctness of sentences in such extensively available Bengali content. As far as we know, no substantial work has been done in the field of Bengali Sentence Validation utilizing deep learning approaches. Due to the lack of linguistic resources, sophisticated Natural Language Processing tools, and benchmark datasets, developing an automated Sentence Validation System for a limited-resource language like Bengali is challenging. Additionally, Bengali Sentences come in two morphological varieties (Sadhu-bhasha and Cholito-bhasha), making the validation process more challenging. The proposed automated Bengali Sentence Validation system contains the CNN-BiLSTM hybrid classifier model. As of now, there is no standard dataset for Bengali sentence validation. Due to the lack of a standard dataset, we collected Bengali sentences from different sources in Bangladesh and developed a Bengali Sentence Validation (BSV) Dataset with around 5000 labelled sentences arranged into two categories such as correct and incorrect. Experimental results demonstrate that the proposed system outperformed other classifier models and existing approaches for Bengali Sentence Validation and is able to categorize a wide range of Bengali sentences based on their correctness. The system’s F1 score for the Bengali Sentence Validation is 98%.

孟加拉语是孟加拉国的官方语言,在孟加拉国和印度西孟加拉邦广泛使用。由于互联网和智能设备的普及,孟加拉语数字文本材料和文档的使用与日俱增。本研究提出了一个自动孟加拉语句子验证系统,以有效确定这些广泛使用的孟加拉语内容中句子的正确性。据我们所知,在孟加拉语句子验证领域还没有利用深度学习方法进行的实质性工作。由于缺乏语言资源、复杂的自然语言处理工具和基准数据集,为孟加拉语这种资源有限的语言开发自动句子验证系统具有挑战性。此外,孟加拉语句子有两种形态(Sadhu-bhasha 和 Cholito-bhasha),这使得验证过程更具挑战性。拟议的孟加拉语句子自动验证系统包含 CNN-BiLSTM 混合分类器模型。到目前为止,还没有孟加拉语句子验证的标准数据集。由于缺乏标准数据集,我们从孟加拉国的不同来源收集了孟加拉语句子,并开发了孟加拉语句子验证(BSV)数据集,其中包含约 5000 个标签句子,分为正确和错误两类。实验结果表明,所提出的系统在孟加拉语句子验证方面的表现优于其他分类器模型和现有方法,能够根据句子的正确性对各种孟加拉语句子进行分类。该系统在孟加拉语句子验证方面的 F1 得分为 98%。
{"title":"A hybrid approach for Bengali sentence validation","authors":"Juel Sikder,&nbsp;Prosenjit Chakraborty,&nbsp;Utpol Kanti Das,&nbsp;Krity Dhar","doi":"10.1007/s10462-024-10795-2","DOIUrl":"10.1007/s10462-024-10795-2","url":null,"abstract":"<div><p>Bengali is the official language of Bangladesh and is widely used in Bangladesh and West Bengal in India. Due to the growing accessibility of the internet and smart devices, the use of digital text material and documents in Bengali is growing with time. An automated Bengali Sentence Validation System is proposed in this study to effectively determine the correctness of sentences in such extensively available Bengali content. As far as we know, no substantial work has been done in the field of Bengali Sentence Validation utilizing deep learning approaches. Due to the lack of linguistic resources, sophisticated Natural Language Processing tools, and benchmark datasets, developing an automated Sentence Validation System for a limited-resource language like Bengali is challenging. Additionally, Bengali Sentences come in two morphological varieties (Sadhu-bhasha and Cholito-bhasha), making the validation process more challenging. The proposed automated Bengali Sentence Validation system contains the CNN-BiLSTM hybrid classifier model. As of now, there is no standard dataset for Bengali sentence validation. Due to the lack of a standard dataset, we collected Bengali sentences from different sources in Bangladesh and developed a Bengali Sentence Validation (BSV) Dataset with around 5000 labelled sentences arranged into two categories such as correct and incorrect. Experimental results demonstrate that the proposed system outperformed other classifier models and existing approaches for Bengali Sentence Validation and is able to categorize a wide range of Bengali sentences based on their correctness. The system’s F1 score for the Bengali Sentence Validation is 98%. </p></div>","PeriodicalId":8449,"journal":{"name":"Artificial Intelligence Review","volume":"57 11","pages":""},"PeriodicalIF":10.7,"publicationDate":"2024-10-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10462-024-10795-2.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142410443","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A review of graph neural network applications in mechanics-related domains 图神经网络在力学相关领域的应用综述
IF 10.7 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-10-04 DOI: 10.1007/s10462-024-10931-y
Yingxue Zhao, Haoran Li, Haosu Zhou, Hamid Reza Attar, Tobias Pfaff, Nan Li

Mechanics-related tasks often present unique challenges in achieving accurate geometric and physical representations, particularly for non-uniform structures. Graph neural networks (GNNs) have emerged as a promising tool to tackle these challenges by adeptly learning from graph data with irregular underlying structures. Consequently, recent years have witnessed a surge in complex mechanics-related applications inspired by the advancements of GNNs. Despite this process, there is a notable absence of a systematic review addressing the recent advancement of GNNs in solving mechanics-related tasks. To bridge this gap, this review article aims to provide an in-depth overview of the GNN applications in mechanics-related domains while identifying key challenges and outlining potential future research directions. In this review article, we begin by introducing the fundamental algorithms of GNNs that are widely employed in mechanics-related applications. We provide a concise explanation of their underlying principles to establish a solid understanding that will serve as a basis for exploring the applications of GNNs in mechanics-related domains. The scope of this paper is intended to cover the categorisation of literature into solid mechanics, fluid mechanics, and interdisciplinary mechanics-related domains, providing a comprehensive summary of graph representation methodologies, GNN architectures, and further discussions in their respective subdomains. Additionally, open data and source codes relevant to these applications are summarised for the convenience of future researchers. This article promotes an interdisciplinary integration of GNNs and mechanics and provides a guide for researchers interested in applying GNNs to solve complex mechanics-related tasks.

与机械相关的任务在实现精确的几何和物理表示方面往往面临独特的挑战,特别是对于非均匀结构。图形神经网络(GNN)通过善于从具有不规则底层结构的图形数据中学习,已成为应对这些挑战的一种有前途的工具。因此,近年来,受 GNN 技术进步的启发,与复杂力学相关的应用激增。尽管如此,目前仍缺乏一篇系统性综述来探讨 GNN 在解决力学相关任务方面的最新进展。为了弥补这一空白,本综述文章旨在深入概述 GNN 在机械相关领域的应用,同时明确关键挑战并概述潜在的未来研究方向。在这篇综述文章中,我们首先介绍了在力学相关应用中广泛使用的 GNN 基本算法。我们简明扼要地解释了这些算法的基本原理,以便为探索 GNN 在力学相关领域的应用奠定坚实的基础。本文的研究范围涵盖了固体力学、流体力学和跨学科力学相关领域的文献分类,全面总结了图表示方法、GNN 架构以及在各自子领域的进一步讨论。此外,还总结了与这些应用相关的开放数据和源代码,以方便未来的研究人员。本文促进了 GNN 与力学的跨学科融合,为有兴趣应用 GNN 解决复杂力学相关任务的研究人员提供了指南。
{"title":"A review of graph neural network applications in mechanics-related domains","authors":"Yingxue Zhao,&nbsp;Haoran Li,&nbsp;Haosu Zhou,&nbsp;Hamid Reza Attar,&nbsp;Tobias Pfaff,&nbsp;Nan Li","doi":"10.1007/s10462-024-10931-y","DOIUrl":"10.1007/s10462-024-10931-y","url":null,"abstract":"<div><p>Mechanics-related tasks often present unique challenges in achieving accurate geometric and physical representations, particularly for non-uniform structures. Graph neural networks (GNNs) have emerged as a promising tool to tackle these challenges by adeptly learning from graph data with irregular underlying structures. Consequently, recent years have witnessed a surge in complex mechanics-related applications inspired by the advancements of GNNs. Despite this process, there is a notable absence of a systematic review addressing the recent advancement of GNNs in solving mechanics-related tasks. To bridge this gap, this review article aims to provide an in-depth overview of the GNN applications in mechanics-related domains while identifying key challenges and outlining potential future research directions. In this review article, we begin by introducing the fundamental algorithms of GNNs that are widely employed in mechanics-related applications. We provide a concise explanation of their underlying principles to establish a solid understanding that will serve as a basis for exploring the applications of GNNs in mechanics-related domains. The scope of this paper is intended to cover the categorisation of literature into solid mechanics, fluid mechanics, and interdisciplinary mechanics-related domains, providing a comprehensive summary of graph representation methodologies, GNN architectures, and further discussions in their respective subdomains. Additionally, open data and source codes relevant to these applications are summarised for the convenience of future researchers. This article promotes an interdisciplinary integration of GNNs and mechanics and provides a guide for researchers interested in applying GNNs to solve complex mechanics-related tasks.</p></div>","PeriodicalId":8449,"journal":{"name":"Artificial Intelligence Review","volume":"57 11","pages":""},"PeriodicalIF":10.7,"publicationDate":"2024-10-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10462-024-10931-y.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142409848","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Improving the Kepler optimization algorithm with chaotic maps: comprehensive performance evaluation and engineering applications 用混沌图改进开普勒优化算法:综合性能评估和工程应用
IF 10.7 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-10-03 DOI: 10.1007/s10462-024-10857-5
Nawal El Ghouate, Ahmed Bencherqui, Hanaa Mansouri, Ahmed El Maloufy, Mohamed Amine Tahiri, Hicham Karmouni, Mhamed Sayyouri, S. S. Askar, Mohamed Abouhawwash

The Kepler Optimisation Algorithm (KOA) is a recently proposed algorithm that is inspired by Kepler’s laws to predict the positions and velocities of planets at a given time. However, although promising, KOA can encounter challenges such as convergence to sub-optimal solutions or slow convergence speed. This paper proposes an improvement to KOA by integrating chaotic maps to solve complex engineering problems. The improved algorithm, named Chaotic Kepler Optimization Algorithm (CKOA), is characterized by a better ability to avoid local minima and to reach globally optimal solutions thanks to a dynamic diversification strategy based on chaotic maps. To confirm the effectiveness of the suggested approach, in-depth statistical analyses were carried out using the CEC2020 and CEC2022 benchmarks. These analyses included mean and standard deviation of fitness, convergence curves, Wilcoxon tests, as well as population diversity assessments. The experimental results, which compare CKOA not only to the original KOA but also to eight other recent optimizers, show that the proposed algorithm performs better in terms of convergence speed and solution quality. In addition, CKOA has been successfully tested on three complex engineering problems, confirming its robustness and practical effectiveness. These results make CKOA a powerful optimisation tool in a variety of complex real-world contexts. After final acceptance, the source code will be uploaded to the Github account: nawal.elghouate@usmba.ac.ma.

开普勒优化算法(KOA)是最近提出的一种算法,它受到开普勒定律的启发,可以预测行星在给定时间内的位置和速度。然而,KOA 虽然前景广阔,但也会遇到收敛到次优解或收敛速度慢等挑战。本文提出了一种改进 KOA 的方法,通过整合混沌图来解决复杂的工程问题。改进后的算法被命名为混沌开普勒优化算法(CKOA),其特点是由于采用了基于混沌图的动态多样化策略,因此能更好地避免局部最小值,并达到全局最优解。为了证实所建议方法的有效性,我们使用 CEC2020 和 CEC2022 基准进行了深入的统计分析。这些分析包括适合度的平均值和标准偏差、收敛曲线、Wilcoxon 检验以及种群多样性评估。实验结果不仅将 CKOA 与原始 KOA 进行了比较,而且还与其他八个最新的优化器进行了比较,结果表明,所提出的算法在收敛速度和解决方案质量方面表现更好。此外,CKOA 还成功地在三个复杂的工程问题上进行了测试,证实了它的鲁棒性和实用性。这些结果使 CKOA 在各种复杂的实际环境中成为一个强大的优化工具。最终验收后,源代码将上传到 Github 账户:nawal.elghouate@usmba.ac.ma。
{"title":"Improving the Kepler optimization algorithm with chaotic maps: comprehensive performance evaluation and engineering applications","authors":"Nawal El Ghouate,&nbsp;Ahmed Bencherqui,&nbsp;Hanaa Mansouri,&nbsp;Ahmed El Maloufy,&nbsp;Mohamed Amine Tahiri,&nbsp;Hicham Karmouni,&nbsp;Mhamed Sayyouri,&nbsp;S. S. Askar,&nbsp;Mohamed Abouhawwash","doi":"10.1007/s10462-024-10857-5","DOIUrl":"10.1007/s10462-024-10857-5","url":null,"abstract":"<div><p>The Kepler Optimisation Algorithm (KOA) is a recently proposed algorithm that is inspired by Kepler’s laws to predict the positions and velocities of planets at a given time. However, although promising, KOA can encounter challenges such as convergence to sub-optimal solutions or slow convergence speed. This paper proposes an improvement to KOA by integrating chaotic maps to solve complex engineering problems. The improved algorithm, named Chaotic Kepler Optimization Algorithm (CKOA), is characterized by a better ability to avoid local minima and to reach globally optimal solutions thanks to a dynamic diversification strategy based on chaotic maps. To confirm the effectiveness of the suggested approach, in-depth statistical analyses were carried out using the CEC2020 and CEC2022 benchmarks. These analyses included mean and standard deviation of fitness, convergence curves, Wilcoxon tests, as well as population diversity assessments. The experimental results, which compare CKOA not only to the original KOA but also to eight other recent optimizers, show that the proposed algorithm performs better in terms of convergence speed and solution quality. In addition, CKOA has been successfully tested on three complex engineering problems, confirming its robustness and practical effectiveness. These results make CKOA a powerful optimisation tool in a variety of complex real-world contexts. After final acceptance, the source code will be uploaded to the Github account: nawal.elghouate@usmba.ac.ma.</p></div>","PeriodicalId":8449,"journal":{"name":"Artificial Intelligence Review","volume":"57 11","pages":""},"PeriodicalIF":10.7,"publicationDate":"2024-10-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10462-024-10857-5.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142409608","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Over-the-air upgrading for enhancing security of intelligent connected vehicles: a survey 增强智能网联汽车安全性的空中升级:一项调查
IF 10.7 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-10-03 DOI: 10.1007/s10462-024-10968-z
Beibei Li, Wei Hu, Lemei Da, Yibing Wu, Xinxin Wang, Yiwei Li, Chaoxuan Yuan

The continuous improvement in the connectivity, automation and autonomy levels of Intelligent Connected Vehicles (ICVs) significantly increases the probability of potential security threats. Over-the-Air (OTA) is a promising technique for upgrading features of ICVs and enhancing their reliability and security against environmental disturbances as well as malicious attacks. To better understand the potential security risks and possible countermeasures, we survey research works in ICV security during OTA from cloud upgrade, terminal upgrade, and object upgrade. We also summarize existing methods in OTA upgrading techniques and systematically investigate the overall framework of OTA upgrading methods from the perspectives of Software-Over-the-Air (SOTA) and Firmware-Over-the-Air (FOTA).We further discuss possible mitigation strategies and open issues yet to be resolved in this research direction. This survey shows that OTA provides a powerful technique for upgrading the ICV features and improving ICV security.

智能网联汽车(ICV)的连接性、自动化和自主化水平不断提高,这大大增加了潜在安全威胁的可能性。空中下载(OTA)是一种很有前途的技术,可用于升级 ICV 的功能,提高其可靠性和安全性,以抵御环境干扰和恶意攻击。为了更好地了解潜在的安全风险和可能的应对措施,我们从云升级、终端升级和对象升级等方面调查了 OTA 期间 ICV 安全方面的研究工作。我们还总结了 OTA 升级技术的现有方法,并从软件空中升级(Software-Over-the-Air,SOTA)和固件空中升级(Firmware-Over-the-Air,FOTA)的角度系统地研究了 OTA 升级方法的整体框架。这项调查表明,OTA 为升级 ICV 功能和提高 ICV 安全性提供了一种强大的技术。
{"title":"Over-the-air upgrading for enhancing security of intelligent connected vehicles: a survey","authors":"Beibei Li,&nbsp;Wei Hu,&nbsp;Lemei Da,&nbsp;Yibing Wu,&nbsp;Xinxin Wang,&nbsp;Yiwei Li,&nbsp;Chaoxuan Yuan","doi":"10.1007/s10462-024-10968-z","DOIUrl":"10.1007/s10462-024-10968-z","url":null,"abstract":"<div><p>The continuous improvement in the connectivity, automation and autonomy levels of Intelligent Connected Vehicles (ICVs) significantly increases the probability of potential security threats. Over-the-Air (OTA) is a promising technique for upgrading features of ICVs and enhancing their reliability and security against environmental disturbances as well as malicious attacks. To better understand the potential security risks and possible countermeasures, we survey research works in ICV security during OTA from cloud upgrade, terminal upgrade, and object upgrade. We also summarize existing methods in OTA upgrading techniques and systematically investigate the overall framework of OTA upgrading methods from the perspectives of Software-Over-the-Air (SOTA) and Firmware-Over-the-Air (FOTA).We further discuss possible mitigation strategies and open issues yet to be resolved in this research direction. This survey shows that OTA provides a powerful technique for upgrading the ICV features and improving ICV security.</p></div>","PeriodicalId":8449,"journal":{"name":"Artificial Intelligence Review","volume":"57 11","pages":""},"PeriodicalIF":10.7,"publicationDate":"2024-10-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10462-024-10968-z.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142409646","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Cost optimization in edge computing: a survey 边缘计算的成本优化:调查
IF 10.7 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-10-01 DOI: 10.1007/s10462-024-10947-4
Liming Cao, Tao Huo, Shaobo Li, Xingxing Zhang, Yanchi Chen, Guangzheng Lin, Fengbin Wu, Yihong Ling, Yaxin Zhou, Qun Xie

The edge computing paradigm is becoming increasingly commercialized due to the widespread adoption of wireless communication technologies and the growing demand for compute-intensive mobile applications. Edge computing complements the cloud computing model by deploying computation, storage, and network resources to the edge locations of wireless access networks, empowering end devices to run resource-intensive applications. In order to promote the commercialization of edge computing, it is important to explore effective ways to reduce the cost of edge computing networks. This paper provides a comprehensive review of the research findings in recent years, offering a clear perspective on the research dynamics. This paper first recalls the architectural framework of edge computing. Then, the main optimization objectives and optimization methods are comprehensively described. Mainstream mathematical models for cost reduction are then shown in depth. The paper also discusses the methods used to evaluate the effectiveness. Then, typical examples of typical application scenarios for edge computing networks are examined in depth. Finally, the paper identifies some unresolved issues. We expect future research to make more attempts in these directions.

由于无线通信技术的广泛应用和计算密集型移动应用需求的不断增长,边缘计算模式正日益商业化。边缘计算是对云计算模式的补充,它将计算、存储和网络资源部署到无线接入网络的边缘位置,使终端设备能够运行资源密集型应用。为了促进边缘计算的商业化,探索降低边缘计算网络成本的有效方法非常重要。本文全面回顾了近年来的研究成果,为研究动态提供了清晰的视角。本文首先回顾了边缘计算的架构框架。然后,全面阐述了主要优化目标和优化方法。然后,深入展示了降低成本的主流数学模型。本文还讨论了用于评估有效性的方法。然后,深入分析了边缘计算网络的典型应用场景。最后,本文指出了一些尚未解决的问题。我们期待未来的研究在这些方向上做出更多尝试。
{"title":"Cost optimization in edge computing: a survey","authors":"Liming Cao,&nbsp;Tao Huo,&nbsp;Shaobo Li,&nbsp;Xingxing Zhang,&nbsp;Yanchi Chen,&nbsp;Guangzheng Lin,&nbsp;Fengbin Wu,&nbsp;Yihong Ling,&nbsp;Yaxin Zhou,&nbsp;Qun Xie","doi":"10.1007/s10462-024-10947-4","DOIUrl":"10.1007/s10462-024-10947-4","url":null,"abstract":"<div><p>The edge computing paradigm is becoming increasingly commercialized due to the widespread adoption of wireless communication technologies and the growing demand for compute-intensive mobile applications. Edge computing complements the cloud computing model by deploying computation, storage, and network resources to the edge locations of wireless access networks, empowering end devices to run resource-intensive applications. In order to promote the commercialization of edge computing, it is important to explore effective ways to reduce the cost of edge computing networks. This paper provides a comprehensive review of the research findings in recent years, offering a clear perspective on the research dynamics. This paper first recalls the architectural framework of edge computing. Then, the main optimization objectives and optimization methods are comprehensively described. Mainstream mathematical models for cost reduction are then shown in depth. The paper also discusses the methods used to evaluate the effectiveness. Then, typical examples of typical application scenarios for edge computing networks are examined in depth. Finally, the paper identifies some unresolved issues. We expect future research to make more attempts in these directions.</p></div>","PeriodicalId":8449,"journal":{"name":"Artificial Intelligence Review","volume":"57 11","pages":""},"PeriodicalIF":10.7,"publicationDate":"2024-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10462-024-10947-4.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142409318","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
AI-enhanced power quality management in distribution systems: implementing a dual-phase UPQC control with adaptive neural networks and optimized PI controllers 配电系统中的人工智能增强型电能质量管理:利用自适应神经网络和优化 PI 控制器实现双相 UPQC 控制
IF 10.7 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-09-30 DOI: 10.1007/s10462-024-10959-0
Arvind R. Singh, Masoud Dashtdar, Mohit Bajaj, Reza Garmsiri, Vojtech Blazek, Lukas Prokop, Stanislav Misak

In the realm of electrical distribution, managing power quality is critical due to its significant impact on infrastructure and customer satisfaction. Addressing issues such as voltage sags and swells, along with current and voltage harmonics, is imperative. The innovative approach proposed in this paper centers on a dual-phase control strategy using a Universal Power Quality Conditioner that integrates series and parallel compensations to rectify these disturbances simultaneously. Our methodology introduces a hybrid control scheme that employs adaptive dynamic neural networks (ADNN), a sinusoidal tracking filter (STF), and a proportional-integral (PI) controller optimized via an improved krill herd (IKH) algorithm. The first phase utilizes the ADNN-based adaptive integrated estimator for quick and accurate disturbance detection and estimation. Subsequently, the second phase employs the STF, omitting the Low Pass Filter and employing a Phase Locking Loop to generate precise reference voltages and currents for the series and parallel active filters based on dynamic load and source conditions. This advanced control mechanism not only enhances system efficacy but also reduces the need for extensive computational resources. Furthermore, the performance of both series and parallel inverters is finely tuned through a PI controller optimized with the IKH algorithm, improving the DC link voltage regulation. Our extensive testing under various conditions, including voltage imbalances and harmonic disturbances, demonstrates the robustness of the proposed solution in both transient and steady-state scenarios.

在配电领域,由于电能质量对基础设施和客户满意度有重大影响,因此电能质量管理至关重要。解决电压骤降和骤升以及电流和电压谐波等问题势在必行。本文提出的创新方法以使用通用电能质量调节器的双相控制策略为核心,该策略集成了串联和并联补偿,可同时纠正这些干扰。我们的方法引入了一种混合控制方案,该方案采用了自适应动态神经网络 (ADNN)、正弦跟踪滤波器 (STF) 和通过改进的磷虾群 (IKH) 算法优化的比例积分 (PI) 控制器。第一阶段利用基于 ADNN 的自适应集成估算器进行快速、准确的干扰检测和估算。随后,第二阶段采用 STF,省略低通滤波器,并采用锁相环,根据动态负载和源条件为串联和并联有源滤波器生成精确的参考电压和电流。这种先进的控制机制不仅提高了系统效率,还减少了对大量计算资源的需求。此外,还通过使用 IKH 算法优化的 PI 控制器对串联和并联逆变器的性能进行了微调,从而改善了直流链路电压调节。我们在包括电压不平衡和谐波干扰在内的各种条件下进行了大量测试,证明了所提出的解决方案在瞬态和稳态情况下的稳健性。
{"title":"AI-enhanced power quality management in distribution systems: implementing a dual-phase UPQC control with adaptive neural networks and optimized PI controllers","authors":"Arvind R. Singh,&nbsp;Masoud Dashtdar,&nbsp;Mohit Bajaj,&nbsp;Reza Garmsiri,&nbsp;Vojtech Blazek,&nbsp;Lukas Prokop,&nbsp;Stanislav Misak","doi":"10.1007/s10462-024-10959-0","DOIUrl":"10.1007/s10462-024-10959-0","url":null,"abstract":"<div><p>In the realm of electrical distribution, managing power quality is critical due to its significant impact on infrastructure and customer satisfaction. Addressing issues such as voltage sags and swells, along with current and voltage harmonics, is imperative. The innovative approach proposed in this paper centers on a dual-phase control strategy using a Universal Power Quality Conditioner that integrates series and parallel compensations to rectify these disturbances simultaneously. Our methodology introduces a hybrid control scheme that employs adaptive dynamic neural networks (ADNN), a sinusoidal tracking filter (STF), and a proportional-integral (PI) controller optimized via an improved krill herd (IKH) algorithm. The first phase utilizes the ADNN-based adaptive integrated estimator for quick and accurate disturbance detection and estimation. Subsequently, the second phase employs the STF, omitting the Low Pass Filter and employing a Phase Locking Loop to generate precise reference voltages and currents for the series and parallel active filters based on dynamic load and source conditions. This advanced control mechanism not only enhances system efficacy but also reduces the need for extensive computational resources. Furthermore, the performance of both series and parallel inverters is finely tuned through a PI controller optimized with the IKH algorithm, improving the DC link voltage regulation. Our extensive testing under various conditions, including voltage imbalances and harmonic disturbances, demonstrates the robustness of the proposed solution in both transient and steady-state scenarios.</p></div>","PeriodicalId":8449,"journal":{"name":"Artificial Intelligence Review","volume":"57 11","pages":""},"PeriodicalIF":10.7,"publicationDate":"2024-09-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10462-024-10959-0.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142415133","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Artificial Intelligence Review
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1