Pub Date : 2024-08-23DOI: 10.1007/s00521-024-10326-8
Engin Tas, Ayca Hatice Atli
Radiological findings play an essential and complementary role in diagnosing Covid-19, assessing its severity, and managing its patients. Artificial intelligence technology based on medical imaging, which has made exciting developments by being applied in many areas, has become an area of interest for the rapid and accurate detection of the disease in the fight against the Covid-19 pandemic. The main difficulty is the inability to obtain a large dataset size with quality and standard images that neural networks need to perform well. Aiming at this problem, this study proposes a Siamese neural network-based deep learning framework for accurate diagnostics of Covid-19 using chest X-ray (CXR) images. The pre-trained VGG16 architecture, based on the transfer learning approach, forms the backbone of the Siamese neural network. The outputs of the backbones are joined together by a merging layer, and then the output passes through a fully connected layer. Based on this structure, category-aware Siamese-based models are produced for each class. The predictions from the models are combined using a voting mechanism to reduce the possibility of misclassification and to make better decisions. The framework was evaluated using a publicly available dataset for the 4-class classification task for Covid-19 pneumonia, lung opacity, normal, and non-Covid-19 viral pneumonia images. The findings reveal the high discrimination ability of the framework, trained using only 10 images per class in less training time, achieving an average test accuracy of 92%. Our framework, which learns a single Siamese-based pairwise model for each class, effectively captures class-specific features. Additionally, it has the potential to deal with data scarcity and long training time problems in multi-class classification tasks.
{"title":"A Siamese neural network-based diagnosis of COVID-19 using chest X-rays","authors":"Engin Tas, Ayca Hatice Atli","doi":"10.1007/s00521-024-10326-8","DOIUrl":"https://doi.org/10.1007/s00521-024-10326-8","url":null,"abstract":"<p>Radiological findings play an essential and complementary role in diagnosing Covid-19, assessing its severity, and managing its patients. Artificial intelligence technology based on medical imaging, which has made exciting developments by being applied in many areas, has become an area of interest for the rapid and accurate detection of the disease in the fight against the Covid-19 pandemic. The main difficulty is the inability to obtain a large dataset size with quality and standard images that neural networks need to perform well. Aiming at this problem, this study proposes a Siamese neural network-based deep learning framework for accurate diagnostics of Covid-19 using chest X-ray (CXR) images. The pre-trained VGG16 architecture, based on the transfer learning approach, forms the backbone of the Siamese neural network. The outputs of the backbones are joined together by a merging layer, and then the output passes through a fully connected layer. Based on this structure, category-aware Siamese-based models are produced for each class. The predictions from the models are combined using a voting mechanism to reduce the possibility of misclassification and to make better decisions. The framework was evaluated using a publicly available dataset for the 4-class classification task for Covid-19 pneumonia, lung opacity, normal, and non-Covid-19 viral pneumonia images. The findings reveal the high discrimination ability of the framework, trained using only 10 images per class in less training time, achieving an average test accuracy of 92%. Our framework, which learns a single Siamese-based pairwise model for each class, effectively captures class-specific features. Additionally, it has the potential to deal with data scarcity and long training time problems in multi-class classification tasks.</p>","PeriodicalId":18925,"journal":{"name":"Neural Computing and Applications","volume":"27 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-08-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142188348","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-23DOI: 10.1007/s00521-024-10309-9
Hsiao-Tien Tsai, Jichong Wu, Puneet Gupta, Eric R. Heinz, Amir Jafari
Coronary Artery Bypass Graft (CABG) is a common cardiac surgery, but it continues to have many associated risks, including the need for blood transfusions. Previous research has shown that blood transfusion during CABG surgery is associated with an increased risk for infection and mortality. The current study aims to use modern techniques, such as deep neural networks and data synthesis, to develop models that can best predict the need for blood transfusion among CABG patients. Results show that neural networks with synthetic data generated by DataSynthesizer have the best performance. Implications of results and future directions are discussed.
{"title":"Predicting blood transfusions for coronary artery bypass graft patients using deep neural networks and synthetic data","authors":"Hsiao-Tien Tsai, Jichong Wu, Puneet Gupta, Eric R. Heinz, Amir Jafari","doi":"10.1007/s00521-024-10309-9","DOIUrl":"https://doi.org/10.1007/s00521-024-10309-9","url":null,"abstract":"<p>Coronary Artery Bypass Graft (CABG) is a common cardiac surgery, but it continues to have many associated risks, including the need for blood transfusions. Previous research has shown that blood transfusion during CABG surgery is associated with an increased risk for infection and mortality. The current study aims to use modern techniques, such as deep neural networks and data synthesis, to develop models that can best predict the need for blood transfusion among CABG patients. Results show that neural networks with synthetic data generated by DataSynthesizer have the best performance. Implications of results and future directions are discussed.</p>","PeriodicalId":18925,"journal":{"name":"Neural Computing and Applications","volume":"7 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-08-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142188287","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-23DOI: 10.1007/s00521-024-10204-3
Eren Duzcu, Bora Yıldırım
The design of a helicopter is an intricate and challenging process. Decisions made during the preliminary design phase can significantly impact subsequent design stages, making it crucial to base these decisions on a solid foundation. A range of methods, including hand calculations, finite element analyses, and experimental tests, can be employed to establish the conceptual design parameters. However, these methods often come with the drawbacks of being time-intensive and costly, especially when testing various structures during the early design phase. To address this issue, this study introduces an artificial neural network-based design tool to evaluate the static structural characteristics of a helicopter’s horizontal stabilizer. The tool was built in Python using the Keras library. The required database for the training of the artificial neural network model was established using finite element analyses of the horizontal stabilizer subjected to the aerodynamic load for diverse design variables. The model’s performance was evaluated, and the model’s outputs were compared to the results derived from the finite element analyses. Moreover, the Hammersley sampling methodology was employed to reduce the size of the database without compromising on accuracy. The study also assessed the impact of decreasing the amount of data fed into the network model.
直升机的设计是一个复杂而具有挑战性的过程。在初步设计阶段做出的决定会对后续设计阶段产生重大影响,因此将这些决定建立在坚实的基础上至关重要。可以采用包括手工计算、有限元分析和实验测试在内的一系列方法来确定概念设计参数。然而,这些方法往往存在耗时长、成本高的缺点,尤其是在早期设计阶段测试各种结构时。为解决这一问题,本研究引入了一种基于人工神经网络的设计工具,用于评估直升机水平安定面的静态结构特性。该工具使用 Keras 库在 Python 中构建。通过对水平稳定器在不同设计变量下的气动载荷进行有限元分析,建立了人工神经网络模型训练所需的数据库。对模型的性能进行了评估,并将模型的输出结果与有限元分析得出的结果进行了比较。此外,还采用了哈默斯利取样方法,在不影响精度的情况下减少了数据库的大小。研究还评估了减少输入网络模型的数据量的影响。
{"title":"Development of a design tool for the horizontal stabilizer of a helicopter using artificial neural networks","authors":"Eren Duzcu, Bora Yıldırım","doi":"10.1007/s00521-024-10204-3","DOIUrl":"https://doi.org/10.1007/s00521-024-10204-3","url":null,"abstract":"<p>The design of a helicopter is an intricate and challenging process. Decisions made during the preliminary design phase can significantly impact subsequent design stages, making it crucial to base these decisions on a solid foundation. A range of methods, including hand calculations, finite element analyses, and experimental tests, can be employed to establish the conceptual design parameters. However, these methods often come with the drawbacks of being time-intensive and costly, especially when testing various structures during the early design phase. To address this issue, this study introduces an artificial neural network-based design tool to evaluate the static structural characteristics of a helicopter’s horizontal stabilizer. The tool was built in Python using the Keras library. The required database for the training of the artificial neural network model was established using finite element analyses of the horizontal stabilizer subjected to the aerodynamic load for diverse design variables. The model’s performance was evaluated, and the model’s outputs were compared to the results derived from the finite element analyses. Moreover, the Hammersley sampling methodology was employed to reduce the size of the database without compromising on accuracy. The study also assessed the impact of decreasing the amount of data fed into the network model.</p>","PeriodicalId":18925,"journal":{"name":"Neural Computing and Applications","volume":"24 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-08-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142188290","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-23DOI: 10.1007/s00521-024-10327-7
Mengdie Lu, Haiyan Lu, Xinyu Hou, Qingyuan Hu
Arithmetic optimization algorithm (AOA) is a recently proposed algorithm inspired by mathematical operations. It has been used to solve a variety of optimization problems due to its simplicity of parameters and ease of implementation. However, it has been found that AOA encounters challenges such as poor exploration and premature convergence. To solve these issues, this paper proposes a self-adaptive AOA with hybrid search modes, named AOAHSM. In this algorithm, two hybrid search modes, i.e., the parallel search mode and the serial search mode, are established by combining AOA and differential evolution (DE) in different ways to enhance the exploration and exploitation abilities, respectively. In the parallel search mode, AOA and DE independently implement on their respective subpopulations to maintain a high distribution of the population. In the serial search mode, DE is embedded into AOA to provide more diversified solutions and thereby help the population jump out of local optima. Then, a self-adaptive conversion strategy is employed to dynamically switch between the two modes so as to achieve a better balance between exploration and exploitation. Additionally, a Levy flight strategy is used to perturb and update the best solution obtained in each iteration to further prevent premature convergence. Lastly, a binary version of AOAHSM is proposed to tackle the 0–1 knapsack problem. The proposed algorithms are evaluated on CEC2019, CEC2020 test functions, two typical engineering design problems and 45 instances of the 0–1 knapsack problem and compared with a number of state-of-the-art meta-heuristic algorithms. The obtained results demonstrate that AOAHSM and its binary version not only significantly outperform the original AOA but also achieve superior performance to the comparison algorithms in most cases.
{"title":"A self-adaptive arithmetic optimization algorithm with hybrid search modes for 0–1 knapsack problem","authors":"Mengdie Lu, Haiyan Lu, Xinyu Hou, Qingyuan Hu","doi":"10.1007/s00521-024-10327-7","DOIUrl":"https://doi.org/10.1007/s00521-024-10327-7","url":null,"abstract":"<p>Arithmetic optimization algorithm (AOA) is a recently proposed algorithm inspired by mathematical operations. It has been used to solve a variety of optimization problems due to its simplicity of parameters and ease of implementation. However, it has been found that AOA encounters challenges such as poor exploration and premature convergence. To solve these issues, this paper proposes a self-adaptive AOA with hybrid search modes, named AOAHSM. In this algorithm, two hybrid search modes, i.e., the parallel search mode and the serial search mode, are established by combining AOA and differential evolution (DE) in different ways to enhance the exploration and exploitation abilities, respectively. In the parallel search mode, AOA and DE independently implement on their respective subpopulations to maintain a high distribution of the population. In the serial search mode, DE is embedded into AOA to provide more diversified solutions and thereby help the population jump out of local optima. Then, a self-adaptive conversion strategy is employed to dynamically switch between the two modes so as to achieve a better balance between exploration and exploitation. Additionally, a Levy flight strategy is used to perturb and update the best solution obtained in each iteration to further prevent premature convergence. Lastly, a binary version of AOAHSM is proposed to tackle the 0–1 knapsack problem. The proposed algorithms are evaluated on CEC2019, CEC2020 test functions, two typical engineering design problems and 45 instances of the 0–1 knapsack problem and compared with a number of state-of-the-art meta-heuristic algorithms. The obtained results demonstrate that AOAHSM and its binary version not only significantly outperform the original AOA but also achieve superior performance to the comparison algorithms in most cases.</p>","PeriodicalId":18925,"journal":{"name":"Neural Computing and Applications","volume":"160 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-08-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142188289","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-22DOI: 10.1007/s00521-024-10158-6
Insha Majeed Wani, Sakshi Arora
Osteoporosis (OP) is the most prevalent and common bone disease, especially knee osteoporosis. It significantly disables sufferers all over the world. Although laborious and prone to user variation, manual diagnosis, segmentation, and annotation of knee joints continue to be the preferred way to diagnose OP in clinical procedures. Therefore, many deep learning algorithms, particularly the convolutional neural network (CNN), have been created to increase clinical workflow efficiency to overcome the shortcomings of the widely used method as above. Medical imaging procedures can show hidden structures in a volumetric view, particularly those that generate three-dimensional (3D) pictures like MRI. We created a dataset of 240 pictures from patients who had knee X-rays and skeletal bone mineral density assessments at the same time. Four convolutional neural networks (CNN) models were used to analyse the X-ray images and deep neural networks for clinical covariances to determine the degree of osteoporosis. Additionally, we investigated ensemble models that included each CNN with a clinical covariance. For every network, scores for accuracy and error rate were computed. ResNet and Alexnet displayed the highest levels of accuracy when the CNN models were tested using knee X-rays with normal, low BMD, and osteoporosis. An ensemble of DNN with Alexnet, ResNet, and both ResNet and Alexnet are employed resulting in improved accuracy. The ensemble of best-performing CNN and DNN is proposed to diagnose osteoporosis more accurately. The proposed method has produced a highly accurate osteoporosis diagnosis.
骨质疏松症(OP)是最普遍、最常见的骨病,尤其是膝关节骨质疏松症。骨质疏松症严重影响着全世界的患者。膝关节的人工诊断、分割和标注虽然费力且易受用户差异的影响,但仍是临床程序中诊断骨质疏松症的首选方法。因此,许多深度学习算法,特别是卷积神经网络(CNN)应运而生,以提高临床工作流程的效率,克服上述广泛使用的方法的缺点。医学成像程序可以显示容积视图中的隐藏结构,尤其是像核磁共振成像这样生成三维(3D)图片的程序。我们创建了一个包含 240 张图片的数据集,这些图片来自同时接受膝关节 X 光检查和骨骼骨矿密度评估的患者。我们使用四个卷积神经网络(CNN)模型分析 X 光图像,并使用深度神经网络分析临床协方差,以确定骨质疏松症的程度。此外,我们还研究了包含每个卷积神经网络和临床协方差的集合模型。我们计算了每个网络的准确率和错误率得分。在使用正常、低 BMD 和骨质疏松症的膝关节 X 光片对 CNN 模型进行测试时,ResNet 和 Alexnet 的准确率最高。使用包含 Alexnet、ResNet 以及 ResNet 和 Alexnet 的 DNN 集合提高了准确率。建议使用表现最佳的 CNN 和 DNN 的集合来更准确地诊断骨质疏松症。所提出的方法对骨质疏松症做出了高度准确的诊断。
{"title":"Deep ensemble learning for osteoporosis diagnosis from knee X-rays: a preliminary cohort study in Kashmir valley","authors":"Insha Majeed Wani, Sakshi Arora","doi":"10.1007/s00521-024-10158-6","DOIUrl":"https://doi.org/10.1007/s00521-024-10158-6","url":null,"abstract":"<p>Osteoporosis (OP) is the most prevalent and common bone disease, especially knee osteoporosis. It significantly disables sufferers all over the world. Although laborious and prone to user variation, manual diagnosis, segmentation, and annotation of knee joints continue to be the preferred way to diagnose OP in clinical procedures. Therefore, many deep learning algorithms, particularly the convolutional neural network (CNN), have been created to increase clinical workflow efficiency to overcome the shortcomings of the widely used method as above. Medical imaging procedures can show hidden structures in a volumetric view, particularly those that generate three-dimensional (3D) pictures like MRI. We created a dataset of 240 pictures from patients who had knee X-rays and skeletal bone mineral density assessments at the same time. Four convolutional neural networks (CNN) models were used to analyse the X-ray images and deep neural networks for clinical covariances to determine the degree of osteoporosis. Additionally, we investigated ensemble models that included each CNN with a clinical covariance. For every network, scores for accuracy and error rate were computed. ResNet and Alexnet displayed the highest levels of accuracy when the CNN models were tested using knee X-rays with normal, low BMD, and osteoporosis. An ensemble of DNN with Alexnet, ResNet, and both ResNet and Alexnet are employed resulting in improved accuracy. The ensemble of best-performing CNN and DNN is proposed to diagnose osteoporosis more accurately. The proposed method has produced a highly accurate osteoporosis diagnosis.</p>","PeriodicalId":18925,"journal":{"name":"Neural Computing and Applications","volume":"18 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-08-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142188347","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Ensuring environmental safety and regulatory compliance at Department of Energy (DOE) sites demands an efficient and reliable detection system for low-level nuclear waste (LLW). Unlike existing methods that rely on human effort, this paper explores the integration of computer vision algorithms to automate the identification of such waste across DOE facilities. We evaluate the effectiveness of multiple algorithms in classifying nuclear waste materials and their adaptability to newly emerging LLW. Our research introduces and implements five state-of-the-art computer vision models, each representing a different approach to the problem. Through rigorous experimentation and validation, we evaluate these algorithms based on performance, speed, and adaptability. The results reveal a noteworthy trade-off between detection performance and adaptability. YOLOv7 shows the best performance and requires the highest effort to detect new LLW. Conversely, OWL-ViT has lower performance than YOLOv7 and requires minimal effort to detect new LLW. The inference speed does not strongly correlate with performance or adaptability. These findings offer valuable insights into the strengths and limitations of current computer vision algorithms for LLW detection. Each developed model provides a specialized solution with distinct advantages and disadvantages, empowering DOE stakeholders to select the algorithm that aligns best with their specific needs.
要确保能源部(DOE)场址的环境安全和合规性,就需要一个高效可靠的低放射性核废料(LLW)检测系统。与依赖人力的现有方法不同,本文探讨了计算机视觉算法的集成,以自动识别 DOE 设施中的此类废物。我们评估了多种算法在核废料材料分类方面的有效性,以及它们对新出现的 LLW 的适应性。我们的研究引入并实施了五种最先进的计算机视觉模型,每种模型都代表了解决问题的不同方法。通过严格的实验和验证,我们根据性能、速度和适应性对这些算法进行了评估。结果显示,在检测性能和适应性之间存在值得注意的权衡。YOLOv7 的性能最好,但检测新 LLW 所需的工作量最大。相反,OWL-ViT 的性能比 YOLOv7 低,但检测新 LLW 所需的工作量却最小。推理速度与性能或适应性的关系不大。这些发现为了解当前计算机视觉算法在检测 LLW 方面的优势和局限性提供了宝贵的见解。每个开发的模型都提供了具有明显优缺点的专门解决方案,使 DOE 利益相关者能够选择最符合其特定需求的算法。
{"title":"AI-based detection and identification of low-level nuclear waste: a comparative analysis","authors":"Aris Duani Rojas, Leonel Lagos, Himanshu Upadhyay, Jayesh Soni, Nagarajan Prabakar","doi":"10.1007/s00521-024-10238-7","DOIUrl":"https://doi.org/10.1007/s00521-024-10238-7","url":null,"abstract":"<p>Ensuring environmental safety and regulatory compliance at Department of Energy (DOE) sites demands an efficient and reliable detection system for low-level nuclear waste (LLW). Unlike existing methods that rely on human effort, this paper explores the integration of computer vision algorithms to automate the identification of such waste across DOE facilities. We evaluate the effectiveness of multiple algorithms in classifying nuclear waste materials and their adaptability to newly emerging LLW. Our research introduces and implements five state-of-the-art computer vision models, each representing a different approach to the problem. Through rigorous experimentation and validation, we evaluate these algorithms based on performance, speed, and adaptability. The results reveal a noteworthy trade-off between detection performance and adaptability. YOLOv7 shows the best performance and requires the highest effort to detect new LLW. Conversely, OWL-ViT has lower performance than YOLOv7 and requires minimal effort to detect new LLW. The inference speed does not strongly correlate with performance or adaptability. These findings offer valuable insights into the strengths and limitations of current computer vision algorithms for LLW detection. Each developed model provides a specialized solution with distinct advantages and disadvantages, empowering DOE stakeholders to select the algorithm that aligns best with their specific needs.</p>","PeriodicalId":18925,"journal":{"name":"Neural Computing and Applications","volume":"37 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-08-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142188349","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-22DOI: 10.1007/s00521-024-10243-w
Sammy Kinga, Tamer F. Megahed, Haruichi Kanaya, Diaa-Eldin A. Mansour
Power electronic converters play a crucial role in integrating distributed generation, renewable energy sources, microgrids, and HVDC transmission networks into the grid. The control technique used in the voltage source inverters (VSI) is essential for handling load variations, system nonlinearity, stability, and fast transient response. This study focuses on improving the robustness and control performance of VSIs by integrating a Kalman filter adaptive observer into a finite control set model predictive control (FCS-MPC), resulting in an improved FCS-MPC strategy (IMPC). The classical FCS-MPC can be affected by inaccuracies due to measurement noise and uncertainties in system models, leading to less accurate predictions and suboptimal control actions. By employing the Kalman filter adaptive observer, real-time estimates of unmeasured variables are provided, compensating for uncertainties, and enhancing control performance. To further enhance flexibility and adaptivity, an artificial neural network (ANN)-based controller is designed. The ANN controller is trained offline using IMPC as baseline thus eliminating the need for online predictions and optimization. The ANN controller directly generates inverter switching configuration states, resulting in high-quality sinusoidal output voltage with low distortions. Comparative analysis is conducted for the classical FCS-MPC, IMPC, support vector machine (SVM), convolutional neural network (CNN), and ANN-based controllers under diverse operating conditions and system parameters. Although it has reduced interpretability, the ANN controller exhibits superior harmonic reduction, outperforming both MPC-based controllers and SVM. Evaluation against CNN-based controls also validates the ANN’s robustness and effectiveness in handling uncertainties, emphasizing its adaptability, efficiency, and practical applicability in power electronic applications.
电力电子变流器在将分布式发电、可再生能源、微电网和高压直流输电网络并入电网方面发挥着至关重要的作用。电压源逆变器(VSI)中使用的控制技术对于处理负载变化、系统非线性、稳定性和快速瞬态响应至关重要。本研究的重点是通过将卡尔曼滤波器自适应观测器集成到有限控制集模型预测控制(FCS-MPC)中,改进 FCS-MPC 策略(IMPC),从而提高 VSI 的鲁棒性和控制性能。传统的 FCS-MPC 可能会受到测量噪声和系统模型不确定性造成的不准确性的影响,导致预测不准确和控制行动不理想。通过采用卡尔曼滤波自适应观测器,可以实时估计未测量的变量,补偿不确定性,提高控制性能。为了进一步提高灵活性和适应性,设计了基于人工神经网络(ANN)的控制器。人工神经网络控制器以 IMPC 为基准进行离线训练,因此无需在线预测和优化。ANN 控制器直接生成逆变器开关配置状态,从而产生低失真、高质量的正弦输出电压。在不同的运行条件和系统参数下,对经典的 FCS-MPC、IMPC、支持向量机(SVM)、卷积神经网络(CNN)和基于 ANN 的控制器进行了比较分析。虽然解释性较差,但 ANN 控制器在减少谐波方面表现出色,优于基于 MPC 的控制器和 SVM。与基于 CNN 的控制器进行的评估还验证了 ANN 在处理不确定性时的鲁棒性和有效性,强调了其在电力电子应用中的适应性、效率和实际应用性。
{"title":"Enhancing robustness and control performance of voltage source inverters using Kalman filter adaptive observer and ANN-based model predictive controller","authors":"Sammy Kinga, Tamer F. Megahed, Haruichi Kanaya, Diaa-Eldin A. Mansour","doi":"10.1007/s00521-024-10243-w","DOIUrl":"https://doi.org/10.1007/s00521-024-10243-w","url":null,"abstract":"<p>Power electronic converters play a crucial role in integrating distributed generation, renewable energy sources, microgrids, and HVDC transmission networks into the grid. The control technique used in the voltage source inverters (VSI) is essential for handling load variations, system nonlinearity, stability, and fast transient response. This study focuses on improving the robustness and control performance of VSIs by integrating a Kalman filter adaptive observer into a finite control set model predictive control (FCS-MPC), resulting in an improved FCS-MPC strategy (IMPC). The classical FCS-MPC can be affected by inaccuracies due to measurement noise and uncertainties in system models, leading to less accurate predictions and suboptimal control actions. By employing the Kalman filter adaptive observer, real-time estimates of unmeasured variables are provided, compensating for uncertainties, and enhancing control performance. To further enhance flexibility and adaptivity, an artificial neural network (ANN)-based controller is designed. The ANN controller is trained offline using IMPC as baseline thus eliminating the need for online predictions and optimization. The ANN controller directly generates inverter switching configuration states, resulting in high-quality sinusoidal output voltage with low distortions. Comparative analysis is conducted for the classical FCS-MPC, IMPC, support vector machine (SVM), convolutional neural network (CNN), and ANN-based controllers under diverse operating conditions and system parameters. Although it has reduced interpretability, the ANN controller exhibits superior harmonic reduction, outperforming both MPC-based controllers and SVM. Evaluation against CNN-based controls also validates the ANN’s robustness and effectiveness in handling uncertainties, emphasizing its adaptability, efficiency, and practical applicability in power electronic applications.</p>","PeriodicalId":18925,"journal":{"name":"Neural Computing and Applications","volume":"7 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-08-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142188295","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-22DOI: 10.1007/s00521-024-10353-5
Shadi Alijani, Jamil Fayyad, Homayoun Najjaran
Deep learning models are often evaluated in scenarios where the data distribution is different from those used in the training and validation phases. The discrepancy presents a challenge for accurately predicting the performance of models once deployed on the target distribution. Domain adaptation and generalization are widely recognized as effective strategies for addressing such shifts, thereby ensuring reliable performance. The recent promising results in applying vision transformers in computer vision tasks, coupled with advancements in self-attention mechanisms, have demonstrated their significant potential for robustness and generalization in handling distribution shifts. Motivated by the increased interest from the research community, our paper investigates the deployment of vision transformers in domain adaptation and domain generalization scenarios. For domain adaptation methods, we categorize research into feature-level, instance-level, model-level adaptations, and hybrid approaches, along with other categorizations with respect to diverse strategies for enhancing domain adaptation. Similarly, for domain generalization, we categorize research into multi-domain learning, meta-learning, regularization techniques, and data augmentation strategies. We further classify diverse strategies in research, underscoring the various approaches researchers have taken to address distribution shifts by integrating vision transformers. The inclusion of comprehensive tables summarizing these categories is a distinct feature of our work, offering valuable insights for researchers. These findings highlight the versatility of vision transformers in managing distribution shifts, crucial for real-world applications, especially in critical safety and decision-making scenarios.
{"title":"Vision transformers in domain adaptation and domain generalization: a study of robustness","authors":"Shadi Alijani, Jamil Fayyad, Homayoun Najjaran","doi":"10.1007/s00521-024-10353-5","DOIUrl":"https://doi.org/10.1007/s00521-024-10353-5","url":null,"abstract":"<p>Deep learning models are often evaluated in scenarios where the data distribution is different from those used in the training and validation phases. The discrepancy presents a challenge for accurately predicting the performance of models once deployed on the target distribution. Domain adaptation and generalization are widely recognized as effective strategies for addressing such shifts, thereby ensuring reliable performance. The recent promising results in applying vision transformers in computer vision tasks, coupled with advancements in self-attention mechanisms, have demonstrated their significant potential for robustness and generalization in handling distribution shifts. Motivated by the increased interest from the research community, our paper investigates the deployment of vision transformers in domain adaptation and domain generalization scenarios. For domain adaptation methods, we categorize research into feature-level, instance-level, model-level adaptations, and hybrid approaches, along with other categorizations with respect to diverse strategies for enhancing domain adaptation. Similarly, for domain generalization, we categorize research into multi-domain learning, meta-learning, regularization techniques, and data augmentation strategies. We further classify diverse strategies in research, underscoring the various approaches researchers have taken to address distribution shifts by integrating vision transformers. The inclusion of comprehensive tables summarizing these categories is a distinct feature of our work, offering valuable insights for researchers. These findings highlight the versatility of vision transformers in managing distribution shifts, crucial for real-world applications, especially in critical safety and decision-making scenarios.</p>","PeriodicalId":18925,"journal":{"name":"Neural Computing and Applications","volume":"1197 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-08-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142188293","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
The rising incidence of cancer underscores the imperative for innovative diagnostic and prognostic methodologies. This study delves into the potential of RNA-Seq gene expression data to enhance cancer classification accuracy. Introducing a pioneering approach, we model gene expression data as point clouds, capitalizing on the data's intrinsic properties to bolster classification performance. Utilizing PointNet, a typical technique for processing point cloud data, as our framework's cornerstone, we incorporate inductive biases pertinent to gene expression and pathways. This integration markedly elevates model efficacy, culminating in developing an end-to-end deep learning classifier with an accuracy rate surpassing 99%. Our findings not only illuminate the capabilities of AI-driven models in the realm of oncology but also highlight the criticality of acknowledging biological dataset nuances in model design. This research provides insights into application of deep learning in medical science, setting the stage for further innovation in cancer classification through sophisticated biological data analysis. The source code for our study is accessible at: https://github.com/cialab/GPNet.
{"title":"Gene pointNet for tumor classification","authors":"Hao Lu, Mostafa Rezapour, Haseebullah Baha, Muhammad Khalid Khan Niazi, Aarthi Narayanan, Metin Nafi Gurcan","doi":"10.1007/s00521-024-10307-x","DOIUrl":"https://doi.org/10.1007/s00521-024-10307-x","url":null,"abstract":"<p>The rising incidence of cancer underscores the imperative for innovative diagnostic and prognostic methodologies. This study delves into the potential of RNA-Seq gene expression data to enhance cancer classification accuracy. Introducing a pioneering approach, we model gene expression data as point clouds, capitalizing on the data's intrinsic properties to bolster classification performance. Utilizing PointNet, a typical technique for processing point cloud data, as our framework's cornerstone, we incorporate inductive biases pertinent to gene expression and pathways. This integration markedly elevates model efficacy, culminating in developing an end-to-end deep learning classifier with an accuracy rate surpassing 99%. Our findings not only illuminate the capabilities of AI-driven models in the realm of oncology but also highlight the criticality of acknowledging biological dataset nuances in model design. This research provides insights into application of deep learning in medical science, setting the stage for further innovation in cancer classification through sophisticated biological data analysis. The source code for our study is accessible at: https://github.com/cialab/GPNet.</p>","PeriodicalId":18925,"journal":{"name":"Neural Computing and Applications","volume":"8 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-08-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142188288","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-22DOI: 10.1007/s00521-024-09813-9
Zhijun Ji
As internet information technology continues to advance, individuals are increasingly encountering and managing a vast volume of data and information. A large and complex amount of information hinders the effective transmission of valuable information, making it difficult to find multimedia content of interest in the vastness of the internet. As the volume of multimedia content rapidly grows, personalized recommendation algorithms play a crucial role in matching relevant content to users. Knowledge graphs, due to their powerful organizational and relationship processing capabilities, are commonly applied in intelligent search engines and recommendation systems. This article focuses on the effective utilization of semantic association information in knowledge graphs for multimedia content recommendation. Two main areas of research are conducted. In this article, two novel approaches are presented. To begin with, the primary objective is to improve the learning of knowledge feature representation. This is achieved by introducing a model based on self-attention, which effectively captures the diverse significance of triplets in determining the semantics of entities. This leads to improved quality of knowledge feature representation, thereby serving as valuable auxiliary information for multimedia content recommendation systems. Secondly, the article addresses the integration of knowledge graphs in multimedia content recommendation applications. This paper proposes a content recommendation algorithm that integrates a combined embedding of behavior and knowledge features. By leveraging past preferences and utilizing the semantic structure of knowledge graphs, this algorithm provides a comprehensive exploration of user interests and hobbies. Ultimately, this conducts extensive experiments to assess the effectiveness and performance of the proposed algorithms. The results validate the feasibility and efficacy of these algorithms in enhancing multimedia content recommendation systems.
{"title":"Multimedia content recommendation algorithm based on behavior and knowledge feature embedding","authors":"Zhijun Ji","doi":"10.1007/s00521-024-09813-9","DOIUrl":"https://doi.org/10.1007/s00521-024-09813-9","url":null,"abstract":"<p>As internet information technology continues to advance, individuals are increasingly encountering and managing a vast volume of data and information. A large and complex amount of information hinders the effective transmission of valuable information, making it difficult to find multimedia content of interest in the vastness of the internet. As the volume of multimedia content rapidly grows, personalized recommendation algorithms play a crucial role in matching relevant content to users. Knowledge graphs, due to their powerful organizational and relationship processing capabilities, are commonly applied in intelligent search engines and recommendation systems. This article focuses on the effective utilization of semantic association information in knowledge graphs for multimedia content recommendation. Two main areas of research are conducted. In this article, two novel approaches are presented. To begin with, the primary objective is to improve the learning of knowledge feature representation. This is achieved by introducing a model based on self-attention, which effectively captures the diverse significance of triplets in determining the semantics of entities. This leads to improved quality of knowledge feature representation, thereby serving as valuable auxiliary information for multimedia content recommendation systems. Secondly, the article addresses the integration of knowledge graphs in multimedia content recommendation applications. This paper proposes a content recommendation algorithm that integrates a combined embedding of behavior and knowledge features. By leveraging past preferences and utilizing the semantic structure of knowledge graphs, this algorithm provides a comprehensive exploration of user interests and hobbies. Ultimately, this conducts extensive experiments to assess the effectiveness and performance of the proposed algorithms. The results validate the feasibility and efficacy of these algorithms in enhancing multimedia content recommendation systems.</p>","PeriodicalId":18925,"journal":{"name":"Neural Computing and Applications","volume":"8 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-08-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142224561","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}