Amir Zarringhalam, Saeid Shiry Ghidary, Ali Mohades, Seyed-Ali Sadegh-Zadeh
We present a novel vector quantization (VQ) module for the two state-of-the-art long-range simultaneous localization and mapping (SLAM) algorithms. The VQ task in SLAM is generally performed using unsupervised methods. We provide an alternative approach trough embedding a semisupervised hyperbolic graph convolutional neural network (HGCN) in the VQ step of the SLAM processes. The SLAM platforms we have utilized for this purpose are fast appearance-based mapping (FABMAP) and oriented fast and rotated short (ORB), both of which rely on extracting the features of the captured images in their loop closure detection (LCD) module. For the first time, we have considered the space formed by these SURF features, robust image descriptors, as a graph, enabling us to apply an HGCN in the VQ section which results in an improved LCD performance. The HGCN vector quantizes the SURF feature space, leading to a bag-of-word (BoW) representation construction of the images. This representation is subsequently used to determine LCD accuracy and recall. Our approaches in this study are referred to as HGCN-FABMAP and HGCN-ORB. The main advantage of using HGCN in the LCD section is that it scales linearly when the features are accumulated. The benchmarking experiments show the superiority of our methods in terms of both trajectory generation accuracy in small-scale paths and LCD accuracy and recall for large-scale problems.
{"title":"Semisupervised Vector Quantization in Visual SLAM Using HGCN","authors":"Amir Zarringhalam, Saeid Shiry Ghidary, Ali Mohades, Seyed-Ali Sadegh-Zadeh","doi":"10.1155/2024/9992159","DOIUrl":"https://doi.org/10.1155/2024/9992159","url":null,"abstract":"<p>We present a novel vector quantization (VQ) module for the two state-of-the-art long-range simultaneous localization and mapping (SLAM) algorithms. The VQ task in SLAM is generally performed using unsupervised methods. We provide an alternative approach trough embedding a semisupervised hyperbolic graph convolutional neural network (HGCN) in the VQ step of the SLAM processes. The SLAM platforms we have utilized for this purpose are fast appearance-based mapping (FABMAP) and oriented fast and rotated short (ORB), both of which rely on extracting the features of the captured images in their loop closure detection (LCD) module. For the first time, we have considered the space formed by these SURF features, robust image descriptors, as a graph, enabling us to apply an HGCN in the VQ section which results in an improved LCD performance. The HGCN vector quantizes the SURF feature space, leading to a bag-of-word (BoW) representation construction of the images. This representation is subsequently used to determine LCD accuracy and recall. Our approaches in this study are referred to as HGCN-FABMAP and HGCN-ORB. The main advantage of using HGCN in the LCD section is that it scales linearly when the features are accumulated. The benchmarking experiments show the superiority of our methods in terms of both trajectory generation accuracy in small-scale paths and LCD accuracy and recall for large-scale problems.</p>","PeriodicalId":14089,"journal":{"name":"International Journal of Intelligent Systems","volume":null,"pages":null},"PeriodicalIF":7.0,"publicationDate":"2024-02-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141165002","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Fan Wu, Liangrui Wei, Haiyong Luo, Fang Zhao, Xin Ma, Bokun Ning
GNSS (global navigation satellite systems) technology enables high-precision single-point positioning (SPP) in open environments. However, the accuracy of GNSS positioning is significantly compromised in complex urban canyons due to signal obstructions and non-line-of-sight propagation errors. To address this challenge, we propose a GNSS displacement estimation algorithm. This method learns nonlinear dependencies between GNSS raw measurements and corresponding position changes, capturing dynamic and layered features in GNSS measurement data for displacement estimation. We introduce a denoising auto-encoder (DAE) to preprocess raw GNSS observations, reducing the impact of noise. The model simultaneously outputs estimated displacement and model confidence. The fusion process dynamically combines positioning results from the SPP algorithm and the D-Tran model, adaptively blending them to achieve accurate and optimal positioning estimation. This approach optimizes the accuracy of estimated positioning results while maintaining confidence in the estimation. Experimental results show a 61% reduction in root mean square error (RMSE) and 100% availability in urban canyon environments compared to traditional single-point positioning techniques.
{"title":"T-SPP: Improving GNSS Single-Point Positioning Performance Using Transformer-Based Correction","authors":"Fan Wu, Liangrui Wei, Haiyong Luo, Fang Zhao, Xin Ma, Bokun Ning","doi":"10.1155/2024/6643723","DOIUrl":"10.1155/2024/6643723","url":null,"abstract":"<p>GNSS (global navigation satellite systems) technology enables high-precision single-point positioning (SPP) in open environments. However, the accuracy of GNSS positioning is significantly compromised in complex urban canyons due to signal obstructions and non-line-of-sight propagation errors. To address this challenge, we propose a GNSS displacement estimation algorithm. This method learns nonlinear dependencies between GNSS raw measurements and corresponding position changes, capturing dynamic and layered features in GNSS measurement data for displacement estimation. We introduce a denoising auto-encoder (DAE) to preprocess raw GNSS observations, reducing the impact of noise. The model simultaneously outputs estimated displacement and model confidence. The fusion process dynamically combines positioning results from the SPP algorithm and the D-Tran model, adaptively blending them to achieve accurate and optimal positioning estimation. This approach optimizes the accuracy of estimated positioning results while maintaining confidence in the estimation. Experimental results show a 61% reduction in root mean square error (RMSE) and 100% availability in urban canyon environments compared to traditional single-point positioning techniques.</p>","PeriodicalId":14089,"journal":{"name":"International Journal of Intelligent Systems","volume":null,"pages":null},"PeriodicalIF":7.0,"publicationDate":"2024-02-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140416063","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Flood disasters occur worldwide, and flood risk prediction is conducive to protecting human life and property safety. Influenced by topographic changes and rainfall, the water level fluctuates randomly and violently during the flood, introducing many noises and directly increasing the difficulty of flood prediction. A data-driven flood forecasting method is proposed based on data preprocessing and a two-layer BiLSTM-Attention network to improve forecast accuracy. First, the Variational Mode Decomposition (VMD) is used to decompose the data for reducing noise and produce suitable Intrinsic Mode Functions (IMFs); Then, an optimized two-layer attention-based Bidirectional Long Sshort-Term memory (BiLSTM-Attention) network is constructed to predict each IMF. Finally, two optimization algorithms are used to obtain the optimized parameters of VMD and BiLSTM intelligently, increasing the self-adaptability. The inertia factor of particle swarm optimization is improved and then used to optimize the five hyperparameters of BiLSTM. The proposed model reduces storage errors for smaller training sets and can achieve good performance. Three water level data sets from the Yangtze River in China are used for comparative experiments. Numerical results show that the peak height absolute error is within 2 cm, and the relative error of peak time arrival is within 30%. Compared with LSTM, BiLSTM, CNN-BiLSTM-attention, etc., the proposed model reduces the root mean square error by at least 50% and has advantages for high-risk forecasting when the water level exceeds the defense line and fluctuates prominently.
{"title":"A Data-Driven Method and Hybrid Deep Learning Model for Flood Risk Prediction","authors":"Chenmin Ni, Pei Shan Fam, Muhammad Fadhil Marsani","doi":"10.1155/2024/3562709","DOIUrl":"10.1155/2024/3562709","url":null,"abstract":"<p>Flood disasters occur worldwide, and flood risk prediction is conducive to protecting human life and property safety. Influenced by topographic changes and rainfall, the water level fluctuates randomly and violently during the flood, introducing many noises and directly increasing the difficulty of flood prediction. A data-driven flood forecasting method is proposed based on data preprocessing and a two-layer BiLSTM-Attention network to improve forecast accuracy. First, the Variational Mode Decomposition (VMD) is used to decompose the data for reducing noise and produce suitable Intrinsic Mode Functions (IMFs); Then, an optimized two-layer attention-based Bidirectional Long Sshort-Term memory (BiLSTM-Attention) network is constructed to predict each IMF. Finally, two optimization algorithms are used to obtain the optimized parameters of VMD and BiLSTM intelligently, increasing the self-adaptability. The inertia factor of particle swarm optimization is improved and then used to optimize the five hyperparameters of BiLSTM. The proposed model reduces storage errors for smaller training sets and can achieve good performance. Three water level data sets from the Yangtze River in China are used for comparative experiments. Numerical results show that the peak height absolute error is within 2 cm, and the relative error of peak time arrival is within 30%. Compared with LSTM, BiLSTM, CNN-BiLSTM-attention, etc., the proposed model reduces the root mean square error by at least 50% and has advantages for high-risk forecasting when the water level exceeds the defense line and fluctuates prominently.</p>","PeriodicalId":14089,"journal":{"name":"International Journal of Intelligent Systems","volume":null,"pages":null},"PeriodicalIF":7.0,"publicationDate":"2024-02-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140425827","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Despite the rapid growth of machine learning (ML) and its far-reaching applications in various fields such as healthcare, finance, and urban heat management, there are still some unresolved challenges in the field of climate change. Reliable subseasonal forecasts of summer temperatures would be a great benefit to society. Although numerical weather prediction (NWP) models are better at capturing relevant sources of predictability, such as temperatures, land, and sea surface conditions, the subseasonal potential is not fully exploited. One such challenge is accurate subseasonal temperature forecasting using cutting-edge ML technology. This study aims to assess and predict the changes in subseasonal temperature during the summer season (from March to June) in Senegal on 2-weeks time scales. Six ML techniques, including linear regression (LR), decision tree (DT), support vector machine (SVM), artificial neural network (ANN), long short-term memory (LSTM), and gated recurrent units (GRU), are used. The experiments utilize a multivariate approach by incorporating variables of the ERA-5 dataset from 1981 to 2022. The results compared all the performances of the methods to assess their overall effectiveness in forecasting air temperature (t2m) values over 2 weeks. Our analysis demonstrates that the GRU model outperforms the other ML models, achieving a Nash–Sutcliffe efficiency (NSE) score of 74.68% and a mean absolute percentage error (MAPE) of 2.51%. The GRU model effectively captures long-term dependencies and exhibits superior performance in temperature forecasting. Furthermore, a comparison between the observed and predicted values confirms the accuracy of the GRU model in aligning with actual temperature trends. Overall, this study contributes an impactful deep learning model to the field of subseasonal temperature forecasting in West Africa (Senegal), which offers local authorities the capability to anticipate climatic events and enact preventive measures accordingly.
尽管机器学习(ML)发展迅速,并在医疗保健、金融和城市供热管理等多个领域得到了广泛应用,但在气候变化领域仍存在一些尚未解决的难题。对夏季气温进行可靠的分季节预报将给社会带来极大的好处。尽管数值天气预报(NWP)模式能更好地捕捉相关的可预测性来源,如气温、陆地和海洋表面条件,但其亚季节潜力并未得到充分利用。其中一个挑战就是利用最先进的 ML 技术准确预报分季节气温。本研究旨在评估和预测塞内加尔夏季(3 月至 6 月)2 周时间尺度上的分季节气温变化。研究采用了六种 ML 技术,包括线性回归 (LR)、决策树 (DT)、支持向量机 (SVM)、人工神经网络 (ANN)、长短期记忆 (LSTM) 和门控递归单元 (GRU)。实验采用多元方法,纳入了 1981 年至 2022 年 ERA-5 数据集的变量。结果比较了所有方法的性能,以评估它们在预报两周内气温(t2m)值方面的总体效果。我们的分析表明,GRU 模型优于其他 ML 模型,其纳什-苏特克利夫效率(NSE)为 74.68%,平均绝对百分比误差(MAPE)为 2.51%。GRU 模型有效地捕捉了长期依赖关系,在气温预测方面表现出卓越的性能。此外,观测值和预测值之间的比较也证实了 GRU 模型在与实际气温趋势保持一致方面的准确性。总之,本研究为西非(塞内加尔)的分季节气温预报领域贡献了一个有影响力的深度学习模型,为当地政府提供了预测气候事件并制定相应预防措施的能力。
{"title":"Subseasonal Prediction of Summer Temperature in West Africa Using Artificial Intelligence: A Case Study of Senegal","authors":"Annine Duclaire Kenne, Mory Toure, Lema Logamou Seknewna, Herve Landry Ketsemen","doi":"10.1155/2024/8869267","DOIUrl":"10.1155/2024/8869267","url":null,"abstract":"<p>Despite the rapid growth of machine learning (ML) and its far-reaching applications in various fields such as healthcare, finance, and urban heat management, there are still some unresolved challenges in the field of climate change. Reliable subseasonal forecasts of summer temperatures would be a great benefit to society. Although numerical weather prediction (NWP) models are better at capturing relevant sources of predictability, such as temperatures, land, and sea surface conditions, the subseasonal potential is not fully exploited. One such challenge is accurate subseasonal temperature forecasting using cutting-edge ML technology. This study aims to assess and predict the changes in subseasonal temperature during the summer season (from March to June) in Senegal on 2-weeks time scales. Six ML techniques, including linear regression (LR), decision tree (DT), support vector machine (SVM), artificial neural network (ANN), long short-term memory (LSTM), and gated recurrent units (GRU), are used. The experiments utilize a multivariate approach by incorporating variables of the ERA-5 dataset from 1981 to 2022. The results compared all the performances of the methods to assess their overall effectiveness in forecasting air temperature (t2m) values over 2 weeks. Our analysis demonstrates that the GRU model outperforms the other ML models, achieving a Nash–Sutcliffe efficiency (NSE) score of 74.68% and a mean absolute percentage error (MAPE) of 2.51%. The GRU model effectively captures long-term dependencies and exhibits superior performance in temperature forecasting. Furthermore, a comparison between the observed and predicted values confirms the accuracy of the GRU model in aligning with actual temperature trends. Overall, this study contributes an impactful deep learning model to the field of subseasonal temperature forecasting in West Africa (Senegal), which offers local authorities the capability to anticipate climatic events and enact preventive measures accordingly.</p>","PeriodicalId":14089,"journal":{"name":"International Journal of Intelligent Systems","volume":null,"pages":null},"PeriodicalIF":7.0,"publicationDate":"2024-02-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140427583","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Mohammad Hossein Moghimi Esfand-Abadi, Mohammad Hassan Djavareshkian, Afshin Madani
In the present study, the effects of the wing fence on the wing tip vortices and control surfaces located at the tip of the wing in a flying wing aircraft have been investigated using a numerical method. For the size of the fences, the average dimensions extracted from the wing tip vortices at different angles of attack are used. The basic determining parameter is the rolling torque coefficient, which is tried to be shown by employing a parametric study of the flow behavior in different situations of fence placement. These effects on the rolling torque of the aircraft are measured due to the presence of the split drag rudder control system. In this study, the fences were installed at three different heights and three different positions along the length of the wing, which were investigated at angles of attack of 7 to 16 degrees. The next stage of the research is to design the dimensions of the fence using the single-objective optimization method (a method to find the best solution for a problem with a specific goal). The designing of the fences at three points based on the dimensions of the wing tip vortex is carried out with the computational fluid dynamics (CFD) method (CFD is a computational method that uses physical laws to predict the behavior of fluids.). The aim of this research is to achieve the best design that converges to an optimal solution with minimum time and cost (CFD solution is long). However, CFD analysis requires a lot of computational time. To address this challenge, we employed a hybrid learning model comprising the radial basis function (RBF), a type of artificial neural network, and Kriging, a Gaussian process-based interpolation technique. The dataset for training the hybrid model was obtained from numerical solutions of CFD simulations involving a fence placed at various locations on the wing. Additionally, a genetic algorithm was employed as the optimization method in all instances where it was required. Using the power of machine learning techniques helped us identify the optimal placement of the fence to prevent it from being engulfed by the vortex and to optimize the utilization of the split drag system, yielding significant improvements.
{"title":"Kriging and Radial Basis Function Models for Optimized Design of UAV Wing Fences to Reduce Rolling Moment","authors":"Mohammad Hossein Moghimi Esfand-Abadi, Mohammad Hassan Djavareshkian, Afshin Madani","doi":"10.1155/2024/4108121","DOIUrl":"https://doi.org/10.1155/2024/4108121","url":null,"abstract":"<p>In the present study, the effects of the wing fence on the wing tip vortices and control surfaces located at the tip of the wing in a flying wing aircraft have been investigated using a numerical method. For the size of the fences, the average dimensions extracted from the wing tip vortices at different angles of attack are used. The basic determining parameter is the rolling torque coefficient, which is tried to be shown by employing a parametric study of the flow behavior in different situations of fence placement. These effects on the rolling torque of the aircraft are measured due to the presence of the split drag rudder control system. In this study, the fences were installed at three different heights and three different positions along the length of the wing, which were investigated at angles of attack of 7 to 16 degrees. The next stage of the research is to design the dimensions of the fence using the single-objective optimization method (a method to find the best solution for a problem with a specific goal). The designing of the fences at three points based on the dimensions of the wing tip vortex is carried out with the computational fluid dynamics (CFD) method (CFD is a computational method that uses physical laws to predict the behavior of fluids.). The aim of this research is to achieve the best design that converges to an optimal solution with minimum time and cost (CFD solution is long). However, CFD analysis requires a lot of computational time. To address this challenge, we employed a hybrid learning model comprising the radial basis function (RBF), a type of artificial neural network, and Kriging, a Gaussian process-based interpolation technique. The dataset for training the hybrid model was obtained from numerical solutions of CFD simulations involving a fence placed at various locations on the wing. Additionally, a genetic algorithm was employed as the optimization method in all instances where it was required. Using the power of machine learning techniques helped us identify the optimal placement of the fence to prevent it from being engulfed by the vortex and to optimize the utilization of the split drag system, yielding significant improvements.</p>","PeriodicalId":14089,"journal":{"name":"International Journal of Intelligent Systems","volume":null,"pages":null},"PeriodicalIF":7.0,"publicationDate":"2024-02-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141164964","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Shaymaa E. Sorour, Abeer A. Wafa, Amr A. Abohany, Reda M. Hussien
The potential of technology to revolutionize healthcare is exemplified by the synergy between artificial intelligence (AI) and early detection of cardiomegaly, demonstrating the power of proactive intervention in cardiovascular health. This paper presents an innovative approach that leverages advanced AI algorithms, specifically deep learning (DL) technology, for the early detection of cardiomegaly. The methodology consists of five key steps, including data collection, image preprocessing, data augmentation, feature extraction, and classification. Utilizing chest X-ray (CXR) images from the National Institutes of Health (NIH), the study applies rigorous image preprocessing operations, including color transformation and normalization. To enhance model generalization, data augmentation is employed, paving the way for two distinct DL models, a convolutional neural network (CNN) developed from scratch and a pretrained residual network with 50 layers (ResNet50), and adapted to the problem domain. Both models are systematically evaluated with five optimizers, revealing the AdaMax optimizer’s superiority for the CNN model and AdaGrad’s efficacy for the modified ResNet50. The proposed CNN with AdaMax achieves an impressive 99.91% accuracy, outperforming recent techniques in precision, recall, and F1 − score. This research underscores the transformative potential of AI in cardiovascular health diagnostics, emphasizing the significance of timely intervention.
{"title":"A Deep Learning System for Detecting Cardiomegaly Disease Based on CXR Image","authors":"Shaymaa E. Sorour, Abeer A. Wafa, Amr A. Abohany, Reda M. Hussien","doi":"10.1155/2024/8997093","DOIUrl":"10.1155/2024/8997093","url":null,"abstract":"<p>The potential of technology to revolutionize healthcare is exemplified by the synergy between artificial intelligence (AI) and early detection of cardiomegaly, demonstrating the power of proactive intervention in cardiovascular health. This paper presents an innovative approach that leverages advanced AI algorithms, specifically deep learning (DL) technology, for the early detection of cardiomegaly. The methodology consists of five key steps, including data collection, image preprocessing, data augmentation, feature extraction, and classification. Utilizing chest X-ray (CXR) images from the National Institutes of Health (NIH), the study applies rigorous image preprocessing operations, including color transformation and normalization. To enhance model generalization, data augmentation is employed, paving the way for two distinct DL models, a convolutional neural network (CNN) developed from scratch and a pretrained residual network with 50 layers (ResNet50), and adapted to the problem domain. Both models are systematically evaluated with five optimizers, revealing the AdaMax optimizer’s superiority for the CNN model and AdaGrad’s efficacy for the modified ResNet50. The proposed CNN with AdaMax achieves an impressive 99.91% accuracy, outperforming recent techniques in precision, recall, and <i>F</i>1 − score. This research underscores the transformative potential of AI in cardiovascular health diagnostics, emphasizing the significance of timely intervention.</p>","PeriodicalId":14089,"journal":{"name":"International Journal of Intelligent Systems","volume":null,"pages":null},"PeriodicalIF":7.0,"publicationDate":"2024-02-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140435516","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Suyu Wang, Quan Yue, Mengyu Zhao, Huazhi Zhang, Yan Li
Facing the implementation problems such like low growth reward, long training time, and poor stability of the multiagent learning methods when dealing with complex environment and more agents, this paper proposes a fast optimal coordination method for multiagent in complex environment (FOC-MACE). Firstly, the environment exploration strategy is introduced into the policy network based on the MADDPG method for higher growth rewards. Then, the parallel computing technology is adopted in the critic network, in purpose to effectively reduce the training time. These tactics together are beneficial to enhance the stability of multiagent learning. Lastly, the optimal resource allocation is carried out to realize optimal coevolution of the multiagents and further improve the learning ability of the agents’ group. To verify the effectiveness of our proposal, the FOC-MACE is compared with several advanced methods at current stage in the MPE environment. Three different experiments prove that by using our method, the growth reward is increased by up to 37.1%, the training is speed up significantly, and the stability of the method, which represented by standardized variance, is also improved. In addition, this paper validated the fast optimal coordination method for multiagent systems in the context of UAV scenarios, demonstrating the practical performance of the approach. Through comprehensive experiments and scenario validations, the study successfully confirmed the effectiveness of the proposed fast optimal coordination method for multiagent systems in complex environments.
{"title":"A Fast Optimal Coordination Method for Multiagent in Complex Environment","authors":"Suyu Wang, Quan Yue, Mengyu Zhao, Huazhi Zhang, Yan Li","doi":"10.1155/2024/5346187","DOIUrl":"10.1155/2024/5346187","url":null,"abstract":"<p>Facing the implementation problems such like low growth reward, long training time, and poor stability of the multiagent learning methods when dealing with complex environment and more agents, this paper proposes a fast optimal coordination method for multiagent in complex environment (FOC-MACE). Firstly, the environment exploration strategy is introduced into the policy network based on the MADDPG method for higher growth rewards. Then, the parallel computing technology is adopted in the critic network, in purpose to effectively reduce the training time. These tactics together are beneficial to enhance the stability of multiagent learning. Lastly, the optimal resource allocation is carried out to realize optimal coevolution of the multiagents and further improve the learning ability of the agents’ group. To verify the effectiveness of our proposal, the FOC-MACE is compared with several advanced methods at current stage in the MPE environment. Three different experiments prove that by using our method, the growth reward is increased by up to 37.1%, the training is speed up significantly, and the stability of the method, which represented by standardized variance, is also improved. In addition, this paper validated the fast optimal coordination method for multiagent systems in the context of UAV scenarios, demonstrating the practical performance of the approach. Through comprehensive experiments and scenario validations, the study successfully confirmed the effectiveness of the proposed fast optimal coordination method for multiagent systems in complex environments.</p>","PeriodicalId":14089,"journal":{"name":"International Journal of Intelligent Systems","volume":null,"pages":null},"PeriodicalIF":7.0,"publicationDate":"2024-02-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140446959","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Breast cancer characterization remains a significant and challenging issue in contemporary medicine. Accurately distinguishing between malignant and benign breast lesions is crucial for effective diagnosis and treatment. The anatomical structure of malignant breast ultrasound images is more chaotic than that of benign images due to disease pathologies. However, texture-based analysis alone often fails to identify the extent of chaoticness in malignant breast ultrasound images due to their vague appearance with normal echo patterns, leading to missed diagnoses and increased mortality rates. To address this issue, we proposed an angular feature-based multilevel breast cancer classification framework that aims to improve the accuracy and efficiency of classification. The proposed framework mimics the radiologist interpretation procedure by identifying the chaoticness on the periphery of the breast lesion in a breast ultrasound image (level-1). If the lesion contains an acute angle in any part of the periphery, it can be characterized as malignant or otherwise benign. However, solely relying on level-1 analysis may result in misclassification, especially when benign lesions exhibit echo patterns that resemble malignant ones. To overcome this limitation and to make the proposed system highly sensitive, advanced texture-based analysis (using combined shape, texture, and angular features) is performed (level-2). Finally, the performance of the proposed system is evaluated using a cross-dataset (consisting of 1293 breast ultrasound images) and compared with the different individual feature extraction techniques. Encouragingly, our system demonstrated an accuracy of 96.99% for classifying malignant and benign tumors, which is also validated using statistical analysis. The implications of our research lie in its potential to significantly improve breast cancer diagnosis by providing a reliable, efficient, and sensitive tool for radiologists.
{"title":": Multilevel Breast Cancer Classification Framework Using Radiomic Features","authors":"Lipismita Panigrahi, Tej Bahadur Chandra, Atul Kumar Srivastava, Neeraj Varshney, Kamred Udham Singh, Shambhu Mahato","doi":"10.1155/2024/6631016","DOIUrl":"10.1155/2024/6631016","url":null,"abstract":"<p>Breast cancer characterization remains a significant and challenging issue in contemporary medicine. Accurately distinguishing between malignant and benign breast lesions is crucial for effective diagnosis and treatment. The anatomical structure of malignant breast ultrasound images is more chaotic than that of benign images due to disease pathologies. However, texture-based analysis alone often fails to identify the extent of chaoticness in malignant breast ultrasound images due to their vague appearance with normal echo patterns, leading to missed diagnoses and increased mortality rates. To address this issue, we proposed an angular feature-based multilevel breast cancer classification framework <span></span><math></math> that aims to improve the accuracy and efficiency of classification. The proposed framework mimics the radiologist interpretation procedure by identifying the chaoticness on the periphery of the breast lesion in a breast ultrasound image (level-1). If the lesion contains an acute angle in any part of the periphery, it can be characterized as malignant or otherwise benign. However, solely relying on level-1 analysis may result in misclassification, especially when benign lesions exhibit echo patterns that resemble malignant ones. To overcome this limitation and to make the proposed system highly sensitive, advanced texture-based analysis (using combined shape, texture, and angular features) is performed (level-2). Finally, the performance of the proposed system is evaluated using a cross-dataset (consisting of 1293 breast ultrasound images) and compared with the different individual feature extraction techniques. Encouragingly, our system demonstrated an accuracy of 96.99% for classifying malignant and benign tumors, which is also validated using statistical analysis. The implications of our research lie in its potential to significantly improve breast cancer diagnosis by providing a reliable, efficient, and sensitive tool for radiologists.</p>","PeriodicalId":14089,"journal":{"name":"International Journal of Intelligent Systems","volume":null,"pages":null},"PeriodicalIF":7.0,"publicationDate":"2024-02-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139960044","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Yongfei Zhang, Xinying Lin, Hong Yang, Jie He, Linbo Qing, Xiaohai He, Yi Li, Honggang Chen
In recent years, remarkable performance improvements have been produced by deep convolutional neural networks (CNN) for single image super-resolution (SISR). Nevertheless, a high proportion of CNN-based SISR models are with quite a few network parameters and high computational complexity for deep or wide architectures. How to more fully utilize deep features to make a balance between model complexity and reconstruction performance is one of the main challenges in this field. To address this problem, on the basis of the well-known information multi-distillation model, a multi-attention feature distillation network termed as MAFDN is developed for lightweight and accurate SISR. Specifically, an effective multi-attention feature distillation block (MAFDB) is designed and used as the basic feature extraction unit in MAFDN. With the help of multi-attention layers including pixel attention, spatial attention, and channel attention, MAFDB uses multiple information distillation branches to learn more discriminative and representative features. Furthermore, MAFDB introduces the depthwise over-parameterized convolutional layer (DO-Conv)-based residual block (OPCRB) to enhance its ability without incurring any parameter and computation increase in the inference stage. The results on commonly used datasets demonstrate that our MAFDN outperforms existing representative lightweight SISR models when taking both reconstruction performance and model complexity into consideration. For example, for ×4 SR on Set5, MAFDN (597K/33.79G) obtains 0.21 dB/0.0037 and 0.10 dB/0.0015 PSNR/SSIM gains over the attention-based SR model AFAN (692K/50.90G) and the feature distillation-based SR model DDistill-SR (675K/32.83G), respectively.
{"title":"A Multi-Attention Feature Distillation Neural Network for Lightweight Single Image Super-Resolution","authors":"Yongfei Zhang, Xinying Lin, Hong Yang, Jie He, Linbo Qing, Xiaohai He, Yi Li, Honggang Chen","doi":"10.1155/2024/3255233","DOIUrl":"10.1155/2024/3255233","url":null,"abstract":"<p>In recent years, remarkable performance improvements have been produced by deep convolutional neural networks (CNN) for single image super-resolution (SISR). Nevertheless, a high proportion of CNN-based SISR models are with quite a few network parameters and high computational complexity for deep or wide architectures. How to more fully utilize deep features to make a balance between model complexity and reconstruction performance is one of the main challenges in this field. To address this problem, on the basis of the well-known information multi-distillation model, a multi-attention feature distillation network termed as MAFDN is developed for lightweight and accurate SISR. Specifically, an effective multi-attention feature distillation block (MAFDB) is designed and used as the basic feature extraction unit in MAFDN. With the help of multi-attention layers including pixel attention, spatial attention, and channel attention, MAFDB uses multiple information distillation branches to learn more discriminative and representative features. Furthermore, MAFDB introduces the depthwise over-parameterized convolutional layer (DO-Conv)-based residual block (OPCRB) to enhance its ability without incurring any parameter and computation increase in the inference stage. The results on commonly used datasets demonstrate that our MAFDN outperforms existing representative lightweight SISR models when taking both reconstruction performance and model complexity into consideration. For example, for ×4 SR on Set5, MAFDN (597K/33.79G) obtains 0.21 dB/0.0037 and 0.10 dB/0.0015 PSNR/SSIM gains over the attention-based SR model AFAN (692K/50.90G) and the feature distillation-based SR model DDistill-SR (675K/32.83G), respectively.</p>","PeriodicalId":14089,"journal":{"name":"International Journal of Intelligent Systems","volume":null,"pages":null},"PeriodicalIF":7.0,"publicationDate":"2024-02-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139776257","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Yongfei Zhang, Xinying Lin, Hong Yang, Jie He, Linbo Qing, Xiaohai He, Yi Li, Honggang Chen
In recent years, remarkable performance improvements have been produced by deep convolutional neural networks (CNN) for single image super-resolution (SISR). Nevertheless, a high proportion of CNN-based SISR models are with quite a few network parameters and high computational complexity for deep or wide architectures. How to more fully utilize deep features to make a balance between model complexity and reconstruction performance is one of the main challenges in this field. To address this problem, on the basis of the well-known information multi-distillation model, a multi-attention feature distillation network termed as MAFDN is developed for lightweight and accurate SISR. Specifically, an effective multi-attention feature distillation block (MAFDB) is designed and used as the basic feature extraction unit in MAFDN. With the help of multi-attention layers including pixel attention, spatial attention, and channel attention, MAFDB uses multiple information distillation branches to learn more discriminative and representative features. Furthermore, MAFDB introduces the depthwise over-parameterized convolutional layer (DO-Conv)-based residual block (OPCRB) to enhance its ability without incurring any parameter and computation increase in the inference stage. The results on commonly used datasets demonstrate that our MAFDN outperforms existing representative lightweight SISR models when taking both reconstruction performance and model complexity into consideration. For example, for ×4 SR on Set5, MAFDN (597K/33.79G) obtains 0.21 dB/0.0037 and 0.10 dB/0.0015 PSNR/SSIM gains over the attention-based SR model AFAN (692K/50.90G) and the feature distillation-based SR model DDistill-SR (675K/32.83G), respectively.
{"title":"A Multi-Attention Feature Distillation Neural Network for Lightweight Single Image Super-Resolution","authors":"Yongfei Zhang, Xinying Lin, Hong Yang, Jie He, Linbo Qing, Xiaohai He, Yi Li, Honggang Chen","doi":"10.1155/2024/3255233","DOIUrl":"https://doi.org/10.1155/2024/3255233","url":null,"abstract":"In recent years, remarkable performance improvements have been produced by deep convolutional neural networks (CNN) for single image super-resolution (SISR). Nevertheless, a high proportion of CNN-based SISR models are with quite a few network parameters and high computational complexity for deep or wide architectures. How to more fully utilize deep features to make a balance between model complexity and reconstruction performance is one of the main challenges in this field. To address this problem, on the basis of the well-known information multi-distillation model, a multi-attention feature distillation network termed as MAFDN is developed for lightweight and accurate SISR. Specifically, an effective multi-attention feature distillation block (MAFDB) is designed and used as the basic feature extraction unit in MAFDN. With the help of multi-attention layers including pixel attention, spatial attention, and channel attention, MAFDB uses multiple information distillation branches to learn more discriminative and representative features. Furthermore, MAFDB introduces the depthwise over-parameterized convolutional layer (DO-Conv)-based residual block (OPCRB) to enhance its ability without incurring any parameter and computation increase in the inference stage. The results on commonly used datasets demonstrate that our MAFDN outperforms existing representative lightweight SISR models when taking both reconstruction performance and model complexity into consideration. For example, for ×4 SR on Set5, MAFDN (597K/33.79G) obtains 0.21 dB/0.0037 and 0.10 dB/0.0015 PSNR/SSIM gains over the attention-based SR model AFAN (692K/50.90G) and the feature distillation-based SR model DDistill-SR (675K/32.83G), respectively.","PeriodicalId":14089,"journal":{"name":"International Journal of Intelligent Systems","volume":null,"pages":null},"PeriodicalIF":7.0,"publicationDate":"2024-02-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139836013","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}