Pub Date : 2024-03-06DOI: 10.1080/0954898X.2024.2309947
Sangeetha Alagumani, Uma Maheswari Natarajan
The 5th generation (5 G) network is required to meet the growing demand for fast data speeds and the expanding number of customers. Apart from offering higher speeds, 5 G will be employed in other industries such as the Internet of Things, broadcast services, and so on. Energy efficiency, scalability, resiliency, interoperability, and high data rate/low delay are the primary requirements and obstacles of 5 G cellular networks. Due to IEEE 802.11p's constraints, such as limited coverage, inability to handle dense vehicle networks, signal congestion, and connectivity outages, efficient data distribution is a big challenge (MAC contention problem). In this research, vehicle-to-vehicle (V2V), vehicle-to-infrastructure (V2I) and vehicle-to-pedestrian (V2P) services are used to overcome bandwidth constraints in very dense network communications from cellular tool to everything (C-V2X). Clustering is done through multi-layered multi-access edge clustering, which helps reduce vehicle contention. Fuzzy logic and Q-learning and intelligence are used for a multi-hop route selection system. The proposed protocol adjusts the number of cluster-head nodes using a Q-learning algorithm, allowing it to quickly adapt to a range of scenarios with varying bandwidths and vehicle densities.
第五代(5 G)网络需要满足日益增长的高速数据需求和不断扩大的客户数量。除了提供更高的速度,5 G 还将应用于其他行业,如物联网、广播服务等。能源效率、可扩展性、弹性、互操作性和高数据速率/低延迟是 5 G 蜂窝网络的主要要求和障碍。由于 IEEE 802.11p 的限制,如有限的覆盖范围、无法处理密集的车辆网络、信号拥塞和连接中断,高效的数据分发是一个巨大的挑战(MAC 竞争问题)。在这项研究中,车对车(V2V)、车对基础设施(V2I)和车对行人(V2P)服务被用来克服从蜂窝工具到万物(C-V2X)的高密度网络通信中的带宽限制。聚类是通过多层多接入边缘聚类完成的,这有助于减少车辆争用。多跳路由选择系统采用了模糊逻辑和 Q 学习与智能。提议的协议使用 Q-learning 算法调整簇头节点的数量,使其能够快速适应带宽和车辆密度不同的各种情况。
{"title":"Q-learning and fuzzy logic multi-tier multi-access edge clustering for 5g v2x communication.","authors":"Sangeetha Alagumani, Uma Maheswari Natarajan","doi":"10.1080/0954898X.2024.2309947","DOIUrl":"https://doi.org/10.1080/0954898X.2024.2309947","url":null,"abstract":"<p><p>The 5th generation (5 G) network is required to meet the growing demand for fast data speeds and the expanding number of customers. Apart from offering higher speeds, 5 G will be employed in other industries such as the Internet of Things, broadcast services, and so on. Energy efficiency, scalability, resiliency, interoperability, and high data rate/low delay are the primary requirements and obstacles of 5 G cellular networks. Due to IEEE 802.11p's constraints, such as limited coverage, inability to handle dense vehicle networks, signal congestion, and connectivity outages, efficient data distribution is a big challenge (MAC contention problem). In this research, vehicle-to-vehicle (V2V), vehicle-to-infrastructure (V2I) and vehicle-to-pedestrian (V2P) services are used to overcome bandwidth constraints in very dense network communications from cellular tool to everything (C-V2X). Clustering is done through multi-layered multi-access edge clustering, which helps reduce vehicle contention. Fuzzy logic and Q-learning and intelligence are used for a multi-hop route selection system. The proposed protocol adjusts the number of cluster-head nodes using a Q-learning algorithm, allowing it to quickly adapt to a range of scenarios with varying bandwidths and vehicle densities.</p>","PeriodicalId":54735,"journal":{"name":"Network-Computation in Neural Systems","volume":" ","pages":"1-24"},"PeriodicalIF":7.8,"publicationDate":"2024-03-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140040908","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-03DOI: 10.1080/0954898X.2024.2323530
Jie Cao, Jiacheng Fan, Chin-Ling Chen, Zhenyu Wu, Qingxuan Jiang, Shikai Li
As the number of patients increases, physicians are dealing with more and more cases of degenerative spine pathologies on a daily basis. To reduce the workload of healthcare professionals, we propose a modified Swin-UNet network model. Firstly, the Swin Transformer Blocks are improved using a residual post-normalization and scaling cosine attention mechanism, which makes the training process of the model more stable and improves the accuracy. Secondly, we use the log-space continuous position biasing method instead of the bicubic interpolation position biasing method. This method solves the problem of performance loss caused by the large difference between the resolution of the pretraining image and the resolution of the spine image. Finally, we introduce a segmentation smooth module (SSM) at the decoder stage. The SSM effectively reduces redundancy, and enhances the segmentation edge processing to improve the model's segmentation accuracy. To validate the proposed method, we conducted experiments on a real dataset provided by hospitals. The average segmentation accuracy is no less than 95%. The experimental results demonstrate the superiority of the proposed method over the original model and other models of the same type in segmenting the spinous processes of the vertebrae and the posterior arch of the spine.
{"title":"A Spinal MRI Image Segmentation Method Based on Improved Swin-UNet.","authors":"Jie Cao, Jiacheng Fan, Chin-Ling Chen, Zhenyu Wu, Qingxuan Jiang, Shikai Li","doi":"10.1080/0954898X.2024.2323530","DOIUrl":"https://doi.org/10.1080/0954898X.2024.2323530","url":null,"abstract":"<p><p>As the number of patients increases, physicians are dealing with more and more cases of degenerative spine pathologies on a daily basis. To reduce the workload of healthcare professionals, we propose a modified Swin-UNet network model. Firstly, the Swin Transformer Blocks are improved using a residual post-normalization and scaling cosine attention mechanism, which makes the training process of the model more stable and improves the accuracy. Secondly, we use the log-space continuous position biasing method instead of the bicubic interpolation position biasing method. This method solves the problem of performance loss caused by the large difference between the resolution of the pretraining image and the resolution of the spine image. Finally, we introduce a segmentation smooth module (SSM) at the decoder stage. The SSM effectively reduces redundancy, and enhances the segmentation edge processing to improve the model's segmentation accuracy. To validate the proposed method, we conducted experiments on a real dataset provided by hospitals. The average segmentation accuracy is no less than 95%. The experimental results demonstrate the superiority of the proposed method over the original model and other models of the same type in segmenting the spinous processes of the vertebrae and the posterior arch of the spine.</p>","PeriodicalId":54735,"journal":{"name":"Network-Computation in Neural Systems","volume":" ","pages":"1-29"},"PeriodicalIF":7.8,"publicationDate":"2024-03-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140023319","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-02-24DOI: 10.1080/0954898X.2024.2316080
N Ananthi, V Balaji, M Mohana, S Gnanapriya
Plant diseases are rising nowadays. Plant diseases lead to high economic losses. Internet of Things (IoT) technology has found its application in various sectors. This led to the introduction of smart farming, in which IoT has been utilized to help identify the exact spot of the diseased affected region on the leaf from the vast farmland in a well-organized and automated manner. Thus, the main focus of this task is the introduction of a novel plant disease detection model that relies on IoT technology. The collected images are given to the Image Transmission phase. Here, the encryption task is performed by employing the Advanced Encryption Standard (AES) and also the decrypted plant images are fed to the pre-processing stage. The Mask Regions with Convolutional Neural Networks (R-CNN) are used to segment the pre-processed images. Then, the segmented images are given to the detection phase in which the Adaptive Dense Hybrid Convolution Network with Attention Mechanism (ADHCN-AM) approach is utilized to perform the detection of plant disease. From the ADHCN-AM, the final detected plant disease outcomes are obtained. Throughout the entire validation, the offered model shows 95% enhancement in terms of MCC showcasing its effectiveness over the existing approaches.
{"title":"Smart plant disease net: Adaptive Dense Hybrid Convolution network with attention mechanism for IoT-based plant disease detection by improved optimization approach.","authors":"N Ananthi, V Balaji, M Mohana, S Gnanapriya","doi":"10.1080/0954898X.2024.2316080","DOIUrl":"https://doi.org/10.1080/0954898X.2024.2316080","url":null,"abstract":"<p><p>Plant diseases are rising nowadays. Plant diseases lead to high economic losses. Internet of Things (IoT) technology has found its application in various sectors. This led to the introduction of smart farming, in which IoT has been utilized to help identify the exact spot of the diseased affected region on the leaf from the vast farmland in a well-organized and automated manner. Thus, the main focus of this task is the introduction of a novel plant disease detection model that relies on IoT technology. The collected images are given to the Image Transmission phase. Here, the encryption task is performed by employing the Advanced Encryption Standard (AES) and also the decrypted plant images are fed to the pre-processing stage. The Mask Regions with Convolutional Neural Networks (R-CNN) are used to segment the pre-processed images. Then, the segmented images are given to the detection phase in which the Adaptive Dense Hybrid Convolution Network with Attention Mechanism (ADHCN-AM) approach is utilized to perform the detection of plant disease. From the ADHCN-AM, the final detected plant disease outcomes are obtained. Throughout the entire validation, the offered model shows 95% enhancement in terms of MCC showcasing its effectiveness over the existing approaches.</p>","PeriodicalId":54735,"journal":{"name":"Network-Computation in Neural Systems","volume":" ","pages":"1-39"},"PeriodicalIF":7.8,"publicationDate":"2024-02-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139944676","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-02-12DOI: 10.1080/0954898X.2024.2310687
Rathinavelu Sathiyaseelan, Krishnamoorthy Ranganathan, Ramesh Ramamoorthy, M Pedda Chennaiah
Retinal haemorrhage stands as an early indicator of diabetic retinopathy, necessitating accurate detection for timely diagnosis. Addressing this need, this study proposes an enhanced machine-based diagnostic test for diabetic retinopathy through an updated UNet framework, adept at scrutinizing fundus images for signs of retinal haemorrhages. The customized UNet underwent GPU training using the IDRiD database, validated against the publicly available DIARETDB1 and IDRiD datasets. Emphasizing the complexity of segmentation, the study employed preprocessing techniques, augmenting image quality and data integrity. Subsequently, the trained neural network showcased a remarkable performance boost, accurately identifying haemorrhage regions with 80% sensitivity, 99.6% specificity, and 98.6% accuracy. The experimental findings solidify the network's reliability, showcasing potential to alleviate ophthalmologists' workload significantly. Notably, achieving an Intersection over Union (IoU) of 76.61% and a Dice coefficient of 86.51% underscores the system's competence. The study's outcomes signify substantial enhancements in diagnosing critical diabetic retinal conditions, promising profound improvements in diagnostic accuracy and efficiency, thereby marking a significant advancement in automated retinal haemorrhage detection for diabetic retinopathy.
{"title":"Haemorrhage diagnosis in colour fundus images using a fast-convolutional neural network based on a modified U-Net.","authors":"Rathinavelu Sathiyaseelan, Krishnamoorthy Ranganathan, Ramesh Ramamoorthy, M Pedda Chennaiah","doi":"10.1080/0954898X.2024.2310687","DOIUrl":"10.1080/0954898X.2024.2310687","url":null,"abstract":"<p><p>Retinal haemorrhage stands as an early indicator of diabetic retinopathy, necessitating accurate detection for timely diagnosis. Addressing this need, this study proposes an enhanced machine-based diagnostic test for diabetic retinopathy through an updated UNet framework, adept at scrutinizing fundus images for signs of retinal haemorrhages. The customized UNet underwent GPU training using the IDRiD database, validated against the publicly available DIARETDB1 and IDRiD datasets. Emphasizing the complexity of segmentation, the study employed preprocessing techniques, augmenting image quality and data integrity. Subsequently, the trained neural network showcased a remarkable performance boost, accurately identifying haemorrhage regions with 80% sensitivity, 99.6% specificity, and 98.6% accuracy. The experimental findings solidify the network's reliability, showcasing potential to alleviate ophthalmologists' workload significantly. Notably, achieving an Intersection over Union (IoU) of 76.61% and a Dice coefficient of 86.51% underscores the system's competence. The study's outcomes signify substantial enhancements in diagnosing critical diabetic retinal conditions, promising profound improvements in diagnostic accuracy and efficiency, thereby marking a significant advancement in automated retinal haemorrhage detection for diabetic retinopathy.</p>","PeriodicalId":54735,"journal":{"name":"Network-Computation in Neural Systems","volume":" ","pages":"1-22"},"PeriodicalIF":7.8,"publicationDate":"2024-02-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139725025","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-02-01Epub Date: 2024-02-08DOI: 10.1080/0954898X.2023.2275720
Srilakshmi Aluri, Sagar S Imambi
Brain tumour (BT) is a dangerous neurological disorder produced by abnormal cell growth within the skull or brain. Nowadays, the death rate of people with BT is linearly growing. The finding of tumours at an early stage is crucial for giving treatment to patients, which improves the survival rate of patients. Hence, the BT classification (BTC) is done in this research using magnetic resonance imaging (MRI) images. In this research, the input MRI image is pre-processed using a non-local means (NLM) filter that denoises the input image. For attaining the effective classified result, the tumour area from the MRI image is segmented by the SegNet model. Furthermore, the BTC is accomplished by the LeNet model whose weight is optimized by the Golden Teacher Learning Optimization Algorithm (GTLO) such that the classified output produced by the LeNet model is Gliomas, Meningiomas, and Pituitary tumours. The experimental outcome displays that the GTLO-LeNet achieved an Accuracy of 0.896, Negative Predictive value (NPV) of 0.907, Positive Predictive value (PPV) of 0.821, True Negative Rate (TNR) of 0.880, and True Positive Rate (TPR) of 0.888.
{"title":"Brain tumour classification using MRI images based on lenet with golden teacher learning optimization.","authors":"Srilakshmi Aluri, Sagar S Imambi","doi":"10.1080/0954898X.2023.2275720","DOIUrl":"10.1080/0954898X.2023.2275720","url":null,"abstract":"<p><p>Brain tumour (BT) is a dangerous neurological disorder produced by abnormal cell growth within the skull or brain. Nowadays, the death rate of people with BT is linearly growing. The finding of tumours at an early stage is crucial for giving treatment to patients, which improves the survival rate of patients. Hence, the BT classification (BTC) is done in this research using magnetic resonance imaging (MRI) images. In this research, the input MRI image is pre-processed using a non-local means (NLM) filter that denoises the input image. For attaining the effective classified result, the tumour area from the MRI image is segmented by the SegNet model. Furthermore, the BTC is accomplished by the LeNet model whose weight is optimized by the Golden Teacher Learning Optimization Algorithm (GTLO) such that the classified output produced by the LeNet model is Gliomas, Meningiomas, and Pituitary tumours. The experimental outcome displays that the GTLO-LeNet achieved an Accuracy of 0.896, Negative Predictive value (NPV) of 0.907, Positive Predictive value (PPV) of 0.821, True Negative Rate (TNR) of 0.880, and True Positive Rate (TPR) of 0.888.</p>","PeriodicalId":54735,"journal":{"name":"Network-Computation in Neural Systems","volume":" ","pages":"27-54"},"PeriodicalIF":7.8,"publicationDate":"2024-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"72016196","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-02-01Epub Date: 2024-02-08DOI: 10.1080/0954898X.2023.2275722
Vaishali Bajait, Nandagopal Malarvizhi
Our approach includes picture preprocessing, feature extraction utilizing the SqueezeNet model, hyperparameter optimisation utilising the Equilibrium Optimizer (EO) algorithm, and classification utilising a Stacked Autoencoder (SAE) model. Each of these processes is carried out in a series of separate steps. During the image preprocessing stage, contrast limited adaptive histogram equalisations (CLAHE) is utilized to improve the contrasts, and Adaptive Bilateral Filtering (ABF) to get rid of any noise that may be present. The SqueezeNet paradigm is utilized to obtain relevant characteristics from the pictures that have been preprocessed, and the EO technique is utilized to fine-tune the hyperparameters. Finally, the SAE model categorises the diseases that affect the grape leaf. The simulation analysis of the EODTL-GLDC technique tested New Plant Diseases Datasets and the results were inspected in many prospects. The results demonstrate that this model outperforms other deep learning techniques and methods that are more often related to machine learning. Specifically, this technique was able to attain a precision of 96.31% on the testing datasets and 96.88% on the training data set that was split 80:20. These results offer more proof that the suggested strategy is successful in automating the detection and categorization of grape leaf diseases.
{"title":"Automated grape leaf nutrition deficiency disease detection and classification Equilibrium Optimizer with deep transfer learning model.","authors":"Vaishali Bajait, Nandagopal Malarvizhi","doi":"10.1080/0954898X.2023.2275722","DOIUrl":"10.1080/0954898X.2023.2275722","url":null,"abstract":"<p><p>Our approach includes picture preprocessing, feature extraction utilizing the SqueezeNet model, hyperparameter optimisation utilising the Equilibrium Optimizer (EO) algorithm, and classification utilising a Stacked Autoencoder (SAE) model. Each of these processes is carried out in a series of separate steps. During the image preprocessing stage, contrast limited adaptive histogram equalisations (CLAHE) is utilized to improve the contrasts, and Adaptive Bilateral Filtering (ABF) to get rid of any noise that may be present. The SqueezeNet paradigm is utilized to obtain relevant characteristics from the pictures that have been preprocessed, and the EO technique is utilized to fine-tune the hyperparameters. Finally, the SAE model categorises the diseases that affect the grape leaf. The simulation analysis of the EODTL-GLDC technique tested New Plant Diseases Datasets and the results were inspected in many prospects. The results demonstrate that this model outperforms other deep learning techniques and methods that are more often related to machine learning. Specifically, this technique was able to attain a precision of 96.31% on the testing datasets and 96.88% on the training data set that was split 80:20. These results offer more proof that the suggested strategy is successful in automating the detection and categorization of grape leaf diseases.</p>","PeriodicalId":54735,"journal":{"name":"Network-Computation in Neural Systems","volume":" ","pages":"55-72"},"PeriodicalIF":7.8,"publicationDate":"2024-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"71489081","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Nowadays, wireless sensor networks (WSN) have gained huge attention worldwide due to their wide applications in different domains. The limited amount of energy resources is considered as the main limitations of WSN, which generally affect the network life time. Hence, a dynamic clustering and routing model is designed to resolve this issue. In this research work, a deep-learning model is employed for the prediction of energy and an optimization algorithmic technique is designed for the determination of optimal routes. Initially, the dynamic cluster WSN is simulated using energy, mobility, trust, and Link Life Time (LLT) models. The deep neuro-fuzzy network (DNFN) is utilized for the prediction of residual energy of nodes and the cluster workloads are dynamically balanced by the dynamic clustering of data using a fuzzy system. The designed Flamingo Jellyfish Search Optimization (FJSO) model is used for tuning the weights of the fuzzy system by considering different fitness parameters. Moreover, routing is performed using FJSO model which is used for the identification of optimal path to transmit data. In addition, the experimentation is done using MATLAB tool and the results proved that the designed FJSO model attained maximum of 0.657J energy, a minimum of 0.739 m distance, 0.649 s delay, 0.849 trust, and 0.885 Mbps throughput.
{"title":"Flamingo Jelly Fish search optimization-based routing with deep-learning enabled energy prediction in WSN data communication.","authors":"Dhanabal Subramanian, Sangeetha Subramaniam, Krishnamoorthy Natarajan, Kumaravel Thangavel","doi":"10.1080/0954898X.2023.2279971","DOIUrl":"10.1080/0954898X.2023.2279971","url":null,"abstract":"<p><p>Nowadays, wireless sensor networks (WSN) have gained huge attention worldwide due to their wide applications in different domains. The limited amount of energy resources is considered as the main limitations of WSN, which generally affect the network life time. Hence, a dynamic clustering and routing model is designed to resolve this issue. In this research work, a deep-learning model is employed for the prediction of energy and an optimization algorithmic technique is designed for the determination of optimal routes. Initially, the dynamic cluster WSN is simulated using energy, mobility, trust, and Link Life Time (LLT) models. The deep neuro-fuzzy network (DNFN) is utilized for the prediction of residual energy of nodes and the cluster workloads are dynamically balanced by the dynamic clustering of data using a fuzzy system. The designed Flamingo Jellyfish Search Optimization (FJSO) model is used for tuning the weights of the fuzzy system by considering different fitness parameters. Moreover, routing is performed using FJSO model which is used for the identification of optimal path to transmit data. In addition, the experimentation is done using MATLAB tool and the results proved that the designed FJSO model attained maximum of 0.657J energy, a minimum of 0.739 m distance, 0.649 s delay, 0.849 trust, and 0.885 Mbps throughput.</p>","PeriodicalId":54735,"journal":{"name":"Network-Computation in Neural Systems","volume":" ","pages":"73-100"},"PeriodicalIF":7.8,"publicationDate":"2024-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138479329","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-02-01Epub Date: 2024-02-08DOI: 10.1080/0954898X.2023.2270040
Rajveer K Shastri, Aparna R Shastri, Prashant P Nitnaware, Digambar M Padulkar
In the diagnosis of cardiac disorders Heart sound has a major role, and early detection is crucial to safeguard the patients. Computerized strategies of heart sound classification advocate intensive and more exact results in a quick and better manner. Using a hybrid optimization-controlled deep learning strategy this paper proposed an automatic heart sound classification module. The parameter tuning of the Deep Neural Network (DNN) classifier in a satisfactory manner is the importance of this research which depends on the Hybrid Sneaky optimization algorithm. The developed sneaky optimization algorithm inherits the traits of questing and societal search agents. Moreover, input data from the Phonocardiogram (PCG) database undergoes the process of feature extraction which extract the important features, like statistical, Heart Rate Variability (HRV), and to enhance the performance of this model, the features of Mel frequency Cepstral coefficients (MFCC) are assisted. The developed Sneaky optimization-based DNN classifier's performance is determined in respect of the metrics, namely precision, accuracy, specificity, and sensitivity, which are around 97%, 96.98%, 97%, and 96.9%, respectively.
{"title":"Hybrid Sneaky algorithm-based deep neural networks for Heart sound classification using phonocardiogram.","authors":"Rajveer K Shastri, Aparna R Shastri, Prashant P Nitnaware, Digambar M Padulkar","doi":"10.1080/0954898X.2023.2270040","DOIUrl":"10.1080/0954898X.2023.2270040","url":null,"abstract":"<p><p>In the diagnosis of cardiac disorders Heart sound has a major role, and early detection is crucial to safeguard the patients. Computerized strategies of heart sound classification advocate intensive and more exact results in a quick and better manner. <u>U</u>sing a hybrid optimization-controlled deep learning strategy this paper proposed an automatic heart sound classification module. The parameter tuning of the Deep Neural Network (DNN) classifier in a satisfactory manner is the importance of this research which depends on the Hybrid Sneaky optimization algorithm. The developed sneaky optimization algorithm inherits the traits of questing and societal search agents. Moreover, input data from the Phonocardiogram (PCG) database undergoes the process of feature extraction which extract the important features, like statistical, Heart Rate Variability (HRV), and to enhance the performance of this model, the features of Mel frequency Cepstral coefficients (MFCC) are assisted. The developed Sneaky optimization-based DNN classifier's performance is determined in respect of the metrics, namely precision, accuracy, specificity, and sensitivity, which are around 97%, 96.98%, 97%, and 96.9%, respectively.</p>","PeriodicalId":54735,"journal":{"name":"Network-Computation in Neural Systems","volume":" ","pages":"1-26"},"PeriodicalIF":7.8,"publicationDate":"2024-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138453076","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-11-21DOI: 10.1080/0954898X.2023.2286002
David Femi, Manapakkam Anandan Mukunthan
Leaf infection detection and diagnosis at an earlier stage can improve agricultural output and reduce monetary costs. An inaccurate segmentation may degrade the accuracy of disease classification due to some different and complex leaf diseases. Also, the disease's adhesion and dimension can overlap, causing partial under-segmentation. Therefore, a novel robust Deep Encoder-Decoder Cascaded Network (DEDCNet) model is proposed in this manuscript for leaf image segmentation that precisely segments the diseased leaf spots and differentiates similar diseases. This model is comprised of an Infected Spot Recognition Network and an Infected Spot Segmentation Network. Initially, ISRN is designed by integrating cascaded CNN with a Feature Pyramid Pooling layer to identify the infected leaf spot and avoid an impact of background details. After that, the ISSN developed using an encoder-decoder network, which uses a multi-scale dilated convolution kernel to precisely segment the infected leaf spot. Moreover, the resultant leaf segments are provided to the pre-learned CNN models to learn texture features followed by the SVM algorithm to categorize leaf disease classes. The ODEDCNet delivers exceptional performance on both the Betel Leaf Image and PlantVillage datasets. On the Betel Leaf Image dataset, it achieves an accuracy of 94.89%, with high precision (94.35%), recall (94.77%), and F-score (94.56%), while maintaining low under-segmentation (6.2%) and over-segmentation rates (2.8%). It also achieves a remarkable Dice coefficient of 0.9822, all in just 0.10 seconds. On the PlantVillage dataset, the ODEDCNet outperforms other existing models with an accuracy of 96.5%, demonstrating high precision (96.61%), recall (96.5%), and F-score (96.56%). It excels in reducing under-segmentation to just 3.12% and over-segmentation to 2.56%. Furthermore, it achieves a Dice coefficient of 0.9834 in a mere 0.09 seconds. It evident for the greater efficiency on both segmentation and categorization of leaf diseases contrasted with the existing models.
{"title":"Plant leaf infected spot segmentation using robust encoder-decoder cascaded deep learning model.","authors":"David Femi, Manapakkam Anandan Mukunthan","doi":"10.1080/0954898X.2023.2286002","DOIUrl":"https://doi.org/10.1080/0954898X.2023.2286002","url":null,"abstract":"<p><p>Leaf infection detection and diagnosis at an earlier stage can improve agricultural output and reduce monetary costs. An inaccurate segmentation may degrade the accuracy of disease classification due to some different and complex leaf diseases. Also, the disease's adhesion and dimension can overlap, causing partial under-segmentation. Therefore, a novel robust Deep Encoder-Decoder Cascaded Network (DEDCNet) model is proposed in this manuscript for leaf image segmentation that precisely segments the diseased leaf spots and differentiates similar diseases. This model is comprised of an Infected Spot Recognition Network and an Infected Spot Segmentation Network. Initially, ISRN is designed by integrating cascaded CNN with a Feature Pyramid Pooling layer to identify the infected leaf spot and avoid an impact of background details. After that, the ISSN developed using an encoder-decoder network, which uses a multi-scale dilated convolution kernel to precisely segment the infected leaf spot. Moreover, the resultant leaf segments are provided to the pre-learned CNN models to learn texture features followed by the SVM algorithm to categorize leaf disease classes. The ODEDCNet delivers exceptional performance on both the Betel Leaf Image and PlantVillage datasets. On the Betel Leaf Image dataset, it achieves an accuracy of 94.89%, with high precision (94.35%), recall (94.77%), and F-score (94.56%), while maintaining low under-segmentation (6.2%) and over-segmentation rates (2.8%). It also achieves a remarkable Dice coefficient of 0.9822, all in just 0.10 seconds. On the PlantVillage dataset, the ODEDCNet outperforms other existing models with an accuracy of 96.5%, demonstrating high precision (96.61%), recall (96.5%), and F-score (96.56%). It excels in reducing under-segmentation to just 3.12% and over-segmentation to 2.56%. Furthermore, it achieves a Dice coefficient of 0.9834 in a mere 0.09 seconds. It evident for the greater efficiency on both segmentation and categorization of leaf diseases contrasted with the existing models.</p>","PeriodicalId":54735,"journal":{"name":"Network-Computation in Neural Systems","volume":" ","pages":"1-19"},"PeriodicalIF":7.8,"publicationDate":"2023-11-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138464478","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}