Pub Date : 2024-12-03DOI: 10.1080/0954898X.2024.2428705
Visu Pandurangan, Smitha Ponnayyan Sarojam, Pughazendi Narayanan, Murugananthan Velayutham
Skin melanin lesions are typically identified as tiny patches on the skin, which are impacted by melanocyte cell overgrowth. The number of people with skin cancer is increasing worldwide. Accurate and timely skin cancer identification is critical to reduce the mortality rates. An incorrect diagnosis can be fatal to the patient. To tackle these issues, this article proposes the Recurrent Prototypical Object Segmentation Network (RPO-SegNet) for the segmentation of skin lesions and a hybrid Deep Learning (DL) - based skin cancer classification. The RPO-SegNet is formed by integrating the Recurrent Prototypical Networks (RP-Net), and Object Segmentation Networks (O-SegNet). At first, the input image is taken from a database and forwarded to image pre-processing. Then, the segmentation of skin lesions is accomplished using the proposed RPO-SegNet. After the segmentation, feature extraction is accomplished. Finally, skin cancer classification and detection are accomplished by employing the Fuzzy-based Shepard Convolutional Maxout Network (FSCMN) by combining the Deep Maxout Network (DMN), and Shepard Convolutional Neural Network (ShCNN). The established RPO-SegNet+FSCMN attained improved accuracy, True Negative Rate (TNR), True Positive Rate (TPR), dice coefficient, Jaccard coefficient, and segmentation analysis of 91.985%, 92.735%, 93.485%, 90.902%, 90.164%, and 91.734%.
{"title":"Hybrid deep learning-based skin cancer classification with RPO-SegNet for skin lesion segmentation.","authors":"Visu Pandurangan, Smitha Ponnayyan Sarojam, Pughazendi Narayanan, Murugananthan Velayutham","doi":"10.1080/0954898X.2024.2428705","DOIUrl":"https://doi.org/10.1080/0954898X.2024.2428705","url":null,"abstract":"<p><p>Skin melanin lesions are typically identified as tiny patches on the skin, which are impacted by melanocyte cell overgrowth. The number of people with skin cancer is increasing worldwide. Accurate and timely skin cancer identification is critical to reduce the mortality rates. An incorrect diagnosis can be fatal to the patient. To tackle these issues, this article proposes the Recurrent Prototypical Object Segmentation Network (RPO-SegNet) for the segmentation of skin lesions and a hybrid Deep Learning (DL) - based skin cancer classification. The RPO-SegNet is formed by integrating the Recurrent Prototypical Networks (RP-Net), and Object Segmentation Networks (O-SegNet). At first, the input image is taken from a database and forwarded to image pre-processing. Then, the segmentation of skin lesions is accomplished using the proposed RPO-SegNet. After the segmentation, feature extraction is accomplished. Finally, skin cancer classification and detection are accomplished by employing the Fuzzy-based Shepard Convolutional Maxout Network (FSCMN) by combining the Deep Maxout Network (DMN), and Shepard Convolutional Neural Network (ShCNN). The established RPO-SegNet+FSCMN attained improved accuracy, True Negative Rate (TNR), True Positive Rate (TPR), dice coefficient, Jaccard coefficient, and segmentation analysis of 91.985%, 92.735%, 93.485%, 90.902%, 90.164%, and 91.734%.</p>","PeriodicalId":54735,"journal":{"name":"Network-Computation in Neural Systems","volume":" ","pages":"1-28"},"PeriodicalIF":1.1,"publicationDate":"2024-12-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142774851","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-11-28DOI: 10.1080/0954898X.2024.2434487
Mohamed Aboukhair, Abdelrahim Koura, Mohammed Kayed
Convolutional neural networks (CNNs) have reached their peak of complex structures, but until now, few researchers have addressed the problem of relying on one filter size. Mainly a 3 × 3 filter is the most common one used in any structure. Only at the first layers of the CNN model, filters bigger than 3 × 3 could be partially used. Most researchers work with filters (size, values, etc.) as a black box. To the best of our knowledge, this research is the first pilot study that proposes a new multi-filter layer in which different filters with variant sizes are used to replace the 3 × 3 filter layers. Our proposed multi-filter layer has yielded encouraging results, demonstrating notable improvements ranging from 1% to 5% in performance. This achievement was realized by developing two innovative structures, namely the fixed structure and the decreasing structure. Both of them leverage the multi-filter layer. Although the two structures exhibit promising outcomes, the later structure offers the additional advantages of reduced computational requirements and enhanced learner strength.
{"title":"A pilot study of novel multi-filter CNN layer.","authors":"Mohamed Aboukhair, Abdelrahim Koura, Mohammed Kayed","doi":"10.1080/0954898X.2024.2434487","DOIUrl":"https://doi.org/10.1080/0954898X.2024.2434487","url":null,"abstract":"<p><p>Convolutional neural networks (CNNs) have reached their peak of complex structures, but until now, few researchers have addressed the problem of relying on one filter size. Mainly a 3 × 3 filter is the most common one used in any structure. Only at the first layers of the CNN model, filters bigger than 3 × 3 could be partially used. Most researchers work with filters (size, values, etc.) as a black box. To the best of our knowledge, this research is the first pilot study that proposes a new multi-filter layer in which different filters with variant sizes are used to replace the 3 × 3 filter layers. Our proposed multi-filter layer has yielded encouraging results, demonstrating notable improvements ranging from 1% to 5% in performance. This achievement was realized by developing two innovative structures, namely the fixed structure and the decreasing structure. Both of them leverage the multi-filter layer. Although the two structures exhibit promising outcomes, the later structure offers the additional advantages of reduced computational requirements and enhanced learner strength.</p>","PeriodicalId":54735,"journal":{"name":"Network-Computation in Neural Systems","volume":" ","pages":"1-22"},"PeriodicalIF":1.1,"publicationDate":"2024-11-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142752332","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-11-27DOI: 10.1080/0954898X.2024.2429721
Love Kumar Thawait, Mukesh Kumar Singh
Thermal Power Plant is a common power plant that generates power by fuel-burning to produce electricity. Being a significant component of the energy sector, the Thermal Power Plant faces several issues that lead to reduced productivity. Conventional researchers have tried using different mechanisms for improvising the production of Thermal Power Plants in varied dimensions. Due to the diverse dimensions considered by existing works, the present review endeavours to afford a comprehensive summary of these works. To achieve this, the study reviews articles in the range (2019-2023) that are allied with the utility of SC methodologies (encompassing AI-ML (Machine Learning) and DL (Deep Learning) in enhancing the productivity of Thermal Power Plants by various dimensions. The conventional AI-based approaches are comparatively evaluated for effective contribution in improvising Thermal Power Plant production. Following this, a critical assessment encompasses the year-wise distribution and varied dimensions focussed by traditional studies in this area. This would support future researchers in determining the dimensions that have attained limited and high focus based on which appropriate research works can be performed. Finally, future suggestions and research gaps are included to offer new stimulus for further investigation of AI in Thermal Power Plants.
{"title":"A review on real time implementation of soft computing techniques in thermal power plant.","authors":"Love Kumar Thawait, Mukesh Kumar Singh","doi":"10.1080/0954898X.2024.2429721","DOIUrl":"https://doi.org/10.1080/0954898X.2024.2429721","url":null,"abstract":"<p><p>Thermal Power Plant is a common power plant that generates power by fuel-burning to produce electricity. Being a significant component of the energy sector, the Thermal Power Plant faces several issues that lead to reduced productivity. Conventional researchers have tried using different mechanisms for improvising the production of Thermal Power Plants in varied dimensions. Due to the diverse dimensions considered by existing works, the present review endeavours to afford a comprehensive summary of these works. To achieve this, the study reviews articles in the range (2019-2023) that are allied with the utility of SC methodologies (encompassing AI-ML (Machine Learning) and DL (Deep Learning) in enhancing the productivity of Thermal Power Plants by various dimensions. The conventional AI-based approaches are comparatively evaluated for effective contribution in improvising Thermal Power Plant production. Following this, a critical assessment encompasses the year-wise distribution and varied dimensions focussed by traditional studies in this area. This would support future researchers in determining the dimensions that have attained limited and high focus based on which appropriate research works can be performed. Finally, future suggestions and research gaps are included to offer new stimulus for further investigation of AI in Thermal Power Plants.</p>","PeriodicalId":54735,"journal":{"name":"Network-Computation in Neural Systems","volume":" ","pages":"1-37"},"PeriodicalIF":1.1,"publicationDate":"2024-11-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142734747","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-11-21DOI: 10.1080/0954898X.2024.2428713
Zhimin Tian, Huijuan Hao, Guowei Dai, Yajuan Li
Tomato harvesting in intelligent greenhouses is crucial for reducing costs and optimizing management. Agricultural robots, as an automated solution, require advanced visual perception. This study proposes a tomato detection and counting algorithm based on YOLOv8 (TCAttn-YOLOv8). To handle small, occluded tomato targets in images, a new detection layer (NDL) is added to the Neck and Head decoupled structure, improving small object recognition. The ColBlock, a dual-branch structure leveraging Transformer advantages, enhances feature extraction and fusion, focusing on densely targeted regions and minimizing small object feature loss in complex backgrounds. C2fGhost and GhostConv are integrated into the Neck network to reduce model parameters and floating-point operations, improving feature expression. The WIoU (Wise-IoU) loss function is adopted to accelerate convergence and increase regression accuracy. Experimental results show that TCAttn-YOLOv8 achieves an mAP@0.5 of 96.31%, with an FPS of 95 and a parameter size of 2.7 M, outperforming seven lightweight YOLO algorithms. For automated tomato counting, the R2 between predicted and actual counts is 0.9282, indicating the algorithm's suitability for replacing manual counting. This method effectively supports tomato detection and counting in intelligent greenhouses, offering valuable insights for robotic harvesting and yield estimation research.
{"title":"Optimizing tomato detection and counting in smart greenhouses: A lightweight YOLOv8 model incorporating high- and low-frequency feature transformer structures.","authors":"Zhimin Tian, Huijuan Hao, Guowei Dai, Yajuan Li","doi":"10.1080/0954898X.2024.2428713","DOIUrl":"https://doi.org/10.1080/0954898X.2024.2428713","url":null,"abstract":"<p><p>Tomato harvesting in intelligent greenhouses is crucial for reducing costs and optimizing management. Agricultural robots, as an automated solution, require advanced visual perception. This study proposes a tomato detection and counting algorithm based on YOLOv8 (TCAttn-YOLOv8). To handle small, occluded tomato targets in images, a new detection layer (NDL) is added to the Neck and Head decoupled structure, improving small object recognition. The ColBlock, a dual-branch structure leveraging Transformer advantages, enhances feature extraction and fusion, focusing on densely targeted regions and minimizing small object feature loss in complex backgrounds. C2fGhost and GhostConv are integrated into the Neck network to reduce model parameters and floating-point operations, improving feature expression. The WIoU (Wise-IoU) loss function is adopted to accelerate convergence and increase regression accuracy. Experimental results show that TCAttn-YOLOv8 achieves an mAP@0.5 of 96.31%, with an FPS of 95 and a parameter size of 2.7 M, outperforming seven lightweight YOLO algorithms. For automated tomato counting, the <i>R<sup>2</sup></i> between predicted and actual counts is 0.9282, indicating the algorithm's suitability for replacing manual counting. This method effectively supports tomato detection and counting in intelligent greenhouses, offering valuable insights for robotic harvesting and yield estimation research.</p>","PeriodicalId":54735,"journal":{"name":"Network-Computation in Neural Systems","volume":" ","pages":"1-37"},"PeriodicalIF":1.1,"publicationDate":"2024-11-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142683524","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-11-17DOI: 10.1080/0954898X.2024.2424248
Annamalai Thiruvengadam Madhavi, Kamal Basha Rahimunnisa
Estimating the optimal answer is expensive for huge data resources that decrease the functionality of the system. To solve these issues, the latest groundnut leaf disorder identification model by deep learning techniques is implemented. The images are collected from traditional databases, and then they are given to the pre-processing stage. Then, relevant features are drawn out from the preprocessed images in two stages. In the first stage, the preprocessed image is segmented using adaptive TransResunet++, where the variables are tuned with the help of designed Hybrid Position of Beluga Whale and Cuttle Fish (HP-BWCF) and finally get the feature set 1 using Kaze Feature Points and Binary Descriptors. In the second stage, the same Kaze feature points and the binary descriptors are extracted from the preprocessed image separately, and then obtain feature set 2. Then, the extracted feature sets 1 and 2 are concatenated and given to the Hybrid Convolution-based Adaptive Resnet with Attention Mechanism (HCAR-AM) to detect the ground nut leaf diseases very effectively. The parameters from this HCAR-AM are tuned via the same HP-BWCF. The experimental outcome is analysed over various recently developed ground nut leaf disease detection approaches in accordance with various performance measures.
{"title":"HCAR-AM ground nut leaf net: Hybrid convolution-based adaptive ResNet with attention mechanism for detecting ground nut leaf diseases with adaptive segmentation.","authors":"Annamalai Thiruvengadam Madhavi, Kamal Basha Rahimunnisa","doi":"10.1080/0954898X.2024.2424248","DOIUrl":"https://doi.org/10.1080/0954898X.2024.2424248","url":null,"abstract":"<p><p>Estimating the optimal answer is expensive for huge data resources that decrease the functionality of the system. To solve these issues, the latest groundnut leaf disorder identification model by deep learning techniques is implemented. The images are collected from traditional databases, and then they are given to the pre-processing stage. Then, relevant features are drawn out from the preprocessed images in two stages. In the first stage, the preprocessed image is segmented using adaptive TransResunet++, where the variables are tuned with the help of designed Hybrid Position of Beluga Whale and Cuttle Fish (HP-BWCF) and finally get the feature set 1 using Kaze Feature Points and Binary Descriptors. In the second stage, the same Kaze feature points and the binary descriptors are extracted from the preprocessed image separately, and then obtain feature set 2. Then, the extracted feature sets 1 and 2 are concatenated and given to the Hybrid Convolution-based Adaptive Resnet with Attention Mechanism (HCAR-AM) to detect the ground nut leaf diseases very effectively. The parameters from this HCAR-AM are tuned via the same HP-BWCF. The experimental outcome is analysed over various recently developed ground nut leaf disease detection approaches in accordance with various performance measures.</p>","PeriodicalId":54735,"journal":{"name":"Network-Computation in Neural Systems","volume":" ","pages":"1-41"},"PeriodicalIF":1.1,"publicationDate":"2024-11-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142649538","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-11-16DOI: 10.1080/0954898X.2024.2426580
Suresh Kumar Krishnamoorthy, Vanitha Cn
Cancer involves abnormal cell growth, with types like intestinal and oesophageal cancer often diagnosed in advanced stages, making them hard to cure. Symptoms are like burning sensations in the stomach and swallowing difficulties are specified as colorectal cancer. Deep learning significantly impacts the medical image processing and diagnosis, offering potential improvements in accuracy and efficiency. The Kruskal Szekeres Generative Adversarial Network Augmented Deep Autoencoder (KSGANA-DA) is introduced for early colorectal cancer detection and it comprises two stages; Initial stage, data augmentation uses Affine Transform via Random Horizontal Rotation and Geometric Transform via Kruskal-Szekeres that coordinates to improve the training dataset diversity, boosting detection performance. The second stage, a Deep Autoencoder Anatomical Landmark-based Image Segmentation preserves edge pixel spatial locations, improving precision and recall for early boundary detection. Experiments validate KSGANA-DA performance and different existing methods are implemented into Python. The results of KSGANA-DA are to provide higher precision by 41%, recall by 7%, and lesser training time by 46% than compared to conventional methods.
{"title":"Kruskal Szekeres generative adversarial network augmented deep autoencoder for colorectal cancer detection.","authors":"Suresh Kumar Krishnamoorthy, Vanitha Cn","doi":"10.1080/0954898X.2024.2426580","DOIUrl":"https://doi.org/10.1080/0954898X.2024.2426580","url":null,"abstract":"<p><p>Cancer involves abnormal cell growth, with types like intestinal and oesophageal cancer often diagnosed in advanced stages, making them hard to cure. Symptoms are like burning sensations in the stomach and swallowing difficulties are specified as colorectal cancer. Deep learning significantly impacts the medical image processing and diagnosis, offering potential improvements in accuracy and efficiency. The Kruskal Szekeres Generative Adversarial Network Augmented Deep Autoencoder (KSGANA-DA) is introduced for early colorectal cancer detection and it comprises two stages; Initial stage, data augmentation uses Affine Transform via Random Horizontal Rotation and Geometric Transform via Kruskal-Szekeres that coordinates to improve the training dataset diversity, boosting detection performance. The second stage, a Deep Autoencoder Anatomical Landmark-based Image Segmentation preserves edge pixel spatial locations, improving precision and recall for early boundary detection. Experiments validate KSGANA-DA performance and different existing methods are implemented into Python. The results of KSGANA-DA are to provide higher precision by 41%, recall by 7%, and lesser training time by 46% than compared to conventional methods.</p>","PeriodicalId":54735,"journal":{"name":"Network-Computation in Neural Systems","volume":" ","pages":"1-27"},"PeriodicalIF":1.1,"publicationDate":"2024-11-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142645136","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-11-12DOI: 10.1080/0954898X.2024.2421196
Maxime Carriere, Rosario Tomasello, Friedemann Pulvermüller
The ability of humans to store spoken words in verbal working memory and build extensive vocabularies is believed to stem from evolutionary changes in cortical connectivity across primate species. However, the underlying neurobiological mechanisms remain unclear. Why can humans acquire vast vocabularies, while non-human primates cannot? This study addresses this question using brain-constrained neural networks that realize between-species differences in cortical connectivity. It investigates how these structural differences support the formation of neural representations for spoken words and the emergence of verbal working memory, crucial for human vocabulary building. We develop comparative models of frontotemporal and occipital cortices, reflecting human and non-human primate neuroanatomy. Using meanfield and spiking neural networks, we simulate auditory word recognition and examine verbal working memory function. The "human models", characterized by denser inter-area connectivity in core language areas, produced larger cell assemblies than the "monkey models", with specific topographies reflecting semantic properties of the represented words. Crucially, longer-lasting reverberant neural activity was observed in human versus monkey architectures, compatible with robust verbal working memory, a necessary condition for vocabulary building. Our findings offer insights into the structural basis of human-specific symbol learning and verbal working memory, shedding light on humans' unique capacity for large vocabulary acquisition.
{"title":"Can human brain connectivity explain verbal working memory?","authors":"Maxime Carriere, Rosario Tomasello, Friedemann Pulvermüller","doi":"10.1080/0954898X.2024.2421196","DOIUrl":"10.1080/0954898X.2024.2421196","url":null,"abstract":"<p><p>The ability of humans to store spoken words in verbal working memory and build extensive vocabularies is believed to stem from evolutionary changes in cortical connectivity across primate species. However, the underlying neurobiological mechanisms remain unclear. Why can humans acquire vast vocabularies, while non-human primates cannot? This study addresses this question using brain-constrained neural networks that realize between-species differences in cortical connectivity. It investigates how these structural differences support the formation of neural representations for spoken words and the emergence of verbal working memory, crucial for human vocabulary building. We develop comparative models of frontotemporal and occipital cortices, reflecting human and non-human primate neuroanatomy. Using meanfield and spiking neural networks, we simulate auditory word recognition and examine verbal working memory function. The \"human models\", characterized by denser inter-area connectivity in core language areas, produced larger cell assemblies than the \"monkey models\", with specific topographies reflecting semantic properties of the represented words. Crucially, longer-lasting reverberant neural activity was observed in human versus monkey architectures, compatible with robust verbal working memory, a necessary condition for vocabulary building. Our findings offer insights into the structural basis of human-specific symbol learning and verbal working memory, shedding light on humans' unique capacity for large vocabulary acquisition.</p>","PeriodicalId":54735,"journal":{"name":"Network-Computation in Neural Systems","volume":" ","pages":"1-42"},"PeriodicalIF":1.1,"publicationDate":"2024-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142632807","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Owing to the epidemic growth of diabetes, ophthalmologists need to examine the huge fundus images for diagnosing the disease of Diabetic Retinopathy (DR). Without proper knowledge, people are too lethargic to detect the DR. Therefore, the early diagnosis system is requisite for treating ailments in the medical industry. Therefore, a novel deep model-based DR detection structure is recommended to fix the aforementioned difficulties. The developed deep model-based diabetic retinopathy detection process is performed adaptively. The DR detection process is imitated by garnering the images from benchmark sources. The gathered images are further preceded by the abnormality segmentation phase. Here, the Residual TransUNet with Enhanced loss function is used to employ the abnormality segmentation, and the loss function in this structure may be helpful to lessen the error in the segmentation procedure. Further, the segmented images are passed to the final phase of retinopathy detection. At this phase, the detection is carried out through the Adaptive Multiscale MobileNet. The variables in the AMMNet are optimized by the Adaptive Puzzle Optimization to obtain better detection performance. Finally, the effectiveness of the offered approach is confirmed by the experimentation procedure over various performance indices.
由于糖尿病的流行性增长,眼科医生需要检查巨大的眼底图像来诊断糖尿病视网膜病变(DR)。由于缺乏适当的知识,人们对糖尿病视网膜病变的检测过于迟钝。因此,早期诊断系统是医疗行业治疗疾病的必要条件。因此,建议采用一种基于深度模型的新型 DR 检测结构来解决上述难题。所开发的基于深度模型的糖尿病视网膜病变检测过程是自适应执行的。DR 检测过程是通过从基准源获取图像来模仿的。收集到的图像将进一步进入异常分割阶段。在此,使用带有增强损失函数的残差 TransUNet 来进行异常分割,这种结构中的损失函数可能有助于减少分割过程中的误差。此外,分割后的图像将进入视网膜病变检测的最后阶段。在这一阶段,检测通过自适应多尺度移动网络进行。自适应拼图优化法对 AMMNet 中的变量进行优化,以获得更好的检测性能。最后,通过对各种性能指标进行实验,确认了所提供方法的有效性。
{"title":"Automatic screening of retinal lesions for detecting diabetic retinopathy using adaptive multiscale MobileNet with abnormality segmentation from public dataset.","authors":"Nandhini Selvaganapathy, Saravanan Siddhan, Parthasarathy Sundararajan, Sathiyaprasad Balasundaram","doi":"10.1080/0954898X.2024.2424242","DOIUrl":"https://doi.org/10.1080/0954898X.2024.2424242","url":null,"abstract":"<p><p>Owing to the epidemic growth of diabetes, ophthalmologists need to examine the huge fundus images for diagnosing the disease of Diabetic Retinopathy (DR). Without proper knowledge, people are too lethargic to detect the DR. Therefore, the early diagnosis system is requisite for treating ailments in the medical industry. Therefore, a novel deep model-based DR detection structure is recommended to fix the aforementioned difficulties. The developed deep model-based diabetic retinopathy detection process is performed adaptively. The DR detection process is imitated by garnering the images from benchmark sources. The gathered images are further preceded by the abnormality segmentation phase. Here, the Residual TransUNet with Enhanced loss function is used to employ the abnormality segmentation, and the loss function in this structure may be helpful to lessen the error in the segmentation procedure. Further, the segmented images are passed to the final phase of retinopathy detection. At this phase, the detection is carried out through the Adaptive Multiscale MobileNet. The variables in the AMMNet are optimized by the Adaptive Puzzle Optimization to obtain better detection performance. Finally, the effectiveness of the offered approach is confirmed by the experimentation procedure over various performance indices.</p>","PeriodicalId":54735,"journal":{"name":"Network-Computation in Neural Systems","volume":" ","pages":"1-33"},"PeriodicalIF":1.1,"publicationDate":"2024-11-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142632789","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-11-04DOI: 10.1080/0954898X.2024.2412673
Yafei Ding, Gaomin Zhang
The human body will show very complex and diversified posture changes in the process of falling, including body posture, limb position, angle and movement trajectory, etc. The coordinates of the key points of the model are mapped to the three-dimensional space to form a three-dimensional model and obtain the three-dimensional coordinates of the key points; The construction decomposition method is used to calculate the rotation matrix of each key point, and the rotation matrix is solved to obtain the angular displacement data of the key points on different degrees of freedom. The method of curve fitting combined with the weight distribution kernel function based on self-organizing mapping theory is used to obtain the motion trajectory prediction equation of the human body falling in different degrees of freedom at random positions in three-dimensional space, determine the key point trajectory of human random fall behaviour. The experimental results show that the mapped 3D model is consistent with the real human body structure. This method can accurately determine whether the human body falls or squats randomly, and the prediction results of the key points of the human fall are consistent with the actions of the human body after the fall.
{"title":"Key point trajectory prediction method of human stochastic posture falls.","authors":"Yafei Ding, Gaomin Zhang","doi":"10.1080/0954898X.2024.2412673","DOIUrl":"https://doi.org/10.1080/0954898X.2024.2412673","url":null,"abstract":"<p><p>The human body will show very complex and diversified posture changes in the process of falling, including body posture, limb position, angle and movement trajectory, etc. The coordinates of the key points of the model are mapped to the three-dimensional space to form a three-dimensional model and obtain the three-dimensional coordinates of the key points; The construction decomposition method is used to calculate the rotation matrix of each key point, and the rotation matrix is solved to obtain the angular displacement data of the key points on different degrees of freedom. The method of curve fitting combined with the weight distribution kernel function based on self-organizing mapping theory is used to obtain the motion trajectory prediction equation of the human body falling in different degrees of freedom at random positions in three-dimensional space, determine the key point trajectory of human random fall behaviour. The experimental results show that the mapped 3D model is consistent with the real human body structure. This method can accurately determine whether the human body falls or squats randomly, and the prediction results of the key points of the human fall are consistent with the actions of the human body after the fall.</p>","PeriodicalId":54735,"journal":{"name":"Network-Computation in Neural Systems","volume":" ","pages":"1-23"},"PeriodicalIF":1.1,"publicationDate":"2024-11-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142570364","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
A brain tumour is an abnormal mass of tissue. Brain tumours vary in size, from tiny to large. Moreover, they display variations in location, shape, and size, which add complexity to their detection. The accurate delineation of tumour regions poses a challenge due to their irregular boundaries. In this research, these issues are overcome by introducing the DTDO-ZFNet for detection of brain tumour. The input Magnetic Resonance Imaging (MRI) image is fed to the pre-processing stage. Tumour areas are segmented by utilizing SegNet in which the factors of SegNet are biased using DTDO. The image augmentation is carried out using eminent techniques, such as geometric transformation and colour space transformation. Here, features such as GIST descriptor, PCA-NGIST, statistical feature and Haralick features, SLBT feature, and CNN features are extricated. Finally, the categorization of the tumour is accomplished based on ZFNet, which is trained by utilizing DTDO. The devised DTDO is a consolidation of DTBO and CDDO. The comparison of proposed DTDO-ZFNet with the existing methods, which results in highest accuracy of 0.944, a positive predictive value (PPV) of 0.936, a true positive rate (TPR) of 0.939, a negative predictive value (NPV) of 0.937, and a minimal false-negative rate (FNR) of 0.061%.
{"title":"DTDO: Driving Training Development Optimization enabled deep learning approach for brain tumour classification using MRI.","authors":"Vadamodula Prasad, Issac Diana Jeba Jingle, Gopalsamy Venkadakrishnan Sriramakrishnan","doi":"10.1080/0954898X.2024.2351159","DOIUrl":"10.1080/0954898X.2024.2351159","url":null,"abstract":"<p><p>A brain tumour is an abnormal mass of tissue. Brain tumours vary in size, from tiny to large. Moreover, they display variations in location, shape, and size, which add complexity to their detection. The accurate delineation of tumour regions poses a challenge due to their irregular boundaries. In this research, these issues are overcome by introducing the DTDO-ZFNet for detection of brain tumour. The input Magnetic Resonance Imaging (MRI) image is fed to the pre-processing stage. Tumour areas are segmented by utilizing SegNet in which the factors of SegNet are biased using DTDO. The image augmentation is carried out using eminent techniques, such as geometric transformation and colour space transformation. Here, features such as GIST descriptor, PCA-NGIST, statistical feature and Haralick features, SLBT feature, and CNN features are extricated. Finally, the categorization of the tumour is accomplished based on ZFNet, which is trained by utilizing DTDO. The devised DTDO is a consolidation of DTBO and CDDO. The comparison of proposed DTDO-ZFNet with the existing methods, which results in highest accuracy of 0.944, a positive predictive value (PPV) of 0.936, a true positive rate (TPR) of 0.939, a negative predictive value (NPV) of 0.937, and a minimal false-negative rate (FNR) of 0.061%.</p>","PeriodicalId":54735,"journal":{"name":"Network-Computation in Neural Systems","volume":" ","pages":"520-561"},"PeriodicalIF":1.1,"publicationDate":"2024-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141154757","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}