Pub Date : 2024-09-19DOI: 10.1007/s00521-024-10373-1
Romulo Augusto Aires Soares, Alexandre Cesar Muniz de Oliveira, Paulo Rogerio de Almeida Ribeiro, Areolino de Almeida Neto
This paper presents a new strategy that uses multiple neural networks in conjunction with the DEtection TRansformer (DETR) network to detect firearms in surveillance images. The strategy developed in this work presents a methodology that promotes collaboration and self-coordination of networks in the fully connected layers of DETR through the technique of multiple self-coordinating artificial neural networks (MANN), which does not require a coordinator. This self-coordination consists of training the networks one after the other and integrating their outputs without an extra element called a coordinator. The results indicate that the proposed network is highly effective, achieving high-level outcomes in firearm detection. The network’s high precision of 84% and its ability to perform classifications are noteworthy.
{"title":"Firearm detection using DETR with multiple self-coordinated neural networks","authors":"Romulo Augusto Aires Soares, Alexandre Cesar Muniz de Oliveira, Paulo Rogerio de Almeida Ribeiro, Areolino de Almeida Neto","doi":"10.1007/s00521-024-10373-1","DOIUrl":"https://doi.org/10.1007/s00521-024-10373-1","url":null,"abstract":"<p>This paper presents a new strategy that uses multiple neural networks in conjunction with the DEtection TRansformer (DETR) network to detect firearms in surveillance images. The strategy developed in this work presents a methodology that promotes collaboration and self-coordination of networks in the fully connected layers of DETR through the technique of multiple self-coordinating artificial neural networks (MANN), which does not require a coordinator. This self-coordination consists of training the networks one after the other and integrating their outputs without an extra element called a coordinator. The results indicate that the proposed network is highly effective, achieving high-level outcomes in firearm detection. The network’s high precision of 84% and its ability to perform classifications are noteworthy.</p>","PeriodicalId":18925,"journal":{"name":"Neural Computing and Applications","volume":"4 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142251068","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-19DOI: 10.1007/s00521-024-10379-9
Ivan David Briceño-Pinzón, Raquel Maria de Oliveira Pires, Geraldo Andrade Carvalho, Flávia Barbosa Silva Botelho, Júlia Lima Baute, Marcela Carlota Nery
The X-ray method, together with image analysis tools, has been used to evaluate the internal structures of seeds and correlate them with the physical, physiological and sanitary quality, providing significant and accurate results. The objective of this study was to analyze radiographic images of rice seeds infested by the rice weevil Sitophilus oryzae (Linnaeus, 1763) (Coleoptera: Curculionidae). Rice seed samples from three different cultivars were infested with S. oryzae for 90 days. Next, seed samples collected at random were analyzed by X-ray testing. The radiographic images were analyzed by ImageJ® software to extract color and shape features. Scanning electron microscopy analyses were also performed. The results showed that X-ray testing was effective in detecting infestation. The gray distribution histograms revealed differences between healthy seeds and those infested by adult insects or empty seeds, confirmed by the significant differences obtained for the area and relative and integrated density variables. The study demonstrated that the analysis of radiographic images can provide quantitative information on insect infestation of rice seeds, which is useful in the evaluation of seed quality and for detecting the presence of pests in rice seeds.
X 射线方法和图像分析工具已被用于评估种子的内部结构,并将其与物理、生理和卫生质量联系起来,从而提供重要而准确的结果。本研究的目的是分析受稻象虫(Sitophilus oryzae,Linnaeus,1763)(鞘翅目:卷须科)侵染的水稻种子的射线图像。来自三个不同栽培品种的水稻种子样本被 S. oryzae 侵染了 90 天。然后,对随机采集的种子样本进行 X 射线检测分析。利用 ImageJ® 软件对射线图像进行分析,提取颜色和形状特征。同时还进行了扫描电子显微镜分析。结果表明,X 射线检测能有效发现虫害。灰色分布直方图显示了健康种子与受成虫侵染种子或空种子之间的差异,面积、相对密度和综合密度变量的显著差异也证实了这一点。该研究表明,射线图像分析可提供有关水稻种子虫害的定量信息,有助于评估种子质量和检测水稻种子中是否存在害虫。
{"title":"Potential analysis of radiographic images to determine infestation of rice seeds","authors":"Ivan David Briceño-Pinzón, Raquel Maria de Oliveira Pires, Geraldo Andrade Carvalho, Flávia Barbosa Silva Botelho, Júlia Lima Baute, Marcela Carlota Nery","doi":"10.1007/s00521-024-10379-9","DOIUrl":"https://doi.org/10.1007/s00521-024-10379-9","url":null,"abstract":"<p>The X-ray method, together with image analysis tools, has been used to evaluate the internal structures of seeds and correlate them with the physical, physiological and sanitary quality, providing significant and accurate results. The objective of this study was to analyze radiographic images of rice seeds infested by the rice weevil <i>Sitophilus oryzae</i> (Linnaeus, 1763) (Coleoptera: Curculionidae). Rice seed samples from three different cultivars were infested with <i>S. oryzae</i> for 90 days. Next, seed samples collected at random were analyzed by X-ray testing. The radiographic images were analyzed by ImageJ® software to extract color and shape features. Scanning electron microscopy analyses were also performed. The results showed that X-ray testing was effective in detecting infestation. The gray distribution histograms revealed differences between healthy seeds and those infested by adult insects or empty seeds, confirmed by the significant differences obtained for the area and relative and integrated density variables. The study demonstrated that the analysis of radiographic images can provide quantitative information on insect infestation of rice seeds, which is useful in the evaluation of seed quality and for detecting the presence of pests in rice seeds.</p>","PeriodicalId":18925,"journal":{"name":"Neural Computing and Applications","volume":"29 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142251065","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-19DOI: 10.1007/s00521-024-10411-y
Delmiro D. Sampaio-Neto, Telmo M. Silva Filho, Renata M. C. R. Souza
Most recommendation systems are implemented using numerical or categorical data, that is, traditional data. This type of data can be a limiting factor when used to model complex concepts where there is internal variability or internal structure in the data. To overcome these limitations, symbolic data is used, where data can be represented by different types of values, such as intervals, lists, or histograms. This work introduces a single approach to constructing recommendation systems based on content or based on collaborative filtering using modal variables for users and items. In the content-based system, user profiles and item profiles are created from modal representations of their features, and a list of items is matched against a user profile. For collaborative filtering, user profiles are built, and users are grouped to form a neighborhood, products rated by users of this neighborhood are recommended based on the similarity between the neighbor and the user who will receive the recommendation. Experiments are carried out, using a movie domain dataset, to evaluate the effectiveness of the proposed approach. The outcomes suggest our ability to generate ranked lists of superior quality compared to previous methods utilizing symbolic data. Specifically, the lists created through the proposed method exhibit higher normalized discounted cumulative gain and, in qualitative terms, showcase more diverse content.
{"title":"Recommendation systems with user and item profiles based on symbolic modal data","authors":"Delmiro D. Sampaio-Neto, Telmo M. Silva Filho, Renata M. C. R. Souza","doi":"10.1007/s00521-024-10411-y","DOIUrl":"https://doi.org/10.1007/s00521-024-10411-y","url":null,"abstract":"<p>Most recommendation systems are implemented using numerical or categorical data, that is, traditional data. This type of data can be a limiting factor when used to model complex concepts where there is internal variability or internal structure in the data. To overcome these limitations, symbolic data is used, where data can be represented by different types of values, such as intervals, lists, or histograms. This work introduces a single approach to constructing recommendation systems based on content or based on collaborative filtering using modal variables for users and items. In the content-based system, user profiles and item profiles are created from modal representations of their features, and a list of items is matched against a user profile. For collaborative filtering, user profiles are built, and users are grouped to form a neighborhood, products rated by users of this neighborhood are recommended based on the similarity between the neighbor and the user who will receive the recommendation. Experiments are carried out, using a movie domain dataset, to evaluate the effectiveness of the proposed approach. The outcomes suggest our ability to generate ranked lists of superior quality compared to previous methods utilizing symbolic data. Specifically, the lists created through the proposed method exhibit higher normalized discounted cumulative gain and, in qualitative terms, showcase more diverse content.</p>","PeriodicalId":18925,"journal":{"name":"Neural Computing and Applications","volume":"21 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142251066","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
This study investigates the use of electroencephalography (EEG) to characterize emotions and provides insights into the consistency between self-reported and machine learning outcomes. Thirty participants engaged in five virtual reality environments designed to elicit specific emotions, while their brain activity was recorded. The participants self-assessed their ground truth emotional state in terms of Arousal and Valence through a Self-Assessment Manikin. Gradient Boosted Decision Tree was adopted as a classification algorithm to test the EEG feasibility in the characterization of emotional states. Distinctive patterns of neural activation corresponding to different levels of Valence and Arousal emerged, and a noteworthy correspondence between the outcomes of the self-assessments and the classifier suggested that EEG-based affective indicators can be successfully applied in emotional characterization, shedding light on the possibility of using them as ground truth measurements. These findings provide compelling evidence for the validity of EEG as a tool for emotion characterization and its contribution to a better understanding of emotional activation.
{"title":"Effective affective EEG-based indicators in emotion-evoking VR environments: an evidence from machine learning","authors":"Ivonne Angelica Castiblanco Jimenez, Elena Carlotta Olivetti, Enrico Vezzetti, Sandro Moos, Alessia Celeghin, Federica Marcolin","doi":"10.1007/s00521-024-10240-z","DOIUrl":"https://doi.org/10.1007/s00521-024-10240-z","url":null,"abstract":"<p>This study investigates the use of electroencephalography (EEG) to characterize emotions and provides insights into the consistency between self-reported and machine learning outcomes. Thirty participants engaged in five virtual reality environments designed to elicit specific emotions, while their brain activity was recorded. The participants self-assessed their ground truth emotional state in terms of Arousal and Valence through a Self-Assessment Manikin. Gradient Boosted Decision Tree was adopted as a classification algorithm to test the EEG feasibility in the characterization of emotional states. Distinctive patterns of neural activation corresponding to different levels of Valence and Arousal emerged, and a noteworthy correspondence between the outcomes of the self-assessments and the classifier suggested that EEG-based affective indicators can be successfully applied in emotional characterization, shedding light on the possibility of using them as ground truth measurements. These findings provide compelling evidence for the validity of EEG as a tool for emotion characterization and its contribution to a better understanding of emotional activation.</p>","PeriodicalId":18925,"journal":{"name":"Neural Computing and Applications","volume":"17 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142251070","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-19DOI: 10.1007/s00521-024-10415-8
Orcun Yildiz, Krishnan Raghavan, Henry Chan, Mathew J. Cherukara, Prasanna Balaprakash, Subramanian Sankaranarayanan, Tom Peterka
X-ray Bragg coherent diffraction imaging is a powerful technique for 3D materials characterization. However, obtaining X-ray diffraction data is difficult and computationally intensive, motivating the need for automated processing of coherent diffraction images, with the goal of minimizing the number of X-ray datasets needed. We automate a machine learning approach to identify crystalline line defects in samples from the raw coherent diffraction data, in a workflow coupling coherent diffraction data generation with training and inference of deep neural network defect classifiers. In particular, we adopt a continual learning approach, where we generate training data as needed based on the accuracy of the defect classifier instead of generating all training data a priori. Moreover, we develop a novel data generation mechanism to improve the efficiency of defect identification beyond the previously published continual learning approach. We call the improved method smart continual learning. The results show that our approach improves the accuracy of defect classifiers and reduces training data requirements by up to 98% compared with prior approaches.
X 射线布拉格相干衍射成像是三维材料表征的一项强大技术。然而,获取 X 射线衍射数据既困难又耗费计算资源,因此需要对相干衍射图像进行自动处理,以最大限度地减少所需的 X 射线数据集数量。我们将相干衍射数据生成与深度神经网络缺陷分类器的训练和推理相结合,在工作流程中采用机器学习方法,从原始相干衍射数据中自动识别样品中的结晶线缺陷。特别是,我们采用了一种持续学习方法,即根据缺陷分类器的准确性在需要时生成训练数据,而不是事先生成所有训练数据。此外,我们还开发了一种新颖的数据生成机制,以提高缺陷识别效率,超越之前发布的持续学习方法。我们将改进后的方法称为智能持续学习。结果表明,与之前的方法相比,我们的方法提高了缺陷分类器的准确性,并减少了高达 98% 的训练数据需求。
{"title":"Automated defect identification in coherent diffraction imaging with smart continual learning","authors":"Orcun Yildiz, Krishnan Raghavan, Henry Chan, Mathew J. Cherukara, Prasanna Balaprakash, Subramanian Sankaranarayanan, Tom Peterka","doi":"10.1007/s00521-024-10415-8","DOIUrl":"https://doi.org/10.1007/s00521-024-10415-8","url":null,"abstract":"<p>X-ray Bragg coherent diffraction imaging is a powerful technique for 3D materials characterization. However, obtaining X-ray diffraction data is difficult and computationally intensive, motivating the need for automated processing of coherent diffraction images, with the goal of minimizing the number of X-ray datasets needed. We automate a machine learning approach to identify crystalline line defects in samples from the raw coherent diffraction data, in a workflow coupling coherent diffraction data generation with training and inference of deep neural network defect classifiers. In particular, we adopt a continual learning approach, where we generate training data as needed based on the accuracy of the defect classifier instead of generating all training data a priori. Moreover, we develop a novel data generation mechanism to improve the efficiency of defect identification beyond the previously published continual learning approach. We call the improved method <i>smart continual learning.</i> The results show that our approach improves the accuracy of defect classifiers and reduces training data requirements by up to 98% compared with prior approaches.</p>","PeriodicalId":18925,"journal":{"name":"Neural Computing and Applications","volume":"198 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142251069","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-19DOI: 10.1007/s00521-024-10422-9
Pedro A. Villa-García, Raúl Alonso-Calvo, Miguel García-Remesal
A novel methodology is introduced for extracting entities from noisy scanned documents by using end-to-end data and reformulating the entity extraction task as a text summarization problem. This approach offers two significant advantages over traditional entity extraction methods while maintaining comparable performance. First, it utilizes preexisting data to construct datasets, thereby eliminating the need for labor-intensive annotation procedures. Second, it employs multitask learning, enabling the training of a model via a single dataset. To evaluate our approach against state-of-the-art methods, we adapted three commonly used datasets, namely, Conference on Natural Language Learning (CoNLL++), few-shot named entity recognition (Few-NERD), and WikiNEuRal domain adaptation (WikiNEuRal + DA), to the format required by our methodology. We subsequently fine-tuned four sequence-to-sequence models: text-to-text transfer transformer (T5), fine-tuned language net T5 (FLAN-T5), bidirectional autoregressive transformer (BART), and pretraining with extracted gap sentences for abstractive summarization sequence-to-sequence models (PEGASUS). The results indicate that, in the absence of optical character recognition (OCR) noise, the BART model performs comparably to state-of-the-art methods. Furthermore, the performance degradation was limited to 3.49–5.23% when 39–62% of the sentences contained OCR noise. This performance is significantly superior to that of previous studies, which reported a 10–20% decrease in the F1 score with texts that had a 20% OCR error rate. Our experimental results demonstrate that a single model trained via our methodology can reliably extract entities from noisy OCRed texts, unlike existing state-of-the-art approaches, which require separate models for correcting OCR errors and extracting entities.
{"title":"End-to-end entity extraction from OCRed texts using summarization models","authors":"Pedro A. Villa-García, Raúl Alonso-Calvo, Miguel García-Remesal","doi":"10.1007/s00521-024-10422-9","DOIUrl":"https://doi.org/10.1007/s00521-024-10422-9","url":null,"abstract":"<p>A novel methodology is introduced for extracting entities from noisy scanned documents by using end-to-end data and reformulating the entity extraction task as a text summarization problem. This approach offers two significant advantages over traditional entity extraction methods while maintaining comparable performance. First, it utilizes preexisting data to construct datasets, thereby eliminating the need for labor-intensive annotation procedures. Second, it employs multitask learning, enabling the training of a model via a single dataset. To evaluate our approach against state-of-the-art methods, we adapted three commonly used datasets, namely, Conference on Natural Language Learning (CoNLL++), few-shot named entity recognition (Few-NERD), and WikiNEuRal domain adaptation (WikiNEuRal + DA), to the format required by our methodology. We subsequently fine-tuned four sequence-to-sequence models: text-to-text transfer transformer (T5), fine-tuned language net T5 (FLAN-T5), bidirectional autoregressive transformer (BART), and pretraining with extracted gap sentences for abstractive summarization sequence-to-sequence models (PEGASUS). The results indicate that, in the absence of optical character recognition (OCR) noise, the BART model performs comparably to state-of-the-art methods. Furthermore, the performance degradation was limited to 3.49–5.23% when 39–62% of the sentences contained OCR noise. This performance is significantly superior to that of previous studies, which reported a 10–20% decrease in the F1 score with texts that had a 20% OCR error rate. Our experimental results demonstrate that a single model trained via our methodology can reliably extract entities from noisy OCRed texts, unlike existing state-of-the-art approaches, which require separate models for correcting OCR errors and extracting entities.</p>","PeriodicalId":18925,"journal":{"name":"Neural Computing and Applications","volume":"15 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142251067","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-19DOI: 10.1007/s00521-024-10427-4
Richard Ningthoujam, Keisham Pritamdas, Loitongbam Surajkumar Singh
The object detection model based on the transfer learning approach comprises feature extraction and detection layers. YOLOv2 is among the fastest detection algorithms, which can utilize various pretrained classifier networks for feature extraction. However, reducing the number of network layers and increasing the mean average precision (mAP) together have challenges. Darknet-19-based YOLOv2 model achieved an mAP of 76.78% by having a smaller number of layers than other existing models. This work proposes modification by adding layers that help enhance feature extraction for further increasing the mAP of the model. Above that, the initial weights of the new layers can be random or deterministic, fine-tuned during training. In our work, we introduce a block of layers initialized with deterministic weights derived from several edge detection filter weights. Integrating such a block to the darknet-19-based object detection model improves the mAP to 85.94%, outperforming the other existing model in terms of mAP and number of layers.
{"title":"Edge detective weights initialization on Darknet-19 model for YOLOv2-based facemask detection","authors":"Richard Ningthoujam, Keisham Pritamdas, Loitongbam Surajkumar Singh","doi":"10.1007/s00521-024-10427-4","DOIUrl":"https://doi.org/10.1007/s00521-024-10427-4","url":null,"abstract":"<p>The object detection model based on the transfer learning approach comprises feature extraction and detection layers. YOLOv2 is among the fastest detection algorithms, which can utilize various pretrained classifier networks for feature extraction. However, reducing the number of network layers and increasing the mean average precision (mAP) together have challenges. Darknet-19-based YOLOv2 model achieved an mAP of 76.78% by having a smaller number of layers than other existing models. This work proposes modification by adding layers that help enhance feature extraction for further increasing the mAP of the model. Above that, the initial weights of the new layers can be random or deterministic, fine-tuned during training. In our work, we introduce a block of layers initialized with deterministic weights derived from several edge detection filter weights. Integrating such a block to the darknet-19-based object detection model improves the mAP to 85.94%, outperforming the other existing model in terms of mAP and number of layers.</p>","PeriodicalId":18925,"journal":{"name":"Neural Computing and Applications","volume":"37 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142251073","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-19DOI: 10.1007/s00521-024-10362-4
Asim Naveed, Syed S. Naqvi, Tariq M. Khan, Shahzaib Iqbal, M. Yaqoob Wani, Haroon Ahmed Khan
In computer-aided diagnosis tools employed for skin cancer treatment and early diagnosis, skin lesion segmentation is important. However, achieving precise segmentation is challenging due to inherent variations in appearance, contrast, texture, and blurry lesion boundaries. This research presents a robust approach utilizing a dilated convolutional residual network, which incorporates an attention-based spatial feature enhancement block (ASFEB) and employs a guided decoder strategy. In each dilated convolutional residual block, dilated convolution is employed to broaden the receptive field with varying dilation rates. To improve the spatial feature information of the encoder, we employed an attention-based spatial feature enhancement block in the skip connections. The ASFEB in our proposed method combines feature maps obtained from average and maximum-pooling operations. These combined features are then weighted using the active outcome of global average pooling and convolution operations. Additionally, we have incorporated a guided decoder strategy, where each decoder block is optimized using an individual loss function to enhance the feature learning process in the proposed AD-Net. The proposed AD-Net presents a significant benefit by necessitating fewer model parameters compared to its peer methods. This reduction in parameters directly impacts the number of labeled data required for training, facilitating faster convergence during the training process. The effectiveness of the proposed AD-Net was evaluated using four public benchmark datasets. We conducted a Wilcoxon signed-rank test to verify the efficiency of the AD-Net. The outcomes suggest that our method surpasses other cutting-edge methods in performance, even without the implementation of data augmentation strategies.
{"title":"AD-Net: Attention-based dilated convolutional residual network with guided decoder for robust skin lesion segmentation","authors":"Asim Naveed, Syed S. Naqvi, Tariq M. Khan, Shahzaib Iqbal, M. Yaqoob Wani, Haroon Ahmed Khan","doi":"10.1007/s00521-024-10362-4","DOIUrl":"https://doi.org/10.1007/s00521-024-10362-4","url":null,"abstract":"<p>In computer-aided diagnosis tools employed for skin cancer treatment and early diagnosis, skin lesion segmentation is important. However, achieving precise segmentation is challenging due to inherent variations in appearance, contrast, texture, and blurry lesion boundaries. This research presents a robust approach utilizing a dilated convolutional residual network, which incorporates an attention-based spatial feature enhancement block (ASFEB) and employs a guided decoder strategy. In each dilated convolutional residual block, dilated convolution is employed to broaden the receptive field with varying dilation rates. To improve the spatial feature information of the encoder, we employed an attention-based spatial feature enhancement block in the skip connections. The ASFEB in our proposed method combines feature maps obtained from average and maximum-pooling operations. These combined features are then weighted using the active outcome of global average pooling and convolution operations. Additionally, we have incorporated a guided decoder strategy, where each decoder block is optimized using an individual loss function to enhance the feature learning process in the proposed AD-Net. The proposed AD-Net presents a significant benefit by necessitating fewer model parameters compared to its peer methods. This reduction in parameters directly impacts the number of labeled data required for training, facilitating faster convergence during the training process. The effectiveness of the proposed AD-Net was evaluated using four public benchmark datasets. We conducted a Wilcoxon signed-rank test to verify the efficiency of the AD-Net. The outcomes suggest that our method surpasses other cutting-edge methods in performance, even without the implementation of data augmentation strategies.</p>","PeriodicalId":18925,"journal":{"name":"Neural Computing and Applications","volume":"29 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142268762","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-18DOI: 10.1007/s00521-024-10352-6
Sangram K. Jena, K. Subramani, Alvaro Velasquez
Kidney transplantation is vital for treating end-stage renal disease, impacting roughly one in a thousand Europeans. The search for a suitable deceased donor often leads to prolonged and uncertain wait times, making living donor transplants a viable alternative. However, approximately 40% of living donors are incompatible with their intended recipients. Therefore, many countries have established kidney exchange programs, allowing patients with incompatible donors to participate in “swap” arrangements, exchanging donors with other patients in similar situations. Several variants of the vertex-disjoint cycle cover problem model the above problem, which deals with different aspects of kidney exchange as required. This paper discusses several specific vertex-disjoint cycle cover variants and deals with finding the exact solution. We employ the dataless neural networks framework to establish single differentiable functions for each variant. Recent research highlights the framework’s effectiveness in representing several combinatorial optimization problems. Inspired by these findings, we propose customized dataless neural networks for vertex-disjoint cycle cover variants. We derive a differentiable function for each variant and prove that the function will attain its minimum value if an exact solution is found for the corresponding problem variant. We also provide proof of the correctness of our approach.
{"title":"Designing dataless neural networks for kidney exchange variants","authors":"Sangram K. Jena, K. Subramani, Alvaro Velasquez","doi":"10.1007/s00521-024-10352-6","DOIUrl":"https://doi.org/10.1007/s00521-024-10352-6","url":null,"abstract":"<p>Kidney transplantation is vital for treating end-stage renal disease, impacting roughly one in a thousand Europeans. The search for a suitable deceased donor often leads to prolonged and uncertain wait times, making living donor transplants a viable alternative. However, approximately 40% of living donors are incompatible with their intended recipients. Therefore, many countries have established kidney exchange programs, allowing patients with incompatible donors to participate in “swap” arrangements, exchanging donors with other patients in similar situations. Several variants of the vertex-disjoint cycle cover problem model the above problem, which deals with different aspects of kidney exchange as required. This paper discusses several specific vertex-disjoint cycle cover variants and deals with finding the exact solution. We employ the dataless neural networks framework to establish single differentiable functions for each variant. Recent research highlights the framework’s effectiveness in representing several combinatorial optimization problems. Inspired by these findings, we propose customized dataless neural networks for vertex-disjoint cycle cover variants. We derive a differentiable function for each variant and prove that the function will attain its minimum value if an exact solution is found for the corresponding problem variant. We also provide proof of the correctness of our approach.</p>","PeriodicalId":18925,"journal":{"name":"Neural Computing and Applications","volume":"3 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142251064","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-18DOI: 10.1007/s00521-024-10139-9
Fabio Merizzi, Andrea Asperti, Stefano Colamonaco
The Copernicus Regional Reanalysis for Europe, CERRA, is a high-resolution regional reanalysis dataset for the European domain. In recent years, it has shown significant utility across various climate-related tasks, ranging from forecasting and climate change research to renewable energy prediction, resource management, air quality risk assessment, and the forecasting of rare events, among others. Unfortunately, the availability of CERRA is lagging 2 years behind the current date, due to constraints in acquiring the requisite external data and the intensive computational demands inherent in its generation. As a solution, this paper introduces a novel method using diffusion models to approximate CERRA downscaling in a data-driven manner, without additional informations. By leveraging the lower resolution ERA5 dataset, which provides boundary conditions for CERRA, we approach this as a super-resolution task. Focusing on wind speed around Italy, our model, trained on existing CERRA data, shows promising results, closely mirroring the original CERRA. Validation with in-situ observations further confirms the model’s accuracy in approximating ground measurements.
{"title":"Wind speed super-resolution and validation: from ERA5 to CERRA via diffusion models","authors":"Fabio Merizzi, Andrea Asperti, Stefano Colamonaco","doi":"10.1007/s00521-024-10139-9","DOIUrl":"https://doi.org/10.1007/s00521-024-10139-9","url":null,"abstract":"<p>The Copernicus Regional Reanalysis for Europe, CERRA, is a high-resolution regional reanalysis dataset for the European domain. In recent years, it has shown significant utility across various climate-related tasks, ranging from forecasting and climate change research to renewable energy prediction, resource management, air quality risk assessment, and the forecasting of rare events, among others. Unfortunately, the availability of CERRA is lagging 2 years behind the current date, due to constraints in acquiring the requisite external data and the intensive computational demands inherent in its generation. As a solution, this paper introduces a novel method using diffusion models to approximate CERRA downscaling in a data-driven manner, without additional informations. By leveraging the lower resolution ERA5 dataset, which provides boundary conditions for CERRA, we approach this as a super-resolution task. Focusing on wind speed around Italy, our model, trained on existing CERRA data, shows promising results, closely mirroring the original CERRA. Validation with in-situ observations further confirms the model’s accuracy in approximating ground measurements.</p>","PeriodicalId":18925,"journal":{"name":"Neural Computing and Applications","volume":"65 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142251072","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}