Pub Date : 2023-01-01DOI: 10.1007/s00521-023-08319-0
S Sivamohan, S S Sridhar
Industry 4.0 enable novel business cases, such as client-specific production, real-time monitoring of process condition and progress, independent decision making and remote maintenance, to name a few. However, they are more susceptible to a broad range of cyber threats because of limited resources and heterogeneous nature. Such risks cause financial and reputational damages for businesses, well as the theft of sensitive information. The higher level of diversity in industrial network prevents the attackers from such attacks. Therefore, to efficiently detect the intrusions, a novel intrusion detection system known as Bidirectional Long Short-Term Memory based Explainable Artificial Intelligence framework (BiLSTM-XAI) is developed. Initially, the preprocessing task using data cleaning and normalization is performed to enhance the data quality for detecting network intrusions. Subsequently, the significant features are selected from the databases using the Krill herd optimization (KHO) algorithm. The proposed BiLSTM-XAI approach provides better security and privacy inside the industry networking system by detecting intrusions very precisely. In this, we utilized SHAP and LIME explainable AI algorithms to improve interpretation of prediction results. The experimental setup is made by MATLAB 2016 software using Honeypot and NSL-KDD datasets as input. The analysis result reveals that the proposed method achieves superior performance in detecting intrusions with a classification accuracy of 98.2%.
{"title":"An optimized model for network intrusion detection systems in industry 4.0 using XAI based Bi-LSTM framework.","authors":"S Sivamohan, S S Sridhar","doi":"10.1007/s00521-023-08319-0","DOIUrl":"https://doi.org/10.1007/s00521-023-08319-0","url":null,"abstract":"<p><p>Industry 4.0 enable novel business cases, such as client-specific production, real-time monitoring of process condition and progress, independent decision making and remote maintenance, to name a few. However, they are more susceptible to a broad range of cyber threats because of limited resources and heterogeneous nature. Such risks cause financial and reputational damages for businesses, well as the theft of sensitive information. The higher level of diversity in industrial network prevents the attackers from such attacks. Therefore, to efficiently detect the intrusions, a novel intrusion detection system known as Bidirectional Long Short-Term Memory based Explainable Artificial Intelligence framework (BiLSTM-XAI) is developed. Initially, the preprocessing task using data cleaning and normalization is performed to enhance the data quality for detecting network intrusions. Subsequently, the significant features are selected from the databases using the Krill herd optimization (KHO) algorithm. The proposed BiLSTM-XAI approach provides better security and privacy inside the industry networking system by detecting intrusions very precisely. In this, we utilized SHAP and LIME explainable AI algorithms to improve interpretation of prediction results. The experimental setup is made by MATLAB 2016 software using Honeypot and NSL-KDD datasets as input. The analysis result reveals that the proposed method achieves superior performance in detecting intrusions with a classification accuracy of 98.2%.</p>","PeriodicalId":49766,"journal":{"name":"Neural Computing & Applications","volume":"35 15","pages":"11459-11475"},"PeriodicalIF":6.0,"publicationDate":"2023-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9999327/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9439695","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Nowadays, deep learning plays a vital role behind many of the emerging technologies. Few applications of deep learning include speech recognition, virtual assistant, healthcare, entertainment, and so on. In healthcare applications, deep learning can be used to predict diseases effectively. It is a type of computer model that learns in conducting classification tasks directly from text, sound, or images. It also provides better accuracy and sometimes outdoes human performance. We presented a novel approach that makes use of the deep learning method in our proposed work. The prediction of pulmonary disease can be performed with the aid of convolutional neural network (CNN) incorporated with auction-based optimization algorithm (ABOA) and DSC process. The traditional CNN ignores the dominant features from the X-ray images while performing the feature extraction process. This can be effectively circumvented by the adoption of ABOA, and the DSC is used to classify the pulmonary disease types such as fibrosis, pneumonia, cardiomegaly, and normal from the X-ray images. We have taken two datasets, namely the NIH Chest X-ray dataset and ChestX-ray8. The performances of the proposed approach are compared with deep learning-based state-of-art works such as BPD, DL, CSS-DL, and Grad-CAM. From the performance analyses, it is confirmed that the proposed approach effectively extracts the features from the X-ray images, and thus, the prediction of pulmonary diseases is more accurate than the state-of-art approaches.
{"title":"ABOA-CNN: auction-based optimization algorithm with convolutional neural network for pulmonary disease prediction.","authors":"Balaji Annamalai, Prabakeran Saravanan, Indumathi Varadharajan","doi":"10.1007/s00521-022-08033-3","DOIUrl":"https://doi.org/10.1007/s00521-022-08033-3","url":null,"abstract":"<p><p>Nowadays, deep learning plays a vital role behind many of the emerging technologies. Few applications of deep learning include speech recognition, virtual assistant, healthcare, entertainment, and so on. In healthcare applications, deep learning can be used to predict diseases effectively. It is a type of computer model that learns in conducting classification tasks directly from text, sound, or images. It also provides better accuracy and sometimes outdoes human performance. We presented a novel approach that makes use of the deep learning method in our proposed work. The prediction of pulmonary disease can be performed with the aid of convolutional neural network (CNN) incorporated with auction-based optimization algorithm (ABOA) and DSC process. The traditional CNN ignores the dominant features from the X-ray images while performing the feature extraction process. This can be effectively circumvented by the adoption of ABOA, and the DSC is used to classify the pulmonary disease types such as fibrosis, pneumonia, cardiomegaly, and normal from the X-ray images. We have taken two datasets, namely the NIH Chest X-ray dataset and ChestX-ray8. The performances of the proposed approach are compared with deep learning-based state-of-art works such as BPD, DL, CSS-DL, and Grad-CAM. From the performance analyses, it is confirmed that the proposed approach effectively extracts the features from the X-ray images, and thus, the prediction of pulmonary diseases is more accurate than the state-of-art approaches.</p>","PeriodicalId":49766,"journal":{"name":"Neural Computing & Applications","volume":"35 10","pages":"7463-7474"},"PeriodicalIF":6.0,"publicationDate":"2023-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9910772/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9506281","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-01-01DOI: 10.1007/s00521-023-08442-y
Md Rajib Hossain, Mohammed Moshiul Hoque, Nazmul Siddique, Iqbal H Sarker
Covid text identification (CTI) is a crucial research concern in natural language processing (NLP). Social and electronic media are simultaneously adding a large volume of Covid-affiliated text on the World Wide Web due to the effortless access to the Internet, electronic gadgets and the Covid outbreak. Most of these texts are uninformative and contain misinformation, disinformation and malinformation that create an infodemic. Thus, Covid text identification is essential for controlling societal distrust and panic. Though very little Covid-related research (such as Covid disinformation, misinformation and fake news) has been reported in high-resource languages (e.g. English), CTI in low-resource languages (like Bengali) is in the preliminary stage to date. However, automatic CTI in Bengali text is challenging due to the deficit of benchmark corpora, complex linguistic constructs, immense verb inflexions and scarcity of NLP tools. On the other hand, the manual processing of Bengali Covid texts is arduous and costly due to their messy or unstructured forms. This research proposes a deep learning-based network (CovTiNet) to identify Covid text in Bengali. The CovTiNet incorporates an attention-based position embedding feature fusion for text-to-feature representation and attention-based CNN for Covid text identification. Experimental results show that the proposed CovTiNet achieved the highest accuracy of 96.61±.001% on the developed dataset (BCovC) compared to the other methods and baselines (i.e. BERT-M, IndicBERT, ELECTRA-Bengali, DistilBERT-M, BiLSTM, DCNN, CNN, LSTM, VDCNN and ACNN).
{"title":"CovTiNet: Covid text identification network using attention-based positional embedding feature fusion.","authors":"Md Rajib Hossain, Mohammed Moshiul Hoque, Nazmul Siddique, Iqbal H Sarker","doi":"10.1007/s00521-023-08442-y","DOIUrl":"https://doi.org/10.1007/s00521-023-08442-y","url":null,"abstract":"<p><p>Covid text identification (CTI) is a crucial research concern in natural language processing (NLP). Social and electronic media are simultaneously adding a large volume of Covid-affiliated text on the World Wide Web due to the effortless access to the Internet, electronic gadgets and the Covid outbreak. Most of these texts are uninformative and contain misinformation, disinformation and malinformation that create an infodemic. Thus, Covid text identification is essential for controlling societal distrust and panic. Though very little Covid-related research (such as Covid disinformation, misinformation and fake news) has been reported in high-resource languages (e.g. English), CTI in low-resource languages (like Bengali) is in the preliminary stage to date. However, automatic CTI in Bengali text is challenging due to the deficit of benchmark corpora, complex linguistic constructs, immense verb inflexions and scarcity of NLP tools. On the other hand, the manual processing of Bengali Covid texts is arduous and costly due to their messy or unstructured forms. This research proposes a deep learning-based network (CovTiNet) to identify Covid text in Bengali. The CovTiNet incorporates an attention-based position embedding feature fusion for text-to-feature representation and attention-based CNN for Covid text identification. Experimental results show that the proposed CovTiNet achieved the highest accuracy of 96.61±.001% on the developed dataset (<i>BCovC</i>) compared to the other methods and baselines (i.e. BERT-M, IndicBERT, ELECTRA-Bengali, DistilBERT-M, BiLSTM, DCNN, CNN, LSTM, VDCNN and ACNN).</p>","PeriodicalId":49766,"journal":{"name":"Neural Computing & Applications","volume":"35 18","pages":"13503-13527"},"PeriodicalIF":6.0,"publicationDate":"2023-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10011801/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9507549","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-01-01Epub Date: 2023-04-10DOI: 10.1007/s00521-023-08407-1
M Murugappan, Ali K Bourisly, N B Prakash, M G Sumithra, U Rajendra Acharya
Lung segmentation algorithms play a significant role in segmenting theinfected regions in the lungs. This work aims to develop a computationally efficient and robust deep learning model for lung segmentation using chest computed tomography (CT) images with DeepLabV3 + networks for two-class (background and lung field) and four-class (ground-glass opacities, background, consolidation, and lung field). In this work, we investigate the performance of the DeepLabV3 + network with five pretrained networks: Xception, ResNet-18, Inception-ResNet-v2, MobileNet-v2 and ResNet-50. A publicly available database for COVID-19 that contains 750 chest CT images and corresponding pixel-labeled images are used to develop the deep learning model. The segmentation performance has been assessed using five performance measures: Intersection of Union (IoU), Weighted IoU, Balance F1 score, pixel accu-racy, and global accuracy. The experimental results of this work confirm that the DeepLabV3 + network with ResNet-18 and a batch size of 8 have a higher performance for two-class segmentation. DeepLabV3 + network coupled with ResNet-50 and a batch size of 16 yielded better results for four-class segmentation compared to other pretrained networks. Besides, the ResNet with a fewer number of layers is highly adequate for developing a more robust lung segmentation network with lesser computational complexity compared to the conventional DeepLabV3 + network with Xception. This present work proposes a unified DeepLabV3 + network to delineate the two and four different regions automatically using CT images for CoVID-19 patients. Our developed automated segmented model can be further developed to be used as a clinical diagnosis system for CoVID-19 as well as assist clinicians in providing an accurate second opinion CoVID-19 diagnosis.
{"title":"Automated semantic lung segmentation in chest CT images using deep neural network.","authors":"M Murugappan, Ali K Bourisly, N B Prakash, M G Sumithra, U Rajendra Acharya","doi":"10.1007/s00521-023-08407-1","DOIUrl":"10.1007/s00521-023-08407-1","url":null,"abstract":"<p><p>Lung segmentation algorithms play a significant role in segmenting theinfected regions in the lungs. This work aims to develop a computationally efficient and robust deep learning model for lung segmentation using chest computed tomography (CT) images with DeepLabV3 + networks for two-class (background and lung field) and four-class (ground-glass opacities, background, consolidation, and lung field). In this work, we investigate the performance of the DeepLabV3 + network with five pretrained networks: Xception, ResNet-18, Inception-ResNet-v2, MobileNet-v2 and ResNet-50. A publicly available database for COVID-19 that contains 750 chest CT images and corresponding pixel-labeled images are used to develop the deep learning model. The segmentation performance has been assessed using five performance measures: Intersection of Union (IoU), Weighted IoU, Balance F1 score, pixel accu-racy, and global accuracy. The experimental results of this work confirm that the DeepLabV3 + network with ResNet-18 and a batch size of 8 have a higher performance for two-class segmentation. DeepLabV3 + network coupled with ResNet-50 and a batch size of 16 yielded better results for four-class segmentation compared to other pretrained networks. Besides, the ResNet with a fewer number of layers is highly adequate for developing a more robust lung segmentation network with lesser computational complexity compared to the conventional DeepLabV3 + network with Xception. This present work proposes a unified DeepLabV3 + network to delineate the two and four different regions automatically using CT images for CoVID-19 patients. Our developed automated segmented model can be further developed to be used as a clinical diagnosis system for CoVID-19 as well as assist clinicians in providing an accurate second opinion CoVID-19 diagnosis.</p>","PeriodicalId":49766,"journal":{"name":"Neural Computing & Applications","volume":"35 21","pages":"15343-15364"},"PeriodicalIF":4.5,"publicationDate":"2023-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10088735/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9570242","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-01-01Epub Date: 2023-04-07DOI: 10.1007/s00521-023-08556-3
Mehmet Şirin Gündüz, Gültekin Işık
The coronavirus disease (COVID-19) is primarily disseminated through physical contact. As a precaution, it is recommended that indoor spaces have a limited number of people and at least one meter apart. This study proposes a real-time method for monitoring physical distancing compliance in indoor spaces using computer vision and deep learning techniques. The proposed method utilizes YOLO (You Only Look Once), a popular convolutional neural network-based object detection model, pre-trained on the Microsoft COCO (Common Objects in Context) dataset to detect persons and estimate their physical distance in real time. The effectiveness of the proposed method was assessed using metrics including accuracy rate, frame per second (FPS), and mean average precision (mAP). The results show that the YOLO v3 model had the most remarkable accuracy (87.07%) and mAP (89.91%). On the other hand, the highest fps rate of up to 18.71 was achieved by the YOLO v5s model. The results demonstrate the potential of the proposed method for effectively monitoring physical distancing compliance in indoor spaces, providing valuable insights for future use in other public health scenarios.
冠状病毒疾病(新冠肺炎)主要通过身体接触传播。作为预防措施,建议室内空间的人数有限,且至少间隔一米。这项研究提出了一种使用计算机视觉和深度学习技术实时监测室内空间物理距离遵守情况的方法。所提出的方法利用YOLO(You Only Look Once),一种流行的基于卷积神经网络的对象检测模型,在Microsoft COCO(Common Objects in Context)数据集上预先训练,实时检测人员并估计他们的物理距离。使用包括准确率、每秒帧数(FPS)和平均精度(mAP)在内的指标来评估所提出方法的有效性。结果表明,YOLO v3模型具有最显著的准确率(87.07%)和mAP(89.91%)。另一方面,YOLO v5s模型获得了高达18.71的最高fps速率。结果证明了所提出的方法在有效监测室内空间物理距离遵守情况方面的潜力,为未来在其他公共卫生场景中的使用提供了有价值的见解。
{"title":"A new YOLO-based method for social distancing from real-time videos.","authors":"Mehmet Şirin Gündüz, Gültekin Işık","doi":"10.1007/s00521-023-08556-3","DOIUrl":"10.1007/s00521-023-08556-3","url":null,"abstract":"<p><p>The coronavirus disease (COVID-19) is primarily disseminated through physical contact. As a precaution, it is recommended that indoor spaces have a limited number of people and at least one meter apart. This study proposes a real-time method for monitoring physical distancing compliance in indoor spaces using computer vision and deep learning techniques. The proposed method utilizes YOLO (You Only Look Once), a popular convolutional neural network-based object detection model, pre-trained on the Microsoft COCO (Common Objects in Context) dataset to detect persons and estimate their physical distance in real time. The effectiveness of the proposed method was assessed using metrics including accuracy rate, frame per second (FPS), and mean average precision (mAP). The results show that the YOLO v3 model had the most remarkable accuracy (87.07%) and mAP (89.91%). On the other hand, the highest fps rate of up to 18.71 was achieved by the YOLO v5s model. The results demonstrate the potential of the proposed method for effectively monitoring physical distancing compliance in indoor spaces, providing valuable insights for future use in other public health scenarios.</p>","PeriodicalId":49766,"journal":{"name":"Neural Computing & Applications","volume":"35 21","pages":"15261-15271"},"PeriodicalIF":6.0,"publicationDate":"2023-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10081816/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9576902","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-01-01DOI: 10.1007/s00521-020-05592-1
Simin Li, Yulan Lin, Tong Zhu, Mengjie Fan, Shicheng Xu, Weihao Qiu, Can Chen, Linfeng Li, Yao Wang, Jun Yan, Justin Wong, Lin Naing, Shabei Xu
To predict the mortality of patients with coronavirus disease 2019 (COVID-19). We collected clinical data of COVID-19 patients between January 18 and March 29 2020 in Wuhan, China . Gradient boosting decision tree (GBDT), logistic regression (LR) model, and simplified LR were built to predict the mortality of COVID-19. We also evaluated different models by computing area under curve (AUC), accuracy, positive predictive value (PPV), and negative predictive value (NPV) under fivefold cross-validation. A total of 2924 patients were included in our evaluation, with 257 (8.8%) died and 2667 (91.2%) survived during hospitalization. Upon admission, there were 21 (0.7%) mild cases, 2051 (70.1%) moderate case, 779 (26.6%) severe cases, and 73 (2.5%) critically severe cases. The GBDT model exhibited the highest fivefold AUC, which was 0.941, followed by LR (0.928) and LR-5 (0.913). The diagnostic accuracies of GBDT, LR, and LR-5 were 0.889, 0.868, and 0.887, respectively. In particular, the GBDT model demonstrated the highest sensitivity (0.899) and specificity (0.889). The NPV of all three models exceeded 97%, while their PPV values were relatively low, resulting in 0.381 for LR, 0.402 for LR-5, and 0.432 for GBDT. Regarding severe and critically severe cases, the GBDT model also performed the best with a fivefold AUC of 0.918. In the external validation test of the LR-5 model using 72 cases of COVID-19 from Brunei, leukomonocyte (%) turned to show the highest fivefold AUC (0.917), followed by urea (0.867), age (0.826), and SPO2 (0.704). The findings confirm that the mortality prediction performance of the GBDT is better than the LR models in confirmed cases of COVID-19. The performance comparison seems independent of disease severity.
Supplementary information: The online version contains supplementary material available at(10.1007/s00521-020-05592-1).
{"title":"Development and external evaluation of predictions models for mortality of COVID-19 patients using machine learning method.","authors":"Simin Li, Yulan Lin, Tong Zhu, Mengjie Fan, Shicheng Xu, Weihao Qiu, Can Chen, Linfeng Li, Yao Wang, Jun Yan, Justin Wong, Lin Naing, Shabei Xu","doi":"10.1007/s00521-020-05592-1","DOIUrl":"https://doi.org/10.1007/s00521-020-05592-1","url":null,"abstract":"<p><p>To predict the mortality of patients with coronavirus disease 2019 (COVID-19). We collected clinical data of COVID-19 patients between January 18 and March 29 2020 in Wuhan, China . Gradient boosting decision tree (GBDT), logistic regression (LR) model, and simplified LR were built to predict the mortality of COVID-19. We also evaluated different models by computing area under curve (AUC), accuracy, positive predictive value (PPV), and negative predictive value (NPV) under fivefold cross-validation. A total of 2924 patients were included in our evaluation, with 257 (8.8%) died and 2667 (91.2%) survived during hospitalization. Upon admission, there were 21 (0.7%) mild cases, 2051 (70.1%) moderate case, 779 (26.6%) severe cases, and 73 (2.5%) critically severe cases. The GBDT model exhibited the highest fivefold AUC, which was 0.941, followed by LR (0.928) and LR-5 (0.913). The diagnostic accuracies of GBDT, LR, and LR-5 were 0.889, 0.868, and 0.887, respectively. In particular, the GBDT model demonstrated the highest sensitivity (0.899) and specificity (0.889). The NPV of all three models exceeded 97%, while their PPV values were relatively low, resulting in 0.381 for LR, 0.402 for LR-5, and 0.432 for GBDT. Regarding severe and critically severe cases, the GBDT model also performed the best with a fivefold AUC of 0.918. In the external validation test of the LR-5 model using 72 cases of COVID-19 from Brunei, leukomonocyte (%) turned to show the highest fivefold AUC (0.917), followed by urea (0.867), age (0.826), and SPO2 (0.704). The findings confirm that the mortality prediction performance of the GBDT is better than the LR models in confirmed cases of COVID-19. The performance comparison seems independent of disease severity.</p><p><strong>Supplementary information: </strong>The online version contains supplementary material available at(10.1007/s00521-020-05592-1).</p>","PeriodicalId":49766,"journal":{"name":"Neural Computing & Applications","volume":"35 18","pages":"13037-13046"},"PeriodicalIF":6.0,"publicationDate":"2023-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1007/s00521-020-05592-1","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9853876","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-01-01DOI: 10.1007/s00521-022-07739-8
Soodeh Nasrollah, S Esmaeil Najafi, Hadi Bagherzadeh, Mohsen Rostamy-Malkhalifeh
In recent years, the hyper-competitive marketplace has led to a drastic enhancement in the importance of the supply chain problem. Hence, the attention of managers and researchers has been attracted to one of the most crucial problems in the supply chain management area called the supply chain network design problem. In this regard, this research attempts to design an integrated forward and backward logistics network by proposing a multi-objective mathematical model. The suggested model aims at minimizing the environmental impacts and the costs while maximizing the resilience and responsiveness of the supply chain. Since uncertainty is a major issue in the supply chain problem, the present paper studies the research problem under the mixed uncertainty and utilizes the robust possibilistic stochastic method to cope with the uncertainty. On the other side, since configuring a supply chain is known as an NP-Hard problem, this research develops an enhanced particle swarm optimization algorithm to obtain optimal/near-optimal solutions in a reasonable time. Based on the achieved results, the developed algorithm can obtain high-quality solutions (i.e. solutions with zero or a very small gap from the optimal solution) in a reasonable amount of time. The achieved results demonstrate the negative impact of the enhancement of the demand on environmental damages and the total cost. Also, according to the outputs, by increasing the service level, the total cost and environmental impacts have increased by 41% and 10%, respectively. On the other hand, the results show that increasing the disrupted capacity parameters has led to a 17% increase in the total costs and a 7% increase in carbon emissions.
Supplementary information: The online version contains supplementary material available at 10.1007/s00521-022-07739-8.
{"title":"An enhanced PSO algorithm to configure a responsive-resilient supply chain network considering environmental issues: a case study of the oxygen concentrator device.","authors":"Soodeh Nasrollah, S Esmaeil Najafi, Hadi Bagherzadeh, Mohsen Rostamy-Malkhalifeh","doi":"10.1007/s00521-022-07739-8","DOIUrl":"https://doi.org/10.1007/s00521-022-07739-8","url":null,"abstract":"<p><p>In recent years, the hyper-competitive marketplace has led to a drastic enhancement in the importance of the supply chain problem. Hence, the attention of managers and researchers has been attracted to one of the most crucial problems in the supply chain management area called the supply chain network design problem. In this regard, this research attempts to design an integrated forward and backward logistics network by proposing a multi-objective mathematical model. The suggested model aims at minimizing the environmental impacts and the costs while maximizing the resilience and responsiveness of the supply chain. Since uncertainty is a major issue in the supply chain problem, the present paper studies the research problem under the mixed uncertainty and utilizes the robust possibilistic stochastic method to cope with the uncertainty. On the other side, since configuring a supply chain is known as an NP-Hard problem, this research develops an enhanced particle swarm optimization algorithm to obtain optimal/near-optimal solutions in a reasonable time. Based on the achieved results, the developed algorithm can obtain high-quality solutions (i.e. solutions with zero or a very small gap from the optimal solution) in a reasonable amount of time. The achieved results demonstrate the negative impact of the enhancement of the demand on environmental damages and the total cost. Also, according to the outputs, by increasing the service level, the total cost and environmental impacts have increased by 41% and 10%, respectively. On the other hand, the results show that increasing the disrupted capacity parameters has led to a 17% increase in the total costs and a 7% increase in carbon emissions.</p><p><strong>Supplementary information: </strong>The online version contains supplementary material available at 10.1007/s00521-022-07739-8.</p>","PeriodicalId":49766,"journal":{"name":"Neural Computing & Applications","volume":"35 3","pages":"2647-2678"},"PeriodicalIF":6.0,"publicationDate":"2023-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9440659/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"10538298","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-01-01DOI: 10.1007/s00521-022-07917-8
Sifan Peng, Baoqun Yin, Qianqian Yang, Qing He, Luyang Wang
Crowd counting has received increasing attention due to its important roles in multiple fields, such as social security, commercial applications, epidemic prevention and control. To this end, we explore two critical issues that seriously affect the performance of crowd counting including nonuniform crowd density distribution and cross-domain problems. Aiming at the nonuniform crowd density distribution issue, we propose a density rectifying network (DRNet) that consists of several dual-layer pyramid fusion modules (DPFM) and a density rectification map (DRmap) auxiliary learning module. The proposed DPFM is embedded into DRNet to integrate multi-scale crowd density features through dual-layer pyramid fusion. The devised DRmap auxiliary learning module further rectifies the incorrect crowd density estimation by adaptively weighting the initial crowd density maps. With respect to the cross-domain issue, we develop a domain adaptation method of randomly cutting mixed dual-domain images, which learns domain-invariance features and decreases the domain gap between the source domain and the target domain from global and local perspectives. Experimental results indicate that the devised DRNet achieves the best mean absolute error (MAE) and competitive mean squared error (MSE) compared with other excellent methods on four benchmark datasets. Additionally, a series of cross-domain experiments are conducted to demonstrate the effectiveness of the proposed domain adaption method. Significantly, when the A and B parts of the Shanghaitech dataset are the source domain and target domain respectively, the proposed domain adaption method decreases the MAE of DRNet by .
{"title":"Exploring density rectification and domain adaption method for crowd counting.","authors":"Sifan Peng, Baoqun Yin, Qianqian Yang, Qing He, Luyang Wang","doi":"10.1007/s00521-022-07917-8","DOIUrl":"https://doi.org/10.1007/s00521-022-07917-8","url":null,"abstract":"<p><p>Crowd counting has received increasing attention due to its important roles in multiple fields, such as social security, commercial applications, epidemic prevention and control. To this end, we explore two critical issues that seriously affect the performance of crowd counting including nonuniform crowd density distribution and cross-domain problems. Aiming at the nonuniform crowd density distribution issue, we propose a density rectifying network (DRNet) that consists of several dual-layer pyramid fusion modules (DPFM) and a density rectification map (DRmap) auxiliary learning module. The proposed DPFM is embedded into DRNet to integrate multi-scale crowd density features through dual-layer pyramid fusion. The devised DRmap auxiliary learning module further rectifies the incorrect crowd density estimation by adaptively weighting the initial crowd density maps. With respect to the cross-domain issue, we develop a domain adaptation method of randomly cutting mixed dual-domain images, which learns domain-invariance features and decreases the domain gap between the source domain and the target domain from global and local perspectives. Experimental results indicate that the devised DRNet achieves the best mean absolute error (MAE) and competitive mean squared error (MSE) compared with other excellent methods on four benchmark datasets. Additionally, a series of cross-domain experiments are conducted to demonstrate the effectiveness of the proposed domain adaption method. Significantly, when the A and B parts of the Shanghaitech dataset are the source domain and target domain respectively, the proposed domain adaption method decreases the MAE of DRNet by <math><mrow><mn>47.6</mn> <mo>%</mo></mrow> </math> .</p>","PeriodicalId":49766,"journal":{"name":"Neural Computing & Applications","volume":"35 4","pages":"3551-3569"},"PeriodicalIF":6.0,"publicationDate":"2023-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9568950/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"10638062","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-01-01Epub Date: 2023-01-28DOI: 10.1007/s00521-023-08285-7
Nora Alturayeif, Hamzah Luqman, Moataz Ahmed
Stance detection is an evolving opinion mining research area motivated by the vast increase in the variety and volume of user-generated content. In this regard, considerable research has been recently carried out in the area of stance detection. In this study, we review the different techniques proposed in the literature for stance detection as well as other applications such as rumor veracity detection. Particularly, we conducted a systematic literature review of empirical research on the machine learning (ML) models for stance detection that were published from January 2015 to October 2022. We analyzed 96 primary studies, which spanned eight categories of ML techniques. In this paper, we categorize the analyzed studies according to a taxonomy of six dimensions: approaches, target dependency, applications, modeling, language, and resources. We further classify and analyze the corresponding techniques from each dimension's perspective and highlight their strengths and weaknesses. The analysis reveals that deep learning models that adopt a mechanism of self-attention have been used more frequently than the other approaches. It is worth noting that emerging ML techniques such as few-shot learning and multitask learning have been used extensively for stance detection. A major conclusion of our analysis is that despite that ML models have shown to be promising in this field, the application of these models in the real world is still limited. Our analysis lists challenges and gaps to be addressed in future research. Furthermore, the taxonomy presented can assist researchers in developing and positioning new techniques for stance detection-related applications.
立场检测是一个不断发展的意见挖掘研究领域,其动因是用户生成内容的种类和数量大幅增加。在这方面,最近在立场检测领域开展了大量研究。在本研究中,我们回顾了文献中提出的用于立场检测的不同技术以及谣言真实性检测等其他应用。特别是,我们对 2015 年 1 月至 2022 年 10 月期间发表的有关立场检测的机器学习(ML)模型的实证研究进行了系统的文献综述。我们分析了 96 项主要研究,涉及八类 ML 技术。在本文中,我们根据六个维度的分类法对所分析的研究进行了分类:方法、目标依赖、应用、建模、语言和资源。我们从每个维度的角度对相应的技术进行了进一步的分类和分析,并强调了它们的优缺点。分析结果显示,采用自我关注机制的深度学习模型比其他方法使用得更频繁。值得注意的是,一些新兴的 ML 技术,如少量学习(few-shot learning)和多任务学习(multitask learning),已被广泛用于姿态检测。我们分析得出的一个主要结论是,尽管 ML 模型在这一领域大有可为,但这些模型在现实世界中的应用仍然有限。我们的分析列出了未来研究中需要应对的挑战和差距。此外,我们提出的分类法可以帮助研究人员为姿态检测相关应用开发和定位新技术。
{"title":"A systematic review of machine learning techniques for stance detection and its applications.","authors":"Nora Alturayeif, Hamzah Luqman, Moataz Ahmed","doi":"10.1007/s00521-023-08285-7","DOIUrl":"10.1007/s00521-023-08285-7","url":null,"abstract":"<p><p>Stance detection is an evolving opinion mining research area motivated by the vast increase in the variety and volume of user-generated content. In this regard, considerable research has been recently carried out in the area of stance detection. In this study, we review the different techniques proposed in the literature for stance detection as well as other applications such as rumor veracity detection. Particularly, we conducted a systematic literature review of empirical research on the machine learning (ML) models for stance detection that were published from January 2015 to October 2022. We analyzed 96 primary studies, which spanned eight categories of ML techniques. In this paper, we categorize the analyzed studies according to a taxonomy of six dimensions: approaches, target dependency, applications, modeling, language, and resources. We further classify and analyze the corresponding techniques from each dimension's perspective and highlight their strengths and weaknesses. The analysis reveals that deep learning models that adopt a mechanism of self-attention have been used more frequently than the other approaches. It is worth noting that emerging ML techniques such as few-shot learning and multitask learning have been used extensively for stance detection. A major conclusion of our analysis is that despite that ML models have shown to be promising in this field, the application of these models in the real world is still limited. Our analysis lists challenges and gaps to be addressed in future research. Furthermore, the taxonomy presented can assist researchers in developing and positioning new techniques for stance detection-related applications.</p>","PeriodicalId":49766,"journal":{"name":"Neural Computing & Applications","volume":"35 7","pages":"5113-5144"},"PeriodicalIF":4.5,"publicationDate":"2023-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9884072/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"10642878","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-01-01DOI: 10.1007/s00521-022-07435-7
Nagore Sagastibeltza, Asier Salazar-Ramirez, Raquel Martinez, Jose Luis Jodra, Javier Muguerza
Nowadays, considering society's highly demanding lifestyles, it is important to consider the usefulness of relaxation from the perspective of both psychology and clinical practice. The response towards relaxation (RResp) is a mind-body interaction that relaxes the organism or compensates for the physiological effects caused by stress. This work aims to automatically detect the different mental states (relaxation, rest and stress) in which RResps may occur so that complete feedback about the quality of the relaxation can be given to the subject itself, the psychologist or the doctor. To this end, an experiment was conducted to induce both states of stress and relaxation in a sample of 20 university students (average age of years old). The electrocardiographic and electrodermal activity signals collected from the participants produced a dataset with 1641 episodes or instances in which the previously mentioned mental states take place. This data was used to extract up to 50 features and train several supervised learning algorithms (rule-based, trees, probabilistic, ensemble classifiers, etc.) using and not using feature selection techniques. Besides, the authors synthesised the cardiac activity information into a single new feature and discretised it down to three levels. The experimentation revealed which features were most discriminating, reaching a classification average accuracy of up to % with the 6 most relevant features for the own-collected dataset. Finally, being restrictive, the same solution/subspace was tested with a dataset referenced in the bibliography (WESAD) and scored an average accuracy of %.
{"title":"Automatic detection of the mental state in responses towards relaxation.","authors":"Nagore Sagastibeltza, Asier Salazar-Ramirez, Raquel Martinez, Jose Luis Jodra, Javier Muguerza","doi":"10.1007/s00521-022-07435-7","DOIUrl":"https://doi.org/10.1007/s00521-022-07435-7","url":null,"abstract":"<p><p>Nowadays, considering society's highly demanding lifestyles, it is important to consider the usefulness of relaxation from the perspective of both psychology and clinical practice. The response towards relaxation (RResp) is a mind-body interaction that relaxes the organism or compensates for the physiological effects caused by stress. This work aims to automatically detect the different mental states (relaxation, rest and stress) in which RResps may occur so that complete feedback about the quality of the relaxation can be given to the subject itself, the psychologist or the doctor. To this end, an experiment was conducted to induce both states of stress and relaxation in a sample of 20 university students (average age of <math><mrow><mn>25.76</mn> <mo>±</mo> <mn>3.7</mn></mrow> </math> years old). The electrocardiographic and electrodermal activity signals collected from the participants produced a dataset with 1641 episodes or instances in which the previously mentioned mental states take place. This data was used to extract up to 50 features and train several supervised learning algorithms (rule-based, trees, probabilistic, ensemble classifiers, etc.) using and not using feature selection techniques. Besides, the authors synthesised the cardiac activity information into a single new feature and discretised it down to three levels. The experimentation revealed which features were most discriminating, reaching a classification average accuracy of up to <math><mrow><mn>94.01</mn> <mo>±</mo> <mn>1.73</mn></mrow> </math> % with the 6 most relevant features for the own-collected dataset. Finally, being restrictive, the same solution/subspace was tested with a dataset referenced in the bibliography (WESAD) and scored an average accuracy of <math><mrow><mn>90.36</mn> <mo>±</mo> <mn>1.62</mn></mrow> </math> %.</p>","PeriodicalId":49766,"journal":{"name":"Neural Computing & Applications","volume":"35 8","pages":"5679-5696"},"PeriodicalIF":6.0,"publicationDate":"2023-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9178946/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"10868777","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}