Pub Date : 2024-06-19DOI: 10.1007/s10586-024-04610-4
Liji Luo, Siwei Wei, Hua Tang, Chunzhi Wang
As the scale of data centers continues to expand, optimizing resource utilization becomes increasingly critical. Employing virtual machine (VM) migration technology to maintain hosts within an appropriate workload range holds substantial promise for enhancing platform resource utilization, workload equilibrium, and energy efficiency. This study endeavors to reframe virtual machine migration as a partition problem and introduces an integrated framework that adeptly evaluate workload status and precisely identifies the optimal migration target, thus mitigating the expenses associated with virtual machine migration. Our framework commences by employing workload prediction to evaluate host status for determining the most opportune timing for migration. Subsequently, we leverage Service Level Agreements (SLA) violation as the optimization objective to ascertain the optimal status threshold, thereby facilitating effective workload partition of the host. Finally, the framework employs multi-dimensional host resource balance as a guide to schedule host migration in diverse areas, ensuring robust resource utilization post-migration. Experimental results show that compared with three benchmark VM allocation algorithms, SESA, PPRG, and ThrRs. Our framework achieves a significant (17%) increase in multidimensional resource utilization across various types of data centers, accompanied by a noteworthy (27%) reduction in SLA violation rate with fewer time consumption and energy expenditure during VM migration.
{"title":"An effective partition-based framework for virtual machine migration in cloud services","authors":"Liji Luo, Siwei Wei, Hua Tang, Chunzhi Wang","doi":"10.1007/s10586-024-04610-4","DOIUrl":"https://doi.org/10.1007/s10586-024-04610-4","url":null,"abstract":"<p>As the scale of data centers continues to expand, optimizing resource utilization becomes increasingly critical. Employing virtual machine (VM) migration technology to maintain hosts within an appropriate workload range holds substantial promise for enhancing platform resource utilization, workload equilibrium, and energy efficiency. This study endeavors to reframe virtual machine migration as a partition problem and introduces an integrated framework that adeptly evaluate workload status and precisely identifies the optimal migration target, thus mitigating the expenses associated with virtual machine migration. Our framework commences by employing workload prediction to evaluate host status for determining the most opportune timing for migration. Subsequently, we leverage Service Level Agreements (SLA) violation as the optimization objective to ascertain the optimal status threshold, thereby facilitating effective workload partition of the host. Finally, the framework employs multi-dimensional host resource balance as a guide to schedule host migration in diverse areas, ensuring robust resource utilization post-migration. Experimental results show that compared with three benchmark VM allocation algorithms, SESA, PPRG, and ThrRs. Our framework achieves a significant <span>(17%)</span> increase in multidimensional resource utilization across various types of data centers, accompanied by a noteworthy <span>(27%)</span> reduction in SLA violation rate with fewer time consumption and energy expenditure during VM migration.</p>","PeriodicalId":501576,"journal":{"name":"Cluster Computing","volume":"20 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-06-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141531733","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-06-18DOI: 10.1007/s10586-024-04551-y
Mohammad Yekta, Hadi Shahriar Shahhoseini
The extensive deployment of large cloud data centers has led to substantial energy consumption. Energy conservation is a critical concern for cloud service providers seeking to lower their operating costs within their data centers. To address this energy consumption challenge, effective approaches such as VM consolidation and VM migration are essential. These approaches must carefully balance the trade-off between energy consumption and Service Level Agreement Violations (SLAV). In this paper, we propose an energy-efficient VM selection algorithm for VM consolidation and call it the Simultaneous CPU–Ram Utilization Balancer (SCRUB) policy. This algorithm takes into account CPU and RAM utilization while trying to maintain a balance between energy consumption and SLAV. To evaluate the performance of our proposed method, we implemented it using the Cloudsim simulation toolkit and conducted simulations using real-world workload traces from PlanetLab and Google over three different days. The results show that the SCRUB VM selection policy has led to improvements in various metrics, including reduced energy consumption and a decreased number of VM migrations compared to existing VM selection policies. Specifically, it achieved a 16.98% decrease in energy consumption and a 46.42% reduction in the number of migrations for the PlanetLab dataset, and a 10.95% decrease in energy consumption and a 43.96% decline in the number of migrations for the Google dataset compared to the baseline algorithm MMT.
大型云数据中心的广泛部署导致了大量能源消耗。对于希望降低数据中心运营成本的云服务提供商来说,节能是一个至关重要的问题。要应对这一能耗挑战,虚拟机整合和虚拟机迁移等有效方法至关重要。这些方法必须在能耗和违反服务级别协议(SLAV)之间谨慎权衡。在本文中,我们提出了一种用于虚拟机整合的高能效虚拟机选择算法,并将其称为 "CPU-内存同时利用平衡器(SCRUB)策略"。该算法考虑了 CPU 和 RAM 的利用率,同时努力保持能耗和 SLAV 之间的平衡。为了评估我们提出的方法的性能,我们使用 Cloudsim 仿真工具包实施了该方法,并使用 PlanetLab 和 Google 在三个不同日期的真实工作负载跟踪进行了仿真。结果表明,与现有的虚拟机选择策略相比,SCRUB 虚拟机选择策略改善了各种指标,包括降低能耗和减少虚拟机迁移次数。具体来说,与基准算法MMT相比,SCRUB在PlanetLab数据集上实现了16.98%的能耗降低和46.42%的迁移次数减少,在Google数据集上实现了10.95%的能耗降低和43.96%的迁移次数减少。
{"title":"SCRUB: a novel energy-efficient virtual machines selection and migration scheme in cloud data centers","authors":"Mohammad Yekta, Hadi Shahriar Shahhoseini","doi":"10.1007/s10586-024-04551-y","DOIUrl":"https://doi.org/10.1007/s10586-024-04551-y","url":null,"abstract":"<p>The extensive deployment of large cloud data centers has led to substantial energy consumption. Energy conservation is a critical concern for cloud service providers seeking to lower their operating costs within their data centers. To address this energy consumption challenge, effective approaches such as VM consolidation and VM migration are essential. These approaches must carefully balance the trade-off between energy consumption and Service Level Agreement Violations (SLAV). In this paper, we propose an energy-efficient VM selection algorithm for VM consolidation and call it the Simultaneous CPU–Ram Utilization Balancer (SCRUB) policy. This algorithm takes into account CPU and RAM utilization while trying to maintain a balance between energy consumption and SLAV. To evaluate the performance of our proposed method, we implemented it using the Cloudsim simulation toolkit and conducted simulations using real-world workload traces from PlanetLab and Google over three different days. The results show that the SCRUB VM selection policy has led to improvements in various metrics, including reduced energy consumption and a decreased number of VM migrations compared to existing VM selection policies. Specifically, it achieved a 16.98% decrease in energy consumption and a 46.42% reduction in the number of migrations for the PlanetLab dataset, and a 10.95% decrease in energy consumption and a 43.96% decline in the number of migrations for the Google dataset compared to the baseline algorithm MMT.</p>","PeriodicalId":501576,"journal":{"name":"Cluster Computing","volume":"26 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-06-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141522112","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
In this paper, aiming at the problem that the balance between exploration and exploitation of traditional Tasmanian devil optimization algorithm is unflexible, and easy to fall into local optimum, an improved Tasmanian devil optimization algorithm (NTDO) based on the sine-cosine strategy of dynamic weighted factors is proposed. The designed method balances the global and local search capabilities of the algorithm, effectively improves the situation that the algorithm falls into local optimum, and integrally improves the optimization performance of the algorithm. Firstly, the good point set theory is used instead of the traditional random method to find the initial individuals, which can render the initial population is more evenly distributed in the search space and the population diversity is improved. Secondly, A sine-cosine strategy based on dynamic weighted factors is proposed to coordinate the global exploration and local optimization capabilities of the algorithm, and enhance the convergence accuracy of the algorithm. Thirdly, since Tasmanian devil is easy to fall into local optimum in the process of hunting prey, a nonlinear decline strategy based on oscillation factor is presented, which increases the search range of the algorithm and improves the ability of the algorithm to jump out of the local optimal value.Finally, 12 evaluation functions, cec2019 and cec2021 test functions commonly used in NTDO and TDO, WOA, DBO, PSO, GWO, DFPSO and PDGWO algorithms are compared and analyzed, and the experimental results show the effectiveness and feasibility of the scheme.
{"title":"An improved Tasmanian devil optimization algorithm based on sine-cosine strategy with dynamic weighting factors","authors":"Huanlong Zhang, Chenglin Guo, Jianwei Zhang, Xin Wang, Jiaxiang Zhang","doi":"10.1007/s10586-024-04443-1","DOIUrl":"https://doi.org/10.1007/s10586-024-04443-1","url":null,"abstract":"<p>In this paper, aiming at the problem that the balance between exploration and exploitation of traditional Tasmanian devil optimization algorithm is unflexible, and easy to fall into local optimum, an improved Tasmanian devil optimization algorithm (NTDO) based on the sine-cosine strategy of dynamic weighted factors is proposed. The designed method balances the global and local search capabilities of the algorithm, effectively improves the situation that the algorithm falls into local optimum, and integrally improves the optimization performance of the algorithm. Firstly, the good point set theory is used instead of the traditional random method to find the initial individuals, which can render the initial population is more evenly distributed in the search space and the population diversity is improved. Secondly, A sine-cosine strategy based on dynamic weighted factors is proposed to coordinate the global exploration and local optimization capabilities of the algorithm, and enhance the convergence accuracy of the algorithm. Thirdly, since Tasmanian devil is easy to fall into local optimum in the process of hunting prey, a nonlinear decline strategy based on oscillation factor is presented, which increases the search range of the algorithm and improves the ability of the algorithm to jump out of the local optimal value.Finally, 12 evaluation functions, cec2019 and cec2021 test functions commonly used in NTDO and TDO, WOA, DBO, PSO, GWO, DFPSO and PDGWO algorithms are compared and analyzed, and the experimental results show the effectiveness and feasibility of the scheme.</p>","PeriodicalId":501576,"journal":{"name":"Cluster Computing","volume":"67 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-06-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141522107","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
In the modern, ever-shifting global agri-food environment, the topmost concern revolves around securing the safety, quality, and authenticity of agri-food products. Blockchain technology is being seen as a revolutionary solution for dealing with these issues, providing a decentralized and transparent ledger for the tracking of agri-food products. By incorporating global positioning system and navigation systems within blockchain-based traceability solutions amplifies the capabilities for real-time monitoring, security, and trust. This paper proposes a layered architecture for an efficient agri-food traceability system. The data layer, manages interactions between various entities in the supply chain management and generates agri-food product-related data. The blockchain layer manages data via transactions and smart contracts, using the interplanetary file system for secure, decentralized storage. The navigation layer, combines navigation with Indian constellation and global positioning system to offer precise real-time positioning and timing services, enhancing product tracking. This integrated approach not only improves food safety but also supports sustainability efforts by reducing food waste and bolstering consumer trust in the agri-food industry. We implement the proposed system using Remix IDE, MetaMask wallet, and the Sepolia test network, summarizing the deployment analysis. Performance evaluation is conducted using the JMeter simulation toolkit.The proposed framework achieves an average throughput of 329.26 transactions per second, latency of 49.3 ms, and response time of 87.9 ms. We conduct a comparative evaluation of the proposed system with related studies. From this comparative analysis, we observed that our proposed framework has better features than other related works.
{"title":"Revolutionizing agri-food supply chain management with blockchain-based traceability and navigation integration","authors":"Manoj Aggarwal, Pritam Rani, Prity Rani, Pratima Sharma","doi":"10.1007/s10586-024-04609-x","DOIUrl":"https://doi.org/10.1007/s10586-024-04609-x","url":null,"abstract":"<p>In the modern, ever-shifting global agri-food environment, the topmost concern revolves around securing the safety, quality, and authenticity of agri-food products. Blockchain technology is being seen as a revolutionary solution for dealing with these issues, providing a decentralized and transparent ledger for the tracking of agri-food products. By incorporating global positioning system and navigation systems within blockchain-based traceability solutions amplifies the capabilities for real-time monitoring, security, and trust. This paper proposes a layered architecture for an efficient agri-food traceability system. The data layer, manages interactions between various entities in the supply chain management and generates agri-food product-related data. The blockchain layer manages data via transactions and smart contracts, using the interplanetary file system for secure, decentralized storage. The navigation layer, combines navigation with Indian constellation and global positioning system to offer precise real-time positioning and timing services, enhancing product tracking. This integrated approach not only improves food safety but also supports sustainability efforts by reducing food waste and bolstering consumer trust in the agri-food industry. We implement the proposed system using Remix IDE, MetaMask wallet, and the Sepolia test network, summarizing the deployment analysis. Performance evaluation is conducted using the JMeter simulation toolkit.The proposed framework achieves an average throughput of 329.26 transactions per second, latency of 49.3 ms, and response time of 87.9 ms. We conduct a comparative evaluation of the proposed system with related studies. From this comparative analysis, we observed that our proposed framework has better features than other related works.</p>","PeriodicalId":501576,"journal":{"name":"Cluster Computing","volume":"239 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-06-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141522020","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-06-17DOI: 10.1007/s10586-024-04570-9
R. Shanmugapriya, S. V. N. Santhosh Kumar
Base station (BS) offers data dissemination as a service to IoT smart devices, enabling efficient reprogramming or reconfiguration for their intended activities in post-deployment. Most of the existing IoT data dissemination schemes rely on flooding, leading to the Redundant Broadcast Storm Problem (RBSP), where multiple sensor nodes repeatedly transmit redundant data to neighbours. RBSP elevates network energy consumption and sender congestion in the network. Given that IoT smart devices communicate through open wireless mediums with the internet as a backbone, they are vulnerable to various malicious threats during data dissemination. Intruders may engage in malicious activities and compromise configuration parameters, leading to device failure to execute intended services. This paper proposes a Secure Cloud-Integrated Data Dissemination Protocol (SCIDP) aimed at ensuring the secure dissemination of data within cloud-integrated environments to mitigate RBSP’s impact and enhances security for performing effective reprogramming of sensor devices in IoT. The proposed protocol is implemented by using NS3 simulator with realistic simulation parameters. Simulation results indicate that the proposed protocol enhances energy efficiency by 12%, dissemination effectiveness by 16%, and network lifespan by 16%. Furthermore, the proposed system decreases communication overhead by 11% and computational costs by 9% compared to alternative existing protocols. From the formal security analysis, the proposed system proves that it can withstand against various kinds of security attacks in the network.
{"title":"SCIDP–Secure cloud-integrated data dissemination protocol for efficient reprogramming in internet of things","authors":"R. Shanmugapriya, S. V. N. Santhosh Kumar","doi":"10.1007/s10586-024-04570-9","DOIUrl":"https://doi.org/10.1007/s10586-024-04570-9","url":null,"abstract":"<p>Base station (BS) offers data dissemination as a service to IoT smart devices, enabling efficient reprogramming or reconfiguration for their intended activities in post-deployment. Most of the existing IoT data dissemination schemes rely on flooding, leading to the Redundant Broadcast Storm Problem (RBSP), where multiple sensor nodes repeatedly transmit redundant data to neighbours. RBSP elevates network energy consumption and sender congestion in the network. Given that IoT smart devices communicate through open wireless mediums with the internet as a backbone, they are vulnerable to various malicious threats during data dissemination. Intruders may engage in malicious activities and compromise configuration parameters, leading to device failure to execute intended services. This paper proposes a Secure Cloud-Integrated Data Dissemination Protocol (SCIDP) aimed at ensuring the secure dissemination of data within cloud-integrated environments to mitigate RBSP’s impact and enhances security for performing effective reprogramming of sensor devices in IoT. The proposed protocol is implemented by using NS3 simulator with realistic simulation parameters. Simulation results indicate that the proposed protocol enhances energy efficiency by 12%, dissemination effectiveness by 16%, and network lifespan by 16%. Furthermore, the proposed system decreases communication overhead by 11% and computational costs by 9% compared to alternative existing protocols. From the formal security analysis, the proposed system proves that it can withstand against various kinds of security attacks in the network.</p>","PeriodicalId":501576,"journal":{"name":"Cluster Computing","volume":"857 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141507580","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-06-17DOI: 10.1007/s10586-024-04581-6
Mamatha Maddu, Yamarthi Narasimha Rao
Software-defined networking (SDN) is known for its enhanced network programmability and adaptability, but maintaining strong safety precautions to protect against emerging cyber-attacks remains a constant issue. Since SDN has logically centralized control, an attack on the controller might paralyze the entire network. For this reason, intrusion detection is very crucial. Many academics have embraced state-of-the-art techniques to assess and identify these assaults. However, the majority of these approaches lack scalability and accuracy. Moreover, they had difficulties with restricted features, low efficiency, incorrect characteristics, and computing complexity. Therefore, to detect network vulnerabilities in SDN-based IoT networks, we developed a practical deep learning approach based on Res2Net and Elman Recurrent Neural Networks (ERNN) technique as a defense solution to detect security issues in SDN. This framework consists of multiple steps and starts by addressing the dataset’s class imbalance issue with a Data Augmentation Generative Adversarial Network (DAGAN). Next, the Res2net and Enhanced Honey Badger Algorithm (EHBA) are used to extract and select features. This lowers the computational expense and lessens the possibility that the model would be misled by unsuitable and negative characteristics. Finally, an ERNN-based technique is used to detect and classify the intrusions in SDN. After seeing the network assaults, a practical mitigation framework is implemented to mitigate the network attacks. Three SDN IoT-focused datasets, InSDN, IoT-23 and ToN-IoT, are used in an experimental investigation to analyze the proposed framework’s performance. The results of numerous trials show that the proposed method outperforms existing techniques regarding several constraints.
软件定义网络(SDN)以其增强的网络可编程性和适应性而闻名,但如何保持强大的安全防范措施以抵御新出现的网络攻击仍是一个长期问题。由于 SDN 具有逻辑上的集中控制,对控制器的攻击可能导致整个网络瘫痪。因此,入侵检测至关重要。许多学者已经采用了最先进的技术来评估和识别这些攻击。然而,这些方法大多缺乏可扩展性和准确性。此外,它们还存在功能受限、效率低、特征不正确和计算复杂等问题。因此,为了检测基于 SDN 的物联网网络中的网络漏洞,我们开发了一种基于 Res2Net 和 Elman 循环神经网络(ERNN)技术的实用深度学习方法,作为检测 SDN 中安全问题的防御解决方案。该框架由多个步骤组成,首先使用数据增强生成对抗网络(DAGAN)解决数据集的类不平衡问题。然后,使用 Res2net 和增强型蜜獾算法(EHBA)来提取和选择特征。这不仅降低了计算成本,还减少了模型被不合适的负面特征误导的可能性。最后,基于 ERNN 的技术被用于检测和分类 SDN 中的入侵。在发现网络攻击后,实施了一个实用的缓解框架来缓解网络攻击。在实验调查中使用了三个以 SDN 物联网为重点的数据集:InSDN、IoT-23 和 ToN-IoT,以分析拟议框架的性能。大量试验结果表明,建议的方法在多个限制条件方面优于现有技术。
{"title":"Res2Net-ERNN: deep learning based cyberattack classification in software defined network","authors":"Mamatha Maddu, Yamarthi Narasimha Rao","doi":"10.1007/s10586-024-04581-6","DOIUrl":"https://doi.org/10.1007/s10586-024-04581-6","url":null,"abstract":"<p>Software-defined networking (SDN) is known for its enhanced network programmability and adaptability, but maintaining strong safety precautions to protect against emerging cyber-attacks remains a constant issue. Since SDN has logically centralized control, an attack on the controller might paralyze the entire network. For this reason, intrusion detection is very crucial. Many academics have embraced state-of-the-art techniques to assess and identify these assaults. However, the majority of these approaches lack scalability and accuracy. Moreover, they had difficulties with restricted features, low efficiency, incorrect characteristics, and computing complexity. Therefore, to detect network vulnerabilities in SDN-based IoT networks, we developed a practical deep learning approach based on Res2Net and Elman Recurrent Neural Networks (ERNN) technique as a defense solution to detect security issues in SDN. This framework consists of multiple steps and starts by addressing the dataset’s class imbalance issue with a Data Augmentation Generative Adversarial Network (DAGAN). Next, the Res2net and Enhanced Honey Badger Algorithm (EHBA) are used to extract and select features. This lowers the computational expense and lessens the possibility that the model would be misled by unsuitable and negative characteristics. Finally, an ERNN-based technique is used to detect and classify the intrusions in SDN. After seeing the network assaults, a practical mitigation framework is implemented to mitigate the network attacks. Three SDN IoT-focused datasets, InSDN, IoT-23 and ToN-IoT, are used in an experimental investigation to analyze the proposed framework’s performance. The results of numerous trials show that the proposed method outperforms existing techniques regarding several constraints.</p>","PeriodicalId":501576,"journal":{"name":"Cluster Computing","volume":"25 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141522022","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-06-17DOI: 10.1007/s10586-024-04540-1
Essam H. Houssein, Doaa A. Abdelkareem, Gang Hu, Mohamed Abdel Hameed, Ibrahim A. Ibrahim, Mina Younan
Skin cancer is one of the most dangerous types of cancer due to its immediate appearance and the possibility of rapid spread. It arises from uncontrollably growing cells, rapidly dividing cells in one area of the body, invading other bodily tissues, and spreading throughout the body. Early detection helps prevent cancer progress from reaching critical levels, reducing the risk of complications and the need for more aggressive treatment options. Convolutional neural networks (CNNs) revolutionize skin cancer diagnosis by extracting intricate features from images, enabling an accurate classification of lesions. Their role extends to early detection, providing a powerful tool for dermatologists to identify abnormalities in their nascent stages, ultimately improving patient outcomes. This study proposes a novel deep convolutional neural network (DCNN) approach to classifying skin cancer lesions. The proposed DCNN model is evaluated using two unbalanced datasets, namely HAM10000 and ISIC-2019. The DCNN model is compared with other transfer learning models, including VGG16, VGG19, DenseNet121, DenseNet201, and MobileNetV2. Its performance is assessed using four widely used evaluation metrics: accuracy, recall, precision, F1-score, specificity, and AUC. The experimental results demonstrate that the proposed DCNN model outperforms other deep learning (DL) models that utilized these datasets. The proposed DCNN model achieved the highest accuracy with the HAM10000 and ISIC-2019 datasets, reaching (98.5%) and (97.1%), respectively. These experimental results show how competitive and successful the DCNN model is in overcoming the problems caused by class imbalance and raising skin cancer classification accuracy. Furthermore, the proposed model demonstrates superior performance, particularly excelling in terms of accuracy, compared to other recent studies that utilize the same datasets, which highlights the robustness and effectiveness of the proposed DCNN.
{"title":"An effective multiclass skin cancer classification approach based on deep convolutional neural network","authors":"Essam H. Houssein, Doaa A. Abdelkareem, Gang Hu, Mohamed Abdel Hameed, Ibrahim A. Ibrahim, Mina Younan","doi":"10.1007/s10586-024-04540-1","DOIUrl":"https://doi.org/10.1007/s10586-024-04540-1","url":null,"abstract":"<p>Skin cancer is one of the most dangerous types of cancer due to its immediate appearance and the possibility of rapid spread. It arises from uncontrollably growing cells, rapidly dividing cells in one area of the body, invading other bodily tissues, and spreading throughout the body. Early detection helps prevent cancer progress from reaching critical levels, reducing the risk of complications and the need for more aggressive treatment options. Convolutional neural networks (CNNs) revolutionize skin cancer diagnosis by extracting intricate features from images, enabling an accurate classification of lesions. Their role extends to early detection, providing a powerful tool for dermatologists to identify abnormalities in their nascent stages, ultimately improving patient outcomes. This study proposes a novel deep convolutional neural network (DCNN) approach to classifying skin cancer lesions. The proposed DCNN model is evaluated using two unbalanced datasets, namely HAM10000 and ISIC-2019. The DCNN model is compared with other transfer learning models, including VGG16, VGG19, DenseNet121, DenseNet201, and MobileNetV2. Its performance is assessed using four widely used evaluation metrics: accuracy, recall, precision, F1-score, specificity, and AUC. The experimental results demonstrate that the proposed DCNN model outperforms other deep learning (DL) models that utilized these datasets. The proposed DCNN model achieved the highest accuracy with the HAM10000 and ISIC-2019 datasets, reaching <span>(98.5%)</span> and <span>(97.1%)</span>, respectively. These experimental results show how competitive and successful the DCNN model is in overcoming the problems caused by class imbalance and raising skin cancer classification accuracy. Furthermore, the proposed model demonstrates superior performance, particularly excelling in terms of accuracy, compared to other recent studies that utilize the same datasets, which highlights the robustness and effectiveness of the proposed DCNN.</p>","PeriodicalId":501576,"journal":{"name":"Cluster Computing","volume":"23 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141522114","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-06-17DOI: 10.1007/s10586-024-04624-y
Alparslan Fişne, M. Mücahit Enes Yurtsever, Süleyman Eken
Thermographic inspection is particularly effective in identifying thermal bridges because it visualizes temperature differences on the building’s surface. The focus of this work is on energy-efficient computing for deep learning-based thermal bridge (anomaly) detection models. In this study, we concentrate on object detection-based models such as Mask R-CNN_FPN_50, Swin-T Transformer, and FSAF. We do benchmark tests on TBRR dataset with varying input sizes. To overcome the energy-efficient design, we apply optimizations such as compression, latency reduction, and pruning to these models. After our proposed improvements, the inference of the anomaly detection model, Mask R-CNN_FPN_50 with compression technique, is approximately 7.5% faster than the original. Also, more acceleration is observed in all models with increasing input size. Another criterion we focus on is total energy gain for optimized models. Swin-T transformer has the most inference energy gains for all input sizes ((approx)27 J for 3000 x 4000 and (approx)14 J for 2400 x 3400). In conclusion, our study presents an optimization of size, weight, and power for vision-based anomaly detection for buildings.
{"title":"Energy-efficient buildings with energy-efficient optimized models: a case study on thermal bridge detection","authors":"Alparslan Fişne, M. Mücahit Enes Yurtsever, Süleyman Eken","doi":"10.1007/s10586-024-04624-y","DOIUrl":"https://doi.org/10.1007/s10586-024-04624-y","url":null,"abstract":"<p>Thermographic inspection is particularly effective in identifying thermal bridges because it visualizes temperature differences on the building’s surface. The focus of this work is on energy-efficient computing for deep learning-based thermal bridge (anomaly) detection models. In this study, we concentrate on object detection-based models such as Mask R-CNN_FPN_50, Swin-T Transformer, and FSAF. We do benchmark tests on TBRR dataset with varying input sizes. To overcome the energy-efficient design, we apply optimizations such as compression, latency reduction, and pruning to these models. After our proposed improvements, the inference of the anomaly detection model, Mask R-CNN_FPN_50 with compression technique, is approximately 7.5% faster than the original. Also, more acceleration is observed in all models with increasing input size. Another criterion we focus on is total energy gain for optimized models. Swin-T transformer has the most inference energy gains for all input sizes (<span>(approx)</span>27 J for 3000 x 4000 and <span>(approx)</span>14 J for 2400 x 3400). In conclusion, our study presents an optimization of size, weight, and power for vision-based anomaly detection for buildings.</p>","PeriodicalId":501576,"journal":{"name":"Cluster Computing","volume":"17 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141522111","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-06-17DOI: 10.1007/s10586-024-04545-w
Ali Mohammadzadeh, Seyedali Mirjalili
This paper proposes a meta-heuristic called Eel and Grouper Optimizer (EGO). The EGO algorithm is inspired by the symbiotic interaction and foraging strategy of eels and groupers in marine ecosystems. The algorithm’s efficacy is demonstrated through rigorous evaluation using nineteen benchmark functions, showcasing its superior performance compared to established meta-heuristic algorithms. The findings and results on the benchmark functions demonstrate that the EGO algorithm outperforms well-known meta-heuristics. This work also considers solving a wide range of real-world practical engineering case studies including tension/compression spring, pressure vessel, piston lever, and car side impact, and the CEC 2020 Real-World Benchmark using EGO to illustrate the practicality of the proposed algorithm when dealing with the challenges of real search spaces with unknown global optima. The results show that the proposed EGO algorithm is a reliable soft computing technique for real-world optimization problems and can efficiently outperform the existing algorithms in the literature.
{"title":"Eel and grouper optimizer: a nature-inspired optimization algorithm","authors":"Ali Mohammadzadeh, Seyedali Mirjalili","doi":"10.1007/s10586-024-04545-w","DOIUrl":"https://doi.org/10.1007/s10586-024-04545-w","url":null,"abstract":"<p>This paper proposes a meta-heuristic called Eel and Grouper Optimizer (EGO). The EGO algorithm is inspired by the symbiotic interaction and foraging strategy of eels and groupers in marine ecosystems. The algorithm’s efficacy is demonstrated through rigorous evaluation using nineteen benchmark functions, showcasing its superior performance compared to established meta-heuristic algorithms. The findings and results on the benchmark functions demonstrate that the EGO algorithm outperforms well-known meta-heuristics. This work also considers solving a wide range of real-world practical engineering case studies including tension/compression spring, pressure vessel, piston lever, and car side impact, and the CEC 2020 Real-World Benchmark using EGO to illustrate the practicality of the proposed algorithm when dealing with the challenges of real search spaces with unknown global optima. The results show that the proposed EGO algorithm is a reliable soft computing technique for real-world optimization problems and can efficiently outperform the existing algorithms in the literature.</p>","PeriodicalId":501576,"journal":{"name":"Cluster Computing","volume":"51 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141522113","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-06-16DOI: 10.1007/s10586-024-04592-3
Wei Hu, Ji Feng, Degang Yang
Some particular shaped datasets, such as manifold datasets, have restrictions on density peak clustering (DPC) performance. The main reason of variations in sample densities between clusters of data and uneven densities is not taken into consideration by the DPC algorithm, which could result in the wrong clustering center selection. Additionally, the use of single assignment method is leads to the domino effect of assignment errors. To address these problems, this paper creates a new, improved density peaks clustering method use the similarity assignment strategy with K nearest Neighbors (IDPC-SKNN). Firstly, a new method for defining local density is proposed. Local density is comprehensively consider in the proportion of the average density inside the region, which realize the precise location of low-density clusters. Then, using the samples’ K-nearest neighbors information, a new similarity allocation method is proposed. Allocation strategy successfully address assignment cascading mistakes and improves algorithms robustness. Finally, based on four evaluation indicators, our algorithm outperforms all the comparative clustering algorithm, according to experiments conducted on synthetic, real world and the Olivetti Faces datasets.
一些特殊形状的数据集(如流形数据集)对密度峰聚类(DPC)性能有限制。主要原因是数据簇之间样本密度的变化和密度不均没有被 DPC 算法考虑在内,这可能会导致聚类中心选择错误。此外,使用单一赋值方法会导致赋值错误的多米诺骨牌效应。为了解决这些问题,本文使用 K 最近邻的相似性赋值策略(IDPC-SKNN)创建了一种新的、改进的密度峰聚类方法。首先,本文提出了一种定义局部密度的新方法。局部密度综合考虑了区域内平均密度的比例,实现了低密度聚类的精确定位。然后,利用样本的 K 近邻信息,提出了一种新的相似性分配方法。分配策略成功地解决了分配级联错误,提高了算法的鲁棒性。最后,根据在合成、真实世界和 Olivetti Faces 数据集上进行的实验,基于四个评价指标,我们的算法优于所有比较聚类算法。
{"title":"An improved density peaks clustering algorithm using similarity assignment strategy with K-nearest neighbors","authors":"Wei Hu, Ji Feng, Degang Yang","doi":"10.1007/s10586-024-04592-3","DOIUrl":"https://doi.org/10.1007/s10586-024-04592-3","url":null,"abstract":"<p>Some particular shaped datasets, such as manifold datasets, have restrictions on density peak clustering (DPC) performance. The main reason of variations in sample densities between clusters of data and uneven densities is not taken into consideration by the DPC algorithm, which could result in the wrong clustering center selection. Additionally, the use of single assignment method is leads to the domino effect of assignment errors. To address these problems, this paper creates a new, improved density peaks clustering method use the similarity assignment strategy with K nearest Neighbors (IDPC-SKNN). Firstly, a new method for defining local density is proposed. Local density is comprehensively consider in the proportion of the average density inside the region, which realize the precise location of low-density clusters. Then, using the samples’ K-nearest neighbors information, a new similarity allocation method is proposed. Allocation strategy successfully address assignment cascading mistakes and improves algorithms robustness. Finally, based on four evaluation indicators, our algorithm outperforms all the comparative clustering algorithm, according to experiments conducted on synthetic, real world and the Olivetti Faces datasets.</p>","PeriodicalId":501576,"journal":{"name":"Cluster Computing","volume":"136 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-06-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141522023","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}