Pub Date : 2024-09-15DOI: 10.1007/s00607-024-01345-3
Syed Ali Haider, Junaid A. Zubairi, Sahar Idwan
Traffic congestion in urban areas poses several challenges to municipal authorities including pollution, productivity loss, reckless driving, and delays in dealing with emergencies. Smart cities can use modern IoT infrastructure to solve the congestion problem and reduce pollution and delays. In this article, we focus on congestion mapping and mitigation for emergency vehicles in smart cities. We use a novel traffic light control technique to change the flow of cars on lights of interest thereby making way for emergency vehicles. We use a simulation model for a selected area of Manhattan to implement congestion mapping and to help find the fastest path for routing emergency vehicles based on the congestion metrics. The system controls traffic lights to block off the roads feeding into congestion and allows flow away from the congested path. This helps in clearing the preferred route to help emergency vehicles reach the destination faster. We show that the proposed algorithm can map congestion on city roads with accuracy thus helping to improve the response time of the emergency services and saving precious lives.
{"title":"Mapping and just-in-time traffic congestion mitigation for emergency vehicles in smart cities","authors":"Syed Ali Haider, Junaid A. Zubairi, Sahar Idwan","doi":"10.1007/s00607-024-01345-3","DOIUrl":"https://doi.org/10.1007/s00607-024-01345-3","url":null,"abstract":"<p>Traffic congestion in urban areas poses several challenges to municipal authorities including pollution, productivity loss, reckless driving, and delays in dealing with emergencies. Smart cities can use modern IoT infrastructure to solve the congestion problem and reduce pollution and delays. In this article, we focus on congestion mapping and mitigation for emergency vehicles in smart cities. We use a novel traffic light control technique to change the flow of cars on lights of interest thereby making way for emergency vehicles. We use a simulation model for a selected area of Manhattan to implement congestion mapping and to help find the fastest path for routing emergency vehicles based on the congestion metrics. The system controls traffic lights to block off the roads feeding into congestion and allows flow away from the congested path. This helps in clearing the preferred route to help emergency vehicles reach the destination faster. We show that the proposed algorithm can map congestion on city roads with accuracy thus helping to improve the response time of the emergency services and saving precious lives.</p>","PeriodicalId":10718,"journal":{"name":"Computing","volume":"40 1","pages":""},"PeriodicalIF":3.7,"publicationDate":"2024-09-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142258095","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-13DOI: 10.1007/s00607-024-01348-0
J. Angela Jennifa Sujana, R. Venitta Raj, V. K. Raja Priya
Street lamp is a great asset for human society with a narrow beam spread light. The extensive proliferation of solar power in street lamps causes power outages due to their variable power-generated profiles. Thus Smart Street Lamp Fog Intelligence (SSLFI) framework based on hierarchical learning was proposed for efficient energy management in solar street lamps. Smart Street Lamp (SSL) shifts its brightness at higher and lower light levels with a comforting, energy-efficient gleam of light. The fog intelligence framework forecasts the SSL output power through short-term probabilistic energy consumption forecasts using Q-NARX-BiLSTM (Quantile Regression-Nonlinear Auto-Regressive Neural Networks with exogenous input-Bidirectional Long short-term memory) model. NARX-BiLSTM of two module types: (1) NARXNN (Nonlinear Auto-Regressive Neural Networks with exogenous input) model generates SSL power consumption and (2) BiLSTM (Bidirectional Long short-term memory) model generates SSL power forecasts. The quantile regression with the NARX-BiLSTM (Nonlinear Auto-Regressive Neural Networks with exogenous input-Bidirectional Long short-term memory) model forecasts the seasonal patterns achieving non-parametric interval predictions. The probabilistic predictions of power consumption are determined based on the conditional quantile using an improved kernel density estimation approach. The fuzzy inference system adopts forecasting results to diagnose fault conditions in street lamps. The experiment results show that the proposed framework SSLFI outperformed the state-of-the-art models forecasting under different weather conditions.
{"title":"Fog intelligence for energy efficient management in smart street lamps","authors":"J. Angela Jennifa Sujana, R. Venitta Raj, V. K. Raja Priya","doi":"10.1007/s00607-024-01348-0","DOIUrl":"https://doi.org/10.1007/s00607-024-01348-0","url":null,"abstract":"<p>Street lamp is a great asset for human society with a narrow beam spread light. The extensive proliferation of solar power in street lamps causes power outages due to their variable power-generated profiles. Thus Smart Street Lamp Fog Intelligence (SSLFI) framework based on hierarchical learning was proposed for efficient energy management in solar street lamps. Smart Street Lamp (SSL) shifts its brightness at higher and lower light levels with a comforting, energy-efficient gleam of light. The fog intelligence framework forecasts the SSL output power through short-term probabilistic energy consumption forecasts using Q-NARX-BiLSTM (Quantile Regression-Nonlinear Auto-Regressive Neural Networks with exogenous input-Bidirectional Long short-term memory) model. NARX-BiLSTM of two module types: (1) NARXNN (Nonlinear Auto-Regressive Neural Networks with exogenous input) model generates SSL power consumption and (2) BiLSTM (Bidirectional Long short-term memory) model generates SSL power forecasts. The quantile regression with the NARX-BiLSTM (Nonlinear Auto-Regressive Neural Networks with exogenous input-Bidirectional Long short-term memory) model forecasts the seasonal patterns achieving non-parametric interval predictions. The probabilistic predictions of power consumption are determined based on the conditional quantile using an improved kernel density estimation approach. The fuzzy inference system adopts forecasting results to diagnose fault conditions in street lamps. The experiment results show that the proposed framework SSLFI outperformed the state-of-the-art models forecasting under different weather conditions.</p>","PeriodicalId":10718,"journal":{"name":"Computing","volume":"7 1","pages":""},"PeriodicalIF":3.7,"publicationDate":"2024-09-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142222718","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-13DOI: 10.1007/s00607-024-01333-7
Divyans Mahansaria, Uttam Kumar Roy
At the time of authentication, confidential data are exchanged between the user/device and the authentication server to determine the legitimacy of the source requesting authentication. Safeguarding the authentication process from security attacks is of utmost importance, and various authentication methods exist depending on the system’s requirements. However, no authentication process can guarantee full-proof security. This research aimed to use the context of users and devices during authentication to detect anomalies and security-related attacks. In particular, denial-of-service (DoS)/distributed denial-of-service (DDoS) attacks and brute-force attacks have been analyzed in detail using contextual information. Extensive simulations were conducted on the benchmark CIC-IDS2017 dataset using the Weka tool. The performance metrics of recall, precision, accuracy, f-score, and model-built time were computed for the four machine-learning classifiers—J48, Random Forest, Multi-Layer Perceptron, and Bayes Net—for different combinations of data splits and groups of data features. For both DoS/DDoS and brute-force attacks, some of the experimental results show a more than 99% value for recall, precision, accuracy, and f-score. The results of the experiments, security analysis, and threat modeling show that the proposed authentication scheme effectively enhances a secure system’s security level.
在进行身份验证时,用户/设备与身份验证服务器之间会交换机密数据,以确定请求身份验证来源的合法性。防止身份验证过程受到安全攻击是最重要的,根据系统的要求有各种身份验证方法。然而,没有任何一种身份验证过程能保证完全安全。这项研究旨在利用用户和设备在身份验证过程中的上下文来检测异常情况和与安全相关的攻击。特别是,利用上下文信息详细分析了拒绝服务(DoS)/分布式拒绝服务(DDoS)攻击和暴力破解攻击。使用 Weka 工具在基准 CIC-IDS2017 数据集上进行了大量模拟。针对不同的数据拆分组合和数据特征组,计算了四种机器学习分类器--J48、随机森林、多层感知器和贝叶斯网的召回率、精确度、准确率、f-分数和建模时间等性能指标。对于 DoS/DDoS 和暴力破解攻击,部分实验结果显示召回率、精确率、准确率和 f 分数均超过 99%。实验、安全分析和威胁建模的结果表明,所提出的认证方案能有效提高安全系统的安全级别。
{"title":"Contextual authentication of users and devices using machine learning","authors":"Divyans Mahansaria, Uttam Kumar Roy","doi":"10.1007/s00607-024-01333-7","DOIUrl":"https://doi.org/10.1007/s00607-024-01333-7","url":null,"abstract":"<p>At the time of authentication, confidential data are exchanged between the user/device and the authentication server to determine the legitimacy of the source requesting authentication. Safeguarding the authentication process from security attacks is of utmost importance, and various authentication methods exist depending on the system’s requirements. However, no authentication process can guarantee full-proof security. This research aimed to use the context of users and devices during authentication to detect anomalies and security-related attacks. In particular, denial-of-service (DoS)/distributed denial-of-service (DDoS) attacks and brute-force attacks have been analyzed in detail using contextual information. Extensive simulations were conducted on the benchmark CIC-IDS2017 dataset using the Weka tool. The performance metrics of recall, precision, accuracy, f-score, and model-built time were computed for the four machine-learning classifiers—J48, Random Forest, Multi-Layer Perceptron, and Bayes Net—for different combinations of data splits and groups of data features. For both DoS/DDoS and brute-force attacks, some of the experimental results show a more than 99% value for recall, precision, accuracy, and f-score. The results of the experiments, security analysis, and threat modeling show that the proposed authentication scheme effectively enhances a secure system’s security level.</p>","PeriodicalId":10718,"journal":{"name":"Computing","volume":"183 1","pages":""},"PeriodicalIF":3.7,"publicationDate":"2024-09-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142222720","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
One of the most well-known names that has recently attained new heights and set a standard is Internet of Things (IoT). IoT aims to connect all physical devices in such a way that they are subject to human control over the Internet.The emergence of IoT in almost all the industries has redesigned them including smart agriculture. In today’s world, the growth in agriculture sector is rapid, smarter and precise than ever. In case of IoT, the objects are termed as services, sometimes with similar functionalities but distinct quality of service parameters. As the user’s requirements are complex, a single service cannot fulfil them efficiently. So, service composition is the solution. These services known as atomic services, are represented as workflow, with each of them having distinct candidate composite services. Fulfilling these Quality of Service (QoS) constraints makes it a NP-hard problem which can’t be solved using traditional approaches. Hence, comes the concept of evolutionary approaches. In this paper one of the evolutionary approach- NSGA-II is used to optimize the production of apple by composing the various services, taking into account the cost and time as multi-objective problem to be solved. This is for the very first time that QoS aware service composition problem has been optimized in smart agriculture as found in the literature. Results are further compared with multi-objective genetic algorithm (MOGA) and it has been found that NSGA-II outperforms MOGA by generating well-proportioned pareto optimal solutions.
{"title":"Multi-objective service composition optimization problem in IoT for agriculture 4.0","authors":"Shalini Sharma, Bhupendra Kumar Pathak, Rajiv Kumar","doi":"10.1007/s00607-024-01346-2","DOIUrl":"https://doi.org/10.1007/s00607-024-01346-2","url":null,"abstract":"<p>One of the most well-known names that has recently attained new heights and set a standard is Internet of Things (IoT). IoT aims to connect all physical devices in such a way that they are subject to human control over the Internet.The emergence of IoT in almost all the industries has redesigned them including smart agriculture. In today’s world, the growth in agriculture sector is rapid, smarter and precise than ever. In case of IoT, the objects are termed as services, sometimes with similar functionalities but distinct quality of service parameters. As the user’s requirements are complex, a single service cannot fulfil them efficiently. So, service composition is the solution. These services known as atomic services, are represented as workflow, with each of them having distinct candidate composite services. Fulfilling these Quality of Service (QoS) constraints makes it a NP-hard problem which can’t be solved using traditional approaches. Hence, comes the concept of evolutionary approaches. In this paper one of the evolutionary approach- NSGA-II is used to optimize the production of apple by composing the various services, taking into account the cost and time as multi-objective problem to be solved. This is for the very first time that QoS aware service composition problem has been optimized in smart agriculture as found in the literature. Results are further compared with multi-objective genetic algorithm (MOGA) and it has been found that NSGA-II outperforms MOGA by generating well-proportioned pareto optimal solutions.</p>","PeriodicalId":10718,"journal":{"name":"Computing","volume":"27 1","pages":""},"PeriodicalIF":3.7,"publicationDate":"2024-09-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142222721","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-06DOI: 10.1007/s00607-024-01344-4
Yingze Liu, Yuanbo Guo, Rajiv Ranjan, Dan Chen
This study introduces a Deep Reinforcement Learning approach (DRL-MD) aimed at optimizing the deployment of mitigations to minimize redundancy while ensuring effective defense against cyberattacks. DRL-MD initially enhances ATT &CK (Adversarial Tactics, Techniques, and Common Knowledge) to underscore the formal relationships between attacks and defenses. Over the enhanced ATT &CK, DRL-MD then operates in two phases: (1) Estimating Node Importance: DRL-MD proposes a model to estimate the importance of deployed nodes in the network, prioritizing mitigation deployment locations for better evaluation of mitigation effectiveness; and (2) Optimizing Mitigation Deployment: A Soft Actor-Critic algorithm finds the optimal mitigation deployment policy through multi-objective optimization of the importance of deployed nodes, the effectiveness of mitigations in preventing cyberattacks, vulnerability repair, and deployment cost. A case study with DRL-MD against the state-of-the-art counterparts has been performed considering the WannaCry threat, and results indicate that: (1) DRL-MD performs the best with 6.4–11% decrease in deployment cost; and (2) DRL-MD can significantly reduce redundancy in mitigation deployments, which partially benefits from the enhanced ATT &CK model. Overall, a comprehensive solution of mitigation deployment has been fostered to significantly lower the redundancy with more effective defenses against cyberattacks sustained.
本研究介绍了一种深度强化学习方法(DRL-MD),旨在优化缓解措施的部署,以尽量减少冗余,同时确保有效防御网络攻击。DRL-MD 最初增强了 ATT &CK (对抗战术、技术和常识),以强调攻击与防御之间的正式关系。在增强 ATT &CK 的基础上,DRL-MD 分两个阶段运行:(1)估计节点重要性:DRL-MD 提出了一个估算网络中部署节点重要性的模型,优先考虑缓解部署位置,以便更好地评估缓解效果;以及 (2) 优化缓解部署:软行为批判算法通过对部署节点的重要性、缓解措施在预防网络攻击方面的有效性、漏洞修复和部署成本进行多目标优化,找到最佳缓解部署策略。考虑到 WannaCry 威胁,利用 DRL-MD 与最先进的同行进行了案例研究,结果表明结果表明:(1) DRL-MD 性能最佳,部署成本降低了 6.4-11%;(2) DRL-MD 可显著减少缓解部署中的冗余,这部分得益于增强型 ATT &CK 模型。总之,一个全面的缓解部署解决方案已经形成,可显著降低冗余,更有效地防御网络攻击。
{"title":"Optimization of mitigation deployment using deep reinforcement learning over an enhanced ATT &CK","authors":"Yingze Liu, Yuanbo Guo, Rajiv Ranjan, Dan Chen","doi":"10.1007/s00607-024-01344-4","DOIUrl":"https://doi.org/10.1007/s00607-024-01344-4","url":null,"abstract":"<p>This study introduces a Deep Reinforcement Learning approach (DRL-MD) aimed at optimizing the deployment of mitigations to minimize redundancy while ensuring effective defense against cyberattacks. DRL-MD initially enhances ATT &CK (Adversarial Tactics, Techniques, and Common Knowledge) to underscore the formal relationships between attacks and defenses. Over the enhanced ATT &CK, DRL-MD then operates in two phases: (1) <i>Estimating Node Importance</i>: DRL-MD proposes a model to estimate the importance of deployed nodes in the network, prioritizing mitigation deployment locations for better evaluation of mitigation effectiveness; and (2) <i>Optimizing Mitigation Deployment</i>: A Soft Actor-Critic algorithm finds the optimal mitigation deployment policy through multi-objective optimization of the importance of deployed nodes, the effectiveness of mitigations in preventing cyberattacks, vulnerability repair, and deployment cost. A case study with DRL-MD against the state-of-the-art counterparts has been performed considering the <i>WannaCry</i> threat, and results indicate that: (1) DRL-MD performs the best with 6.4–11% decrease in deployment cost; and (2) DRL-MD can significantly reduce redundancy in mitigation deployments, which partially benefits from the enhanced ATT &CK model. Overall, a comprehensive solution of mitigation deployment has been fostered to significantly lower the redundancy with more effective defenses against cyberattacks sustained.</p>","PeriodicalId":10718,"journal":{"name":"Computing","volume":"437 1","pages":""},"PeriodicalIF":3.7,"publicationDate":"2024-09-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142222724","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Evaluating GPU compute instances for High Performance Computing (HPC) and Artificial Intelligence (AI) applications in the cloud involves complex decision-making processes. This research applies the Technique for Order of Preference by Similarity to Ideal Solution (TOPSIS) to rank various GPU compute instances for HPC and AI from leading cloud providers: Amazon Web Services (AWS), Microsoft Azure, Google Cloud Platform (GCP), and Oracle Cloud Infrastructure (OCI). The analysis incorporates a sensitivity examination, bootstrapping, and non-parametric tests to ensure robust and reliable rankings. Sensitivity analysis reveals the stability of the TOPSIS framework despite variations in criteria weights, while bootstrap analysis provides confidence intervals for the rankings, highlighting their consistency. The Friedman test confirms that ranking stability persists across different scenarios, indicating minimal impact from weight adjustments. Despite these insights, limitations such as interdependencies among criteria, data accuracy, and generalizability constraints must be acknowledged. This comprehensive approach ensures informed decision-making for selecting optimal GPU instances for cloud-based HPC and AI tasks.
{"title":"Robust evaluation of GPU compute instances for HPC and AI in the cloud: a TOPSIS approach with sensitivity, bootstrapping, and non-parametric analysis","authors":"Mandeep Kumar, Gagandeep Kaur, Prashant Singh Rana","doi":"10.1007/s00607-024-01342-6","DOIUrl":"https://doi.org/10.1007/s00607-024-01342-6","url":null,"abstract":"<p>Evaluating GPU compute instances for High Performance Computing (HPC) and Artificial Intelligence (AI) applications in the cloud involves complex decision-making processes. This research applies the Technique for Order of Preference by Similarity to Ideal Solution (TOPSIS) to rank various GPU compute instances for HPC and AI from leading cloud providers: Amazon Web Services (AWS), Microsoft Azure, Google Cloud Platform (GCP), and Oracle Cloud Infrastructure (OCI). The analysis incorporates a sensitivity examination, bootstrapping, and non-parametric tests to ensure robust and reliable rankings. Sensitivity analysis reveals the stability of the TOPSIS framework despite variations in criteria weights, while bootstrap analysis provides confidence intervals for the rankings, highlighting their consistency. The Friedman test confirms that ranking stability persists across different scenarios, indicating minimal impact from weight adjustments. Despite these insights, limitations such as interdependencies among criteria, data accuracy, and generalizability constraints must be acknowledged. This comprehensive approach ensures informed decision-making for selecting optimal GPU instances for cloud-based HPC and AI tasks.\u0000</p>","PeriodicalId":10718,"journal":{"name":"Computing","volume":"40 1","pages":""},"PeriodicalIF":3.7,"publicationDate":"2024-09-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142222722","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-05DOI: 10.1007/s00607-024-01343-5
Hyejin Cha, In Kee Kim, Taeseok Kim
Efficient management of the write buffer in solid-state drives (SSDs) can be achieved by predicting future I/O request patterns using machine learning techniques. However, the computational demands posed by sophisticated approaches like deep learning remain significant, despite the increasing computational power of SSDs. This paper presents a novel approach to write buffer management that addresses these challenges. Our method employs a lightweight yet accurate random forest classifier to predict the forward reuse distances (FRDs) of I/O requests, indicating the likelihood of recurring identical I/O requests. Our key insight is that, rather than aiming for exact FRD predictions for future individual requests, we focus on identifying whether the predicted FRD exceeds the buffer size. With this insight, our method implements efficient buffer management operations, including bypassing the buffer storage when necessary. To achieve this, we introduce a banding method that quantizes FRDs according to the buffer size. This enables predictions at the band level, forming the foundation for a lightweight machine learning model. Subsequently, we assign high caching priority to write requests that are anticipated to have a short FRD band. Through extensive evaluations utilizing a simulator, we demonstrate that our method achieves results comparable to those of the optimal algorithm in terms of hit rate in most scenarios. Moreover, our approach outperforms state-of-the-art algorithms, which depend on past I/O reference patterns, by up to 27%.
{"title":"Using a random forest to predict quantized reuse distance in an SSD write buffer","authors":"Hyejin Cha, In Kee Kim, Taeseok Kim","doi":"10.1007/s00607-024-01343-5","DOIUrl":"https://doi.org/10.1007/s00607-024-01343-5","url":null,"abstract":"<p>Efficient management of the write buffer in solid-state drives (SSDs) can be achieved by predicting future I/O request patterns using machine learning techniques. However, the computational demands posed by sophisticated approaches like deep learning remain significant, despite the increasing computational power of SSDs. This paper presents a novel approach to write buffer management that addresses these challenges. Our method employs a lightweight yet accurate random forest classifier to predict the forward reuse distances (FRDs) of I/O requests, indicating the likelihood of recurring identical I/O requests. Our key insight is that, rather than aiming for exact FRD predictions for future individual requests, we focus on identifying whether the predicted FRD exceeds the buffer size. With this insight, our method implements efficient buffer management operations, including bypassing the buffer storage when necessary. To achieve this, we introduce a banding method that quantizes FRDs according to the buffer size. This enables predictions at the band level, forming the foundation for a lightweight machine learning model. Subsequently, we assign high caching priority to write requests that are anticipated to have a short FRD band. Through extensive evaluations utilizing a simulator, we demonstrate that our method achieves results comparable to those of the optimal algorithm in terms of hit rate in most scenarios. Moreover, our approach outperforms state-of-the-art algorithms, which depend on past I/O reference patterns, by up to 27%.</p>","PeriodicalId":10718,"journal":{"name":"Computing","volume":"29 1","pages":""},"PeriodicalIF":3.7,"publicationDate":"2024-09-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142222723","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-02DOI: 10.1007/s00607-024-01320-y
Santiago Mejía, Jose Aguilar
Retail companies face major problems in the estimation of their product’s future demand due to the high diversity of sales behavior that each good presents. Different forecasting models are implemented to meet the demand requirements for efficient inventory management. However, in most of the proposed works, a single model approach is applied to forecast all products, ignoring that some methods are better adapted for certain features of the demand time series of each product. The proposed forecasting system addresses this problem, by implementing a two-phase methodology that initially clusters the products with the application of an unsupervised learning approach using the extracted demand features of each good, and then, implements a second phase where, after a feature engineering process, a set of different forecasting methods are evaluated to identify those with best performs for each cluster. Finally, ensemble machine learning models are implemented using the top-performing models of each cluster to carry out the demand estimation. The results indicate that the proposed forecasting system improves the demand estimation over the single forecasting approaches when evaluating the R2, MSE, and MASE quality measures.
由于每种商品的销售行为千差万别,零售公司在估计其产品的未来需求时面临着重大问题。为了满足高效库存管理的需求要求,人们采用了不同的预测模型。然而,在大多数提议的工作中,都是采用单一模型方法来预测所有产品,而忽略了有些方法更适合每种产品需求时间序列的某些特征。为解决这一问题,所提出的预测系统分为两个阶段:首先,采用无监督学习方法,利用提取的每种商品的需求特征对产品进行聚类;然后,实施第二阶段,在特征工程流程之后,对一系列不同的预测方法进行评估,以确定哪些方法在每个聚类中表现最佳。最后,使用每个群组中表现最好的模型来实施集合机器学习模型,以进行需求预测。结果表明,在评估 R2、MSE 和 MASE 质量指标时,与单一预测方法相比,建议的预测系统提高了需求预测效果。
{"title":"A demand forecasting system of product categories defined by their time series using a hybrid approach of ensemble learning with feature engineering","authors":"Santiago Mejía, Jose Aguilar","doi":"10.1007/s00607-024-01320-y","DOIUrl":"https://doi.org/10.1007/s00607-024-01320-y","url":null,"abstract":"<p>Retail companies face major problems in the estimation of their product’s future demand due to the high diversity of sales behavior that each good presents. Different forecasting models are implemented to meet the demand requirements for efficient inventory management. However, in most of the proposed works, a single model approach is applied to forecast all products, ignoring that some methods are better adapted for certain features of the demand time series of each product. The proposed forecasting system addresses this problem, by implementing a two-phase methodology that initially clusters the products with the application of an unsupervised learning approach using the extracted demand features of each good, and then, implements a second phase where, after a feature engineering process, a set of different forecasting methods are evaluated to identify those with best performs for each cluster. Finally, ensemble machine learning models are implemented using the top-performing models of each cluster to carry out the demand estimation. The results indicate that the proposed forecasting system improves the demand estimation over the single forecasting approaches when evaluating the R<sup>2</sup>, MSE, and MASE quality measures.</p>","PeriodicalId":10718,"journal":{"name":"Computing","volume":"99 1","pages":""},"PeriodicalIF":3.7,"publicationDate":"2024-09-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142222725","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-25DOI: 10.1007/s00607-024-01340-8
Tassawar Ali, Hikmat Ullah Khan, Fawaz Khaled Alarfaj, Mohammed AlReshoodi
Cloud computing offers demand-based allocation of required resources to its clients ensuring optimal use of resources in a cost-effective manner. However, due to the massive increase in demand for physical resources by datacenters cloud management suffers from inefficient resource management. To enhance efficiency by reducing resource setup time, workload prediction has become an active research area. It helps to make management decisions proactively and enables the cloud management system to better respond to spikes in the workload. This study proposes a hybrid model combining both state-of-the-art deep learning models and evolutionary algorithms for workload prediction. The proposed cluster-based differential evolution neural network model utilizes differential evolution for the optimization of feature weights of the deep neural network to predict the future workloads of a cloud datacenter. The proposed model uses a novel mutation strategy that clusters the population based on an agglomerative technique and chooses the best gene from randomly chosen clusters. Thus, the strategy creates a balance between the exploration and exploitation of the population and enables the model to avoid local optima and converge rapidly. The datasets used for the experiments are created from Google’s real-world traces and the Alibaba platform. The model is compared with backpropagation, Adam optimizer-based LSTM, and an evolutionary neural network-based three-mutation policy. We evaluated the performance of the proposed model in terms of root mean squared error in predicting the upcoming CPU, RAM, and BW usage. The proposed model achieved an error rate as low as 0.0002 to outperform the existing studies in the relevant literature. To further authenticate the results, we performed the statistical analysis of the obtained results in terms of R-squared, mean bias deviation, 90th percentile score, and Theil’s U statistics. The high accuracy and automaticity of the proposed model have paved the way for its application in diverse areas of cloud computing, including real-time applications.
云计算根据客户需求分配所需资源,确保以经济高效的方式优化使用资源。然而,由于数据中心对物理资源的需求大量增加,云计算管理存在资源管理效率低下的问题。为了通过减少资源设置时间来提高效率,工作量预测已成为一个活跃的研究领域。它有助于主动做出管理决策,使云管理系统能够更好地应对工作负载的激增。本研究提出了一种混合模型,将最先进的深度学习模型和进化算法相结合,用于工作量预测。所提出的基于集群的差分进化神经网络模型利用差分进化来优化深度神经网络的特征权重,从而预测云数据中心未来的工作负载。所提出的模型采用了一种新颖的突变策略,该策略基于聚类技术对种群进行聚类,并从随机选择的聚类中选择最佳基因。因此,该策略在种群的探索和利用之间建立了平衡,使模型能够避免局部最优并快速收敛。实验所使用的数据集来自谷歌的真实跟踪和阿里巴巴平台。我们将该模型与反向传播、基于亚当优化器的 LSTM 以及基于进化神经网络的三突变策略进行了比较。我们从均方根误差的角度评估了所提模型在预测即将到来的 CPU、RAM 和 BW 使用率方面的性能。拟议模型的误差率低至 0.0002,优于相关文献中的现有研究。为了进一步验证结果,我们对所获得的结果进行了统计分析,包括 R 方、平均偏差、第 90 百分位数得分和 Theil's U 统计量。所提模型的高准确性和自动性为其在云计算各领域(包括实时应用)的应用铺平了道路。
{"title":"Hybrid deep learning and evolutionary algorithms for accurate cloud workload prediction","authors":"Tassawar Ali, Hikmat Ullah Khan, Fawaz Khaled Alarfaj, Mohammed AlReshoodi","doi":"10.1007/s00607-024-01340-8","DOIUrl":"https://doi.org/10.1007/s00607-024-01340-8","url":null,"abstract":"<p>Cloud computing offers demand-based allocation of required resources to its clients ensuring optimal use of resources in a cost-effective manner. However, due to the massive increase in demand for physical resources by datacenters cloud management suffers from inefficient resource management. To enhance efficiency by reducing resource setup time, workload prediction has become an active research area. It helps to make management decisions proactively and enables the cloud management system to better respond to spikes in the workload. This study proposes a hybrid model combining both state-of-the-art deep learning models and evolutionary algorithms for workload prediction. The proposed cluster-based differential evolution neural network model utilizes differential evolution for the optimization of feature weights of the deep neural network to predict the future workloads of a cloud datacenter. The proposed model uses a novel mutation strategy that clusters the population based on an agglomerative technique and chooses the best gene from randomly chosen clusters. Thus, the strategy creates a balance between the exploration and exploitation of the population and enables the model to avoid local optima and converge rapidly. The datasets used for the experiments are created from Google’s real-world traces and the Alibaba platform. The model is compared with backpropagation, Adam optimizer-based LSTM, and an evolutionary neural network-based three-mutation policy. We evaluated the performance of the proposed model in terms of root mean squared error in predicting the upcoming CPU, RAM, and BW usage. The proposed model achieved an error rate as low as 0.0002 to outperform the existing studies in the relevant literature. To further authenticate the results, we performed the statistical analysis of the obtained results in terms of R-squared, mean bias deviation, 90th percentile score, and Theil’s U statistics. The high accuracy and automaticity of the proposed model have paved the way for its application in diverse areas of cloud computing, including real-time applications.</p>","PeriodicalId":10718,"journal":{"name":"Computing","volume":"10 1","pages":""},"PeriodicalIF":3.7,"publicationDate":"2024-08-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142222728","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-22DOI: 10.1007/s00607-024-01331-9
David Rodriguez, Ian Yang, Jose M. Del Alamo, Norman Sadeh
The number and dynamic nature of web sites and mobile applications present regulators and app store operators with significant challenges when it comes to enforcing compliance with applicable privacy and data protection laws. Over the past several years, people have turned to Natural Language Processing (NLP) techniques to automate privacy compliance analysis (e.g., comparing statements in privacy policies with analysis of the code and behavior of mobile apps) and to answer people’s privacy questions. Traditionally, these NLP techniques have relied on labor-intensive and potentially error-prone manual annotation processes to build the corpora necessary to train them. This article explores and evaluates the use of Large Language Models (LLMs) as an alternative for effectively and efficiently identifying and categorizing a variety of data practice disclosures found in the text of privacy policies. Specifically, we report on the performance of ChatGPT and Llama 2, two particularly popular LLM-based tools. This includes engineering prompts and evaluating different configurations of these LLM techniques. Evaluation of the resulting techniques on well-known corpora of privacy policy annotations yields an F1 score exceeding 93%. This score is higher than scores reported earlier in the literature on these benchmarks. This performance is obtained at minimal marginal cost (excluding the cost required to train the foundational models themselves). These results, which are consistent with those reported in other domains, suggest that LLMs offer a particularly promising approach to automated privacy policy analysis at scale.
{"title":"Large language models: a new approach for privacy policy analysis at scale","authors":"David Rodriguez, Ian Yang, Jose M. Del Alamo, Norman Sadeh","doi":"10.1007/s00607-024-01331-9","DOIUrl":"https://doi.org/10.1007/s00607-024-01331-9","url":null,"abstract":"<p>The number and dynamic nature of web sites and mobile applications present regulators and app store operators with significant challenges when it comes to enforcing compliance with applicable privacy and data protection laws. Over the past several years, people have turned to Natural Language Processing (NLP) techniques to automate privacy compliance analysis (e.g., comparing statements in privacy policies with analysis of the code and behavior of mobile apps) and to answer people’s privacy questions. Traditionally, these NLP techniques have relied on labor-intensive and potentially error-prone manual annotation processes to build the corpora necessary to train them. This article explores and evaluates the use of Large Language Models (LLMs) as an alternative for effectively and efficiently identifying and categorizing a variety of data practice disclosures found in the text of privacy policies. Specifically, we report on the performance of ChatGPT and Llama 2, two particularly popular LLM-based tools. This includes engineering prompts and evaluating different configurations of these LLM techniques. Evaluation of the resulting techniques on well-known corpora of privacy policy annotations yields an F1 score exceeding 93%. This score is higher than scores reported earlier in the literature on these benchmarks. This performance is obtained at minimal marginal cost (excluding the cost required to train the foundational models themselves). These results, which are consistent with those reported in other domains, suggest that LLMs offer a particularly promising approach to automated privacy policy analysis at scale.</p>","PeriodicalId":10718,"journal":{"name":"Computing","volume":"8 1","pages":""},"PeriodicalIF":3.7,"publicationDate":"2024-08-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142222726","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}