首页 > 最新文献

The International Arab Journal of Information Technology最新文献

英文 中文
Deep Learning Based Feature Discriminability Boosted Concurrent Metal Surface Defect Detection System Using YOLOv-5s-FRN 使用 YOLOv-5s-FRN 的基于深度学习的特征可区分性增强型并发金属表面缺陷检测系统
Pub Date : 2024-01-01 DOI: 10.34028//iajit/21/1/9
Reshma P. Vengaloor, Roopa Muralidhar
Computer vision and deep learning techniques are the most emerging technologies in this era. Both of these can greatly raise the rate at which defects on metal surfaces are identified while performing industrial quality checks. The identification of faults over metal surfaces can be viewed as a significant challenge since they are easily impacted by ambient factors including illumination and light reflections. This paper proposes novel metal surface defect detection network called as YOLOv-5s-FRN in response to the problems of ineffective detection brought by the conventional manual inspection system. The proposed system is developed through the integration of a novel architectural module called as Feature Recalibration Network (FRN) to the You Only Look Once-version-5 small network )YOLOv-5s(. In order to extract the global feature information from the provided image, FRN is able to evaluate the interdependencies between the channels. This improves the feature discrimination capability and prediction accuracy of the defect detection system. The incorporation of FRN structure makes YOLOv-5s architecture to selectively enhance the necessary features and discard the unwanted ones. Therefore, the proposed novel method will efficiently detect and classify the metal surface defects such as crazing, patches, inclusions, scratches, pitted surfaces and rolled in scale. North Eastern University Surface Defect Database (NEU-DET) has been used to train and test the proposed architectural model. The suggested system has been compared with alternative models based on several performance matrices such as precision, recall and Mean Average Precision (mAP). It is observed that the proposed YOLOv-5s-FRN architecture provides significant performance improvement than state-of-the-art methods. The proposed system has been provided satisfactory results by means of improvement in mAP and time consumption. The proposed model has delivered value of mAP_0.5 as 98.05% and that of mAP_0.5:0.95 as 89.03%.
计算机视觉和深度学习技术是当今时代最新兴的技术。在进行工业质量检查时,这两种技术都能大大提高金属表面缺陷的识别率。由于金属表面容易受到光照和光反射等环境因素的影响,因此金属表面缺陷的识别是一项重大挑战。针对传统人工检测系统检测效果不佳的问题,本文提出了名为 YOLOv-5s-FRN 的新型金属表面缺陷检测网络。为了从提供的图像中提取全局特征信息,FRN 能够评估通道之间的相互依赖关系。这就提高了缺陷检测系统的特征辨别能力和预测精度。FRN 结构的加入使 YOLOv-5s 架构能够选择性地增强必要的特征,而舍弃不需要的特征。因此,所提出的新方法将有效地检测和分类金属表面缺陷,如裂纹、斑块、夹杂物、划痕、凹坑表面和轧制。东北大学表面缺陷数据库(NEU-DET)被用来训练和测试所提出的建筑模型。根据精确度、召回率和平均精确度 (mAP) 等性能矩阵,将建议的系统与其他模型进行了比较。结果表明,与最先进的方法相比,建议的 YOLOv-5s-FRN 架构能显著提高性能。通过改善 mAP 和耗时,拟议系统取得了令人满意的结果。拟议模型的 mAP_0.5 值为 98.05%,mAP_0.5:0.95 值为 89.03%。
{"title":"Deep Learning Based Feature Discriminability Boosted Concurrent Metal Surface Defect Detection System Using YOLOv-5s-FRN","authors":"Reshma P. Vengaloor, Roopa Muralidhar","doi":"10.34028//iajit/21/1/9","DOIUrl":"https://doi.org/10.34028//iajit/21/1/9","url":null,"abstract":"Computer vision and deep learning techniques are the most emerging technologies in this era. Both of these can greatly raise the rate at which defects on metal surfaces are identified while performing industrial quality checks. The identification of faults over metal surfaces can be viewed as a significant challenge since they are easily impacted by ambient factors including illumination and light reflections. This paper proposes novel metal surface defect detection network called as YOLOv-5s-FRN in response to the problems of ineffective detection brought by the conventional manual inspection system. The proposed system is developed through the integration of a novel architectural module called as Feature Recalibration Network (FRN) to the You Only Look Once-version-5 small network )YOLOv-5s(. In order to extract the global feature information from the provided image, FRN is able to evaluate the interdependencies between the channels. This improves the feature discrimination capability and prediction accuracy of the defect detection system. The incorporation of FRN structure makes YOLOv-5s architecture to selectively enhance the necessary features and discard the unwanted ones. Therefore, the proposed novel method will efficiently detect and classify the metal surface defects such as crazing, patches, inclusions, scratches, pitted surfaces and rolled in scale. North Eastern University Surface Defect Database (NEU-DET) has been used to train and test the proposed architectural model. The suggested system has been compared with alternative models based on several performance matrices such as precision, recall and Mean Average Precision (mAP). It is observed that the proposed YOLOv-5s-FRN architecture provides significant performance improvement than state-of-the-art methods. The proposed system has been provided satisfactory results by means of improvement in mAP and time consumption. The proposed model has delivered value of mAP_0.5 as 98.05% and that of mAP_0.5:0.95 as 89.03%.","PeriodicalId":161392,"journal":{"name":"The International Arab Journal of Information Technology","volume":"10 2","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139129337","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
ERDAP: A Novel Method of Event Relation Data Augmentation Based on Relation Prediction ERDAP:基于关系预测的事件关系数据扩充新方法
Pub Date : 2024-01-01 DOI: 10.34028/iajit/21/1/6
Ruijuan Hu, Yue Chen, Haiyan Liu
Event relation extraction is a key aspect in the fields of event evolutionary graph construction, knowledge question and answer, and intelligence analysis, etc. Currently, supervised learning methods that rely on large amounts of labeled data are mostly used; however, the size of existing event relation datasets is small and cannot provide sufficient training data for the models. To alleviate this challenging research question, this study proposes a novel data augmentation model, called Event Relation Data Augmentation based on relationship Prediction (ERDAP), that allows both semantic and structural features to be taken into account without changing the semantic relation label compatibility, uses event relation graph convolutional neural networks to predict event relations, and expands the generated high-quality event relation triples as new training data for the event relation texts. Experimental evaluation using event causality extraction method on Chinese emergent event dataset shows that our model significantly outperforms existing text augmentation methods and achieves desirable performance, which provides a new idea for event relation data augmentation
事件关系提取是事件演化图构建、知识问答、情报分析等领域的关键环节。然而,现有的事件关系数据集规模较小,无法为模型提供足够的训练数据。为了解决这个具有挑战性的研究问题,本研究提出了一种新颖的数据增强模型,称为基于关系预测的事件关系数据增强(ERDAP),该模型允许在不改变语义关系标签兼容性的情况下同时考虑语义和结构特征,使用事件关系图卷积神经网络预测事件关系,并将生成的高质量事件关系三元组扩展为事件关系文本的新训练数据。利用事件因果关系提取方法在中国突发事件数据集上进行的实验评估表明,我们的模型明显优于现有的文本扩充方法,并取得了理想的性能,为事件关系数据扩充提供了一种新思路。
{"title":"ERDAP: A Novel Method of Event Relation Data Augmentation Based on Relation Prediction","authors":"Ruijuan Hu, Yue Chen, Haiyan Liu","doi":"10.34028/iajit/21/1/6","DOIUrl":"https://doi.org/10.34028/iajit/21/1/6","url":null,"abstract":"Event relation extraction is a key aspect in the fields of event evolutionary graph construction, knowledge question and answer, and intelligence analysis, etc. Currently, supervised learning methods that rely on large amounts of labeled data are mostly used; however, the size of existing event relation datasets is small and cannot provide sufficient training data for the models. To alleviate this challenging research question, this study proposes a novel data augmentation model, called Event Relation Data Augmentation based on relationship Prediction (ERDAP), that allows both semantic and structural features to be taken into account without changing the semantic relation label compatibility, uses event relation graph convolutional neural networks to predict event relations, and expands the generated high-quality event relation triples as new training data for the event relation texts. Experimental evaluation using event causality extraction method on Chinese emergent event dataset shows that our model significantly outperforms existing text augmentation methods and achieves desirable performance, which provides a new idea for event relation data augmentation","PeriodicalId":161392,"journal":{"name":"The International Arab Journal of Information Technology","volume":"9 14","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139129675","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Comprehensive Microservice Extraction Approach Integrating Business Functions and Database Entities 整合业务功能和数据库实体的综合微服务提取方法
Pub Date : 2024-01-01 DOI: 10.34028/iajit/21/1/3
Deepali Bajaj, Anita Goel, Suresh Gupta
Cloud application practitioners are building large-scale enterprise applications as microservices, to leverage scalability, performance, and availability. Microservices architecture allows a large monolithic application to be split into small, loosely coupled services. A service communicates with other services using lightweight protocols such as RESTful APIs. Extracting microservices from the monolith is a challenging task and is mostly performed manually by system architects based on their skills. This extraction involves both: 1) Partitioning of business logic, 2) Partitioning of database. For partitioning of business logic, the existing research studies focus on decomposition by considering the dependencies in the application at the class-level. However, with the passage of time, monolith application classes outgrow their size defying the Single Responsibility Principle (SRP). So, there is a need to consider the code within the classes when identifying microservices. Current studies also lack the partitioning of database and ignore the mapping of Database Entities (DE) to the microservices. In this paper, we present a Comprehensive Microservice Extraction Approach (CMEA) that considers: 1) Both classes and their methods to define and refine microservices, 2) Associate the DE to microservices using newly devised eight guiding rules handling ownership conflicts. This approach has been applied to three benchmark web applications implemented in Java and one in-house application implemented in both Java and Python. Our results demonstrate better or similar software quality attributes in comparison to the existing related studies. CMEA improves software quality attributes by 22%. System architects can easily identify microservices along with their DE using our approach. The CMEA is generic and language-independent so it can be used for any application
云应用程序从业人员正在以微服务的形式构建大型企业应用程序,以充分利用可扩展性、性能和可用性。微服务架构允许将大型单体应用程序拆分为小型、松散耦合的服务。服务使用 RESTful API 等轻量级协议与其他服务通信。从单体中提取微服务是一项具有挑战性的任务,大多由系统架构师根据自身技能手动完成。这种提取涉及两个方面:1)业务逻辑分区;2)数据库分区。对于业务逻辑的分区,现有的研究侧重于通过考虑类级应用中的依赖关系来进行分解。然而,随着时间的推移,单体应用类的规模会不断扩大,从而违背单一责任原则(SRP)。因此,在识别微服务时,需要考虑类内的代码。目前的研究还缺乏数据库分区,忽略了数据库实体(DE)与微服务的映射。在本文中,我们提出了一种全面的微服务抽取方法(CMEA),它考虑到了以下几点:1)同时使用类及其方法来定义和完善微服务;2)使用新设计的八条处理所有权冲突的指导规则将数据库实体与微服务关联起来。这种方法已应用于三个用 Java 实现的基准网络应用程序和一个用 Java 和 Python 实现的内部应用程序。与现有的相关研究相比,我们的研究结果表明了更好或相似的软件质量属性。CMEA 将软件质量属性提高了 22%。使用我们的方法,系统架构师可以轻松识别微服务及其 DE。CMEA 是通用的,与语言无关,因此可用于任何应用程序。
{"title":"A Comprehensive Microservice Extraction Approach Integrating Business Functions and Database Entities","authors":"Deepali Bajaj, Anita Goel, Suresh Gupta","doi":"10.34028/iajit/21/1/3","DOIUrl":"https://doi.org/10.34028/iajit/21/1/3","url":null,"abstract":"Cloud application practitioners are building large-scale enterprise applications as microservices, to leverage scalability, performance, and availability. Microservices architecture allows a large monolithic application to be split into small, loosely coupled services. A service communicates with other services using lightweight protocols such as RESTful APIs. Extracting microservices from the monolith is a challenging task and is mostly performed manually by system architects based on their skills. This extraction involves both: 1) Partitioning of business logic, 2) Partitioning of database. For partitioning of business logic, the existing research studies focus on decomposition by considering the dependencies in the application at the class-level. However, with the passage of time, monolith application classes outgrow their size defying the Single Responsibility Principle (SRP). So, there is a need to consider the code within the classes when identifying microservices. Current studies also lack the partitioning of database and ignore the mapping of Database Entities (DE) to the microservices. In this paper, we present a Comprehensive Microservice Extraction Approach (CMEA) that considers: 1) Both classes and their methods to define and refine microservices, 2) Associate the DE to microservices using newly devised eight guiding rules handling ownership conflicts. This approach has been applied to three benchmark web applications implemented in Java and one in-house application implemented in both Java and Python. Our results demonstrate better or similar software quality attributes in comparison to the existing related studies. CMEA improves software quality attributes by 22%. System architects can easily identify microservices along with their DE using our approach. The CMEA is generic and language-independent so it can be used for any application","PeriodicalId":161392,"journal":{"name":"The International Arab Journal of Information Technology","volume":"75 4","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139128150","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Classification of Breast Cancer using Ensemble Filter Feature Selection with Triplet Attention Based Efficient Net Classifier 使用基于三重注意的高效网络分类器的集合过滤器特征选择进行乳腺癌分类
Pub Date : 2024-01-01 DOI: 10.34028/iajit/21/1/2
Madhukar Bangalore Nagaraj, Bharathi Shivanandamurthy Hiremath, Ashwin Matta Polnaya
In medical imaging, the effective detection and classification of Breast Cancer (BC) is a current research important task because of the still existing difficulty to distinguish abnormalities from normal breast tissues due to their subtle appearance and ambiguous margins and distinguish abnormalities from the normal breast. Moreover, BC detection based on an automated detection model is needed, because manual diagnosis faces problems due to cost and shortage of skilled manpower, and also takes a very long time. Using deep learning and ensemble feature selection techniques, in this paper, a novel framework is introduced for classifying BC from histopathology images. The five primary steps of the suggested framework are as follows: 1) to make the largest original dataset and then deep learning model with data augmentation to improve the learning. 2) The best features are selected by an Ensemble Filter Feature selection Method (EFFM) which combines the best feature subsets to produce the final feature subsets. 3) Then the pruned Convolution Neural Network (CNN) model is utilized to extract the optimal features. 4) Finally, the classification is done through the Triplet Attention based Efficient Network (TAENet) classifier. The suggested model produces a 98% accuracy rate after being trained and tested on two different histopathology imaging datasets including images from four different data cohorts. Subsequently, the suggested strategy outperforms the conventional ones since the ensemble filter habitually acquires the best features, and experimental results demonstrate the importance of the proposed approach
在医学影像领域,乳腺癌(BC)的有效检测和分类是当前研究的一项重要任务,因为异常乳腺组织的外观细微,边缘模糊不清,将异常乳腺组织与正常乳腺组织区分开来仍然存在困难。此外,由于人工诊断面临成本高、技术人才短缺等问题,而且耗时很长,因此需要基于自动检测模型的乳腺癌检测。本文利用深度学习和集合特征选择技术,介绍了一种从组织病理学图像中对 BC 进行分类的新型框架。建议框架的五个主要步骤如下:1) 制作最大的原始数据集,然后通过数据增强的深度学习模型来提高学习效果。2)通过集合过滤特征选择法(EFFM)选出最佳特征,该方法将最佳特征子集结合起来,产生最终特征子集。3) 然后利用剪枝卷积神经网络(CNN)模型提取最佳特征。4) 最后,通过基于三重注意的高效网络(TAENet)分类器进行分类。建议的模型在两个不同的组织病理学成像数据集(包括来自四个不同数据队列的图像)上经过训练和测试后,准确率达到 98%。随后,由于集合滤波器习惯性地获取最佳特征,因此建议的策略优于传统策略,实验结果证明了建议方法的重要性。
{"title":"Classification of Breast Cancer using Ensemble Filter Feature Selection with Triplet Attention Based Efficient Net Classifier","authors":"Madhukar Bangalore Nagaraj, Bharathi Shivanandamurthy Hiremath, Ashwin Matta Polnaya","doi":"10.34028/iajit/21/1/2","DOIUrl":"https://doi.org/10.34028/iajit/21/1/2","url":null,"abstract":"In medical imaging, the effective detection and classification of Breast Cancer (BC) is a current research important task because of the still existing difficulty to distinguish abnormalities from normal breast tissues due to their subtle appearance and ambiguous margins and distinguish abnormalities from the normal breast. Moreover, BC detection based on an automated detection model is needed, because manual diagnosis faces problems due to cost and shortage of skilled manpower, and also takes a very long time. Using deep learning and ensemble feature selection techniques, in this paper, a novel framework is introduced for classifying BC from histopathology images. The five primary steps of the suggested framework are as follows: 1) to make the largest original dataset and then deep learning model with data augmentation to improve the learning. 2) The best features are selected by an Ensemble Filter Feature selection Method (EFFM) which combines the best feature subsets to produce the final feature subsets. 3) Then the pruned Convolution Neural Network (CNN) model is utilized to extract the optimal features. 4) Finally, the classification is done through the Triplet Attention based Efficient Network (TAENet) classifier. The suggested model produces a 98% accuracy rate after being trained and tested on two different histopathology imaging datasets including images from four different data cohorts. Subsequently, the suggested strategy outperforms the conventional ones since the ensemble filter habitually acquires the best features, and experimental results demonstrate the importance of the proposed approach","PeriodicalId":161392,"journal":{"name":"The International Arab Journal of Information Technology","volume":"46 5","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139126993","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
XAI-PDF: A Robust Framework for Malicious PDF Detection Leveraging SHAP-Based Feature Engineering XAI-PDF:利用基于 SHAP 的特征工程进行恶意 PDF 检测的稳健框架
Pub Date : 2024-01-01 DOI: 10.34028/iajit/21/1/12
Mustafa Al-Fayoumi, Q. Abu Al-haija, Rakan Armoush, Christine Amareen
With the increasing number of malicious PDF files used for cyberattacks, it is essential to develop efficient and accurate classifiers to detect and prevent these threats. Machine Learning (ML) models have successfully detected malicious PDF files. This paper presents XAI-PDF, an efficient system for malicious PDF detection designed to enhance accuracy and minimize decision-making time on a modern dataset, the Evasive-PDFMal2022 dataset. The proposed method optimizes malicious PDF classifier performance by employing feature engineering guided by Shapley Additive Explanations (SHAP). Particularly, the model development approach comprises four phases: data preparation, model building, explainability of the models, and derived features. Utilizing the interpretability of SHAP values, crucial features are identified, and new ones are generated, resulting in an improved classification model that showcases the effectiveness of interpretable AI techniques in enhancing model performance. Various interpretable ML models were implemented, with the Lightweight Gradient Boosting Machine (LGBM) outperforming other classifiers. The Explainable Artificial Intelligence (XAI) global surrogate model generated explanations for LGBM predictions. Experimental comparisons of XAI-PDF with baseline methods revealed its superiority in achieving higher accuracy, precision, and F1-scores with minimal False Positive (FP) and False Negative (FN) rates (99.9%, 100%, 99.89%,0.000, and 0.002, respectively). Additionally, XAI-PDF requires only 1.36 milliseconds per record for predictions, demonstrating increased resilience in detecting evasive malicious PDF files compared to state-of-the-art methods
随着用于网络攻击的恶意 PDF 文件数量不断增加,开发高效、准确的分类器来检测和预防这些威胁至关重要。机器学习(ML)模型已成功检测出恶意 PDF 文件。本文介绍了 XAI-PDF,这是一种高效的恶意 PDF 检测系统,旨在提高准确性并最大限度地减少现代数据集(Evasive-PDFMal2022 数据集)上的决策时间。所提出的方法通过采用以 Shapley Additive Explanations (SHAP) 为指导的特征工程来优化恶意 PDF 分类器的性能。特别是,模型开发方法包括四个阶段:数据准备、模型构建、模型的可解释性和衍生特征。利用 SHAP 值的可解释性,确定关键特征并生成新特征,从而改进分类模型,展示可解释人工智能技术在提高模型性能方面的有效性。我们实施了各种可解释的 ML 模型,其中轻量级梯度提升机(LGBM)的表现优于其他分类器。可解释人工智能(XAI)全局代理模型为 LGBM 预测生成了解释。XAI-PDF 与基线方法的实验比较显示,XAI-PDF 在实现更高的准确度、精确度和 F1 分数以及最低的假阳性(FP)和假阴性(FN)率(分别为 99.9%、100%、99.89%、0.000 和 0.002)方面更具优势。此外,XAI-PDF 的每条预测记录仅需 1.36 毫秒,这表明与最先进的方法相比,XAI-PDF 在检测躲避性恶意 PDF 文件方面具有更强的适应性。
{"title":"XAI-PDF: A Robust Framework for Malicious PDF Detection Leveraging SHAP-Based Feature Engineering","authors":"Mustafa Al-Fayoumi, Q. Abu Al-haija, Rakan Armoush, Christine Amareen","doi":"10.34028/iajit/21/1/12","DOIUrl":"https://doi.org/10.34028/iajit/21/1/12","url":null,"abstract":"With the increasing number of malicious PDF files used for cyberattacks, it is essential to develop efficient and accurate classifiers to detect and prevent these threats. Machine Learning (ML) models have successfully detected malicious PDF files. This paper presents XAI-PDF, an efficient system for malicious PDF detection designed to enhance accuracy and minimize decision-making time on a modern dataset, the Evasive-PDFMal2022 dataset. The proposed method optimizes malicious PDF classifier performance by employing feature engineering guided by Shapley Additive Explanations (SHAP). Particularly, the model development approach comprises four phases: data preparation, model building, explainability of the models, and derived features. Utilizing the interpretability of SHAP values, crucial features are identified, and new ones are generated, resulting in an improved classification model that showcases the effectiveness of interpretable AI techniques in enhancing model performance. Various interpretable ML models were implemented, with the Lightweight Gradient Boosting Machine (LGBM) outperforming other classifiers. The Explainable Artificial Intelligence (XAI) global surrogate model generated explanations for LGBM predictions. Experimental comparisons of XAI-PDF with baseline methods revealed its superiority in achieving higher accuracy, precision, and F1-scores with minimal False Positive (FP) and False Negative (FN) rates (99.9%, 100%, 99.89%,0.000, and 0.002, respectively). Additionally, XAI-PDF requires only 1.36 milliseconds per record for predictions, demonstrating increased resilience in detecting evasive malicious PDF files compared to state-of-the-art methods","PeriodicalId":161392,"journal":{"name":"The International Arab Journal of Information Technology","volume":"43 8","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139126539","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
The Coupling of a Multiplicative Group and the Theory of Chaos in the Encryptions of Images 图像加密中的乘法群耦合与混沌理论
Pub Date : 2024-01-01 DOI: 10.34028/iajit/21/1/1
F. ElAzzaby, N. Elakkad, Khalid Sabour
This study is centered on a significant scientific contribution within the realm of image cryptography. The chosen approach involves employing the bidimensional Arnold Cat Map transformation to reposition and modify pixel locations, guided by parameters derived from the original image. The construction of the multiplicative group Z/nZ, comprising equivalence classes modulo n, relies on a hyper-chaotic sequence derived from the 2D sinusoidal logistic modulation map. The correlation between this sequence and the preceding step yields an unpredictable blurring pattern, effectively altering the statistical properties of resulting matrices and distributing the influence of individual bits across the entire encrypted image. For each pixel, the encryption process entails an XOR operation with the Z/nZ group, followed by a right shift based on the three Least Significant Bits (LSB) of the preceding pixel. This meticulous procedure is iterated for every pixel, leaving no trace of similarity or association with the original plaintext image, effectively rendering it blurred and indecipherable. To gauge the efficacy of our algorithm, we subjected it to thorough evaluation utilizing diverse criteria, including histogram analysis, which unveiled a nearly uniform pattern in the encrypted images. Entropy values were found to be close to 8, while the correlation analysis exhibited a pronounced proximity to 0. Moreover, we subjected our approach to differential attacks, and the calculated values of the Number of Changing Pixel Rate (NPCR > 99.6) and the Unified Averaged Changed Intensity (UACI > 33.2) corroborated the strength and resilience of our methodology. In addition, to establish its comparative standing, we undertook a comprehensive assessment, meticulously comparing our method to various existing approaches from the literature, including those proposed by Hua, Es-sabry, and Faragallah. This systematic process accentuated the high level of responsiveness and sensitivity inherent in our approach, thus underscoring its innovative and promising nature
这项研究的核心是在图像密码学领域做出重大科学贡献。所选择的方法包括利用二维阿诺德猫图变换,在原始图像参数的引导下,重新定位和修改像素位置。乘法组 Z/nZ 的构建由 modulo n 的等价类组成,它依赖于从二维正弦对数调制图衍生出的超混沌序列。该序列与前一步骤之间的相关性产生了一种不可预测的模糊模式,有效地改变了结果矩阵的统计特性,并在整个加密图像中分散了各个比特的影响。对于每个像素,加密过程都需要与 Z/nZ 组进行 XOR 运算,然后根据前一个像素的三个最小有效位(LSB)进行右移。每一个像素都要重复这一细致的过程,不留任何与原始明文图像相似或关联的痕迹,从而有效地使原始明文图像变得模糊不清、难以破译。为了衡量我们算法的有效性,我们利用不同的标准对其进行了全面评估,包括直方图分析,该分析揭示了加密图像中几乎一致的模式。熵值接近 8,而相关性分析则明显接近 0。此外,我们还对我们的方法进行了差分攻击,计算得出的像素变化率(NPCR > 99.6)和统一平均变化强度(UACI > 33.2)值证实了我们方法的强度和弹性。此外,为了确定其比较地位,我们进行了全面评估,将我们的方法与文献中现有的各种方法进行了细致比较,包括 Hua、Es-sabry 和 Faragallah 提出的方法。这一系统化的过程凸显了我们的方法所固有的高度响应性和敏感性,从而强调了其创新性和前景广阔的性质
{"title":"The Coupling of a Multiplicative Group and the Theory of Chaos in the Encryptions of Images","authors":"F. ElAzzaby, N. Elakkad, Khalid Sabour","doi":"10.34028/iajit/21/1/1","DOIUrl":"https://doi.org/10.34028/iajit/21/1/1","url":null,"abstract":"This study is centered on a significant scientific contribution within the realm of image cryptography. The chosen approach involves employing the bidimensional Arnold Cat Map transformation to reposition and modify pixel locations, guided by parameters derived from the original image. The construction of the multiplicative group Z/nZ, comprising equivalence classes modulo n, relies on a hyper-chaotic sequence derived from the 2D sinusoidal logistic modulation map. The correlation between this sequence and the preceding step yields an unpredictable blurring pattern, effectively altering the statistical properties of resulting matrices and distributing the influence of individual bits across the entire encrypted image. For each pixel, the encryption process entails an XOR operation with the Z/nZ group, followed by a right shift based on the three Least Significant Bits (LSB) of the preceding pixel. This meticulous procedure is iterated for every pixel, leaving no trace of similarity or association with the original plaintext image, effectively rendering it blurred and indecipherable. To gauge the efficacy of our algorithm, we subjected it to thorough evaluation utilizing diverse criteria, including histogram analysis, which unveiled a nearly uniform pattern in the encrypted images. Entropy values were found to be close to 8, while the correlation analysis exhibited a pronounced proximity to 0. Moreover, we subjected our approach to differential attacks, and the calculated values of the Number of Changing Pixel Rate (NPCR > 99.6) and the Unified Averaged Changed Intensity (UACI > 33.2) corroborated the strength and resilience of our methodology. In addition, to establish its comparative standing, we undertook a comprehensive assessment, meticulously comparing our method to various existing approaches from the literature, including those proposed by Hua, Es-sabry, and Faragallah. This systematic process accentuated the high level of responsiveness and sensitivity inherent in our approach, thus underscoring its innovative and promising nature","PeriodicalId":161392,"journal":{"name":"The International Arab Journal of Information Technology","volume":"44 7","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139127951","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Cohesive Pair-Wises Constrained Deep Embedding for Semi-Supervised Clustering with Very Few Labeled Samples* 利用极少量标记样本进行半监督聚类的内聚性对维约束深度嵌入技术*
Pub Date : 2024-01-01 DOI: 10.34028/iajit/21/1/7
Zhang Jing, Guiyan Wei, Yonggong Ren
Semi-supervised learning is a powerful paradigm for excavating latent structures of between labeled and unlabeled samples under the view of models constructing. Currently, graph-based models solve the approximate matrix that directly represent distributions of samples by the spatial metric. The crux lies in optimizing connections of samples, which is achieved by subjecting to must-links or cannot-links. Unfortunately, to find links are rather difficult for semi-supervised clustering with very few labeled samples, therefore, significantly impairs the robustness and accuracy in such scenario. To address this problem, we propose the Cohesive Pair-wises Constrained deep Embedding model (CPCE) to obtain an optimal embedding for representing the category distribution of samples and avoid the failed graph-structure of the global samples. CPCE designs the deep network framework based on CNN-Autoencoder by minimizing reconstruct errors of samples, and build up constrains both of the sample distribution for within-class and the category distribution for intra-class to optimal the latent embedding. Then, leverage the strong supervised information obtained from cohesive pair-wises to pull samples into within-class, which avoid the similarity of high-dimension features located in different categories to achieve more the compact solution. We demonstrate the proposed method in popular datasets and compare the superiority with popular methods
半监督学习(Semi-supervised learning)是在模型构建视角下挖掘已标注和未标注样本之间潜在结构的一种强大范式。目前,基于图的模型解决了用空间度量直接表示样本分布的近似矩阵问题。其关键在于优化样本的连接,通过必须连接或不能连接来实现。遗憾的是,对于只有极少数标注样本的半监督聚类来说,要找到链接是相当困难的,因此在这种情况下会严重影响鲁棒性和准确性。为了解决这个问题,我们提出了内聚对智约束深度嵌入模型(Cohesive Pair-wises Constrained deep Embedding model,CPCE),以获得代表样本类别分布的最优嵌入,避免全局样本的图结构失效。CPCE 通过最小化样本重构误差来设计基于 CNN-Autoencoder 的深度网络框架,并同时对类内样本分布和类内类别分布建立约束,以优化潜在嵌入。然后,利用从内聚配对中获得的强监督信息,将样本拉入类内,从而避免了位于不同类别中的高维特征的相似性,实现了更紧凑的解决方案。我们在流行的数据集中演示了所提出的方法,并与流行的方法比较了其优越性
{"title":"Cohesive Pair-Wises Constrained Deep Embedding for Semi-Supervised Clustering with Very Few Labeled Samples*","authors":"Zhang Jing, Guiyan Wei, Yonggong Ren","doi":"10.34028/iajit/21/1/7","DOIUrl":"https://doi.org/10.34028/iajit/21/1/7","url":null,"abstract":"Semi-supervised learning is a powerful paradigm for excavating latent structures of between labeled and unlabeled samples under the view of models constructing. Currently, graph-based models solve the approximate matrix that directly represent distributions of samples by the spatial metric. The crux lies in optimizing connections of samples, which is achieved by subjecting to must-links or cannot-links. Unfortunately, to find links are rather difficult for semi-supervised clustering with very few labeled samples, therefore, significantly impairs the robustness and accuracy in such scenario. To address this problem, we propose the Cohesive Pair-wises Constrained deep Embedding model (CPCE) to obtain an optimal embedding for representing the category distribution of samples and avoid the failed graph-structure of the global samples. CPCE designs the deep network framework based on CNN-Autoencoder by minimizing reconstruct errors of samples, and build up constrains both of the sample distribution for within-class and the category distribution for intra-class to optimal the latent embedding. Then, leverage the strong supervised information obtained from cohesive pair-wises to pull samples into within-class, which avoid the similarity of high-dimension features located in different categories to achieve more the compact solution. We demonstrate the proposed method in popular datasets and compare the superiority with popular methods","PeriodicalId":161392,"journal":{"name":"The International Arab Journal of Information Technology","volume":"85 21","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139125135","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Scrupulous SCGAN Framework for Recognition of Restored Images with Caffe based PCA Filtration 利用基于 Caffe 的 PCA 过滤技术识别修复图像的 Scrupulous SCGAN 框架
Pub Date : 2024-01-01 DOI: 10.34028//iajit/21/1/10
Khushboo Agarwal, Manish Dixit
Computer vision enables to detect many objects in any scenario which helps in various real time application but still face recognition and detection remains a tedious process due to the low resolution, blurriness, noise, diverse pose and expression and occlusions. This proposal develops a novel scrupulous Standardized Convolute Generative Adversarial Network (SCGAN) framework for performing accurate face recognition automatically by restoring the occluded region including blind face restoration. Initially, a scrupulous image refining technique is utilised to offer the appropriate input to the network in the subsequent process. Following the pre-processing stage, a Caffe based Principle Component Analysis (PCA) filtration is conducted which uses convolutional architecture for fast feature embedding that collects spatial information and significant differentiating characteristics to counteract the loss of information existing in pooling operations. Then a filtration method identifies the specific match of the face based on the extracted features, creating uncorrelated variables that optimise variance across time while minimising information loss. To handle all the diversification occurring in the image and accurately recognise the face with occlusion in any part of the face, a novel Standardized Convolute GAN network is used to restore the image and recognise the face using novel Generative Adversarial Network (GAN) networks are modelled. This GAN ensures the normal distribution along with parametric optimization contributing to the high performance with accuracy of 96.05% and Peak Signal to Noise Ratio (PSNR) of 18 and Structural Similarity Index Metric (SSIM) of 98% for restored face recognition. Thus, the performance of the framework based on properly recognizing the face from the generated images is evaluated and discussed.
计算机视觉能够检测任何场景中的许多物体,这有助于各种实时应用,但由于分辨率低、模糊、噪声、姿势和表情各异以及遮挡等原因,人脸识别和检测仍然是一个繁琐的过程。本提案开发了一种新颖的无差别标准化卷积生成对抗网络(SCGAN)框架,通过恢复遮挡区域(包括盲目的人脸修复)自动执行准确的人脸识别。首先,利用严格的图像细化技术,在后续过程中为网络提供适当的输入。在预处理阶段之后,进行基于 Caffe 的主成分分析(PCA)过滤,该过滤使用卷积架构进行快速特征嵌入,收集空间信息和重要的区分特征,以抵消池化操作中存在的信息损失。然后,过滤方法根据提取的特征识别人脸的特定匹配,创建不相关的变量,优化跨时间的方差,同时最大限度地减少信息丢失。为了处理图像中出现的所有多样化情况,并在人脸的任何部分出现闭塞的情况下准确识别人脸,使用了一种新颖的标准化卷积 GAN 网络来还原图像,并使用新颖的生成对抗网络 (GAN) 网络模型来识别人脸。该 GAN 网络确保了正态分布,并对参数进行了优化,从而使还原的人脸识别准确率达到 96.05%,峰值信噪比(PSNR)达到 18%,结构相似度指标(SSIM)达到 98%。因此,我们对基于从生成的图像中正确识别人脸的框架的性能进行了评估和讨论。
{"title":"Scrupulous SCGAN Framework for Recognition of Restored Images with Caffe based PCA Filtration","authors":"Khushboo Agarwal, Manish Dixit","doi":"10.34028//iajit/21/1/10","DOIUrl":"https://doi.org/10.34028//iajit/21/1/10","url":null,"abstract":"Computer vision enables to detect many objects in any scenario which helps in various real time application but still face recognition and detection remains a tedious process due to the low resolution, blurriness, noise, diverse pose and expression and occlusions. This proposal develops a novel scrupulous Standardized Convolute Generative Adversarial Network (SCGAN) framework for performing accurate face recognition automatically by restoring the occluded region including blind face restoration. Initially, a scrupulous image refining technique is utilised to offer the appropriate input to the network in the subsequent process. Following the pre-processing stage, a Caffe based Principle Component Analysis (PCA) filtration is conducted which uses convolutional architecture for fast feature embedding that collects spatial information and significant differentiating characteristics to counteract the loss of information existing in pooling operations. Then a filtration method identifies the specific match of the face based on the extracted features, creating uncorrelated variables that optimise variance across time while minimising information loss. To handle all the diversification occurring in the image and accurately recognise the face with occlusion in any part of the face, a novel Standardized Convolute GAN network is used to restore the image and recognise the face using novel Generative Adversarial Network (GAN) networks are modelled. This GAN ensures the normal distribution along with parametric optimization contributing to the high performance with accuracy of 96.05% and Peak Signal to Noise Ratio (PSNR) of 18 and Structural Similarity Index Metric (SSIM) of 98% for restored face recognition. Thus, the performance of the framework based on properly recognizing the face from the generated images is evaluated and discussed.","PeriodicalId":161392,"journal":{"name":"The International Arab Journal of Information Technology","volume":"39 18","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139125635","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Healthcare Data Security in Cloud Storage Using Light Weight Symmetric Key Algorithm 使用轻量级对称密钥算法确保云存储中的医疗保健数据安全
Pub Date : 2024-01-01 DOI: 10.34028/iajit/21/1/5
Vimala Devi Parthasarathy, Kalaichelvi Viswalingam
Now-a-days, healthcare monitoring system is very much important system in the medical field to know the patient‘s health status immediately. In the proposed system, the sensors are fixed over the patient’s body or placed at some distances around the body to collect patient’s significant parameters like blood pressure, temperature, heart beat rate, etc. These parameters are collected by the healthcare professionals through some connectivity mechanisms like Bluetooth, ZigBee, etc. These significant data will be outsourced to the cloud storage in a secure way to avoid attack from the attackers. So, we need some protection mechanism to safeguard this information. This article proposes a light weight cryptographic algorithm (symmetric key) via random number key generation using Diffie-Hellman key exchange based on Elliptic Curve (ECDH) cryptography. As a result of substituting bytes (S-box) and folding (horizontal and vertical) operations, the proposed symmetric key algorithm achieves the foremost properties of cryptography such as confusion and diffusion quite well. Experimental results showed that the overall execution time of the proposed algorithm is superior to the standard Advanced Encryption Standard (AES) algorithm. The throughput rate of the proposed algorithm is 20.525095 KB/seconds whereas for the standard AES algorithm throughput rate is 18.727215 KB/seconds. So, the proposed algorithm is faster than the existing AES algorithm. Moreover, the construction of S-box, IS-box and the key generation procedures are entirely different in the proposed algorithm so, it increases the complexity for the attacker and it will create confusion to the attacker
如今,医疗保健监测系统是医疗领域中非常重要的系统,可以即时了解病人的健康状况。在提议的系统中,传感器被固定在病人身体上或放置在身体周围一定距离的地方,以收集病人的重要参数,如血压、体温、心跳速率等。医护人员通过蓝牙、ZigBee 等连接机制收集这些参数。这些重要数据将以安全的方式外包给云存储,以避免攻击者的攻击。因此,我们需要一些保护机制来保护这些信息。本文提出了一种轻量级加密算法(对称密钥),通过使用基于椭圆曲线(ECDH)加密技术的 Diffie-Hellman 密钥交换法生成随机数密钥。通过替换字节(S-box)和折叠(水平和垂直)操作,所提出的对称密钥算法很好地实现了密码学的首要特性,如混淆和扩散。实验结果表明,拟议算法的总体执行时间优于标准高级加密标准(AES)算法。建议算法的吞吐率为 20.525095 KB/秒,而标准 AES 算法的吞吐率为 18.727215 KB/秒。因此,提议的算法比现有的 AES 算法更快。此外,在拟议算法中,S 盒、IS 盒的构造和密钥生成程序完全不同,因此增加了攻击者的复杂性,并会给攻击者造成困惑。
{"title":"Healthcare Data Security in Cloud Storage Using Light Weight Symmetric Key Algorithm","authors":"Vimala Devi Parthasarathy, Kalaichelvi Viswalingam","doi":"10.34028/iajit/21/1/5","DOIUrl":"https://doi.org/10.34028/iajit/21/1/5","url":null,"abstract":"Now-a-days, healthcare monitoring system is very much important system in the medical field to know the patient‘s health status immediately. In the proposed system, the sensors are fixed over the patient’s body or placed at some distances around the body to collect patient’s significant parameters like blood pressure, temperature, heart beat rate, etc. These parameters are collected by the healthcare professionals through some connectivity mechanisms like Bluetooth, ZigBee, etc. These significant data will be outsourced to the cloud storage in a secure way to avoid attack from the attackers. So, we need some protection mechanism to safeguard this information. This article proposes a light weight cryptographic algorithm (symmetric key) via random number key generation using Diffie-Hellman key exchange based on Elliptic Curve (ECDH) cryptography. As a result of substituting bytes (S-box) and folding (horizontal and vertical) operations, the proposed symmetric key algorithm achieves the foremost properties of cryptography such as confusion and diffusion quite well. Experimental results showed that the overall execution time of the proposed algorithm is superior to the standard Advanced Encryption Standard (AES) algorithm. The throughput rate of the proposed algorithm is 20.525095 KB/seconds whereas for the standard AES algorithm throughput rate is 18.727215 KB/seconds. So, the proposed algorithm is faster than the existing AES algorithm. Moreover, the construction of S-box, IS-box and the key generation procedures are entirely different in the proposed algorithm so, it increases the complexity for the attacker and it will create confusion to the attacker","PeriodicalId":161392,"journal":{"name":"The International Arab Journal of Information Technology","volume":"60 9","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139126903","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Dual-Objective Approach for Allocation of Virtual Machine with improved Job Scheduling in Cloud Computing 在云计算中分配虚拟机并改进作业调度的双目标方法
Pub Date : 2024-01-01 DOI: 10.34028/iajit/21/1/4
Sandeep Sutar, Manjunathswamy Byranahallieraiah, Kumarswamy Shivashankaraiah
In Cloud Computing (CC) environment, requests of user are maintained via workloads that are allocated to Virtual Machines (VMs) using scheduling techniques which primarily focus on reducing the time for processing by generating efficient schedules of smaller lengths. The efficient processing of requests also needs larger usage of resources that incurs higher overhead in the form of utilization of energy and optimization of cost utilized by Physical Machines (PMs). Assignment of VMs optimally in the environment of CC for jobs submitted by users is a challenge. In order to obtain better solution involving scheduling of jobs to VMs, considering two parameters utilization of energy and cost, we present a dual-objective approach for VM allocation with improved scheduling of jobs in CC environment. The proposed work aimed to build a dual-objective scheduling model for improved job scheduling, focusing on minimization of cost and utilization of energy at a time. For evaluating performance of dual-objective approach, we utilized two types of benchmark datasets and compared with existing approaches such as Whale, Artificial Bee Colony (ABC), Particle Swarm Optimization (PSO) and Metaheuristic Dynamic VM Allocation (MDVMA) techniques. The results obtained from simulation demonstrated that dual-objective approach performs better in the form of minimization of utilization of energy and cost
在云计算(CC)环境中,用户的请求通过使用调度技术分配给虚拟机(VM)的工作负载来维持,调度技术主要侧重于通过生成长度较小的高效调度来减少处理时间。高效处理请求还需要使用更多的资源,从而产生更高的能源利用率和物理机(PM)成本优化形式的开销。在 CC 环境中为用户提交的任务优化分配虚拟机是一项挑战。为了在考虑能源利用率和成本这两个参数的情况下获得更好的解决方案,我们提出了一种双目标方法,用于在 CC 环境中改进作业调度的虚拟机分配。所提出的工作旨在为改进作业调度建立一个双目标调度模型,同时关注成本最小化和能源利用率。为了评估双目标方法的性能,我们使用了两种类型的基准数据集,并与现有的方法进行了比较,如 Whale、人工蜂群(ABC)、粒子群优化(PSO)和元启发式动态虚拟机分配(MDVMA)技术。模拟结果表明,双目标方法在能源利用率和成本最小化方面表现更好。
{"title":"A Dual-Objective Approach for Allocation of Virtual Machine with improved Job Scheduling in Cloud Computing","authors":"Sandeep Sutar, Manjunathswamy Byranahallieraiah, Kumarswamy Shivashankaraiah","doi":"10.34028/iajit/21/1/4","DOIUrl":"https://doi.org/10.34028/iajit/21/1/4","url":null,"abstract":"In Cloud Computing (CC) environment, requests of user are maintained via workloads that are allocated to Virtual Machines (VMs) using scheduling techniques which primarily focus on reducing the time for processing by generating efficient schedules of smaller lengths. The efficient processing of requests also needs larger usage of resources that incurs higher overhead in the form of utilization of energy and optimization of cost utilized by Physical Machines (PMs). Assignment of VMs optimally in the environment of CC for jobs submitted by users is a challenge. In order to obtain better solution involving scheduling of jobs to VMs, considering two parameters utilization of energy and cost, we present a dual-objective approach for VM allocation with improved scheduling of jobs in CC environment. The proposed work aimed to build a dual-objective scheduling model for improved job scheduling, focusing on minimization of cost and utilization of energy at a time. For evaluating performance of dual-objective approach, we utilized two types of benchmark datasets and compared with existing approaches such as Whale, Artificial Bee Colony (ABC), Particle Swarm Optimization (PSO) and Metaheuristic Dynamic VM Allocation (MDVMA) techniques. The results obtained from simulation demonstrated that dual-objective approach performs better in the form of minimization of utilization of energy and cost","PeriodicalId":161392,"journal":{"name":"The International Arab Journal of Information Technology","volume":"22 20","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139127187","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
The International Arab Journal of Information Technology
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1