首页 > 最新文献

IET Software最新文献

英文 中文
Unsupervised Person Reidentification Using Stripe-Driven Fusion Transformer Network 基于条纹驱动融合变压器网络的无监督人员再识别
IF 1.3 4区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-10-22 DOI: 10.1049/sfw2/6394038
Zeyu Zang, Yang Liu, Shuang Liu, Zhong Zhang, Xinshan Zhu

In recent years, some methods utilize a transformer as the backbone to model the long-range context dependencies, reflecting a prevailing trend in unsupervised person reidentification (Re-ID) tasks. However, they only explore the global information through interactive learning in the framework of the transformer, which ignores the learning of the part information in the interaction process for pedestrian images. In this study, we present a novel transformer network for unsupervised person Re-ID, a stripe-driven fusion transformer (SDFT), designed to simultaneously capture the global interaction and the part interaction when modeling the long-range context dependencies. Meanwhile, we present a stripe-driven regularization (SDR) to constrain the part aggregation features and the global features by considering the consistency principle from the aspects of the features and the clusters, aiming to improve the representational capacity of the features. Furthermore, to investigate the relationships between local regions of pedestrian images, we present a stripe-driven contrastive loss (SDCL) to learn discriminative part features from the perspectives of pedestrian identity and stripes. The proposed method has undergone extensive validations on publicly available unsupervised person Re-ID benchmarks, and the experimental results confirm its superiority and effectiveness.

近年来,一些方法利用转换器作为主干来对远程上下文依赖性进行建模,这反映了无监督人员重新识别(Re-ID)任务的流行趋势。然而,他们只是在变压器的框架中通过交互学习来探索全局信息,而忽略了行人图像交互过程中局部信息的学习。在这项研究中,我们提出了一种新的无监督人Re-ID变压器网络,一种条带驱动的融合变压器(SDFT),设计用于在建模远程上下文依赖时同时捕获全局交互和部分交互。同时,从特征和聚类两个方面考虑一致性原则,提出了一种条带驱动正则化(SDR)方法来约束零件聚集特征和全局特征,以提高特征的表示能力。此外,为了研究行人图像局部区域之间的关系,我们提出了一种条纹驱动的对比损失(SDCL)方法,从行人身份和条纹的角度学习区分部分特征。该方法在公开的无监督人身份识别基准上进行了大量的验证,实验结果证实了该方法的优越性和有效性。
{"title":"Unsupervised Person Reidentification Using Stripe-Driven Fusion Transformer Network","authors":"Zeyu Zang,&nbsp;Yang Liu,&nbsp;Shuang Liu,&nbsp;Zhong Zhang,&nbsp;Xinshan Zhu","doi":"10.1049/sfw2/6394038","DOIUrl":"https://doi.org/10.1049/sfw2/6394038","url":null,"abstract":"<p>In recent years, some methods utilize a transformer as the backbone to model the long-range context dependencies, reflecting a prevailing trend in unsupervised person reidentification (Re-ID) tasks. However, they only explore the global information through interactive learning in the framework of the transformer, which ignores the learning of the part information in the interaction process for pedestrian images. In this study, we present a novel transformer network for unsupervised person Re-ID, a stripe-driven fusion transformer (SDFT), designed to simultaneously capture the global interaction and the part interaction when modeling the long-range context dependencies. Meanwhile, we present a stripe-driven regularization (SDR) to constrain the part aggregation features and the global features by considering the consistency principle from the aspects of the features and the clusters, aiming to improve the representational capacity of the features. Furthermore, to investigate the relationships between local regions of pedestrian images, we present a stripe-driven contrastive loss (SDCL) to learn discriminative part features from the perspectives of pedestrian identity and stripes. The proposed method has undergone extensive validations on publicly available unsupervised person Re-ID benchmarks, and the experimental results confirm its superiority and effectiveness.</p>","PeriodicalId":50378,"journal":{"name":"IET Software","volume":"2025 1","pages":""},"PeriodicalIF":1.3,"publicationDate":"2025-10-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ietresearch.onlinelibrary.wiley.com/doi/epdf/10.1049/sfw2/6394038","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145366659","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Blockchain-Based Model to Predict Agile Software Estimation Using Machine Learning Techniques 使用机器学习技术预测敏捷软件评估的基于区块链的模型
IF 1.3 4区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-10-22 DOI: 10.1049/sfw2/9238663
Mohammad Ayub Latif, Muhammad Khalid Khan, Maaz Bin Ahmad, Toqeer Mahmood, Muhammad Tariq Mahmood, Young-Bok Joo

The importance of software estimation is utmost, as it is one of the most crucial activities for software project management. Although numerous software estimation techniques exist, the accuracy achieved by these techniques is questionable. This work studies the existing software estimation techniques for Agile software development (ASD), identifies the gap, and proposes a decentralized framework for estimation of ASD using machine-learning (ML) algorithms, which utilize the blockchain technology. The estimation model uses nearest neighbors with four ML techniques for ASD. Using an available ASD dataset, after the augmentation on the dataset, the proposed model emits results for the completion time prediction of software. Use of another popular dataset for ASD predicts the software effort using the same proposed model. The crux of the proposed model is that it simulates blockchain technology to predict the completion time and the effort of a software using ML algorithms. This type of estimation model, using ML, making use of blockchain technology, does not exist in the literature, and this is the core novelty of this proposed model. The final prediction of the software effort integrates another technique for improving the calculated estimation, the standard deviation technique proposed by the authors previously. This model helped lessening the overall mean magnitude of relative error (MMRE) of the original model from 6.82% to 1.73% for the augmented dataset of 126 projects. All four ML techniques used for the proposed model give a better p-value than the original model using statistical testing through the Wilcoxon test. The average of the MMRE for effort estimation of all four techniques is below 25% on a dataset of 136 projects. The application of the standard deviation technique further helps in lessening the MMRE of the proposed model at 70%, 80%, and 90% confidence levels. The work will give insight to researchers and experts and open the doors for new research in this area.

软件评估的重要性是最大的,因为它是软件项目管理中最关键的活动之一。尽管存在许多软件评估技术,但这些技术所达到的准确性是值得怀疑的。本文研究了敏捷软件开发(ASD)的现有软件估计技术,识别了差距,并提出了一个使用机器学习(ML)算法的分散的ASD估计框架,该框架利用区块链技术。该估计模型使用最近邻和四种ML技术对ASD进行估计。该模型利用现有的ASD数据集,对数据集进行增强后,输出结果用于软件的完成时间预测。使用另一个流行的ASD数据集来预测使用相同建议模型的软件工作量。该模型的核心是模拟区块链技术,利用ML算法预测软件的完成时间和工作量。这种使用ML,利用区块链技术的估计模型在文献中是不存在的,这是本文提出的模型的核心新颖之处。软件工作的最终预测集成了另一种改进计算估计的技术,即作者先前提出的标准偏差技术。该模型帮助126个项目的增强数据集将原始模型的总体平均相对误差幅度(MMRE)从6.82%降低到1.73%。使用通过Wilcoxon检验进行统计检验的模型中使用的所有四种ML技术都比原始模型提供了更好的p值。在136个项目的数据集上,所有四种技术的工作量估计的MMRE平均值低于25%。标准差技术的应用进一步降低了模型在70%、80%和90%置信水平下的最小最小方差。这项工作将为研究人员和专家提供见解,并为该领域的新研究打开大门。
{"title":"Blockchain-Based Model to Predict Agile Software Estimation Using Machine Learning Techniques","authors":"Mohammad Ayub Latif,&nbsp;Muhammad Khalid Khan,&nbsp;Maaz Bin Ahmad,&nbsp;Toqeer Mahmood,&nbsp;Muhammad Tariq Mahmood,&nbsp;Young-Bok Joo","doi":"10.1049/sfw2/9238663","DOIUrl":"https://doi.org/10.1049/sfw2/9238663","url":null,"abstract":"<p>The importance of software estimation is utmost, as it is one of the most crucial activities for software project management. Although numerous software estimation techniques exist, the accuracy achieved by these techniques is questionable. This work studies the existing software estimation techniques for Agile software development (ASD), identifies the gap, and proposes a decentralized framework for estimation of ASD using machine-learning (ML) algorithms, which utilize the blockchain technology. The estimation model uses nearest neighbors with four ML techniques for ASD. Using an available ASD dataset, after the augmentation on the dataset, the proposed model emits results for the completion time prediction of software. Use of another popular dataset for ASD predicts the software effort using the same proposed model. The crux of the proposed model is that it simulates blockchain technology to predict the completion time and the effort of a software using ML algorithms. This type of estimation model, using ML, making use of blockchain technology, does not exist in the literature, and this is the core novelty of this proposed model. The final prediction of the software effort integrates another technique for improving the calculated estimation, the standard deviation technique proposed by the authors previously. This model helped lessening the overall mean magnitude of relative error (MMRE) of the original model from 6.82% to 1.73% for the augmented dataset of 126 projects. All four ML techniques used for the proposed model give a better <i>p</i>-value than the original model using statistical testing through the Wilcoxon test. The average of the MMRE for effort estimation of all four techniques is below 25% on a dataset of 136 projects. The application of the standard deviation technique further helps in lessening the MMRE of the proposed model at 70%, 80%, and 90% confidence levels. The work will give insight to researchers and experts and open the doors for new research in this area.</p>","PeriodicalId":50378,"journal":{"name":"IET Software","volume":"2025 1","pages":""},"PeriodicalIF":1.3,"publicationDate":"2025-10-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ietresearch.onlinelibrary.wiley.com/doi/epdf/10.1049/sfw2/9238663","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145366661","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MMF: A Lightweight Approach of Multimodel Fusion for Malware Detection MMF:一种用于恶意软件检测的轻量级多模型融合方法
IF 1.3 4区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-10-14 DOI: 10.1049/sfw2/1046015
Bo Yang, Mengbo Li, Li Li, Huai Liu

Nowadays, the Android system is widely used in mobile devices. The existence of malware in the Android system has posed serious security risks. Therefore, detecting malware has become a main research focus for Android devices. The existing malware detection methods include those based on static analysis, dynamic analysis, and hybrid analysis. The dynamic analysis and hybrid analysis methods require the simulation of malware’s execution in a certain environment, which often incurs high costs. With the aid of contemporary deep learning technology, static method can provide comparably good results without running software. To address these challenges, we propose a novel and efficient multimodel fusion (MMF) malware detection method. MMF innovatively integrates various static features, including application programming interface (API) call characteristics, request permission (RP) features, and bytecode image features. This fusion approach allows MMF to achieve high detection performance without the need for dynamic execution of the software. Compared to existing methods, MMF exhibits a higher accuracy rate of 99.4% and demonstrates superiority over baseline techniques in various metrics. Our comprehensive analysis and experiments confirm MMF’s effectiveness and efficiency in detecting malware, making a significant contribution to the field of Android malware detection.

如今,Android系统在移动设备上得到了广泛的应用。Android系统中恶意软件的存在带来了严重的安全隐患。因此,检测恶意软件已成为Android设备的主要研究重点。现有的恶意软件检测方法包括基于静态分析、动态分析和混合分析的检测方法。动态分析和混合分析方法需要在一定的环境下模拟恶意软件的执行情况,这往往会产生很高的成本。在当代深度学习技术的帮助下,静态方法可以在不运行软件的情况下提供相当好的结果。为了解决这些挑战,我们提出了一种新颖高效的多模型融合(MMF)恶意软件检测方法。MMF创新性地集成了各种静态特性,包括应用程序编程接口(API)调用特性、请求权限(RP)特性和字节码映像特性。这种融合方法使MMF无需动态执行软件即可实现高检测性能。与现有方法相比,MMF的准确率高达99.4%,在各种指标上都优于基线技术。我们的综合分析和实验证实了MMF检测恶意软件的有效性和效率,为Android恶意软件检测领域做出了重大贡献。
{"title":"MMF: A Lightweight Approach of Multimodel Fusion for Malware Detection","authors":"Bo Yang,&nbsp;Mengbo Li,&nbsp;Li Li,&nbsp;Huai Liu","doi":"10.1049/sfw2/1046015","DOIUrl":"https://doi.org/10.1049/sfw2/1046015","url":null,"abstract":"<p>Nowadays, the Android system is widely used in mobile devices. The existence of malware in the Android system has posed serious security risks. Therefore, detecting malware has become a main research focus for Android devices. The existing malware detection methods include those based on static analysis, dynamic analysis, and hybrid analysis. The dynamic analysis and hybrid analysis methods require the simulation of malware’s execution in a certain environment, which often incurs high costs. With the aid of contemporary deep learning technology, static method can provide comparably good results without running software. To address these challenges, we propose a novel and efficient multimodel fusion (MMF) malware detection method. MMF innovatively integrates various static features, including application programming interface (API) call characteristics, request permission (RP) features, and bytecode image features. This fusion approach allows MMF to achieve high detection performance without the need for dynamic execution of the software. Compared to existing methods, MMF exhibits a higher accuracy rate of 99.4% and demonstrates superiority over baseline techniques in various metrics. Our comprehensive analysis and experiments confirm MMF’s effectiveness and efficiency in detecting malware, making a significant contribution to the field of Android malware detection.</p>","PeriodicalId":50378,"journal":{"name":"IET Software","volume":"2025 1","pages":""},"PeriodicalIF":1.3,"publicationDate":"2025-10-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ietresearch.onlinelibrary.wiley.com/doi/epdf/10.1049/sfw2/1046015","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145316864","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Automated NLP-Based Classification of Nonfunctional Requirements in Blockchain and Cross-Domain Software Systems Using BERT and Machine Learning 基于nlp的区块链和跨域软件系统非功能需求自动分类
IF 1.3 4区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-10-12 DOI: 10.1049/sfw2/9996509
Touseef Tahir, Bilal Hassan, Hamid Jahankhani, Nimra Zia, Muhammad Sharjeel

Automated nonfunctional requirements (NFRs) classification enhances consistency and traceability by systematically labeling requirements, saving effort, supporting early architectural and testing decisions, improving stakeholder communication, and enabling quality across diverse software domains. While prior work has applied natural language processing (NLP) and machine learning (ML) to NFR classification, existing datasets are often limited in size, domain diversity, and contextual richness. This study presents a novel dataset comprising over 2400 NFRs spanning 269 software projects across 26 software application domains, including nine blockchain projects. The raw requirements are standardized using Rupp’s boilerplate to reduce vagueness and ambiguity, and the classification of NFRs types follows ISO/IEC 25,010 definitions. We employ a range of traditional ML, deep learning (DL), and a transformer-based model (i.e., BERT-base) for automated classification of NFRs, evaluating performance across cross-domain and blockchain-specific NFRs. Results highlight that domain-aware adaptation significantly enhances classification accuracy, with traditional ML and DL models showing strong performance on blockchain requirements. This work contributes a publicly available, context-rich dataset and provides empirical insights into the effectiveness of NLP-based NFR classification in both general and blockchain-specific settings.

自动化的非功能需求(NFRs)分类通过系统地标记需求、节省工作、支持早期架构和测试决策、改进涉众沟通以及支持跨不同软件领域的质量来增强一致性和可追溯性。虽然之前的工作已经将自然语言处理(NLP)和机器学习(ML)应用于NFR分类,但现有的数据集通常在规模、领域多样性和上下文丰富性方面受到限制。本研究提出了一个新的数据集,包含2400多个nfr,跨越26个软件应用领域的269个软件项目,包括9个区块链项目。原始需求使用Rupp的样板进行标准化,以减少模糊性和模糊性,nfr类型的分类遵循ISO/IEC 25,010定义。我们采用了一系列传统的ML、深度学习(DL)和基于转换器的模型(即BERT-base)来对nfr进行自动分类,评估跨域和区块链特定的nfr的性能。结果表明,领域感知自适应显著提高了分类精度,传统的ML和DL模型在区块链要求上表现出色。这项工作提供了一个公开可用的、上下文丰富的数据集,并为基于nlp的NFR分类在一般和区块链特定设置中的有效性提供了经验见解。
{"title":"Automated NLP-Based Classification of Nonfunctional Requirements in Blockchain and Cross-Domain Software Systems Using BERT and Machine Learning","authors":"Touseef Tahir,&nbsp;Bilal Hassan,&nbsp;Hamid Jahankhani,&nbsp;Nimra Zia,&nbsp;Muhammad Sharjeel","doi":"10.1049/sfw2/9996509","DOIUrl":"https://doi.org/10.1049/sfw2/9996509","url":null,"abstract":"<p>Automated nonfunctional requirements (NFRs) classification enhances consistency and traceability by systematically labeling requirements, saving effort, supporting early architectural and testing decisions, improving stakeholder communication, and enabling quality across diverse software domains. While prior work has applied natural language processing (NLP) and machine learning (ML) to NFR classification, existing datasets are often limited in size, domain diversity, and contextual richness. This study presents a novel dataset comprising over 2400 NFRs spanning 269 software projects across 26 software application domains, including nine blockchain projects. The raw requirements are standardized using Rupp’s boilerplate to reduce vagueness and ambiguity, and the classification of NFRs types follows ISO/IEC 25,010 definitions. We employ a range of traditional ML, deep learning (DL), and a transformer-based model (i.e., BERT-base) for automated classification of NFRs, evaluating performance across cross-domain and blockchain-specific NFRs. Results highlight that domain-aware adaptation significantly enhances classification accuracy, with traditional ML and DL models showing strong performance on blockchain requirements. This work contributes a publicly available, context-rich dataset and provides empirical insights into the effectiveness of NLP-based NFR classification in both general and blockchain-specific settings.</p>","PeriodicalId":50378,"journal":{"name":"IET Software","volume":"2025 1","pages":""},"PeriodicalIF":1.3,"publicationDate":"2025-10-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ietresearch.onlinelibrary.wiley.com/doi/epdf/10.1049/sfw2/9996509","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145316635","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Design of Minimal Spanning Tree and Analytic Hierarchical Process (SAHP) Based Hybrid Technique for Software Requirements Prioritization 基于最小生成树和层次分析法(SAHP)的软件需求优先级混合技术设计
IF 1.3 4区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-10-09 DOI: 10.1049/sfw2/8819735
Muhammad Yaseen, Esraa Ali, Nadeem Sarwar, Leila Jamel, Irfanud Din, Farrukh Yuldashev, Foongli Law

Prioritizing software requirements in a sustainable manner can significantly contribute to the success of a software project, adding substantial value throughout its development lifecycle. Analytic hierarchical process (AHP) is considered to yield more accurate prioritized results, but due to high pairwise comparisons, it is not considered to be scalable for prioritization of high number of requirements. To address scalability issue, a hybrid approach of minimal spanning trees (MSTs) and AHP, called as spanning tree and AHP (SAHP), is designed for prioritizing large set of functional requirements (FRs) with fewer comparisons, and thus scalability issue is solved. In this research, on-demand open object (ODOO) enterprise resource planning (ERP) system FRs are prioritized, and the results are compared with AHP. The results of the case study proved that SAHP is more scalable that can prioritize any type of requirement with only n–1 pairs of requirements. Total FRs considered for case from ODOO were 100, where 18 spanning trees were constructed from it. With only 90 pairwise comparisons, these FRs were prioritized with more consistency compared to AHP. Total pairwise comparisons with AHP reach 4950, which is 55 times more compared with SAHP. Consistency of results is measured from average consistency index (CI) value, which was below 0.1. The consistency ratio (CR) value below 0.1 shows results are consistent and acceptable.

以可持续的方式对软件需求进行优先级排序可以显著地促进软件项目的成功,在整个开发生命周期中增加实质性的价值。分析层次过程(AHP)被认为产生更准确的优先级结果,但是由于高度两两比较,它不被认为是可伸缩的,用于大量需求的优先级排序。为了解决可伸缩性问题,设计了一种最小生成树(MSTs)和层次分析法的混合方法,称为生成树和层次分析法(SAHP),以较少的比较对大型功能需求(FRs)进行优先级排序,从而解决了可伸缩性问题。本研究对按需开放对象(ODOO)企业资源规划(ERP)系统中的FRs进行了优先级排序,并将结果与层次分析法进行了比较。案例研究的结果证明,SAHP具有更高的可伸缩性,可以用n-1对需求对任何类型的需求进行优先级排序。对于来自ODOO的案例,考虑的总FRs为100,其中18棵生成树是由它构建的。只有90个两两比较,与AHP相比,这些fr的优先级更具一致性。与AHP的两两比较总数达到4950,是SAHP的55倍。结果的一致性以平均一致性指数(CI)值衡量,CI值小于0.1。一致性比(CR)值小于0.1表示结果一致且可接受。
{"title":"Design of Minimal Spanning Tree and Analytic Hierarchical Process (SAHP) Based Hybrid Technique for Software Requirements Prioritization","authors":"Muhammad Yaseen,&nbsp;Esraa Ali,&nbsp;Nadeem Sarwar,&nbsp;Leila Jamel,&nbsp;Irfanud Din,&nbsp;Farrukh Yuldashev,&nbsp;Foongli Law","doi":"10.1049/sfw2/8819735","DOIUrl":"https://doi.org/10.1049/sfw2/8819735","url":null,"abstract":"<p>Prioritizing software requirements in a sustainable manner can significantly contribute to the success of a software project, adding substantial value throughout its development lifecycle. Analytic hierarchical process (AHP) is considered to yield more accurate prioritized results, but due to high pairwise comparisons, it is not considered to be scalable for prioritization of high number of requirements. To address scalability issue, a hybrid approach of minimal spanning trees (MSTs) and AHP, called as spanning tree and AHP (SAHP), is designed for prioritizing large set of functional requirements (FRs) with fewer comparisons, and thus scalability issue is solved. In this research, on-demand open object (ODOO) enterprise resource planning (ERP) system FRs are prioritized, and the results are compared with AHP. The results of the case study proved that SAHP is more scalable that can prioritize any type of requirement with only <i>n</i>–1 pairs of requirements. Total FRs considered for case from ODOO were 100, where 18 spanning trees were constructed from it. With only 90 pairwise comparisons, these FRs were prioritized with more consistency compared to AHP. Total pairwise comparisons with AHP reach 4950, which is 55 times more compared with SAHP. Consistency of results is measured from average consistency index (CI) value, which was below 0.1. The consistency ratio (CR) value below 0.1 shows results are consistent and acceptable.</p>","PeriodicalId":50378,"journal":{"name":"IET Software","volume":"2025 1","pages":""},"PeriodicalIF":1.3,"publicationDate":"2025-10-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ietresearch.onlinelibrary.wiley.com/doi/epdf/10.1049/sfw2/8819735","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145272241","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Word-Level Nonequivalence and Translation Strategies in English–Chinese Translation Based on Image Processing Technology 基于图像处理技术的英汉翻译中的词级不对等及翻译策略
IF 1.3 4区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-09-29 DOI: 10.1049/sfw2/5511556
Haihua Tu, Lingbo Han

The process of translation is the process of accurately understanding the original work. It uses other languages to express the meaning of the original work and reproduce the original text in other languages. However, translation equivalence is a relative term, and there is no complete equivalence. In translation practice, translators often face different inequalities. The inequality of lexical levels means that no words matching the original text can be found in the specified language. These equivalence relationships are different to some extent, which brings great difficulties to translation. This paper first made a relevant interpretation of the common phenomenon of word-level inequality in English–Chinese translation, and analyzed the differences of source language concepts in translation. It made a relevant study on the lexical inequality in English–Chinese translation, and described the cultural inequality. After that, this paper studied and planned the equivalence requirements and solutions in English–Chinese translation. It was proposed to strengthen the learning and understanding of Chinese and Western cultures, and to translate based on the cultural characteristics of different regions. It was also proposed that transliteration should be used to ensure the accuracy of English–Chinese translation and reduce the nonequivalence between word levels. Subsequently, this paper introduced image processing technology into translation and used image processing technology to strengthen translation strategies. It also focused on analyzing the main types of image processing technology and used image processing technology to fully understand the translation process. It was necessary to use image processing technology to correctly express the translation. Finally, image processing technology was used to strengthen translation strategies and research. According to experiments and surveys, the use of image processing technology to create new English–Chinese translation strategies could effectively improve the satisfaction of 18% of translators.

翻译的过程就是准确理解原作的过程。它用其他语言来表达原作的意思,用其他语言再现原作。然而,翻译对等是一个相对的术语,不存在完全的对等。在翻译实践中,译者经常面临不同的不平等现象。词汇层次的不平等意味着在指定语言中找不到与原文匹配的单词。这些对等关系在一定程度上是不同的,这给翻译带来了很大的困难。本文首先对英汉翻译中常见的词级不平等现象进行了相关解释,并分析了翻译中源语概念的差异。对英汉翻译中的词汇不平等现象进行了相关研究,并对文化不平等现象进行了描述。然后,对英汉翻译中的对等要求和解决方案进行了研究和规划。建议加强对中西方文化的学习和了解,根据不同地区的文化特点进行翻译。为了保证英汉翻译的准确性,减少词层之间的不对等现象,应采用音译的方法。随后,本文将图像处理技术引入到翻译中,并利用图像处理技术加强翻译策略。重点分析了图像处理技术的主要类型,并利用图像处理技术全面了解翻译过程。使用图像处理技术来正确表达翻译内容是必要的。最后,利用图像处理技术加强翻译策略和研究。根据实验和调查,利用图像处理技术创建新的英汉翻译策略可以有效提高18%的译者的满意度。
{"title":"Word-Level Nonequivalence and Translation Strategies in English–Chinese Translation Based on Image Processing Technology","authors":"Haihua Tu,&nbsp;Lingbo Han","doi":"10.1049/sfw2/5511556","DOIUrl":"https://doi.org/10.1049/sfw2/5511556","url":null,"abstract":"<p>The process of translation is the process of accurately understanding the original work. It uses other languages to express the meaning of the original work and reproduce the original text in other languages. However, translation equivalence is a relative term, and there is no complete equivalence. In translation practice, translators often face different inequalities. The inequality of lexical levels means that no words matching the original text can be found in the specified language. These equivalence relationships are different to some extent, which brings great difficulties to translation. This paper first made a relevant interpretation of the common phenomenon of word-level inequality in English–Chinese translation, and analyzed the differences of source language concepts in translation. It made a relevant study on the lexical inequality in English–Chinese translation, and described the cultural inequality. After that, this paper studied and planned the equivalence requirements and solutions in English–Chinese translation. It was proposed to strengthen the learning and understanding of Chinese and Western cultures, and to translate based on the cultural characteristics of different regions. It was also proposed that transliteration should be used to ensure the accuracy of English–Chinese translation and reduce the nonequivalence between word levels. Subsequently, this paper introduced image processing technology into translation and used image processing technology to strengthen translation strategies. It also focused on analyzing the main types of image processing technology and used image processing technology to fully understand the translation process. It was necessary to use image processing technology to correctly express the translation. Finally, image processing technology was used to strengthen translation strategies and research. According to experiments and surveys, the use of image processing technology to create new English–Chinese translation strategies could effectively improve the satisfaction of 18% of translators.</p>","PeriodicalId":50378,"journal":{"name":"IET Software","volume":"2025 1","pages":""},"PeriodicalIF":1.3,"publicationDate":"2025-09-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ietresearch.onlinelibrary.wiley.com/doi/epdf/10.1049/sfw2/5511556","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145224163","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Systematic Mapping of AI-Based Approaches for Requirements Prioritization 基于人工智能的需求优先排序方法的系统映射
IF 1.3 4区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-09-27 DOI: 10.1049/sfw2/8953863
María-Isabel Limaylla-Lunarejo, Nelly Condori-Fernandez, Miguel Rodríguez Luaces

Context and Motivation: Requirements prioritization (RP) is a main concern of requirements engineering (RE). Traditional prioritization techniques, while effective, often involve manual effort and are time-consuming. In recent years, thanks to the advances in AI-based techniques and algorithms, several promising alternatives have emerged to optimize this process.

Question: The main goal of this work is to review the current state of requirement prioritization, focusing on AI-based techniques and a classification scheme to provide a comprehensive overview. Additionally, we examine the criteria utilized by these AI-based techniques, as well as the datasets and evaluation metrics employed. For this purpose, we conducted a systematic mapping study (SMS) of studies published between 2011 and 2023.

Results: Our analysis reveals a diverse range of AI-based techniques in use, with fuzzy logic being the most commonly applied. Moreover, most studies continue to depend on stakeholder input as a key criterion, limiting the potential for full automation of the prioritization process. Finally, there appears to be no standardized evaluation metric or dataset across the reviewed papers, focusing on the need for standardized approaches across studies.

Contribution: This work provides a systematic categorization of current AI-based techniques used for automating RP. Additionally, it updates and expands existing reviews, offering a valuable resource for practitioners and nonspecialists.

背景和动机:需求优先级(RP)是需求工程(RE)的主要关注点。传统的优先级划分技术虽然有效,但往往需要人工操作,而且耗时。近年来,由于基于人工智能的技术和算法的进步,出现了一些有希望的替代方案来优化这一过程。问题:这项工作的主要目标是回顾需求优先级的当前状态,关注基于人工智能的技术和分类方案,以提供一个全面的概述。此外,我们还研究了这些基于人工智能的技术所使用的标准,以及所采用的数据集和评估指标。为此,我们对2011年至2023年间发表的研究进行了系统的地图研究(SMS)。结果:我们的分析揭示了使用的各种基于人工智能的技术,模糊逻辑是最常用的。此外,大多数研究仍然依赖利益相关者的输入作为关键标准,限制了优先排序过程完全自动化的潜力。最后,在审查的论文中似乎没有标准化的评估指标或数据集,重点是需要标准化的研究方法。贡献:这项工作提供了当前用于自动化RP的基于ai的技术的系统分类。此外,它更新并扩展了现有的评论,为从业者和非专业人士提供了有价值的资源。
{"title":"Systematic Mapping of AI-Based Approaches for Requirements Prioritization","authors":"María-Isabel Limaylla-Lunarejo,&nbsp;Nelly Condori-Fernandez,&nbsp;Miguel Rodríguez Luaces","doi":"10.1049/sfw2/8953863","DOIUrl":"https://doi.org/10.1049/sfw2/8953863","url":null,"abstract":"<p><b>Context and Motivation:</b> Requirements prioritization (RP) is a main concern of requirements engineering (RE). Traditional prioritization techniques, while effective, often involve manual effort and are time-consuming. In recent years, thanks to the advances in AI-based techniques and algorithms, several promising alternatives have emerged to optimize this process.</p><p><b>Question:</b> The main goal of this work is to review the current state of requirement prioritization, focusing on AI-based techniques and a classification scheme to provide a comprehensive overview. Additionally, we examine the criteria utilized by these AI-based techniques, as well as the datasets and evaluation metrics employed. For this purpose, we conducted a systematic mapping study (SMS) of studies published between 2011 and 2023.</p><p><b>Results:</b> Our analysis reveals a diverse range of AI-based techniques in use, with fuzzy logic being the most commonly applied. Moreover, most studies continue to depend on stakeholder input as a key criterion, limiting the potential for full automation of the prioritization process. Finally, there appears to be no standardized evaluation metric or dataset across the reviewed papers, focusing on the need for standardized approaches across studies.</p><p><b>Contribution:</b> This work provides a systematic categorization of current AI-based techniques used for automating RP. Additionally, it updates and expands existing reviews, offering a valuable resource for practitioners and nonspecialists.</p>","PeriodicalId":50378,"journal":{"name":"IET Software","volume":"2025 1","pages":""},"PeriodicalIF":1.3,"publicationDate":"2025-09-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ietresearch.onlinelibrary.wiley.com/doi/epdf/10.1049/sfw2/8953863","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145146844","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Web-Based Early Dementia Detection Using Deep Learning, Ensemble Machine Learning, and Model Explainability Through LIME and SHAP 基于网络的早期痴呆检测使用深度学习,集成机器学习,并通过LIME和SHAP模型的可解释性
IF 1.3 4区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-09-27 DOI: 10.1049/sfw2/5455082
Khandaker Mohammad Mohi Uddin, Abir Chowdhury, Md Mahbubur Rahman Druvo, Md. Shariful Islam, Md Ashraf Uddin

Dementia is a gradual and incapacitating illness that impairs cognitive abilities and causes memory loss, disorientation, and challenges with daily tasks. Treatment of the disease and better patient outcomes depend on early identification of dementia. In this paper, the study uses a publicly available dataset to develop a comprehensive ensemble model of machine learning (ML) and deep learning (DL) framework for classifying the dementia stages. Before using SMOTE to balance the data, the procedure starts with data preprocessing which includes handling missing values, normalization and encoding. F-value and p-value help to select the best seven features, and the dataset is divided into training (70%) and testing (30%) portions. In addition, four DL models like long short-term memory (LSTM), convolutional neural networks (CNNs), multilayer perceptron (MLP), artificial neural networks (ANNs), and 12 ML models are trained such as logistic regression (LR), random forest (RF) and support vector machine (SVM). Hyperparameter tuning was utilized to further enhance each model’s performance and an ensemble voting technique was applied to aggregate predictions from several ML and DL algorithms, providing more reliable and accurate outcomes. For ensuring model transparency, interpretability strategies like as shapley additive explanations (SHAP) and local interpretable model-agnostic explanations (LIME) are applied in ANN and LR. The suggested model’s ANN shows a promising accuracy of 97.32% demonstrating its efficacy in the early diagnosis and categorization of dementia which can support clinical decisions. Furthermore, the proposed work, created a web-based solution for diagnosing dementia in real-time.

痴呆症是一种逐渐丧失能力的疾病,会损害认知能力,导致记忆丧失、定向障碍和日常工作困难。这种疾病的治疗和更好的患者预后取决于痴呆症的早期识别。在本文中,该研究使用公开可用的数据集开发了一个全面的机器学习(ML)和深度学习(DL)框架集成模型,用于对痴呆阶段进行分类。在使用SMOTE平衡数据之前,该过程从数据预处理开始,包括处理缺失值、规范化和编码。f值和p值帮助选择最好的7个特征,数据集被分为训练(70%)和测试(30%)部分。此外,还训练了长短期记忆(LSTM)、卷积神经网络(cnn)、多层感知器(MLP)、人工神经网络(ann)等4种深度学习模型,以及逻辑回归(LR)、随机森林(RF)和支持向量机(SVM)等12种ML模型。利用超参数调优来进一步提高每个模型的性能,并应用集成投票技术来聚合来自多个ML和DL算法的预测,提供更可靠和准确的结果。为了确保模型的透明性,在人工神经网络和LR中应用了shapley加性解释(SHAP)和局部可解释模型不可知解释(LIME)等可解释性策略。该模型的人工神经网络显示出97.32%的准确率,表明其在痴呆症的早期诊断和分类方面的有效性,可以支持临床决策。此外,这项工作还创建了一个基于网络的实时诊断痴呆症的解决方案。
{"title":"Web-Based Early Dementia Detection Using Deep Learning, Ensemble Machine Learning, and Model Explainability Through LIME and SHAP","authors":"Khandaker Mohammad Mohi Uddin,&nbsp;Abir Chowdhury,&nbsp;Md Mahbubur Rahman Druvo,&nbsp;Md. Shariful Islam,&nbsp;Md Ashraf Uddin","doi":"10.1049/sfw2/5455082","DOIUrl":"https://doi.org/10.1049/sfw2/5455082","url":null,"abstract":"<p>Dementia is a gradual and incapacitating illness that impairs cognitive abilities and causes memory loss, disorientation, and challenges with daily tasks. Treatment of the disease and better patient outcomes depend on early identification of dementia. In this paper, the study uses a publicly available dataset to develop a comprehensive ensemble model of machine learning (ML) and deep learning (DL) framework for classifying the dementia stages. Before using SMOTE to balance the data, the procedure starts with data preprocessing which includes handling missing values, normalization and encoding. <i>F</i>-value and <i>p</i>-value help to select the best seven features, and the dataset is divided into training (70%) and testing (30%) portions. In addition, four DL models like long short-term memory (LSTM), convolutional neural networks (CNNs), multilayer perceptron (MLP), artificial neural networks (ANNs), and 12 ML models are trained such as logistic regression (LR), random forest (RF) and support vector machine (SVM). Hyperparameter tuning was utilized to further enhance each model’s performance and an ensemble voting technique was applied to aggregate predictions from several ML and DL algorithms, providing more reliable and accurate outcomes. For ensuring model transparency, interpretability strategies like as shapley additive explanations (SHAP) and local interpretable model-agnostic explanations (LIME) are applied in ANN and LR. The suggested model’s ANN shows a promising accuracy of 97.32% demonstrating its efficacy in the early diagnosis and categorization of dementia which can support clinical decisions. Furthermore, the proposed work, created a web-based solution for diagnosing dementia in real-time.</p>","PeriodicalId":50378,"journal":{"name":"IET Software","volume":"2025 1","pages":""},"PeriodicalIF":1.3,"publicationDate":"2025-09-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ietresearch.onlinelibrary.wiley.com/doi/epdf/10.1049/sfw2/5455082","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145146843","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Systematic Literature Review on Application of Agile Software Development Process Models for the Development of Safety-Critical Systems in Multiple Domains 敏捷软件开发过程模型在多领域安全关键系统开发中的应用综述
IF 1.3 4区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-09-19 DOI: 10.1049/sfw2/5227350
Hafiza Maria Maqsood, Joelma Choma, Eduardo Guerra, Andrea Bondavalli

This paper presents a literature review on using agile for safety-critical systems (SCSs). We have systematically selected and evaluated relevant literature to find out major areas of concern for adapting agile in the development of SCSs. In the paper, we have listed the most used Agile process models and reasons for their suitability for SCS, then we have outlined phases of the software development life cycle (SDLC) where changes are required to make an agile process suitable for the development of SCSs. Thirdly, we have elaborated on problems and other important aspects according to specific domains where agile is used for SCS. This paper serves as an insight into the latest trends and problems regarding the use of Agile process models to develop SCSs.

本文介绍了在安全关键系统(scs)中使用敏捷的文献综述。我们系统地选择和评估了相关文献,以找出在scs开发中采用敏捷的主要关注领域。在本文中,我们列出了最常用的敏捷过程模型以及它们适合于SCS的原因,然后我们概述了软件开发生命周期(SDLC)的各个阶段,在这些阶段中,需要进行更改才能使敏捷过程适合于SCS的开发。第三,我们根据敏捷在SCS中应用的具体领域,详细阐述了问题和其他重要方面。本文提供了关于使用敏捷过程模型开发scs的最新趋势和问题的见解。
{"title":"A Systematic Literature Review on Application of Agile Software Development Process Models for the Development of Safety-Critical Systems in Multiple Domains","authors":"Hafiza Maria Maqsood,&nbsp;Joelma Choma,&nbsp;Eduardo Guerra,&nbsp;Andrea Bondavalli","doi":"10.1049/sfw2/5227350","DOIUrl":"10.1049/sfw2/5227350","url":null,"abstract":"<p>This paper presents a literature review on using agile for safety-critical systems (SCSs). We have systematically selected and evaluated relevant literature to find out major areas of concern for adapting agile in the development of SCSs. In the paper, we have listed the most used Agile process models and reasons for their suitability for SCS, then we have outlined phases of the software development life cycle (SDLC) where changes are required to make an agile process suitable for the development of SCSs. Thirdly, we have elaborated on problems and other important aspects according to specific domains where agile is used for SCS. This paper serves as an insight into the latest trends and problems regarding the use of Agile process models to develop SCSs.</p>","PeriodicalId":50378,"journal":{"name":"IET Software","volume":"2025 1","pages":""},"PeriodicalIF":1.3,"publicationDate":"2025-09-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ietresearch.onlinelibrary.wiley.com/doi/epdf/10.1049/sfw2/5227350","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145101897","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
BLOCKVISA: A Blockchain-Based System for Efficient and Secure Visa, Passport, and Immigration Verification BLOCKVISA:基于区块链的高效安全签证、护照和移民验证系统
IF 1.3 4区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-09-19 DOI: 10.1049/sfw2/5567569
Faraz Masood, Ali Haider Shamsan, Arman Rasool Faridi

In the fast-changing landscape of global mobility, the need for secure, efficient, and interoperable visa, passport, and immigration verification systems has never been higher. Traditional systems are inefficient, have security vulnerabilities, and exhibit poor interoperability. This study introduces a novel approach for the blockchain solution in passport verification inefficiencies-BLOCKVISA. BLOCKVISA, in its nature, uses decentralized and immutable blockchain technology to make the system more secure, automate the verification process, and ensure data sharing frictionlessly across jurisdictions. Core components of the system include smart contracts developed in Solidity, a user interface (UI) created with Next.js, and integration with MetaMask and Web3.js for safe interactions with the blockchain. Rigorous testing was done using Mocha, and more intensive benchmarking was done using Hyperledger Caliper against Ganache, Hyperledger Besu, as well as all the test networks, that is, Rinkeby, Ropsten, Goerli, Kovan, among others. Experiments showed that with BLOCKVISA, high throughput and low latency in controlled settings can be achieved, with almost perfect success rates being recorded. It also gave insights into how it would perform even better when deployed on a public network. The article undertakes a comparative analysis of performance metrics, brings out robust security features of the system, and discusses its scalability and feasibility for real-world implementation. By integrating advanced blockchain technology into the visa, passport, and immigration verification process, BLOCKVISA sets a new standard for global mobility solutions, promising enhanced efficiency, security, and interoperability.

在快速变化的全球流动性环境中,对安全、高效和可互操作的签证、护照和移民验证系统的需求从未如此高涨。传统系统效率低下,存在安全漏洞,互操作性差。本研究提出了一种新颖的方法来解决护照验证效率低下的区块链解决方案- blockvisa。从本质上讲,BLOCKVISA使用分散和不可变的区块链技术,使系统更加安全,自动化验证过程,并确保跨司法管辖区的数据无摩擦共享。该系统的核心组件包括在Solidity中开发的智能合约,用Next.js创建的用户界面(UI),以及与MetaMask和Web3.js的集成,以便与区块链进行安全交互。使用Mocha进行了严格的测试,并使用Hyperledger Caliper对Ganache, Hyperledger Besu以及所有测试网络(即Rinkeby, Ropsten, Goerli, Kovan等)进行了更密集的基准测试。实验表明,使用BLOCKVISA可以在受控设置下实现高吞吐量和低延迟,并记录了几乎完美的成功率。它还提供了在公共网络上部署时如何更好地执行的见解。本文对性能指标进行了比较分析,提出了系统的健壮的安全特性,并讨论了其在实际实现中的可伸缩性和可行性。通过将先进的区块链技术集成到签证、护照和移民验证流程中,BLOCKVISA为全球移动解决方案树立了新的标准,有望提高效率、安全性和互操作性。
{"title":"BLOCKVISA: A Blockchain-Based System for Efficient and Secure Visa, Passport, and Immigration Verification","authors":"Faraz Masood,&nbsp;Ali Haider Shamsan,&nbsp;Arman Rasool Faridi","doi":"10.1049/sfw2/5567569","DOIUrl":"10.1049/sfw2/5567569","url":null,"abstract":"<p>In the fast-changing landscape of global mobility, the need for secure, efficient, and interoperable visa, passport, and immigration verification systems has never been higher. Traditional systems are inefficient, have security vulnerabilities, and exhibit poor interoperability. This study introduces a novel approach for the blockchain solution in passport verification inefficiencies-BLOCKVISA. BLOCKVISA, in its nature, uses decentralized and immutable blockchain technology to make the system more secure, automate the verification process, and ensure data sharing frictionlessly across jurisdictions. Core components of the system include smart contracts developed in Solidity, a user interface (UI) created with Next.js, and integration with MetaMask and Web3.js for safe interactions with the blockchain. Rigorous testing was done using Mocha, and more intensive benchmarking was done using Hyperledger Caliper against Ganache, Hyperledger Besu, as well as all the test networks, that is, Rinkeby, Ropsten, Goerli, Kovan, among others. Experiments showed that with BLOCKVISA, high throughput and low latency in controlled settings can be achieved, with almost perfect success rates being recorded. It also gave insights into how it would perform even better when deployed on a public network. The article undertakes a comparative analysis of performance metrics, brings out robust security features of the system, and discusses its scalability and feasibility for real-world implementation. By integrating advanced blockchain technology into the visa, passport, and immigration verification process, BLOCKVISA sets a new standard for global mobility solutions, promising enhanced efficiency, security, and interoperability.</p>","PeriodicalId":50378,"journal":{"name":"IET Software","volume":"2025 1","pages":""},"PeriodicalIF":1.3,"publicationDate":"2025-09-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ietresearch.onlinelibrary.wiley.com/doi/epdf/10.1049/sfw2/5567569","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145101898","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IET Software
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1