首页 > 最新文献

Automated Software Engineering最新文献

英文 中文
Mobile application review summarization using chain of density prompting 使用密度链提示的移动应用审查汇总
IF 3.1 2区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-06-24 DOI: 10.1007/s10515-025-00533-5
Shristi Shrestha, Anas Mahmoud

Mobile app users commonly rely on app store ratings and reviews to find apps that suit their needs. However, the sheer volume of reviews available on app stores can lead to information overload, thus impeding users’ ability to make informed app selection decisions. To overcome this limitation, in this paper, we leverage Large Language Models (LLMs) to summarize mobile app reviews. In particular, we use the Chain of Density (CoD) prompt to guide OpenAI GPT-4 to generate abstractive, semantically dense, and readable summaries of mobile app reviews. The CoD prompt is engineered to iteratively extract salient entities from the source text and fuse them into a fixed-length summary. We evaluate the performance of our approach using a large dataset of mobile app reviews. We further conduct an empirical evaluation with 48 study participants to assess the readability of the generated CoD summaries. Our results show that an altered CoD prompt can correctly identify the main themes in user reviews and consolidate them into a natural language summary that is intended for end-user consumption. The prompt also manages to maintain the readability of the generated summaries while increasing their density. Our work in this paper aims to substantially improve mobile app users’ experience by providing an effective mechanism for summarizing important user feedback in the review stream.

手机应用用户通常依靠应用商店的评级和评论来寻找适合自己需求的应用。然而,应用商店中大量的评论可能会导致信息过载,从而阻碍用户做出明智的应用选择决策。为了克服这一限制,在本文中,我们利用大型语言模型(llm)来总结手机应用评论。特别是,我们使用密度链(CoD)提示来指导OpenAI GPT-4生成抽象的、语义密集的、可读的移动应用评论摘要。CoD提示被设计为迭代地从源文本中提取重要实体,并将它们融合到固定长度的摘要中。我们使用大量手机应用评论数据集来评估我们方法的性能。我们进一步对48名研究参与者进行了实证评估,以评估生成的CoD摘要的可读性。我们的研究结果表明,修改后的CoD提示符可以正确识别用户评论中的主题,并将它们整合到一个自然语言摘要中,以供最终用户使用。提示符还设法保持生成摘要的可读性,同时增加它们的密度。我们在本文中的工作旨在通过提供一种有效的机制来总结评论流中的重要用户反馈,从而大幅改善移动应用程序用户的体验。
{"title":"Mobile application review summarization using chain of density prompting","authors":"Shristi Shrestha,&nbsp;Anas Mahmoud","doi":"10.1007/s10515-025-00533-5","DOIUrl":"10.1007/s10515-025-00533-5","url":null,"abstract":"<div><p>Mobile app users commonly rely on app store ratings and reviews to find apps that suit their needs. However, the sheer volume of reviews available on app stores can lead to information overload, thus impeding users’ ability to make informed app selection decisions. To overcome this limitation, in this paper, we leverage Large Language Models (LLMs) to summarize mobile app reviews. In particular, we use the Chain of Density (CoD) prompt to guide OpenAI GPT-4 to generate abstractive, semantically dense, and readable summaries of mobile app reviews. The CoD prompt is engineered to iteratively extract salient entities from the source text and fuse them into a fixed-length summary. We evaluate the performance of our approach using a large dataset of mobile app reviews. We further conduct an empirical evaluation with 48 study participants to assess the readability of the generated CoD summaries. Our results show that an altered CoD prompt can correctly identify the main themes in user reviews and consolidate them into a natural language summary that is intended for end-user consumption. The prompt also manages to maintain the readability of the generated summaries while increasing their density. Our work in this paper aims to substantially improve mobile app users’ experience by providing an effective mechanism for summarizing important user feedback in the review stream.</p></div>","PeriodicalId":55414,"journal":{"name":"Automated Software Engineering","volume":"32 2","pages":""},"PeriodicalIF":3.1,"publicationDate":"2025-06-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145168100","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
DC-GAR: detecting vulnerabilities by utilizing graph properties and random walks to uncover richer features DC-GAR:通过利用图形属性和随机漫步来发现更丰富的特征来检测漏洞
IF 3.1 2区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-06-14 DOI: 10.1007/s10515-025-00532-6
Meng Wang, Xiao Han, Hong Zhang, Yiran Guo, Jiangfan Guo

Deep learning has become prominent in source code vulnerability detection due to its ability to automatically extract complex feature representations from code, eliminating the need for manually defined rules or patterns. Some methods treat code as text sequences, however, they often overlook its inherent structural information. In contrast, graph-based approaches effectively capture structural relationships, but the sparseness and inconsistency of structures may lead to uneven feature vector extraction, which means that the model may not be able to adequately characterize important nodes or paths. To address this issue, we propose an approach called Dual-channel Graph Neural Network combining Graph properties and Random walks (DC-GAR). This approach integrates graph properties and random walks within a dual-channel graph neural network framework to enhance vulnerability detection. Specifically, graph properties capture global semantic features, while random walks provide context-dependent node structure information. The combination of these features is then leveraged by the dual-channel graph neural network for detection and classification. We have implemented DC-GAR and evaluated it on a dataset of 29,514 functions. Experimental results demonstrate that DC-GAR surpasses state-of-the-art vulnerability detectors, including FlawFinder, SySeVR, Devign, VulCNN, AMPLE, HardVD, CodeBERT, and GraphCodeBERT in terms of accuracy and F1-Score. Moreover, DC-GAR has proven effective and practical in real-world open-source projects.

深度学习在源代码漏洞检测方面已经变得突出,因为它能够自动从代码中提取复杂的特征表示,从而消除了手动定义规则或模式的需要。一些方法将代码视为文本序列,然而,它们经常忽略其固有的结构信息。相比之下,基于图的方法可以有效地捕获结构关系,但结构的稀疏性和不一致性可能导致特征向量提取不均匀,这意味着模型可能无法充分表征重要节点或路径。为了解决这个问题,我们提出了一种称为双通道图神经网络结合图属性和随机漫步(DC-GAR)的方法。该方法在双通道图神经网络框架内集成了图属性和随机游走,增强了漏洞检测能力。具体来说,图属性捕获全局语义特征,而随机漫步提供与上下文相关的节点结构信息。然后,双通道图神经网络利用这些特征的组合进行检测和分类。我们已经实现了DC-GAR,并在包含29,514个函数的数据集上对其进行了评估。实验结果表明,DC-GAR在准确率和F1-Score方面超过了最先进的漏洞检测器,包括FlawFinder、SySeVR、Devign、VulCNN、AMPLE、HardVD、CodeBERT和GraphCodeBERT。此外,DC-GAR已经在现实世界的开源项目中被证明是有效和实用的。
{"title":"DC-GAR: detecting vulnerabilities by utilizing graph properties and random walks to uncover richer features","authors":"Meng Wang,&nbsp;Xiao Han,&nbsp;Hong Zhang,&nbsp;Yiran Guo,&nbsp;Jiangfan Guo","doi":"10.1007/s10515-025-00532-6","DOIUrl":"10.1007/s10515-025-00532-6","url":null,"abstract":"<div><p>Deep learning has become prominent in source code vulnerability detection due to its ability to automatically extract complex feature representations from code, eliminating the need for manually defined rules or patterns. Some methods treat code as text sequences, however, they often overlook its inherent structural information. In contrast, graph-based approaches effectively capture structural relationships, but the sparseness and inconsistency of structures may lead to uneven feature vector extraction, which means that the model may not be able to adequately characterize important nodes or paths. To address this issue, we propose an approach called <b>D</b>ual-<b>c</b>hannel Graph Neural Network combining <b>G</b>raph properties <b>a</b>nd <b>R</b>andom walks (<b>DC-GAR</b>). This approach integrates graph properties and random walks within a dual-channel graph neural network framework to enhance vulnerability detection. Specifically, graph properties capture global semantic features, while random walks provide context-dependent node structure information. The combination of these features is then leveraged by the dual-channel graph neural network for detection and classification. We have implemented DC-GAR and evaluated it on a dataset of 29,514 functions. Experimental results demonstrate that DC-GAR surpasses state-of-the-art vulnerability detectors, including <i>FlawFinder</i>, <i>SySeVR</i>, <i>Devign</i>, <i>VulCNN</i>, <i>AMPLE</i>, <i>HardVD</i>, <i>CodeBERT</i>, and <i>GraphCodeBERT</i> in terms of accuracy and F1-Score. Moreover, DC-GAR has proven effective and practical in real-world open-source projects.</p></div>","PeriodicalId":55414,"journal":{"name":"Automated Software Engineering","volume":"32 2","pages":""},"PeriodicalIF":3.1,"publicationDate":"2025-06-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145165624","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
On-the-fly unfolding with optimal exploration for linear temporal logic model checking of concurrent software and systems 并行软件和系统线性时序逻辑模型检验的动态展开与优化探索
IF 3.1 2区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-06-11 DOI: 10.1007/s10515-025-00511-x
Shuo Li, Li’ao Zheng, Ru Yang, Zhijun Ding

Linear temporal logic (LTL) model checking faces a significant challenge known as the state-explosion problem. The on-the-fly method is a solution that constructs and checks the state space simultaneously, avoiding generating all states in advance. But it is not effective for concurrent interleaving. Unfolding based on Petri nets is a succinct structure covering all states that can mitigate this problem caused by concurrency. Many state-of-the-art methods optimally explore a complete unfolding structure using a tree-like structure. However, it is difficult to apply such a tree-like structure directly to the traditional on-the-fly method of LTL. At the same time, constructing a complete unfolding structure in advance and then checking LTL is also wasteful. Thus, the existing optimal exploration methods are not applicable to the on-the-fly unfolding. To solve these challenges, we propose an LTL model-checking method called on-the-fly unfolding with optimal exploration. This method is based on program dependence net (PDNet) proposed in the previous work. Firstly, we define conflict transitions of PDNet and an exploration tree with a novel notion of delayed transitions, which differs from the existing tree-like structure. The tree improves the on-the-fly unfolding by exploring each partial-order run only once and avoiding enumerating all possible combinations. Then, we propose an on-the-fly unfolding algorithm that simultaneously constructs the exploration tree and generates the unfolding structure while checking LTL. We implement a tool for verifying LTL properties of concurrent programs. It also improves traditional unfolding generations and performs better than SPIN and DiVine on the used benchmarks. The core contribution of this paper is that we propose an on-the-fly unfolding with an optimal exploration method for LTL. It avoids the complete enumeration of concurrent combinations from traditional unfolding generation.

线性时间逻辑(LTL)模型检验面临着状态爆炸问题。动态方法是一种同时构造和检查状态空间的解决方案,避免提前生成所有状态。但对并发交错的处理效果不理想。基于Petri网的展开是一种简洁的结构,涵盖了所有状态,可以减轻并发性引起的这个问题。许多最先进的方法使用树状结构最佳地探索完整的展开结构。然而,这种树状结构很难直接应用到传统的实时LTL方法中。同时,提前构造一个完整的展开结构,然后再检查LTL也是一种浪费。因此,现有的最优勘探方法不适用于动态展开。为了解决这些挑战,我们提出了一种LTL模型检查方法,称为最优探索的动态展开。该方法基于先前提出的程序依赖网络(PDNet)。首先,我们定义了PDNet的冲突转换,并提出了一种新的探索树,该树与现有的树状结构不同,具有延迟转换的概念。树通过只探索每个部分顺序运行一次,避免枚举所有可能的组合,从而改进了动态展开。然后,我们提出了一种实时展开算法,在检查LTL的同时构建探索树并生成展开结构。我们实现了一个验证并发程序的LTL属性的工具。它还改进了传统的展开代,并且在使用的基准测试中比SPIN和DiVine表现得更好。本文的核心贡献在于,我们提出了一种动态展开LTL的最优探索方法。它避免了传统展开生成中并发组合的完整枚举。
{"title":"On-the-fly unfolding with optimal exploration for linear temporal logic model checking of concurrent software and systems","authors":"Shuo Li,&nbsp;Li’ao Zheng,&nbsp;Ru Yang,&nbsp;Zhijun Ding","doi":"10.1007/s10515-025-00511-x","DOIUrl":"10.1007/s10515-025-00511-x","url":null,"abstract":"<div><p>Linear temporal logic (LTL) model checking faces a significant challenge known as the state-explosion problem. The on-the-fly method is a solution that constructs and checks the state space simultaneously, avoiding generating all states in advance. But it is not effective for concurrent interleaving. Unfolding based on Petri nets is a succinct structure covering all states that can mitigate this problem caused by concurrency. Many state-of-the-art methods optimally explore a complete unfolding structure using a tree-like structure. However, it is difficult to apply such a tree-like structure directly to the traditional on-the-fly method of LTL. At the same time, constructing a complete unfolding structure in advance and then checking LTL is also wasteful. Thus, the existing optimal exploration methods are not applicable to the on-the-fly unfolding. To solve these challenges, we propose an LTL model-checking method called on-the-fly unfolding with optimal exploration. This method is based on program dependence net (PDNet) proposed in the previous work. Firstly, we define conflict transitions of PDNet and an exploration tree with a novel notion of delayed transitions, which differs from the existing tree-like structure. The tree improves the on-the-fly unfolding by exploring each partial-order run only once and avoiding enumerating all possible combinations. Then, we propose an on-the-fly unfolding algorithm that simultaneously constructs the exploration tree and generates the unfolding structure while checking LTL. We implement a tool for verifying LTL properties of concurrent programs. It also improves traditional unfolding generations and performs better than <i>SPIN</i> and <i>DiVine</i> on the used benchmarks. The core contribution of this paper is that we propose an on-the-fly unfolding with an optimal exploration method for LTL. It avoids the complete enumeration of concurrent combinations from traditional unfolding generation.</p></div>","PeriodicalId":55414,"journal":{"name":"Automated Software Engineering","volume":"32 2","pages":""},"PeriodicalIF":3.1,"publicationDate":"2025-06-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145165164","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
NexuSym: Marrying symbolic path finders with large language models NexuSym:将符号寻路器与大型语言模型相结合
IF 3.1 2区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-06-07 DOI: 10.1007/s10515-025-00529-1
Jiayi Wang, Ping Yu, Yi Qin, Yanyan Jiang, Yuan Yao, Xiaoxing Ma

Symbolic execution is a powerful technique for automated test case generation, ensuring comprehensive coverage of potential scenarios. However, it often struggles with complex, deep paths due to path explosion. Conversely, large language models (LLMs) utilize vast training data to generate test cases that can uncover intricate program behaviors that symbolic execution might miss. Despite their complementary strengths, integrating the systematic nature of symbolic execution with the creative capabilities of LLMs presents a significant challenge. We introduce NexuSym, an innovative tool that integrates symbolic execution with LLMs to facilitate the automatic generation of test cases. To effectively bridge the gap between these two approaches, we have developed a test case reducer, which normalizes the LLM-generated test cases to make them compatible with symbolic execution. Additionally, we propose a search space summarizer, which abstracts and condenses the search space explored by symbolic execution, enabling the LLM to focus on the most promising areas for further exploration. We instantiated NexuSym on KLEE and ChatGPT. Our evaluation of NexuSym involved 99 coreutils programs and 9 large GNU programs. The experimental results demonstrate that NexuSym significantly enhances program test coverage, with improvements of up to 20% in certain cases. Furthermore, we conducted an analysis of the monetary costs associated with using the LLM API, revealing that NexuSym is a highly cost-effective solution.

符号执行是自动化测试用例生成的强大技术,确保了潜在场景的全面覆盖。然而,由于路径爆炸,它经常与复杂而深刻的路径作斗争。相反,大型语言模型(llm)利用大量的训练数据来生成测试用例,这些测试用例可以揭示符号执行可能错过的复杂程序行为,尽管它们具有互补的优势,但将符号执行的系统性质与llm的创造性能力相集成是一个重大挑战。我们介绍NexuSym,一个创新的工具,集成了符号执行与llm,以促进测试用例的自动生成。为了有效地弥合这两种方法之间的差距,我们开发了一个测试用例减速器,它规范了llm生成的测试用例,使它们与符号执行兼容。此外,我们提出了一个搜索空间摘要器,它抽象和压缩了符号执行所探索的搜索空间,使LLM能够专注于最有前途的领域进行进一步的探索。我们在KLEE和ChatGPT上实例化了nexusyum。我们对NexuSym的评估涉及99个内核程序和9个大型GNU程序。实验结果表明,NexuSym显著提高了程序测试覆盖率,在某些情况下提高了20%。此外,我们对使用LLM API的成本进行了分析,发现NexuSym是一种极具成本效益的解决方案。
{"title":"NexuSym: Marrying symbolic path finders with large language models","authors":"Jiayi Wang,&nbsp;Ping Yu,&nbsp;Yi Qin,&nbsp;Yanyan Jiang,&nbsp;Yuan Yao,&nbsp;Xiaoxing Ma","doi":"10.1007/s10515-025-00529-1","DOIUrl":"10.1007/s10515-025-00529-1","url":null,"abstract":"<div><p>Symbolic execution is a powerful technique for automated test case generation, ensuring comprehensive coverage of potential scenarios. However, it often struggles with complex, deep paths due to path explosion. Conversely, large language models (LLMs) utilize vast training data to generate test cases that can uncover intricate program behaviors that symbolic execution might miss. Despite their complementary strengths, integrating the systematic nature of symbolic execution with the creative capabilities of LLMs presents a significant challenge. We introduce <span>NexuSym</span>, an innovative tool that integrates symbolic execution with LLMs to facilitate the automatic generation of test cases. To effectively bridge the gap between these two approaches, we have developed a test case reducer, which normalizes the LLM-generated test cases to make them compatible with symbolic execution. Additionally, we propose a search space summarizer, which abstracts and condenses the search space explored by symbolic execution, enabling the LLM to focus on the most promising areas for further exploration. We instantiated <span>NexuSym</span> on KLEE and ChatGPT. Our evaluation of <span>NexuSym</span> involved 99 coreutils programs and 9 large GNU programs. The experimental results demonstrate that <span>NexuSym</span> significantly enhances program test coverage, with improvements of up to 20% in certain cases. Furthermore, we conducted an analysis of the monetary costs associated with using the LLM API, revealing that <span>NexuSym</span> is a highly cost-effective solution.</p></div>","PeriodicalId":55414,"journal":{"name":"Automated Software Engineering","volume":"32 2","pages":""},"PeriodicalIF":3.1,"publicationDate":"2025-06-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145163313","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
What information contributes to log-based anomaly detection? Insights from a configurable transformer-based approach 哪些信息有助于基于日志的异常检测?来自基于可配置转换器的方法的见解
IF 3.1 2区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-06-03 DOI: 10.1007/s10515-025-00527-3
Xingfang Wu, Heng Li, Foutse Khomh

Log data are generated from logging statements in the source code, providing insights into the execution processes of software applications and systems. State-of-the-art log-based anomaly detection approaches typically leverage deep learning models to capture the semantic or sequential information in the log data and detect anomalous runtime behaviors. However, the impacts of these different types of information are not clear. In addition, most existing approaches ignore the timestamps in log data, which can potentially provide fine-grained sequential and temporal information. In this work, we propose a configurable Transformer-based anomaly detection model that can capture the semantic, sequential, and temporal information in the log data and allows us to configure the different types of information as the model’s features. Additionally, we train and evaluate the proposed model using log sequences of different lengths, thus overcoming the constraint of existing methods that rely on fixed-length or time-windowed log sequences as inputs. With the proposed model, we conduct a series of experiments with different combinations of input features to evaluate the roles of different types of information (i.e., sequential, temporal, semantic information) in anomaly detection. The model can attain competitive and consistently stable performance compared to the baselines when presented with log sequences of varying lengths. The results indicate that the event occurrence information plays a key role in identifying anomalies, while the impact of the sequential and temporal information is not significant for anomaly detection on the studied public datasets. On the other hand, the findings also reveal the simplicity of the studied public datasets and highlight the importance of constructing new datasets that contain different types of anomalies to better evaluate the performance of anomaly detection models.

日志数据是从源代码中的日志语句生成的,提供了对软件应用程序和系统执行过程的洞察。最先进的基于日志的异常检测方法通常利用深度学习模型来捕获日志数据中的语义或顺序信息,并检测异常的运行时行为。然而,这些不同类型信息的影响尚不清楚。此外,大多数现有的方法都忽略了日志数据中的时间戳,这可能会提供细粒度的顺序和时间信息。在这项工作中,我们提出了一个可配置的基于transformer的异常检测模型,该模型可以捕获日志数据中的语义、顺序和时间信息,并允许我们将不同类型的信息配置为模型的特征。此外,我们使用不同长度的对数序列来训练和评估所提出的模型,从而克服了依赖固定长度或时间窗对数序列作为输入的现有方法的约束。利用提出的模型,我们对输入特征的不同组合进行了一系列实验,以评估不同类型的信息(即顺序信息、时间信息、语义信息)在异常检测中的作用。当呈现不同长度的对数序列时,与基线相比,该模型可以获得具有竞争力和持续稳定的性能。结果表明,事件发生信息在异常识别中起关键作用,序列信息和时间信息对异常检测的影响不显著。另一方面,研究结果也揭示了所研究的公共数据集的简单性,并强调了构建包含不同类型异常的新数据集的重要性,以便更好地评估异常检测模型的性能。
{"title":"What information contributes to log-based anomaly detection? Insights from a configurable transformer-based approach","authors":"Xingfang Wu,&nbsp;Heng Li,&nbsp;Foutse Khomh","doi":"10.1007/s10515-025-00527-3","DOIUrl":"10.1007/s10515-025-00527-3","url":null,"abstract":"<div><p>Log data are generated from logging statements in the source code, providing insights into the execution processes of software applications and systems. State-of-the-art log-based anomaly detection approaches typically leverage deep learning models to capture the semantic or sequential information in the log data and detect anomalous runtime behaviors. However, the impacts of these different types of information are not clear. In addition, most existing approaches ignore the timestamps in log data, which can potentially provide fine-grained sequential and temporal information. In this work, we propose a configurable Transformer-based anomaly detection model that can capture the semantic, sequential, and temporal information in the log data and allows us to configure the different types of information as the model’s features. Additionally, we train and evaluate the proposed model using log sequences of different lengths, thus overcoming the constraint of existing methods that rely on fixed-length or time-windowed log sequences as inputs. With the proposed model, we conduct a series of experiments with different combinations of input features to evaluate the roles of different types of information (i.e., sequential, temporal, semantic information) in anomaly detection. The model can attain competitive and consistently stable performance compared to the baselines when presented with log sequences of varying lengths. The results indicate that the event occurrence information plays a key role in identifying anomalies, while the impact of the sequential and temporal information is not significant for anomaly detection on the studied public datasets. On the other hand, the findings also reveal the simplicity of the studied public datasets and highlight the importance of constructing new datasets that contain different types of anomalies to better evaluate the performance of anomaly detection models.</p></div>","PeriodicalId":55414,"journal":{"name":"Automated Software Engineering","volume":"32 2","pages":""},"PeriodicalIF":3.1,"publicationDate":"2025-06-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145161241","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Semi-supervised software vulnerability assessment via code lexical and structural information fusion 基于代码词法和结构信息融合的半监督软件漏洞评估
IF 3.1 2区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-06-03 DOI: 10.1007/s10515-025-00526-4
Wenlong Pei, Yilin Huang, Xiang Chen, Guilong Lu, Yong Liu, Chao Ni

In

recent years, data-driven approaches have become popular for software vulnerability assessment (SVA). However, these approaches need a large amount of labeled SVA data to construct effective SVA models. This process demands security expertise for accurate labeling, incurring significant costs and introducing potential errors. Therefore, collecting the training datasets for SVA can be a challenging task. To effectively alleviate the SVA data labeling cost, we propose an approach SURF, which makes full use of a limited amount of labeled SVA data combined with a large amount of unlabeled SVA data to train the SVA model via semi-supervised learning. Furthermore, SURF incorporates lexical information (i.e., treat the code as plain text) and structural information (i.e., treat the code as the code property graph) as bimodal inputs for the SVA model training, which can further improve the performance of SURF. Through extensive experiments, we evaluated the effectiveness of SURF on a dataset that contains C/C++ vulnerable functions from real-world software projects. The results show that only by labeling 30% of the SVA data, SURF can reach or even exceed the performance of state-of-the-art SVA baselines (such as DeepCVA and Func), even if these supervised baselines use 100% of the labeled SVA data. Furthermore, SURF can also exceed the performance of the state-of-the-art Positive-unlabeled learning baseline PILOT when both are trained on 30% of the labeled SVA data.

近年来,数据驱动方法在软件漏洞评估(SVA)中越来越流行。然而,这些方法需要大量标记的SVA数据来构建有效的SVA模型。这个过程需要安全方面的专业知识来进行准确的标记,这会产生巨大的成本并引入潜在的错误。因此,收集SVA的训练数据集可能是一项具有挑战性的任务。为了有效减轻SVA数据标注成本,我们提出了一种SURF方法,该方法充分利用有限的标记SVA数据结合大量未标记SVA数据,通过半监督学习训练SVA模型。此外,SURF将词法信息(即将代码视为纯文本)和结构信息(即将代码视为代码属性图)作为SVA模型训练的双峰输入,可以进一步提高SURF的性能。通过广泛的实验,我们评估了SURF在包含来自真实软件项目的C/ c++脆弱函数的数据集上的有效性。结果表明,只要标记30%的SVA数据,SURF就可以达到甚至超过最先进的SVA基线(如DeepCVA和Func)的性能,即使这些监督基线使用100%标记的SVA数据。此外,SURF也可以超过最先进的Positive-unlabeled学习基线PILOT,当两者都在30%的标记SVA数据上训练时。
{"title":"Semi-supervised software vulnerability assessment via code lexical and structural information fusion","authors":"Wenlong Pei,&nbsp;Yilin Huang,&nbsp;Xiang Chen,&nbsp;Guilong Lu,&nbsp;Yong Liu,&nbsp;Chao Ni","doi":"10.1007/s10515-025-00526-4","DOIUrl":"10.1007/s10515-025-00526-4","url":null,"abstract":"<div><p>In </p><p>recent years, data-driven approaches have become popular for software vulnerability assessment (SVA). However, these approaches need a large amount of labeled SVA data to construct effective SVA models. This process demands security expertise for accurate labeling, incurring significant costs and introducing potential errors. Therefore, collecting the training datasets for SVA can be a challenging task. To effectively alleviate the SVA data labeling cost, we propose an approach SURF, which makes full use of a limited amount of labeled SVA data combined with a large amount of unlabeled SVA data to train the SVA model via semi-supervised learning. Furthermore, SURF incorporates lexical information (i.e., treat the code as plain text) and structural information (i.e., treat the code as the code property graph) as bimodal inputs for the SVA model training, which can further improve the performance of SURF. Through extensive experiments, we evaluated the effectiveness of SURF on a dataset that contains C/C++ vulnerable functions from real-world software projects. The results show that only by labeling 30% of the SVA data, SURF can reach or even exceed the performance of state-of-the-art SVA baselines (such as DeepCVA and Func), even if these supervised baselines use 100% of the labeled SVA data. Furthermore, SURF can also exceed the performance of the state-of-the-art Positive-unlabeled learning baseline PILOT when both are trained on 30% of the labeled SVA data.</p></div>","PeriodicalId":55414,"journal":{"name":"Automated Software Engineering","volume":"32 2","pages":""},"PeriodicalIF":3.1,"publicationDate":"2025-06-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145161568","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Software testing for extended reality applications: a systematic mapping study 扩展现实应用的软件测试:系统的映射研究
IF 3.1 2区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-06-03 DOI: 10.1007/s10515-025-00523-7
Ruizhen Gu, José Miguel Rojas, Donghwan Shin

Extended Reality (XR) is an emerging technology spanning diverse application domains and offering immersive user experiences. However, its unique characteristics, such as six degrees of freedom interactions, present significant testing challenges distinct from traditional 2D GUI applications, demanding novel testing techniques to build high-quality XR applications. This paper presents the first systematic mapping study on software testing for XR applications. We selected 34 studies focusing on techniques and empirical approaches in XR software testing for detailed examination. The studies are classified and reviewed to address the current research landscape, test facets, and evaluation methodologies in the XR testing domain. Additionally, we provide a repository summarising the mapping study, including datasets and tools referenced in the selected studies, to support future research and practical applications. Our study highlights open challenges in XR testing and proposes actionable future research directions to address the gaps and advance the field of XR software testing.

扩展现实(XR)是一种新兴的技术,跨越了不同的应用领域,并提供了沉浸式的用户体验。然而,其独特的特性,如六自由度交互,与传统的2D GUI应用程序不同,提出了重大的测试挑战,需要新颖的测试技术来构建高质量的XR应用程序。本文首次对XR应用的软件测试进行了系统的映射研究。本文选取了34篇研究XR软件测试技术和实证方法的研究进行详细考察。这些研究被分类和回顾,以解决当前的研究前景,测试方面,和评估方法在XR测试领域。此外,我们提供了一个存储库,总结了地图研究,包括在选定的研究中引用的数据集和工具,以支持未来的研究和实际应用。我们的研究强调了XR测试中的开放挑战,并提出了可操作的未来研究方向,以解决差距并推进XR软件测试领域。
{"title":"Software testing for extended reality applications: a systematic mapping study","authors":"Ruizhen Gu,&nbsp;José Miguel Rojas,&nbsp;Donghwan Shin","doi":"10.1007/s10515-025-00523-7","DOIUrl":"10.1007/s10515-025-00523-7","url":null,"abstract":"<div><p>Extended Reality (XR) is an emerging technology spanning diverse application domains and offering immersive user experiences. However, its unique characteristics, such as six degrees of freedom interactions, present significant testing challenges distinct from traditional 2D GUI applications, demanding novel testing techniques to build high-quality XR applications. This paper presents the first systematic mapping study on software testing for XR applications. We selected 34 studies focusing on techniques and empirical approaches in XR software testing for detailed examination. The studies are classified and reviewed to address the current research landscape, test facets, and evaluation methodologies in the XR testing domain. Additionally, we provide a repository summarising the mapping study, including datasets and tools referenced in the selected studies, to support future research and practical applications. Our study highlights open challenges in XR testing and proposes actionable future research directions to address the gaps and advance the field of XR software testing.</p></div>","PeriodicalId":55414,"journal":{"name":"Automated Software Engineering","volume":"32 2","pages":""},"PeriodicalIF":3.1,"publicationDate":"2025-06-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10515-025-00523-7.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145161567","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
HGNNLink: recovering requirements-code traceability links with text and dependency-aware heterogeneous graph neural networks HGNNLink:使用文本和依赖关系感知的异构图神经网络恢复需求代码可追溯性链接
IF 3.1 2区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-05-31 DOI: 10.1007/s10515-025-00528-2
Bangchao Wang, Zhiyuan Zou, Xuanxuan Liang, Huan Jin, Peng Liang

Manually recovering traceability links between requirements and code artifacts often consumes substantial human resources. To address this, researchers have proposed automated methods based on textual similarity between requirements and code artifacts, such as information retrieval (IR) and pre-trained models, to determine whether traceability links exist between requirements and code artifacts. However, in the same system, developers often follow similar naming conventions and repeatedly use the same frameworks and template code, resulting in high textual similarity between code artifacts that are functionally unrelated. This makes it difficult to accurately identify the corresponding code artifacts for requirements artifacts solely based on textual similarity. Therefore, it is necessary to leverage the dependency relationships between code artifacts to assist in the requirements-code traceability link recovery process. Existing methods often treat dependency relationships as a post-processing step to refine textual similarity, overlooking the importance of textual similarity and dependency relationships in generating requirements-code traceability links. To address these limitations, we proposed Heterogeneous Graph Neural Network Link (HGNNLink), a requirements traceability approach that uses vectors generated by pre-trained models as node features and considers IR similarity and dependency relationships as edge features. By employing a heterogeneous graph neural network, HGNNLink aggregates and dynamically evaluates the impact of textual similarity and code dependencies on link generation. The experimental results show that HGNNLink improves the average F1 score by 13.36% compared to the current state-of-the-art (SOTA) method GA-XWCoDe in a dataset collected from ten open source software (OSS) projects. HGNNLink can extend IR methods by using high similarity candidate links as edges, and the extended HGNNLink achieves a 2.48% improvement in F1 compared to the original IR method after threshold parameter configuration using a genetic algorithm.

手动恢复需求和代码工件之间的可追溯性链接通常会消耗大量的人力资源。为了解决这个问题,研究人员提出了基于需求和代码工件之间的文本相似性的自动化方法,例如信息检索(IR)和预训练模型,以确定需求和代码工件之间是否存在可追溯性链接。然而,在相同的系统中,开发人员经常遵循相似的命名约定,并重复使用相同的框架和模板代码,从而导致功能不相关的代码工件之间的文本高度相似。这使得仅仅基于文本相似性的需求工件难以准确地识别相应的代码工件。因此,有必要利用代码工件之间的依赖关系来协助需求-代码可跟踪性链接恢复过程。现有的方法通常将依赖关系视为细化文本相似性的后处理步骤,忽略了文本相似性和依赖关系在生成需求-代码可追溯性链接中的重要性。为了解决这些限制,我们提出了异构图神经网络链接(HGNNLink),这是一种需求可追溯性方法,使用预训练模型生成的向量作为节点特征,并将IR相似性和依赖关系作为边缘特征。通过采用异构图神经网络,HGNNLink聚合并动态评估文本相似性和代码依赖性对链接生成的影响。实验结果表明,与目前最先进的GA-XWCoDe方法相比,HGNNLink在10个开源软件(OSS)项目数据集中的F1平均得分提高了13.36%。HGNNLink利用高相似度候选链路作为边缘对红外方法进行扩展,采用遗传算法配置阈值参数后,扩展后的HGNNLink在F1上比原红外方法提高了2.48%。
{"title":"HGNNLink: recovering requirements-code traceability links with text and dependency-aware heterogeneous graph neural networks","authors":"Bangchao Wang,&nbsp;Zhiyuan Zou,&nbsp;Xuanxuan Liang,&nbsp;Huan Jin,&nbsp;Peng Liang","doi":"10.1007/s10515-025-00528-2","DOIUrl":"10.1007/s10515-025-00528-2","url":null,"abstract":"<div><p>Manually recovering traceability links between requirements and code artifacts often consumes substantial human resources. To address this, researchers have proposed automated methods based on textual similarity between requirements and code artifacts, such as information retrieval (IR) and pre-trained models, to determine whether traceability links exist between requirements and code artifacts. However, in the same system, developers often follow similar naming conventions and repeatedly use the same frameworks and template code, resulting in high textual similarity between code artifacts that are functionally unrelated. This makes it difficult to accurately identify the corresponding code artifacts for requirements artifacts solely based on textual similarity. Therefore, it is necessary to leverage the dependency relationships between code artifacts to assist in the requirements-code traceability link recovery process. Existing methods often treat dependency relationships as a post-processing step to refine textual similarity, overlooking the importance of textual similarity and dependency relationships in generating requirements-code traceability links. To address these limitations, we proposed Heterogeneous Graph Neural Network Link (HGNNLink), a requirements traceability approach that uses vectors generated by pre-trained models as node features and considers IR similarity and dependency relationships as edge features. By employing a heterogeneous graph neural network, HGNNLink aggregates and dynamically evaluates the impact of textual similarity and code dependencies on link generation. The experimental results show that HGNNLink improves the average F1 score by 13.36% compared to the current state-of-the-art (SOTA) method GA-XWCoDe in a dataset collected from ten open source software (OSS) projects. HGNNLink can extend IR methods by using high similarity candidate links as edges, and the extended HGNNLink achieves a 2.48% improvement in F1 compared to the original IR method after threshold parameter configuration using a genetic algorithm.</p></div>","PeriodicalId":55414,"journal":{"name":"Automated Software Engineering","volume":"32 2","pages":""},"PeriodicalIF":3.1,"publicationDate":"2025-05-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145171197","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Continuous integration of architectural performance models with parametric dependencies – the CIPM approach 结构性能模型与参数依赖性的持续集成——CIPM方法
IF 3.1 2区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-05-29 DOI: 10.1007/s10515-025-00521-9
Manar Mazkatli, David Monschein, Martin Armbruster, Robert Heinrich, Anne Koziolek

The explicit consideration of the software architecture supports system evolution and efficient quality assurance. In particular, Architecture-based Performance Prediction (AbPP) assesses the performance for future scenarios (e.g., alternative workload, design, deployment) without expensive measurements for all such alternatives. However, accurate AbPP requires an up-to-date architectural Performance Model (aPM) that is parameterized over factors impacting the performance (e.g., input data characteristics). Especially in agile development, keeping such a parametric aPM consistent with software artifacts is challenging due to frequent evolutionary, adaptive, and usage-related changes. Existing approaches do not address the impact of all aforementioned changes. Moreover, the extraction of a complete aPM after each impacting change causes unnecessary monitoring overhead and may overwrite previous manual adjustments. In this article, we present the Continuous Integration of architectural Performance Model (CIPM) approach, which automatically updates a parametric aPM after each evolutionary, adaptive, or usage change. To reduce the monitoring overhead, CIPM only calibrates the affected performance parameters (e.g., resource demand) using adaptive monitoring. Moreover, a self-validation process in CIPM validates the accuracy, manages the monitoring to reduce overhead, and recalibrates inaccurate parts. Consequently, CIPM will automatically keep the aPM up-to-date throughout the development and operation, which enables AbPP for a proactive identification of upcoming performance problems and for evaluating alternatives at low costs. We evaluate the applicability of CIPM in terms of accuracy, monitoring overhead, and scalability using six cases (four Java-based open source applications and two industrial Lua-based sensor applications). Regarding accuracy, we observed that CIPM correctly keeps an aPM up-to-date and estimates performance parameters well so that it supports accurate performance predictions. Regarding the monitoring overhead in our experiments, CIPM’s adaptive instrumentation demonstrated a significant reduction in the number of required instrumentation probes, ranging from 12.6 % to 83.3 %, depending on the specific cases evaluated. Finally, we found out that CIPM’s execution time is reasonable and scales well with an increasing number of model elements and monitoring data.

对软件体系结构的明确考虑支持系统演化和有效的质量保证。特别是,基于体系结构的性能预测(AbPP)评估未来场景(例如,可选工作负载、设计、部署)的性能,而无需对所有这些可选方案进行昂贵的度量。然而,准确的AbPP需要一个最新的体系结构性能模型(aPM),该模型参数化了影响性能的因素(例如,输入数据特征)。特别是在敏捷开发中,由于频繁的进化、自适应和与使用相关的更改,保持这样一个参数化aPM与软件工件的一致性是具有挑战性的。现有的方法不能解决上述所有变化的影响。此外,在每次影响更改之后提取完整的aPM会导致不必要的监视开销,并可能覆盖以前的手动调整。在本文中,我们介绍了体系结构性能模型的持续集成(Continuous Integration of architectural Performance Model, CIPM)方法,该方法在每次进化、自适应或使用变化之后自动更新参数aPM。为了减少监视开销,CIPM只使用自适应监视校准受影响的性能参数(例如,资源需求)。此外,CIPM中的自验证过程验证准确性,管理监视以减少开销,并重新校准不准确的部件。因此,在整个开发和操作过程中,CIPM将自动使aPM保持最新状态,这使AbPP能够主动识别即将出现的性能问题,并以低成本评估替代方案。我们使用六个案例(四个基于java的开源应用程序和两个基于lua的工业传感器应用程序)来评估CIPM在准确性、监视开销和可伸缩性方面的适用性。关于准确性,我们观察到CIPM正确地使aPM保持最新状态,并很好地估计性能参数,从而支持准确的性能预测。关于我们实验中的监控开销,CIPM的自适应仪器显示所需仪器探针的数量显著减少,范围从12.6%到83.3%,具体取决于评估的具体情况。最后,我们发现CIPM的执行时间是合理的,并且随着模型元素和监控数据数量的增加而具有良好的可扩展性。
{"title":"Continuous integration of architectural performance models with parametric dependencies – the CIPM approach","authors":"Manar Mazkatli,&nbsp;David Monschein,&nbsp;Martin Armbruster,&nbsp;Robert Heinrich,&nbsp;Anne Koziolek","doi":"10.1007/s10515-025-00521-9","DOIUrl":"10.1007/s10515-025-00521-9","url":null,"abstract":"<p>The explicit consideration of the software architecture supports system evolution and efficient quality assurance. In particular, Architecture-based Performance Prediction (AbPP) assesses the performance for future scenarios (e.g., alternative workload, design, deployment) without expensive measurements for all such alternatives. However, accurate AbPP requires an up-to-date architectural Performance Model (aPM) that is parameterized over factors impacting the performance (e.g., input data characteristics). Especially in agile development, keeping such a parametric aPM consistent with software artifacts is challenging due to frequent evolutionary, adaptive, and usage-related changes. Existing approaches do not address the impact of all aforementioned changes. Moreover, the extraction of a complete aPM after each impacting change causes unnecessary monitoring overhead and may overwrite previous manual adjustments. In this article, we present the Continuous Integration of architectural Performance Model (CIPM) approach, which automatically updates a parametric aPM after each evolutionary, adaptive, or usage change. To reduce the monitoring overhead, CIPM only calibrates the affected performance parameters (e.g., resource demand) using adaptive monitoring. Moreover, a self-validation process in CIPM validates the accuracy, manages the monitoring to reduce overhead, and recalibrates inaccurate parts. Consequently, CIPM will automatically keep the aPM up-to-date throughout the development and operation, which enables AbPP for a proactive identification of upcoming performance problems and for evaluating alternatives at low costs. We evaluate the applicability of CIPM in terms of accuracy, monitoring overhead, and scalability using six cases (four Java-based open source applications and two industrial Lua-based sensor applications). Regarding accuracy, we observed that CIPM correctly keeps an aPM up-to-date and estimates performance parameters well so that it supports accurate performance predictions. Regarding the monitoring overhead in our experiments, CIPM’s adaptive instrumentation demonstrated a significant reduction in the number of required instrumentation probes, ranging from 12.6 % to 83.3 %, depending on the specific cases evaluated. Finally, we found out that CIPM’s execution time is reasonable and scales well with an increasing number of model elements and monitoring data.</p>","PeriodicalId":55414,"journal":{"name":"Automated Software Engineering","volume":"32 2","pages":""},"PeriodicalIF":3.1,"publicationDate":"2025-05-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10515-025-00521-9.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145171381","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Understanding the privacy-realisticness dilemma of the metaverse 理解虚拟世界的隐私-现实困境
IF 2 2区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-05-26 DOI: 10.1007/s10515-025-00516-6
Xiaolu Zhang, Tahmid Rafi, Yuejun Guan, Shuqing Li, Michael R. Lyu

Metaverse is a form of next-generation human–computer interaction and social networks based on virtual and augmented reality. Both the research and industry community have invested much in this area to develop useful applications and enhance user experience. Meanwhile, the expanded human–computer interface which enables the immersive experience in the Metaverse will also inevitably expand the interface of potential privacy leaks. This dilemma between immersive user experience and higher privacy risks has not been well studied and it is not clear how different users would make decisions when facing such a dilemma. In this research work, we systematically studied this dilemma in different usage scenarios of the Metaverse and performed a study on 177 users to understand the factors that may affect users’ decision making. From the study, we found that user preference on immersive experience and privacy protection can be very different in different usage scenarios and we expect our study results can provide some insights and guidance for the design of privacy protection mechanisms in Metaverse platforms and applications.

虚拟世界是基于虚拟和增强现实的下一代人机交互和社交网络的一种形式。为了开发有用的应用程序和增强用户体验,研究和工业界都在这一领域投入了大量资金。与此同时,扩展的人机界面使虚拟世界的沉浸式体验成为可能,也不可避免地扩大了潜在隐私泄露的界面。这种沉浸式用户体验和更高隐私风险之间的困境尚未得到很好的研究,也不清楚不同的用户在面对这种困境时会如何做出决定。在本研究中,我们系统地研究了在不同的虚拟世界使用场景下的这一困境,并对177名用户进行了研究,以了解可能影响用户决策的因素。通过研究,我们发现用户对沉浸式体验和隐私保护的偏好在不同的使用场景下会有很大的差异,我们希望我们的研究结果可以为Metaverse平台和应用中隐私保护机制的设计提供一些见解和指导。
{"title":"Understanding the privacy-realisticness dilemma of the metaverse","authors":"Xiaolu Zhang,&nbsp;Tahmid Rafi,&nbsp;Yuejun Guan,&nbsp;Shuqing Li,&nbsp;Michael R. Lyu","doi":"10.1007/s10515-025-00516-6","DOIUrl":"10.1007/s10515-025-00516-6","url":null,"abstract":"<div><p>Metaverse is a form of next-generation human–computer interaction and social networks based on virtual and augmented reality. Both the research and industry community have invested much in this area to develop useful applications and enhance user experience. Meanwhile, the expanded human–computer interface which enables the immersive experience in the Metaverse will also inevitably expand the interface of potential privacy leaks. This dilemma between immersive user experience and higher privacy risks has not been well studied and it is not clear how different users would make decisions when facing such a dilemma. In this research work, we systematically studied this dilemma in different usage scenarios of the Metaverse and performed a study on 177 users to understand the factors that may affect users’ decision making. From the study, we found that user preference on immersive experience and privacy protection can be very different in different usage scenarios and we expect our study results can provide some insights and guidance for the design of privacy protection mechanisms in Metaverse platforms and applications.</p></div>","PeriodicalId":55414,"journal":{"name":"Automated Software Engineering","volume":"32 2","pages":""},"PeriodicalIF":2.0,"publicationDate":"2025-05-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144135273","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Automated Software Engineering
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1