首页 > 最新文献

IEEE Transactions on Big Data最新文献

英文 中文
Guest Editorial Special Issue on Federated Learning for Big Data Applications 大数据应用的联邦学习特刊
IF 5.7 3区 计算机科学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2025-09-03 DOI: 10.1109/TBDATA.2024.3417057
Xiaowen Chu;Wei Wang;Cong Wang;Yang Liu;Rongfei Zeng;Christopher G. Brinton
{"title":"Guest Editorial Special Issue on Federated Learning for Big Data Applications","authors":"Xiaowen Chu;Wei Wang;Cong Wang;Yang Liu;Rongfei Zeng;Christopher G. Brinton","doi":"10.1109/TBDATA.2024.3417057","DOIUrl":"https://doi.org/10.1109/TBDATA.2024.3417057","url":null,"abstract":"","PeriodicalId":13106,"journal":{"name":"IEEE Transactions on Big Data","volume":"11 5","pages":"2099-2101"},"PeriodicalIF":5.7,"publicationDate":"2025-09-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11149636","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144990054","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multi-Modal Entity in One Word: Aligning Multi-Level Semantics for Multi-Modal Knowledge Graph Completion 一个词中的多模态实体:多模态知识图补全的多层语义对齐
IF 5.7 3区 计算机科学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2025-08-19 DOI: 10.1109/TBDATA.2025.3600014
Lan Zhao;Boyue Wang;Junbin Gao;Xiaoyan Li;Yongli Hu;Baocai Yin
Current multi-modal knowledge graph completion often incorporates simple fusion neural networks to achieve multi-modal alignment and knowledge completion tasks, which face three major challenges: 1) Inconsistent semantics between images and texts corresponding to the same entity; 2) Discrepancies in semantic spaces resulting from the use of diverse uni-modal feature extractors; 3) Inadequate evaluation of semantic alignment using only energy functions or basic contrastive learning losses. To address these challenges, we propose the Multi-modal Entity in One Word (MEOW) model. This model ensures alignment at various levels, including text-image match alignment, feature alignment and distribution alignment. Specificially, the entity image filtering module utilizes a visual-language model to exclude unrelated images by aligning their captions with corresponding text descriptions. A pre-trained CLIP-based encoder is utilized for encoding dense semantic relationships, while a graph attention network based structure encoder handles sparse semantic relationships, yielding a comprehensive semantic representation and enhancing convergence speed. Additionally, a diffusion model is integrated to enhance denoising capabilities. The proposed MEOW further includes a distribution alignment module equipped with dense alignment constraint, integrity alignment constraint, and fusion fidelity constraint to effectively align multi-modal representations. Experiments on two public multi-modal knowledge graph datasets show that MEOW significantly improves link prediction performance.
当前的多模态知识图谱补全通常采用简单的融合神经网络来完成多模态对齐和知识补全任务,这面临着三个主要挑战:1)同一实体对应的图像和文本之间语义不一致;2)使用不同的单模态特征提取器导致的语义空间差异;3)仅使用能量函数或基本对比学习损失对语义对齐的评价不足。为了应对这些挑战,我们提出了一个词中的多模态实体(MEOW)模型。该模型保证了不同层次的对齐,包括文本-图像匹配对齐、特征对齐和分布对齐。具体来说,实体图像过滤模块利用一种视觉语言模型,通过将图像的标题与相应的文本描述对齐来排除不相关的图像。利用预训练的基于clip的编码器对密集语义关系进行编码,利用基于图注意网络的结构编码器对稀疏语义关系进行编码,得到了全面的语义表示,提高了收敛速度。此外,还集成了扩散模型来增强去噪能力。提出的MEOW进一步包括一个分布对齐模块,该模块配备密集对齐约束、完整性对齐约束和融合保真度约束,以有效地对齐多模态表示。在两个公共多模态知识图数据集上的实验表明,MEOW显著提高了链路预测性能。
{"title":"Multi-Modal Entity in One Word: Aligning Multi-Level Semantics for Multi-Modal Knowledge Graph Completion","authors":"Lan Zhao;Boyue Wang;Junbin Gao;Xiaoyan Li;Yongli Hu;Baocai Yin","doi":"10.1109/TBDATA.2025.3600014","DOIUrl":"https://doi.org/10.1109/TBDATA.2025.3600014","url":null,"abstract":"Current multi-modal knowledge graph completion often incorporates simple fusion neural networks to achieve multi-modal alignment and knowledge completion tasks, which face three major challenges: 1) Inconsistent semantics between images and texts corresponding to the same entity; 2) Discrepancies in semantic spaces resulting from the use of diverse uni-modal feature extractors; 3) Inadequate evaluation of semantic alignment using only energy functions or basic contrastive learning losses. To address these challenges, we propose the Multi-modal Entity in One Word (MEOW) model. This model ensures alignment at various levels, including text-image match alignment, feature alignment and distribution alignment. Specificially, the entity image filtering module utilizes a visual-language model to exclude unrelated images by aligning their captions with corresponding text descriptions. A pre-trained CLIP-based encoder is utilized for encoding dense semantic relationships, while a graph attention network based structure encoder handles sparse semantic relationships, yielding a comprehensive semantic representation and enhancing convergence speed. Additionally, a diffusion model is integrated to enhance denoising capabilities. The proposed MEOW further includes a distribution alignment module equipped with dense alignment constraint, integrity alignment constraint, and fusion fidelity constraint to effectively align multi-modal representations. Experiments on two public multi-modal knowledge graph datasets show that MEOW significantly improves link prediction performance.","PeriodicalId":13106,"journal":{"name":"IEEE Transactions on Big Data","volume":"11 6","pages":"3539-3552"},"PeriodicalIF":5.7,"publicationDate":"2025-08-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145493294","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Editorial High-Performance Recommender Systems Based on Spatiotemporal Data 基于时空数据的编辑高性能推荐系统
IF 7.5 3区 计算机科学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2025-07-10 DOI: 10.1109/TBDATA.2024.3451088
Shuo Shang;Qi Liu;Renhe Jiang;Ryosuke Shibasaki;Panos Kalnis;Christian S. Jensen
{"title":"Editorial High-Performance Recommender Systems Based on Spatiotemporal Data","authors":"Shuo Shang;Qi Liu;Renhe Jiang;Ryosuke Shibasaki;Panos Kalnis;Christian S. Jensen","doi":"10.1109/TBDATA.2024.3451088","DOIUrl":"https://doi.org/10.1109/TBDATA.2024.3451088","url":null,"abstract":"","PeriodicalId":13106,"journal":{"name":"IEEE Transactions on Big Data","volume":"11 4","pages":"1588-1588"},"PeriodicalIF":7.5,"publicationDate":"2025-07-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11077801","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144597786","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Editorial: Big Data Analytics in Complex Social Information Networks 社论:复杂社会信息网络中的大数据分析
IF 7.5 3区 计算机科学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2025-07-10 DOI: 10.1109/TBDATA.2024.3485316
Desheng Dash Wu;David L. Olson
This special issue deals with research related to applications of and methods to support Big Data analytics in complex social information networks. The digital age and the rise of social media have sped up changes to social systems with unforeseen consequences. However, there are major challenges created.
本期特刊讨论了在复杂的社会信息网络中支持大数据分析的应用和方法。数字时代和社交媒体的兴起加速了社会体系的变革,带来了不可预见的后果。然而,也产生了重大挑战。
{"title":"Editorial: Big Data Analytics in Complex Social Information Networks","authors":"Desheng Dash Wu;David L. Olson","doi":"10.1109/TBDATA.2024.3485316","DOIUrl":"https://doi.org/10.1109/TBDATA.2024.3485316","url":null,"abstract":"This special issue deals with research related to applications of and methods to support Big Data analytics in complex social information networks. The digital age and the rise of social media have sped up changes to social systems with unforeseen consequences. However, there are major challenges created.","PeriodicalId":13106,"journal":{"name":"IEEE Transactions on Big Data","volume":"11 4","pages":"1650-1651"},"PeriodicalIF":7.5,"publicationDate":"2025-07-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11077792","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144598007","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
GE-GNN: Gated Edge-Augmented Graph Neural Network for Fraud Detection 用于欺诈检测的门控边缘增强图神经网络
IF 7.5 3区 计算机科学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2025-04-18 DOI: 10.1109/TBDATA.2025.3562486
Wenxin Zhang;Cuicui Luo
Graph Neural Networks (GNNs) play a significant role and have been widely applied in fraud detection tasks, exhibiting substantial improvements in detection performance compared to conventional methodologies. However, within the intricate structure of fraud graphs, fraudsters usually camouflage themselves among a large number of benign entities. An effective solution to address the camouflage problem involves the incorporation of complex and abundant edge information. Nevertheless, existing GNN-based methods frequently neglect to integrate this crucial information into the message passing process, thereby limiting their efficacy. To address the above issues, this study proposes a novel Gated Edge-augmented Graph Neural Network(GE-GNN). Our approach begins with an edge-based feature augmentation mechanism that leverages both node and edge features within a single relation. Subsequently, we apply the augmented representation to the message passing process to update the node embeddings. Furthermore, we design a gate logistic to regulate the expression of augmented information. Finally, we integrate node features across different relations to obtain a comprehensive representation. Extensive experimental results on two real-world datasets demonstrate that the proposed method outperforms several state-of-the-art methods.
图神经网络(gnn)在欺诈检测任务中发挥着重要作用,并已被广泛应用,与传统方法相比,在检测性能方面有了实质性的改进。然而,在错综复杂的欺诈图结构中,欺诈者往往隐藏在大量良性实体之中。利用复杂而丰富的边缘信息是解决伪装问题的有效方法。然而,现有的基于gnn的方法经常忽略将这一关键信息集成到消息传递过程中,从而限制了它们的有效性。为了解决上述问题,本研究提出了一种新的门控边缘增强图神经网络(GE-GNN)。我们的方法从基于边缘的特征增强机制开始,该机制在单个关系中利用节点和边缘特征。随后,我们将增强表示应用到消息传递过程中,以更新节点嵌入。此外,我们还设计了一个门逻辑来调节增广信息的表达。最后,对不同关系的节点特征进行整合,得到一个综合的表示。在两个真实数据集上的广泛实验结果表明,所提出的方法优于几种最先进的方法。
{"title":"GE-GNN: Gated Edge-Augmented Graph Neural Network for Fraud Detection","authors":"Wenxin Zhang;Cuicui Luo","doi":"10.1109/TBDATA.2025.3562486","DOIUrl":"https://doi.org/10.1109/TBDATA.2025.3562486","url":null,"abstract":"Graph Neural Networks (GNNs) play a significant role and have been widely applied in fraud detection tasks, exhibiting substantial improvements in detection performance compared to conventional methodologies. However, within the intricate structure of fraud graphs, fraudsters usually camouflage themselves among a large number of benign entities. An effective solution to address the camouflage problem involves the incorporation of complex and abundant edge information. Nevertheless, existing GNN-based methods frequently neglect to integrate this crucial information into the message passing process, thereby limiting their efficacy. To address the above issues, this study proposes a novel Gated Edge-augmented Graph Neural Network(GE-GNN). Our approach begins with an edge-based feature augmentation mechanism that leverages both node and edge features within a single relation. Subsequently, we apply the augmented representation to the message passing process to update the node embeddings. Furthermore, we design a gate logistic to regulate the expression of augmented information. Finally, we integrate node features across different relations to obtain a comprehensive representation. Extensive experimental results on two real-world datasets demonstrate that the proposed method outperforms several state-of-the-art methods.","PeriodicalId":13106,"journal":{"name":"IEEE Transactions on Big Data","volume":"11 4","pages":"1664-1676"},"PeriodicalIF":7.5,"publicationDate":"2025-04-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144598065","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Topology-Based Node-Level Membership Inference Attacks on Graph Neural Networks 基于拓扑的图神经网络节点级隶属推理攻击
IF 5.7 3区 计算机科学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2025-04-08 DOI: 10.1109/TBDATA.2025.3558855
Faqian Guan;Tianqing Zhu;Wanlei Zhou;Philip S. Yu
Graph neural networks (GNNs) have obtained considerable attention due to their ability to leverage the inherent topological and node information present in graph data. While extensive research has been conducted on privacy attacks targeting machine learning models, the exploration of privacy risks associated with node-level membership inference attacks on GNNs remains relatively limited. GNNs learn representations that encapsulate valuable information about the nodes. These learned representations can be exploited by attackers to infer whether a specific node belongs to the training dataset, leading to the disclosure of sensitive information. The insidious nature of such privacy breaches often leads to an underestimation of the associated risks. Furthermore, the inherent challenges posed by node membership inference attacks make it difficult to develop effective attack models for GNNs that can successfully infer node membership. We propose a more efficient approach that specifically targets node-level membership inference attacks on GNNs. Initially, we combine nodes and their respective neighbors to carry out node membership inference attacks. To address the challenge of variable-length features arising from the differing number of neighboring nodes, we introduce an effective feature processing strategy. Furthermore, we propose two strategies: multiple training of shadow models and random selection of non-membership data, to enhance the performance of the attack model. We empirically evaluate the efficacy of our proposed method using three benchmark datasets. Additionally, we explore two potential defense mechanisms against node-level membership inference attacks.
图神经网络(gnn)由于能够利用图数据中存在的固有拓扑和节点信息而获得了相当大的关注。虽然针对机器学习模型的隐私攻击进行了广泛的研究,但对gnn上节点级成员推理攻击相关的隐私风险的探索仍然相对有限。gnn学习封装有关节点的有价值信息的表示。这些学习到的表示可以被攻击者利用来推断特定节点是否属于训练数据集,从而导致敏感信息的泄露。此类隐私泄露的隐蔽性往往导致对相关风险的低估。此外,节点隶属度推理攻击所带来的固有挑战使得很难开发出能够成功推断节点隶属度的gnn有效攻击模型。我们提出了一种更有效的方法,专门针对gnn的节点级成员推理攻击。最初,我们将节点和它们各自的邻居结合起来进行节点隶属推理攻击。为了解决因相邻节点数量不同而产生的变长特征的挑战,我们引入了一种有效的特征处理策略。为了提高攻击模型的性能,我们提出了多重训练阴影模型和随机选择非隶属性数据两种策略。我们使用三个基准数据集对我们提出的方法的有效性进行了实证评估。此外,我们还探讨了针对节点级成员推理攻击的两种潜在防御机制。
{"title":"Topology-Based Node-Level Membership Inference Attacks on Graph Neural Networks","authors":"Faqian Guan;Tianqing Zhu;Wanlei Zhou;Philip S. Yu","doi":"10.1109/TBDATA.2025.3558855","DOIUrl":"https://doi.org/10.1109/TBDATA.2025.3558855","url":null,"abstract":"Graph neural networks (GNNs) have obtained considerable attention due to their ability to leverage the inherent topological and node information present in graph data. While extensive research has been conducted on privacy attacks targeting machine learning models, the exploration of privacy risks associated with node-level membership inference attacks on GNNs remains relatively limited. GNNs learn representations that encapsulate valuable information about the nodes. These learned representations can be exploited by attackers to infer whether a specific node belongs to the training dataset, leading to the disclosure of sensitive information. The insidious nature of such privacy breaches often leads to an underestimation of the associated risks. Furthermore, the inherent challenges posed by node membership inference attacks make it difficult to develop effective attack models for GNNs that can successfully infer node membership. We propose a more efficient approach that specifically targets node-level membership inference attacks on GNNs. Initially, we combine nodes and their respective neighbors to carry out node membership inference attacks. To address the challenge of variable-length features arising from the differing number of neighboring nodes, we introduce an effective feature processing strategy. Furthermore, we propose two strategies: multiple training of shadow models and random selection of non-membership data, to enhance the performance of the attack model. We empirically evaluate the efficacy of our proposed method using three benchmark datasets. Additionally, we explore two potential defense mechanisms against node-level membership inference attacks.","PeriodicalId":13106,"journal":{"name":"IEEE Transactions on Big Data","volume":"11 5","pages":"2809-2826"},"PeriodicalIF":5.7,"publicationDate":"2025-04-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144934350","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Revocable DSSE in Healthcare Systems With Range Query Support 具有范围查询支持的医疗保健系统中可撤销的DSSE
IF 5.7 3区 计算机科学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2025-04-01 DOI: 10.1109/TBDATA.2025.3556636
Hanqi Zhang;Yandong Zheng;Chang Xu;Liehuang Zhu;Jiayin Wang
With the rapid development of cloud computing, online health monitoring systems are becoming increasingly prevalent. To protect medical data privacy while supporting search operations, Dynamic Searchable Symmetric Encryption (DSSE) technology has been widely used in health monitoring systems. For better monitoring of patient status, keyword range query is also a necessary requirement for the DSSE scheme. Furthermore, in the multi-user setting, user revocation usually leads the owner to download and re-encrypt all indexes, resulting in significant computational overhead. In this paper, we propose a lightweight revocable DSSE scheme with range query support. First, we propose a novel and privacy-preserving range query algorithm that defends plaintext inference attacks. Second, we design a singly linked list structure based on delegatable pseudorandom functions and key-updatable pseudorandom functions, which support lightweight user revocation. Rigorous security analysis proves the security of our proposed range query scheme and demonstrates that our scheme can achieve forward and backward privacy. Experimental evaluations show that our scheme is highly efficient.
随着云计算的快速发展,在线健康监测系统越来越普遍。为了在支持搜索操作的同时保护医疗数据隐私,动态可搜索对称加密(DSSE)技术已广泛应用于健康监测系统中。为了更好地监测患者状态,关键字范围查询也是DSSE方案的必要要求。此外,在多用户设置中,用户撤销通常会导致所有者下载并重新加密所有索引,从而导致大量的计算开销。本文提出了一种支持范围查询的轻量级可撤销DSSE方案。首先,我们提出了一种新的保护隐私的范围查询算法来防御明文推理攻击。其次,基于可委派伪随机函数和可键更新伪随机函数设计了支持轻量级用户撤销的单链表结构。严格的安全性分析证明了我们提出的范围查询方案的安全性,并证明了我们的方案可以实现前向和后向隐私。实验结果表明,该方案是高效的。
{"title":"Revocable DSSE in Healthcare Systems With Range Query Support","authors":"Hanqi Zhang;Yandong Zheng;Chang Xu;Liehuang Zhu;Jiayin Wang","doi":"10.1109/TBDATA.2025.3556636","DOIUrl":"https://doi.org/10.1109/TBDATA.2025.3556636","url":null,"abstract":"With the rapid development of cloud computing, online health monitoring systems are becoming increasingly prevalent. To protect medical data privacy while supporting search operations, Dynamic Searchable Symmetric Encryption (DSSE) technology has been widely used in health monitoring systems. For better monitoring of patient status, keyword range query is also a necessary requirement for the DSSE scheme. Furthermore, in the multi-user setting, user revocation usually leads the owner to download and re-encrypt all indexes, resulting in significant computational overhead. In this paper, we propose a lightweight revocable DSSE scheme with range query support. First, we propose a novel and privacy-preserving range query algorithm that defends plaintext inference attacks. Second, we design a singly linked list structure based on delegatable pseudorandom functions and key-updatable pseudorandom functions, which support lightweight user revocation. Rigorous security analysis proves the security of our proposed range query scheme and demonstrates that our scheme can achieve forward and backward privacy. Experimental evaluations show that our scheme is highly efficient.","PeriodicalId":13106,"journal":{"name":"IEEE Transactions on Big Data","volume":"11 5","pages":"2764-2778"},"PeriodicalIF":5.7,"publicationDate":"2025-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144934474","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Utility-Driven Data Analytics Algorithm for Transaction Modifications Using Pre-Large Concept With Single Database Scan 使用单个数据库扫描的Pre-Large概念的事务修改的效用驱动数据分析算法
IF 5.7 3区 计算机科学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2025-04-01 DOI: 10.1109/TBDATA.2025.3556615
Unil Yun;Hanju Kim;Myungha Cho;Taewoong Ryu;Seungwan Park;Doyoon Kim;Doyoung Kim;Chanhee Lee;Witold Pedrycz
Utility-driven pattern analysis is a fundamental method for analyzing noteworthy patterns with high utility for diverse quantitative transactional databases. Recently, various approaches have emerged to handle large, dynamic database environments more efficiently by reducing the number of data scans and pattern expansion operations with the pre-large concept. However, existing pre-large-based high utility pattern mining methods either fail to handle real-time transaction modifications or require additional data scans to validate candidate patterns. In this paper, we propose a novel efficient utility-driven pattern mining algorithm using the pre-large concept for transaction modifications. Our method incorporates a single-scan-based framework through the management of actual utility values and discovers high utility patterns without candidate generation for efficient utility-driven dynamic data analysis in the modification environment. We compared the performance of the proposed method with state-of-the-art methods through extensive performance evaluation utilizing real and synthetic datasets. According to the evaluation results and a case study, the suggested method performs a minimum of 1.5 times faster than state-of-the-art methods alongside minimal compromise in memory, and it scaled well with increases in database size. Further statistical analyses indicate that the proposed method reduces the pattern search space compared to the previous method while delivering a complete set of accurate results without loss.
效用驱动的模式分析是分析各种定量事务数据库中具有高效用的重要模式的基本方法。最近,出现了各种方法,通过使用pre-large概念减少数据扫描和模式展开操作的数量,从而更有效地处理大型动态数据库环境。然而,现有的pre-large-based高实用模式挖掘方法要么无法处理实时事务修改,要么需要额外的数据扫描来验证候选模式。在本文中,我们提出了一种新的高效实用驱动的模式挖掘算法,该算法使用pre-large概念进行事务修改。我们的方法结合了一个基于单一扫描的框架,通过对实际效用值的管理,发现高效用模式,而不需要在修改环境中为有效的效用驱动的动态数据分析生成候选模式。我们通过利用真实和合成数据集进行广泛的性能评估,将所提出的方法的性能与最先进的方法进行了比较。根据评估结果和一个案例研究,建议的方法的执行速度比最先进的方法至少快1.5倍,同时对内存的损害最小,并且随着数据库大小的增加而扩展得很好。进一步的统计分析表明,与之前的方法相比,所提出的方法减少了模式搜索空间,同时提供了一组完整的准确结果而没有损失。
{"title":"Utility-Driven Data Analytics Algorithm for Transaction Modifications Using Pre-Large Concept With Single Database Scan","authors":"Unil Yun;Hanju Kim;Myungha Cho;Taewoong Ryu;Seungwan Park;Doyoon Kim;Doyoung Kim;Chanhee Lee;Witold Pedrycz","doi":"10.1109/TBDATA.2025.3556615","DOIUrl":"https://doi.org/10.1109/TBDATA.2025.3556615","url":null,"abstract":"Utility-driven pattern analysis is a fundamental method for analyzing noteworthy patterns with high utility for diverse quantitative transactional databases. Recently, various approaches have emerged to handle large, dynamic database environments more efficiently by reducing the number of data scans and pattern expansion operations with the pre-large concept. However, existing pre-large-based high utility pattern mining methods either fail to handle real-time transaction modifications or require additional data scans to validate candidate patterns. In this paper, we propose a novel efficient utility-driven pattern mining algorithm using the pre-large concept for transaction modifications. Our method incorporates a single-scan-based framework through the management of actual utility values and discovers high utility patterns without candidate generation for efficient utility-driven dynamic data analysis in the modification environment. We compared the performance of the proposed method with state-of-the-art methods through extensive performance evaluation utilizing real and synthetic datasets. According to the evaluation results and a case study, the suggested method performs a minimum of 1.5 times faster than state-of-the-art methods alongside minimal compromise in memory, and it scaled well with increases in database size. Further statistical analyses indicate that the proposed method reduces the pattern search space compared to the previous method while delivering a complete set of accurate results without loss.","PeriodicalId":13106,"journal":{"name":"IEEE Transactions on Big Data","volume":"11 5","pages":"2792-2808"},"PeriodicalIF":5.7,"publicationDate":"2025-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144934407","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Novel Concept-Cognitive Learning Model Oriented to Three-Way Concept for Knowledge Acquisition 一种新的概念认知学习模式——面向知识获取的三向概念
IF 5.7 3区 计算机科学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2025-04-01 DOI: 10.1109/TBDATA.2025.3556637
Weihua Xu;Di Jiang
Concept-cognitive learning (CCL) is the process of enabling machines to simulate the concept learning of the human brain. Existing CCL models focus on formal context while neglecting the importance of skill context. Furthermore, CCL models, which solely focus on positive information, restrict the learning capacity by neglecting negative information, and greatly impeding the acquisition of knowledge. To overcome these issues, we proposes a novel concept-cognitive learning model oriented to three-way concept for knowledge acquisition. First, this paper explains and investigates the relationship between skills and knowledge based on the three-way concept and its properties. Then, in order to simultaneously consider positive and negative information, describe more detailed information, learn more skills, and acquire accurate knowledge, a three-way information granule is described from the perspective of cognitive learning. Then, a transformation method is proposed to transform between different three-way information granules, allowing for the transformation of arbitrary three-way information granule into necessary, sufficient, sufficient and necessary three-way information granules. Finally, algorithm corresponding to the transformation method is designed, and subsequently tested across diverse UCI datasets. The experimental outcomes affirm the effectiveness and excellence of the suggested model and algorithm.
概念认知学习(CCL)是使机器能够模拟人脑概念学习的过程。现有的CCL模型侧重于形式语境,而忽视了技能语境的重要性。此外,CCL模型只关注正面信息,忽略了负面信息,限制了学习能力,极大地阻碍了知识的获取。为了克服这些问题,我们提出了一种新的基于三向概念的知识获取概念认知学习模型。首先,本文从“三向”概念及其性质出发,对技能与知识的关系进行了解释和探讨。然后,为了同时考虑正面和负面信息,描述更详细的信息,学习更多的技能,获得准确的知识,从认知学习的角度描述了一个三向信息颗粒。然后,提出了一种在不同三向信息粒之间转换的转换方法,实现了任意三向信息粒向必要、充分、充分、必要四向信息粒的转换。最后,设计了与转换方法相对应的算法,并在不同的UCI数据集上进行了测试。实验结果证实了所提模型和算法的有效性和优越性。
{"title":"A Novel Concept-Cognitive Learning Model Oriented to Three-Way Concept for Knowledge Acquisition","authors":"Weihua Xu;Di Jiang","doi":"10.1109/TBDATA.2025.3556637","DOIUrl":"https://doi.org/10.1109/TBDATA.2025.3556637","url":null,"abstract":"Concept-cognitive learning (CCL) is the process of enabling machines to simulate the concept learning of the human brain. Existing CCL models focus on formal context while neglecting the importance of skill context. Furthermore, CCL models, which solely focus on positive information, restrict the learning capacity by neglecting negative information, and greatly impeding the acquisition of knowledge. To overcome these issues, we proposes a novel concept-cognitive learning model oriented to three-way concept for knowledge acquisition. First, this paper explains and investigates the relationship between skills and knowledge based on the three-way concept and its properties. Then, in order to simultaneously consider positive and negative information, describe more detailed information, learn more skills, and acquire accurate knowledge, a three-way information granule is described from the perspective of cognitive learning. Then, a transformation method is proposed to transform between different three-way information granules, allowing for the transformation of arbitrary three-way information granule into necessary, sufficient, sufficient and necessary three-way information granules. Finally, algorithm corresponding to the transformation method is designed, and subsequently tested across diverse UCI datasets. The experimental outcomes affirm the effectiveness and excellence of the suggested model and algorithm.","PeriodicalId":13106,"journal":{"name":"IEEE Transactions on Big Data","volume":"11 5","pages":"2779-2791"},"PeriodicalIF":5.7,"publicationDate":"2025-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144934563","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
PViTGAtt-IP: Severity Quantification of Lung Infections in Chest X-Rays and CT Scans via Parallel and Cross-Attended Encoders PViTGAtt-IP:通过平行和交叉编码器对胸部x线和CT扫描中肺部感染的严重程度进行量化
IF 5.7 3区 计算机科学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2025-03-31 DOI: 10.1109/TBDATA.2025.3556612
Bouthaina Slika;Fadi Dornaika;Fares Bougourzi;Karim Hammoudi
The development of a robust and adaptive deep learning technique for the diagnosis of pneumonia and the assessment of its severity was a major challenge. Indeed, both chest X-rays (CXR) and CT scans have been widely studied for the diagnosis, detection and quantification of pneumonia. In this paper, a novel approach (PViTGAtt-IP) based on a parallel array of vision transformers is presented, in which the input image is divided into regions of interest. Each region is fed into an individual model and the collective output gives the severity score. Three parallel architectures were also derived and tested. The proposed models were subjected to rigorous tests on two different datasets: RALO CXRs and Per COVID-19 CT scans. The experimental results showed that the proposed models exhibited high performance in accurately predicting scores for both datasets. In particular, the parallel transformers with multi-gate attention proved to be the best performing model. Furthermore, a comparative analysis using state-of-the-art methods showed that our proposed approach consistently achieved competitive or even better performance in terms of the Mean Absolute Error (MAE) and the Pearson Correlation Coefficient (PC). This emphasizes the effectiveness and superiority of our models in the context of diagnosing and assessing the severity of pneumonia.
开发一种用于肺炎诊断和严重程度评估的强大且自适应的深度学习技术是一项重大挑战。事实上,胸部x光片(CXR)和CT扫描已被广泛研究用于肺炎的诊断、检测和量化。本文提出了一种基于视觉变压器并行阵列的PViTGAtt-IP方法,该方法将输入图像划分为感兴趣的区域。每个地区都被输入到一个单独的模型中,集体输出给出了严重程度评分。还推导并测试了三种并行架构。提出的模型在两个不同的数据集上进行了严格的测试:RALO cxr和Per COVID-19 CT扫描。实验结果表明,所提出的模型在准确预测两个数据集的分数方面表现出很高的性能。其中,具有多栅极关注的并联变压器是性能最好的模型。此外,使用最先进的方法进行的比较分析表明,我们提出的方法在平均绝对误差(MAE)和皮尔逊相关系数(PC)方面始终具有竞争力甚至更好的性能。这强调了我们的模型在诊断和评估肺炎严重程度方面的有效性和优越性。
{"title":"PViTGAtt-IP: Severity Quantification of Lung Infections in Chest X-Rays and CT Scans via Parallel and Cross-Attended Encoders","authors":"Bouthaina Slika;Fadi Dornaika;Fares Bougourzi;Karim Hammoudi","doi":"10.1109/TBDATA.2025.3556612","DOIUrl":"https://doi.org/10.1109/TBDATA.2025.3556612","url":null,"abstract":"The development of a robust and adaptive deep learning technique for the diagnosis of pneumonia and the assessment of its severity was a major challenge. Indeed, both chest X-rays (CXR) and CT scans have been widely studied for the diagnosis, detection and quantification of pneumonia. In this paper, a novel approach (PViTGAtt-IP) based on a parallel array of vision transformers is presented, in which the input image is divided into regions of interest. Each region is fed into an individual model and the collective output gives the severity score. Three parallel architectures were also derived and tested. The proposed models were subjected to rigorous tests on two different datasets: RALO CXRs and Per COVID-19 CT scans. The experimental results showed that the proposed models exhibited high performance in accurately predicting scores for both datasets. In particular, the parallel transformers with multi-gate attention proved to be the best performing model. Furthermore, a comparative analysis using state-of-the-art methods showed that our proposed approach consistently achieved competitive or even better performance in terms of the Mean Absolute Error (MAE) and the Pearson Correlation Coefficient (PC). This emphasizes the effectiveness and superiority of our models in the context of diagnosing and assessing the severity of pneumonia.","PeriodicalId":13106,"journal":{"name":"IEEE Transactions on Big Data","volume":"11 5","pages":"2736-2748"},"PeriodicalIF":5.7,"publicationDate":"2025-03-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144934547","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE Transactions on Big Data
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1