首页 > 最新文献

AI Open最新文献

英文 中文
Methodologies and their comparison in complex compound aspect-based sentiment analysis: A survey 复杂复合面向情感分析方法及其比较:综述
Pub Date : 2025-01-01 Epub Date: 2025-03-07 DOI: 10.1016/j.aiopen.2025.02.002
Faiz Ghifari Haznitrama, Ho-Jin Choi, Chin-Wan Chung
Sentiment analysis as a part of natural language processing (NLP) has received much attention following the demand to understand people’s opinions. Aspect-based sentiment analysis (ABSA) is a fine-grained task from sentiment analysis that aims to classify the sentiment at the aspect level. Throughout the years, researchers have formulated ABSA into various tasks for different scenarios. Unlike early works, current ABSA tasks utilize many elements to provide more details to produce informative results. However, it is difficult to completely explore the works of ABSA because of the many different tasks, terms, and results. This paper surveyed recent studies on ABSA, specifically on its complex compound tasks. We investigated some key elements, problem formulations, and datasets currently utilized by most ABSA communities. We focused on reviewing the latest methodologies and worked to find the current state-of-the-art methodologies by performing a comparative analysis. From our study, we found that there has been a shift to generative methods in solving the ABSA problem, which signifies the evolving emphasis on holistic, end-to-end approaches. Finally, we identified some open challenges and future directions for ABSA research.
情感分析作为自然语言处理(NLP)的一部分,随着人们对理解人们观点的需求而受到越来越多的关注。基于方面的情感分析(ABSA)是一种来自情感分析的细粒度任务,旨在对方面级别的情感进行分类。多年来,研究人员已经将ABSA制定为不同场景的各种任务。与早期的工作不同,当前的ABSA任务利用许多元素来提供更多细节,以产生信息丰富的结果。然而,由于许多不同的任务、术语和结果,很难完全探索ABSA的作品。本文综述了近年来对ABSA的研究,特别是对其复杂复合任务的研究。我们调查了目前大多数ABSA社区使用的一些关键要素、问题表述和数据集。我们专注于审查最新的方法,并通过进行比较分析,努力找到当前最先进的方法。从我们的研究中,我们发现在解决ABSA问题时已经转向生成方法,这意味着对整体端到端方法的不断强调。最后,我们确定了ABSA研究的一些开放挑战和未来方向。
{"title":"Methodologies and their comparison in complex compound aspect-based sentiment analysis: A survey","authors":"Faiz Ghifari Haznitrama,&nbsp;Ho-Jin Choi,&nbsp;Chin-Wan Chung","doi":"10.1016/j.aiopen.2025.02.002","DOIUrl":"10.1016/j.aiopen.2025.02.002","url":null,"abstract":"<div><div>Sentiment analysis as a part of natural language processing (NLP) has received much attention following the demand to understand people’s opinions. Aspect-based sentiment analysis (ABSA) is a fine-grained task from sentiment analysis that aims to classify the sentiment at the aspect level. Throughout the years, researchers have formulated ABSA into various tasks for different scenarios. Unlike early works, current ABSA tasks utilize many elements to provide more details to produce informative results. However, it is difficult to completely explore the works of ABSA because of the many different tasks, terms, and results. This paper surveyed recent studies on ABSA, specifically on its complex compound tasks. We investigated some key elements, problem formulations, and datasets currently utilized by most ABSA communities. We focused on reviewing the latest methodologies and worked to find the current <em>state-of-the-art</em> methodologies by performing a comparative analysis. From our study, we found that there has been a shift to generative methods in solving the ABSA problem, which signifies the evolving emphasis on holistic, end-to-end approaches. Finally, we identified some open challenges and future directions for ABSA research.</div></div>","PeriodicalId":100068,"journal":{"name":"AI Open","volume":"6 ","pages":"Pages 53-69"},"PeriodicalIF":0.0,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143601111","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
DFM: Dialogue foundation model for universal large-scale dialogue-oriented task learning 面向大范围对话的任务学习对话基础模型
IF 14.8 Pub Date : 2025-01-01 Epub Date: 2025-07-01 DOI: 10.1016/j.aiopen.2025.04.001
Zhi Chen , Da Ma , Hanqi Li , Lu Chen , Jiabao Ji , Yuncong Liu , Bei Chen , Mengyue Wu , Su Zhu , Xin Dong , Fujiang Ge , Qingliang Miao , Jian-Guang Lou , Shuai Fan , Kai Yu
Building a universal conversational agent has been a long-standing goal of the dialogue research community. Most previous works only focus on a small set of dialogue tasks. In this work, we aim to build a unified dialogue foundation model (DFM) which can be used to solve massive diverse dialogue tasks. To achieve this goal, a large-scale well-annotated dialogue dataset with rich task diversity (DialogZoo) is collected. We introduce a framework to unify all dialogue tasks and propose novel auxiliary self-supervised tasks to achieve stable training of DFM on the highly diverse large scale DialogZoo corpus. Experiments show that, compared with models of the same size, DFM can achieve competitive performance on very rich cross-domain downstream dialogue tasks. Furthermore, when scaling to large language models, DFM remains effective. This demonstrates that DFM largely extends the ability of unified dialogue pre-trained model.
建立一个通用的会话代理一直是对话研究界的长期目标。大多数以前的作品只关注一小部分对话任务。在这项工作中,我们的目标是建立一个统一的对话基础模型(DFM),该模型可以用于解决大量不同的对话任务。为了实现这一目标,收集了一个具有丰富任务多样性、标注良好的大规模对话数据集(DialogZoo)。我们引入了一个框架来统一所有的对话任务,并提出了新的辅助自监督任务,以实现DFM在高度多样化的大型DialogZoo语料库上的稳定训练。实验表明,与相同规模的模型相比,DFM在非常丰富的跨域下游对话任务上可以取得具有竞争力的性能。此外,当扩展到大型语言模型时,DFM仍然有效。这表明DFM在很大程度上扩展了统一对话预训练模型的能力。
{"title":"DFM: Dialogue foundation model for universal large-scale dialogue-oriented task learning","authors":"Zhi Chen ,&nbsp;Da Ma ,&nbsp;Hanqi Li ,&nbsp;Lu Chen ,&nbsp;Jiabao Ji ,&nbsp;Yuncong Liu ,&nbsp;Bei Chen ,&nbsp;Mengyue Wu ,&nbsp;Su Zhu ,&nbsp;Xin Dong ,&nbsp;Fujiang Ge ,&nbsp;Qingliang Miao ,&nbsp;Jian-Guang Lou ,&nbsp;Shuai Fan ,&nbsp;Kai Yu","doi":"10.1016/j.aiopen.2025.04.001","DOIUrl":"10.1016/j.aiopen.2025.04.001","url":null,"abstract":"<div><div>Building a universal conversational agent has been a long-standing goal of the dialogue research community. Most previous works only focus on a small set of dialogue tasks. In this work, we aim to build a unified dialogue foundation model (DFM) which can be used to solve massive diverse dialogue tasks. To achieve this goal, a large-scale well-annotated dialogue dataset with rich task diversity (DialogZoo) is collected. We introduce a framework to unify all dialogue tasks and propose novel auxiliary self-supervised tasks to achieve stable training of DFM on the highly diverse large scale DialogZoo corpus. Experiments show that, compared with models of the same size, DFM can achieve competitive performance on very rich cross-domain downstream dialogue tasks. Furthermore, when scaling to large language models, DFM remains effective. This demonstrates that DFM largely extends the ability of unified dialogue pre-trained model.</div></div>","PeriodicalId":100068,"journal":{"name":"AI Open","volume":"6 ","pages":"Pages 108-117"},"PeriodicalIF":14.8,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144828236","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multi-scale texture loss for CT denoising with GANs gan对CT去噪的多尺度纹理损失
IF 14.8 Pub Date : 2025-01-01 Epub Date: 2025-09-23 DOI: 10.1016/j.aiopen.2025.09.001
Francesco Di Feola , Lorenzo Tronchin , Valerio Guarrasi , Paolo Soda
Generative Adversarial Networks (GANs) have proved as a powerful framework for denoising applications in medical imaging. However, GAN-based denoising algorithms still suffer from limitations in capturing complex relationships within the images. In this regard, the loss function plays a crucial role in guiding the image generation process, encompassing how much a synthetic image differs from a real image. To grasp highly complex and non-linear textural relationships in the training process, this work presents a novel approach to capture and embed multi-scale texture information into the loss function. Our method introduces a differentiable multi-scale texture representation of the images dynamically aggregated by a self-attention layer, thus exploiting end-to-end gradient-based optimization. We validate our approach by carrying out extensive experiments in the context of low-dose CT denoising, a challenging application that aims to enhance the quality of noisy CT scans. We utilize three publicly available datasets, including one simulated and two real datasets. The results are promising as compared to other well-established loss functions, being also consistent across three different GAN architectures. The code is available at: https://github.com/trainlab/MSTLF-TextureLoss.
生成对抗网络(GANs)已被证明是医学成像降噪应用的强大框架。然而,基于gan的去噪算法在捕获图像中的复杂关系方面仍然存在局限性。在这方面,损失函数在指导图像生成过程中起着至关重要的作用,它包含了合成图像与真实图像的差异。为了在训练过程中掌握高度复杂和非线性的纹理关系,本文提出了一种新的方法来捕获和嵌入多尺度纹理信息到损失函数中。我们的方法引入了由自关注层动态聚合的图像的可微多尺度纹理表示,从而利用了端到端基于梯度的优化。我们通过在低剂量CT去噪背景下进行广泛的实验来验证我们的方法,低剂量CT去噪是一项具有挑战性的应用,旨在提高噪声CT扫描的质量。我们使用三个公开可用的数据集,包括一个模拟数据集和两个真实数据集。与其他已建立的损失函数相比,结果是有希望的,并且在三种不同的GAN架构中也是一致的。代码可从https://github.com/trainlab/MSTLF-TextureLoss获得。
{"title":"Multi-scale texture loss for CT denoising with GANs","authors":"Francesco Di Feola ,&nbsp;Lorenzo Tronchin ,&nbsp;Valerio Guarrasi ,&nbsp;Paolo Soda","doi":"10.1016/j.aiopen.2025.09.001","DOIUrl":"10.1016/j.aiopen.2025.09.001","url":null,"abstract":"<div><div>Generative Adversarial Networks (GANs) have proved as a powerful framework for denoising applications in medical imaging. However, GAN-based denoising algorithms still suffer from limitations in capturing complex relationships within the images. In this regard, the loss function plays a crucial role in guiding the image generation process, encompassing how much a synthetic image differs from a real image. To grasp highly complex and non-linear textural relationships in the training process, this work presents a novel approach to capture and embed multi-scale texture information into the loss function. Our method introduces a differentiable multi-scale texture representation of the images dynamically aggregated by a self-attention layer, thus exploiting end-to-end gradient-based optimization. We validate our approach by carrying out extensive experiments in the context of low-dose CT denoising, a challenging application that aims to enhance the quality of noisy CT scans. We utilize three publicly available datasets, including one simulated and two real datasets. The results are promising as compared to other well-established loss functions, being also consistent across three different GAN architectures. The code is available at: <span><span>https://github.com/trainlab/MSTLF-TextureLoss</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":100068,"journal":{"name":"AI Open","volume":"6 ","pages":"Pages 142-154"},"PeriodicalIF":14.8,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145157276","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
ChatLLM network: More brains, more intelligence ChatLLM网络:更多的大脑,更多的智慧
Pub Date : 2025-01-01 Epub Date: 2025-02-12 DOI: 10.1016/j.aiopen.2025.01.001
Rui Hao , Linmei Hu , Weijian Qi , Qingliu Wu , Yirui Zhang , Liqiang Nie
Dialogue-based language models mark a huge milestone in the field of artificial intelligence, by their impressive ability to interact with users, as well as a series of challenging tasks prompted by customized instructions. However, the prevalent large-scale dialogue-based language models like ChatGPT still have room for improvement, such as unstable responses to questions and the inability to think cooperatively like humans. Considering the ability of dialogue-based language models in conversation and their inherent randomness in thinking, we propose ChatLLM network that allows multiple dialogue-based language models to interact, provide feedback, and think together. We design a network of ChatLLMs, consisting multiple layers of language models. Specifically, individual instances of language model may possess distinct perspectives towards the same problem, and by consolidating these diverse viewpoints via a separate language model, the ChatLLM network system can conduct decision-making more objectively and comprehensively. In addition, a language-based feedback mechanism comparable to backpropagation is devised to update the outputs of the language models within the network. This stratified system of interaction can be analogized to the relationship between leaders and employees in a social organization, where collective decision-making often yields superior judgments or resolutions. Experiments on datasets demonstrate that our network attains significant improvements in problem-solving, leading to observable progress amongst each member.
基于对话的语言模型在人工智能领域是一个巨大的里程碑,它们与用户互动的能力令人印象深刻,还能根据定制指令完成一系列具有挑战性的任务。然而,目前流行的大型对话式语言模型(如 ChatGPT)仍有改进的余地,如对问题的回答不稳定,无法像人类一样合作思考等。考虑到基于对话的语言模型在对话中的能力及其固有的思维随机性,我们提出了 ChatLLM 网络,它允许多个基于对话的语言模型进行互动、反馈和共同思考。我们设计了一个由多层语言模型组成的 ChatLLM 网络。具体来说,语言模型的各个实例可能对同一问题持有不同的观点,而通过单独的语言模型整合这些不同的观点,ChatLLM 网络系统可以更客观、更全面地进行决策。此外,还设计了一种类似于反向传播的基于语言的反馈机制,用于更新网络内语言模型的输出。这种分层互动系统可以类比为社会组织中领导与员工之间的关系,在这种关系中,集体决策往往会产生更优越的判断或解决方案。数据集上的实验表明,我们的网络在解决问题方面取得了显著的进步,每个成员都取得了可观的进步。
{"title":"ChatLLM network: More brains, more intelligence","authors":"Rui Hao ,&nbsp;Linmei Hu ,&nbsp;Weijian Qi ,&nbsp;Qingliu Wu ,&nbsp;Yirui Zhang ,&nbsp;Liqiang Nie","doi":"10.1016/j.aiopen.2025.01.001","DOIUrl":"10.1016/j.aiopen.2025.01.001","url":null,"abstract":"<div><div>Dialogue-based language models mark a huge milestone in the field of artificial intelligence, by their impressive ability to interact with users, as well as a series of challenging tasks prompted by customized instructions. However, the prevalent large-scale dialogue-based language models like ChatGPT still have room for improvement, such as unstable responses to questions and the inability to think cooperatively like humans. Considering the ability of dialogue-based language models in conversation and their inherent randomness in thinking, we propose ChatLLM network that allows multiple dialogue-based language models to interact, provide feedback, and think together. We design a network of ChatLLMs, consisting multiple layers of language models. Specifically, individual instances of language model may possess distinct perspectives towards the same problem, and by consolidating these diverse viewpoints via a separate language model, the ChatLLM network system can conduct decision-making more objectively and comprehensively. In addition, a language-based feedback mechanism comparable to backpropagation is devised to update the outputs of the language models within the network. This stratified system of interaction can be analogized to the relationship between leaders and employees in a social organization, where collective decision-making often yields superior judgments or resolutions. Experiments on datasets demonstrate that our network attains significant improvements in problem-solving, leading to observable progress amongst each member.</div></div>","PeriodicalId":100068,"journal":{"name":"AI Open","volume":"6 ","pages":"Pages 45-52"},"PeriodicalIF":0.0,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143420049","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Erratum regarding Declaration of Competing Interest statements in previously published articles 关于先前发表的文章中竞争利益声明的勘误表
IF 14.8 Pub Date : 2025-01-01 Epub Date: 2024-01-09 DOI: 10.1016/j.aiopen.2024.01.002
{"title":"Erratum regarding Declaration of Competing Interest statements in previously published articles","authors":"","doi":"10.1016/j.aiopen.2024.01.002","DOIUrl":"10.1016/j.aiopen.2024.01.002","url":null,"abstract":"","PeriodicalId":100068,"journal":{"name":"AI Open","volume":"6 ","pages":"Pages 331-332"},"PeriodicalIF":14.8,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145839253","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Solving the enigma: Enhancing faithfulness and comprehensibility in explanations of deep networks 解谜:增强深度网络解释的忠实性和可理解性
Pub Date : 2025-01-01 Epub Date: 2025-03-03 DOI: 10.1016/j.aiopen.2025.02.001
Michail Mamalakis , Antonios Mamalakis , Ingrid Agartz , Lynn Egeland Mørch-Johnsen , Graham K. Murray , John Suckling , Pietro Lio
The accelerated progress of artificial intelligence (AI) has popularized deep learning models across various domains, yet their inherent opacity poses challenges, particularly in critical fields like healthcare, medicine, and the geosciences. Explainable AI (XAI) has emerged to shed light on these ’black box’ models, aiding in deciphering their decision-making processes. However, different XAI methods often produce significantly different explanations, leading to high inter-method variability that increases uncertainty and undermines trust in deep networks’ predictions. In this study, we address this challenge by introducing a novel framework designed to enhance the explainability of deep networks through a dual focus on maximizing both accuracy and comprehensibility in the explanations. Our framework integrates outputs from multiple established XAI methods and leverages a non-linear neural network model, termed the ‘Explanation optimizer,’ to construct a unified, optimal explanation. The optimizer uses two primary metrics — faithfulness and complexity — to evaluate the quality of the explanations. Faithfulness measures the accuracy with which the explanation reflects the network’s decision-making, while complexity assesses the comprehensibility of the explanation. By balancing these metrics, the optimizer provides explanations that are both accurate and accessible, addressing a central limitation in current XAI methods. Through experiments on multi-class and binary classification tasks in both 2D object and 3D neuroscience imaging, we validate the efficacy of our approach. Our explanation optimizer achieved superior faithfulness scores, averaging 155% and 63% higher than the best-performing individual XAI methods in the 3D and 2D applications, respectively, while also reducing complexity to enhance comprehensibility. These results demonstrate that optimal explanations based on specific quality criteria are achievable, offering a solution to the issue of inter-method variability in the current XAI literature and supporting more trustworthy deep network predictions.
人工智能(AI)的加速发展使深度学习模型在各个领域得到普及,但其固有的不透明性带来了挑战,特别是在医疗保健、医学和地球科学等关键领域。可解释人工智能(XAI)的出现,揭示了这些“黑匣子”模型,帮助破译它们的决策过程。然而,不同的XAI方法通常会产生显著不同的解释,导致方法间的高可变性,增加了不确定性,破坏了对深度网络预测的信任。在本研究中,我们通过引入一个新的框架来解决这一挑战,该框架旨在通过双重关注最大化解释的准确性和可理解性来增强深度网络的可解释性。我们的框架集成了多个已建立的XAI方法的输出,并利用称为“解释优化器”的非线性神经网络模型来构建统一的最佳解释。优化器使用两个主要指标——忠实度和复杂性——来评估解释的质量。信度衡量的是解释反映网络决策的准确性,而复杂性评估的是解释的可理解性。通过平衡这些指标,优化器提供了既准确又可访问的解释,解决了当前XAI方法中的一个主要限制。通过对二维目标和三维神经科学成像的多类和二元分类任务进行实验,验证了该方法的有效性。我们的解释优化器获得了卓越的忠实度分数,在3D和2D应用中,平均比表现最好的单个XAI方法分别高出155%和63%,同时还降低了复杂性以提高可理解性。这些结果表明,基于特定质量标准的最佳解释是可以实现的,为当前XAI文献中的方法间变异性问题提供了解决方案,并支持更可信的深度网络预测。
{"title":"Solving the enigma: Enhancing faithfulness and comprehensibility in explanations of deep networks","authors":"Michail Mamalakis ,&nbsp;Antonios Mamalakis ,&nbsp;Ingrid Agartz ,&nbsp;Lynn Egeland Mørch-Johnsen ,&nbsp;Graham K. Murray ,&nbsp;John Suckling ,&nbsp;Pietro Lio","doi":"10.1016/j.aiopen.2025.02.001","DOIUrl":"10.1016/j.aiopen.2025.02.001","url":null,"abstract":"<div><div>The accelerated progress of artificial intelligence (AI) has popularized deep learning models across various domains, yet their inherent opacity poses challenges, particularly in critical fields like healthcare, medicine, and the geosciences. Explainable AI (XAI) has emerged to shed light on these ’black box’ models, aiding in deciphering their decision-making processes. However, different XAI methods often produce significantly different explanations, leading to high inter-method variability that increases uncertainty and undermines trust in deep networks’ predictions. In this study, we address this challenge by introducing a novel framework designed to enhance the explainability of deep networks through a dual focus on maximizing both accuracy and comprehensibility in the explanations. Our framework integrates outputs from multiple established XAI methods and leverages a non-linear neural network model, termed the ‘Explanation optimizer,’ to construct a unified, optimal explanation. The optimizer uses two primary metrics — faithfulness and complexity — to evaluate the quality of the explanations. Faithfulness measures the accuracy with which the explanation reflects the network’s decision-making, while complexity assesses the comprehensibility of the explanation. By balancing these metrics, the optimizer provides explanations that are both accurate and accessible, addressing a central limitation in current XAI methods. Through experiments on multi-class and binary classification tasks in both 2D object and 3D neuroscience imaging, we validate the efficacy of our approach. Our explanation optimizer achieved superior faithfulness scores, averaging 155% and 63% higher than the best-performing individual XAI methods in the 3D and 2D applications, respectively, while also reducing complexity to enhance comprehensibility. These results demonstrate that optimal explanations based on specific quality criteria are achievable, offering a solution to the issue of inter-method variability in the current XAI literature and supporting more trustworthy deep network predictions.</div></div>","PeriodicalId":100068,"journal":{"name":"AI Open","volume":"6 ","pages":"Pages 70-81"},"PeriodicalIF":0.0,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143628964","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Scalable graph attention-based instance selection via mini-batch sampling and hierarchical hashing 通过小批量采样和分层哈希进行可扩展的基于关注的图实例选择
IF 14.8 Pub Date : 2025-01-01 Epub Date: 2025-09-22 DOI: 10.1016/j.aiopen.2025.08.004
Zahiriddin Rustamov , Ayham Zaitouny , Nazar Zaki
Instance selection (IS) addresses the critical challenge of reducing dataset size while keeping informative characteristics, becoming increasingly important as datasets grow to millions of instances. Current IS methods often struggle with capturing complex relationships in high-dimensional spaces and scale with large datasets. This paper introduces a graph attention-based instance selection (GAIS) method that uses attention mechanisms to identify informative instances through their structural relationships in graph representations. We present two approaches for scalable graph construction: a distance-based mini-batch sampling technique that achieves dataset-size-independent complexity through strategic batch processing, and a hierarchical hashing approach that enables efficient similarity computation through random projections. The mini-batch approach keeps class distributions through stratified sampling, while the hierarchical hashing method captures relationships at multiple granularities through single-level, multi-level, and multi-view variants. Experiments across 39 datasets show that GAIS achieves reduction rates above 96% while maintaining or improving model performance relative to state-of-the-art IS methods. The findings show that the distance-based mini-batch approach offers an optimal efficiency for large-scale datasets, while multi-view variants excel on complex, high-dimensional data, demonstrating that attention-based importance scoring can effectively identify instances important for maintaining decision boundaries while avoiding computationally prohibitive pairwise comparisons. The code is publicly available at https://github.com/zahiriddin-rustamov/gais.
实例选择(IS)解决了在保持信息特征的同时减少数据集大小的关键挑战,随着数据集增长到数百万个实例,它变得越来越重要。当前的IS方法通常难以在高维空间中捕获复杂的关系,并且难以在大型数据集上进行扩展。本文介绍了一种基于图注意的实例选择(GAIS)方法,该方法利用注意机制通过图表示中的结构关系来识别信息实例。我们提出了两种可扩展图构建方法:一种基于距离的小批量采样技术,通过战略性批处理实现与数据集大小无关的复杂性,以及一种分层哈希方法,通过随机投影实现高效的相似性计算。小批处理方法通过分层抽样保持类分布,而分层散列方法通过单级、多级和多视图变体捕获多粒度的关系。39个数据集的实验表明,相对于最先进的IS方法,GAIS在保持或提高模型性能的同时,实现了96%以上的降噪率。研究结果表明,基于距离的小批量方法为大规模数据集提供了最佳效率,而多视图变体在复杂的高维数据上表现出色,这表明基于注意力的重要性评分可以有效地识别对维持决策边界重要的实例,同时避免了计算上的两两比较。该代码可在https://github.com/zahiriddin-rustamov/gais上公开获得。
{"title":"Scalable graph attention-based instance selection via mini-batch sampling and hierarchical hashing","authors":"Zahiriddin Rustamov ,&nbsp;Ayham Zaitouny ,&nbsp;Nazar Zaki","doi":"10.1016/j.aiopen.2025.08.004","DOIUrl":"10.1016/j.aiopen.2025.08.004","url":null,"abstract":"<div><div>Instance selection (IS) addresses the critical challenge of reducing dataset size while keeping informative characteristics, becoming increasingly important as datasets grow to millions of instances. Current IS methods often struggle with capturing complex relationships in high-dimensional spaces and scale with large datasets. This paper introduces a graph attention-based instance selection (GAIS) method that uses attention mechanisms to identify informative instances through their structural relationships in graph representations. We present two approaches for scalable graph construction: a distance-based mini-batch sampling technique that achieves dataset-size-independent complexity through strategic batch processing, and a hierarchical hashing approach that enables efficient similarity computation through random projections. The mini-batch approach keeps class distributions through stratified sampling, while the hierarchical hashing method captures relationships at multiple granularities through single-level, multi-level, and multi-view variants. Experiments across 39 datasets show that GAIS achieves reduction rates above 96% while maintaining or improving model performance relative to state-of-the-art IS methods. The findings show that the distance-based mini-batch approach offers an optimal efficiency for large-scale datasets, while multi-view variants excel on complex, high-dimensional data, demonstrating that attention-based importance scoring can effectively identify instances important for maintaining decision boundaries while avoiding computationally prohibitive pairwise comparisons. The code is publicly available at <span><span>https://github.com/zahiriddin-rustamov/gais</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":100068,"journal":{"name":"AI Open","volume":"6 ","pages":"Pages 167-182"},"PeriodicalIF":14.8,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145219483","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Client: Cross-variable linear integrated enhanced transformer for multivariate long-term time series forecasting 客户:用于多变量长期时间序列预测的交叉变量线性集成增强变压器
Pub Date : 2025-01-01 Epub Date: 2025-07-12 DOI: 10.1016/j.aiopen.2025.06.001
Jiaxin Gao , Wenbo Hu , Dongxiao Zhang , Yuntian Chen
Long-term time series forecasting (LTSF) is crucial in modern society, playing a pivotal role in facilitating long-term planning and developing early warning systems. While many Transformer-based models have recently been introduced for LTSF, a doubt has been raised regarding the effectiveness of attention modules in capturing cross-time dependencies. In this study, we design a mask-series experiment to validate this assumption and subsequently propose the ”Cross-variable Linear Integrated ENhanced Transformer for Multivariate Long-Term Time Series Forecasting” (Client), an advanced model that outperforms both traditional Transformer-based models and linear models. Client employs the linear module to learn trend information and the enhanced Transformer module to capture cross-variable dependencies. Meanwhile, the cross-variable Transformer module in Client simplifies the embedding and position encoding layers and replaces the decoder module with a projection layer. Extensive experiments with nine real-world datasets have confirmed the SOTA performance of Client with the least computation time and memory consumption compared with the previous Transformer-based models. Our code is available at https://github.com/daxin007/Client.
长期时间序列预测(LTSF)在现代社会中至关重要,在促进长期规划和建立预警系统方面发挥着关键作用。虽然最近为LTSF引入了许多基于transformer的模型,但是对于注意力模块在捕获跨时间依赖性方面的有效性提出了疑问。在本研究中,我们设计了一个掩模系列实验来验证这一假设,并随后提出了“用于多元长期时间序列预测的交叉变量线性集成增强型变压器”(Client),这是一个优于传统基于变压器的模型和线性模型的先进模型。客户端使用线性模块来学习趋势信息,使用增强的Transformer模块来捕获跨变量依赖关系。同时,Client中的跨变量Transformer模块简化了嵌入层和位置编码层,并将解码器模块替换为投影层。在9个真实数据集上的大量实验证明,与之前基于transformer的模型相比,Client的SOTA性能具有最小的计算时间和内存消耗。我们的代码可在https://github.com/daxin007/Client上获得。
{"title":"Client: Cross-variable linear integrated enhanced transformer for multivariate long-term time series forecasting","authors":"Jiaxin Gao ,&nbsp;Wenbo Hu ,&nbsp;Dongxiao Zhang ,&nbsp;Yuntian Chen","doi":"10.1016/j.aiopen.2025.06.001","DOIUrl":"10.1016/j.aiopen.2025.06.001","url":null,"abstract":"<div><div>Long-term time series forecasting (LTSF) is crucial in modern society, playing a pivotal role in facilitating long-term planning and developing early warning systems. While many Transformer-based models have recently been introduced for LTSF, a doubt has been raised regarding the effectiveness of attention modules in capturing cross-time dependencies. In this study, we design a mask-series experiment to validate this assumption and subsequently propose the ”Cross-variable Linear Integrated ENhanced Transformer for Multivariate Long-Term Time Series Forecasting” (<em>Client</em>), an advanced model that outperforms both traditional Transformer-based models and linear models. <em>Client</em> employs the linear module to learn trend information and the enhanced Transformer module to capture cross-variable dependencies. Meanwhile, the cross-variable Transformer module in <em>Client</em> simplifies the embedding and position encoding layers and replaces the decoder module with a projection layer. Extensive experiments with nine real-world datasets have confirmed the SOTA performance of <em>Client</em> with the least computation time and memory consumption compared with the previous Transformer-based models. Our code is available at <span><span>https://github.com/daxin007/Client</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":100068,"journal":{"name":"AI Open","volume":"6 ","pages":"Pages 93-107"},"PeriodicalIF":0.0,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144656936","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Robust emotion recognition using hybrid Bayesian LSTM based on Laban movement analysis 基于Laban动作分析的混合贝叶斯LSTM鲁棒情绪识别
IF 14.8 Pub Date : 2025-01-01 Epub Date: 2025-09-29 DOI: 10.1016/j.aiopen.2025.09.002
Shuang Wu , Daniela M. Romano
Emotion recognition has become increasingly significant in artificial intelligence; however, the impact of body movements on emotion interpretation remains under-explored. This paper presents a novel Hybrid Bayesian Pre-trained Long Short-Term Memory (HBP-LSTM) framework that combines low-level pose data with high-level kinematic features, utilising Bayesian inference to enhance the accuracy and robustness of emotion recognition. The proposed model is trained on high-quality laboratory data to capture the fundamental patterns of emotional expression through body movements. We introduce noise and employ adversarial attack methods such as the Fast Gradient Sign Method (FGSM) to evaluate the model’s robustness during testing. This approach assesses the HBP-LSTM’s ability to maintain performance under data degradation and adversarial conditions, common challenges in real-world scenarios. We validated the HBP-LSTM on two public datasets, EGBM and KDAEE, demonstrating that the model exhibits high robustness against noise and adversarial perturbations, outperforming traditional models. The HBP-LSTM accurately identifies seven basic emotions (happiness, sadness, surprise, fear, anger, disgust, and neutrality) with accuracies of 98% and 88% on the EGBM and KDAEE datasets, respectively. HBP-LSTM is a noise-resistant model with a reliable emotion recognition framework, which lays the foundation for future applications of emotion recognition technology in more challenging real-world environments.
情感识别在人工智能中变得越来越重要;然而,身体动作对情绪解释的影响仍未得到充分探讨。本文提出了一种新的混合贝叶斯预训练长短期记忆(HBP-LSTM)框架,该框架将低级姿态数据与高级运动特征相结合,利用贝叶斯推理来提高情绪识别的准确性和鲁棒性。所提出的模型是在高质量的实验室数据上训练的,以捕捉通过身体动作表达情感的基本模式。在测试过程中,我们引入噪声并采用对抗攻击方法(如快速梯度符号法(FGSM))来评估模型的鲁棒性。该方法评估了HBP-LSTM在数据退化和对抗条件下保持性能的能力,这是现实场景中的常见挑战。我们在两个公共数据集(EGBM和KDAEE)上验证了HBP-LSTM,结果表明该模型对噪声和对抗性扰动具有很高的鲁棒性,优于传统模型。HBP-LSTM准确识别七种基本情绪(快乐、悲伤、惊讶、恐惧、愤怒、厌恶和中立),在EGBM和KDAEE数据集上的准确率分别为98%和88%。HBP-LSTM是一种具有可靠情绪识别框架的抗噪声模型,为未来情绪识别技术在更具挑战性的现实环境中的应用奠定了基础。
{"title":"Robust emotion recognition using hybrid Bayesian LSTM based on Laban movement analysis","authors":"Shuang Wu ,&nbsp;Daniela M. Romano","doi":"10.1016/j.aiopen.2025.09.002","DOIUrl":"10.1016/j.aiopen.2025.09.002","url":null,"abstract":"<div><div>Emotion recognition has become increasingly significant in artificial intelligence; however, the impact of body movements on emotion interpretation remains under-explored. This paper presents a novel Hybrid Bayesian Pre-trained Long Short-Term Memory (HBP-LSTM) framework that combines low-level pose data with high-level kinematic features, utilising Bayesian inference to enhance the accuracy and robustness of emotion recognition. The proposed model is trained on high-quality laboratory data to capture the fundamental patterns of emotional expression through body movements. We introduce noise and employ adversarial attack methods such as the Fast Gradient Sign Method (FGSM) to evaluate the model’s robustness during testing. This approach assesses the HBP-LSTM’s ability to maintain performance under data degradation and adversarial conditions, common challenges in real-world scenarios. We validated the HBP-LSTM on two public datasets, EGBM and KDAEE, demonstrating that the model exhibits high robustness against noise and adversarial perturbations, outperforming traditional models. The HBP-LSTM accurately identifies seven basic emotions (happiness, sadness, surprise, fear, anger, disgust, and neutrality) with accuracies of 98% and 88% on the EGBM and KDAEE datasets, respectively. HBP-LSTM is a noise-resistant model with a reliable emotion recognition framework, which lays the foundation for future applications of emotion recognition technology in more challenging real-world environments.</div></div>","PeriodicalId":100068,"journal":{"name":"AI Open","volume":"6 ","pages":"Pages 183-203"},"PeriodicalIF":14.8,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145218905","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Erratum regarding Declaration of Competing Interest statements in previously published articles 关于以前发表的文章中竞争利益声明的勘误
IF 14.8 Pub Date : 2025-01-01 Epub Date: 2024-01-06 DOI: 10.1016/j.aiopen.2024.01.001
{"title":"Erratum regarding Declaration of Competing Interest statements in previously published articles","authors":"","doi":"10.1016/j.aiopen.2024.01.001","DOIUrl":"10.1016/j.aiopen.2024.01.001","url":null,"abstract":"","PeriodicalId":100068,"journal":{"name":"AI Open","volume":"6 ","pages":"Pages 329-330"},"PeriodicalIF":14.8,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139394934","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
AI Open
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1