首页 > 最新文献

Proceedings of the ... AAAI Conference on Artificial Intelligence. AAAI Conference on Artificial Intelligence最新文献

英文 中文
Learning Noise-Induced Reward Functions for Surpassing Demonstrations in Imitation Learning 模仿学习中超越示范的学习噪声诱导奖励函数
Liangyu Huo, Zulin Wang, Mai Xu
Imitation learning (IL) has recently shown impressive performance in training a reinforcement learning agent with human demonstrations, eliminating the difficulty of designing elaborate reward functions in complex environments. However, most IL methods work under the assumption of the optimality of the demonstrations and thus cannot learn policies to surpass the demonstrators. Some methods have been investigated to obtain better-than-demonstration (BD) performance with inner human feedback or preference labels. In this paper, we propose a method to learn rewards from suboptimal demonstrations via a weighted preference learning technique (LERP). Specifically, we first formulate the suboptimality of demonstrations as the inaccurate estimation of rewards. The inaccuracy is modeled with a reward noise random variable following the Gumbel distribution. Moreover, we derive an upper bound of the expected return with different noise coefficients and propose a theorem to surpass the demonstrations. Unlike existing literature, our analysis does not depend on the linear reward constraint. Consequently, we develop a BD model with a weighted preference learning technique. Experimental results on continuous control and high-dimensional discrete control tasks show the superiority of our LERP method over other state-of-the-art BD methods.
模仿学习(IL)最近在训练具有人类演示的强化学习代理方面显示出令人印象深刻的表现,消除了在复杂环境中设计复杂奖励函数的困难。然而,大多数IL方法都是在假设演示的最优性下工作的,因此无法学习策略以超越演示。已经研究了一些方法,以获得比演示(BD)性能更好的内在人的反馈或偏好标签。在本文中,我们提出了一种通过加权偏好学习技术(LERP)从次优演示中学习奖励的方法。具体来说,我们首先将演示的次最优性表述为对奖励的不准确估计。不准确性用奖励噪声随机变量按照甘贝尔分布建模。此外,我们还推导了不同噪声系数下期望收益的上界,并提出了一个定理来超越这些证明。与现有文献不同,我们的分析不依赖于线性奖励约束。因此,我们开发了一个带有加权偏好学习技术的BD模型。在连续控制和高维离散控制任务上的实验结果表明,LERP方法优于其他最先进的BD方法。
{"title":"Learning Noise-Induced Reward Functions for Surpassing Demonstrations in Imitation Learning","authors":"Liangyu Huo, Zulin Wang, Mai Xu","doi":"10.1609/aaai.v37i7.25962","DOIUrl":"https://doi.org/10.1609/aaai.v37i7.25962","url":null,"abstract":"Imitation learning (IL) has recently shown impressive performance in training a reinforcement learning agent with human demonstrations, eliminating the difficulty of designing elaborate reward functions in complex environments. However, most IL methods work under the assumption of the optimality of the demonstrations and thus cannot learn policies to surpass the demonstrators. Some methods have been investigated to obtain better-than-demonstration (BD) performance with inner human feedback or preference labels. In this paper, we propose a method to learn rewards from suboptimal demonstrations via a weighted preference learning technique (LERP). Specifically, we first formulate the suboptimality of demonstrations as the inaccurate estimation of rewards. The inaccuracy is modeled with a reward noise random variable following the Gumbel distribution. Moreover, we derive an upper bound of the expected return with different noise coefficients and propose a theorem to surpass the demonstrations. Unlike existing literature, our analysis does not depend on the linear reward constraint. Consequently, we develop a BD model with a weighted preference learning technique. Experimental results on continuous control and high-dimensional discrete control tasks show the superiority of our LERP method over other state-of-the-art BD methods.","PeriodicalId":74506,"journal":{"name":"Proceedings of the ... AAAI Conference on Artificial Intelligence. AAAI Conference on Artificial Intelligence","volume":"50 1","pages":"7953-7961"},"PeriodicalIF":0.0,"publicationDate":"2023-06-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"87259615","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Unsupervised Legal Evidence Retrieval via Contrastive Learning with Approximate Aggregated Positive 基于近似聚合正的对比学习的无监督法律证据检索
Feng Yao, Jingyuan Zhang, Yating Zhang, Xiaozhong Liu, Changlong Sun, Yun Liu, Weixing Shen
Verifying the facts alleged by the prosecutors before the trial requires the judges to retrieve evidence within the massive materials accompanied.Existing Legal AI applications often assume the facts are already determined and fail to notice the difficulty of reconstructing them. To build a practical Legal AI application and free the judges from the manually searching work, we introduce the task of Legal Evidence Retrieval, which aims at automatically retrieving the precise fact-related verbal evidence within a single case. We formulate the task in a dense retrieval paradigm, and jointly learn the constrastive representations and alignments between facts and evidence. To get rid of the tedious annotations, we construct an approximated positive vector for a given fact by aggregating a set of evidence from the same case. An entropy-based denoise technique is further applied to mitigate the impact of false positive samples. We train our models on tens of thousands of unlabeled cases and evaluate them on a labeled dataset containing 919 cases and 4,336 queries. Experimental results indicate that our approach is effective and outperforms other state-of-the-art representation and retrieval models. The dataset and code are available at https://github.com/yaof20/LER.
为了在审判前验证检方提出的事实,法官需要从大量的材料中提取证据。现有的法律人工智能应用程序通常假设事实已经确定,并且没有注意到重建它们的困难。为了构建一个实用的法律人工智能应用程序,将法官从手动搜索工作中解放出来,我们引入了法律证据检索任务,旨在自动检索单个案件中与事实相关的精确口头证据。我们在密集检索范式中制定任务,并共同学习事实和证据之间的约束表示和对齐。为了摆脱繁琐的注释,我们通过聚合来自同一案例的一组证据来为给定事实构造一个近似的正向量。进一步应用基于熵的去噪技术来减轻假阳性样本的影响。我们在成千上万个未标记的案例上训练我们的模型,并在包含919个案例和4336个查询的标记数据集上对它们进行评估。实验结果表明,我们的方法是有效的,并且优于其他最先进的表示和检索模型。数据集和代码可在https://github.com/yaof20/LER上获得。
{"title":"Unsupervised Legal Evidence Retrieval via Contrastive Learning with Approximate Aggregated Positive","authors":"Feng Yao, Jingyuan Zhang, Yating Zhang, Xiaozhong Liu, Changlong Sun, Yun Liu, Weixing Shen","doi":"10.1609/aaai.v37i4.25603","DOIUrl":"https://doi.org/10.1609/aaai.v37i4.25603","url":null,"abstract":"Verifying the facts alleged by the prosecutors before the trial requires the judges to retrieve evidence within the massive materials accompanied.\u0000Existing Legal AI applications often assume the facts are already determined and fail to notice the difficulty of reconstructing them. To build a practical Legal AI application and free the judges from the manually searching work, we introduce the task of Legal Evidence Retrieval, which aims at automatically retrieving the precise fact-related verbal evidence within a single case. We formulate the task in a dense retrieval paradigm, and jointly learn the constrastive representations and alignments between facts and evidence. To get rid of the tedious annotations, we construct an approximated positive vector for a given fact by aggregating a set of evidence from the same case. An entropy-based denoise technique is further applied to mitigate the impact of false positive samples. We train our models on tens of thousands of unlabeled cases and evaluate them on a labeled dataset containing 919 cases and 4,336 queries. Experimental results indicate that our approach is effective and outperforms other state-of-the-art representation and retrieval models. The dataset and code are available at https://github.com/yaof20/LER.","PeriodicalId":74506,"journal":{"name":"Proceedings of the ... AAAI Conference on Artificial Intelligence. AAAI Conference on Artificial Intelligence","volume":"21 1","pages":"4783-4791"},"PeriodicalIF":0.0,"publicationDate":"2023-06-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"87256005","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Tensorized Incomplete Multi-View Clustering with Intrinsic Graph Completion 具有内在图补全的张紧化不完全多视图聚类
Shuping Zhao, Jie Wen, Lunke Fei, Bob Zhang
Most of the existing incomplete multi-view clustering (IMVC) methods focus on attaining a consensus representation from different views but ignore the important information hidden in the missing views and the latent intrinsic structures in each view. To tackle these issues, in this paper, a unified and novel framework, named tensorized incomplete multi-view clustering with intrinsic graph completion (TIMVC_IGC) is proposed. Firstly, owing to the effectiveness of the low-rank representation in revealing the inherent structure of the data, we exploit it to infer the missing instances and construct the complete graph for each view. Afterwards, inspired by the structural consistency, a between-view consistency constraint is imposed to guarantee the similarity of the graphs from different views. More importantly, the TIMVC_IGC simultaneously learns the low-rank structures of the different views and explores the correlations of the different graphs in a latent manifold sub-space using a low-rank tensor constraint, such that the intrinsic graphs of the different views can be obtained. Finally, a consensus representation for each sample is gained with a co-regularization term for final clustering. Experimental results on several real-world databases illustrates that the proposed method can outperform the other state-of-the-art related methods for incomplete multi-view clustering.
现有的不完全多视图聚类(IMVC)方法大多侧重于从不同的视图中获得一致的表示,而忽略了隐藏在缺失视图中的重要信息和每个视图中潜在的内在结构。为了解决这些问题,本文提出了一个统一的、新颖的框架,即具有内在图补全的张张化不完全多视图聚类(TIMVC_IGC)。首先,由于低秩表示在揭示数据固有结构方面的有效性,我们利用它来推断缺失实例并为每个视图构建完整图。然后,受结构一致性的启发,施加视图间一致性约束以保证不同视图图的相似性。更重要的是,TIMVC_IGC同时学习不同视图的低秩结构,并利用低秩张量约束在潜在流形子空间中探索不同图的相关性,从而获得不同视图的内在图。最后,用协正则化项获得每个样本的一致表示,用于最终聚类。在多个真实数据库上的实验结果表明,该方法在不完全多视图聚类方面优于其他相关方法。
{"title":"Tensorized Incomplete Multi-View Clustering with Intrinsic Graph Completion","authors":"Shuping Zhao, Jie Wen, Lunke Fei, Bob Zhang","doi":"10.1609/aaai.v37i9.26340","DOIUrl":"https://doi.org/10.1609/aaai.v37i9.26340","url":null,"abstract":"Most of the existing incomplete multi-view clustering (IMVC) methods focus on attaining a consensus representation from different views but ignore the important information hidden in the missing views and the latent intrinsic structures in each view. To tackle these issues, in this paper, a unified and novel framework, named tensorized incomplete multi-view clustering with intrinsic graph completion (TIMVC_IGC) is proposed. Firstly, owing to the effectiveness of the low-rank representation in revealing the inherent structure of the data, we exploit it to infer the missing instances and construct the complete graph for each view. Afterwards, inspired by the structural consistency, a between-view consistency constraint is imposed to guarantee the similarity of the graphs from different views. More importantly, the TIMVC_IGC simultaneously learns the low-rank structures of the different views and explores the correlations of the different graphs in a latent manifold sub-space using a low-rank tensor constraint, such that the intrinsic graphs of the different views can be obtained. Finally, a consensus representation for each sample is gained with a co-regularization term for final clustering. Experimental results on several real-world databases illustrates that the proposed method can outperform the other state-of-the-art related methods for incomplete multi-view clustering.","PeriodicalId":74506,"journal":{"name":"Proceedings of the ... AAAI Conference on Artificial Intelligence. AAAI Conference on Artificial Intelligence","volume":"24 1","pages":"11327-11335"},"PeriodicalIF":0.0,"publicationDate":"2023-06-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"87345086","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Bayesian Cross-Modal Alignment Learning for Few-Shot Out-of-Distribution Generalization 基于贝叶斯跨模态学习的少镜头分布外泛化
Lin Zhu, Xinbing Wang, Cheng Zhou, Nanyang Ye
Recent advances in large pre-trained models showed promising results in few-shot learning. However, their generalization ability on two-dimensional Out-of-Distribution (OoD) data, i.e., correlation shift and diversity shift, has not been thoroughly investigated. Researches have shown that even with a significant amount of training data, few methods can achieve better performance than the standard empirical risk minimization method (ERM) in OoD generalization. This few-shot OoD generalization dilemma emerges as a challenging direction in deep neural network generalization research, where the performance suffers from overfitting on few-shot examples and OoD generalization errors. In this paper, leveraging a broader supervision source, we explore a novel Bayesian cross-modal image-text alignment learning method (Bayes-CAL) to address this issue. Specifically, the model is designed as only text representations are fine-tuned via a Bayesian modelling approach with gradient orthogonalization loss and invariant risk minimization (IRM) loss. The Bayesian approach is essentially introduced to avoid overfitting the base classes observed during training and improve generalization to broader unseen classes. The dedicated loss is introduced to achieve better image-text alignment by disentangling the causal and non-casual parts of image features. Numerical experiments demonstrate that Bayes-CAL achieved state-of-the-art OoD generalization performances on two-dimensional distribution shifts. Moreover, compared with CLIP-like models, Bayes-CAL yields more stable generalization performances on unseen classes. Our code is available at https://github.com/LinLLLL/BayesCAL.
大型预训练模型的最新进展显示,在少量学习中取得了令人鼓舞的结果。然而,它们对二维离分布(out - distribution, OoD)数据的泛化能力,即相关性偏移和多样性偏移,尚未得到深入的研究。研究表明,即使有大量的训练数据,在OoD泛化中也很少有方法能比标准的经验风险最小化方法(ERM)取得更好的性能。在深度神经网络泛化研究中,小样本OoD泛化困境是一个具有挑战性的研究方向,其性能受到小样本过拟合和OoD泛化误差的影响。在本文中,利用更广泛的监督源,我们探索了一种新的贝叶斯跨模态图像-文本对齐学习方法(Bayes-CAL)来解决这个问题。具体来说,该模型被设计为只有文本表示通过具有梯度正交化损失和不变风险最小化(IRM)损失的贝叶斯建模方法进行微调。引入贝叶斯方法本质上是为了避免在训练期间观察到的基类过拟合,并提高对更广泛的未见类的泛化。通过分离图像特征的因果部分和非因果部分,引入专用损失来实现更好的图像-文本对齐。数值实验表明,Bayes-CAL在二维分布位移上取得了较好的泛化性能。此外,与类clip模型相比,Bayes-CAL在未见过的类上具有更稳定的泛化性能。我们的代码可在https://github.com/LinLLLL/BayesCAL上获得。
{"title":"Bayesian Cross-Modal Alignment Learning for Few-Shot Out-of-Distribution Generalization","authors":"Lin Zhu, Xinbing Wang, Cheng Zhou, Nanyang Ye","doi":"10.1609/aaai.v37i9.26355","DOIUrl":"https://doi.org/10.1609/aaai.v37i9.26355","url":null,"abstract":"Recent advances in large pre-trained models showed promising results in few-shot learning. However, their generalization ability on two-dimensional Out-of-Distribution (OoD) data, i.e., correlation shift and diversity shift, has not been thoroughly investigated. Researches have shown that even with a significant amount of training data, few methods can achieve better performance than the standard empirical risk minimization method (ERM) in OoD generalization. This few-shot OoD generalization dilemma emerges as a challenging direction in deep neural network generalization research, where the performance suffers from overfitting on few-shot examples and OoD generalization errors. In this paper, leveraging a broader supervision source, we explore a novel Bayesian cross-modal image-text alignment learning method (Bayes-CAL) to address this issue. Specifically, the model is designed as only text representations are fine-tuned via a Bayesian modelling approach with gradient orthogonalization loss and invariant risk minimization (IRM) loss. The Bayesian approach is essentially introduced to avoid overfitting the base classes observed during training and improve generalization to broader unseen classes. The dedicated loss is introduced to achieve better image-text alignment by disentangling the causal and non-casual parts of image features. Numerical experiments demonstrate that Bayes-CAL achieved state-of-the-art OoD generalization performances on two-dimensional distribution shifts. Moreover, compared with CLIP-like models, Bayes-CAL yields more stable generalization performances on unseen classes. Our code is available at https://github.com/LinLLLL/BayesCAL.","PeriodicalId":74506,"journal":{"name":"Proceedings of the ... AAAI Conference on Artificial Intelligence. AAAI Conference on Artificial Intelligence","volume":"26 1","pages":"11461-11469"},"PeriodicalIF":0.0,"publicationDate":"2023-06-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"80659733","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
WaveForM: Graph Enhanced Wavelet Learning for Long Sequence Forecasting of Multivariate Time Series 多变量时间序列长序列预测的波形图增强小波学习
Fu-qiang Yang, Xin Li, Min Wang, Hongyu Zang, W. Pang, Mingzhong Wang
Multivariate time series (MTS) analysis and forecasting are crucial in many real-world applications, such as smart traffic management and weather forecasting. However, most existing work either focuses on short sequence forecasting or makes predictions predominantly with time domain features, which is not effective at removing noises with irregular frequencies in MTS. Therefore, we propose WaveForM, an end-to-end graph enhanced Wavelet learning framework for long sequence FORecasting of MTS. WaveForM first utilizes Discrete Wavelet Transform (DWT) to represent MTS in the wavelet domain, which captures both frequency and time domain features with a sound theoretical basis. To enable the effective learning in the wavelet domain, we further propose a graph constructor, which learns a global graph to represent the relationships between MTS variables, and graph-enhanced prediction modules, which utilize dilated convolution and graph convolution to capture the correlations between time series and predict the wavelet coefficients at different levels. Extensive experiments on five real-world forecasting datasets show that our model can achieve considerable performance improvement over different prediction lengths against the most competitive baseline of each dataset.
多元时间序列(MTS)分析和预测在许多实际应用中至关重要,例如智能交通管理和天气预报。然而,现有的工作大多集中在短序列预测或主要利用时域特征进行预测,无法有效去除MTS中不规则频率的噪声。因此,我们提出了用于MTS长序列预测的端到端图增强小波学习框架波形。波形首先利用离散小波变换(DWT)在小波域表示MTS;该方法同时捕捉了频域和时域特征,具有良好的理论基础。为了实现小波域的有效学习,我们进一步提出了一个图构造器,它学习一个全局图来表示MTS变量之间的关系,以及图增强预测模块,它利用扩展卷积和图卷积来捕获时间序列之间的相关性并预测不同层次的小波系数。在五个真实世界预测数据集上的大量实验表明,我们的模型可以在每个数据集的最具竞争力基线的不同预测长度上取得相当大的性能改进。
{"title":"WaveForM: Graph Enhanced Wavelet Learning for Long Sequence Forecasting of Multivariate Time Series","authors":"Fu-qiang Yang, Xin Li, Min Wang, Hongyu Zang, W. Pang, Mingzhong Wang","doi":"10.1609/aaai.v37i9.26276","DOIUrl":"https://doi.org/10.1609/aaai.v37i9.26276","url":null,"abstract":"Multivariate time series (MTS) analysis and forecasting are crucial in many real-world applications, such as smart traffic management and weather forecasting. However, most existing work either focuses on short sequence forecasting or makes predictions predominantly with time domain features, which is not effective at removing noises with irregular frequencies in MTS. Therefore, we propose WaveForM, an end-to-end graph enhanced Wavelet learning framework for long sequence FORecasting of MTS. WaveForM first utilizes Discrete Wavelet Transform (DWT) to represent MTS in the wavelet domain, which captures both frequency and time domain features with a sound theoretical basis. To enable the effective learning in the wavelet domain, we further propose a graph constructor, which learns a global graph to represent the relationships between MTS variables, and graph-enhanced prediction modules, which utilize dilated convolution and graph convolution to capture the correlations between time series and predict the wavelet coefficients at different levels. Extensive experiments on five real-world forecasting datasets show that our model can achieve considerable performance improvement over different prediction lengths against the most competitive baseline of each dataset.","PeriodicalId":74506,"journal":{"name":"Proceedings of the ... AAAI Conference on Artificial Intelligence. AAAI Conference on Artificial Intelligence","volume":"9 1","pages":"10754-10761"},"PeriodicalIF":0.0,"publicationDate":"2023-06-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"85171922","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Augmenting Affective Dependency Graph via Iterative Incongruity Graph Learning for Sarcasm Detection 基于迭代不一致图学习的情感依赖图增强反讽检测
Xiaobao Wang, Yiqi Dong, Di Jin, Yawen Li, Longbiao Wang, J. Dang
Recently, progress has been made towards improving automatic sarcasm detection in computer science. Among existing models, manually constructing static graphs for texts and then using graph neural networks (GNNs) is one of the most effective approaches for drawing long-range incongruity patterns. However, the manually constructed graph structure might be prone to errors (e.g., noisy or incomplete) and not optimal for the sarcasm detection task. Errors produced during the graph construction step cannot be remedied and may accrue to the following stages, resulting in poor performance. To surmount the above limitations, we explore a novel Iterative Augmenting Affective Graph and Dependency Graph (IAAD) framework to jointly and iteratively learn the incongruity graph structure. IAAD can alternatively update the incongruity graph structure and node representation until the learning graph structure is optimal for the metrics of sarcasm detection. More concretely, we begin with deriving an affective and a dependency graph for each instance, then an iterative incongruity graph learning module is employed to augment affective and dependency graphs for obtaining the optimal inconsistent semantic graph with the goal of optimizing the graph for the sarcasm detection task. Extensive experiments on three datasets demonstrate that the proposed model outperforms state-of-the-art baselines for sarcasm detection with significant margins.
近年来,计算机科学在改进自动讽刺检测方面取得了一些进展。在现有模型中,手工构建文本静态图,然后使用图神经网络(gnn)来绘制远程不一致模式是最有效的方法之一。然而,手工构建的图结构可能容易出错(例如,有噪声或不完整),并且不适合讽刺检测任务。在图形构建步骤中产生的错误无法纠正,并可能累积到以下阶段,导致性能不佳。为了克服上述局限性,我们探索了一种新的迭代增强情感图和依赖图(IAAD)框架,用于联合迭代学习不一致图结构。IAAD可以交替地更新不一致图结构和节点表示,直到学习图结构对于讽刺检测的度量是最优的。更具体地说,我们首先推导每个实例的情感图和依赖图,然后使用迭代不一致图学习模块对情感图和依赖图进行扩充,以获得最优的不一致语义图,以优化图用于讽刺检测任务。在三个数据集上进行的大量实验表明,所提出的模型在讽刺检测方面的性能优于最先进的基线。
{"title":"Augmenting Affective Dependency Graph via Iterative Incongruity Graph Learning for Sarcasm Detection","authors":"Xiaobao Wang, Yiqi Dong, Di Jin, Yawen Li, Longbiao Wang, J. Dang","doi":"10.1609/aaai.v37i4.25594","DOIUrl":"https://doi.org/10.1609/aaai.v37i4.25594","url":null,"abstract":"Recently, progress has been made towards improving automatic sarcasm detection in computer science. Among existing models, manually constructing static graphs for texts and then using graph neural networks (GNNs) is one of the most effective approaches for drawing long-range incongruity patterns. However, the manually constructed graph structure might be prone to errors (e.g., noisy or incomplete) and not optimal for the sarcasm detection task. Errors produced during the graph construction step cannot be remedied and may accrue to the following stages, resulting in poor performance. To surmount the above limitations, we explore a novel Iterative Augmenting Affective Graph and Dependency Graph (IAAD) framework to jointly and iteratively learn the incongruity graph structure. IAAD can alternatively update the incongruity graph structure and node representation until the learning graph structure is optimal for the metrics of sarcasm detection. More concretely, we begin with deriving an affective and a dependency graph for each instance, then an iterative incongruity graph learning module is employed to augment affective and dependency graphs for obtaining the optimal inconsistent semantic graph with the goal of optimizing the graph for the sarcasm detection task. Extensive experiments on three datasets demonstrate that the proposed model outperforms state-of-the-art baselines for sarcasm detection with significant margins.","PeriodicalId":74506,"journal":{"name":"Proceedings of the ... AAAI Conference on Artificial Intelligence. AAAI Conference on Artificial Intelligence","volume":"31 1","pages":"4702-4710"},"PeriodicalIF":0.0,"publicationDate":"2023-06-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"86997075","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
NL2LTL - a Python Package for Converting Natural Language (NL) Instructions to Linear Temporal Logic (LTL) Formulas NL2LTL -一个Python包,用于将自然语言(NL)指令转换为线性时间逻辑(LTL)公式
Francesco Fuggitti, T. Chakraborti
This is a demonstration of our newly released Python package NL2LTL which leverages the latest in natural language understanding (NLU) and large language models (LLMs) to translate natural language instructions to linear temporal logic (LTL) formulas. This allows direct translation to formal languages that a reasoning system can use, while at the same time, allowing the end-user to provide inputs in natural language without having to understand any details of an underlying formal language. The package comes with support for a set of default LTL patterns, corresponding to popular DECLARE templates, but is also fully extensible to new formulas and user inputs. The package is open-source and is free to use for the AI community under the MIT license. Open Source: https://github.com/IBM/nl2ltl. Video Link: https://bit.ly/3dHW5b1
这是我们最新发布的Python包NL2LTL的演示,它利用最新的自然语言理解(NLU)和大型语言模型(llm)将自然语言指令转换为线性时间逻辑(LTL)公式。这允许直接翻译为推理系统可以使用的形式语言,同时允许最终用户以自然语言提供输入,而无需了解底层形式语言的任何细节。该包支持一组默认的LTL模式,与流行的DECLARE模板相对应,但也可以完全扩展到新的公式和用户输入。该软件包是开源的,可以在MIT许可下免费用于AI社区。开源:https://github.com/IBM/nl2ltl。视频链接:https://bit.ly/3dHW5b1
{"title":"NL2LTL - a Python Package for Converting Natural Language (NL) Instructions to Linear Temporal Logic (LTL) Formulas","authors":"Francesco Fuggitti, T. Chakraborti","doi":"10.1609/aaai.v37i13.27068","DOIUrl":"https://doi.org/10.1609/aaai.v37i13.27068","url":null,"abstract":"This is a demonstration of our newly released Python package NL2LTL which leverages the latest in natural language understanding (NLU) and large language models (LLMs) to translate natural language instructions to linear temporal logic (LTL) formulas. This allows direct translation to formal languages that a reasoning system can use, while at the same time, allowing the end-user to provide inputs in natural language without having to understand any details of an underlying \u0000formal language. The package comes with support for a set of default LTL patterns, corresponding to popular DECLARE templates, but is also fully extensible to new formulas and user inputs. The package is open-source and is free to use for the AI community under the MIT license. Open Source: https://github.com/IBM/nl2ltl. Video Link: https://bit.ly/3dHW5b1","PeriodicalId":74506,"journal":{"name":"Proceedings of the ... AAAI Conference on Artificial Intelligence. AAAI Conference on Artificial Intelligence","volume":"21 1","pages":"16428-16430"},"PeriodicalIF":0.0,"publicationDate":"2023-06-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"90754750","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Script, Language, and Labels: Overcoming Three Discrepancies for Low-Resource Language Specialization 文字、语言和标签:克服低资源语言专门化的三个差异
Jaeseong Lee, Dohyeon Lee, Seung-won Hwang
Although multilingual pretrained models (mPLMs) enabled support of various natural language processing in diverse languages, its limited coverage of 100+ languages lets 6500+ languages remain ‘unseen’. One common approach for an unseen language is specializing the model for it as target, by performing additional masked language modeling (MLM) with the target language corpus. However, we argue that, due to the discrepancy from multilingual MLM pretraining, a naive specialization as such can be suboptimal. Specifically, we pose three discrepancies to overcome. Script and linguistic discrepancy of the target language from the related seen languages, hinder a positive transfer, for which we propose to maximize representation similarity, unlike existing approaches maximizing overlaps. In addition, label space for MLM prediction can vary across languages, for which we propose to reinitialize top layers for a more effective adaptation. Experiments over four different language families and three tasks shows that our method improves the task performance of unseen languages with statistical significance, while previous approach fails to.
尽管多语言预训练模型(mplm)支持多种语言的各种自然语言处理,但其对100多种语言的有限覆盖使6500多种语言仍然“看不见”。对于看不见的语言,一种常见的方法是通过使用目标语言语料库执行额外的掩码语言建模(MLM),将其模型专门化为目标语言。然而,我们认为,由于多语言传销预训练的差异,这样的朴素专业化可能是次优的。具体来说,我们提出了三个需要克服的差异。目标语的文字和语言差异会阻碍正向迁移,因此我们建议最大化表征相似性,而不是现有的最大化重叠的方法。此外,传销预测的标签空间可能因语言而异,为此我们建议重新初始化顶层以获得更有效的适应。在四个不同的语族和三个任务上的实验表明,我们的方法对未见过的语言的任务性能有显著的提高,而以前的方法没有。
{"title":"Script, Language, and Labels: Overcoming Three Discrepancies for Low-Resource Language Specialization","authors":"Jaeseong Lee, Dohyeon Lee, Seung-won Hwang","doi":"10.1609/aaai.v37i11.26528","DOIUrl":"https://doi.org/10.1609/aaai.v37i11.26528","url":null,"abstract":"Although multilingual pretrained models (mPLMs) enabled support of various natural language processing in diverse languages, its limited coverage of 100+ languages lets 6500+ languages remain ‘unseen’. One common approach for an unseen language is specializing the model for it as target, by performing additional masked language modeling (MLM) with the target language corpus. However, we argue that, due to the discrepancy from multilingual MLM pretraining, a naive specialization as such can be suboptimal. Specifically, we pose three discrepancies to overcome. Script and linguistic discrepancy of the target language from the related seen languages, hinder a positive transfer, for which we propose to maximize representation similarity, unlike existing approaches maximizing overlaps. In addition, label space for MLM prediction can vary across languages, for which we propose to reinitialize top layers for a more effective adaptation. Experiments over four different language families and three tasks shows that our method improves the task performance of unseen languages with statistical significance, while previous approach fails to.","PeriodicalId":74506,"journal":{"name":"Proceedings of the ... AAAI Conference on Artificial Intelligence. AAAI Conference on Artificial Intelligence","volume":"33 1","pages":"13004-13013"},"PeriodicalIF":0.0,"publicationDate":"2023-06-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"91254954","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
See How You Read? Multi-Reading Habits Fusion Reasoning for Multi-Modal Fake News Detection 看看你是如何阅读的?多阅读习惯融合推理的多模态假新闻检测
Lianwei Wu, Pusheng Liu, Yanning Zhang
The existing approaches based on different neural networks automatically capture and fuse the multimodal semantics of news, which have achieved great success for fake news detection. However, they still suffer from the limitations of both shallow fusion of multimodal features and less attention to the inconsistency between different modalities. To overcome them, we propose multi-reading habits fusion reasoning networks (MRHFR) for multi-modal fake news detection. In MRHFR, inspired by people's different reading habits for multimodal news, we summarize three basic cognitive reading habits and put forward cognition-aware fusion layer to learn the dependencies between multimodal features of news, so as to deepen their semantic-level integration. To explore the inconsistency of different modalities of news, we develop coherence constraint reasoning layer from two perspectives, which first measures the semantic consistency between the comments and different modal features of the news, and then probes the semantic deviation caused by unimodal features to the multimodal news content through constraint strategy. Experiments on two public datasets not only demonstrate that MRHFR not only achieves the excellent performance but also provides a new paradigm for capturing inconsistencies between multi-modal news.
现有的基于不同神经网络的方法自动捕获和融合新闻的多模态语义,在假新闻检测中取得了很大的成功。然而,它们仍然存在多模态特征的浅融合和对不同模态之间不一致性关注不足的局限性。为了克服这些问题,我们提出了多阅读习惯融合推理网络(MRHFR)用于多模态假新闻检测。在MRHFR中,受人们对多模态新闻的不同阅读习惯的启发,我们总结了三种基本的认知阅读习惯,并提出了认知感知融合层来学习新闻多模态特征之间的依赖关系,从而加深它们在语义层面的融合。为了探究新闻不同模态的不一致性,我们从两个角度构建了连贯约束推理层,首先衡量新闻评论与不同模态特征之间的语义一致性,然后通过约束策略探究单模态特征对多模态新闻内容造成的语义偏差。在两个公共数据集上的实验表明,MRHFR不仅取得了优异的性能,而且为捕获多模态新闻之间的不一致性提供了一种新的范式。
{"title":"See How You Read? Multi-Reading Habits Fusion Reasoning for Multi-Modal Fake News Detection","authors":"Lianwei Wu, Pusheng Liu, Yanning Zhang","doi":"10.1609/aaai.v37i11.26609","DOIUrl":"https://doi.org/10.1609/aaai.v37i11.26609","url":null,"abstract":"The existing approaches based on different neural networks automatically capture and fuse the multimodal semantics of news, which have achieved great success for fake news detection. However, they still suffer from the limitations of both shallow fusion of multimodal features and less attention to the inconsistency between different modalities. To overcome them, we propose multi-reading habits fusion reasoning networks (MRHFR) for multi-modal fake news detection. In MRHFR, inspired by people's different reading habits for multimodal news, we summarize three basic cognitive reading habits and put forward cognition-aware fusion layer to learn the dependencies between multimodal features of news, so as to deepen their semantic-level integration. To explore the inconsistency of different modalities of news, we develop coherence constraint reasoning layer from two perspectives, which first measures the semantic consistency between the comments and different modal features of the news, and then probes the semantic deviation caused by unimodal features to the multimodal news content through constraint strategy. Experiments on two public datasets not only demonstrate that MRHFR not only achieves the excellent performance but also provides a new paradigm for capturing inconsistencies between multi-modal news.","PeriodicalId":74506,"journal":{"name":"Proceedings of the ... AAAI Conference on Artificial Intelligence. AAAI Conference on Artificial Intelligence","volume":"10 1","pages":"13736-13744"},"PeriodicalIF":0.0,"publicationDate":"2023-06-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"89942549","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
FeedFormer: Revisiting Transformer Decoder for Efficient Semantic Segmentation FeedFormer:用于高效语义分割的重访变压器解码器
J. Shim, Hyunwoo Yu, Kyeongbo Kong, Suk-Ju Kang
With the success of Vision Transformer (ViT) in image classification, its variants have yielded great success in many downstream vision tasks. Among those, the semantic segmentation task has also benefited greatly from the advance of ViT variants. However, most studies of the transformer for semantic segmentation only focus on designing efficient transformer encoders, rarely giving attention to designing the decoder. Several studies make attempts in using the transformer decoder as the segmentation decoder with class-wise learnable query. Instead, we aim to directly use the encoder features as the queries. This paper proposes the Feature Enhancing Decoder transFormer (FeedFormer) that enhances structural information using the transformer decoder. Our goal is to decode the high-level encoder features using the lowest-level encoder feature. We do this by formulating high-level features as queries, and the lowest-level feature as the key and value. This enhances the high-level features by collecting the structural information from the lowest-level feature. Additionally, we use a simple reformation trick of pushing the encoder blocks to take the place of the existing self-attention module of the decoder to improve efficiency. We show the superiority of our decoder with various light-weight transformer-based decoders on popular semantic segmentation datasets. Despite the minute computation, our model has achieved state-of-the-art performance in the performance computation trade-off. Our model FeedFormer-B0 surpasses SegFormer-B0 with 1.8% higher mIoU and 7.1% less computation on ADE20K, and 1.7% higher mIoU and 14.4% less computation on Cityscapes, respectively. Code will be released at: https://github.com/jhshim1995/FeedFormer.
随着视觉转换器(Vision Transformer, ViT)在图像分类中的成功,其变体在许多下游视觉任务中也取得了巨大的成功。其中,语义分割任务也得益于ViT变体的进步。然而,大多数用于语义分割的变压器的研究只关注于设计高效的变压器编码器,很少关注解码器的设计。一些研究尝试将变换解码器作为分段解码器,并提出了类可学习查询。相反,我们的目标是直接使用编码器特性作为查询。本文提出了一种特征增强解码器变压器(FeedFormer),利用变压器解码器增强结构信息。我们的目标是使用最低级编码器特征来解码高级编码器特征。我们通过将高级特性表述为查询,将最低级特性表述为键和值来实现这一点。这通过从最低级别的特征中收集结构信息来增强高级特征。此外,我们还采用了一种简单的改造技巧,即推入编码器块来取代现有的解码器自关注模块,以提高效率。我们在常用的语义分割数据集上展示了我们的解码器与各种轻量级的基于转换器的解码器的优势。尽管计算时间很短,但我们的模型在性能计算权衡方面取得了最先进的性能。我们的模型FeedFormer-B0超过SegFormer-B0,在ADE20K上的mIoU提高1.8%,计算量减少7.1%,在cityscape上的mIoU提高1.7%,计算量减少14.4%。代码将在https://github.com/jhshim1995/FeedFormer上发布。
{"title":"FeedFormer: Revisiting Transformer Decoder for Efficient Semantic Segmentation","authors":"J. Shim, Hyunwoo Yu, Kyeongbo Kong, Suk-Ju Kang","doi":"10.1609/aaai.v37i2.25321","DOIUrl":"https://doi.org/10.1609/aaai.v37i2.25321","url":null,"abstract":"With the success of Vision Transformer (ViT) in image classification, its variants have yielded great success in many downstream vision tasks. Among those, the semantic segmentation task has also benefited greatly from the advance of ViT variants. However, most studies of the transformer for semantic segmentation only focus on designing efficient transformer encoders, rarely giving attention to designing the decoder. Several studies make attempts in using the transformer decoder as the segmentation decoder with class-wise learnable query. Instead, we aim to directly use the encoder features as the queries. This paper proposes the Feature Enhancing Decoder transFormer (FeedFormer) that enhances structural information using the transformer decoder. Our goal is to decode the high-level encoder features using the lowest-level encoder feature. We do this by formulating high-level features as queries, and the lowest-level feature as the key and value. This enhances the high-level features by collecting the structural information from the lowest-level feature. Additionally, we use a simple reformation trick of pushing the encoder blocks to take the place of the existing self-attention module of the decoder to improve efficiency. We show the superiority of our decoder with various light-weight transformer-based decoders on popular semantic segmentation datasets. Despite the minute computation, our model has achieved state-of-the-art performance in the performance computation trade-off. Our model FeedFormer-B0 surpasses SegFormer-B0 with 1.8% higher mIoU and 7.1% less computation on ADE20K, and 1.7% higher mIoU and 14.4% less computation on Cityscapes, respectively. Code will be released at: https://github.com/jhshim1995/FeedFormer.","PeriodicalId":74506,"journal":{"name":"Proceedings of the ... AAAI Conference on Artificial Intelligence. AAAI Conference on Artificial Intelligence","volume":"5 1","pages":"2263-2271"},"PeriodicalIF":0.0,"publicationDate":"2023-06-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"90036945","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Proceedings of the ... AAAI Conference on Artificial Intelligence. AAAI Conference on Artificial Intelligence
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1