首页 > 最新文献

AAAI Conference on Artificial Intelligence最新文献

英文 中文
Learning to Stop Cut Generation for Efficient Mixed-Integer Linear Programming 高效混合整数线性规划的停止切分生成学习
Pub Date : 2024-01-31 DOI: 10.48550/arXiv.2401.17527
Haotian Ling, Zhihai Wang, Jie Wang
Cutting planes (cuts) play an important role in solving mixed-integer linear programs (MILPs), as they significantly tighten the dual bounds and improve the solving performance. A key problem for cuts is when to stop cuts generation, which is important for the efficiency of solving MILPs. However, many modern MILP solvers employ hard-coded heuristics to tackle this problem, which tends to neglect underlying patterns among MILPs from certain applications. To address this challenge, we formulate the cuts generation stopping problem as a reinforcement learning problem and propose a novel hybrid graph representation model (HYGRO) to learn effective stopping strategies. An appealing feature of HYGRO is that it can effectively capture both the dynamic and static features of MILPs, enabling dynamic decision-making for the stopping strategies. To the best of our knowledge, HYGRO is the first data-driven method to tackle the cuts generation stopping problem. By integrating our approach with modern solvers, experiments demonstrate that HYGRO significantly improves the efficiency of solving MILPs compared to competitive baselines, achieving up to 31% improvement.
切割平面(切割)在求解混合整数线性方程组(MILPs)中发挥着重要作用,因为它们能显著收窄对偶边界并提高求解性能。切平面的一个关键问题是何时停止生成切平面,这对提高 MILP 的求解效率非常重要。然而,许多现代 MILP 求解器采用硬编码启发式方法来解决这个问题,这往往会忽略某些应用中 MILP 之间的基本模式。为了应对这一挑战,我们将切分生成停止问题表述为强化学习问题,并提出了一种新型混合图表示模型(HYGRO)来学习有效的停止策略。HYGRO 的一个吸引人的特点是,它能有效捕捉 MILPs 的动态和静态特征,从而实现停止策略的动态决策。据我们所知,HYGRO 是第一种以数据为驱动的方法来解决切分生成停止问题。通过将我们的方法与现代求解器相结合,实验证明,与竞争基线相比,HYGRO 显著提高了 MILPs 的求解效率,提高幅度高达 31%。
{"title":"Learning to Stop Cut Generation for Efficient Mixed-Integer Linear Programming","authors":"Haotian Ling, Zhihai Wang, Jie Wang","doi":"10.48550/arXiv.2401.17527","DOIUrl":"https://doi.org/10.48550/arXiv.2401.17527","url":null,"abstract":"Cutting planes (cuts) play an important role in solving mixed-integer linear programs (MILPs), as they significantly tighten the dual bounds and improve the solving performance. A key problem for cuts is when to stop cuts generation, which is important for the efficiency of solving MILPs. However, many modern MILP solvers employ hard-coded heuristics to tackle this problem, which tends to neglect underlying patterns among MILPs from certain applications. To address this challenge, we formulate the cuts generation stopping problem as a reinforcement learning problem and propose a novel hybrid graph representation model (HYGRO) to learn effective stopping strategies. An appealing feature of HYGRO is that it can effectively capture both the dynamic and static features of MILPs, enabling dynamic decision-making for the stopping strategies. To the best of our knowledge, HYGRO is the first data-driven method to tackle the cuts generation stopping problem. By integrating our approach with modern solvers, experiments demonstrate that HYGRO significantly improves the efficiency of solving MILPs compared to competitive baselines, achieving up to 31% improvement.","PeriodicalId":518480,"journal":{"name":"AAAI Conference on Artificial Intelligence","volume":"206 ","pages":"20759-20767"},"PeriodicalIF":0.0,"publicationDate":"2024-01-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140529474","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Cross-View Hierarchical Graph Learning Hypernetwork for Skill Demand-Supply Joint Prediction 用于技能供需联合预测的跨视图层次图学习超网络
Pub Date : 2024-01-31 DOI: 10.48550/arXiv.2401.17838
Wenshuo Chao, Zhaopeng Qiu, Likang Wu, Zhuoning Guo, Zhi Zheng, Hengshu Zhu, Hao Liu
The rapidly changing landscape of technology and industries leads to dynamic skill requirements, making it crucial for employees and employers to anticipate such shifts to maintain a competitive edge in the labor market. Existing efforts in this area either relies on domain-expert knowledge or regarding the skill evolution as a simplified time series forecasting problem. However, both approaches overlook the sophisticated relationships among different skills and the inner-connection between skill demand and supply variations. In this paper, we propose a Cross-view Hierarchical Graph learning Hypernetwork (CHGH) framework for joint skill demand-supply prediction. Specifically, CHGH is an encoder-decoder network consisting of i) a cross-view graph encoder to capture the interconnection between skill demand and supply, ii) a hierarchical graph encoder to model the co-evolution of skills from a cluster-wise perspective, and iii) a conditional hyper-decoder to jointly predict demand and supply variations by incorporating historical demand-supply gaps. Extensive experiments on three real-world datasets demonstrate the superiority of the proposed framework compared to seven baselines and the effectiveness of the three modules.
技术和行业的快速变化导致技能要求的动态变化,因此,雇员和雇主必须预测这种变化,以保持在劳动力市场上的竞争优势。该领域的现有研究要么依赖领域专家的知识,要么将技能演变视为一个简化的时间序列预测问题。然而,这两种方法都忽略了不同技能之间的复杂关系以及技能需求和供给变化之间的内在联系。在本文中,我们提出了一种用于技能供需联合预测的跨视图分层图学习超网络(Cross-view Hierarchical Graph learning Hypernetwork,CGH)框架。具体来说,CHGH 是一个编码器-解码器网络,包括 i) 一个跨视图图编码器,用于捕捉技能需求和供给之间的相互联系;ii) 一个分层图编码器,用于从集群的角度对技能的共同演化进行建模;iii) 一个条件超解码器,用于通过纳入历史供需缺口来联合预测需求和供给的变化。在三个真实世界数据集上进行的广泛实验证明,与七个基线相比,所提出的框架更有优势,而且三个模块都很有效。
{"title":"A Cross-View Hierarchical Graph Learning Hypernetwork for Skill Demand-Supply Joint Prediction","authors":"Wenshuo Chao, Zhaopeng Qiu, Likang Wu, Zhuoning Guo, Zhi Zheng, Hengshu Zhu, Hao Liu","doi":"10.48550/arXiv.2401.17838","DOIUrl":"https://doi.org/10.48550/arXiv.2401.17838","url":null,"abstract":"The rapidly changing landscape of technology and industries leads to dynamic skill requirements, making it crucial for employees and employers to anticipate such shifts to maintain a competitive edge in the labor market. Existing efforts in this area either relies on domain-expert knowledge or regarding the skill evolution as a simplified time series forecasting problem. However, both approaches overlook the sophisticated relationships among different skills and the inner-connection between skill demand and supply variations. \u0000In this paper, we propose a Cross-view Hierarchical Graph learning Hypernetwork (CHGH) framework for joint skill demand-supply prediction. Specifically, CHGH is an encoder-decoder network consisting of i) a cross-view graph encoder to capture the interconnection between skill demand and supply, ii) a hierarchical graph encoder to model the co-evolution of skills from a cluster-wise perspective, and iii) a conditional hyper-decoder to jointly predict demand and supply variations by incorporating historical demand-supply gaps. Extensive experiments on three real-world datasets demonstrate the superiority of the proposed framework compared to seven baselines and the effectiveness of the three modules.","PeriodicalId":518480,"journal":{"name":"AAAI Conference on Artificial Intelligence","volume":"173 ","pages":"19813-19822"},"PeriodicalIF":0.0,"publicationDate":"2024-01-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140529291","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Knowledge-Aware Neuron Interpretation for Scene Classification 用于场景分类的知识感知神经元解读技术
Pub Date : 2024-01-29 DOI: 10.48550/arXiv.2401.15820
Yong Guan, Freddy Lecue, Jiaoyan Chen, Ru Li, Jeff Z. Pan
Although neural models have achieved remarkable performance, they still encounter doubts due to the intransparency. To this end, model prediction explanation is attracting more and more attentions. However, current methods rarely incorporate external knowledge and still suffer from three limitations: (1) Neglecting concept completeness. Merely selecting concepts may not sufficient for prediction. (2) Lacking concept fusion. Failure to merge semantically-equivalent concepts. (3) Difficult in manipulating model behavior. Lack of verification for explanation on original model. To address these issues, we propose a novel knowledge-aware neuron interpretation framework to explain model predictions for image scene classification. Specifically, for concept completeness, we present core concepts of a scene based on knowledge graph, ConceptNet, to gauge the completeness of concepts. Our method, incorporating complete concepts, effectively provides better prediction explanations compared to baselines. Furthermore, for concept fusion, we introduce a knowledge graph-based method known as Concept Filtering, which produces over 23% point gain on neuron behaviors for neuron interpretation. At last, we propose Model Manipulation, which aims to study whether the core concepts based on ConceptNet could be employed to manipulate model behavior. The results show that core concepts can effectively improve the performance of original model by over 26%.
尽管神经模型已经取得了不俗的成绩,但由于其不透明性,仍会受到质疑。为此,模型预测解释受到越来越多的关注。然而,目前的方法很少结合外部知识,仍然存在三个局限性:(1)忽视概念的完整性。仅仅选择概念可能不足以进行预测。(2) 缺乏概念融合。无法合并语义相等的概念。(3) 难以操作模型行为。缺乏对原始模型解释的验证。为了解决这些问题,我们提出了一个新颖的知识感知神经元解释框架来解释图像场景分类的模型预测。具体来说,针对概念的完整性,我们基于知识图谱 ConceptNet 提出了场景的核心概念,以衡量概念的完整性。与基线方法相比,我们的方法结合了完整概念,能有效地提供更好的预测解释。此外,在概念融合方面,我们引入了一种基于知识图谱的方法,即概念过滤法(Concept Filtering),该方法在神经元行为解释方面的增益超过 23%。最后,我们提出了 "模型操纵"(Model Manipulation),旨在研究是否可以利用基于 ConceptNet 的核心概念来操纵模型行为。结果表明,核心概念可以有效提高原始模型的性能,提高幅度超过 26%。
{"title":"Knowledge-Aware Neuron Interpretation for Scene Classification","authors":"Yong Guan, Freddy Lecue, Jiaoyan Chen, Ru Li, Jeff Z. Pan","doi":"10.48550/arXiv.2401.15820","DOIUrl":"https://doi.org/10.48550/arXiv.2401.15820","url":null,"abstract":"Although neural models have achieved remarkable performance, they still encounter doubts due to the intransparency. To this end, model prediction explanation is attracting more and more attentions. However, current methods rarely incorporate external knowledge and still suffer from three limitations: (1) Neglecting concept completeness. Merely selecting concepts may not sufficient for prediction. (2) Lacking concept fusion. Failure to merge semantically-equivalent concepts. (3) Difficult in manipulating model behavior. Lack of verification for explanation on original model. To address these issues, we propose a novel knowledge-aware neuron interpretation framework to explain model predictions for image scene classification. Specifically, for concept completeness, we present core concepts of a scene based on knowledge graph, ConceptNet, to gauge the completeness of concepts. Our method, incorporating complete concepts, effectively provides better prediction explanations compared to baselines. Furthermore, for concept fusion, we introduce a knowledge graph-based method known as Concept Filtering, which produces over 23% point gain on neuron behaviors for neuron interpretation. At last, we propose Model Manipulation, which aims to study whether the core concepts based on ConceptNet could be employed to manipulate model behavior. The results show that core concepts can effectively improve the performance of original model by over 26%.","PeriodicalId":518480,"journal":{"name":"AAAI Conference on Artificial Intelligence","volume":"72 5","pages":"1950-1958"},"PeriodicalIF":0.0,"publicationDate":"2024-01-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140529588","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Continuous Treatment Effect Estimation Using Gradient Interpolation and Kernel Smoothing 利用梯度插值和核平滑法进行连续治疗效果估计
Pub Date : 2024-01-27 DOI: 10.48550/arXiv.2401.15447
Lokesh Nagalapatti, Akshay Iyer, Abir De, Sunita Sarawagi
We address the Individualized continuous treatment effect(ICTE) estimation problem where we predict the effect ofany continuous valued treatment on an individual using ob-servational data. The main challenge in this estimation taskis the potential confounding of treatment assignment with in-dividual’s covariates in the training data, whereas during in-ference ICTE requires prediction on independently sampledtreatments. In contrast to prior work that relied on regularizersor unstable GAN training, we advocate the direct approachof augmenting training individuals with independently sam-pled treatments and inferred counterfactual outcomes. We in-fer counterfactual outcomes using a two-pronged strategy: aGradient Interpolation for close-to-observed treatments, anda Gaussian Process based Kernel Smoothing which allowsus to down weigh high variance inferences. We evaluate ourmethod on five benchmarks and show that our method out-performs six state-of-the-art methods on the counterfactualestimation error. We analyze the superior performance of ourmethod by showing that (1) our inferred counterfactual re-sponses are more accurate, and (2) adding them to the train-ing data reduces the distributional distance between the con-founded training distribution and test distribution where treat-ment is independent of covariates. Our proposed method ismodel-agnostic and we show that it improves ICTE accuracyof several existing models.
我们要解决个体化连续治疗效果(ICTE)估计问题,即利用观察数据预测任何连续治疗对个体的效果。这一估计任务的主要挑战在于治疗分配可能与训练数据中的个体协变量混淆,而在推断过程中,ICTE 需要对独立采样的治疗进行预测。与之前依赖正则化器或不稳定 GAN 训练的工作不同,我们主张直接使用独立采样的治疗方法和推断出的反事实结果来增强训练个体。我们使用双管齐下的策略来推断反事实结果:梯度插值法用于接近观察到的处理方法,而基于高斯过程的核平滑法则允许我们降低高方差推断的权重。我们在五个基准上对我们的方法进行了评估,结果表明我们的方法在反事实估计误差方面优于六种最先进的方法。我们分析了我们的方法的优越性能,表明:(1) 我们推断出的反事实再反应更准确;(2) 将它们添加到训练数据中,可以减少在处理与协变量无关的情况下,有根据的训练分布与测试分布之间的分布距离。我们提出的方法与模型无关,而且我们的研究表明,它提高了几个现有模型的 ICTE 准确性。
{"title":"Continuous Treatment Effect Estimation Using Gradient Interpolation and Kernel Smoothing","authors":"Lokesh Nagalapatti, Akshay Iyer, Abir De, Sunita Sarawagi","doi":"10.48550/arXiv.2401.15447","DOIUrl":"https://doi.org/10.48550/arXiv.2401.15447","url":null,"abstract":"We address the Individualized continuous treatment effect\u0000(ICTE) estimation problem where we predict the effect of\u0000any continuous valued treatment on an individual using ob-\u0000servational data. The main challenge in this estimation task\u0000is the potential confounding of treatment assignment with in-\u0000dividual’s covariates in the training data, whereas during in-\u0000ference ICTE requires prediction on independently sampled\u0000treatments. In contrast to prior work that relied on regularizers\u0000or unstable GAN training, we advocate the direct approach\u0000of augmenting training individuals with independently sam-\u0000pled treatments and inferred counterfactual outcomes. We in-\u0000fer counterfactual outcomes using a two-pronged strategy: a\u0000Gradient Interpolation for close-to-observed treatments, and\u0000a Gaussian Process based Kernel Smoothing which allows\u0000us to down weigh high variance inferences. We evaluate our\u0000method on five benchmarks and show that our method out-\u0000performs six state-of-the-art methods on the counterfactual\u0000estimation error. We analyze the superior performance of our\u0000method by showing that (1) our inferred counterfactual re-\u0000sponses are more accurate, and (2) adding them to the train-\u0000ing data reduces the distributional distance between the con-\u0000founded training distribution and test distribution where treat-\u0000ment is independent of covariates. Our proposed method is\u0000model-agnostic and we show that it improves ICTE accuracy\u0000of several existing models.","PeriodicalId":518480,"journal":{"name":"AAAI Conference on Artificial Intelligence","volume":"178 6","pages":"14397-14404"},"PeriodicalIF":0.0,"publicationDate":"2024-01-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140530287","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Sketch and Refine: Towards Fast and Accurate Lane Detection 素描与提炼:实现快速准确的车道检测
Pub Date : 2024-01-26 DOI: 10.48550/arXiv.2401.14729
Chao Chen, Jie Liu, Chang Zhou, Jie Tang, Gangshan Wu
Lane detection is to determine the precise location and shape of lanes on the road. Despite efforts made by current methods, it remains a challenging task due to the complexity of real-world scenarios. Existing approaches, whether proposal-based or keypoint-based, suffer from depicting lanes effectively and efficiently. Proposal-based methods detect lanes by distinguishing and regressing a collection of proposals in a streamlined top-down way, yet lack sufficient flexibility in lane representation. Keypoint-based methods, on the other hand, construct lanes flexibly from local descriptors, which typically entail complicated post-processing. In this paper, we present a “Sketch-and-Refine” paradigm that utilizes the merits of both keypoint-based and proposal-based methods. The motivation is that local directions of lanes are semantically simple and clear. At the “Sketch” stage, local directions of keypoints can be easily estimated by fast convolutional layers. Then we can build a set of lane proposals accordingly with moderate accuracy. At the “Refine” stage, we further optimize these proposals via a novel Lane Segment Association Module (LSAM), which allows adaptive lane segment adjustment. Last but not least, we propose multi-level feature integration to enrich lane feature representations more efficiently. Based on the proposed “Sketch-and-Refine” paradigm, we propose a fast yet effective lane detector dubbed “SRLane”. Experiments show that our SRLane can run at a fast speed (i.e., 278 FPS) while yielding an F1 score of 78.9%. The source code is available at: https://github.com/passerer/SRLane.
车道检测是指确定道路上车道的精确位置和形状。尽管目前的方法已经做出了努力,但由于现实世界场景的复杂性,这仍然是一项具有挑战性的任务。无论是基于建议的方法还是基于关键点的方法,现有方法都无法有效、高效地描绘车道。基于建议的方法通过自上而下的精简方式对建议集合进行区分和回归来检测车道,但在车道表示方面缺乏足够的灵活性。另一方面,基于关键点的方法可根据局部描述符灵活构建车道,但通常需要复杂的后处理。在本文中,我们提出了一种 "勾勒-再细化 "范式,它同时利用了基于关键点和基于建议的方法的优点。其动机在于,车道的局部方向在语义上简单明了。在 "草图 "阶段,关键点的局部方向可以很容易地通过快速卷积层估算出来。然后,我们就可以相应地建立一套车道建议,准确度适中。在 "细化 "阶段,我们通过新颖的车道分段关联模块(LSAM)进一步优化这些建议,该模块允许自适应车道分段调整。最后,我们提出了多层次特征整合,以更有效地丰富车道特征表征。基于所提出的 "勾勒-提炼 "范式,我们提出了一种快速而有效的车道检测器,称为 "SRLane"。实验表明,我们的 SRLane 能以较快的速度运行(即 278 FPS),同时获得 78.9% 的 F1 分数。源代码见:https://github.com/passerer/SRLane。
{"title":"Sketch and Refine: Towards Fast and Accurate Lane Detection","authors":"Chao Chen, Jie Liu, Chang Zhou, Jie Tang, Gangshan Wu","doi":"10.48550/arXiv.2401.14729","DOIUrl":"https://doi.org/10.48550/arXiv.2401.14729","url":null,"abstract":"Lane detection is to determine the precise location and shape of lanes on the road. Despite efforts made by current methods, it remains a challenging task due to the complexity of real-world scenarios. Existing approaches, whether proposal-based or keypoint-based, suffer from depicting lanes effectively and efficiently. Proposal-based methods detect lanes by distinguishing and regressing a collection of proposals in a streamlined top-down way, yet lack sufficient flexibility in lane representation. Keypoint-based methods, on the other hand, construct lanes flexibly from local descriptors, which typically entail complicated post-processing. In this paper, we present a “Sketch-and-Refine” paradigm that utilizes the merits of both keypoint-based and proposal-based methods. The motivation is that local directions of lanes are semantically simple and clear. At the “Sketch” stage, local directions of keypoints can be easily estimated by fast convolutional layers. Then we can build a set of lane proposals accordingly with moderate accuracy. At the “Refine” stage, we further optimize these proposals via a novel Lane Segment Association Module (LSAM), which allows adaptive lane segment adjustment. Last but not least, we propose multi-level feature integration to enrich lane feature representations more efficiently. Based on the proposed “Sketch-and-Refine” paradigm, we propose a fast yet effective lane detector dubbed “SRLane”. Experiments show that our SRLane can run at a fast speed (i.e., 278 FPS) while yielding an F1 score of 78.9%. The source code is available at: https://github.com/passerer/SRLane.","PeriodicalId":518480,"journal":{"name":"AAAI Conference on Artificial Intelligence","volume":"23 2","pages":"1001-1009"},"PeriodicalIF":0.0,"publicationDate":"2024-01-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140530110","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Friendly Attacks to Improve Channel Coding Reliability 提高信道编码可靠性的友好攻击
Pub Date : 2024-01-25 DOI: 10.48550/arXiv.2401.14184
Anastasia Kurmukova, Deniz Gündüz
This paper introduces a novel approach called "friendly attack" aimed at enhancing the performance of error correction channel codes. Inspired by the concept of adversarial attacks, our method leverages the idea of introducing slight perturbations to the neural network input, resulting in a substantial impact on the network's performance. By introducing small perturbations to fixed-point modulated codewords before transmission, we effectively improve the decoder's performance without violating the input power constraint. The perturbation design is accomplished by a modified iterative fast gradient method. This study investigates various decoder architectures suitable for computing gradients to obtain the desired perturbations. Specifically, we consider belief propagation (BP) for LDPC codes; the error correcting code transformer, BP and neural BP (NBP) for polar codes, and neural BCJR for convolutional codes. We demonstrate that the proposed friendly attack method can improve the reliability across different channels, modulations, codes, and decoders. This method allows us to increase the reliability of communication with a legacy receiver by simply modifying the transmitted codeword appropriately.
本文介绍了一种名为 "友好攻击 "的新方法,旨在提高纠错信道编码的性能。受对抗攻击概念的启发,我们的方法利用了在神经网络输入中引入微小扰动,从而对网络性能产生重大影响的思想。通过在传输前对定点调制码字引入微小扰动,我们可以在不违反输入功率约束的情况下有效提高解码器的性能。扰动设计是通过改进的迭代快速梯度法完成的。本研究探讨了适合计算梯度以获得所需扰动的各种解码器架构。具体来说,我们考虑了 LDPC 码的信念传播 (BP)、纠错码变换器、极性码的 BP 和神经 BP (NBP),以及卷积码的神经 BCJR。我们证明,所提出的友好攻击方法可以提高不同信道、调制、编码和解码器的可靠性。通过这种方法,我们只需适当修改传输的码字,就能提高与传统接收器通信的可靠性。
{"title":"Friendly Attacks to Improve Channel Coding Reliability","authors":"Anastasia Kurmukova, Deniz Gündüz","doi":"10.48550/arXiv.2401.14184","DOIUrl":"https://doi.org/10.48550/arXiv.2401.14184","url":null,"abstract":"This paper introduces a novel approach called \"friendly attack\" aimed at enhancing the performance of error correction channel codes. Inspired by the concept of adversarial attacks, our method leverages the idea of introducing slight perturbations to the neural network input, resulting in a substantial impact on the network's performance. By introducing small perturbations to fixed-point modulated codewords before transmission, we effectively improve the decoder's performance without violating the input power constraint. The perturbation design is accomplished by a modified iterative fast gradient method. This study investigates various decoder architectures suitable for computing gradients to obtain the desired perturbations. Specifically, we consider belief propagation (BP) for LDPC codes; the error correcting code transformer, BP and neural BP (NBP) for polar codes, and neural BCJR for convolutional codes. We demonstrate that the proposed friendly attack method can improve the reliability across different channels, modulations, codes, and decoders. This method allows us to increase the reliability of communication with a legacy receiver by simply modifying the transmitted codeword appropriately.","PeriodicalId":518480,"journal":{"name":"AAAI Conference on Artificial Intelligence","volume":"26 1","pages":"13292-13300"},"PeriodicalIF":0.0,"publicationDate":"2024-01-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140530685","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
On the Affinity, Rationality, and Diversity of Hierarchical Topic Modeling 论分层主题建模的亲和性、合理性和多样性
Pub Date : 2024-01-25 DOI: 10.48550/arXiv.2401.14113
Xiaobao Wu, Fengjun Pan, Thong Nguyen, Yichao Feng, Chaoqun Liu, Cong-Duy Nguyen, A. Luu
Hierarchical topic modeling aims to discover latent topics from a corpus and organize them into a hierarchy to understand documents with desirable semantic granularity. However, existing work struggles with producing topic hierarchies of low affinity, rationality, and diversity, which hampers document understanding. To overcome these challenges, we in this paper propose Transport Plan and Context-aware Hierarchical Topic Model (TraCo). Instead of early simple topic dependencies, we propose a transport plan dependency method. It constrains dependencies to ensure their sparsity and balance, and also regularizes topic hierarchy building with them. This improves affinity and diversity of hierarchies. We further propose a context-aware disentangled decoder. Rather than previously entangled decoding, it distributes different semantic granularity to topics at different levels by disentangled decoding. This facilitates the rationality of hierarchies. Experiments on benchmark datasets demonstrate that our method surpasses state-of-the-art baselines, effectively improving the affinity, rationality, and diversity of hierarchical topic modeling with better performance on downstream tasks.
分层主题建模旨在从语料库中发现潜在主题,并将其组织成一个层次结构,从而以理想的语义粒度理解文档。然而,现有的工作难以产生亲和性、合理性和多样性较低的主题层次,这阻碍了对文档的理解。为了克服这些挑战,我们在本文中提出了运输计划和上下文感知分层主题模型(TraCo)。我们提出了一种运输计划依赖方法,而不是早期的简单主题依赖。该方法对依赖关系进行约束,以确保其稀疏性和平衡性,并利用依赖关系对话题层次结构进行规范化构建。这提高了层次结构的亲和性和多样性。我们进一步提出了一种上下文感知的纠缠解码器。它不是以前的纠缠解码器,而是通过非纠缠解码将不同语义粒度分配给不同层次的主题。这促进了分层的合理性。在基准数据集上进行的实验证明,我们的方法超越了最先进的基线,有效改善了分层主题建模的亲和性、合理性和多样性,在下游任务中表现更佳。
{"title":"On the Affinity, Rationality, and Diversity of Hierarchical Topic Modeling","authors":"Xiaobao Wu, Fengjun Pan, Thong Nguyen, Yichao Feng, Chaoqun Liu, Cong-Duy Nguyen, A. Luu","doi":"10.48550/arXiv.2401.14113","DOIUrl":"https://doi.org/10.48550/arXiv.2401.14113","url":null,"abstract":"Hierarchical topic modeling aims to discover latent topics from a corpus and organize them into a hierarchy to understand documents with desirable semantic granularity. However, existing work struggles with producing topic hierarchies of low affinity, rationality, and diversity, which hampers document understanding. To overcome these challenges, we in this paper propose Transport Plan and Context-aware Hierarchical Topic Model (TraCo). Instead of early simple topic dependencies, we propose a transport plan dependency method. It constrains dependencies to ensure their sparsity and balance, and also regularizes topic hierarchy building with them. This improves affinity and diversity of hierarchies. We further propose a context-aware disentangled decoder. Rather than previously entangled decoding, it distributes different semantic granularity to topics at different levels by disentangled decoding. This facilitates the rationality of hierarchies. Experiments on benchmark datasets demonstrate that our method surpasses state-of-the-art baselines, effectively improving the affinity, rationality, and diversity of hierarchical topic modeling with better performance on downstream tasks.","PeriodicalId":518480,"journal":{"name":"AAAI Conference on Artificial Intelligence","volume":"279 1-2","pages":"19261-19269"},"PeriodicalIF":0.0,"publicationDate":"2024-01-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140530359","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Multi-level Cross-modal Alignment for Image Clustering 图像聚类的多级跨模态对齐
Pub Date : 2024-01-22 DOI: 10.48550/arXiv.2401.11740
Liping Qiu, Qin Zhang, Xiaojun Chen, Shao-Qian Cai
Recently, the cross-modal pretraining model has been employed to produce meaningful pseudo-labels to supervise the training of an image clustering model. However, numerous erroneous alignments in a cross-modal pretraining model could produce poor-quality pseudo labels and degrade clustering performance. To solve the aforementioned issue, we propose a novel Multi-level Cross-modal Alignment method to improve the alignments in a cross-modal pretraining model for downstream tasks, by building a smaller but better semantic space and aligning the images and texts in three levels, i.e., instance-level, prototype-level, and semantic-level. Theoretical results show that our proposed method converges, and suggests effective means to reduce the expected clustering risk of our method. Experimental results on five benchmark datasets clearly show the superiority of our new method.
最近,跨模态预训练模型被用来生成有意义的伪标签,以监督图像聚类模型的训练。然而,跨模态预训练模型中的大量错误配准可能会产生劣质的伪标签并降低聚类性能。为了解决上述问题,我们提出了一种新颖的多层次跨模态对齐方法,通过建立一个更小但更好的语义空间,并在三个层次(即实例层次、原型层次和语义层次)上对图像和文本进行对齐,从而改进下游任务的跨模态预训练模型中的对齐。理论结果表明,我们提出的方法是收敛的,并提出了有效的方法来降低我们方法的预期聚类风险。在五个基准数据集上的实验结果清楚地表明了我们的新方法的优越性。
{"title":"Multi-level Cross-modal Alignment for Image Clustering","authors":"Liping Qiu, Qin Zhang, Xiaojun Chen, Shao-Qian Cai","doi":"10.48550/arXiv.2401.11740","DOIUrl":"https://doi.org/10.48550/arXiv.2401.11740","url":null,"abstract":"Recently, the cross-modal pretraining model has been employed to produce meaningful pseudo-labels to supervise the training of an image clustering model. However, numerous erroneous alignments in a cross-modal pretraining model could produce poor-quality pseudo labels and degrade clustering performance. To solve the aforementioned issue, we propose a novel Multi-level Cross-modal Alignment method to improve the alignments in a cross-modal pretraining model for downstream tasks, by building a smaller but better semantic space and aligning the images and texts in three levels, i.e., instance-level, prototype-level, and semantic-level. Theoretical results show that our proposed method converges, and suggests effective means to reduce the expected clustering risk of our method. Experimental results on five benchmark datasets clearly show the superiority of our new method.","PeriodicalId":518480,"journal":{"name":"AAAI Conference on Artificial Intelligence","volume":"290 17","pages":"14695-14703"},"PeriodicalIF":0.0,"publicationDate":"2024-01-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140530968","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Enhancing Evolving Domain Generalization through Dynamic Latent Representations 通过动态潜在表征增强不断发展的领域泛化能力
Pub Date : 2024-01-16 DOI: 10.48550/arXiv.2401.08464
Binghui Xie, Yongqiang Chen, Jiaqi Wang, Kaiwen Zhou, Bo Han, Wei Meng, James Cheng
Domain generalization is a critical challenge for machine learning systems. Prior domain generalization methods focus on extracting domain-invariant features across several stationary domains to enable generalization to new domains. However, in non-stationary tasks where new domains evolve in an underlying continuous structure, such as time, merely extracting the invariant features is insufficient for generalization to the evolving new domains. Nevertheless, it is non-trivial to learn both evolving and invariant features within a single model due to their conflicts. To bridge this gap, we build causal models to characterize the distribution shifts concerning the two patterns, and propose to learn both dynamic and invariant features via a new framework called Mutual Information-Based Sequential Autoencoders (MISTS). MISTS adopts information theoretic constraints onto sequential autoencoders to disentangle the dynamic and invariant features, and leverage an adaptive classifier to make predictions based on both evolving and invariant information. Our experimental results on both synthetic and real-world datasets demonstrate that MISTS succeeds in capturing both evolving and invariant information, and present promising results in evolving domain generalization tasks.
领域泛化是机器学习系统面临的一个重要挑战。先前的领域泛化方法侧重于提取多个静态领域的领域不变特征,从而实现对新领域的泛化。然而,在非静态任务中,新领域在时间等底层连续结构中不断演化,仅仅提取不变特征不足以泛化到不断演化的新领域。然而,由于演化特征和不变特征之间存在冲突,要在一个模型中同时学习这两种特征并非易事。为了弥补这一缺陷,我们建立了因果模型来描述这两种模式的分布变化,并提出通过一种名为 "基于互信息的序列自动编码器(MISTS)"的新框架来学习动态和不变特征。MISTS 在序列自动编码器上采用信息论约束来区分动态特征和不变特征,并利用自适应分类器根据演化信息和不变信息进行预测。我们在合成数据集和真实世界数据集上的实验结果表明,MISTS 成功地捕捉到了演化信息和不变信息,并在演化领域泛化任务中取得了可喜的成果。
{"title":"Enhancing Evolving Domain Generalization through Dynamic Latent Representations","authors":"Binghui Xie, Yongqiang Chen, Jiaqi Wang, Kaiwen Zhou, Bo Han, Wei Meng, James Cheng","doi":"10.48550/arXiv.2401.08464","DOIUrl":"https://doi.org/10.48550/arXiv.2401.08464","url":null,"abstract":"Domain generalization is a critical challenge for machine learning systems. Prior domain generalization methods focus on extracting domain-invariant features across several stationary domains to enable generalization to new domains. However, in non-stationary tasks where new domains evolve in an underlying continuous structure, such as time, merely extracting the invariant features is insufficient for generalization to the evolving new domains. Nevertheless, it is non-trivial to learn both evolving and invariant features within a single model due to their conflicts. To bridge this gap, we build causal models to characterize the distribution shifts concerning the two patterns, and propose to learn both dynamic and invariant features via a new framework called Mutual Information-Based Sequential Autoencoders (MISTS). MISTS adopts information theoretic constraints onto sequential autoencoders to disentangle the dynamic and invariant features, and leverage an adaptive classifier to make predictions based on both evolving and invariant information. Our experimental results on both synthetic and real-world datasets demonstrate that MISTS succeeds in capturing both evolving and invariant information, and present promising results in evolving domain generalization tasks.","PeriodicalId":518480,"journal":{"name":"AAAI Conference on Artificial Intelligence","volume":"24 4","pages":"16040-16048"},"PeriodicalIF":0.0,"publicationDate":"2024-01-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140531699","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MICA: Towards Explainable Skin Lesion Diagnosis via Multi-Level Image-Concept Alignment MICA:通过多层次图像概念对齐实现可解释的皮肤病变诊断
Pub Date : 2024-01-16 DOI: 10.48550/arXiv.2401.08527
Yequan Bie, Luyang Luo, Hao Chen
Black-box deep learning approaches have showcased significant potential in the realm of medical image analysis. However, the stringent trustworthiness requirements intrinsic to the medical field have catalyzed research into the utilization of Explainable Artificial Intelligence (XAI), with a particular focus on concept-based methods. Existing concept-based methods predominantly apply concept annotations from a single perspective (e.g., global level), neglecting the nuanced semantic relationships between sub-regions and concepts embedded within medical images. This leads to underutilization of the valuable medical information and may cause models to fall short in harmoniously balancing interpretability and performance when employing inherently interpretable architectures such as Concept Bottlenecks. To mitigate these shortcomings, we propose a multi-modal explainable disease diagnosis framework that meticulously aligns medical images and clinical-related concepts semantically at multiple strata, encompassing the image level, token level, and concept level. Moreover, our method allows for model intervention and offers both textual and visual explanations in terms of human-interpretable concepts. Experimental results on three skin image datasets demonstrate that our method, while preserving model interpretability, attains high performance and label efficiency for concept detection and disease diagnosis. The code is available at https://github.com/Tommy-Bie/MICA.
黑盒子深度学习方法已在医学图像分析领域展现出巨大潜力。然而,医疗领域固有的严格可信度要求促进了对可解释人工智能(XAI)的研究,尤其是对基于概念的方法的研究。现有的基于概念的方法主要是从单一角度(如全局水平)应用概念注释,忽略了医疗图像中嵌入的子区域和概念之间的细微语义关系。这导致宝贵的医学信息未得到充分利用,并可能导致模型在采用固有的可解释架构(如概念瓶颈)时,无法在可解释性和性能之间取得和谐的平衡。为了缓解这些缺陷,我们提出了一种多模态可解释疾病诊断框架,该框架在多个层面(包括图像层面、标记层面和概念层面)对医学图像和临床相关概念的语义进行了细致的调整。此外,我们的方法允许模型干预,并根据人类可理解的概念提供文本和视觉解释。在三个皮肤图像数据集上的实验结果表明,我们的方法在保留模型可解释性的同时,在概念检测和疾病诊断方面实现了高性能和标签效率。代码见 https://github.com/Tommy-Bie/MICA。
{"title":"MICA: Towards Explainable Skin Lesion Diagnosis via Multi-Level Image-Concept Alignment","authors":"Yequan Bie, Luyang Luo, Hao Chen","doi":"10.48550/arXiv.2401.08527","DOIUrl":"https://doi.org/10.48550/arXiv.2401.08527","url":null,"abstract":"Black-box deep learning approaches have showcased significant potential in the realm of medical image analysis. However, the stringent trustworthiness requirements intrinsic to the medical field have catalyzed research into the utilization of Explainable Artificial Intelligence (XAI), with a particular focus on concept-based methods. Existing concept-based methods predominantly apply concept annotations from a single perspective (e.g., global level), neglecting the nuanced semantic relationships between sub-regions and concepts embedded within medical images. This leads to underutilization of the valuable medical information and may cause models to fall short in harmoniously balancing interpretability and performance when employing inherently interpretable architectures such as Concept Bottlenecks. To mitigate these shortcomings, we propose a multi-modal explainable disease diagnosis framework that meticulously aligns medical images and clinical-related concepts semantically at multiple strata, encompassing the image level, token level, and concept level. Moreover, our method allows for model intervention and offers both textual and visual explanations in terms of human-interpretable concepts. Experimental results on three skin image datasets demonstrate that our method, while preserving model interpretability, attains high performance and label efficiency for concept detection and disease diagnosis. The code is available at https://github.com/Tommy-Bie/MICA.","PeriodicalId":518480,"journal":{"name":"AAAI Conference on Artificial Intelligence","volume":"52 9","pages":"837-845"},"PeriodicalIF":0.0,"publicationDate":"2024-01-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140531532","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
期刊
AAAI Conference on Artificial Intelligence
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1