首页 > 最新文献

International Joint Conference on Artificial Intelligence最新文献

英文 中文
Character As Pixels: A Controllable Prompt Adversarial Attacking Framework for Black-Box Text Guided Image Generation Models 字符作为像素:黑箱文本引导图像生成模型的可控提示对抗攻击框架
Pub Date : 2023-08-01 DOI: 10.24963/ijcai.2023/109
Ziyi Kou, Shichao Pei, Yijun Tian, Xiangliang Zhang
In this paper, we study a controllable prompt adversarial attacking problem for text guided image generation (Text2Image) models in the black-box scenario, where the goal is to attack specific visual subjects (e.g., changing a brown dog to white) in a generated image by slightly, if not imperceptibly, perturbing the characters of the driven prompt (e.g., ``brown'' to ``br0wn''). Our study is motivated by the limitations of current Text2Image attacking approaches that still rely on manual trials to create adversarial prompts. To address such limitations, we develop CharGrad, a character-level gradient based attacking framework that replaces specific characters of a prompt with pixel-level similar ones by interactively learning the perturbation direction for the prompt and updating the attacking examiner for the generated image based on a novel proxy perturbation representation for characters. We evaluate CharGrad using the texts from two public image captioning datasets. Results demonstrate that CharGrad outperforms existing text adversarial attacking approaches on attacking various subjects of generated images by black-box Text2Image models in a more effective and efficient way with less perturbation on the characters of the prompts.
在本文中,我们研究了黑盒场景下文本引导图像生成(Text2Image)模型的可控提示对抗性攻击问题,其目标是通过轻微(如果不是难以察觉的话)干扰驱动提示的字符(例如,“棕色”到“棕色”)来攻击生成图像中的特定视觉对象(例如,将棕色狗变为白色)。我们的研究是由当前Text2Image攻击方法的局限性所激发的,这些方法仍然依赖于手动试验来创建对抗性提示。为了解决这些限制,我们开发了CharGrad,这是一个基于字符级梯度的攻击框架,通过交互式地学习提示符的扰动方向,并用像素级相似的字符替换提示符的特定字符,并基于字符的新型代理扰动表示更新生成图像的攻击检查器。我们使用来自两个公共图像字幕数据集的文本来评估CharGrad。结果表明,CharGrad算法在对黑盒Text2Image模型生成的各种主题图像进行攻击时,对提示符字符的扰动较小,且攻击效果更佳,优于现有的文本对抗性攻击方法。
{"title":"Character As Pixels: A Controllable Prompt Adversarial Attacking Framework for Black-Box Text Guided Image Generation Models","authors":"Ziyi Kou, Shichao Pei, Yijun Tian, Xiangliang Zhang","doi":"10.24963/ijcai.2023/109","DOIUrl":"https://doi.org/10.24963/ijcai.2023/109","url":null,"abstract":"In this paper, we study a controllable prompt adversarial attacking problem for text guided image generation (Text2Image) models in the black-box scenario, where the goal is to attack specific visual subjects (e.g., changing a brown dog to white) in a generated image by slightly, if not imperceptibly, perturbing the characters of the driven prompt (e.g., ``brown'' to ``br0wn''). Our study is motivated by the limitations of current Text2Image attacking approaches that still rely on manual trials to create adversarial prompts. To address such limitations, we develop CharGrad, a character-level gradient based attacking framework that replaces specific characters of a prompt with pixel-level similar ones by interactively learning the perturbation direction for the prompt and updating the attacking examiner for the generated image based on a novel proxy perturbation representation for characters. We evaluate CharGrad using the texts from two public image captioning datasets. Results demonstrate that CharGrad outperforms existing text adversarial attacking approaches on attacking various subjects of generated images by black-box Text2Image models in a more effective and efficient way with less perturbation on the characters of the prompts.","PeriodicalId":394530,"journal":{"name":"International Joint Conference on Artificial Intelligence","volume":"317 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115947210","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Counting and Sampling Models in First-Order Logic 一阶逻辑中的计数和抽样模型
Pub Date : 2023-08-01 DOI: 10.24963/ijcai.2023/801
Ondřej Kuželka
First-order model counting (FOMC) is the task of counting models of a first-order logic sentence over a given set of domain elements. Its weighted variant, WFOMC, generalizes FOMC by assigning weights to the models and has many applications in statistical relational learning. More than ten years of research by various authors has led to identification of non-trivial classes of WFOMC problems that can be solved in time polynomial in the number of domain elements. In this paper, we describe recent works on WFOMC and the related problem of weighted first-order model sampling (WFOMS). We also discuss possible applications of WFOMC and WFOMS within statistical relational learning and beyond, e.g., automated solving of problems from enumerative combinatorics and elementary probability theory. Finally, we mention research problems that still need to be tackled in order to make applications of these methods really practical more broadly.
一阶模型计数(FOMC)是对给定域元素集合上的一阶逻辑句子进行模型计数的任务。它的加权变体WFOMC通过为模型分配权重来推广FOMC,并且在统计关系学习中有许多应用。经过十多年的研究,许多作者已经发现了一些非平凡类的WFOMC问题,这些问题可以用域元素数量的时间多项式来解决。本文介绍了WFOMC和加权一阶模型抽样(WFOMS)相关问题的最新研究成果。我们还讨论了WFOMC和WFOMS在统计关系学习及其他领域的可能应用,例如,从枚举组合学和初等概率论中自动解决问题。最后,我们提到了仍然需要解决的研究问题,以便使这些方法的应用真正具有更广泛的实用性。
{"title":"Counting and Sampling Models in First-Order Logic","authors":"Ondřej Kuželka","doi":"10.24963/ijcai.2023/801","DOIUrl":"https://doi.org/10.24963/ijcai.2023/801","url":null,"abstract":"First-order model counting (FOMC) is the task of counting models of a first-order logic sentence over a given set of domain elements. Its weighted variant, WFOMC, generalizes FOMC by assigning weights to the models and has many applications in statistical relational learning. More than ten years of research by various authors has led to identification of non-trivial classes of WFOMC problems that can be solved in time polynomial in the number of domain elements. In this paper, we describe recent works on WFOMC and the related problem of weighted first-order model sampling (WFOMS). We also discuss possible applications of WFOMC and WFOMS within statistical relational learning and beyond, e.g., automated solving of problems from enumerative combinatorics and elementary probability theory. Finally, we mention research problems that still need to be tackled in order to make applications of these methods really practical more broadly.","PeriodicalId":394530,"journal":{"name":"International Joint Conference on Artificial Intelligence","volume":"18 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121026596","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Finding an ϵ-Close Minimal Variation of Parameters in Bayesian Networks 在贝叶斯网络中寻找ϵ-Close最小参数变化
Pub Date : 2023-08-01 DOI: 10.24963/ijcai.2023/635
Bahar Salmani, J. Katoen
This paper addresses the ε-close parameter tuning problem for Bayesiannetworks (BNs): find a minimal ε-close amendment of probability entriesin a given set of (rows in) conditional probability tables that make agiven quantitative constraint on the BN valid. Based on thestate-of-the-art “region verification” techniques for parametric Markovchains, we propose an algorithm whose capabilities gobeyond any existing techniques. Our experiments show that ε-close tuningof large BN benchmarks with up to eight parameters is feasible. Inparticular, by allowing (i) varied parameters in multiple CPTs and (ii)inter-CPT parameter dependencies, we treat subclasses of parametric BNsthat have received scant attention so far.
本文解决了贝叶斯网络(BN)的ε-close参数调整问题:在给定的一组条件概率表中找到一个最小的ε-close修正,使给定的定量约束对BN有效。基于参数马尔可夫链最先进的“区域验证”技术,我们提出了一种超越任何现有技术的算法。我们的实验表明,具有多达8个参数的大型BN基准的ε-接近调谐是可行的。特别是,通过允许(i)多个cpt中的不同参数和(ii) cpt间参数依赖,我们处理了迄今为止很少受到关注的参数bns的子类。
{"title":"Finding an ϵ-Close Minimal Variation of Parameters in Bayesian Networks","authors":"Bahar Salmani, J. Katoen","doi":"10.24963/ijcai.2023/635","DOIUrl":"https://doi.org/10.24963/ijcai.2023/635","url":null,"abstract":"This paper addresses the ε-close parameter tuning problem for Bayesian\u0000\u0000networks (BNs): find a minimal ε-close amendment of probability entries\u0000\u0000in a given set of (rows in) conditional probability tables that make a\u0000\u0000given quantitative constraint on the BN valid. Based on the\u0000\u0000state-of-the-art “region verification” techniques for parametric Markov\u0000\u0000chains, we propose an algorithm whose capabilities go\u0000\u0000beyond any existing techniques. Our experiments show that ε-close tuning\u0000\u0000of large BN benchmarks with up to eight parameters is feasible. In\u0000\u0000particular, by allowing (i) varied parameters in multiple CPTs and (ii)\u0000\u0000inter-CPT parameter dependencies, we treat subclasses of parametric BNs\u0000\u0000that have received scant attention so far.","PeriodicalId":394530,"journal":{"name":"International Joint Conference on Artificial Intelligence","volume":"37 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124939924","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
What Lies beyond the Pareto Front? A Survey on Decision-Support Methods for Multi-Objective Optimization 帕累托前线之外是什么?多目标优化决策支持方法综述
Pub Date : 2023-08-01 DOI: 10.24963/ijcai.2023/755
Zuzanna Osika, J. Z. Salazar, Diederik M. Roijers, F. Oliehoek, P. Murukannaiah
We present a review that unifies decision-support methods for exploring the solutions produced by multi-objective optimization (MOO) algorithms. As MOO is applied to solve diverse problems, approaches for analyzing the trade-offs offered by these algorithms are scattered across fields. We provide an overview of the current advances on this topic, including methods for visualization, mining the solution set, and uncertainty exploration as well as emerging research directions, including interactivity, explainability, and support on ethical aspects. We synthesize these methods drawing from different fields of research to enable building a unified approach, independent of the application. Our goals are to reduce the entry barrier for researchers and practitioners on using MOO algorithms and to provide novel research directions.
我们提出了一种统一的决策支持方法来探索多目标优化(MOO)算法产生的解决方案。由于MOO被应用于解决各种各样的问题,分析这些算法所提供的权衡的方法分散在各个领域。我们概述了该主题的当前进展,包括可视化方法,挖掘解决方案集,不确定性探索以及新兴的研究方向,包括交互性,可解释性和对伦理方面的支持。我们综合了这些来自不同研究领域的方法,以建立一个独立于应用程序的统一方法。我们的目标是减少研究人员和实践者使用mooo算法的进入壁垒,并提供新的研究方向。
{"title":"What Lies beyond the Pareto Front? A Survey on Decision-Support Methods for Multi-Objective Optimization","authors":"Zuzanna Osika, J. Z. Salazar, Diederik M. Roijers, F. Oliehoek, P. Murukannaiah","doi":"10.24963/ijcai.2023/755","DOIUrl":"https://doi.org/10.24963/ijcai.2023/755","url":null,"abstract":"We present a review that unifies decision-support methods for exploring the solutions produced by multi-objective optimization (MOO) algorithms. As MOO is applied to solve diverse problems, approaches for analyzing the trade-offs offered by these algorithms are scattered across fields. We provide an overview of the current advances on this topic, including methods for visualization, mining the solution set, and uncertainty exploration as well as emerging research directions, including interactivity, explainability, and support on ethical aspects. We synthesize these methods drawing from different fields of research to enable building a unified approach, independent of the application. Our goals are to reduce the entry barrier for researchers and practitioners on using MOO algorithms and to provide novel research directions.","PeriodicalId":394530,"journal":{"name":"International Joint Conference on Artificial Intelligence","volume":"24 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125119149","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Contrastive Learning and Reward Smoothing for Deep Portfolio Management 深度投资组合管理的对比学习与奖励平滑
Pub Date : 2023-08-01 DOI: 10.24963/ijcai.2023/441
Yun-Hsuan Lien, Yuan-kui Li, Yu-Shuen Wang
In this study, we used reinforcement learning (RL) models to invest assets in order to earn returns. The models were trained to interact with a simulated environment based on historical market data and learn trading strategies. However, using deep neural networks based on the returns of each period can be challenging due to the unpredictability of financial markets. As a result, the policies learned from training data may not be effective when tested in real-world situations. To address this issue, we incorporated contrastive learning and reward smoothing into our training process. Contrastive learning allows the RL models to recognize patterns in asset states that may indicate future price movements. Reward smoothing, on the other hand, serves as a regularization technique to prevent the models from seeking immediate but uncertain profits. We tested our method against various traditional financial techniques and other deep RL methods, and found it to be effective in both the U.S. stock market and the cryptocurrency market. Our source code is available at https://github.com/sophialien/FinTech-DPM.
在本研究中,我们使用强化学习(RL)模型来投资资产以获得回报。这些模型经过训练,可以与基于历史市场数据的模拟环境进行交互,并学习交易策略。然而,由于金融市场的不可预测性,使用基于每个时期回报的深度神经网络可能具有挑战性。因此,从训练数据中学到的策略在实际情况中进行测试时可能并不有效。为了解决这个问题,我们将对比学习和奖励平滑融入到我们的培训过程中。对比学习允许强化学习模型识别资产状态中的模式,这些模式可能表明未来的价格走势。另一方面,奖励平滑作为一种正则化技术,防止模型寻求即时但不确定的利润。我们将我们的方法与各种传统金融技术和其他深度强化学习方法进行了测试,发现它在美国股市和加密货币市场都是有效的。我们的源代码可从https://github.com/sophialien/FinTech-DPM获得。
{"title":"Contrastive Learning and Reward Smoothing for Deep Portfolio Management","authors":"Yun-Hsuan Lien, Yuan-kui Li, Yu-Shuen Wang","doi":"10.24963/ijcai.2023/441","DOIUrl":"https://doi.org/10.24963/ijcai.2023/441","url":null,"abstract":"In this study, we used reinforcement learning (RL) models to invest assets in order to earn returns. The models were trained to interact with a simulated environment based on historical market data and learn trading strategies. However, using deep neural networks based on the returns of each period can be challenging due to the unpredictability of financial markets. As a result, the policies learned from training data may not be effective when tested in real-world situations. To address this issue, we incorporated contrastive learning and reward smoothing into our training process. Contrastive learning allows the RL models to recognize patterns in asset states that may indicate future price movements. Reward smoothing, on the other hand, serves as a regularization technique to prevent the models from seeking immediate but uncertain profits. We tested our method against various traditional financial techniques and other deep RL methods, and found it to be effective in both the U.S. stock market and the cryptocurrency market. Our source code is available at https://github.com/sophialien/FinTech-DPM.","PeriodicalId":394530,"journal":{"name":"International Joint Conference on Artificial Intelligence","volume":"391 5","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114008326","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A New ANN-SNN Conversion Method with High Accuracy, Low Latency and Good Robustness 一种高精度、低时延、鲁棒性好的ANN-SNN转换新方法
Pub Date : 2023-08-01 DOI: 10.24963/ijcai.2023/342
Bingsen Wang, Jian Cao, Jue Chen, Shuo Feng, Yuan Wang
Due to the advantages of low energy consumption, high robustness and fast inference speed, Spiking Neural Networks (SNNs), with good biological interpretability and the potential to be applied on neuromorphic hardware, are regarded as the third generation of Artificial Neural Networks (ANNs). Despite having so many advantages, the biggest challenge encountered by spiking neural networks is training difficulty caused by the non-differentiability of spike signals. ANN-SNN conversion is an effective method that solves the training difficulty by converting parameters in ANNs to those in SNNs through a specific algorithm. However, the ANN-SNN conversion method also suffers from accuracy degradation and long inference time. In this paper, we reanalyzed the relationship between Integrate-and-Fire (IF) neuron model and ReLU activation function, proposed a StepReLU activation function more suitable for SNNs under membrane potential encoding, and used it to train ANNs. Then we converted the ANNs to SNNs with extremely small conversion error and introduced leakage mechanism to the SNNs and get the final models, which have high accuracy, low latency and good robustness, and have achieved the state-of-the-art performance on various datasets such as CIFAR and ImageNet.
脉冲神经网络(SNNs)由于具有低能耗、高鲁棒性和快速推理速度等优点,具有良好的生物可解释性和在神经形态硬件上的应用潜力,被认为是第三代人工神经网络(ann)。尽管有很多优点,但脉冲神经网络面临的最大挑战是由于脉冲信号的不可微性而导致的训练困难。ANN-SNN转换是一种解决训练困难的有效方法,通过特定的算法将ann网络中的参数转换为snn中的参数。然而,ANN-SNN转换方法也存在精度下降和推理时间长的问题。本文重新分析了IF (Integrate-and-Fire)神经元模型与ReLU激活函数之间的关系,提出了膜电位编码下更适合snn的StepReLU激活函数,并将其用于人工神经网络的训练。然后以极小的转换误差将人工神经网络转换为snn,并在snn中引入泄漏机制,得到精度高、延迟低、鲁棒性好的最终模型,并在CIFAR和ImageNet等各种数据集上取得了最先进的性能。
{"title":"A New ANN-SNN Conversion Method with High Accuracy, Low Latency and Good Robustness","authors":"Bingsen Wang, Jian Cao, Jue Chen, Shuo Feng, Yuan Wang","doi":"10.24963/ijcai.2023/342","DOIUrl":"https://doi.org/10.24963/ijcai.2023/342","url":null,"abstract":"Due to the advantages of low energy consumption, high robustness and fast inference speed, Spiking Neural Networks (SNNs), with good biological interpretability and the potential to be applied on neuromorphic hardware, are regarded as the third generation of Artificial Neural Networks (ANNs). Despite having so many advantages, the biggest challenge encountered by spiking neural networks is training difficulty caused by the non-differentiability of spike signals. ANN-SNN conversion is an effective method that solves the training difficulty by converting parameters in ANNs to those in SNNs through a specific algorithm. However, the ANN-SNN conversion method also suffers from accuracy degradation and long inference time. In this paper, we reanalyzed the relationship between Integrate-and-Fire (IF) neuron model and ReLU activation function, proposed a StepReLU activation function more suitable for SNNs under membrane potential encoding, and used it to train ANNs. Then we converted the ANNs to SNNs with extremely small conversion error and introduced leakage mechanism to the SNNs and get the final models, which have high accuracy, low latency and good robustness, and have achieved the state-of-the-art performance on various datasets such as CIFAR and ImageNet.","PeriodicalId":394530,"journal":{"name":"International Joint Conference on Artificial Intelligence","volume":"195 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122521538","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Generalization Bounds for Adversarial Metric Learning 对抗性度量学习的泛化界限
Pub Date : 2023-08-01 DOI: 10.24963/ijcai.2023/489
Wen Wen, Han Li, H. Chen, Rui Wu, Lingjuan Wu, Liangxuan Zhu
Recently, adversarial metric learning has been proposed to enhance the robustness of the learned distance metric against adversarial perturbations. Despite rapid progress in validating its effectiveness empirically, theoretical guarantees on adversarial robustness and generalization are far less understood. To fill this gap, this paper focuses on unveiling the generalization properties of adversarial metric learning by developing the uniform convergence analysis techniques. Based on the capacity estimation of covering numbers, we establish the first high-probability generalization bounds with order O(n^{-1/2}) for adversarial metric learning with pairwise perturbations and general losses, where n is the number of training samples. Moreover, we obtain the refined generalization bounds with order O(n^{-1}) for the smooth loss by using local Rademacher complexity, which is faster than the previous result of adversarial pairwise learning, e.g., adversarial bipartite ranking. Experimental evaluation on real-world datasets validates our theoretical findings.
最近,人们提出了对抗度量学习来增强学习到的距离度量对对抗扰动的鲁棒性。尽管在实证验证其有效性方面进展迅速,但对对抗鲁棒性和泛化的理论保证却知之甚少。为了填补这一空白,本文着重于通过发展一致收敛分析技术来揭示对抗性度量学习的泛化性质。基于覆盖数的容量估计,我们建立了具有两两扰动和一般损失的对抗性度量学习的第一个O(n^{-1/2})阶的高概率泛化界,其中n为训练样本的数量。此外,我们利用局部Rademacher复杂度得到了光滑损失的O(n^{-1})阶精细泛化界,这比之前的对抗性两两学习(例如对抗性二部排序)的结果更快。在真实世界数据集上的实验评估验证了我们的理论发现。
{"title":"Generalization Bounds for Adversarial Metric Learning","authors":"Wen Wen, Han Li, H. Chen, Rui Wu, Lingjuan Wu, Liangxuan Zhu","doi":"10.24963/ijcai.2023/489","DOIUrl":"https://doi.org/10.24963/ijcai.2023/489","url":null,"abstract":"Recently, adversarial metric learning has been proposed to enhance the robustness of the learned distance metric against adversarial perturbations. Despite rapid progress in validating its effectiveness empirically, theoretical guarantees on adversarial robustness and generalization are far less understood. To fill this gap, this paper focuses on unveiling the generalization properties of adversarial metric learning by developing the uniform convergence analysis techniques. Based on the capacity estimation of covering numbers, we establish the first high-probability generalization bounds with order O(n^{-1/2}) for adversarial metric learning with pairwise perturbations and general losses, where n is the number of training samples. Moreover, we obtain the refined generalization bounds with order O(n^{-1}) for the smooth loss by using local Rademacher complexity, which is faster than the previous result of adversarial pairwise learning, e.g., adversarial bipartite ranking. Experimental evaluation on real-world datasets validates our theoretical findings.","PeriodicalId":394530,"journal":{"name":"International Joint Conference on Artificial Intelligence","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122716666","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Learning to Binarize Continuous Features for Neuro-Rule Networks 学习神经规则网络的连续特征二值化
Pub Date : 2023-08-01 DOI: 10.24963/ijcai.2023/510
Wei Zhang, Y. Liu, Zhuo Wang, Jianyong Wang
Neuro-Rule Networks (NRNs) emerge as a promising neuro-symbolic method, enjoyed by the ability to equate fully-connected neural networks with logic rules. To support learning logic rules consisting of boolean variables, converting input features into binary representations is required. Different from discrete features that could be directly transformed by one-hot encodings, continuous features need to be binarized based on some numerical intervals. Existing studies usually select the bound values of intervals based on empirical strategies (e.g., equal-width interval). However, it is not optimal since the bounds are fixed and cannot be optimized to accommodate the ultimate training target. In this paper, we propose AutoInt, an approach that automatically binarizes continuous features and enables the intervals to be optimized with NRNs in an end-to-end fashion. Specifically, AutoInt automatically selects an interval for a given continuous feature in a soft manner to enable a differentiable learning procedure of interval-related parameters. Moreover, it introduces an additional soft K-means clustering loss to make the interval centres approach the original feature value distribution, thus reducing the risk of overfitting intervals. We conduct comprehensive experiments on public datasets and demonstrate the effectiveness of AutoInt in boosting the performance of NRNs.
神经规则网络(NRNs)作为一种很有前途的神经符号方法出现,因为它能够将完全连接的神经网络与逻辑规则等同起来。为了支持学习由布尔变量组成的逻辑规则,需要将输入特征转换为二进制表示。与单热编码可以直接变换的离散特征不同,连续特征需要基于一定的数值区间进行二值化。现有研究通常基于经验策略(如等宽区间)选择区间的界值。然而,它不是最优的,因为边界是固定的,不能优化以适应最终的训练目标。在本文中,我们提出了AutoInt,一种自动二值化连续特征并使区间能够以端到端方式使用nrn进行优化的方法。具体来说,AutoInt以软方式自动选择给定连续特征的区间,从而实现区间相关参数的可微分学习过程。此外,它还引入了一个额外的软k均值聚类损失,使区间中心接近原始特征值分布,从而降低了区间过拟合的风险。我们在公共数据集上进行了全面的实验,证明了AutoInt在提高自然神经网络性能方面的有效性。
{"title":"Learning to Binarize Continuous Features for Neuro-Rule Networks","authors":"Wei Zhang, Y. Liu, Zhuo Wang, Jianyong Wang","doi":"10.24963/ijcai.2023/510","DOIUrl":"https://doi.org/10.24963/ijcai.2023/510","url":null,"abstract":"Neuro-Rule Networks (NRNs) emerge as a promising neuro-symbolic method, enjoyed by the ability to equate fully-connected neural networks with logic rules. To support learning logic rules consisting of boolean variables, converting input features into binary representations is required. Different from discrete features that could be directly transformed by one-hot encodings, continuous features need to be binarized based on some numerical intervals. Existing studies usually select the bound values of intervals based on empirical strategies (e.g., equal-width interval). However, it is not optimal since the bounds are fixed and cannot be optimized to accommodate the ultimate training target. In this paper, we propose AutoInt, an approach that automatically binarizes continuous features and enables the intervals to be optimized with NRNs in an end-to-end fashion. Specifically, AutoInt automatically selects an interval for a given continuous feature in a soft manner to enable a differentiable learning procedure of interval-related parameters. Moreover, it introduces an additional soft K-means clustering loss to make the interval centres approach the original feature value distribution, thus reducing the risk of overfitting intervals. We conduct comprehensive experiments on public datasets and demonstrate the effectiveness of AutoInt in boosting the performance of NRNs.","PeriodicalId":394530,"journal":{"name":"International Joint Conference on Artificial Intelligence","volume":"40 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122723225","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Vision Language Navigation with Knowledge-driven Environmental Dreamer 视觉语言导航与知识驱动的环境梦想家
Pub Date : 2023-08-01 DOI: 10.24963/ijcai.2023/204
Fengda Zhu, Vincent CS Lee, Xiaojun Chang, Xiaodan Liang
Vision-language navigation (VLN) requires an agent to perceive visual observation in a house scene and navigate step-by-step following natural language instruction. Due to the high cost of data annotation and data collection, current VLN datasets provide limited instruction-trajectory data samples. Learning vision-language alignment for VLN from limited data is challenging since visual observation and language instruction are both complex and diverse. Previous works only generate augmented data based on original scenes while failing to generate data samples from unseen scenes, which limits the generalization ability of the navigation agent. In this paper, we introduce the Knowledge-driven Environmental Dreamer (KED), a method that leverages the knowledge of the embodied environment and generates unseen scenes for a navigation agent to learn. Generating an unseen environment with texture consistency and structure consistency is challenging. To address this problem, we incorporate three knowledge-driven regularization objectives into the KED and adopt a reweighting mechanism for self-adaptive optimization. Our KED method is able to generate unseen embodied environments without extra annotations. We use KED to successfully generate 270 houses and 500K instruction-trajectory pairs. The navigation agent with the KED method outperforms the state-of-the-art methods on various VLN benchmarks, such as R2R, R4R, and RxR. Both qualitative and quantitative experiments prove that our proposed KED method is able to high-quality augmentation data with texture consistency and structure consistency.
视觉语言导航(VLN)要求智能体感知房屋场景中的视觉观察,并按照自然语言指令逐步导航。由于数据标注和数据收集成本高,当前VLN数据集提供的指令轨迹数据样本有限。由于视觉观察和语言教学既复杂又多样,因此从有限的数据中学习VLN的视觉语言对齐是具有挑战性的。以往的研究只能在原始场景的基础上生成增强数据,而不能从未见过的场景中生成数据样本,这限制了导航agent的泛化能力。在本文中,我们介绍了知识驱动的环境梦想者(knowledge -driven Environmental Dreamer,简称KED),这是一种利用具体环境的知识生成未知场景供导航智能体学习的方法。生成具有纹理一致性和结构一致性的看不见的环境是具有挑战性的。为了解决这一问题,我们将三个知识驱动的正则化目标纳入到KED中,并采用了自适应优化的重加权机制。我们的KED方法能够在没有额外注释的情况下生成看不见的具体化环境。我们使用KED成功地生成了270个房屋和500K个指令轨迹对。采用KED方法的导航代理在各种VLN基准测试(如R2R、R4R和RxR)上的性能优于最先进的方法。定性和定量实验都证明了我们提出的KED方法能够实现纹理一致性和结构一致性的高质量增强数据。
{"title":"Vision Language Navigation with Knowledge-driven Environmental Dreamer","authors":"Fengda Zhu, Vincent CS Lee, Xiaojun Chang, Xiaodan Liang","doi":"10.24963/ijcai.2023/204","DOIUrl":"https://doi.org/10.24963/ijcai.2023/204","url":null,"abstract":"Vision-language navigation (VLN) requires an agent to perceive visual observation in a house scene and navigate step-by-step following natural language instruction. Due to the high cost of data annotation and data collection, current VLN datasets provide limited instruction-trajectory data samples. Learning vision-language alignment for VLN from limited data is challenging since visual observation and language instruction are both complex and diverse. Previous works only generate augmented data based on original scenes while failing to generate data samples from unseen scenes, which limits the generalization ability of the navigation agent. In this paper, we introduce the Knowledge-driven Environmental Dreamer (KED), a method that leverages the knowledge of the embodied environment and generates unseen scenes for a navigation agent to learn. Generating an unseen environment with texture consistency and structure consistency is challenging. To address this problem, we incorporate three knowledge-driven regularization objectives into the KED and adopt a reweighting mechanism for self-adaptive optimization. Our KED method is able to generate unseen embodied environments without extra annotations. We use KED to successfully generate 270 houses and 500K instruction-trajectory pairs. The navigation agent with the KED method outperforms the state-of-the-art methods on various VLN benchmarks, such as R2R, R4R, and RxR. Both qualitative and quantitative experiments prove that our proposed KED method is able to high-quality augmentation data with texture consistency and structure consistency.","PeriodicalId":394530,"journal":{"name":"International Joint Conference on Artificial Intelligence","volume":"65 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122932507","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Measuring a Priori Voting Power in Liquid Democracy 衡量流动民主中的先验投票权
Pub Date : 2023-08-01 DOI: 10.24963/ijcai.2023/290
Rachael Colley, Théo Delemazure, Hugo Gilbert
We introduce new power indices to measure the a priori voting power of voters in liquid democracy elections where an underlying network restricts delegations. We argue that our power indices are natural extensions of the standard Penrose-Banzhaf index in simple voting games.We show that computing the criticality of a voter is #P-hard even in weighted games with weights polynomially-bounded in the size of the instance.However, for specific settings, such as when the underlying network is a bipartite or complete graph, recursive formulas can compute these indices for weighted voting games in pseudo-polynomial time.We highlight their theoretical properties and provide numerical results to illustrate how restricting the possible delegations can alter voters' voting power.
我们引入了新的权力指数来衡量选民在流动民主选举中的先验投票权,其中潜在的网络限制了代表团。我们认为我们的权力指数是简单投票游戏中标准Penrose-Banzhaf指数的自然延伸。我们证明,即使在权重以实例的大小为多项式界的加权博弈中,计算选民的临界性也是# p -困难的。然而,对于特定的设置,例如当底层网络是二部图或完全图时,递归公式可以在伪多项式时间内计算加权投票游戏的这些指标。我们强调了它们的理论性质,并提供了数值结果来说明限制可能的代表团如何改变选民的投票权。
{"title":"Measuring a Priori Voting Power in Liquid Democracy","authors":"Rachael Colley, Théo Delemazure, Hugo Gilbert","doi":"10.24963/ijcai.2023/290","DOIUrl":"https://doi.org/10.24963/ijcai.2023/290","url":null,"abstract":"We introduce new power indices to measure the a priori voting power of voters in liquid democracy elections where an underlying network restricts delegations. We argue that our power indices are natural extensions of the standard Penrose-Banzhaf index in simple voting games.\u0000\u0000We show that computing the criticality of a voter is #P-hard even in weighted games with weights polynomially-bounded in the size of the instance.\u0000\u0000However, for specific settings, such as when the underlying network is a bipartite or complete graph, recursive formulas can compute these indices for weighted voting games in pseudo-polynomial time.\u0000\u0000We highlight their theoretical properties and provide numerical results to illustrate how restricting the possible delegations can alter voters' voting power.","PeriodicalId":394530,"journal":{"name":"International Joint Conference on Artificial Intelligence","volume":"142 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131743570","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
International Joint Conference on Artificial Intelligence
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1