首页 > 最新文献

Proceedings of the ... International Conference on Machine Learning. International Conference on Machine Learning最新文献

英文 中文
Near-optimal Conservative Exploration in Reinforcement Learning under Episode-wise Constraints 情景约束下强化学习的近最优保守探索
Donghao Li, Ruiquan Huang, Cong Shen, Jing Yang
This paper investigates conservative exploration in reinforcement learning where the performance of the learning agent is guaranteed to be above a certain threshold throughout the learning process. It focuses on the tabular episodic Markov Decision Process (MDP) setting that has finite states and actions. With the knowledge of an existing safe baseline policy, an algorithm termed as StepMix is proposed to balance the exploitation and exploration while ensuring that the conservative constraint is never violated in each episode with high probability. StepMix features a unique design of a mixture policy that adaptively and smoothly interpolates between the baseline policy and the optimistic policy. Theoretical analysis shows that StepMix achieves near-optimal regret order as in the constraint-free setting, indicating that obeying the stringent episode-wise conservative constraint does not compromise the learning performance. Besides, a randomization-based EpsMix algorithm is also proposed and shown to achieve the same performance as StepMix. The algorithm design and theoretical analysis are further extended to the setting where the baseline policy is not given a priori but must be learned from an offline dataset, and it is proved that similar conservative guarantee and regret can be achieved if the offline dataset is sufficiently large. Experiment results corroborate the theoretical analysis and demonstrate the effectiveness of the proposed conservative exploration strategies.
本文研究了强化学习中的保守探索,在强化学习中,学习代理的性能在整个学习过程中保证在一定的阈值以上。它侧重于具有有限状态和动作的表格情景马尔可夫决策过程(MDP)设置。在了解现有的安全基线策略的基础上,提出了一种称为StepMix的算法来平衡开发和探索,同时确保在每个大概率事件中都不会违反保守约束。StepMix具有独特的混合策略设计,可以自适应地平滑地在基线策略和乐观策略之间进行插值。理论分析表明,StepMix在无约束条件下达到了近似最优的后悔顺序,表明服从严格的情节保守约束并不影响学习性能。此外,本文还提出了一种基于随机化的EpsMix算法,该算法与StepMix算法的性能相当。将算法设计和理论分析进一步扩展到基线策略没有先验而必须从离线数据集中学习的情况,并证明了当离线数据集足够大时,可以实现类似的保守保证和遗憾。实验结果证实了理论分析,并证明了所提出的保守勘探策略的有效性。
{"title":"Near-optimal Conservative Exploration in Reinforcement Learning under Episode-wise Constraints","authors":"Donghao Li, Ruiquan Huang, Cong Shen, Jing Yang","doi":"10.48550/arXiv.2306.06265","DOIUrl":"https://doi.org/10.48550/arXiv.2306.06265","url":null,"abstract":"This paper investigates conservative exploration in reinforcement learning where the performance of the learning agent is guaranteed to be above a certain threshold throughout the learning process. It focuses on the tabular episodic Markov Decision Process (MDP) setting that has finite states and actions. With the knowledge of an existing safe baseline policy, an algorithm termed as StepMix is proposed to balance the exploitation and exploration while ensuring that the conservative constraint is never violated in each episode with high probability. StepMix features a unique design of a mixture policy that adaptively and smoothly interpolates between the baseline policy and the optimistic policy. Theoretical analysis shows that StepMix achieves near-optimal regret order as in the constraint-free setting, indicating that obeying the stringent episode-wise conservative constraint does not compromise the learning performance. Besides, a randomization-based EpsMix algorithm is also proposed and shown to achieve the same performance as StepMix. The algorithm design and theoretical analysis are further extended to the setting where the baseline policy is not given a priori but must be learned from an offline dataset, and it is proved that similar conservative guarantee and regret can be achieved if the offline dataset is sufficiently large. Experiment results corroborate the theoretical analysis and demonstrate the effectiveness of the proposed conservative exploration strategies.","PeriodicalId":74529,"journal":{"name":"Proceedings of the ... International Conference on Machine Learning. International Conference on Machine Learning","volume":"27 1","pages":"19527-19564"},"PeriodicalIF":0.0,"publicationDate":"2023-06-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"83764283","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Self-Interpretable Time Series Prediction with Counterfactual Explanations 具有反事实解释的自解释时间序列预测
Jingquan Yan, Hao Wang
Interpretable time series prediction is crucial for safety-critical areas such as healthcare and autonomous driving. Most existing methods focus on interpreting predictions by assigning important scores to segments of time series. In this paper, we take a different and more challenging route and aim at developing a self-interpretable model, dubbed Counterfactual Time Series (CounTS), which generates counterfactual and actionable explanations for time series predictions. Specifically, we formalize the problem of time series counterfactual explanations, establish associated evaluation protocols, and propose a variational Bayesian deep learning model equipped with counterfactual inference capability of time series abduction, action, and prediction. Compared with state-of-the-art baselines, our self-interpretable model can generate better counterfactual explanations while maintaining comparable prediction accuracy.
可解释的时间序列预测对于医疗保健和自动驾驶等安全关键领域至关重要。大多数现有的方法侧重于通过为时间序列的片段分配重要分数来解释预测。在本文中,我们采取了一种不同的更具挑战性的路线,旨在开发一种自我解释的模型,称为反事实时间序列(计数),它为时间序列预测生成反事实和可操作的解释。具体而言,我们形式化了时间序列反事实解释问题,建立了相关的评估协议,并提出了一个具有时间序列溯因、行动和预测反事实推理能力的变分贝叶斯深度学习模型。与最先进的基线相比,我们的自解释模型可以产生更好的反事实解释,同时保持相当的预测准确性。
{"title":"Self-Interpretable Time Series Prediction with Counterfactual Explanations","authors":"Jingquan Yan, Hao Wang","doi":"10.48550/arXiv.2306.06024","DOIUrl":"https://doi.org/10.48550/arXiv.2306.06024","url":null,"abstract":"Interpretable time series prediction is crucial for safety-critical areas such as healthcare and autonomous driving. Most existing methods focus on interpreting predictions by assigning important scores to segments of time series. In this paper, we take a different and more challenging route and aim at developing a self-interpretable model, dubbed Counterfactual Time Series (CounTS), which generates counterfactual and actionable explanations for time series predictions. Specifically, we formalize the problem of time series counterfactual explanations, establish associated evaluation protocols, and propose a variational Bayesian deep learning model equipped with counterfactual inference capability of time series abduction, action, and prediction. Compared with state-of-the-art baselines, our self-interpretable model can generate better counterfactual explanations while maintaining comparable prediction accuracy.","PeriodicalId":74529,"journal":{"name":"Proceedings of the ... International Conference on Machine Learning. International Conference on Machine Learning","volume":"1 1","pages":"39110-39125"},"PeriodicalIF":0.0,"publicationDate":"2023-06-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"88633998","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Prefer to Classify: Improving Text Classifiers via Auxiliary Preference Learning 偏好分类:通过辅助偏好学习改进文本分类器
Jaehyung Kim, Jinwoo Shin, Dongyeop Kang
The development of largely human-annotated benchmarks has driven the success of deep neural networks in various NLP tasks. To enhance the effectiveness of existing benchmarks, collecting new additional input-output pairs is often too costly and challenging, particularly considering their marginal impact on improving the current model accuracy. Instead, additional or complementary annotations on the existing input texts in the benchmarks can be preferable as an efficient way to pay the additional human cost. In this paper, we investigate task-specific preferences between pairs of input texts as a new alternative way for such auxiliary data annotation. From 'pair-wise' comparisons with respect to the task, the auxiliary preference learning enables the model to learn an additional informative training signal that cannot be captured with 'instance-wise' task labels. To this end, we propose a novel multi-task learning framework, called prefer-to-classify (P2C), which can enjoy the cooperative effect of learning both the given classification task and the auxiliary preferences. Here, we provide three different ways to collect preference signals in practice: (a) implicitly extracting from annotation records (for free, but often unavailable), (b) collecting explicitly from crowd workers (high paid), or (c) pre-trained large language models such as GPT-3 (low paid). Given existing classification NLP benchmarks, we demonstrate that the proposed auxiliary preference learning via P2C on them is effective in improving text classifiers. Our codes are publicly available.
大量人工注释基准的发展推动了深度神经网络在各种自然语言处理任务中的成功。为了提高现有基准的有效性,收集新的额外的输入-输出对通常过于昂贵和具有挑战性,特别是考虑到它们对提高当前模型准确性的边际影响。相反,在基准测试中对现有输入文本进行额外的或补充的注释可能是一种支付额外人力成本的有效方法。在本文中,我们研究了输入文本对之间的任务特定偏好,作为这种辅助数据注释的新替代方法。从任务的“成对”比较中,辅助偏好学习使模型能够学习到一个额外的信息训练信号,这是无法用“实例”任务标签捕获的。为此,我们提出了一种新的多任务学习框架,称为偏好-分类(P2C),它可以同时学习给定的分类任务和辅助偏好。在这里,我们提供了三种不同的方法来在实践中收集偏好信号:(a)隐式地从注释记录中提取(免费,但通常不可用),(b)明确地从人群工作者中收集(高薪),或(c)预训练的大型语言模型,如GPT-3(低薪)。给定现有的分类NLP基准,我们证明了通过P2C对它们进行辅助偏好学习在改进文本分类器方面是有效的。我们的代码是公开的。
{"title":"Prefer to Classify: Improving Text Classifiers via Auxiliary Preference Learning","authors":"Jaehyung Kim, Jinwoo Shin, Dongyeop Kang","doi":"10.48550/arXiv.2306.04925","DOIUrl":"https://doi.org/10.48550/arXiv.2306.04925","url":null,"abstract":"The development of largely human-annotated benchmarks has driven the success of deep neural networks in various NLP tasks. To enhance the effectiveness of existing benchmarks, collecting new additional input-output pairs is often too costly and challenging, particularly considering their marginal impact on improving the current model accuracy. Instead, additional or complementary annotations on the existing input texts in the benchmarks can be preferable as an efficient way to pay the additional human cost. In this paper, we investigate task-specific preferences between pairs of input texts as a new alternative way for such auxiliary data annotation. From 'pair-wise' comparisons with respect to the task, the auxiliary preference learning enables the model to learn an additional informative training signal that cannot be captured with 'instance-wise' task labels. To this end, we propose a novel multi-task learning framework, called prefer-to-classify (P2C), which can enjoy the cooperative effect of learning both the given classification task and the auxiliary preferences. Here, we provide three different ways to collect preference signals in practice: (a) implicitly extracting from annotation records (for free, but often unavailable), (b) collecting explicitly from crowd workers (high paid), or (c) pre-trained large language models such as GPT-3 (low paid). Given existing classification NLP benchmarks, we demonstrate that the proposed auxiliary preference learning via P2C on them is effective in improving text classifiers. Our codes are publicly available.","PeriodicalId":74529,"journal":{"name":"Proceedings of the ... International Conference on Machine Learning. International Conference on Machine Learning","volume":"6 1","pages":"16807-16828"},"PeriodicalIF":0.0,"publicationDate":"2023-06-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"78905568","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
CoCo: A Coupled Contrastive Framework for Unsupervised Domain Adaptive Graph Classification CoCo:无监督域自适应图分类的耦合对比框架
Nan Yin, Libin Shen, Mengzhu Wang, L. Lan, Zeyu Ma, C. Chen, Xiansheng Hua, Xiao Luo
Although graph neural networks (GNNs) have achieved impressive achievements in graph classification, they often need abundant task-specific labels, which could be extensively costly to acquire. A credible solution is to explore additional labeled graphs to enhance unsupervised learning on the target domain. However, how to apply GNNs to domain adaptation remains unsolved owing to the insufficient exploration of graph topology and the significant domain discrepancy. In this paper, we propose Coupled Contrastive Graph Representation Learning (CoCo), which extracts the topological information from coupled learning branches and reduces the domain discrepancy with coupled contrastive learning. CoCo contains a graph convolutional network branch and a hierarchical graph kernel network branch, which explore graph topology in implicit and explicit manners. Besides, we incorporate coupled branches into a holistic multi-view contrastive learning framework, which not only incorporates graph representations learned from complementary views for enhanced understanding, but also encourages the similarity between cross-domain example pairs with the same semantics for domain alignment. Extensive experiments on popular datasets show that our CoCo outperforms these competing baselines in different settings generally.
尽管图神经网络(gnn)在图分类方面取得了令人瞩目的成就,但它们往往需要大量的任务特定标签,而这些标签的获取成本可能非常高。一个可靠的解决方案是探索额外的标记图来增强目标域上的无监督学习。然而,由于对图拓扑的探索不足和领域差异较大,如何将gnn应用于领域自适应仍然是一个未解决的问题。在本文中,我们提出了耦合对比图表示学习(CoCo),它从耦合学习分支中提取拓扑信息,并通过耦合对比学习减少域差异。CoCo包含一个图卷积网络分支和一个分层图核网络分支,它们以隐式和显式的方式探索图拓扑。此外,我们将耦合分支合并到一个整体的多视图对比学习框架中,该框架不仅结合了从互补视图中学习的图表示以增强理解,而且还鼓励具有相同语义的跨域示例对之间的相似性以进行域对齐。在流行数据集上进行的大量实验表明,我们的CoCo在不同设置下的表现通常优于这些竞争基线。
{"title":"CoCo: A Coupled Contrastive Framework for Unsupervised Domain Adaptive Graph Classification","authors":"Nan Yin, Libin Shen, Mengzhu Wang, L. Lan, Zeyu Ma, C. Chen, Xiansheng Hua, Xiao Luo","doi":"10.48550/arXiv.2306.04979","DOIUrl":"https://doi.org/10.48550/arXiv.2306.04979","url":null,"abstract":"Although graph neural networks (GNNs) have achieved impressive achievements in graph classification, they often need abundant task-specific labels, which could be extensively costly to acquire. A credible solution is to explore additional labeled graphs to enhance unsupervised learning on the target domain. However, how to apply GNNs to domain adaptation remains unsolved owing to the insufficient exploration of graph topology and the significant domain discrepancy. In this paper, we propose Coupled Contrastive Graph Representation Learning (CoCo), which extracts the topological information from coupled learning branches and reduces the domain discrepancy with coupled contrastive learning. CoCo contains a graph convolutional network branch and a hierarchical graph kernel network branch, which explore graph topology in implicit and explicit manners. Besides, we incorporate coupled branches into a holistic multi-view contrastive learning framework, which not only incorporates graph representations learned from complementary views for enhanced understanding, but also encourages the similarity between cross-domain example pairs with the same semantics for domain alignment. Extensive experiments on popular datasets show that our CoCo outperforms these competing baselines in different settings generally.","PeriodicalId":74529,"journal":{"name":"Proceedings of the ... International Conference on Machine Learning. International Conference on Machine Learning","volume":"14 1","pages":"40040-40053"},"PeriodicalIF":0.0,"publicationDate":"2023-06-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"73405961","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Task-specific experimental design for treatment effect estimation 治疗效果评估的特定任务实验设计
Beth D. Connolly, Kim Moore, Tobias Schwedes, Alexander Adam, Gary Willis, Ilya Feige, Christopher Frye
Understanding causality should be a core requirement of any attempt to build real impact through AI. Due to the inherent unobservability of counterfactuals, large randomised trials (RCTs) are the standard for causal inference. But large experiments are generically expensive, and randomisation carries its own costs, e.g. when suboptimal decisions are trialed. Recent work has proposed more sample-efficient alternatives to RCTs, but these are not adaptable to the downstream application for which the causal effect is sought. In this work, we develop a task-specific approach to experimental design and derive sampling strategies customised to particular downstream applications. Across a range of important tasks, real-world datasets, and sample sizes, our method outperforms other benchmarks, e.g. requiring an order-of-magnitude less data to match RCT performance on targeted marketing tasks.
理解因果关系应该是任何试图通过人工智能产生真正影响的核心要求。由于反事实的固有不可观察性,大型随机试验(rct)是因果推理的标准。但大型实验通常是昂贵的,随机化也有它自己的成本,例如,在试验次优决策时。最近的工作提出了更多的样本效率替代随机对照试验,但这些不适应下游应用的因果关系是寻求。在这项工作中,我们开发了一种特定任务的实验设计方法,并推导出针对特定下游应用定制的采样策略。在一系列重要任务、真实世界数据集和样本量方面,我们的方法优于其他基准测试,例如,在目标营销任务上,我们需要的数据比RCT少一个数量级。
{"title":"Task-specific experimental design for treatment effect estimation","authors":"Beth D. Connolly, Kim Moore, Tobias Schwedes, Alexander Adam, Gary Willis, Ilya Feige, Christopher Frye","doi":"10.48550/arXiv.2306.05484","DOIUrl":"https://doi.org/10.48550/arXiv.2306.05484","url":null,"abstract":"Understanding causality should be a core requirement of any attempt to build real impact through AI. Due to the inherent unobservability of counterfactuals, large randomised trials (RCTs) are the standard for causal inference. But large experiments are generically expensive, and randomisation carries its own costs, e.g. when suboptimal decisions are trialed. Recent work has proposed more sample-efficient alternatives to RCTs, but these are not adaptable to the downstream application for which the causal effect is sought. In this work, we develop a task-specific approach to experimental design and derive sampling strategies customised to particular downstream applications. Across a range of important tasks, real-world datasets, and sample sizes, our method outperforms other benchmarks, e.g. requiring an order-of-magnitude less data to match RCT performance on targeted marketing tasks.","PeriodicalId":74529,"journal":{"name":"Proceedings of the ... International Conference on Machine Learning. International Conference on Machine Learning","volume":"25 1","pages":"6384-6401"},"PeriodicalIF":0.0,"publicationDate":"2023-06-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82125463","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Efficient and Equivariant Graph Networks for Predicting Quantum Hamiltonian 预测量子哈密顿量的高效等变图网络
Haiyang Yu, Zhao Xu, X. Qian, Xiaoning Qian, Shuiwang Ji
We consider the prediction of the Hamiltonian matrix, which finds use in quantum chemistry and condensed matter physics. Efficiency and equivariance are two important, but conflicting factors. In this work, we propose a SE(3)-equivariant network, named QHNet, that achieves efficiency and equivariance. Our key advance lies at the innovative design of QHNet architecture, which not only obeys the underlying symmetries, but also enables the reduction of number of tensor products by 92%. In addition, QHNet prevents the exponential growth of channel dimension when more atom types are involved. We perform experiments on MD17 datasets, including four molecular systems. Experimental results show that our QHNet can achieve comparable performance to the state of the art methods at a significantly faster speed. Besides, our QHNet consumes 50% less memory due to its streamlined architecture. Our code is publicly available as part of the AIRS library (url{https://github.com/divelab/AIRS}).
我们考虑了在量子化学和凝聚态物理中使用的哈密顿矩阵的预测。效率和等方差是两个重要但相互矛盾的因素。在这项工作中,我们提出了一个SE(3)-等变网络,命名为QHNet,实现了效率和等变。我们的关键进步在于QHNet架构的创新设计,它不仅遵循底层的对称性,而且使张量积的数量减少了92%。此外,当涉及更多原子类型时,QHNet可以防止通道尺寸的指数增长。我们在MD17数据集上进行实验,包括四个分子系统。实验结果表明,我们的QHNet可以以更快的速度获得与最先进方法相当的性能。此外,我们的QHNet由于其流线型架构而减少了50%的内存消耗。我们的代码作为AIRS库的一部分公开提供(url{https://github.com/divelab/AIRS})。
{"title":"Efficient and Equivariant Graph Networks for Predicting Quantum Hamiltonian","authors":"Haiyang Yu, Zhao Xu, X. Qian, Xiaoning Qian, Shuiwang Ji","doi":"10.48550/arXiv.2306.04922","DOIUrl":"https://doi.org/10.48550/arXiv.2306.04922","url":null,"abstract":"We consider the prediction of the Hamiltonian matrix, which finds use in quantum chemistry and condensed matter physics. Efficiency and equivariance are two important, but conflicting factors. In this work, we propose a SE(3)-equivariant network, named QHNet, that achieves efficiency and equivariance. Our key advance lies at the innovative design of QHNet architecture, which not only obeys the underlying symmetries, but also enables the reduction of number of tensor products by 92%. In addition, QHNet prevents the exponential growth of channel dimension when more atom types are involved. We perform experiments on MD17 datasets, including four molecular systems. Experimental results show that our QHNet can achieve comparable performance to the state of the art methods at a significantly faster speed. Besides, our QHNet consumes 50% less memory due to its streamlined architecture. Our code is publicly available as part of the AIRS library (url{https://github.com/divelab/AIRS}).","PeriodicalId":74529,"journal":{"name":"Proceedings of the ... International Conference on Machine Learning. International Conference on Machine Learning","volume":"29 1","pages":"40412-40424"},"PeriodicalIF":0.0,"publicationDate":"2023-06-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"75151680","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Federated Linear Contextual Bandits with User-level Differential Privacy 具有用户级差分隐私的联邦线性上下文强盗
Ruiquan Huang, Huanyu Zhang, Luca Melis, Milan Shen, Meisam Hajzinia, J. Yang
This paper studies federated linear contextual bandits under the notion of user-level differential privacy (DP). We first introduce a unified federated bandits framework that can accommodate various definitions of DP in the sequential decision-making setting. We then formally introduce user-level central DP (CDP) and local DP (LDP) in the federated bandits framework, and investigate the fundamental trade-offs between the learning regrets and the corresponding DP guarantees in a federated linear contextual bandits model. For CDP, we propose a federated algorithm termed as $texttt{ROBIN}$ and show that it is near-optimal in terms of the number of clients $M$ and the privacy budget $varepsilon$ by deriving nearly-matching upper and lower regret bounds when user-level DP is satisfied. For LDP, we obtain several lower bounds, indicating that learning under user-level $(varepsilon,delta)$-LDP must suffer a regret blow-up factor at least $min{1/varepsilon,M}$ or $min{1/sqrt{varepsilon},sqrt{M}}$ under different conditions.
本文在用户级差分隐私(DP)的概念下研究了联邦线性上下文强盗。我们首先引入了一个统一的联邦强盗框架,它可以适应顺序决策设置中DP的各种定义。然后,我们在联邦盗匪框架中正式引入了用户级中心DP (CDP)和本地DP (LDP),并研究了在联邦线性上下文盗匪模型中学习遗憾和相应DP保证之间的基本权衡。对于CDP,我们提出了一种称为$texttt{ROBIN}$的联邦算法,并通过在满足用户级DP时推导出几乎匹配的上后悔界和下后悔界,表明它在客户端数量$M$和隐私预算$varepsilon$方面是接近最优的。对于LDP,我们得到了几个下界,表明在不同条件下,在用户级$(varepsilon,delta)$ -LDP下的学习必须至少有一个后悔膨胀因子$min{1/varepsilon,M}$或$min{1/sqrt{varepsilon},sqrt{M}}$。
{"title":"Federated Linear Contextual Bandits with User-level Differential Privacy","authors":"Ruiquan Huang, Huanyu Zhang, Luca Melis, Milan Shen, Meisam Hajzinia, J. Yang","doi":"10.48550/arXiv.2306.05275","DOIUrl":"https://doi.org/10.48550/arXiv.2306.05275","url":null,"abstract":"This paper studies federated linear contextual bandits under the notion of user-level differential privacy (DP). We first introduce a unified federated bandits framework that can accommodate various definitions of DP in the sequential decision-making setting. We then formally introduce user-level central DP (CDP) and local DP (LDP) in the federated bandits framework, and investigate the fundamental trade-offs between the learning regrets and the corresponding DP guarantees in a federated linear contextual bandits model. For CDP, we propose a federated algorithm termed as $texttt{ROBIN}$ and show that it is near-optimal in terms of the number of clients $M$ and the privacy budget $varepsilon$ by deriving nearly-matching upper and lower regret bounds when user-level DP is satisfied. For LDP, we obtain several lower bounds, indicating that learning under user-level $(varepsilon,delta)$-LDP must suffer a regret blow-up factor at least $min{1/varepsilon,M}$ or $min{1/sqrt{varepsilon},sqrt{M}}$ under different conditions.","PeriodicalId":74529,"journal":{"name":"Proceedings of the ... International Conference on Machine Learning. International Conference on Machine Learning","volume":"89 1","pages":"14060-14095"},"PeriodicalIF":0.0,"publicationDate":"2023-06-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"79466144","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
On Computing Optimal Tree Ensembles 关于计算最优树集成
Christian Komusiewicz, Pascal Kunz, Frank Sommer, Manuel Sorge
Random forests and, more generally, (decisionnobreakdash-)tree ensembles are widely used methods for classification and regression. Recent algorithmic advances allow to compute decision trees that are optimal for various measures such as their size or depth. We are not aware of such research for tree ensembles and aim to contribute to this area. Mainly, we provide two novel algorithms and corresponding lower bounds. First, we are able to carry over and substantially improve on tractability results for decision trees, obtaining a $(6delta D S)^S cdot poly$-time algorithm, where $S$ is the number of cuts in the tree ensemble, $D$ the largest domain size, and $delta$ is the largest number of features in which two examples differ. To achieve this, we introduce the witness-tree technique which also seems promising for practice. Second, we show that dynamic programming, which has been successful for decision trees, may also be viable for tree ensembles, providing an $ell^n cdot poly$-time algorithm, where $ell$ is the number of trees and $n$ the number of examples. Finally, we compare the number of cuts necessary to classify training data sets for decision trees and tree ensembles, showing that ensembles may need exponentially fewer cuts for increasing number of trees.
随机森林和更一般的(decisionnobreakdash-)树集成是广泛使用的分类和回归方法。最近的算法进步允许计算决策树,这是最优的各种措施,如他们的大小或深度。我们不知道这样的研究树的集合和目标是贡献这一领域。主要给出了两种新的算法和相应的下界。首先,我们能够延续并大幅改进决策树的可追溯性结果,获得$(6delta D S)^S cdot poly$ time算法,其中$S$是树集合中的切割数,$D$是最大域大小,$delta$是两个示例不同的最大特征数。为了实现这一目标,我们引入了似乎也很有希望用于实践的见证树技术。其次,我们展示了动态规划,这已经成功的决策树,也可能是可行的树集成,提供了一个$ well ^n cdot聚$时间算法,其中$ well $是树的数量,$n$是例子的数量。最后,我们比较了决策树和树集成分类训练数据集所需的切割次数,表明随着树数量的增加,集成可能需要的切割次数呈指数级减少。
{"title":"On Computing Optimal Tree Ensembles","authors":"Christian Komusiewicz, Pascal Kunz, Frank Sommer, Manuel Sorge","doi":"10.48550/arXiv.2306.04423","DOIUrl":"https://doi.org/10.48550/arXiv.2306.04423","url":null,"abstract":"Random forests and, more generally, (decisionnobreakdash-)tree ensembles are widely used methods for classification and regression. Recent algorithmic advances allow to compute decision trees that are optimal for various measures such as their size or depth. We are not aware of such research for tree ensembles and aim to contribute to this area. Mainly, we provide two novel algorithms and corresponding lower bounds. First, we are able to carry over and substantially improve on tractability results for decision trees, obtaining a $(6delta D S)^S cdot poly$-time algorithm, where $S$ is the number of cuts in the tree ensemble, $D$ the largest domain size, and $delta$ is the largest number of features in which two examples differ. To achieve this, we introduce the witness-tree technique which also seems promising for practice. Second, we show that dynamic programming, which has been successful for decision trees, may also be viable for tree ensembles, providing an $ell^n cdot poly$-time algorithm, where $ell$ is the number of trees and $n$ the number of examples. Finally, we compare the number of cuts necessary to classify training data sets for decision trees and tree ensembles, showing that ensembles may need exponentially fewer cuts for increasing number of trees.","PeriodicalId":74529,"journal":{"name":"Proceedings of the ... International Conference on Machine Learning. International Conference on Machine Learning","volume":"100 1","pages":"17364-17374"},"PeriodicalIF":0.0,"publicationDate":"2023-06-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"89677484","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Training-Free Neural Active Learning with Initialization-Robustness Guarantees 具有初始化-鲁棒性保证的无训练神经主动学习
Apivich Hemachandra, Zhongxiang Dai, Jasraj Singh, See-Kiong Ng, K. H. Low
Existing neural active learning algorithms have aimed to optimize the predictive performance of neural networks (NNs) by selecting data for labelling. However, other than a good predictive performance, being robust against random parameter initializations is also a crucial requirement in safety-critical applications. To this end, we introduce our expected variance with Gaussian processes (EV-GP) criterion for neural active learning, which is theoretically guaranteed to select data points which lead to trained NNs with both (a) good predictive performances and (b) initialization robustness. Importantly, our EV-GP criterion is training-free, i.e., it does not require any training of the NN during data selection, which makes it computationally efficient. We empirically demonstrate that our EV-GP criterion is highly correlated with both initialization robustness and generalization performance, and show that it consistently outperforms baseline methods in terms of both desiderata, especially in situations with limited initial data or large batch sizes.
现有的神经主动学习算法旨在通过选择数据进行标记来优化神经网络(nn)的预测性能。然而,除了良好的预测性能之外,在安全关键型应用程序中,对随机参数初始化的鲁棒性也是一个至关重要的要求。为此,我们引入了神经主动学习的高斯过程期望方差(EV-GP)准则,理论上保证选择数据点,从而使训练好的神经网络具有(a)良好的预测性能和(b)初始化鲁棒性。重要的是,我们的EV-GP准则是无训练的,即在数据选择过程中不需要对神经网络进行任何训练,这使得它的计算效率很高。我们的经验证明,我们的EV-GP标准与初始化鲁棒性和泛化性能高度相关,并表明它在两种理想情况下始终优于基线方法,特别是在初始数据有限或批量较大的情况下。
{"title":"Training-Free Neural Active Learning with Initialization-Robustness Guarantees","authors":"Apivich Hemachandra, Zhongxiang Dai, Jasraj Singh, See-Kiong Ng, K. H. Low","doi":"10.48550/arXiv.2306.04454","DOIUrl":"https://doi.org/10.48550/arXiv.2306.04454","url":null,"abstract":"Existing neural active learning algorithms have aimed to optimize the predictive performance of neural networks (NNs) by selecting data for labelling. However, other than a good predictive performance, being robust against random parameter initializations is also a crucial requirement in safety-critical applications. To this end, we introduce our expected variance with Gaussian processes (EV-GP) criterion for neural active learning, which is theoretically guaranteed to select data points which lead to trained NNs with both (a) good predictive performances and (b) initialization robustness. Importantly, our EV-GP criterion is training-free, i.e., it does not require any training of the NN during data selection, which makes it computationally efficient. We empirically demonstrate that our EV-GP criterion is highly correlated with both initialization robustness and generalization performance, and show that it consistently outperforms baseline methods in terms of both desiderata, especially in situations with limited initial data or large batch sizes.","PeriodicalId":74529,"journal":{"name":"Proceedings of the ... International Conference on Machine Learning. International Conference on Machine Learning","volume":"29 1","pages":"12931-12971"},"PeriodicalIF":0.0,"publicationDate":"2023-06-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"90006971","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Generalized Teacher Forcing for Learning Chaotic Dynamics 混沌动力学学习的广义教师强迫
F. Hess, Z. Monfared, Manuela Brenner, D. Durstewitz
Chaotic dynamical systems (DS) are ubiquitous in nature and society. Often we are interested in reconstructing such systems from observed time series for prediction or mechanistic insight, where by reconstruction we mean learning geometrical and invariant temporal properties of the system in question (like attractors). However, training reconstruction algorithms like recurrent neural networks (RNNs) on such systems by gradient-descent based techniques faces severe challenges. This is mainly due to exploding gradients caused by the exponential divergence of trajectories in chaotic systems. Moreover, for (scientific) interpretability we wish to have as low dimensional reconstructions as possible, preferably in a model which is mathematically tractable. Here we report that a surprisingly simple modification of teacher forcing leads to provably strictly all-time bounded gradients in training on chaotic systems, and, when paired with a simple architectural rearrangement of a tractable RNN design, piecewise-linear RNNs (PLRNNs), allows for faithful reconstruction in spaces of at most the dimensionality of the observed system. We show on several DS that with these amendments we can reconstruct DS better than current SOTA algorithms, in much lower dimensions. Performance differences were particularly compelling on real world data with which most other methods severely struggled. This work thus led to a simple yet powerful DS reconstruction algorithm which is highly interpretable at the same time.
混沌动力系统(DS)在自然界和社会中普遍存在。通常,我们对从观察到的时间序列中重建这样的系统以进行预测或机械洞察感兴趣,通过重建,我们意味着学习所讨论系统的几何和不变的时间特性(如吸引子)。然而,基于梯度下降技术在此类系统上训练递归神经网络(rnn)等重构算法面临严峻挑战。这主要是由于混沌系统中轨迹的指数发散引起的爆炸梯度。此外,为了(科学的)可解释性,我们希望有尽可能低维的重建,最好是在数学上易于处理的模型中。在这里,我们报告了一种令人惊讶的简单的教师强迫修改导致在混沌系统的训练中可以证明严格的时间有界梯度,并且,当与可处理RNN设计的简单架构重排相结合时,分段线性RNN (plrnn)允许在最多观察系统维度的空间中进行忠实重建。我们在几个DS上表明,通过这些修正,我们可以在更低的维度上比当前的SOTA算法更好地重建DS。在处理真实世界的数据时,性能差异尤其引人注目,而大多数其他方法都很难处理这些数据。这项工作导致了一个简单而强大的DS重建算法,同时具有高度的可解释性。
{"title":"Generalized Teacher Forcing for Learning Chaotic Dynamics","authors":"F. Hess, Z. Monfared, Manuela Brenner, D. Durstewitz","doi":"10.48550/arXiv.2306.04406","DOIUrl":"https://doi.org/10.48550/arXiv.2306.04406","url":null,"abstract":"Chaotic dynamical systems (DS) are ubiquitous in nature and society. Often we are interested in reconstructing such systems from observed time series for prediction or mechanistic insight, where by reconstruction we mean learning geometrical and invariant temporal properties of the system in question (like attractors). However, training reconstruction algorithms like recurrent neural networks (RNNs) on such systems by gradient-descent based techniques faces severe challenges. This is mainly due to exploding gradients caused by the exponential divergence of trajectories in chaotic systems. Moreover, for (scientific) interpretability we wish to have as low dimensional reconstructions as possible, preferably in a model which is mathematically tractable. Here we report that a surprisingly simple modification of teacher forcing leads to provably strictly all-time bounded gradients in training on chaotic systems, and, when paired with a simple architectural rearrangement of a tractable RNN design, piecewise-linear RNNs (PLRNNs), allows for faithful reconstruction in spaces of at most the dimensionality of the observed system. We show on several DS that with these amendments we can reconstruct DS better than current SOTA algorithms, in much lower dimensions. Performance differences were particularly compelling on real world data with which most other methods severely struggled. This work thus led to a simple yet powerful DS reconstruction algorithm which is highly interpretable at the same time.","PeriodicalId":74529,"journal":{"name":"Proceedings of the ... International Conference on Machine Learning. International Conference on Machine Learning","volume":"27 1","pages":"13017-13049"},"PeriodicalIF":0.0,"publicationDate":"2023-06-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"74683388","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
期刊
Proceedings of the ... International Conference on Machine Learning. International Conference on Machine Learning
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1