首页 > 最新文献

Proceedings of the ... International Conference on Machine Learning. International Conference on Machine Learning最新文献

英文 中文
Continuation Path Learning for Homotopy Optimization 同伦优化的延拓路径学习
Xi Lin, Zhiyuan Yang, Xiao-Yan Zhang, Qingfu Zhang
Homotopy optimization is a traditional method to deal with a complicated optimization problem by solving a sequence of easy-to-hard surrogate subproblems. However, this method can be very sensitive to the continuation schedule design and might lead to a suboptimal solution to the original problem. In addition, the intermediate solutions, often ignored by classic homotopy optimization, could be useful for many real-world applications. In this work, we propose a novel model-based approach to learn the whole continuation path for homotopy optimization, which contains infinite intermediate solutions for any surrogate subproblems. Rather than the classic unidirectional easy-to-hard optimization, our method can simultaneously optimize the original problem and all surrogate subproblems in a collaborative manner. The proposed model also supports real-time generation of any intermediate solution, which could be desirable for many applications. Experimental studies on different problems show that our proposed method can significantly improve the performance of homotopy optimization and provide extra helpful information to support better decision-making.
同伦优化是一种传统的解决复杂优化问题的方法,它通过求解一系列难易的代理子问题。然而,该方法对连续调度设计非常敏感,可能导致原问题的次优解。此外,通常被经典同伦优化所忽略的中间解可能对许多实际应用程序很有用。在这项工作中,我们提出了一种新的基于模型的方法来学习同伦优化的整个连续路径,该路径包含任何代理子问题的无限中间解。与传统的单向易难优化不同,该方法能够以协作的方式同时优化原始问题和所有代理子问题。所提出的模型还支持任何中间解决方案的实时生成,这可能是许多应用程序所需要的。对不同问题的实验研究表明,本文提出的方法可以显著提高同伦优化的性能,并为更好的决策提供额外的有用信息。
{"title":"Continuation Path Learning for Homotopy Optimization","authors":"Xi Lin, Zhiyuan Yang, Xiao-Yan Zhang, Qingfu Zhang","doi":"10.48550/arXiv.2307.12551","DOIUrl":"https://doi.org/10.48550/arXiv.2307.12551","url":null,"abstract":"Homotopy optimization is a traditional method to deal with a complicated optimization problem by solving a sequence of easy-to-hard surrogate subproblems. However, this method can be very sensitive to the continuation schedule design and might lead to a suboptimal solution to the original problem. In addition, the intermediate solutions, often ignored by classic homotopy optimization, could be useful for many real-world applications. In this work, we propose a novel model-based approach to learn the whole continuation path for homotopy optimization, which contains infinite intermediate solutions for any surrogate subproblems. Rather than the classic unidirectional easy-to-hard optimization, our method can simultaneously optimize the original problem and all surrogate subproblems in a collaborative manner. The proposed model also supports real-time generation of any intermediate solution, which could be desirable for many applications. Experimental studies on different problems show that our proposed method can significantly improve the performance of homotopy optimization and provide extra helpful information to support better decision-making.","PeriodicalId":74529,"journal":{"name":"Proceedings of the ... International Conference on Machine Learning. International Conference on Machine Learning","volume":"65 1","pages":"21288-21311"},"PeriodicalIF":0.0,"publicationDate":"2023-07-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82599543","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
On the Effectiveness of Offline RL for Dialogue Response Generation 离线强化学习在对话响应生成中的有效性研究
Paloma Sodhi, Felix Wu, Ethan R. Elenberg, Kilian Q. Weinberger, Ryan T. McDonald
A common training technique for language models is teacher forcing (TF). TF attempts to match human language exactly, even though identical meanings can be expressed in different ways. This motivates use of sequence-level objectives for dialogue response generation. In this paper, we study the efficacy of various offline reinforcement learning (RL) methods to maximize such objectives. We present a comprehensive evaluation across multiple datasets, models, and metrics. Offline RL shows a clear performance improvement over teacher forcing while not inducing training instability or sacrificing practical training budgets.
一种常见的语言模型训练技术是教师强迫(TF)。TF试图完全匹配人类语言,即使相同的意思可以用不同的方式表达。这促使使用序列级目标生成对话响应。在本文中,我们研究了各种离线强化学习(RL)方法的有效性,以最大化这些目标。我们提出了跨多个数据集、模型和指标的综合评估。离线强化学习在不导致培训不稳定或牺牲实际培训预算的情况下,比教师强迫表现出明显的绩效改善。
{"title":"On the Effectiveness of Offline RL for Dialogue Response Generation","authors":"Paloma Sodhi, Felix Wu, Ethan R. Elenberg, Kilian Q. Weinberger, Ryan T. McDonald","doi":"10.48550/arXiv.2307.12425","DOIUrl":"https://doi.org/10.48550/arXiv.2307.12425","url":null,"abstract":"A common training technique for language models is teacher forcing (TF). TF attempts to match human language exactly, even though identical meanings can be expressed in different ways. This motivates use of sequence-level objectives for dialogue response generation. In this paper, we study the efficacy of various offline reinforcement learning (RL) methods to maximize such objectives. We present a comprehensive evaluation across multiple datasets, models, and metrics. Offline RL shows a clear performance improvement over teacher forcing while not inducing training instability or sacrificing practical training budgets.","PeriodicalId":74529,"journal":{"name":"Proceedings of the ... International Conference on Machine Learning. International Conference on Machine Learning","volume":"3 1","pages":"32088-32104"},"PeriodicalIF":0.0,"publicationDate":"2023-07-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"90876190","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Model-based Offline Reinforcement Learning with Count-based Conservatism 基于模型的基于计数保守的离线强化学习
Byeongchang Kim, Min-hwan Oh
In this paper, we propose a model-based offline reinforcement learning method that integrates count-based conservatism, named $texttt{Count-MORL}$. Our method utilizes the count estimates of state-action pairs to quantify model estimation error, marking the first algorithm of demonstrating the efficacy of count-based conservatism in model-based offline deep RL to the best of our knowledge. For our proposed method, we first show that the estimation error is inversely proportional to the frequency of state-action pairs. Secondly, we demonstrate that the learned policy under the count-based conservative model offers near-optimality performance guarantees. Through extensive numerical experiments, we validate that $texttt{Count-MORL}$ with hash code implementation significantly outperforms existing offline RL algorithms on the D4RL benchmark datasets. The code is accessible at $href{https://github.com/oh-lab/Count-MORL}{https://github.com/oh-lab/Count-MORL}$.
在本文中,我们提出了一种基于模型的离线强化学习方法,该方法集成了基于计数的保守性,命名为$texttt{Count-MORL}$。我们的方法利用状态-动作对的计数估计来量化模型估计误差,这标志着据我们所知,第一个在基于模型的离线深度强化学习中证明基于计数的保守性有效性的算法。对于我们提出的方法,我们首先证明了估计误差与状态-动作对的频率成反比。其次,我们证明了在基于计数的保守模型下学习的策略提供了接近最优的性能保证。通过大量的数值实验,我们验证了$texttt{Count-MORL}$与哈希码实现在D4RL基准数据集上显著优于现有的离线RL算法。代码可在$href{https://github.com/oh-lab/Count-MORL}{https://github.com/oh-lab/Count-MORL}$上访问。
{"title":"Model-based Offline Reinforcement Learning with Count-based Conservatism","authors":"Byeongchang Kim, Min-hwan Oh","doi":"10.48550/arXiv.2307.11352","DOIUrl":"https://doi.org/10.48550/arXiv.2307.11352","url":null,"abstract":"In this paper, we propose a model-based offline reinforcement learning method that integrates count-based conservatism, named $texttt{Count-MORL}$. Our method utilizes the count estimates of state-action pairs to quantify model estimation error, marking the first algorithm of demonstrating the efficacy of count-based conservatism in model-based offline deep RL to the best of our knowledge. For our proposed method, we first show that the estimation error is inversely proportional to the frequency of state-action pairs. Secondly, we demonstrate that the learned policy under the count-based conservative model offers near-optimality performance guarantees. Through extensive numerical experiments, we validate that $texttt{Count-MORL}$ with hash code implementation significantly outperforms existing offline RL algorithms on the D4RL benchmark datasets. The code is accessible at $href{https://github.com/oh-lab/Count-MORL}{https://github.com/oh-lab/Count-MORL}$.","PeriodicalId":74529,"journal":{"name":"Proceedings of the ... International Conference on Machine Learning. International Conference on Machine Learning","volume":"38 1","pages":"16728-16746"},"PeriodicalIF":0.0,"publicationDate":"2023-07-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"86091968","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Reparameterized Policy Learning for Multimodal Trajectory Optimization 多模态轨迹优化的再参数化策略学习
Zhiao Huang, Litian Liang, Z. Ling, Xuanlin Li, Chuang Gan, H. Su
We investigate the challenge of parametrizing policies for reinforcement learning (RL) in high-dimensional continuous action spaces. Our objective is to develop a multimodal policy that overcomes limitations inherent in the commonly-used Gaussian parameterization. To achieve this, we propose a principled framework that models the continuous RL policy as a generative model of optimal trajectories. By conditioning the policy on a latent variable, we derive a novel variational bound as the optimization objective, which promotes exploration of the environment. We then present a practical model-based RL method, called Reparameterized Policy Gradient (RPG), which leverages the multimodal policy parameterization and learned world model to achieve strong exploration capabilities and high data efficiency. Empirical results demonstrate that our method can help agents evade local optima in tasks with dense rewards and solve challenging sparse-reward environments by incorporating an object-centric intrinsic reward. Our method consistently outperforms previous approaches across a range of tasks. Code and supplementary materials are available on the project page https://haosulab.github.io/RPG/
我们研究了在高维连续动作空间中参数化强化学习(RL)策略的挑战。我们的目标是开发一种多模态策略,克服常用高斯参数化固有的局限性。为了实现这一目标,我们提出了一个原则性框架,将连续RL策略建模为最优轨迹的生成模型。通过对潜在变量的约束,我们得到了一个新的变分界作为优化目标,从而促进了对环境的探索。然后,我们提出了一种实用的基于模型的RL方法,称为重参数化策略梯度(RPG),该方法利用多模态策略参数化和学习的世界模型来实现强大的探索能力和高数据效率。实证结果表明,我们的方法可以帮助智能体在具有密集奖励的任务中逃避局部最优,并通过结合以对象为中心的内在奖励来解决具有挑战性的稀疏奖励环境。我们的方法在一系列任务中始终优于以前的方法。代码和补充材料可在项目页面https://haosulab.github.io/RPG/上获得
{"title":"Reparameterized Policy Learning for Multimodal Trajectory Optimization","authors":"Zhiao Huang, Litian Liang, Z. Ling, Xuanlin Li, Chuang Gan, H. Su","doi":"10.48550/arXiv.2307.10710","DOIUrl":"https://doi.org/10.48550/arXiv.2307.10710","url":null,"abstract":"We investigate the challenge of parametrizing policies for reinforcement learning (RL) in high-dimensional continuous action spaces. Our objective is to develop a multimodal policy that overcomes limitations inherent in the commonly-used Gaussian parameterization. To achieve this, we propose a principled framework that models the continuous RL policy as a generative model of optimal trajectories. By conditioning the policy on a latent variable, we derive a novel variational bound as the optimization objective, which promotes exploration of the environment. We then present a practical model-based RL method, called Reparameterized Policy Gradient (RPG), which leverages the multimodal policy parameterization and learned world model to achieve strong exploration capabilities and high data efficiency. Empirical results demonstrate that our method can help agents evade local optima in tasks with dense rewards and solve challenging sparse-reward environments by incorporating an object-centric intrinsic reward. Our method consistently outperforms previous approaches across a range of tasks. Code and supplementary materials are available on the project page https://haosulab.github.io/RPG/","PeriodicalId":74529,"journal":{"name":"Proceedings of the ... International Conference on Machine Learning. International Conference on Machine Learning","volume":"23 1","pages":"13957-13975"},"PeriodicalIF":0.0,"publicationDate":"2023-07-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"87664226","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Private Federated Learning with Autotuned Compression 具有自调优压缩的私有联邦学习
Enayat Ullah, Christopher A. Choquette-Choo, P. Kairouz, Sewoong Oh
We propose new techniques for reducing communication in private federated learning without the need for setting or tuning compression rates. Our on-the-fly methods automatically adjust the compression rate based on the error induced during training, while maintaining provable privacy guarantees through the use of secure aggregation and differential privacy. Our techniques are provably instance-optimal for mean estimation, meaning that they can adapt to the ``hardness of the problem"with minimal interactivity. We demonstrate the effectiveness of our approach on real-world datasets by achieving favorable compression rates without the need for tuning.
我们提出了在不需要设置或调整压缩率的情况下减少私有联邦学习中的通信的新技术。我们的实时方法根据训练过程中产生的错误自动调整压缩率,同时通过使用安全聚合和差分隐私来保持可证明的隐私保证。我们的技术对于平均估计来说是可证明的实例最优的,这意味着它们可以以最小的交互性适应“问题的难度”。我们通过在不需要调优的情况下获得有利的压缩率,证明了我们的方法在真实数据集上的有效性。
{"title":"Private Federated Learning with Autotuned Compression","authors":"Enayat Ullah, Christopher A. Choquette-Choo, P. Kairouz, Sewoong Oh","doi":"10.48550/arXiv.2307.10999","DOIUrl":"https://doi.org/10.48550/arXiv.2307.10999","url":null,"abstract":"We propose new techniques for reducing communication in private federated learning without the need for setting or tuning compression rates. Our on-the-fly methods automatically adjust the compression rate based on the error induced during training, while maintaining provable privacy guarantees through the use of secure aggregation and differential privacy. Our techniques are provably instance-optimal for mean estimation, meaning that they can adapt to the ``hardness of the problem\"with minimal interactivity. We demonstrate the effectiveness of our approach on real-world datasets by achieving favorable compression rates without the need for tuning.","PeriodicalId":74529,"journal":{"name":"Proceedings of the ... International Conference on Machine Learning. International Conference on Machine Learning","volume":"68 1","pages":"34668-34708"},"PeriodicalIF":0.0,"publicationDate":"2023-07-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"80066362","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Fractional Denoising for 3D Molecular Pre-training 三维分子预训练的分数去噪
Shi Feng, Yuyan Ni, Yanyan Lan, Zhiming Ma, Wei-Ying Ma
Coordinate denoising is a promising 3D molecular pre-training method, which has achieved remarkable performance in various downstream drug discovery tasks. Theoretically, the objective is equivalent to learning the force field, which is revealed helpful for downstream tasks. Nevertheless, there are two challenges for coordinate denoising to learn an effective force field, i.e. low coverage samples and isotropic force field. The underlying reason is that molecular distributions assumed by existing denoising methods fail to capture the anisotropic characteristic of molecules. To tackle these challenges, we propose a novel hybrid noise strategy, including noises on both dihedral angel and coordinate. However, denoising such hybrid noise in a traditional way is no more equivalent to learning the force field. Through theoretical deductions, we find that the problem is caused by the dependency of the input conformation for covariance. To this end, we propose to decouple the two types of noise and design a novel fractional denoising method (Frad), which only denoises the latter coordinate part. In this way, Frad enjoys both the merits of sampling more low-energy structures and the force field equivalence. Extensive experiments show the effectiveness of Frad in molecular representation, with a new state-of-the-art on 9 out of 12 tasks of QM9 and on 7 out of 8 targets of MD17.
坐标去噪是一种很有前途的三维分子预训练方法,在各种下游药物发现任务中取得了显著的效果。从理论上讲,目标相当于学习力场,这对后续任务有帮助。然而,坐标去噪学习有效力场存在两个挑战,即低覆盖样本和各向同性力场。其根本原因是现有的去噪方法所假设的分子分布未能捕捉到分子的各向异性特征。为了解决这些问题,我们提出了一种新的混合噪声策略,包括二面角和坐标上的噪声。然而,用传统的方法去噪这种混合噪声并不等同于学习力场。通过理论推导,我们发现这个问题是由输入构象对协方差的依赖性引起的。为此,我们提出将两种类型的噪声解耦,并设计了一种新的分数阶去噪方法(Frad),该方法只对后一种坐标部分进行去噪。这样,Frad既具有采样更多低能结构的优点,又具有力场等效性。大量的实验表明,Frad在分子表征方面是有效的,在QM9的12个任务中有9个任务和MD17的8个目标中有7个目标具有最新的技术水平。
{"title":"Fractional Denoising for 3D Molecular Pre-training","authors":"Shi Feng, Yuyan Ni, Yanyan Lan, Zhiming Ma, Wei-Ying Ma","doi":"10.48550/arXiv.2307.10683","DOIUrl":"https://doi.org/10.48550/arXiv.2307.10683","url":null,"abstract":"Coordinate denoising is a promising 3D molecular pre-training method, which has achieved remarkable performance in various downstream drug discovery tasks. Theoretically, the objective is equivalent to learning the force field, which is revealed helpful for downstream tasks. Nevertheless, there are two challenges for coordinate denoising to learn an effective force field, i.e. low coverage samples and isotropic force field. The underlying reason is that molecular distributions assumed by existing denoising methods fail to capture the anisotropic characteristic of molecules. To tackle these challenges, we propose a novel hybrid noise strategy, including noises on both dihedral angel and coordinate. However, denoising such hybrid noise in a traditional way is no more equivalent to learning the force field. Through theoretical deductions, we find that the problem is caused by the dependency of the input conformation for covariance. To this end, we propose to decouple the two types of noise and design a novel fractional denoising method (Frad), which only denoises the latter coordinate part. In this way, Frad enjoys both the merits of sampling more low-energy structures and the force field equivalence. Extensive experiments show the effectiveness of Frad in molecular representation, with a new state-of-the-art on 9 out of 12 tasks of QM9 and on 7 out of 8 targets of MD17.","PeriodicalId":74529,"journal":{"name":"Proceedings of the ... International Conference on Machine Learning. International Conference on Machine Learning","volume":"14 1","pages":"9938-9961"},"PeriodicalIF":0.0,"publicationDate":"2023-07-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"89062630","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
From Adaptive Query Release to Machine Unlearning 从自适应查询释放到机器学习
Enayat Ullah, R. Arora
We formalize the problem of machine unlearning as design of efficient unlearning algorithms corresponding to learning algorithms which perform a selection of adaptive queries from structured query classes. We give efficient unlearning algorithms for linear and prefix-sum query classes. As applications, we show that unlearning in many problems, in particular, stochastic convex optimization (SCO), can be reduced to the above, yielding improved guarantees for the problem. In particular, for smooth Lipschitz losses and any $rho>0$, our results yield an unlearning algorithm with excess population risk of $tilde Obig(frac{1}{sqrt{n}}+frac{sqrt{d}}{nrho}big)$ with unlearning query (gradient) complexity $tilde O(rho cdot text{Retraining Complexity})$, where $d$ is the model dimensionality and $n$ is the initial number of samples. For non-smooth Lipschitz losses, we give an unlearning algorithm with excess population risk $tilde Obig(frac{1}{sqrt{n}}+big(frac{sqrt{d}}{nrho}big)^{1/2}big)$ with the same unlearning query (gradient) complexity. Furthermore, in the special case of Generalized Linear Models (GLMs), such as those in linear and logistic regression, we get dimension-independent rates of $tilde Obig(frac{1}{sqrt{n}} +frac{1}{(nrho)^{2/3}}big)$ and $tilde Obig(frac{1}{sqrt{n}} +frac{1}{(nrho)^{1/3}}big)$ for smooth Lipschitz and non-smooth Lipschitz losses respectively. Finally, we give generalizations of the above from one unlearning request to textit{dynamic} streams consisting of insertions and deletions.
我们将机器学习问题形式化为有效的学习算法的设计,该算法对应于从结构化查询类中选择自适应查询的学习算法。给出了线性查询类和前缀和查询类的有效学习算法。作为应用,我们证明了在许多问题中,特别是随机凸优化(SCO)中的学习可以简化为上述问题,从而提高了对问题的保证。特别是,对于平滑Lipschitz损失和任何$rho>0$,我们的结果产生了一个具有过度人口风险的取消学习算法$tilde Obig(frac{1}{sqrt{n}}+frac{sqrt{d}}{nrho}big)$和取消学习查询(梯度)复杂性$tilde O(rho cdot text{Retraining Complexity})$,其中$d$是模型维数,$n$是初始样本数。对于非光滑Lipschitz损失,我们给出了一种具有超额人口风险$tilde Obig(frac{1}{sqrt{n}}+big(frac{sqrt{d}}{nrho}big)^{1/2}big)$的学习算法,该算法具有相同的学习查询(梯度)复杂度。此外,在广义线性模型(GLMs)的特殊情况下,如线性回归和逻辑回归,我们分别得到光滑Lipschitz和非光滑Lipschitz损失的$tilde Obig(frac{1}{sqrt{n}} +frac{1}{(nrho)^{2/3}}big)$和$tilde Obig(frac{1}{sqrt{n}} +frac{1}{(nrho)^{1/3}}big)$的维无关率。最后,我们给出了上述的概括,从一个遗忘请求到由插入和删除组成的textit{动态}流。
{"title":"From Adaptive Query Release to Machine Unlearning","authors":"Enayat Ullah, R. Arora","doi":"10.48550/arXiv.2307.11228","DOIUrl":"https://doi.org/10.48550/arXiv.2307.11228","url":null,"abstract":"We formalize the problem of machine unlearning as design of efficient unlearning algorithms corresponding to learning algorithms which perform a selection of adaptive queries from structured query classes. We give efficient unlearning algorithms for linear and prefix-sum query classes. As applications, we show that unlearning in many problems, in particular, stochastic convex optimization (SCO), can be reduced to the above, yielding improved guarantees for the problem. In particular, for smooth Lipschitz losses and any $rho>0$, our results yield an unlearning algorithm with excess population risk of $tilde Obig(frac{1}{sqrt{n}}+frac{sqrt{d}}{nrho}big)$ with unlearning query (gradient) complexity $tilde O(rho cdot text{Retraining Complexity})$, where $d$ is the model dimensionality and $n$ is the initial number of samples. For non-smooth Lipschitz losses, we give an unlearning algorithm with excess population risk $tilde Obig(frac{1}{sqrt{n}}+big(frac{sqrt{d}}{nrho}big)^{1/2}big)$ with the same unlearning query (gradient) complexity. Furthermore, in the special case of Generalized Linear Models (GLMs), such as those in linear and logistic regression, we get dimension-independent rates of $tilde Obig(frac{1}{sqrt{n}} +frac{1}{(nrho)^{2/3}}big)$ and $tilde Obig(frac{1}{sqrt{n}} +frac{1}{(nrho)^{1/3}}big)$ for smooth Lipschitz and non-smooth Lipschitz losses respectively. Finally, we give generalizations of the above from one unlearning request to textit{dynamic} streams consisting of insertions and deletions.","PeriodicalId":74529,"journal":{"name":"Proceedings of the ... International Conference on Machine Learning. International Conference on Machine Learning","volume":"11 1","pages":"34642-34667"},"PeriodicalIF":0.0,"publicationDate":"2023-07-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82605525","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Sequential Multi-Dimensional Self-Supervised Learning for Clinical Time Series 临床时间序列的顺序多维自监督学习
Aniruddh Raghu, P. Chandak, Ridwan Alam, John Guttag, Collin M. Stultz
Self-supervised learning (SSL) for clinical time series data has received significant attention in recent literature, since these data are highly rich and provide important information about a patient's physiological state. However, most existing SSL methods for clinical time series are limited in that they are designed for unimodal time series, such as a sequence of structured features (e.g., lab values and vitals signs) or an individual high-dimensional physiological signal (e.g., an electrocardiogram). These existing methods cannot be readily extended to model time series that exhibit multimodality, with structured features and high-dimensional data being recorded at each timestep in the sequence. In this work, we address this gap and propose a new SSL method -- Sequential Multi-Dimensional SSL -- where a SSL loss is applied both at the level of the entire sequence and at the level of the individual high-dimensional data points in the sequence in order to better capture information at both scales. Our strategy is agnostic to the specific form of loss function used at each level -- it can be contrastive, as in SimCLR, or non-contrastive, as in VICReg. We evaluate our method on two real-world clinical datasets, where the time series contains sequences of (1) high-frequency electrocardiograms and (2) structured data from lab values and vitals signs. Our experimental results indicate that pre-training with our method and then fine-tuning on downstream tasks improves performance over baselines on both datasets, and in several settings, can lead to improvements across different self-supervised loss functions.
临床时间序列数据的自监督学习(SSL)在最近的文献中受到了极大的关注,因为这些数据非常丰富,提供了关于患者生理状态的重要信息。然而,大多数现有的用于临床时间序列的SSL方法都是有限的,因为它们是为单峰时间序列设计的,例如一系列结构化特征(例如,实验室值和生命体征)或单个高维生理信号(例如,心电图)。这些现有的方法不能很容易地扩展到表现出多模态的时间序列,在序列的每个时间步记录结构化特征和高维数据。在这项工作中,我们解决了这一差距,并提出了一种新的SSL方法——顺序多维SSL——其中SSL损失在整个序列和序列中单个高维数据点的水平上都被应用,以便更好地在两个尺度上捕获信息。我们的策略与每个级别使用的损失函数的特定形式无关——它可以是对比的,如SimCLR,也可以是非对比的,如VICReg。我们在两个真实世界的临床数据集上评估了我们的方法,其中时间序列包含(1)高频心电图序列和(2)来自实验室值和生命体征的结构化数据。我们的实验结果表明,使用我们的方法进行预训练,然后对下游任务进行微调,可以提高两个数据集在基线上的性能,并且在一些设置中,可以导致不同自监督损失函数的改进。
{"title":"Sequential Multi-Dimensional Self-Supervised Learning for Clinical Time Series","authors":"Aniruddh Raghu, P. Chandak, Ridwan Alam, John Guttag, Collin M. Stultz","doi":"10.48550/arXiv.2307.10923","DOIUrl":"https://doi.org/10.48550/arXiv.2307.10923","url":null,"abstract":"Self-supervised learning (SSL) for clinical time series data has received significant attention in recent literature, since these data are highly rich and provide important information about a patient's physiological state. However, most existing SSL methods for clinical time series are limited in that they are designed for unimodal time series, such as a sequence of structured features (e.g., lab values and vitals signs) or an individual high-dimensional physiological signal (e.g., an electrocardiogram). These existing methods cannot be readily extended to model time series that exhibit multimodality, with structured features and high-dimensional data being recorded at each timestep in the sequence. In this work, we address this gap and propose a new SSL method -- Sequential Multi-Dimensional SSL -- where a SSL loss is applied both at the level of the entire sequence and at the level of the individual high-dimensional data points in the sequence in order to better capture information at both scales. Our strategy is agnostic to the specific form of loss function used at each level -- it can be contrastive, as in SimCLR, or non-contrastive, as in VICReg. We evaluate our method on two real-world clinical datasets, where the time series contains sequences of (1) high-frequency electrocardiograms and (2) structured data from lab values and vitals signs. Our experimental results indicate that pre-training with our method and then fine-tuning on downstream tasks improves performance over baselines on both datasets, and in several settings, can lead to improvements across different self-supervised loss functions.","PeriodicalId":74529,"journal":{"name":"Proceedings of the ... International Conference on Machine Learning. International Conference on Machine Learning","volume":"68 1","pages":"28531-28548"},"PeriodicalIF":0.0,"publicationDate":"2023-07-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"89500786","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Contextual Reliability: When Different Features Matter in Different Contexts 上下文可靠性:当不同的特征在不同的上下文中起作用时
Gaurav R. Ghosal, Amrith Rajagopal Setlur, Daniel S. Brown, A. Dragan, Aditi Raghunathan
Deep neural networks often fail catastrophically by relying on spurious correlations. Most prior work assumes a clear dichotomy into spurious and reliable features; however, this is often unrealistic. For example, most of the time we do not want an autonomous car to simply copy the speed of surrounding cars -- we don't want our car to run a red light if a neighboring car does so. However, we cannot simply enforce invariance to next-lane speed, since it could provide valuable information about an unobservable pedestrian at a crosswalk. Thus, universally ignoring features that are sometimes (but not always) reliable can lead to non-robust performance. We formalize a new setting called contextual reliability which accounts for the fact that the"right"features to use may vary depending on the context. We propose and analyze a two-stage framework called Explicit Non-spurious feature Prediction (ENP) which first identifies the relevant features to use for a given context, then trains a model to rely exclusively on these features. Our work theoretically and empirically demonstrates the advantages of ENP over existing methods and provides new benchmarks for contextual reliability.
深度神经网络常常因为依赖于虚假的相关性而导致灾难性的失败。大多数先前的工作假设了虚假和可靠特征的明确二分法;然而,这通常是不现实的。例如,大多数时候,我们不希望一辆自动驾驶汽车简单地模仿周围汽车的速度——我们不希望我们的汽车在邻近汽车闯红灯的情况下也闯红灯。然而,我们不能简单地强制下一车道速度的不变性,因为它可以提供有关人行横道上不可观察的行人的有价值的信息。因此,普遍忽略有时(但不总是)可靠的特性可能导致性能不健壮。我们形式化了一种称为上下文可靠性的新设置,它说明了使用“正确”的功能可能因上下文而异的事实。我们提出并分析了一个称为显式非伪特征预测(ENP)的两阶段框架,该框架首先识别用于给定上下文的相关特征,然后训练模型完全依赖这些特征。我们的工作从理论上和经验上证明了ENP相对于现有方法的优势,并为上下文可靠性提供了新的基准。
{"title":"Contextual Reliability: When Different Features Matter in Different Contexts","authors":"Gaurav R. Ghosal, Amrith Rajagopal Setlur, Daniel S. Brown, A. Dragan, Aditi Raghunathan","doi":"10.48550/arXiv.2307.10026","DOIUrl":"https://doi.org/10.48550/arXiv.2307.10026","url":null,"abstract":"Deep neural networks often fail catastrophically by relying on spurious correlations. Most prior work assumes a clear dichotomy into spurious and reliable features; however, this is often unrealistic. For example, most of the time we do not want an autonomous car to simply copy the speed of surrounding cars -- we don't want our car to run a red light if a neighboring car does so. However, we cannot simply enforce invariance to next-lane speed, since it could provide valuable information about an unobservable pedestrian at a crosswalk. Thus, universally ignoring features that are sometimes (but not always) reliable can lead to non-robust performance. We formalize a new setting called contextual reliability which accounts for the fact that the\"right\"features to use may vary depending on the context. We propose and analyze a two-stage framework called Explicit Non-spurious feature Prediction (ENP) which first identifies the relevant features to use for a given context, then trains a model to rely exclusively on these features. Our work theoretically and empirically demonstrates the advantages of ENP over existing methods and provides new benchmarks for contextual reliability.","PeriodicalId":74529,"journal":{"name":"Proceedings of the ... International Conference on Machine Learning. International Conference on Machine Learning","volume":"173 1","pages":"11300-11320"},"PeriodicalIF":0.0,"publicationDate":"2023-07-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"74155802","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Convex Geometry of ReLU-layers, Injectivity on the Ball and Local Reconstruction relu层的凸几何、球上的注入性与局部重建
Daniel Haider, M. Ehler, P. Balázs
The paper uses a frame-theoretic setting to study the injectivity of a ReLU-layer on the closed ball of $mathbb{R}^n$ and its non-negative part. In particular, the interplay between the radius of the ball and the bias vector is emphasized. Together with a perspective from convex geometry, this leads to a computationally feasible method of verifying the injectivity of a ReLU-layer under reasonable restrictions in terms of an upper bound of the bias vector. Explicit reconstruction formulas are provided, inspired by the duality concept from frame theory. All this gives rise to the possibility of quantifying the invertibility of a ReLU-layer and a concrete reconstruction algorithm for any input vector on the ball.
本文利用框架理论研究了一个relu层在$mathbb{R}^n$闭球上的注入性及其非负部分。特别强调了球半径和偏置矢量之间的相互作用。从凸几何的角度来看,这导致了一种计算上可行的方法来验证在合理的限制下,根据偏置向量的上界relu层的注入性。受框架理论中对偶概念的启发,给出了明确的重构公式。所有这些都为量化relu层的可逆性提供了可能,并为球上任何输入向量提供了具体的重建算法。
{"title":"Convex Geometry of ReLU-layers, Injectivity on the Ball and Local Reconstruction","authors":"Daniel Haider, M. Ehler, P. Balázs","doi":"10.48550/arXiv.2307.09672","DOIUrl":"https://doi.org/10.48550/arXiv.2307.09672","url":null,"abstract":"The paper uses a frame-theoretic setting to study the injectivity of a ReLU-layer on the closed ball of $mathbb{R}^n$ and its non-negative part. In particular, the interplay between the radius of the ball and the bias vector is emphasized. Together with a perspective from convex geometry, this leads to a computationally feasible method of verifying the injectivity of a ReLU-layer under reasonable restrictions in terms of an upper bound of the bias vector. Explicit reconstruction formulas are provided, inspired by the duality concept from frame theory. All this gives rise to the possibility of quantifying the invertibility of a ReLU-layer and a concrete reconstruction algorithm for any input vector on the ball.","PeriodicalId":74529,"journal":{"name":"Proceedings of the ... International Conference on Machine Learning. International Conference on Machine Learning","volume":"24 1","pages":"12339-12350"},"PeriodicalIF":0.0,"publicationDate":"2023-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"83476070","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Proceedings of the ... International Conference on Machine Learning. International Conference on Machine Learning
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1