首页 > 最新文献

Proceedings of machine learning research最新文献

英文 中文
On Distribution Dependent Sub-Logarithmic Query Time of Learned Indexing. 学习型索引的分布依赖次对数查询时间。
Sepanta Zeighami, Cyrus Shahabi

A fundamental problem in data management is to find the elements in an array that match a query. Recently, learned indexes are being extensively used to solve this problem, where they learn a model to predict the location of the items in the array. They are empirically shown to outperform non-learned methods (e.g., B-trees or binary search that answer queries in O(logn) time) by orders of magnitude. However, success of learned indexes has not been theoretically justified. Only existing attempt shows the same query time of O(logn), but with a constant factor improvement in space complexity over non-learned methods, under some assumptions on data distribution. In this paper, we significantly strengthen this result, showing that under mild assumptions on data distribution, and the same space complexity as non-learned methods, learned indexes can answer queries in O(loglogn) expected query time. We also show that allowing for slightly larger but still near-linear space overhead, a learned index can achieve O(1) expected query time. Our results theoretically prove learned indexes are orders of magnitude faster than non-learned methods, theoretically grounding their empirical success.

数据管理中的一个基本问题是在数组中查找与查询匹配的元素。最近,学习的索引被广泛用于解决这个问题,它们学习一个模型来预测数组中项目的位置。经验表明,它们在数量级上优于非学习方法(例如,在O(logn)时间内回答查询的B树或二进制搜索)。然而,学习指数的成功并没有得到理论上的证明。只有现有的尝试显示了相同的查询时间O(logn),但在数据分布的一些假设下,与非学习方法相比,空间复杂性不断提高。在本文中,我们显著地加强了这一结果,表明在对数据分布的温和假设下,以及与非学习方法相同的空间复杂性下,学习索引可以在O(loglogn)预期查询时间内回答查询。我们还表明,考虑到稍大但仍接近线性的空间开销,学习的索引可以实现O(1)的预期查询时间。我们的结果从理论上证明了学习指数比非学习方法快几个数量级,这在理论上奠定了它们的经验成功基础。
{"title":"On Distribution Dependent Sub-Logarithmic Query Time of Learned Indexing.","authors":"Sepanta Zeighami, Cyrus Shahabi","doi":"","DOIUrl":"","url":null,"abstract":"<p><p>A fundamental problem in data management is to find the elements in an array that match a query. Recently, learned indexes are being extensively used to solve this problem, where they learn a model to predict the location of the items in the array. They are empirically shown to outperform non-learned methods (e.g., B-trees or binary search that answer queries in <math><mi>O</mi><mo>(</mo><mi>l</mi><mi>o</mi><mi>g</mi><mspace></mspace><mi>n</mi><mo>)</mo></math> time) by orders of magnitude. However, success of learned indexes has not been theoretically justified. Only existing attempt shows the same query time of <math><mi>O</mi><mo>(</mo><mi>l</mi><mi>o</mi><mi>g</mi><mspace></mspace><mi>n</mi><mo>)</mo></math>, but with a constant factor improvement in space complexity over non-learned methods, under some assumptions on data distribution. In this paper, we significantly strengthen this result, showing that under mild assumptions on data distribution, and the same space complexity as non-learned methods, learned indexes can answer queries in <math><mi>O</mi><mo>(</mo><mi>l</mi><mi>o</mi><mi>g</mi><mi>l</mi><mi>o</mi><mi>g</mi><mspace></mspace><mi>n</mi><mo>)</mo></math> expected query time. We also show that allowing for slightly larger but still near-linear space overhead, a learned index can achieve <math><mi>O</mi><mo>(</mo><mn>1</mn><mo>)</mo></math> expected query time. Our results theoretically prove learned indexes are orders of magnitude faster than non-learned methods, theoretically grounding their empirical success.</p>","PeriodicalId":74504,"journal":{"name":"Proceedings of machine learning research","volume":"202 ","pages":"40669-40680"},"PeriodicalIF":0.0,"publicationDate":"2023-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10627073/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"71489774","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Estimating Causal Effects using a Multi-task Deep Ensemble. 使用多任务深度集合估计因果效应。
Ziyang Jiang, Zhuoran Hou, Yiling Liu, Yiman Ren, Keyu Li, David Carlson

A number of methods have been proposed for causal effect estimation, yet few have demonstrated efficacy in handling data with complex structures, such as images. To fill this gap, we propose Causal Multi-task Deep Ensemble (CMDE), a novel framework that learns both shared and group-specific information from the study population. We provide proofs demonstrating equivalency of CDME to a multi-task Gaussian process (GP) with a coregionalization kernel a priori. Compared to multi-task GP, CMDE efficiently handles high-dimensional and multi-modal covariates and provides pointwise uncertainty estimates of causal effects. We evaluate our method across various types of datasets and tasks and find that CMDE outperforms state-of-the-art methods on a majority of these tasks.

针对因果效应估计提出了很多方法,但很少有方法能有效处理图像等结构复杂的数据。为了填补这一空白,我们提出了因果多任务深度集合(CMDE),这是一种新颖的框架,可以从研究人群中学习共享信息和特定群体信息。我们提供了证明,证明 CDME 等同于带有先验核心区域化内核的多任务高斯过程(GP)。与多任务 GP 相比,CMDE 能有效处理高维和多模态协变量,并提供因果效应的点式不确定性估计。我们在各种类型的数据集和任务中对我们的方法进行了评估,发现 CMDE 在大多数任务中的表现都优于最先进的方法。
{"title":"Estimating Causal Effects using a Multi-task Deep Ensemble.","authors":"Ziyang Jiang, Zhuoran Hou, Yiling Liu, Yiman Ren, Keyu Li, David Carlson","doi":"","DOIUrl":"","url":null,"abstract":"<p><p>A number of methods have been proposed for causal effect estimation, yet few have demonstrated efficacy in handling data with complex structures, such as images. To fill this gap, we propose Causal Multi-task Deep Ensemble (CMDE), a novel framework that learns both shared and group-specific information from the study population. We provide proofs demonstrating equivalency of CDME to a multi-task Gaussian process (GP) with a coregionalization kernel <i>a priori</i>. Compared to multi-task GP, CMDE efficiently handles high-dimensional and multi-modal covariates and provides pointwise uncertainty estimates of causal effects. We evaluate our method across various types of datasets and tasks and find that CMDE outperforms state-of-the-art methods on a majority of these tasks.</p>","PeriodicalId":74504,"journal":{"name":"Proceedings of machine learning research","volume":"202 ","pages":"15023-15040"},"PeriodicalIF":0.0,"publicationDate":"2023-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10759931/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139089657","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Actor-Critic Alignment for Offline-to-Online Reinforcement Learning. 离线到在线强化学习的演员-评论家对齐。
Zishun Yu, Xinhua Zhang

Deep offline reinforcement learning has recently demonstrated considerable promises in leveraging offline datasets, providing high-quality models that significantly reduce the online interactions required for fine-tuning. However, such a benefit is often diminished due to the marked state-action distribution shift, which causes significant bootstrap error and wipes out the good initial policy Existing solutions resort to constraining the policy shift or balancing the sample replay based on their online-ness. However, they require online estimation of distribution divergence or density ratio. To avoid such complications, we propose deviating from existing actor-critic approaches that directly transfer the state-action value functions. Instead, we post-process them by aligning with the offline learned policy, so that the Q -values for actions outside the offline policy are also tamed. As a result, the online fine-tuning can be simply performed as in the standard actor-critic algorithms. We show empirically that the proposed method improves the performance of the fine-tuned robotic agents on various simulated tasks.

深度离线强化学习最近在利用离线数据集方面展现出了巨大的前景,它提供了高质量的模型,大大减少了微调所需的在线交互。然而,这种优势往往会因为明显的状态-行动分布偏移而被削弱,因为这种偏移会导致显著的引导误差,并抹去良好的初始策略。但是,它们需要在线估计分布发散或密度比。为了避免这种复杂性,我们提出偏离现有的直接转移状态-行动值函数的行动者批评方法。取而代之的是,我们通过与离线学习的策略保持一致来对其进行后处理,这样离线策略之外的行动 Q 值也会被驯服。因此,在线微调可以像标准演员批评算法一样简单地执行。我们通过实验证明,所提出的方法提高了经过微调的机器人代理在各种模拟任务中的性能。
{"title":"Actor-Critic Alignment for Offline-to-Online Reinforcement Learning.","authors":"Zishun Yu, Xinhua Zhang","doi":"","DOIUrl":"","url":null,"abstract":"<p><p>Deep offline reinforcement learning has recently demonstrated considerable promises in leveraging offline datasets, providing high-quality models that significantly reduce the online interactions required for fine-tuning. However, such a benefit is often diminished due to the marked state-action distribution shift, which causes significant bootstrap error and wipes out the good initial policy Existing solutions resort to constraining the policy shift or balancing the sample replay based on their online-ness. However, they require online estimation of distribution divergence or density ratio. To avoid such complications, we propose deviating from existing actor-critic approaches that directly transfer the state-action value functions. Instead, we post-process them by aligning with the offline learned policy, so that the <math><mi>Q</mi></math> -values for actions outside the offline policy are also tamed. As a result, the online fine-tuning can be simply performed as in the standard actor-critic algorithms. We show empirically that the proposed method improves the performance of the fine-tuned robotic agents on various simulated tasks.</p>","PeriodicalId":74504,"journal":{"name":"Proceedings of machine learning research","volume":"202 ","pages":"40452-40474"},"PeriodicalIF":0.0,"publicationDate":"2023-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11232493/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141565256","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
The Unintended Consequences of Discount Regularization: Improving Regularization in Certainty Equivalence Reinforcement Learning 折扣正则化的意外后果:改进确定性等价强化学习中的正则化
Pub Date : 2023-06-20 DOI: 10.48550/arXiv.2306.11208
Sarah Rathnam, S. Parbhoo, Weiwei Pan, Susan A. Murphy, F. Doshi-Velez
Discount regularization, using a shorter planning horizon when calculating the optimal policy, is a popular choice to restrict planning to a less complex set of policies when estimating an MDP from sparse or noisy data (Jiang et al., 2015). It is commonly understood that discount regularization functions by de-emphasizing or ignoring delayed effects. In this paper, we reveal an alternate view of discount regularization that exposes unintended consequences. We demonstrate that planning under a lower discount factor produces an identical optimal policy to planning using any prior on the transition matrix that has the same distribution for all states and actions. In fact, it functions like a prior with stronger regularization on state-action pairs with more transition data. This leads to poor performance when the transition matrix is estimated from data sets with uneven amounts of data across state-action pairs. Our equivalence theorem leads to an explicit formula to set regularization parameters locally for individual state-action pairs rather than globally. We demonstrate the failures of discount regularization and how we remedy them using our state-action-specific method across simple empirical examples as well as a medical cancer simulator.
折扣正则化,在计算最优策略时使用更短的规划范围,是一种流行的选择,当从稀疏或噪声数据估计MDP时,将规划限制在一组不太复杂的策略上(Jiang等人,2015)。人们通常认为,贴现正则化函数是通过不强调或忽略延迟效应来实现的。在本文中,我们揭示了折扣正则化的另一种观点,它暴露了意想不到的后果。我们证明了在较低折扣因子下的规划与在所有状态和动作具有相同分布的转移矩阵上使用任何先验的规划产生相同的最优策略。事实上,它的功能就像一个具有更强正则化的先验,对具有更多转移数据的状态-动作对。当从跨状态-动作对的数据量不均匀的数据集估计转移矩阵时,这会导致性能差。我们的等价定理给出了一个显式公式,可以为单个状态-动作对局部设置正则化参数,而不是全局设置。我们演示了折扣正则化的失败,以及我们如何通过简单的经验示例以及医疗癌症模拟器使用我们的状态-行动特定方法来补救它们。
{"title":"The Unintended Consequences of Discount Regularization: Improving Regularization in Certainty Equivalence Reinforcement Learning","authors":"Sarah Rathnam, S. Parbhoo, Weiwei Pan, Susan A. Murphy, F. Doshi-Velez","doi":"10.48550/arXiv.2306.11208","DOIUrl":"https://doi.org/10.48550/arXiv.2306.11208","url":null,"abstract":"Discount regularization, using a shorter planning horizon when calculating the optimal policy, is a popular choice to restrict planning to a less complex set of policies when estimating an MDP from sparse or noisy data (Jiang et al., 2015). It is commonly understood that discount regularization functions by de-emphasizing or ignoring delayed effects. In this paper, we reveal an alternate view of discount regularization that exposes unintended consequences. We demonstrate that planning under a lower discount factor produces an identical optimal policy to planning using any prior on the transition matrix that has the same distribution for all states and actions. In fact, it functions like a prior with stronger regularization on state-action pairs with more transition data. This leads to poor performance when the transition matrix is estimated from data sets with uneven amounts of data across state-action pairs. Our equivalence theorem leads to an explicit formula to set regularization parameters locally for individual state-action pairs rather than globally. We demonstrate the failures of discount regularization and how we remedy them using our state-action-specific method across simple empirical examples as well as a medical cancer simulator.","PeriodicalId":74504,"journal":{"name":"Proceedings of machine learning research","volume":"202 1","pages":"28746-28767"},"PeriodicalIF":0.0,"publicationDate":"2023-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"47482371","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
On Distribution Dependent Sub-Logarithmic Query Time of Learned Indexing 学习索引的分布相关次对数查询时间
Pub Date : 2023-06-19 DOI: 10.48550/arXiv.2306.10651
Sepanta Zeighami, C. Shahabi
A fundamental problem in data management is to find the elements in an array that match a query. Recently, learned indexes are being extensively used to solve this problem, where they learn a model to predict the location of the items in the array. They are empirically shown to outperform non-learned methods (e.g., B-trees or binary search that answer queries in O(logn) time) by orders of magnitude. However, success of learned indexes has not been theoretically justified. Only existing attempt shows the same query time of O(logn), but with a constant factor improvement in space complexity over non-learned methods, under some assumptions on data distribution. In this paper, we significantly strengthen this result, showing that under mild assumptions on data distribution, and the same space complexity as non-learned methods, learned indexes can answer queries in O(loglogn) expected query time. We also show that allowing for slightly larger but still near-linear space overhead, a learned index can achieve O(1) expected query time. Our results theoretically prove learned indexes are orders of magnitude faster than non-learned methods, theoretically grounding their empirical success.
数据管理中的一个基本问题是在数组中找到与查询匹配的元素。最近,学习索引被广泛用于解决这个问题,它们学习一个模型来预测数组中项的位置。经验表明,它们比非学习方法(例如,在O(logn)时间内回答查询的b树或二叉搜索)的性能要好几个数量级。然而,学习指标的成功并没有从理论上得到证明。只有现有的尝试显示相同的查询时间为O(logn),但在数据分布的某些假设下,空间复杂度比非学习方法有恒定的提高。在本文中,我们显著加强了这一结果,表明在对数据分布的温和假设下,在与非学习方法相同的空间复杂度下,学习索引可以在O(loglog)期望查询时间内回答查询。我们还表明,考虑到稍大但仍然接近线性的空间开销,学习索引可以实现O(1)预期查询时间。我们的结果从理论上证明了学习索引比非学习方法快几个数量级,理论上为他们的经验成功奠定了基础。
{"title":"On Distribution Dependent Sub-Logarithmic Query Time of Learned Indexing","authors":"Sepanta Zeighami, C. Shahabi","doi":"10.48550/arXiv.2306.10651","DOIUrl":"https://doi.org/10.48550/arXiv.2306.10651","url":null,"abstract":"A fundamental problem in data management is to find the elements in an array that match a query. Recently, learned indexes are being extensively used to solve this problem, where they learn a model to predict the location of the items in the array. They are empirically shown to outperform non-learned methods (e.g., B-trees or binary search that answer queries in O(logn) time) by orders of magnitude. However, success of learned indexes has not been theoretically justified. Only existing attempt shows the same query time of O(logn), but with a constant factor improvement in space complexity over non-learned methods, under some assumptions on data distribution. In this paper, we significantly strengthen this result, showing that under mild assumptions on data distribution, and the same space complexity as non-learned methods, learned indexes can answer queries in O(loglogn) expected query time. We also show that allowing for slightly larger but still near-linear space overhead, a learned index can achieve O(1) expected query time. Our results theoretically prove learned indexes are orders of magnitude faster than non-learned methods, theoretically grounding their empirical success.","PeriodicalId":74504,"journal":{"name":"Proceedings of machine learning research","volume":"261 1","pages":"40669-40680"},"PeriodicalIF":0.0,"publicationDate":"2023-06-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"79545250","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Assessing the Impact of Context Inference Error and Partial Observability on RL Methods for Just-In-Time Adaptive Interventions 评估上下文推断误差和部分可观察性对实时自适应干预RL方法的影响
Pub Date : 2023-05-17 DOI: 10.48550/arXiv.2305.09913
Karine Karine, P. Klasnja, Susan A. Murphy, Benjamin M Marlin
Just-in-Time Adaptive Interventions (JITAIs) are a class of personalized health interventions developed within the behavioral science community. JITAIs aim to provide the right type and amount of support by iteratively selecting a sequence of intervention options from a pre-defined set of components in response to each individual's time varying state. In this work, we explore the application of reinforcement learning methods to the problem of learning intervention option selection policies. We study the effect of context inference error and partial observability on the ability to learn effective policies. Our results show that the propagation of uncertainty from context inferences is critical to improving intervention efficacy as context uncertainty increases, while policy gradient algorithms can provide remarkable robustness to partially observed behavioral state information.
实时适应性干预(JITAIs)是行为科学界开发的一类个性化健康干预措施。JITAI旨在通过从预定义的一组组件中迭代选择一系列干预选项来提供正确类型和数量的支持,以响应每个个体的时变状态。在这项工作中,我们探索了强化学习方法在学习干预选项选择策略问题中的应用。我们研究了上下文推理误差和部分可观察性对学习有效策略能力的影响。我们的结果表明,随着上下文不确定性的增加,上下文推断的不确定性的传播对于提高干预效果至关重要,而策略梯度算法可以对部分观察到的行为状态信息提供显著的鲁棒性。
{"title":"Assessing the Impact of Context Inference Error and Partial Observability on RL Methods for Just-In-Time Adaptive Interventions","authors":"Karine Karine, P. Klasnja, Susan A. Murphy, Benjamin M Marlin","doi":"10.48550/arXiv.2305.09913","DOIUrl":"https://doi.org/10.48550/arXiv.2305.09913","url":null,"abstract":"Just-in-Time Adaptive Interventions (JITAIs) are a class of personalized health interventions developed within the behavioral science community. JITAIs aim to provide the right type and amount of support by iteratively selecting a sequence of intervention options from a pre-defined set of components in response to each individual's time varying state. In this work, we explore the application of reinforcement learning methods to the problem of learning intervention option selection policies. We study the effect of context inference error and partial observability on the ability to learn effective policies. Our results show that the propagation of uncertainty from context inferences is critical to improving intervention efficacy as context uncertainty increases, while policy gradient algorithms can provide remarkable robustness to partially observed behavioral state information.","PeriodicalId":74504,"journal":{"name":"Proceedings of machine learning research","volume":"216 1","pages":"1047-1057"},"PeriodicalIF":0.0,"publicationDate":"2023-05-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"46832340","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
DIET: Conditional independence testing with marginal dependence measures of residual information. DIET:利用残差信息的边际依赖性测量进行条件独立性检验。
Mukund Sudarshan, Aahlad Puli, Wesley Tansey, Rajesh Ranganath

Conditional randomization tests (CRTs) assess whether a variable x is predictive of another variable y, having observed covariates z. CRTs require fitting a large number of predictive models, which is often computationally intractable. Existing solutions to reduce the cost of CRTs typically split the dataset into a train and test portion, or rely on heuristics for interactions, both of which lead to a loss in power. We propose the decoupled independence test (DIET), an algorithm that avoids both of these issues by leveraging marginal independence statistics to test conditional independence relationships. DIET tests the marginal independence of two random variables: Fxz(xz) and Fyz(yz) where Fz(z) is a conditional cumulative distribution function (CDF) for the distribution p(z). These variables are termed "information residuals." We give sufficient conditions for DIET to achieve finite sample type-1 error control and power greater than the type-1 error rate. We then prove that when using the mutual information between the information residuals as a test statistic, DIET yields the most powerful conditionally valid test. Finally, we show DIET achieves higher power than other tractable CRTs on several synthetic and real benchmarks.

条件随机化检验(CRTs)评估的是一个变量 x 是否能预测另一个变量 y 以及观察到的协变量 z。CRT 需要拟合大量的预测模型,这在计算上往往难以实现。现有的降低 CRT 成本的解决方案通常是将数据集分成训练和测试两部分,或依赖启发式方法进行交互,这两种方法都会导致预测能力下降。我们提出的解耦独立性测试(DIET)算法利用边际独立性统计来测试条件独立性关系,从而避免了上述两个问题。DIET 测试两个随机变量的边际独立性:Fx∣z(x∣z)和 Fy∣z(y∣z),其中 F∣z(⋅∣z)是分布 p(⋅∣z)的条件累积分布函数(CDF)。这些变量被称为 "信息残差"。我们给出了 DIET 实现有限样本类型-1 错误控制和功率大于类型-1 错误率的充分条件。然后,我们证明了当使用信息残差之间的互信息作为检验统计量时,DIET 会产生最强大的条件有效检验。最后,我们展示了 DIET 在几个合成和真实基准上比其他可行的 CRT 获得了更高的功率。
{"title":"DIET: Conditional independence testing with marginal dependence measures of residual information.","authors":"Mukund Sudarshan, Aahlad Puli, Wesley Tansey, Rajesh Ranganath","doi":"","DOIUrl":"","url":null,"abstract":"<p><p>Conditional randomization tests (CRTs) assess whether a variable <math><mi>x</mi></math> is predictive of another variable <math><mi>y</mi></math>, having observed covariates <math><mrow><mi>z</mi></mrow></math>. CRTs require fitting a large number of predictive models, which is often computationally intractable. Existing solutions to reduce the cost of CRTs typically split the dataset into a train and test portion, or rely on heuristics for interactions, both of which lead to a loss in power. We propose the decoupled independence test (DIET), an algorithm that avoids both of these issues by leveraging marginal independence statistics to test conditional independence relationships. DIET tests the marginal independence of two random variables: <math><mrow><msub><mi>F</mi><mrow><mi>x</mi><mo>∣</mo><mi>z</mi></mrow></msub><mo>(</mo><mi>x</mi><mo>∣</mo><mi>z</mi><mo>)</mo></mrow></math> and <math><mrow><msub><mi>F</mi><mrow><mi>y</mi><mo>∣</mo><mi>z</mi></mrow></msub><mo>(</mo><mi>y</mi><mo>∣</mo><mi>z</mi><mo>)</mo></mrow></math> where <math><mrow><msub><mi>F</mi><mrow><mo>⋅</mo><mo>∣</mo><mi>z</mi></mrow></msub><mo>(</mo><mo>⋅</mo><mo>∣</mo><mi>z</mi><mo>)</mo></mrow></math> is a conditional cumulative distribution function (CDF) for the distribution <math><mrow><mi>p</mi><mo>(</mo><mo>⋅</mo><mo>∣</mo><mi>z</mi><mo>)</mo></mrow></math>. These variables are termed \"information residuals.\" We give sufficient conditions for DIET to achieve finite sample type-1 error control and power greater than the type-1 error rate. We then prove that when using the mutual information between the information residuals as a test statistic, DIET yields the most powerful conditionally valid test. Finally, we show DIET achieves higher power than other tractable CRTs on several synthetic and real benchmarks.</p>","PeriodicalId":74504,"journal":{"name":"Proceedings of machine learning research","volume":"206 ","pages":"10343-10367"},"PeriodicalIF":0.0,"publicationDate":"2023-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10484293/pdf/nihms-1899844.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"10577745","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Covariate-informed Representation Learning to Prevent Posterior Collapse of iVAE. 协变量知情表示学习预防iVAE后塌陷。
Young-Geun Kim, Ying Liu, Xue-Xin Wei

The recently proposed identifiable variational autoencoder (iVAE) framework provides a promising approach for learning latent independent components (ICs). iVAEs use auxiliary covariates to build an identifiable generation structure from covariates to ICs to observations, and the posterior network approximates ICs given observations and covariates. Though the identifiability is appealing, we show that iVAEs could have local minimum solution where observations and the approximated ICs are independent given covariates.-a phenomenon we referred to as the posterior collapse problem of iVAEs. To overcome this problem, we develop a new approach, covariate-informed iVAE (CI-iVAE) by considering a mixture of encoder and posterior distributions in the objective function. In doing so, the objective function prevents the posterior collapse, resulting latent representations that contain more information of the observations. Furthermore, CI-iVAE extends the original iVAE objective function to a larger class and finds the optimal one among them, thus having tighter evidence lower bounds than the original iVAE. Experiments on simulation datasets, EMNIST, Fashion-MNIST, and a large-scale brain imaging dataset demonstrate the effectiveness of our new method.

最近提出的可识别变分自编码器(iVAE)框架为学习潜在独立分量(ic)提供了一种很有前途的方法。ivae使用辅助协变量构建从协变量到ic到观测值的可识别生成结构,后验网络近似给定观测值和协变量的ic。虽然可辨识性很吸引人,但我们表明ivae可以具有局部最小解,其中观测值和近似的ic是独立的给定协变量。我们将这种现象称为ivae后塌陷问题。为了克服这个问题,我们开发了一种新的方法,协变量通知iVAE (CI-iVAE),通过考虑目标函数中编码器和后验分布的混合。这样做,目标函数可以防止后验崩溃,从而产生包含更多观察信息的潜在表征。此外,CI-iVAE将原有的iVAE目标函数扩展到更大的类中,并从中找到最优的一类,从而比原有的iVAE具有更严格的证据下界。在仿真数据集、EMNIST、Fashion-MNIST和大规模脑成像数据集上的实验证明了该方法的有效性。
{"title":"Covariate-informed Representation Learning to Prevent Posterior Collapse of iVAE.","authors":"Young-Geun Kim,&nbsp;Ying Liu,&nbsp;Xue-Xin Wei","doi":"","DOIUrl":"","url":null,"abstract":"<p><p>The recently proposed identifiable variational autoencoder (iVAE) framework provides a promising approach for learning latent independent components (ICs). iVAEs use auxiliary covariates to build an identifiable generation structure from covariates to ICs to observations, and the posterior network approximates ICs given observations and covariates. Though the identifiability is appealing, we show that iVAEs could have local minimum solution where observations and the approximated ICs are independent given covariates.-a phenomenon we referred to as the posterior collapse problem of iVAEs. To overcome this problem, we develop a new approach, covariate-informed iVAE (CI-iVAE) by considering a mixture of encoder and posterior distributions in the objective function. In doing so, the objective function prevents the posterior collapse, resulting latent representations that contain more information of the observations. Furthermore, CI-iVAE extends the original iVAE objective function to a larger class and finds the optimal one among them, thus having tighter evidence lower bounds than the original iVAE. Experiments on simulation datasets, EMNIST, Fashion-MNIST, and a large-scale brain imaging dataset demonstrate the effectiveness of our new method.</p>","PeriodicalId":74504,"journal":{"name":"Proceedings of machine learning research","volume":"206 ","pages":"2641-2660"},"PeriodicalIF":0.0,"publicationDate":"2023-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10226378/pdf/nihms-1902106.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9908011","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Continuous-Time Decision Transformer for Healthcare Applications. 用于医疗保健应用的连续时间决策变压器。
Zhiyue Zhang, Hongyuan Mei, Yanxun Xu

Offline reinforcement learning (RL) is a promising approach for training intelligent medical agents to learn treatment policies and assist decision making in many healthcare applications, such as scheduling clinical visits and assigning dosages for patients with chronic conditions. In this paper, we investigate the potential usefulness of Decision Transformer (Chen et al., 2021)-a new offline RL paradigm-in medical domains where decision making in continuous time is desired. As Decision Transformer only handles discrete-time (or turn-based) sequential decision making scenarios, we generalize it to Continuous-Time Decision Transformer that not only considers the past clinical measurements and treatments but also the timings of previous visits, and learns to suggest the timings of future visits as well as the treatment plan at each visit. Extensive experiments on synthetic datasets and simulators motivated by real-world medical applications demonstrate that Continuous-Time Decision Transformer is able to outperform competitors and has clinical utility in terms of improving patients' health and prolonging their survival by learning high-performance policies from logged data generated using policies of different levels of quality.

离线强化学习(RL)是一种很有前途的方法,可用于训练智能医疗代理学习治疗策略,并在许多医疗保健应用中协助决策制定,例如为慢性病患者安排门诊和分配剂量。在本文中,我们研究了决策转换器(Chen 等人,2021 年)--一种新的离线 RL 范例--在需要连续时间决策的医疗领域中的潜在用途。由于决策转换器只能处理离散时间(或基于回合的)顺序决策场景,我们将其推广到连续时间决策转换器,它不仅考虑了过去的临床测量和治疗,还考虑了以前就诊的时间,并学会建议未来就诊的时间以及每次就诊的治疗方案。在合成数据集和现实世界医疗应用模拟器上进行的大量实验表明,连续时间决策转换器能够超越竞争对手,并通过从使用不同质量水平的策略生成的日志数据中学习高性能策略,在改善患者健康状况和延长患者生存期方面具有临床实用性。
{"title":"Continuous-Time Decision Transformer for Healthcare Applications.","authors":"Zhiyue Zhang, Hongyuan Mei, Yanxun Xu","doi":"","DOIUrl":"","url":null,"abstract":"<p><p>Offline reinforcement learning (RL) is a promising approach for training intelligent medical agents to learn treatment policies and assist decision making in many healthcare applications, such as scheduling clinical visits and assigning dosages for patients with chronic conditions. In this paper, we investigate the potential usefulness of Decision Transformer (Chen et al., 2021)-a new offline RL paradigm-in medical domains where decision making in continuous time is desired. As Decision Transformer only handles discrete-time (or turn-based) sequential decision making scenarios, we generalize it to Continuous-Time Decision Transformer that not only considers the past clinical measurements and treatments but also the timings of previous visits, and learns to suggest the timings of future visits as well as the treatment plan at each visit. Extensive experiments on synthetic datasets and simulators motivated by real-world medical applications demonstrate that Continuous-Time Decision Transformer is able to outperform competitors and has clinical utility in terms of improving patients' health and prolonging their survival by learning high-performance policies from logged data generated using policies of different levels of quality.</p>","PeriodicalId":74504,"journal":{"name":"Proceedings of machine learning research","volume":"206 ","pages":"6245-6262"},"PeriodicalIF":0.0,"publicationDate":"2023-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10907982/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140023967","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Causal Learning through Deliberate Undersampling. 通过故意减少取样进行因果学习
Kseniya Solovyeva, David Danks, Mohammadsajad Abavisani, Sergey Plis

Domain scientists interested in causal mechanisms are usually limited by the frequency at which they can collect the measurements of social, physical, or biological systems. A common and plausible assumption is that higher measurement frequencies are the only way to gain more informative data about the underlying dynamical causal structure. This assumption is a strong driver for designing new, faster instruments, but such instruments might not be feasible or even possible. In this paper, we show that this assumption is incorrect: there are situations in which we can gain additional information about the causal structure by measuring more slowly than our current instruments. We present an algorithm that uses graphs at multiple measurement timescales to infer underlying causal structure, and show that inclusion of structures at slower timescales can nonetheless reduce the size of the equivalence class of possible causal structures. We provide simulation data about the probability of cases in which deliberate undersampling yields a gain, as well as the size of this gain.

对因果机制感兴趣的领域科学家通常受限于他们收集社会、物理或生物系统测量数据的频率。一个常见且合理的假设是,只有更高的测量频率才能获得更多关于底层动态因果结构的信息数据。这一假设是设计更快的新型仪器的强大动力,但这种仪器可能并不可行,甚至不可能实现。在本文中,我们证明了这一假设是错误的:在某些情况下,我们可以通过比现有仪器更慢的测量速度来获得更多关于因果结构的信息。我们提出了一种算法,利用多个测量时间尺度上的图形来推断潜在的因果结构,并证明将较慢时间尺度上的结构包含在内仍然可以减少可能的因果结构等价类的大小。我们提供了关于刻意减少取样而产生增益的概率以及增益大小的模拟数据。
{"title":"Causal Learning through Deliberate Undersampling.","authors":"Kseniya Solovyeva, David Danks, Mohammadsajad Abavisani, Sergey Plis","doi":"","DOIUrl":"","url":null,"abstract":"<p><p>Domain scientists interested in causal mechanisms are usually limited by the frequency at which they can collect the measurements of social, physical, or biological systems. A common and plausible assumption is that higher measurement frequencies are the only way to gain more informative data about the underlying dynamical causal structure. This assumption is a strong driver for designing new, faster instruments, but such instruments might not be feasible or even possible. In this paper, we show that this assumption is incorrect: there are situations in which we can gain additional information about the causal structure by measuring more <i>slowly</i> than our current instruments. We present an algorithm that uses graphs at multiple measurement timescales to infer underlying causal structure, and show that inclusion of structures at slower timescales can nonetheless reduce the size of the equivalence class of possible causal structures. We provide simulation data about the probability of cases in which deliberate undersampling yields a gain, as well as the size of this gain.</p>","PeriodicalId":74504,"journal":{"name":"Proceedings of machine learning research","volume":"213 ","pages":"518-530"},"PeriodicalIF":0.0,"publicationDate":"2023-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10972601/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140308206","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Proceedings of machine learning research
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1