首页 > 最新文献

JMLR workshop and conference proceedings最新文献

英文 中文
Scalable Convex Multiple Sequence Alignment via Entropy-Regularized Dual Decomposition. 基于熵正则化对偶分解的可伸缩凸多序列对齐。
Jiong Zhang, Ian E H Yen, Pradeep Ravikumar, Inderjit S Dhillon

Multiple Sequence Alignment (MSA) is one of the fundamental tasks in biological sequence analysis that underlies applications such as phylogenetic trees, profiles, and structure prediction. The task, however, is NP-hard, and the current practice resorts to heuristic and local-search methods. Recently, a convex optimization approach for MSA was proposed based on the concept of atomic norm [23], which demonstrates significant improvement over existing methods in the quality of alignments. However, the convex program is challenging to solve due to the constraint given by the intersection of two atomic-norm balls, for which the existing algorithm can only handle sequences of length up to 50, with an iteration complexity subject to constants of unknown relation to the natural parameters of MSA. In this work, we propose an accelerated dual decomposition algorithm that exploits entropy regularization to induce closed-form solutions for each atomic-norm-constrained subproblem, giving a single-loop algorithm of iteration complexity linear to the problem size (total length of all sequences). The proposed algorithm gives significantly better alignments than existing methods on sequences of length up to hundreds, where the existing convex programming method fails to converge in one day.

多序列比对(MSA)是生物序列分析的基础任务之一,是系统发育树、基因图谱和结构预测等应用的基础。然而,这项任务是np困难的,目前的实践采用启发式和局部搜索方法。最近,一种基于原子范数概念的MSA凸优化方法被提出[23],该方法在对齐质量上比现有方法有了显著提高。然而,由于两个原子范数球相交的约束,现有算法只能处理长度不超过50的序列,且迭代复杂度受与MSA自然参数关系未知的常数的影响,使得凸规划的求解具有挑战性。在这项工作中,我们提出了一种加速的对偶分解算法,该算法利用熵正则化来诱导每个原子规范约束子问题的封闭形式解,给出了迭代复杂度与问题大小(所有序列的总长度)线性的单环算法。对于长度达数百的序列,该算法比现有的凸规划方法在一天内不能收敛的情况下具有更好的对齐效果。
{"title":"Scalable Convex Multiple Sequence Alignment via Entropy-Regularized Dual Decomposition.","authors":"Jiong Zhang,&nbsp;Ian E H Yen,&nbsp;Pradeep Ravikumar,&nbsp;Inderjit S Dhillon","doi":"","DOIUrl":"","url":null,"abstract":"<p><p><i>Multiple Sequence Alignment (MSA)</i> is one of the fundamental tasks in biological sequence analysis that underlies applications such as phylogenetic trees, profiles, and structure prediction. The task, however, is NP-hard, and the current practice resorts to heuristic and local-search methods. Recently, a convex optimization approach for MSA was proposed based on the concept of atomic norm [23], which demonstrates significant improvement over existing methods in the quality of alignments. However, the convex program is challenging to solve due to the constraint given by the intersection of two atomic-norm balls, for which the existing algorithm can only handle sequences of length up to 50, with an iteration complexity subject to constants of unknown relation to the natural parameters of MSA. In this work, we propose an <i>accelerated dual decomposition</i> algorithm that exploits <i>entropy regularization</i> to induce closed-form solutions for each atomic-norm-constrained subproblem, giving a single-loop algorithm of iteration complexity linear to the problem size (total length of all sequences). The proposed algorithm gives significantly better alignments than existing methods on sequences of length up to hundreds, where the existing convex programming method fails to converge in one day.</p>","PeriodicalId":89793,"journal":{"name":"JMLR workshop and conference proceedings","volume":"54 ","pages":"1514-1522"},"PeriodicalIF":0.0,"publicationDate":"2017-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5581665/pdf/nihms896524.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"35472764","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Greedy Direction Method of Multiplier for MAP Inference of Large Output Domain. 大输出域MAP推理的乘法器贪心方向法。
Xiangru Huang, Qixing Huang, Ian E H Yen, Pradeep Ravikumar, Ruohan Zhang, Inderjit S Dhillon

Maximum-a-Posteriori (MAP) inference lies at the heart of Graphical Models and Structured Prediction. Despite the intractability of exact MAP inference, approximate methods based on LP relaxations have exhibited superior performance across a wide range of applications. Yet for problems involving large output domains (i.e., the state space for each variable is large), standard LP relaxations can easily give rise to a large number of variables and constraints which are beyond the limit of existing optimization algorithms. In this paper, we introduce an effective MAP inference method for problems with large output domains. The method builds upon alternating minimization of an Augmented Lagrangian that exploits the sparsity of messages through greedy optimization techniques. A key feature of our greedy approach is to introduce variables in an on-demand manner with a pre-built data structure over local factors. This results in a single-loop algorithm of sublinear cost per iteration and O(log(1))-type iteration complexity to achieve ε sub-optimality. In addition, we introduce a variant of GDMM for binary MAP inference problems with a large number of factors. Empirically, the proposed algorithms demonstrate orders of magnitude speedup over state-of-the-art MAP inference techniques on MAP inference problems including Segmentation, Protein Folding, Graph Matching, and Multilabel prediction with pairwise interaction.

最大后验推理(MAP)是图形模型和结构化预测的核心。尽管精确MAP推理很棘手,但基于LP松弛的近似方法在广泛的应用中表现出优异的性能。然而,对于涉及大输出域(即每个变量的状态空间都很大)的问题,标准LP松弛很容易产生大量超出现有优化算法极限的变量和约束。本文针对具有大输出域的问题,提出了一种有效的MAP推理方法。该方法建立在增广拉格朗日量交替最小化的基础上,通过贪婪优化技术利用消息的稀疏性。我们的贪心方法的一个关键特征是,在本地因素上使用预先构建的数据结构以随需应变的方式引入变量。这导致了每次迭代的次线性代价和O(log(1/ε))型迭代复杂度的单循环算法来实现ε次最优性。此外,我们还引入了GDMM的一种变体,用于具有大量因素的二元MAP推理问题。从经验上看,所提出的算法在MAP推理问题上的速度比最先进的MAP推理技术快了几个数量级,包括分割、蛋白质折叠、图匹配和具有成对交互的多标签预测。
{"title":"Greedy Direction Method of Multiplier for MAP Inference of Large Output Domain.","authors":"Xiangru Huang,&nbsp;Qixing Huang,&nbsp;Ian E H Yen,&nbsp;Pradeep Ravikumar,&nbsp;Ruohan Zhang,&nbsp;Inderjit S Dhillon","doi":"","DOIUrl":"","url":null,"abstract":"<p><p>Maximum-a-Posteriori (MAP) inference lies at the heart of Graphical Models and Structured Prediction. Despite the intractability of exact MAP inference, approximate methods based on LP relaxations have exhibited superior performance across a wide range of applications. Yet for problems involving large output domains (i.e., the state space for each variable is large), standard LP relaxations can easily give rise to a large number of variables and constraints which are beyond the limit of existing optimization algorithms. In this paper, we introduce an effective MAP inference method for problems with large output domains. The method builds upon alternating minimization of an Augmented Lagrangian that exploits the sparsity of messages through greedy optimization techniques. A key feature of our greedy approach is to introduce variables in an on-demand manner with a pre-built data structure over local factors. This results in a single-loop algorithm of sublinear cost per iteration and <i>O</i>(log(1<i>/ε</i>))-type iteration complexity to achieve <i>ε</i> sub-optimality. In addition, we introduce a variant of GDMM for binary MAP inference problems with a large number of factors. Empirically, the proposed algorithms demonstrate orders of magnitude speedup over state-of-the-art MAP inference techniques on MAP inference problems including Segmentation, Protein Folding, Graph Matching, and Multilabel prediction with pairwise interaction.</p>","PeriodicalId":89793,"journal":{"name":"JMLR workshop and conference proceedings","volume":"54 ","pages":"1550-1559"},"PeriodicalIF":0.0,"publicationDate":"2017-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5581664/pdf/nihms896523.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"35472765","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Doctor AI: Predicting Clinical Events via Recurrent Neural Networks. 人工智能医生:通过循环神经网络预测临床事件。
Pub Date : 2016-08-01 Epub Date: 2016-12-10
Edward Choi, Mohammad Taha Bahadori, Andy Schuetz, Walter F Stewart, Jimeng Sun

Leveraging large historical data in electronic health record (EHR), we developed Doctor AI, a generic predictive model that covers observed medical conditions and medication uses. Doctor AI is a temporal model using recurrent neural networks (RNN) and was developed and applied to longitudinal time stamped EHR data from 260K patients over 8 years. Encounter records (e.g. diagnosis codes, medication codes or procedure codes) were input to RNN to predict (all) the diagnosis and medication categories for a subsequent visit. Doctor AI assesses the history of patients to make multilabel predictions (one label for each diagnosis or medication category). Based on separate blind test set evaluation, Doctor AI can perform differential diagnosis with up to 79% recall@30, significantly higher than several baselines. Moreover, we demonstrate great generalizability of Doctor AI by adapting the resulting models from one institution to another without losing substantial accuracy.

利用电子健康记录(EHR)中的大量历史数据,我们开发了医生人工智能,这是一种涵盖观察到的医疗状况和药物使用的通用预测模型。医生AI是一个使用循环神经网络(RNN)的时间模型,开发并应用于8年来260K名患者的纵向时间戳电子病历数据。将就诊记录(如诊断代码、用药代码或程序代码)输入到RNN中,以预测后续就诊的(所有)诊断和用药类别。医生AI会评估患者的病史,做出多标签预测(每个诊断或药物类别都有一个标签)。基于单独的盲测试集评估,医生AI可以进行高达79%的鉴别诊断recall@30,显著高于几个基线。此外,我们通过将结果模型从一个机构调整到另一个机构,而不会失去实质性的准确性,证明了医生人工智能的良好泛化性。
{"title":"Doctor AI: Predicting Clinical Events via Recurrent Neural Networks.","authors":"Edward Choi,&nbsp;Mohammad Taha Bahadori,&nbsp;Andy Schuetz,&nbsp;Walter F Stewart,&nbsp;Jimeng Sun","doi":"","DOIUrl":"","url":null,"abstract":"<p><p>Leveraging large historical data in electronic health record (EHR), we developed Doctor AI, a generic predictive model that covers observed medical conditions and medication uses. Doctor AI is a temporal model using recurrent neural networks (RNN) and was developed and applied to longitudinal time stamped EHR data from 260K patients over 8 years. Encounter records (e.g. diagnosis codes, medication codes or procedure codes) were input to RNN to predict (all) the diagnosis and medication categories for a subsequent visit. Doctor AI assesses the history of patients to make multilabel predictions (one label for each diagnosis or medication category). Based on separate blind test set evaluation, Doctor AI can perform differential diagnosis with up to 79% recall@30, significantly higher than several baselines. Moreover, we demonstrate great generalizability of Doctor AI by adapting the resulting models from one institution to another without losing substantial accuracy.</p>","PeriodicalId":89793,"journal":{"name":"JMLR workshop and conference proceedings","volume":"56 ","pages":"301-318"},"PeriodicalIF":0.0,"publicationDate":"2016-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5341604/pdf/nihms-845642.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"34806178","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Hybrid Causal Search Algorithm for Latent Variable Models. 潜在变量模型的混合因果搜索算法。
Juan Miguel Ogarrio, Peter Spirtes, Joe Ramsey

Existing score-based causal model search algorithms such as GES (and a speeded up version, FGS) are asymptotically correct, fast, and reliable, but make the unrealistic assumption that the true causal graph does not contain any unmeasured confounders. There are several constraint-based causal search algorithms (e.g RFCI, FCI, or FCI+) that are asymptotically correct without assuming that there are no unmeasured confounders, but often perform poorly on small samples. We describe a combined score and constraint-based algorithm, GFCI, that we prove is asymptotically correct. On synthetic data, GFCI is only slightly slower than RFCI but more accurate than FCI, RFCI and FCI+.

现有的基于分数的因果模型搜索算法,如GES(和一个加速版本,FGS)是渐进正确、快速和可靠的,但它做出了一个不切实际的假设,即真正的因果图不包含任何不可测量的混杂因素。有几种基于约束的因果搜索算法(如RFCI、FCI或FCI+)在不假设没有未测量混杂因素的情况下是渐近正确的,但在小样本上往往表现不佳。我们描述了一个结合分数和约束的算法,GFCI,我们证明了它是渐近正确的。在综合数据上,GFCI仅比RFCI略慢,但比FCI、RFCI和FCI+更准确。
{"title":"A Hybrid Causal Search Algorithm for Latent Variable Models.","authors":"Juan Miguel Ogarrio,&nbsp;Peter Spirtes,&nbsp;Joe Ramsey","doi":"","DOIUrl":"","url":null,"abstract":"<p><p>Existing score-based causal model search algorithms such as <i>GES</i> (and a speeded up version, <i>FGS</i>) are asymptotically correct, fast, and reliable, but make the unrealistic assumption that the true causal graph does not contain any unmeasured confounders. There are several constraint-based causal search algorithms (e.g <i>RFCI, FCI</i>, or <i>FCI</i>+) that are asymptotically correct without assuming that there are no unmeasured confounders, but often perform poorly on small samples. We describe a combined score and constraint-based algorithm, <i>GFCI</i>, that we prove is asymptotically correct. On synthetic data, <i>GFCI</i> is only slightly slower than <i>RFCI</i> but more accurate than <i>FCI, RFCI</i> and <i>FCI</i>+.</p>","PeriodicalId":89793,"journal":{"name":"JMLR workshop and conference proceedings","volume":"52 ","pages":"368-379"},"PeriodicalIF":0.0,"publicationDate":"2016-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5325717/pdf/nihms845582.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"34766374","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Uncovering Voice Misuse Using Symbolic Mismatch. 利用符号错配揭露语音误用。
Marzyeh Ghassemi, Zeeshan Syed, Daryush D Mehta, Jarrad H Van Stan, Robert E Hillman, John V Guttag

Voice disorders affect an estimated 14 million working-aged Americans, and many more worldwide. We present the first large scale study of vocal misuse based on long-term ambulatory data collected by an accelerometer placed on the neck. We investigate an unsupervised data mining approach to uncovering latent information about voice misuse. We segment signals from over 253 days of data from 22 subjects into over a hundred million single glottal pulses (closures of the vocal folds), cluster segments into symbols, and use symbolic mismatch to uncover differences between patients and matched controls, and between patients pre- and post-treatment. Our results show significant behavioral differences between patients and controls, as well as between some pre- and post-treatment patients. Our proposed approach provides an objective basis for helping diagnose behavioral voice disorders, and is a first step towards a more data-driven understanding of the impact of voice therapy.

据估计,嗓音障碍影响着1400万处于工作年龄的美国人,在世界范围内影响的人数更多。我们提出了基于放置在脖子上的加速度计收集的长期动态数据的第一个大规模的声乐滥用研究。我们研究了一种无监督的数据挖掘方法来发现关于语音滥用的潜在信息。我们将来自22名受试者的253天数据中的信号分割成超过1亿个单个声门脉冲(声带闭合),将片段聚类成符号,并使用符号不匹配来揭示患者与匹配对照组之间以及患者治疗前后之间的差异。我们的研究结果显示患者和对照组之间,以及一些治疗前和治疗后的患者之间存在显著的行为差异。我们提出的方法为帮助诊断行为性语音障碍提供了客观基础,并且是对语音治疗影响的数据驱动理解的第一步。
{"title":"Uncovering Voice Misuse Using Symbolic Mismatch.","authors":"Marzyeh Ghassemi,&nbsp;Zeeshan Syed,&nbsp;Daryush D Mehta,&nbsp;Jarrad H Van Stan,&nbsp;Robert E Hillman,&nbsp;John V Guttag","doi":"","DOIUrl":"","url":null,"abstract":"<p><p>Voice disorders affect an estimated 14 million working-aged Americans, and many more worldwide. We present the first large scale study of vocal misuse based on long-term ambulatory data collected by an accelerometer placed on the neck. We investigate an unsupervised data mining approach to uncovering latent information about voice misuse. We segment signals from over 253 days of data from 22 subjects into over a hundred million single glottal pulses (closures of the vocal folds), cluster segments into symbols, and use symbolic mismatch to uncover differences between patients and matched controls, and between patients pre- and post-treatment. Our results show significant behavioral differences between patients and controls, as well as between some pre- and post-treatment patients. Our proposed approach provides an objective basis for helping diagnose behavioral voice disorders, and is a first step towards a more data-driven understanding of the impact of voice therapy.</p>","PeriodicalId":89793,"journal":{"name":"JMLR workshop and conference proceedings","volume":"56 ","pages":"239-252"},"PeriodicalIF":0.0,"publicationDate":"2016-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8693775/pdf/nihms-1069009.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"39871655","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Causal Discovery from Subsampled Time Series Data by Constraint Optimization. 通过约束优化从子采样时间序列数据中发现因果关系
Antti Hyttinen, Sergey Plis, Matti Järvisalo, Frederick Eberhardt, David Danks

This paper focuses on causal structure estimation from time series data in which measurements are obtained at a coarser timescale than the causal timescale of the underlying system. Previous work has shown that such subsampling can lead to significant errors about the system's causal structure if not properly taken into account. In this paper, we first consider the search for the system timescale causal structures that correspond to a given measurement timescale structure. We provide a constraint satisfaction procedure whose computational performance is several orders of magnitude better than previous approaches. We then consider finite-sample data as input, and propose the first constraint optimization approach for recovering the system timescale causal structure. This algorithm optimally recovers from possible conflicts due to statistical errors. More generally, these advances allow for a robust and non-parametric estimation of system timescale causal structures from subsampled time series data.

本文重点研究从时间序列数据中估算因果结构,在这些数据中,测量值的时间尺度比基本系统的因果时间尺度更粗。以往的研究表明,如果不适当考虑这种子采样,会导致系统因果结构出现重大误差。在本文中,我们首先考虑寻找与给定测量时标结构相对应的系统时标因果结构。我们提供了一种约束满足程序,其计算性能比以前的方法高出几个数量级。然后,我们将有限样本数据作为输入,提出了第一种恢复系统时标因果结构的约束优化方法。该算法能从统计误差导致的可能冲突中优化恢复。更广泛地说,这些进展允许从子采样时间序列数据中对系统时标因果结构进行稳健的非参数估计。
{"title":"Causal Discovery from Subsampled Time Series Data by Constraint Optimization.","authors":"Antti Hyttinen, Sergey Plis, Matti Järvisalo, Frederick Eberhardt, David Danks","doi":"","DOIUrl":"","url":null,"abstract":"<p><p>This paper focuses on causal structure estimation from time series data in which measurements are obtained at a coarser timescale than the causal timescale of the underlying system. Previous work has shown that such subsampling can lead to significant errors about the system's causal structure if not properly taken into account. In this paper, we first consider the search for the system timescale causal structures that correspond to a given measurement timescale structure. We provide a constraint satisfaction procedure whose computational performance is several orders of magnitude better than previous approaches. We then consider finite-sample data as input, and propose the first constraint optimization approach for recovering the system timescale causal structure. This algorithm optimally recovers from possible conflicts due to statistical errors. More generally, these advances allow for a robust and non-parametric estimation of system timescale causal structures from subsampled time series data.</p>","PeriodicalId":89793,"journal":{"name":"JMLR workshop and conference proceedings","volume":"52 ","pages":"216-227"},"PeriodicalIF":0.0,"publicationDate":"2016-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5305170/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140195246","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Square Root Graphical Models: Multivariate Generalizations of Univariate Exponential Families that Permit Positive Dependencies. 平方根图形模型:允许正相关的单变量指数族的多元推广。
David I Inouye, Pradeep Ravikumar, Inderjit S Dhillon

We develop Square Root Graphical Models (SQR), a novel class of parametric graphical models that provides multivariate generalizations of univariate exponential family distributions. Previous multivariate graphical models (Yang et al., 2015) did not allow positive dependencies for the exponential and Poisson generalizations. However, in many real-world datasets, variables clearly have positive dependencies. For example, the airport delay time in New York-modeled as an exponential distribution-is positively related to the delay time in Boston. With this motivation, we give an example of our model class derived from the univariate exponential distribution that allows for almost arbitrary positive and negative dependencies with only a mild condition on the parameter matrix-a condition akin to the positive definiteness of the Gaussian covariance matrix. Our Poisson generalization allows for both positive and negative dependencies without any constraints on the parameter values. We also develop parameter estimation methods using node-wise regressions with 1 regularization and likelihood approximation methods using sampling. Finally, we demonstrate our exponential generalization on a synthetic dataset and a real-world dataset of airport delay times.

我们开发了平方根图形模型(SQR),这是一类新的参数图形模型,它提供了单变量指数族分布的多元推广。以前的多变量图形模型(Yang et al., 2015)不允许指数和泊松推广的正依赖关系。然而,在许多真实世界的数据集中,变量显然具有正相关性。例如,纽约机场的延误时间(建模为指数分布)与波士顿的延误时间呈正相关。有了这个动机,我们给出了一个模型类的例子,该模型类来源于单变量指数分布,它允许几乎任意的正相关和负相关,而参数矩阵只有一个温和的条件——一个类似于高斯协方差矩阵的正确定性的条件。我们的泊松泛化允许正依赖和负依赖,而不受参数值的任何约束。我们也发展了参数估计方法使用节点明智的回归与1正则化和似然逼近方法使用抽样。最后,我们在一个合成数据集和一个真实的机场延误时间数据集上证明了我们的指数泛化。
{"title":"Square Root Graphical Models: Multivariate Generalizations of Univariate Exponential Families that Permit Positive Dependencies.","authors":"David I Inouye,&nbsp;Pradeep Ravikumar,&nbsp;Inderjit S Dhillon","doi":"","DOIUrl":"","url":null,"abstract":"<p><p>We develop Square Root Graphical Models (SQR), a novel class of parametric graphical models that provides multivariate generalizations of univariate exponential family distributions. Previous multivariate graphical models (Yang et al., 2015) did not allow positive dependencies for the exponential and Poisson generalizations. However, in many real-world datasets, variables clearly have positive dependencies. For example, the airport delay time in New York-modeled as an exponential distribution-is positively related to the delay time in Boston. With this motivation, we give an example of our model class derived from the univariate exponential distribution that allows for almost arbitrary positive and negative dependencies with only a mild condition on the parameter matrix-a condition akin to the positive definiteness of the Gaussian covariance matrix. Our Poisson generalization allows for both positive and negative dependencies without any constraints on the parameter values. We also develop parameter estimation methods using node-wise regressions with <i>ℓ</i><sub>1</sub> regularization and likelihood approximation methods using sampling. Finally, we demonstrate our exponential generalization on a synthetic dataset and a real-world dataset of airport delay times.</p>","PeriodicalId":89793,"journal":{"name":"JMLR workshop and conference proceedings","volume":"48 ","pages":"2445-2453"},"PeriodicalIF":0.0,"publicationDate":"2016-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4995108/pdf/nihms808904.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"34338585","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Domain Adaptation with Conditional Transferable Components. 具有条件可转移组件的领域自适应。
Mingming Gong, Kun Zhang, Tongliang Liu, Dacheng Tao, Clark Glymour, Bernhard Schölkopf

Domain adaptation arises in supervised learning when the training (source domain) and test (target domain) data have different distributions. Let X and Y denote the features and target, respectively, previous work on domain adaptation mainly considers the covariate shift situation where the distribution of the features P(X) changes across domains while the conditional distribution P(YX) stays the same. To reduce domain discrepancy, recent methods try to find invariant components [Formula: see text] that have similar [Formula: see text] on different domains by explicitly minimizing a distribution discrepancy measure. However, it is not clear if [Formula: see text] in different domains is also similar when P(YX) changes. Furthermore, transferable components do not necessarily have to be invariant. If the change in some components is identifiable, we can make use of such components for prediction in the target domain. In this paper, we focus on the case where P(XY) and P(Y) both change in a causal system in which Y is the cause for X. Under appropriate assumptions, we aim to extract conditional transferable components whose conditional distribution [Formula: see text] is invariant after proper location-scale (LS) transformations, and identify how P(Y) changes between domains simultaneously. We provide theoretical analysis and empirical evaluation on both synthetic and real-world data to show the effectiveness of our method.

在监督学习中,当训练数据(源域)和测试数据(目标域)具有不同的分布时,就会产生领域自适应。设X和Y分别表示特征和目标,以往关于域自适应的工作主要考虑协变量移位情况,即特征P(X)的分布跨域变化,而条件分布P(Y∣X)保持不变。为了减少域差异,最近的方法试图通过显式地最小化分布差异度量来找到在不同域中具有相似的不变分量[公式:见文本]。然而,尚不清楚当P(Y∣X)变化时,不同领域中的[公式:见文本]是否也相似。此外,可转移的组件不一定是不变的。如果某些组件的变化是可识别的,我们可以利用这些组件在目标域中进行预测。在本文中,我们重点研究了P(X∣Y)和P(Y)在一个因果系统中同时变化的情况,其中Y是X的原因。在适当的假设下,我们旨在提取条件分布[公式:见文本]经过适当的位置尺度(LS)变换后不变的条件可转移分量,并确定P(Y)如何在域之间同时变化。我们对合成数据和实际数据进行了理论分析和实证评估,以证明我们的方法的有效性。
{"title":"Domain Adaptation with Conditional Transferable Components.","authors":"Mingming Gong,&nbsp;Kun Zhang,&nbsp;Tongliang Liu,&nbsp;Dacheng Tao,&nbsp;Clark Glymour,&nbsp;Bernhard Schölkopf","doi":"","DOIUrl":"","url":null,"abstract":"<p><p>Domain adaptation arises in supervised learning when the training (source domain) and test (target domain) data have different distributions. Let <i>X</i> and <i>Y</i> denote the features and target, respectively, previous work on domain adaptation mainly considers the covariate shift situation where the distribution of the features <i>P</i>(<i>X</i>) changes across domains while the conditional distribution <i>P</i>(<i>Y</i>∣<i>X</i>) stays the same. To reduce domain discrepancy, recent methods try to find invariant components [Formula: see text] that have similar [Formula: see text] on different domains by explicitly minimizing a distribution discrepancy measure. However, it is not clear if [Formula: see text] in different domains is also similar when <i>P</i>(<i>Y</i>∣<i>X</i>) changes. Furthermore, transferable components do not necessarily have to be invariant. If the change in some components is identifiable, we can make use of such components for prediction in the target domain. In this paper, we focus on the case where <i>P</i>(<i>X</i>∣<i>Y</i>) and <i>P</i>(<i>Y</i>) both change in a causal system in which <i>Y</i> is the cause for <i>X</i>. Under appropriate assumptions, we aim to extract conditional transferable components whose conditional distribution [Formula: see text] is invariant after proper location-scale (LS) transformations, and identify how <i>P</i>(<i>Y</i>) changes between domains simultaneously. We provide theoretical analysis and empirical evaluation on both synthetic and real-world data to show the effectiveness of our method.</p>","PeriodicalId":89793,"journal":{"name":"JMLR workshop and conference proceedings","volume":"48 ","pages":"2839-2848"},"PeriodicalIF":0.0,"publicationDate":"2016-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5321138/pdf/nihms-846268.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"34766373","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Anytime Exploration for Multi-armed Bandits using Confidence Information. 利用信任信息随时探索多武装土匪。
Kwang-Sung Jun, Robert Nowak

We introduce anytime Explore-m, a pure exploration problem for multi-armed bandits (MAB) that requires making a prediction of the top-m arms at every time step. Anytime Explore-m is more practical than fixed budget or fixed confidence formulations of the top-m problem, since many applications involve a finite, but unpredictable, budget. However, the development and analysis of anytime algorithms present many challenges. We propose AT-LUCB (AnyTime Lower and Upper Confidence Bound), the first nontrivial algorithm that provably solves anytime Explore-m. Our analysis shows that the sample complexity of AT-LUCB is competitive to anytime variants of existing algorithms. Moreover, our empirical evaluation on AT-LUCB shows that AT-LUCB performs as well as or better than state-of-the-art baseline methods for anytime Explore-m.

我们引入了anytime Explore-m,这是一个针对多臂土匪(MAB)的纯探索问题,它需要在每个时间步长对最上面的m个臂进行预测。无论何时,Explore-m都比top-m问题的固定预算或固定置信度公式更实用,因为许多应用都涉及有限但不可预测的预算。然而,任意时间算法的开发和分析面临着许多挑战。我们提出了AT-LUCB (AnyTime Lower and Upper Confidence Bound)算法,这是第一个可以证明解决AnyTime Explore-m问题的非平凡算法。我们的分析表明,AT-LUCB的样本复杂度与现有算法的任何变体相比都具有竞争力。此外,我们对AT-LUCB的实证评估表明,AT-LUCB在任何时候都与最先进的基线方法一样好,甚至更好。
{"title":"Anytime Exploration for Multi-armed Bandits using Confidence Information.","authors":"Kwang-Sung Jun,&nbsp;Robert Nowak","doi":"","DOIUrl":"","url":null,"abstract":"<p><p>We introduce anytime Explore-<i>m</i>, a pure exploration problem for multi-armed bandits (MAB) that requires making a prediction of the top-<i>m</i> arms at every time step. Anytime Explore-<i>m</i> is more practical than fixed budget or fixed confidence formulations of the top-<i>m</i> problem, since many applications involve a finite, but unpredictable, budget. However, the development and analysis of anytime algorithms present many challenges. We propose AT-LUCB (AnyTime Lower and Upper Confidence Bound), the first nontrivial algorithm that provably solves anytime Explore-<i>m</i>. Our analysis shows that the sample complexity of AT-LUCB is competitive to anytime variants of existing algorithms. Moreover, our empirical evaluation on AT-LUCB shows that AT-LUCB performs as well as or better than state-of-the-art baseline methods for anytime Explore-<i>m</i>.</p>","PeriodicalId":89793,"journal":{"name":"JMLR workshop and conference proceedings","volume":"48 ","pages":"974-982"},"PeriodicalIF":0.0,"publicationDate":"2016-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5846129/pdf/nihms894213.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"35915066","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Experimental Design on a Budget for Sparse Linear Models and Applications. 稀疏线性模型和应用的预算实验设计
Sathya N Ravi, Vamsi K Ithapu, Sterling C Johnson, Vikas Singh

Budget constrained optimal design of experiments is a well studied problem. Although the literature is very mature, not many strategies are available when these design problems appear in the context of sparse linear models commonly encountered in high dimensional machine learning. In this work, we study this budget constrained design where the underlying regression model involves a 1-regularized linear function. We propose two novel strategies: the first is motivated geometrically whereas the second is algebraic in nature. We obtain tractable algorithms for this problem which also hold for a more general class of sparse linear models. We perform a detailed set of experiments, on benchmarks and a large neuroimaging study, showing that the proposed models are effective in practice. The latter experiment suggests that these ideas may play a small role in informing enrollment strategies for similar scientific studies in the future.

预算受限的实验优化设计是一个经过深入研究的问题。虽然相关文献已经非常成熟,但当这些设计问题出现在高维机器学习中常见的稀疏线性模型背景下时,可用的策略并不多。在这项工作中,我们研究了底层回归模型涉及 ℓ1-regularized 线性函数时的预算受限设计。我们提出了两种新颖的策略:第一种从几何角度出发,而第二种从代数角度出发。我们获得了解决这个问题的可行算法,这些算法也适用于更一般的稀疏线性模型。我们在基准和大型神经成像研究中进行了一系列详细的实验,结果表明所提出的模型在实践中是有效的。后一项实验表明,这些想法可能会在未来类似科学研究的招生策略中发挥微小的作用。
{"title":"Experimental Design on a Budget for Sparse Linear Models and Applications.","authors":"Sathya N Ravi, Vamsi K Ithapu, Sterling C Johnson, Vikas Singh","doi":"","DOIUrl":"","url":null,"abstract":"<p><p>Budget constrained optimal design of experiments is a well studied problem. Although the literature is very mature, not many strategies are available when these design problems appear in the context of sparse linear models commonly encountered in high dimensional machine learning. In this work, we study this budget constrained design where the underlying regression model involves a <i>ℓ</i><sub>1</sub>-regularized linear function. We propose two novel strategies: the first is motivated geometrically whereas the second is algebraic in nature. We obtain tractable algorithms for this problem which also hold for a more general class of sparse linear models. We perform a detailed set of experiments, on benchmarks and a large neuroimaging study, showing that the proposed models are effective in practice. The latter experiment suggests that these ideas may play a small role in informing enrollment strategies for similar scientific studies in the future.</p>","PeriodicalId":89793,"journal":{"name":"JMLR workshop and conference proceedings","volume":"48 ","pages":"583-592"},"PeriodicalIF":0.0,"publicationDate":"2016-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5415092/pdf/nihms855967.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"34974183","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
JMLR workshop and conference proceedings
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1