首页 > 最新文献

Journal of Machine Learning Research最新文献

英文 中文
SnapVX: A Network-Based Convex Optimization Solver. 基于网络的凸优化求解器。
IF 6 3区 计算机科学 Q1 Mathematics Pub Date : 2017-01-01
David Hallac, Christopher Wong, Steven Diamond, Abhijit Sharang, Rok Sosič, Stephen Boyd, Jure Leskovec

SnapVX is a high-performance solver for convex optimization problems defined on networks. For problems of this form, SnapVX provides a fast and scalable solution with guaranteed global convergence. It combines the capabilities of two open source software packages: Snap.py and CVXPY. Snap.py is a large scale graph processing library, and CVXPY provides a general modeling framework for small-scale subproblems. SnapVX offers a customizable yet easy-to-use Python interface with "out-of-the-box" functionality. Based on the Alternating Direction Method of Multipliers (ADMM), it is able to efficiently store, analyze, parallelize, and solve large optimization problems from a variety of different applications. Documentation, examples, and more can be found on the SnapVX website at http://snap.stanford.edu/snapvx.

SnapVX是针对网络上定义的凸优化问题的高性能求解器。对于这种形式的问题,SnapVX提供了一个快速和可扩展的解决方案,并保证了全局收敛。它结合了两个开源软件包的功能:Snap.py和cvvxpy。Snap.py是一个大规模图形处理库,CVXPY为小规模子问题提供了通用的建模框架。SnapVX提供了一个可定制且易于使用的Python界面,具有“开箱即用”的功能。它基于乘法器的交替方向法(ADMM),能够有效地存储、分析、并行化和解决来自各种不同应用的大型优化问题。可以在SnapVX网站http://snap.stanford.edu/snapvx上找到文档、示例和更多内容。
{"title":"SnapVX: A Network-Based Convex Optimization Solver.","authors":"David Hallac,&nbsp;Christopher Wong,&nbsp;Steven Diamond,&nbsp;Abhijit Sharang,&nbsp;Rok Sosič,&nbsp;Stephen Boyd,&nbsp;Jure Leskovec","doi":"","DOIUrl":"","url":null,"abstract":"<p><p>SnapVX is a high-performance solver for convex optimization problems defined on networks. For problems of this form, SnapVX provides a fast and scalable solution with guaranteed global convergence. It combines the capabilities of two open source software packages: Snap.py and CVXPY. Snap.py is a large scale graph processing library, and CVXPY provides a general modeling framework for small-scale subproblems. SnapVX offers a customizable yet easy-to-use Python interface with \"out-of-the-box\" functionality. Based on the Alternating Direction Method of Multipliers (ADMM), it is able to efficiently store, analyze, parallelize, and solve large optimization problems from a variety of different applications. Documentation, examples, and more can be found on the SnapVX website at http://snap.stanford.edu/snapvx.</p>","PeriodicalId":50161,"journal":{"name":"Journal of Machine Learning Research","volume":null,"pages":null},"PeriodicalIF":6.0,"publicationDate":"2017-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5870756/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"35960855","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Bridging supervised learning and test-based co-optimization 桥梁监督学习和基于测试的协同优化
IF 6 3区 计算机科学 Q1 Mathematics Pub Date : 2017-01-01 DOI: 10.5555/3122009.3122047
PopoviciElena
This paper takes a close look at the important commonalities and subtle differences between the well-established field of supervised learning and the much younger one of cooptimization. It explains...
本文仔细研究了建立良好的监督学习领域与更年轻的协同优化领域之间的重要共同点和细微差异。它解释了……
{"title":"Bridging supervised learning and test-based co-optimization","authors":"PopoviciElena","doi":"10.5555/3122009.3122047","DOIUrl":"https://doi.org/10.5555/3122009.3122047","url":null,"abstract":"This paper takes a close look at the important commonalities and subtle differences between the well-established field of supervised learning and the much younger one of cooptimization. It explains...","PeriodicalId":50161,"journal":{"name":"Journal of Machine Learning Research","volume":null,"pages":null},"PeriodicalIF":6.0,"publicationDate":"2017-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"71140012","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
The DFS fused lasso DFS熔接套索
IF 6 3区 计算机科学 Q1 Mathematics Pub Date : 2017-01-01 DOI: 10.5555/3122009.3242033
PadillaOscar Hernan Madrid, SharpnackJames, G. ScottJames
The fused lasso, also known as (anisotropic) total variation denoising, is widely used for piecewise constant signal estimation with respect to a given undirected graph. The fused lasso estimate is...
融合套索,也称为(各向异性)全变分去噪,广泛用于对给定无向图的分段常数信号估计。融合套索估计是…
{"title":"The DFS fused lasso","authors":"PadillaOscar Hernan Madrid, SharpnackJames, G. ScottJames","doi":"10.5555/3122009.3242033","DOIUrl":"https://doi.org/10.5555/3122009.3242033","url":null,"abstract":"The fused lasso, also known as (anisotropic) total variation denoising, is widely used for piecewise constant signal estimation with respect to a given undirected graph. The fused lasso estimate is...","PeriodicalId":50161,"journal":{"name":"Journal of Machine Learning Research","volume":null,"pages":null},"PeriodicalIF":6.0,"publicationDate":"2017-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"71139798","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Structure-Leveraged Methods in Breast Cancer Risk Prediction. 乳腺癌风险预测中的结构杠杆方法。
IF 6 3区 计算机科学 Q1 Mathematics Pub Date : 2016-12-01
Jun Fan, Yirong Wu, Ming Yuan, David Page, Jie Liu, Irene M Ong, Peggy Peissig, Elizabeth Burnside

Predicting breast cancer risk has long been a goal of medical research in the pursuit of precision medicine. The goal of this study is to develop novel penalized methods to improve breast cancer risk prediction by leveraging structure information in electronic health records. We conducted a retrospective case-control study, garnering 49 mammography descriptors and 77 high-frequency/low-penetrance single-nucleotide polymorphisms (SNPs) from an existing personalized medicine data repository. Structured mammography reports and breast imaging features have long been part of a standard electronic health record (EHR), and genetic markers likely will be in the near future. Lasso and its variants are widely used approaches to integrated learning and feature selection, and our methodological contribution is to incorporate the dependence structure among the features into these approaches. More specifically, we propose a new methodology by combining group penalty and [Formula: see text] (1 ≤ p ≤ 2) fusion penalty to improve breast cancer risk prediction, taking into account structure information in mammography descriptors and SNPs. We demonstrate that our method provides benefits that are both statistically significant and potentially significant to people's lives.

在追求精准医学的过程中,预测乳腺癌风险一直是医学研究的一个目标。本研究的目的是开发新的惩罚方法,利用电子健康记录中的结构信息来提高乳腺癌风险预测。我们进行了一项回顾性病例对照研究,从现有的个性化医学数据库中收集了49个乳房x线摄影描述符和77个高频/低外显率单核苷酸多态性(snp)。结构化乳房x光检查报告和乳房成像特征长期以来一直是标准电子健康记录(EHR)的一部分,遗传标记可能在不久的将来也会成为标准电子健康记录的一部分。Lasso及其变体是广泛使用的集成学习和特征选择方法,我们的方法贡献是将特征之间的依赖结构纳入这些方法中。更具体地说,我们提出了一种新的方法,结合群体惩罚和[公式:见文本](1≤p≤2)融合惩罚来提高乳腺癌风险预测,同时考虑到乳房x光描述符和snp的结构信息。我们证明,我们的方法对人们的生活既有统计学意义,也有潜在意义。
{"title":"Structure-Leveraged Methods in Breast Cancer Risk Prediction.","authors":"Jun Fan,&nbsp;Yirong Wu,&nbsp;Ming Yuan,&nbsp;David Page,&nbsp;Jie Liu,&nbsp;Irene M Ong,&nbsp;Peggy Peissig,&nbsp;Elizabeth Burnside","doi":"","DOIUrl":"","url":null,"abstract":"<p><p>Predicting breast cancer risk has long been a goal of medical research in the pursuit of precision medicine. The goal of this study is to develop novel penalized methods to improve breast cancer risk prediction by leveraging structure information in electronic health records. We conducted a retrospective case-control study, garnering 49 mammography descriptors and 77 high-frequency/low-penetrance single-nucleotide polymorphisms (SNPs) from an existing personalized medicine data repository. Structured mammography reports and breast imaging features have long been part of a standard electronic health record (EHR), and genetic markers likely will be in the near future. Lasso and its variants are widely used approaches to integrated learning and feature selection, and our methodological contribution is to incorporate the dependence structure among the features into these approaches. More specifically, we propose a new methodology by combining group penalty and [Formula: see text] (1 ≤ <i>p</i> ≤ 2) fusion penalty to improve breast cancer risk prediction, taking into account structure information in mammography descriptors and SNPs. We demonstrate that our method provides benefits that are both statistically significant and potentially significant to people's lives.</p>","PeriodicalId":50161,"journal":{"name":"Journal of Machine Learning Research","volume":null,"pages":null},"PeriodicalIF":6.0,"publicationDate":"2016-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5446896/pdf/nihms-826646.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"35042470","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Double or Nothing 要么加倍要么一无所获
IF 6 3区 计算机科学 Q1 Mathematics Pub Date : 2016-08-24 DOI: 10.5555/2946645.3053447
Carol Sutton
Crowdsourcing has gained immense popularity in machine learning applications for obtaining large amounts of labeled data. Crowdsourcing is cheap and fast, but suffers from the problem of low-qualit...
众包在机器学习应用中获得大量标记数据获得了极大的普及。众包成本低、速度快,但存在质量低下的问题……
{"title":"Double or Nothing","authors":"Carol Sutton","doi":"10.5555/2946645.3053447","DOIUrl":"https://doi.org/10.5555/2946645.3053447","url":null,"abstract":"Crowdsourcing has gained immense popularity in machine learning applications for obtaining large amounts of labeled data. Crowdsourcing is cheap and fast, but suffers from the problem of low-qualit...","PeriodicalId":50161,"journal":{"name":"Journal of Machine Learning Research","volume":null,"pages":null},"PeriodicalIF":6.0,"publicationDate":"2016-08-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"71138965","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Convex Regression with Interpretable Sharp Partitions. 带可解释锐分区的凸回归
IF 6 3区 计算机科学 Q1 Mathematics Pub Date : 2016-06-01
Ashley Petersen, Noah Simon, Daniela Witten

We consider the problem of predicting an outcome variable on the basis of a small number of covariates, using an interpretable yet non-additive model. We propose convex regression with interpretable sharp partitions (CRISP) for this task. CRISP partitions the covariate space into blocks in a data-adaptive way, and fits a mean model within each block. Unlike other partitioning methods, CRISP is fit using a non-greedy approach by solving a convex optimization problem, resulting in low-variance fits. We explore the properties of CRISP, and evaluate its performance in a simulation study and on a housing price data set.

我们考虑的问题是,利用可解释但非相加的模型,在少量协变量的基础上预测结果变量。针对这一任务,我们提出了可解释锐分区凸回归(CRISP)。CRISP 以数据适应的方式将协变量空间划分为若干区块,并在每个区块内拟合一个均值模型。与其他分区方法不同的是,CRISP 是通过求解一个凸优化问题,采用非贪心方法拟合的,从而获得低方差拟合结果。我们探讨了 CRISP 的特性,并通过模拟研究和住房价格数据集对其性能进行了评估。
{"title":"Convex Regression with Interpretable Sharp Partitions.","authors":"Ashley Petersen, Noah Simon, Daniela Witten","doi":"","DOIUrl":"","url":null,"abstract":"<p><p>We consider the problem of predicting an outcome variable on the basis of a small number of covariates, using an interpretable yet non-additive model. We propose <i>convex regression with interpretable sharp partitions</i> (CRISP) for this task. CRISP partitions the covariate space into blocks in a data-adaptive way, and fits a mean model within each block. Unlike other partitioning methods, CRISP is fit using a non-greedy approach by solving a convex optimization problem, resulting in low-variance fits. We explore the properties of CRISP, and evaluate its performance in a simulation study and on a housing price data set.</p>","PeriodicalId":50161,"journal":{"name":"Journal of Machine Learning Research","volume":null,"pages":null},"PeriodicalIF":6.0,"publicationDate":"2016-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5021451/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140208103","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
L1-Regularized Least Squares for Support Recovery of High Dimensional Single Index Models with Gaussian Designs. 高斯设计的高维单指数模型支持恢复的 L1-Regularized Least Squares。
IF 6 3区 计算机科学 Q1 Mathematics Pub Date : 2016-05-01
Matey Neykov, Jun S Liu, Tianxi Cai

It is known that for a certain class of single index models (SIMs) [Formula: see text], support recovery is impossible when X ~ 𝒩(0, 𝕀 p×p ) and a model complexity adjusted sample size is below a critical threshold. Recently, optimal algorithms based on Sliced Inverse Regression (SIR) were suggested. These algorithms work provably under the assumption that the design X comes from an i.i.d. Gaussian distribution. In the present paper we analyze algorithms based on covariance screening and least squares with L1 penalization (i.e. LASSO) and demonstrate that they can also enjoy optimal (up to a scalar) rescaled sample size in terms of support recovery, albeit under slightly different assumptions on f and ε compared to the SIR based algorithms. Furthermore, we show more generally, that LASSO succeeds in recovering the signed support of β0 if X ~ 𝒩 (0, Σ), and the covariance Σ satisfies the irrepresentable condition. Our work extends existing results on the support recovery of LASSO for the linear model, to a more general class of SIMs.

众所周知,对于某一类单指标模型(SIMs)[公式:见正文],当 X ~ 𝒩(0, 𝕀 p×p ) 和模型复杂度调整样本量低于临界阈值时,支持恢复是不可能的。最近,有人提出了基于切片反回归(SIR)的最优算法。这些算法是在设计 X 来自 i.i.d. 高斯分布的假设条件下证明有效的。在本文中,我们分析了基于协方差筛选和 L1 惩罚最小二乘法(即 LASSO)的算法,并证明它们在支持恢复方面也能获得最佳(达到标量)重标样本大小,尽管与基于 SIR 的算法相比,对 f 和 ε 的假设略有不同。此外,我们还更广泛地表明,如果 X ~ 𝒩 (0, Σ),并且协方差 Σ 满足不可呈现条件,那么 LASSO 就能成功地恢复 β0 的有符号支持。我们的工作将现有的线性模型 LASSO 支持恢复结果扩展到了更一般的 SIMs 类别。
{"title":"<i>L</i><sub>1</sub>-Regularized Least Squares for Support Recovery of High Dimensional Single Index Models with Gaussian Designs.","authors":"Matey Neykov, Jun S Liu, Tianxi Cai","doi":"","DOIUrl":"","url":null,"abstract":"<p><p>It is known that for a certain class of single index models (SIMs) [Formula: see text], support recovery is impossible when <b><i>X</i></b> ~ 𝒩(0, 𝕀 <i><sub>p</sub></i><sub>×</sub><i><sub>p</sub></i> ) and a <i>model complexity adjusted sample size</i> is below a critical threshold. Recently, optimal algorithms based on Sliced Inverse Regression (SIR) were suggested. These algorithms work provably under the assumption that the design <b><i>X</i></b> comes from an i.i.d. Gaussian distribution. In the present paper we analyze algorithms based on covariance screening and least squares with <i>L</i><sub>1</sub> penalization (i.e. LASSO) and demonstrate that they can also enjoy optimal (up to a scalar) rescaled sample size in terms of support recovery, albeit under slightly different assumptions on <i>f</i> and <i>ε</i> compared to the SIR based algorithms. Furthermore, we show more generally, that LASSO succeeds in recovering the signed support of <b><i>β</i></b><sub>0</sub> if <b><i>X</i></b> ~ 𝒩 (0, <b>Σ</b>), and the covariance <b>Σ</b> satisfies the irrepresentable condition. Our work extends existing results on the support recovery of LASSO for the linear model, to a more general class of SIMs.</p>","PeriodicalId":50161,"journal":{"name":"Journal of Machine Learning Research","volume":null,"pages":null},"PeriodicalIF":6.0,"publicationDate":"2016-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5426818/pdf/nihms851690.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"34994441","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
CVXPY: A Python-Embedded Modeling Language for Convex Optimization. 一种用于凸优化的python嵌入式建模语言。
IF 6 3区 计算机科学 Q1 Mathematics Pub Date : 2016-04-01
Steven Diamond, Stephen Boyd

CVXPY is a domain-specific language for convex optimization embedded in Python. It allows the user to express convex optimization problems in a natural syntax that follows the math, rather than in the restrictive standard form required by solvers. CVXPY makes it easy to combine convex optimization with high-level features of Python such as parallelism and object-oriented design. CVXPY is available at http://www.cvxpy.org/ under the GPL license, along with documentation and examples.

CVXPY是一种特定于领域的语言,用于在Python中嵌入凸优化。它允许用户用遵循数学的自然语法来表达凸优化问题,而不是用求解器所要求的限制性标准形式。CVXPY可以很容易地将凸优化与Python的高级特性(如并行性和面向对象设计)结合起来。CVXPY在GPL许可下可在http://www.cvxpy.org/获得,以及文档和示例。
{"title":"CVXPY: A Python-Embedded Modeling Language for Convex Optimization.","authors":"Steven Diamond,&nbsp;Stephen Boyd","doi":"","DOIUrl":"","url":null,"abstract":"<p><p>CVXPY is a domain-specific language for convex optimization embedded in Python. It allows the user to express convex optimization problems in a natural syntax that follows the math, rather than in the restrictive standard form required by solvers. CVXPY makes it easy to combine convex optimization with high-level features of Python such as parallelism and object-oriented design. CVXPY is available at http://www.cvxpy.org/ under the GPL license, along with documentation and examples.</p>","PeriodicalId":50161,"journal":{"name":"Journal of Machine Learning Research","volume":null,"pages":null},"PeriodicalIF":6.0,"publicationDate":"2016-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4927437/pdf/nihms772320.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"34633294","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Gibbs Sampler for Learning DAGs. 学习 DAG 的 Gibbs 采样器
IF 6 3区 计算机科学 Q1 Mathematics Pub Date : 2016-04-01
Robert J B Goudie, Sach Mukherjee

We propose a Gibbs sampler for structure learning in directed acyclic graph (DAG) models. The standard Markov chain Monte Carlo algorithms used for learning DAGs are random-walk Metropolis-Hastings samplers. These samplers are guaranteed to converge asymptotically but often mix slowly when exploring the large graph spaces that arise in structure learning. In each step, the sampler we propose draws entire sets of parents for multiple nodes from the appropriate conditional distribution. This provides an efficient way to make large moves in graph space, permitting faster mixing whilst retaining asymptotic guarantees of convergence. The conditional distribution is related to variable selection with candidate parents playing the role of covariates or inputs. We empirically examine the performance of the sampler using several simulated and real data examples. The proposed method gives robust results in diverse settings, outperforming several existing Bayesian and frequentist methods. In addition, our empirical results shed some light on the relative merits of Bayesian and constraint-based methods for structure learning.

我们提出了一种用于有向无环图(DAG)模型结构学习的吉布斯采样器。用于 DAG 学习的标准马尔可夫链蒙特卡罗算法是随机漫步 Metropolis-Hastings 采样器。这些采样器保证渐进收敛,但在探索结构学习中出现的大型图空间时,往往混合缓慢。在每一步中,我们提出的采样器都会从适当的条件分布中为多个节点抽取整套父节点。这提供了一种在图空间中进行大规模移动的有效方法,既能加快混合速度,又能保持渐进保证的收敛性。条件分布与变量选择有关,候选父节点扮演协变量或输入的角色。我们利用几个模拟和真实数据实例对采样器的性能进行了实证检验。所提出的方法在不同的环境下都能提供稳健的结果,其性能优于现有的几种贝叶斯和频数方法。此外,我们的实证结果还揭示了贝叶斯方法和基于约束的结构学习方法的相对优势。
{"title":"A Gibbs Sampler for Learning DAGs.","authors":"Robert J B Goudie, Sach Mukherjee","doi":"","DOIUrl":"","url":null,"abstract":"<p><p>We propose a Gibbs sampler for structure learning in directed acyclic graph (DAG) models. The standard Markov chain Monte Carlo algorithms used for learning DAGs are random-walk Metropolis-Hastings samplers. These samplers are guaranteed to converge asymptotically but often mix slowly when exploring the large graph spaces that arise in structure learning. In each step, the sampler we propose draws entire sets of parents for multiple nodes from the appropriate conditional distribution. This provides an efficient way to make large moves in graph space, permitting faster mixing whilst retaining asymptotic guarantees of convergence. The conditional distribution is related to variable selection with candidate parents playing the role of covariates or inputs. We empirically examine the performance of the sampler using several simulated and real data examples. The proposed method gives robust results in diverse settings, outperforming several existing Bayesian and frequentist methods. In addition, our empirical results shed some light on the relative merits of Bayesian and constraint-based methods for structure learning.</p>","PeriodicalId":50161,"journal":{"name":"Journal of Machine Learning Research","volume":null,"pages":null},"PeriodicalIF":6.0,"publicationDate":"2016-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5358773/pdf/emss-67582.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"34845238","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
On Quantile Regression in Reproducing Kernel Hilbert Spaces with Data Sparsity Constraint. 数据稀疏性约束下核希尔伯特空间再现的分位数回归。
IF 6 3区 计算机科学 Q1 Mathematics Pub Date : 2016-04-01
Chong Zhang, Yufeng Liu, Yichao Wu

For spline regressions, it is well known that the choice of knots is crucial for the performance of the estimator. As a general learning framework covering the smoothing splines, learning in a Reproducing Kernel Hilbert Space (RKHS) has a similar issue. However, the selection of training data points for kernel functions in the RKHS representation has not been carefully studied in the literature. In this paper we study quantile regression as an example of learning in a RKHS. In this case, the regular squared norm penalty does not perform training data selection. We propose a data sparsity constraint that imposes thresholding on the kernel function coefficients to achieve a sparse kernel function representation. We demonstrate that the proposed data sparsity method can have competitive prediction performance for certain situations, and have comparable performance in other cases compared to that of the traditional squared norm penalty. Therefore, the data sparsity method can serve as a competitive alternative to the squared norm penalty method. Some theoretical properties of our proposed method using the data sparsity constraint are obtained. Both simulated and real data sets are used to demonstrate the usefulness of our data sparsity constraint.

对于样条回归,众所周知,结点的选择对估计器的性能至关重要。作为覆盖光滑样条的一般学习框架,在再现核希尔伯特空间(RKHS)中学习也存在类似的问题。然而,在RKHS表示中,核函数的训练数据点的选择并没有在文献中得到仔细的研究。本文研究了分位数回归作为RKHS学习的一个例子。在这种情况下,正则平方范数惩罚不执行训练数据选择。我们提出了一种数据稀疏性约束,对核函数系数施加阈值以实现稀疏核函数表示。我们证明了所提出的数据稀疏性方法在某些情况下可以具有竞争性的预测性能,并且在其他情况下与传统的平方范数惩罚相比具有可比的性能。因此,数据稀疏性方法可以作为平方范数惩罚方法的竞争性替代方法。给出了该方法在数据稀疏性约束下的一些理论性质。模拟数据集和真实数据集都被用来证明我们的数据稀疏性约束的有效性。
{"title":"On Quantile Regression in Reproducing Kernel Hilbert Spaces with Data Sparsity Constraint.","authors":"Chong Zhang,&nbsp;Yufeng Liu,&nbsp;Yichao Wu","doi":"","DOIUrl":"","url":null,"abstract":"<p><p>For spline regressions, it is well known that the choice of knots is crucial for the performance of the estimator. As a general learning framework covering the smoothing splines, learning in a Reproducing Kernel Hilbert Space (RKHS) has a similar issue. However, the selection of training data points for kernel functions in the RKHS representation has not been carefully studied in the literature. In this paper we study quantile regression as an example of learning in a RKHS. In this case, the regular squared norm penalty does not perform training data selection. We propose a data sparsity constraint that imposes thresholding on the kernel function coefficients to achieve a sparse kernel function representation. We demonstrate that the proposed data sparsity method can have competitive prediction performance for certain situations, and have comparable performance in other cases compared to that of the traditional squared norm penalty. Therefore, the data sparsity method can serve as a competitive alternative to the squared norm penalty method. Some theoretical properties of our proposed method using the data sparsity constraint are obtained. Both simulated and real data sets are used to demonstrate the usefulness of our data sparsity constraint.</p>","PeriodicalId":50161,"journal":{"name":"Journal of Machine Learning Research","volume":null,"pages":null},"PeriodicalIF":6.0,"publicationDate":"2016-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4850041/pdf/nihms729829.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"34446126","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Journal of Machine Learning Research
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1