首页 > 最新文献

Journal of Machine Learning Research最新文献

英文 中文
Multiplicative Multitask Feature Learning. 乘法多任务特征学习
IF 6 3区 计算机科学 Q1 Mathematics Pub Date : 2016-04-01
Xin Wang, Jinbo Bi, Shipeng Yu, Jiangwen Sun, Minghu Song

We investigate a general framework of multiplicative multitask feature learning which decomposes individual task's model parameters into a multiplication of two components. One of the components is used across all tasks and the other component is task-specific. Several previous methods can be proved to be special cases of our framework. We study the theoretical properties of this framework when different regularization conditions are applied to the two decomposed components. We prove that this framework is mathematically equivalent to the widely used multitask feature learning methods that are based on a joint regularization of all model parameters, but with a more general form of regularizers. Further, an analytical formula is derived for the across-task component as related to the task-specific component for all these regularizers, leading to a better understanding of the shrinkage effects of different regularizers. Study of this framework motivates new multitask learning algorithms. We propose two new learning formulations by varying the parameters in the proposed framework. An efficient blockwise coordinate descent algorithm is developed suitable for solving the entire family of formulations with rigorous convergence analysis. Simulation studies have identified the statistical properties of data that would be in favor of the new formulations. Extensive empirical studies on various classification and regression benchmark data sets have revealed the relative advantages of the two new formulations by comparing with the state of the art, which provides instructive insights into the feature learning problem with multiple tasks.

我们研究了乘法多任务特征学习的一般框架,该框架将单个任务的模型参数分解为两个分量的乘法。其中一个分量用于所有任务,另一个分量则针对特定任务。之前的几种方法都可以证明是我们框架的特例。我们研究了对两个分解分量应用不同正则化条件时该框架的理论特性。我们证明,该框架在数学上等同于广泛使用的多任务特征学习方法,后者基于所有模型参数的联合正则化,但正则化形式更为普遍。此外,对于所有这些正则化器,我们还推导出了跨任务分量与特定任务分量的分析公式,从而更好地理解了不同正则化器的收缩效果。对这一框架的研究激发了新的多任务学习算法。我们通过改变拟议框架中的参数,提出了两种新的学习方案。我们开发了一种高效的顺时针坐标下降算法,适用于求解整个公式系列,并进行了严格的收敛分析。模拟研究确定了有利于新公式的数据统计特性。在各种分类和回归基准数据集上进行的广泛实证研究,通过与现有技术的比较,揭示了这两种新公式的相对优势,从而为多任务特征学习问题提供了具有启发性的见解。
{"title":"Multiplicative Multitask Feature Learning.","authors":"Xin Wang, Jinbo Bi, Shipeng Yu, Jiangwen Sun, Minghu Song","doi":"","DOIUrl":"","url":null,"abstract":"<p><p>We investigate a general framework of multiplicative multitask feature learning which decomposes individual task's model parameters into a multiplication of two components. One of the components is used across all tasks and the other component is task-specific. Several previous methods can be proved to be special cases of our framework. We study the theoretical properties of this framework when different regularization conditions are applied to the two decomposed components. We prove that this framework is mathematically equivalent to the widely used multitask feature learning methods that are based on a joint regularization of all model parameters, but with a more general form of regularizers. Further, an analytical formula is derived for the across-task component as related to the task-specific component for all these regularizers, leading to a better understanding of the shrinkage effects of different regularizers. Study of this framework motivates new multitask learning algorithms. We propose two new learning formulations by varying the parameters in the proposed framework. An efficient blockwise coordinate descent algorithm is developed suitable for solving the entire family of formulations with rigorous convergence analysis. Simulation studies have identified the statistical properties of data that would be in favor of the new formulations. Extensive empirical studies on various classification and regression benchmark data sets have revealed the relative advantages of the two new formulations by comparing with the state of the art, which provides instructive insights into the feature learning problem with multiple tasks.</p>","PeriodicalId":50161,"journal":{"name":"Journal of Machine Learning Research","volume":null,"pages":null},"PeriodicalIF":6.0,"publicationDate":"2016-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5395291/pdf/nihms814714.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"34930505","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Cross-corpora unsupervised learning of trajectories in autism spectrum disorders 自闭症谱系障碍轨迹的跨语料库无监督学习
IF 6 3区 计算机科学 Q1 Mathematics Pub Date : 2016-01-01 DOI: 10.5555/2946645.3007086
ElibolHuseyin Melih, NguyenVincent, LindermanScott, JohnsonMatthew, HashmiAmna, Doshi-VelezFinale
Patients with developmental disorders, such as autism spectrum disorder (ASD), present with symptoms that change with time even if the named diagnosis remains fixed. For example, language impairmen...
患有发育障碍的患者,如自闭症谱系障碍(ASD),即使命名的诊断仍然固定,其症状也会随着时间而变化。例如,有语言障碍的人……
{"title":"Cross-corpora unsupervised learning of trajectories in autism spectrum disorders","authors":"ElibolHuseyin Melih, NguyenVincent, LindermanScott, JohnsonMatthew, HashmiAmna, Doshi-VelezFinale","doi":"10.5555/2946645.3007086","DOIUrl":"https://doi.org/10.5555/2946645.3007086","url":null,"abstract":"Patients with developmental disorders, such as autism spectrum disorder (ASD), present with symptoms that change with time even if the named diagnosis remains fixed. For example, language impairmen...","PeriodicalId":50161,"journal":{"name":"Journal of Machine Learning Research","volume":null,"pages":null},"PeriodicalIF":6.0,"publicationDate":"2016-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"71138847","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Guarding against Spurious Discoveries in High Dimensions. 在高维中防范虚假的发现。
IF 6 3区 计算机科学 Q1 Mathematics Pub Date : 2016-01-01
Jianqing Fan, Wen-Xin Zhou

Many data-mining and statistical machine learning algorithms have been developed to select a subset of covariates to associate with a response variable. Spurious discoveries can easily arise in high-dimensional data analysis due to enormous possibilities of such selections. How can we know statistically our discoveries better than those by chance? In this paper, we define a measure of goodness of spurious fit, which shows how good a response variable can be fitted by an optimally selected subset of covariates under the null model, and propose a simple and effective LAMM algorithm to compute it. It coincides with the maximum spurious correlation for linear models and can be regarded as a generalized maximum spurious correlation. We derive the asymptotic distribution of such goodness of spurious fit for generalized linear models and L1-regression. Such an asymptotic distribution depends on the sample size, ambient dimension, the number of variables used in the fit, and the covariance information. It can be consistently estimated by multiplier bootstrapping and used as a benchmark to guard against spurious discoveries. It can also be applied to model selection, which considers only candidate models with goodness of fits better than those by spurious fits. The theory and method are convincingly illustrated by simulated examples and an application to the binary outcomes from German Neuroblastoma Trials.

已经开发了许多数据挖掘和统计机器学习算法来选择协变量的子集与响应变量相关联。由于这种选择的巨大可能性,在高维数据分析中很容易出现虚假的发现。我们怎么能在统计上比偶然发现更了解我们的发现呢?在本文中,我们定义了伪拟合优度的度量,它显示了在零模型下,一个最优选择的协变量子集可以很好地拟合一个响应变量,并提出了一个简单有效的LAMM算法来计算它。它与线性模型的最大伪相关一致,可以看作是广义的最大伪相关。我们得到了广义线性模型和l1回归的伪拟合优度的渐近分布。这种渐近分布取决于样本大小、环境维度、拟合中使用的变量数量和协方差信息。它可以通过乘法器自举来一致地估计,并用作防止虚假发现的基准。它还可以应用于模型选择,即只考虑拟合优度优于伪拟合的候选模型。通过模拟实例和德国神经母细胞瘤试验二元结果的应用,令人信服地说明了该理论和方法。
{"title":"Guarding against Spurious Discoveries in High Dimensions.","authors":"Jianqing Fan,&nbsp;Wen-Xin Zhou","doi":"","DOIUrl":"","url":null,"abstract":"<p><p>Many data-mining and statistical machine learning algorithms have been developed to select a subset of covariates to associate with a response variable. Spurious discoveries can easily arise in high-dimensional data analysis due to enormous possibilities of such selections. How can we know statistically our discoveries better than those by chance? In this paper, we define a measure of goodness of spurious fit, which shows how good a response variable can be fitted by an optimally selected subset of covariates under the null model, and propose a simple and effective LAMM algorithm to compute it. It coincides with the maximum spurious correlation for linear models and can be regarded as a generalized maximum spurious correlation. We derive the asymptotic distribution of such goodness of spurious fit for generalized linear models and <i>L</i><sub>1</sub>-regression. Such an asymptotic distribution depends on the sample size, ambient dimension, the number of variables used in the fit, and the covariance information. It can be consistently estimated by multiplier bootstrapping and used as a benchmark to guard against spurious discoveries. It can also be applied to model selection, which considers only candidate models with goodness of fits better than those by spurious fits. The theory and method are convincingly illustrated by simulated examples and an application to the binary outcomes from German Neuroblastoma Trials.</p>","PeriodicalId":50161,"journal":{"name":"Journal of Machine Learning Research","volume":null,"pages":null},"PeriodicalIF":6.0,"publicationDate":"2016-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5603346/pdf/nihms842539.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"35535102","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Structure discovery in Bayesian networks by sampling partial orders 抽样偏阶贝叶斯网络的结构发现
IF 6 3区 计算机科学 Q1 Mathematics Pub Date : 2016-01-01 DOI: 10.5555/2946645.2946702
NiinimäkiTeppo, ParviainenPekka, KoivistoMikko
We present methods based on Metropolis-coupled Markov chain Monte Carlo (MC3) and annealed importance sampling (AIS) for estimating the posterior distribution of Bayesian networks. The methods draw...
提出了基于大都市耦合马尔可夫链蒙特卡罗(MC3)和退火重要抽样(AIS)的贝叶斯网络后验分布估计方法。方法绘制…
{"title":"Structure discovery in Bayesian networks by sampling partial orders","authors":"NiinimäkiTeppo, ParviainenPekka, KoivistoMikko","doi":"10.5555/2946645.2946702","DOIUrl":"https://doi.org/10.5555/2946645.2946702","url":null,"abstract":"We present methods based on Metropolis-coupled Markov chain Monte Carlo (MC3) and annealed importance sampling (AIS) for estimating the posterior distribution of Bayesian networks. The methods draw...","PeriodicalId":50161,"journal":{"name":"Journal of Machine Learning Research","volume":null,"pages":null},"PeriodicalIF":6.0,"publicationDate":"2016-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"71138831","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Choice of V for V-fold cross-validation in least-squares density estimation 最小二乘密度估计中V-fold交叉验证V值的选择
IF 6 3区 计算机科学 Q1 Mathematics Pub Date : 2016-01-01 DOI: 10.5555/2946645.3053490
ArlotSylvain, LerasleMatthieu
This paper studies V-fold cross-validation for model selection in least-squares density estimation. The goal is to provide theoretical grounds for choosing V in order to minimize the least-squares ...
研究了最小二乘密度估计模型选择的v折交叉验证方法。目标是为选择V以最小化最小二乘提供理论依据……
{"title":"Choice of V for V-fold cross-validation in least-squares density estimation","authors":"ArlotSylvain, LerasleMatthieu","doi":"10.5555/2946645.3053490","DOIUrl":"https://doi.org/10.5555/2946645.3053490","url":null,"abstract":"This paper studies V-fold cross-validation for model selection in least-squares density estimation. The goal is to provide theoretical grounds for choosing V in order to minimize the least-squares ...","PeriodicalId":50161,"journal":{"name":"Journal of Machine Learning Research","volume":null,"pages":null},"PeriodicalIF":6.0,"publicationDate":"2016-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"71138913","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MOCCA: Mirrored Convex/Concave Optimization for Nonconvex Composite Functions. 非凸复合函数的镜像凸/凹优化。
IF 6 3区 计算机科学 Q1 Mathematics Pub Date : 2016-01-01
Rina Foygel Barber, Emil Y Sidky

Many optimization problems arising in high-dimensional statistics decompose naturally into a sum of several terms, where the individual terms are relatively simple but the composite objective function can only be optimized with iterative algorithms. In this paper, we are interested in optimization problems of the form F(Kx) + G(x), where K is a fixed linear transformation, while F and G are functions that may be nonconvex and/or nondifferentiable. In particular, if either of the terms are nonconvex, existing alternating minimization techniques may fail to converge; other types of existing approaches may instead be unable to handle nondifferentiability. We propose the MOCCA (mirrored convex/concave) algorithm, a primal/dual optimization approach that takes a local convex approximation to each term at every iteration. Inspired by optimization problems arising in computed tomography (CT) imaging, this algorithm can handle a range of nonconvex composite optimization problems, and offers theoretical guarantees for convergence when the overall problem is approximately convex (that is, any concavity in one term is balanced out by convexity in the other term). Empirical results show fast convergence for several structured signal recovery problems.

高维统计中出现的许多优化问题自然分解为若干项的和,其中单个项相对简单,而复合目标函数只能通过迭代算法进行优化。本文研究F(Kx) + G(x)形式的最优化问题,其中K是一个固定的线性变换,而F和G是非凸和/或不可微的函数。特别是,如果任何一项都是非凸的,现有的交替最小化技术可能无法收敛;其他类型的现有方法可能无法处理不可微性。我们提出了MOCCA(镜像凸/凹)算法,这是一种原始/对偶优化方法,在每次迭代中对每个项进行局部凸逼近。受计算机断层扫描(CT)成像中的优化问题的启发,该算法可以处理一系列非凸复合优化问题,并为整体问题近似凸时的收敛性提供理论保证(即一项的任何凹性被另一项的凸性抵消)。实验结果表明,该方法具有较快的收敛速度。
{"title":"MOCCA: Mirrored Convex/Concave Optimization for Nonconvex Composite Functions.","authors":"Rina Foygel Barber,&nbsp;Emil Y Sidky","doi":"","DOIUrl":"","url":null,"abstract":"<p><p>Many optimization problems arising in high-dimensional statistics decompose naturally into a sum of several terms, where the individual terms are relatively simple but the composite objective function can only be optimized with iterative algorithms. In this paper, we are interested in optimization problems of the form F(<i>Kx</i>) + G(<i>x</i>), where <i>K</i> is a fixed linear transformation, while F and G are functions that may be nonconvex and/or nondifferentiable. In particular, if either of the terms are nonconvex, existing alternating minimization techniques may fail to converge; other types of existing approaches may instead be unable to handle nondifferentiability. We propose the MOCCA (mirrored convex/concave) algorithm, a primal/dual optimization approach that takes a local convex approximation to each term at every iteration. Inspired by optimization problems arising in computed tomography (CT) imaging, this algorithm can handle a range of nonconvex composite optimization problems, and offers theoretical guarantees for convergence when the overall problem is approximately convex (that is, any concavity in one term is balanced out by convexity in the other term). Empirical results show fast convergence for several structured signal recovery problems.</p>","PeriodicalId":50161,"journal":{"name":"Journal of Machine Learning Research","volume":null,"pages":null},"PeriodicalIF":6.0,"publicationDate":"2016-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5789814/pdf/nihms870482.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"35785739","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Support Vector Hazards Machine: A Counting Process Framework for Learning Risk Scores for Censored Outcomes. 支持向量危险机:一个计算过程框架,用于学习审查结果的风险评分。
IF 6 3区 计算机科学 Q1 Mathematics Pub Date : 2016-01-01 Epub Date: 2016-08-01
Yuanjia Wang, Tianle Chen, Donglin Zeng

Learning risk scores to predict dichotomous or continuous outcomes using machine learning approaches has been studied extensively. However, how to learn risk scores for time-to-event outcomes subject to right censoring has received little attention until recently. Existing approaches rely on inverse probability weighting or rank-based regression, which may be inefficient. In this paper, we develop a new support vector hazards machine (SVHM) approach to predict censored outcomes. Our method is based on predicting the counting process associated with the time-to-event outcomes among subjects at risk via a series of support vector machines. Introducing counting processes to represent time-to-event data leads to a connection between support vector machines in supervised learning and hazards regression in standard survival analysis. To account for different at risk populations at observed event times, a time-varying offset is used in estimating risk scores. The resulting optimization is a convex quadratic programming problem that can easily incorporate non-linearity using kernel trick. We demonstrate an interesting link from the profiled empirical risk function of SVHM to the Cox partial likelihood. We then formally show that SVHM is optimal in discriminating covariate-specific hazard function from population average hazard function, and establish the consistency and learning rate of the predicted risk using the estimated risk scores. Simulation studies show improved prediction accuracy of the event times using SVHM compared to existing machine learning methods and standard conventional approaches. Finally, we analyze two real world biomedical study data where we use clinical markers and neuroimaging biomarkers to predict age-at-onset of a disease, and demonstrate superiority of SVHM in distinguishing high risk versus low risk subjects.

使用机器学习方法来预测二分或连续结果的学习风险评分已经被广泛研究。然而,直到最近,如何在严格审查的情况下学习时间到事件结果的风险评分才受到关注。现有的方法依赖于逆概率加权或基于秩的回归,这可能是低效的。在本文中,我们开发了一种新的支持向量危险机(SVHM)方法来预测审查结果。我们的方法基于通过一系列支持向量机预测风险受试者中与事件发生时间结果相关的计数过程。引入计数过程来表示事件数据的时间,导致了监督学习中的支持向量机和标准生存分析中的危险回归之间的联系。为了说明观察到的事件时间的不同风险人群,在估计风险评分时使用了时变偏移。由此产生的优化是一个凸二次规划问题,可以使用核技巧很容易地结合非线性。我们证明了SVHM的经验风险函数与Cox偏似然之间的有趣联系。然后,我们正式证明了SVHM在区分协变量特定风险函数和人群平均风险函数方面是最优的,并使用估计的风险得分建立了预测风险的一致性和学习率。仿真研究表明,与现有的机器学习方法和标准的传统方法相比,使用SVHM可以提高事件时间的预测精度。最后,我们分析了两个真实世界的生物医学研究数据,其中我们使用临床标志物和神经成像生物标志物来预测疾病发作时的年龄,并证明了SVHM在区分高风险和低风险受试者方面的优越性。
{"title":"Support Vector Hazards Machine: A Counting Process Framework for Learning Risk Scores for Censored Outcomes.","authors":"Yuanjia Wang,&nbsp;Tianle Chen,&nbsp;Donglin Zeng","doi":"","DOIUrl":"","url":null,"abstract":"<p><p>Learning risk scores to predict dichotomous or continuous outcomes using machine learning approaches has been studied extensively. However, how to learn risk scores for time-to-event outcomes subject to right censoring has received little attention until recently. Existing approaches rely on inverse probability weighting or rank-based regression, which may be inefficient. In this paper, we develop a new support vector hazards machine (SVHM) approach to predict censored outcomes. Our method is based on predicting the counting process associated with the time-to-event outcomes among subjects at risk via a series of support vector machines. Introducing counting processes to represent time-to-event data leads to a connection between support vector machines in supervised learning and hazards regression in standard survival analysis. To account for different at risk populations at observed event times, a time-varying offset is used in estimating risk scores. The resulting optimization is a convex quadratic programming problem that can easily incorporate non-linearity using kernel trick. We demonstrate an interesting link from the profiled empirical risk function of SVHM to the Cox partial likelihood. We then formally show that SVHM is optimal in discriminating covariate-specific hazard function from population average hazard function, and establish the consistency and learning rate of the predicted risk using the estimated risk scores. Simulation studies show improved prediction accuracy of the event times using SVHM compared to existing machine learning methods and standard conventional approaches. Finally, we analyze two real world biomedical study data where we use clinical markers and neuroimaging biomarkers to predict age-at-onset of a disease, and demonstrate superiority of SVHM in distinguishing high risk versus low risk subjects.</p>","PeriodicalId":50161,"journal":{"name":"Journal of Machine Learning Research","volume":null,"pages":null},"PeriodicalIF":6.0,"publicationDate":"2016-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5210213/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"71434774","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
The optimal sample complexity OF PAC learning PAC学习的最优样本复杂度
IF 6 3区 计算机科学 Q1 Mathematics Pub Date : 2016-01-01 DOI: 10.5555/2946645.2946683
HannekeSteve
This work establishes a new upper bound on the number of samples sufficient for PAC learning in the realizable case. The bound matches known lower bounds up to numerical constant factors. This solv...
这项工作建立了在可实现的情况下足以进行PAC学习的样本数量的新上界。该边界与已知的下界匹配,直到数值常数因子。这求解…
{"title":"The optimal sample complexity OF PAC learning","authors":"HannekeSteve","doi":"10.5555/2946645.2946683","DOIUrl":"https://doi.org/10.5555/2946645.2946683","url":null,"abstract":"This work establishes a new upper bound on the number of samples sufficient for PAC learning in the realizable case. The bound matches known lower bounds up to numerical constant factors. This solv...","PeriodicalId":50161,"journal":{"name":"Journal of Machine Learning Research","volume":null,"pages":null},"PeriodicalIF":6.0,"publicationDate":"2016-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"71138825","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Gradients weights improve regression and classification 梯度权重改进了回归和分类
IF 6 3区 计算机科学 Q1 Mathematics Pub Date : 2016-01-01 DOI: 10.5555/2946645.2946667
KpotufeSamory, BoulariasAbdeslam, SchultzThomas, KimKyoungok
In regression problems over Rd, the unknown function f often varies more in some coordinates than in others. We show that weighting each coordinate i according to an estimate of the variation of f ...
在Rd上的回归问题中,未知函数f在某些坐标上的变化往往比在其他坐标上的变化更大。我们表明,根据f的变化估计对每个坐标i进行加权…
{"title":"Gradients weights improve regression and classification","authors":"KpotufeSamory, BoulariasAbdeslam, SchultzThomas, KimKyoungok","doi":"10.5555/2946645.2946667","DOIUrl":"https://doi.org/10.5555/2946645.2946667","url":null,"abstract":"In regression problems over Rd, the unknown function f often varies more in some coordinates than in others. We show that weighting each coordinate i according to an estimate of the variation of f ...","PeriodicalId":50161,"journal":{"name":"Journal of Machine Learning Research","volume":null,"pages":null},"PeriodicalIF":6.0,"publicationDate":"2016-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"71138716","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Fused lasso approach in regression coefficients clustering 回归系数聚类的融合套索方法
IF 6 3区 计算机科学 Q1 Mathematics Pub Date : 2016-01-01 DOI: 10.5555/2946645.3007066
TangLu
As data sets of related studies become more easily accessible, combining data sets of similar studies is often undertaken in practice to achieve a larger sample size and higher power. A major chall...
随着相关研究的数据集越来越容易获取,在实践中往往会将类似研究的数据集进行组合,以获得更大的样本量和更高的功率。一个重大挑战……
{"title":"Fused lasso approach in regression coefficients clustering","authors":"TangLu","doi":"10.5555/2946645.3007066","DOIUrl":"https://doi.org/10.5555/2946645.3007066","url":null,"abstract":"As data sets of related studies become more easily accessible, combining data sets of similar studies is often undertaken in practice to achieve a larger sample size and higher power. A major chall...","PeriodicalId":50161,"journal":{"name":"Journal of Machine Learning Research","volume":null,"pages":null},"PeriodicalIF":6.0,"publicationDate":"2016-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"71138837","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
期刊
Journal of Machine Learning Research
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1