首页 > 最新文献

J. Mach. Learn. Res.最新文献

英文 中文
Learning theory of randomized Kaczmarz algorithm 随机化Kaczmarz算法的学习理论
Pub Date : 2015-01-01 DOI: 10.5555/2789272.2912105
Junhong Lin, Ding-Xuan Zhou
A relaxed randomized Kaczmarz algorithm is investigated in a least squares regression setting by a learning theory approach. When the sampling values are accurate and the regression function (conditional means) is linear, such an algorithm has been well studied in the community of non-uniform sampling. In this paper, we are mainly interested in the different case of either noisy random measurements or a nonlinear regression function. In this case, we show that relaxation is needed. A necessary and sufficient condition on the sequence of relaxation parameters or step sizes for the convergence of the algorithm in expectation is presented. Moreover, polynomial rates of convergence, both in expectation and in probability, are provided explicitly. As a result, the almost sure convergence of the algorithm is proved by applying the Borel-Cantelli Lemma.
利用学习理论方法研究了最小二乘回归环境下的松弛随机化Kaczmarz算法。在采样值准确且回归函数(条件均值)为线性的情况下,该算法在非均匀采样界得到了较好的研究。在本文中,我们主要对噪声随机测量或非线性回归函数的不同情况感兴趣。在这种情况下,我们表明松弛是必要的。给出了松弛参数序列或步长在期望范围内收敛的充分必要条件。此外,还明确地给出了期望和概率的多项式收敛率。最后,利用Borel-Cantelli引理证明了该算法的收敛性。
{"title":"Learning theory of randomized Kaczmarz algorithm","authors":"Junhong Lin, Ding-Xuan Zhou","doi":"10.5555/2789272.2912105","DOIUrl":"https://doi.org/10.5555/2789272.2912105","url":null,"abstract":"A relaxed randomized Kaczmarz algorithm is investigated in a least squares regression setting by a learning theory approach. When the sampling values are accurate and the regression function (conditional means) is linear, such an algorithm has been well studied in the community of non-uniform sampling. In this paper, we are mainly interested in the different case of either noisy random measurements or a nonlinear regression function. In this case, we show that relaxation is needed. A necessary and sufficient condition on the sequence of relaxation parameters or step sizes for the convergence of the algorithm in expectation is presented. Moreover, polynomial rates of convergence, both in expectation and in probability, are provided explicitly. As a result, the almost sure convergence of the algorithm is proved by applying the Borel-Cantelli Lemma.","PeriodicalId":14794,"journal":{"name":"J. Mach. Learn. Res.","volume":"188 1","pages":"3341-3365"},"PeriodicalIF":0.0,"publicationDate":"2015-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"73265887","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 30
CEKA: a tool for mining the wisdom of crowds CEKA:挖掘群体智慧的工具
Pub Date : 2015-01-01 DOI: 10.5555/2789272.2912090
J. Zhang, V. Sheng, B. Nicholson, Xindong Wu
CEKA is a software package for developers and researchers to mine the wisdom of crowds. It makes the entire knowledge discovery procedure much easier, including analyzing qualities of workers, simulating labeling behaviors, inferring true class labels of instances, filtering and correcting mislabeled instances (noise), building learning models and evaluating them. It integrates a set of state-of-the-art inference algorithms, a set of general noise handling algorithms, and abundant functions for model training and evaluation. CEKA is written in Java with core classes being compatible with the well-known machine learning tool WEKA, which makes the utilization of the functions in WEKA much easier.
CEKA是开发人员和研究人员挖掘群体智慧的软件包。它使整个知识发现过程变得更加容易,包括分析工作者的素质,模拟标记行为,推断实例的真实类标签,过滤和纠正错误标记的实例(噪声),建立学习模型并对其进行评估。它集成了一套最先进的推理算法,一套通用的噪声处理算法,以及丰富的模型训练和评估功能。CEKA是用Java编写的,其核心类与著名的机器学习工具WEKA兼容,这使得使用WEKA中的函数更加容易。
{"title":"CEKA: a tool for mining the wisdom of crowds","authors":"J. Zhang, V. Sheng, B. Nicholson, Xindong Wu","doi":"10.5555/2789272.2912090","DOIUrl":"https://doi.org/10.5555/2789272.2912090","url":null,"abstract":"CEKA is a software package for developers and researchers to mine the wisdom of crowds. It makes the entire knowledge discovery procedure much easier, including analyzing qualities of workers, simulating labeling behaviors, inferring true class labels of instances, filtering and correcting mislabeled instances (noise), building learning models and evaluating them. It integrates a set of state-of-the-art inference algorithms, a set of general noise handling algorithms, and abundant functions for model training and evaluation. CEKA is written in Java with core classes being compatible with the well-known machine learning tool WEKA, which makes the utilization of the functions in WEKA much easier.","PeriodicalId":14794,"journal":{"name":"J. Mach. Learn. Res.","volume":"10 1","pages":"2853-2858"},"PeriodicalIF":0.0,"publicationDate":"2015-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"85227326","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 29
Decision boundary for discrete Bayesian network classifiers 离散贝叶斯网络分类器的决策边界
Pub Date : 2015-01-01 DOI: 10.5555/2789272.2912086
Gherardo Varando, C. Bielza, P. Larrañaga
Bayesian network classifiers are a powerful machine learning tool. In order to evaluate the expressive power of these models, we compute families of polynomials that sign-represent decision functions induced by Bayesian network classifiers. We prove that those families are linear combinations of products of Lagrange basis polynomials. In absence of V-structures in the predictor sub-graph, we are also able to prove that this family of polynomials does indeed characterize the specific classifier considered. We then use this representation to bound the number of decision functions representable by Bayesian network classifiers with a given structure.
贝叶斯网络分类器是一种强大的机器学习工具。为了评估这些模型的表达能力,我们计算了由贝叶斯网络分类器诱导的符号表示决策函数的多项式族。我们证明了这些族是拉格朗日基多项式乘积的线性组合。在预测子图中没有v结构的情况下,我们也能够证明这个多项式族确实表征了所考虑的特定分类器。然后,我们使用这种表示来约束具有给定结构的贝叶斯网络分类器可表示的决策函数的数量。
{"title":"Decision boundary for discrete Bayesian network classifiers","authors":"Gherardo Varando, C. Bielza, P. Larrañaga","doi":"10.5555/2789272.2912086","DOIUrl":"https://doi.org/10.5555/2789272.2912086","url":null,"abstract":"Bayesian network classifiers are a powerful machine learning tool. In order to evaluate the expressive power of these models, we compute families of polynomials that sign-represent decision functions induced by Bayesian network classifiers. We prove that those families are linear combinations of products of Lagrange basis polynomials. In absence of V-structures in the predictor sub-graph, we are also able to prove that this family of polynomials does indeed characterize the specific classifier considered. We then use this representation to bound the number of decision functions representable by Bayesian network classifiers with a given structure.","PeriodicalId":14794,"journal":{"name":"J. Mach. Learn. Res.","volume":"62 1","pages":"2725-2749"},"PeriodicalIF":0.0,"publicationDate":"2015-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"77806302","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 22
Geometric intuition and algorithms for Ev-SVM Ev-SVM的几何直觉与算法
Pub Date : 2015-01-01 DOI: 10.5555/2789272.2789283
Á. Jiménez, A. Takeda, J. Lázaro
In this work we address the Ev-SVM model proposed by Perez-Cruz et al. as an extension of the traditional v support vector classification model (v-SVM). Through an enhancement of the range of admissible values for the regularization parameter v, the Ev-SVM has been shown to be able to produce a wider variety of decision functions, giving rise to a better adaptability to the data. However, while a clear and intuitive geometric interpretation can be given for the v-SVM model as a nearest-point problem in reduced convex hulls (RCH-NPP), no previous work has been made in developing such intuition for the Ev-SVM model. In this paper we show how Ev-SVM can be reformulated as a geometrical problem that generalizes RCH-NPP, providing new insights into this model. Under this novel point of view, we propose the RapMinos algorithm, able to solve Ev-SVM more efficiently than the current methods. Furthermore, we show how RapMinos is able to address the Ev-SVM model for any choice of regularization norm lp ≥1 seamlessly, which further extends the SVM model flexibility beyond the usual Ev-SVM models.
在这项工作中,我们解决了Perez-Cruz等人提出的Ev-SVM模型,作为传统v支持向量分类模型(v- svm)的扩展。通过增强正则化参数v的容许值范围,Ev-SVM已被证明能够产生更广泛的决策函数,从而对数据具有更好的适应性。然而,虽然v-SVM模型可以作为简化凸包(RCH-NPP)的最近点问题给出清晰直观的几何解释,但之前没有工作为Ev-SVM模型开发这种直觉。在本文中,我们展示了Ev-SVM如何可以被重新表述为概括RCH-NPP的几何问题,为该模型提供了新的见解。在这种新观点下,我们提出了RapMinos算法,能够比现有方法更有效地求解Ev-SVM。此外,我们展示了RapMinos如何能够无缝地解决Ev-SVM模型中任意正则化范数lp≥1的选择,这进一步扩展了SVM模型的灵活性,超出了通常的Ev-SVM模型。
{"title":"Geometric intuition and algorithms for Ev-SVM","authors":"Á. Jiménez, A. Takeda, J. Lázaro","doi":"10.5555/2789272.2789283","DOIUrl":"https://doi.org/10.5555/2789272.2789283","url":null,"abstract":"In this work we address the Ev-SVM model proposed by Perez-Cruz et al. as an extension of the traditional v support vector classification model (v-SVM). Through an enhancement of the range of admissible values for the regularization parameter v, the Ev-SVM has been shown to be able to produce a wider variety of decision functions, giving rise to a better adaptability to the data. However, while a clear and intuitive geometric interpretation can be given for the v-SVM model as a nearest-point problem in reduced convex hulls (RCH-NPP), no previous work has been made in developing such intuition for the Ev-SVM model. In this paper we show how Ev-SVM can be reformulated as a geometrical problem that generalizes RCH-NPP, providing new insights into this model. Under this novel point of view, we propose the RapMinos algorithm, able to solve Ev-SVM more efficiently than the current methods. Furthermore, we show how RapMinos is able to address the Ev-SVM model for any choice of regularization norm lp ≥1 seamlessly, which further extends the SVM model flexibility beyond the usual Ev-SVM models.","PeriodicalId":14794,"journal":{"name":"J. Mach. Learn. Res.","volume":"29 1","pages":"323-369"},"PeriodicalIF":0.0,"publicationDate":"2015-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"81136110","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
PAC optimal MDP planning with application to invasive species management PAC优化MDP规划及其在入侵物种管理中的应用
Pub Date : 2015-01-01 DOI: 10.5555/2789272.2912119
Majid Alkaee Taleghan, Thomas G. Dietterich, Mark Crowley, K. Hall, H. Albers
In a simulator-defined MDP, the Markovian dynamics and rewards are provided in the form of a simulator from which samples can be drawn. This paper studies MDP planning algorithms that attempt to minimize the number of simulator calls before terminating and outputting a policy that is approximately optimal with high probability. The paper introduces two heuristics for efficient exploration and an improved confidence interval that enables earlier termination with probabilistic guarantees. We prove that the heuristics and the confidence interval are sound and produce with high probability an approximately optimal policy in polynomial time. Experiments on two benchmark problems and two instances of an invasive species management problem show that the improved confidence intervals and the new search heuristics yield reductions of between 8% and 47% in the number of simulator calls required to reach near-optimal policies.
在模拟器定义的MDP中,马尔可夫动态和奖励以模拟器的形式提供,从中可以提取样本。本文研究了MDP规划算法,该算法试图在终止和输出高概率近似最优策略之前最小化模拟器调用的数量。本文介绍了两种有效探索的启发式方法和一种改进的置信区间,可以在概率保证的情况下实现更早的终止。我们证明了启发式和置信区间是合理的,并在多项式时间内以高概率产生近似最优策略。在两个基准问题和两个入侵物种管理问题实例上的实验表明,改进的置信区间和新的搜索启发式方法使达到接近最优策略所需的模拟器调用次数减少了8%至47%。
{"title":"PAC optimal MDP planning with application to invasive species management","authors":"Majid Alkaee Taleghan, Thomas G. Dietterich, Mark Crowley, K. Hall, H. Albers","doi":"10.5555/2789272.2912119","DOIUrl":"https://doi.org/10.5555/2789272.2912119","url":null,"abstract":"In a simulator-defined MDP, the Markovian dynamics and rewards are provided in the form of a simulator from which samples can be drawn. This paper studies MDP planning algorithms that attempt to minimize the number of simulator calls before terminating and outputting a policy that is approximately optimal with high probability. The paper introduces two heuristics for efficient exploration and an improved confidence interval that enables earlier termination with probabilistic guarantees. We prove that the heuristics and the confidence interval are sound and produce with high probability an approximately optimal policy in polynomial time. Experiments on two benchmark problems and two instances of an invasive species management problem show that the improved confidence intervals and the new search heuristics yield reductions of between 8% and 47% in the number of simulator calls required to reach near-optimal policies.","PeriodicalId":14794,"journal":{"name":"J. Mach. Learn. Res.","volume":"31 1","pages":"3877-3903"},"PeriodicalIF":0.0,"publicationDate":"2015-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"79032748","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 22
Learning using privileged information: similarity control and knowledge transfer 利用特权信息学习:相似性控制和知识转移
Pub Date : 2015-01-01 DOI: 10.5555/2789272.2886814
V. Vapnik, R. Izmailov
This paper describes a new paradigm of machine learning, in which Intelligent Teacher is involved. During training stage, Intelligent Teacher provides Student with information that contains, along with classification of each example, additional privileged information (for example, explanation) of this example. The paper describes two mechanisms that can be used for significantly accelerating the speed of Student's learning using privileged information: (1) correction of Student's concepts of similarity between examples, and (2) direct Teacher-Student knowledge transfer.
本文描述了一种新的机器学习范式,其中涉及智能教师。在训练阶段,智能教师向学生提供信息,这些信息包含每个示例的分类,以及该示例的附加特权信息(例如,解释)。本文描述了两种可用于利用特权信息显著加快学生学习速度的机制:(1)纠正学生对示例之间相似性的概念,以及(2)直接师生知识转移。
{"title":"Learning using privileged information: similarity control and knowledge transfer","authors":"V. Vapnik, R. Izmailov","doi":"10.5555/2789272.2886814","DOIUrl":"https://doi.org/10.5555/2789272.2886814","url":null,"abstract":"This paper describes a new paradigm of machine learning, in which Intelligent Teacher is involved. During training stage, Intelligent Teacher provides Student with information that contains, along with classification of each example, additional privileged information (for example, explanation) of this example. The paper describes two mechanisms that can be used for significantly accelerating the speed of Student's learning using privileged information: (1) correction of Student's concepts of similarity between examples, and (2) direct Teacher-Student knowledge transfer.","PeriodicalId":14794,"journal":{"name":"J. Mach. Learn. Res.","volume":"7 1","pages":"2023-2049"},"PeriodicalIF":0.0,"publicationDate":"2015-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"76458118","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 344
Predicting a switching sequence of graph labelings 预测图标记的切换顺序
Pub Date : 2015-01-01 DOI: 10.5555/2789272.2886813
M. Herbster, Stephen Pasteris, M. Pontil
We study the problem of predicting online the labeling of a graph. We consider a novel setting for this problem in which, in addition to observing vertices and labels on the graph, we also observe a sequence of just vertices on a second graph. A latent labeling of the second graph selects one of K labelings to be active on the first graph. We propose a polynomial time algorithm for online prediction in this setting and derive a mistake bound for the algorithm. The bound is controlled by the geometric cut of the observed and latent labelings, as well as the resistance diameters of the graphs. When specialized to multitask prediction and online switching problems the bound gives new and sharper results under certain conditions.
我们研究了在线预测图的标注问题。我们考虑了这个问题的一种新设置,除了观察图上的顶点和标签外,我们还观察了另一个图上的一个顶点序列。第二个图的潜在标记从K个标记中选择一个在第一个图上活动。在这种情况下,我们提出了一种多项式时间的在线预测算法,并推导了该算法的错误界。边界由观察到的和潜在的标记的几何切割以及图的阻力直径控制。当专门用于多任务预测和在线切换问题时,该界在一定条件下给出了新的、更清晰的结果。
{"title":"Predicting a switching sequence of graph labelings","authors":"M. Herbster, Stephen Pasteris, M. Pontil","doi":"10.5555/2789272.2886813","DOIUrl":"https://doi.org/10.5555/2789272.2886813","url":null,"abstract":"We study the problem of predicting online the labeling of a graph. We consider a novel setting for this problem in which, in addition to observing vertices and labels on the graph, we also observe a sequence of just vertices on a second graph. A latent labeling of the second graph selects one of K labelings to be active on the first graph. We propose a polynomial time algorithm for online prediction in this setting and derive a mistake bound for the algorithm. The bound is controlled by the geometric cut of the observed and latent labelings, as well as the resistance diameters of the graphs. When specialized to multitask prediction and online switching problems the bound gives new and sharper results under certain conditions.","PeriodicalId":14794,"journal":{"name":"J. Mach. Learn. Res.","volume":"95 1","pages":"2003-2022"},"PeriodicalIF":0.0,"publicationDate":"2015-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"79547396","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 15
Absent data generating classifier for imbalanced class sizes 不平衡类大小的缺席数据生成分类器
Pub Date : 2015-01-01 DOI: 10.5555/2789272.2912085
Arash Pourhabib, B. Mallick, Yu Ding
We propose an algorithm for two-class classification problems when the training data are imbalanced. This means the number of training instances in one of the classes is so low that the conventional classification algorithms become ineffective in detecting the minority class. We present a modification of the kernel Fisher discriminant analysis such that the imbalanced nature of the problem is explicitly addressed in the new algorithm formulation. The new algorithm exploits the properties of the existing minority points to learn the effects of other minority data points, had they actually existed. The algorithm proceeds iteratively by employing the learned properties and conditional sampling in such a way that it generates sufficient artificial data points for the minority set, thus enhancing the detection probability of the minority class. Implementing the proposed method on a number of simulated and real data sets, we show that our proposed method performs competitively compared to a set of alternative state-of-the-art imbalanced classification algorithms.
针对训练数据不平衡时的两类分类问题,提出了一种算法。这意味着一个类的训练实例数量非常少,以至于传统的分类算法在检测少数类时变得无效。我们提出了一个核费雪判别分析的修改,使问题的不平衡性质在新的算法公式中得到明确的解决。新算法利用现有少数数据点的特性来学习其他少数数据点在实际存在时的效果。该算法利用学习到的属性和条件采样进行迭代,为少数派集生成足够的人工数据点,从而提高了少数派类的检测概率。在许多模拟和真实数据集上实现所提出的方法,我们表明,与一组替代的最先进的不平衡分类算法相比,我们所提出的方法具有竞争力。
{"title":"Absent data generating classifier for imbalanced class sizes","authors":"Arash Pourhabib, B. Mallick, Yu Ding","doi":"10.5555/2789272.2912085","DOIUrl":"https://doi.org/10.5555/2789272.2912085","url":null,"abstract":"We propose an algorithm for two-class classification problems when the training data are imbalanced. This means the number of training instances in one of the classes is so low that the conventional classification algorithms become ineffective in detecting the minority class. We present a modification of the kernel Fisher discriminant analysis such that the imbalanced nature of the problem is explicitly addressed in the new algorithm formulation. The new algorithm exploits the properties of the existing minority points to learn the effects of other minority data points, had they actually existed. The algorithm proceeds iteratively by employing the learned properties and conditional sampling in such a way that it generates sufficient artificial data points for the minority set, thus enhancing the detection probability of the minority class. Implementing the proposed method on a number of simulated and real data sets, we show that our proposed method performs competitively compared to a set of alternative state-of-the-art imbalanced classification algorithms.","PeriodicalId":14794,"journal":{"name":"J. Mach. Learn. Res.","volume":"100 1","pages":"2695-2724"},"PeriodicalIF":0.0,"publicationDate":"2015-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"85908770","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 24
pyGPs: a Python library for Gaussian process regression and classification pyGPs:用于高斯过程回归和分类的Python库
Pub Date : 2015-01-01 DOI: 10.5555/2789272.2912082
Marion Neumann, Shan Huang, D. Marthaler, K. Kersting
We introduce pyGPs, an object-oriented implementation of Gaussian processes (gps) for machine learning. The library provides a wide range of functionalities reaching from simple gp specification via mean and covariance and gp inference to more complex implementations of hyperparameter optimization, sparse approximations, and graph based learning. Using Python we focus on usability for both "users" and "researchers". Our main goal is to offer a user-friendly and flexible implementation of GPs for machine learning.
我们介绍pyGPs,一个用于机器学习的高斯过程(gps)的面向对象实现。该库提供了广泛的功能,从简单的gp规范(通过均值、协方差和gp推理)到更复杂的超参数优化、稀疏逼近和基于图的学习实现。使用Python时,我们关注的是“用户”和“研究人员”的可用性。我们的主要目标是为机器学习提供一个用户友好和灵活的GPs实现。
{"title":"pyGPs: a Python library for Gaussian process regression and classification","authors":"Marion Neumann, Shan Huang, D. Marthaler, K. Kersting","doi":"10.5555/2789272.2912082","DOIUrl":"https://doi.org/10.5555/2789272.2912082","url":null,"abstract":"We introduce pyGPs, an object-oriented implementation of Gaussian processes (gps) for machine learning. The library provides a wide range of functionalities reaching from simple gp specification via mean and covariance and gp inference to more complex implementations of hyperparameter optimization, sparse approximations, and graph based learning. Using Python we focus on usability for both \"users\" and \"researchers\". Our main goal is to offer a user-friendly and flexible implementation of GPs for machine learning.","PeriodicalId":14794,"journal":{"name":"J. Mach. Learn. Res.","volume":"6 1","pages":"2611-2616"},"PeriodicalIF":0.0,"publicationDate":"2015-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"74080103","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 24
SAMOA: scalable advanced massive online analysis 萨摩亚:可扩展的高级大规模在线分析
Pub Date : 2015-01-01 DOI: 10.5555/2789272.2789277
G. D. F. Morales, A. Bifet
SAMOA (SCALABLE ADVANCED MASSIVE ONLINE ANALYSIS) is a platform for mining big data streams. It provides a collection of distributed streaming algorithms for the most common data mining and machine learning tasks such as classification, clustering, and regression, as well as programming abstractions to develop new algorithms. It features a pluggable architecture that allows it to run on several distributed stream processing engines such as Storm, S4, and Samza. samoa is written in Java, is open source, and is available at http://samoa-project.net under the Apache Software License version 2.0.
SAMOA (SCALABLE ADVANCED MASSIVE ONLINE ANALYSIS)是一个挖掘大数据流的平台。它为最常见的数据挖掘和机器学习任务(如分类、聚类和回归)提供了一组分布式流算法,并为开发新算法提供了编程抽象。它的特点是一个可插拔的架构,允许它在多个分布式流处理引擎上运行,如Storm、S4和Samza。samoa是用Java编写的,是开源的,可以在http://samoa-project.net上获得,使用Apache软件许可证2.0版本。
{"title":"SAMOA: scalable advanced massive online analysis","authors":"G. D. F. Morales, A. Bifet","doi":"10.5555/2789272.2789277","DOIUrl":"https://doi.org/10.5555/2789272.2789277","url":null,"abstract":"SAMOA (SCALABLE ADVANCED MASSIVE ONLINE ANALYSIS) is a platform for mining big data streams. It provides a collection of distributed streaming algorithms for the most common data mining and machine learning tasks such as classification, clustering, and regression, as well as programming abstractions to develop new algorithms. It features a pluggable architecture that allows it to run on several distributed stream processing engines such as Storm, S4, and Samza. samoa is written in Java, is open source, and is available at http://samoa-project.net under the Apache Software License version 2.0.","PeriodicalId":14794,"journal":{"name":"J. Mach. Learn. Res.","volume":"28 1","pages":"149-153"},"PeriodicalIF":0.0,"publicationDate":"2015-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"74337170","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 177
期刊
J. Mach. Learn. Res.
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1