首页 > 最新文献

Proceedings of the 23rd international conference on Machine learning最新文献

英文 中文
Nonstationary kernel combination 非平稳核组合
Pub Date : 2006-06-25 DOI: 10.1145/1143844.1143914
Darrin P. Lewis, T. Jebara, William Stafford Noble
The power and popularity of kernel methods stem in part from their ability to handle diverse forms of structured inputs, including vectors, graphs and strings. Recently, several methods have been proposed for combining kernels from heterogeneous data sources. However, all of these methods produce stationary combinations; i.e., the relative weights of the various kernels do not vary among input examples. This article proposes a method for combining multiple kernels in a nonstationary fashion. The approach uses a large-margin latent-variable generative model within the maximum entropy discrimination (MED) framework. Latent parameter estimation is rendered tractable by variational bounds and an iterative optimization procedure. The classifier we use is a log-ratio of Gaussian mixtures, in which each component is implicitly mapped via a Mercer kernel function. We show that the support vector machine is a special case of this model. In this approach, discriminative parameter estimation is feasible via a fast sequential minimal optimization algorithm. Empirical results are presented on synthetic data, several benchmarks, and on a protein function annotation task.
核方法的强大和流行部分源于它们处理各种形式的结构化输入的能力,包括向量、图和字符串。最近,人们提出了几种方法来组合来自异构数据源的核。然而,所有这些方法产生平稳组合;也就是说,各种核的相对权重在不同的输入示例中不会变化。本文提出了一种以非平稳方式组合多个核的方法。该方法在最大熵判别(MED)框架内使用大边际潜变量生成模型。隐参数估计通过变分边界和迭代优化过程变得易于处理。我们使用的分类器是高斯混合物的对数比,其中每个成分都通过默瑟核函数隐式映射。我们证明了支持向量机是该模型的一个特例。在该方法中,通过快速的顺序最小优化算法,判别参数估计是可行的。实证结果提出了合成数据,几个基准,并在蛋白质功能注释任务。
{"title":"Nonstationary kernel combination","authors":"Darrin P. Lewis, T. Jebara, William Stafford Noble","doi":"10.1145/1143844.1143914","DOIUrl":"https://doi.org/10.1145/1143844.1143914","url":null,"abstract":"The power and popularity of kernel methods stem in part from their ability to handle diverse forms of structured inputs, including vectors, graphs and strings. Recently, several methods have been proposed for combining kernels from heterogeneous data sources. However, all of these methods produce stationary combinations; i.e., the relative weights of the various kernels do not vary among input examples. This article proposes a method for combining multiple kernels in a nonstationary fashion. The approach uses a large-margin latent-variable generative model within the maximum entropy discrimination (MED) framework. Latent parameter estimation is rendered tractable by variational bounds and an iterative optimization procedure. The classifier we use is a log-ratio of Gaussian mixtures, in which each component is implicitly mapped via a Mercer kernel function. We show that the support vector machine is a special case of this model. In this approach, discriminative parameter estimation is feasible via a fast sequential minimal optimization algorithm. Empirical results are presented on synthetic data, several benchmarks, and on a protein function annotation task.","PeriodicalId":124011,"journal":{"name":"Proceedings of the 23rd international conference on Machine learning","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134365092","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 91
Statistical debugging: simultaneous identification of multiple bugs 统计调试:同时识别多个bug
Pub Date : 2006-06-25 DOI: 10.1145/1143844.1143983
A. Zheng, Michael I. Jordan, B. Liblit, M. Naik, A. Aiken
We describe a statistical approach to software debugging in the presence of multiple bugs. Due to sparse sampling issues and complex interaction between program predicates, many generic off-the-shelf algorithms fail to select useful bug predictors. Taking inspiration from bi-clustering algorithms, we propose an iterative collective voting scheme for the program runs and predicates. We demonstrate successful debugging results on several real world programs and a large debugging benchmark suite.
我们描述了一种在存在多个错误的情况下进行软件调试的统计方法。由于稀疏的采样问题和程序谓词之间复杂的交互,许多通用的现成算法无法选择有用的错误预测器。受双聚类算法的启发,我们提出了一个迭代的程序运行和谓词的集体投票方案。我们在几个真实世界的程序和一个大型调试基准套件上演示了成功的调试结果。
{"title":"Statistical debugging: simultaneous identification of multiple bugs","authors":"A. Zheng, Michael I. Jordan, B. Liblit, M. Naik, A. Aiken","doi":"10.1145/1143844.1143983","DOIUrl":"https://doi.org/10.1145/1143844.1143983","url":null,"abstract":"We describe a statistical approach to software debugging in the presence of multiple bugs. Due to sparse sampling issues and complex interaction between program predicates, many generic off-the-shelf algorithms fail to select useful bug predictors. Taking inspiration from bi-clustering algorithms, we propose an iterative collective voting scheme for the program runs and predicates. We demonstrate successful debugging results on several real world programs and a large debugging benchmark suite.","PeriodicalId":124011,"journal":{"name":"Proceedings of the 23rd international conference on Machine learning","volume":"31 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128344488","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 168
Feature subset selection bias for classification learning 分类学习的特征子集选择偏差
Pub Date : 2006-06-25 DOI: 10.1145/1143844.1143951
Surendra K. Singhi, Huan Liu
Feature selection is often applied to high-dimensional data prior to classification learning. Using the same training dataset in both selection and learning can result in so-called feature subset selection bias. This bias putatively can exacerbate data over-fitting and negatively affect classification performance. However, in current practice separate datasets are seldom employed for selection and learning, because dividing the training data into two datasets for feature selection and classifier learning respectively reduces the amount of data that can be used in either task. This work attempts to address this dilemma. We formalize selection bias for classification learning, analyze its statistical properties, and study factors that affect selection bias, as well as how the bias impacts classification learning via various experiments. This research endeavors to provide illustration and explanation why the bias may not cause negative impact in classification as much as expected in regression.
特征选择通常在分类学习之前应用于高维数据。在选择和学习中使用相同的训练数据集会导致所谓的特征子集选择偏差。这种偏差可能会加剧数据的过度拟合,并对分类性能产生负面影响。然而,在目前的实践中,很少使用单独的数据集进行选择和学习,因为将训练数据分成两个数据集分别进行特征选择和分类器学习会减少两项任务中可用的数据量。这项工作试图解决这一困境。我们形式化了分类学习的选择偏差,分析了选择偏差的统计特性,并通过各种实验研究了影响选择偏差的因素,以及选择偏差如何影响分类学习。本研究试图提供说明和解释为什么偏差可能不会像回归中预期的那样对分类产生负面影响。
{"title":"Feature subset selection bias for classification learning","authors":"Surendra K. Singhi, Huan Liu","doi":"10.1145/1143844.1143951","DOIUrl":"https://doi.org/10.1145/1143844.1143951","url":null,"abstract":"Feature selection is often applied to high-dimensional data prior to classification learning. Using the same training dataset in both selection and learning can result in so-called feature subset selection bias. This bias putatively can exacerbate data over-fitting and negatively affect classification performance. However, in current practice separate datasets are seldom employed for selection and learning, because dividing the training data into two datasets for feature selection and classifier learning respectively reduces the amount of data that can be used in either task. This work attempts to address this dilemma. We formalize selection bias for classification learning, analyze its statistical properties, and study factors that affect selection bias, as well as how the bias impacts classification learning via various experiments. This research endeavors to provide illustration and explanation why the bias may not cause negative impact in classification as much as expected in regression.","PeriodicalId":124011,"journal":{"name":"Proceedings of the 23rd international conference on Machine learning","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130285109","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 98
Two-dimensional solution path for support vector regression 支持向量回归的二维解路径
Pub Date : 2006-06-25 DOI: 10.1145/1143844.1143969
G. Wang, D. Yeung, F. Lochovsky
Recently, a very appealing approach was proposed to compute the entire solution path for support vector classification (SVC) with very low extra computational cost. This approach was later extended to a support vector regression (SVR) model called ε-SVR. However, the method requires that the error parameter ε be set a priori, which is only possible if the desired accuracy of the approximation can be specified in advance. In this paper, we show that the solution path for ε-SVR is also piecewise linear with respect to ε. We further propose an efficient algorithm for exploring the two-dimensional solution space defined by the regularization and error parameters. As opposed to the algorithm for SVC, our proposed algorithm for ε-SVR initializes the number of support vectors to zero and then increases it gradually as the algorithm proceeds. As such, a good regression function possessing the sparseness property can be obtained after only a few iterations.
最近,人们提出了一种计算支持向量分类(SVC)整个解路径的方法,其额外计算成本非常低。该方法后来被扩展为支持向量回归(SVR)模型,称为ε-SVR。然而,该方法需要先验地设置误差参数ε,这只有在可以提前指定所需的近似精度时才有可能。在本文中,我们证明了ε- svr的解路径对于ε也是分段线性的。我们进一步提出了一种有效的算法来探索由正则化和误差参数定义的二维解空间。与SVC算法相反,我们提出的ε-SVR算法将支持向量的数量初始化为零,然后随着算法的进行逐渐增加。因此,只需几次迭代就可以得到具有稀疏性的良好回归函数。
{"title":"Two-dimensional solution path for support vector regression","authors":"G. Wang, D. Yeung, F. Lochovsky","doi":"10.1145/1143844.1143969","DOIUrl":"https://doi.org/10.1145/1143844.1143969","url":null,"abstract":"Recently, a very appealing approach was proposed to compute the entire solution path for support vector classification (SVC) with very low extra computational cost. This approach was later extended to a support vector regression (SVR) model called ε-SVR. However, the method requires that the error parameter ε be set a priori, which is only possible if the desired accuracy of the approximation can be specified in advance. In this paper, we show that the solution path for ε-SVR is also piecewise linear with respect to ε. We further propose an efficient algorithm for exploring the two-dimensional solution space defined by the regularization and error parameters. As opposed to the algorithm for SVC, our proposed algorithm for ε-SVR initializes the number of support vectors to zero and then increases it gradually as the algorithm proceeds. As such, a good regression function possessing the sparseness property can be obtained after only a few iterations.","PeriodicalId":124011,"journal":{"name":"Proceedings of the 23rd international conference on Machine learning","volume":"87 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131450524","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 36
Autonomous shaping: knowledge transfer in reinforcement learning 自主塑造:强化学习中的知识转移
Pub Date : 2006-06-25 DOI: 10.1145/1143844.1143906
G. Konidaris, A. Barto
We introduce the use of learned shaping rewards in reinforcement learning tasks, where an agent uses prior experience on a sequence of tasks to learn a portable predictor that estimates intermediate rewards, resulting in accelerated learning in later tasks that are related but distinct. Such agents can be trained on a sequence of relatively easy tasks in order to develop a more informative measure of reward that can be transferred to improve performance on more difficult tasks without requiring a hand coded shaping function. We use a rod positioning task to show that this significantly improves performance even after a very brief training period.
我们在强化学习任务中引入了学习成型奖励的使用,其中智能体使用一系列任务的先验经验来学习估计中间奖励的便携式预测器,从而加速了后期相关但不同的任务的学习。这样的智能体可以在一系列相对简单的任务上进行训练,以便开发一种更有信息的奖励措施,这种奖励措施可以转移到更困难的任务上,而不需要手工编码的塑造函数。我们使用一个杆定位任务来证明,即使在很短的训练时间后,这也能显著提高表现。
{"title":"Autonomous shaping: knowledge transfer in reinforcement learning","authors":"G. Konidaris, A. Barto","doi":"10.1145/1143844.1143906","DOIUrl":"https://doi.org/10.1145/1143844.1143906","url":null,"abstract":"We introduce the use of learned shaping rewards in reinforcement learning tasks, where an agent uses prior experience on a sequence of tasks to learn a portable predictor that estimates intermediate rewards, resulting in accelerated learning in later tasks that are related but distinct. Such agents can be trained on a sequence of relatively easy tasks in order to develop a more informative measure of reward that can be transferred to improve performance on more difficult tasks without requiring a hand coded shaping function. We use a rod positioning task to show that this significantly improves performance even after a very brief training period.","PeriodicalId":124011,"journal":{"name":"Proceedings of the 23rd international conference on Machine learning","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129912713","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 225
Kernel information embeddings 核信息嵌入
Pub Date : 2006-06-25 DOI: 10.1145/1143844.1143924
R. Memisevic
We describe a family of embedding algorithms that are based on nonparametric estimates of mutual information (MI). Using Parzen window estimates of the distribution in the joint (input, embedding)-space, we derive a MI-based objective function for dimensionality reduction that can be optimized directly with respect to a set of latent data representatives. Various types of supervision signal can be introduced within the framework by replacing plain MI with several forms of conditional MI. Examples of the semi-(un)supervised algorithms that we obtain this way are a new model for manifold alignment, and a new type of embedding method that performs 'conditional dimensionality reduction'.
我们描述了一系列基于互信息(MI)的非参数估计的嵌入算法。使用联合(输入,嵌入)空间中分布的Parzen窗口估计,我们导出了一个基于mi的降维目标函数,该目标函数可以直接针对一组潜在数据代表进行优化。通过用几种形式的条件MI替换普通MI,可以在框架内引入各种类型的监督信号。我们以这种方式获得的半(非)监督算法的示例是一种新的流形排列模型,以及一种执行“条件降维”的新型嵌入方法。
{"title":"Kernel information embeddings","authors":"R. Memisevic","doi":"10.1145/1143844.1143924","DOIUrl":"https://doi.org/10.1145/1143844.1143924","url":null,"abstract":"We describe a family of embedding algorithms that are based on nonparametric estimates of mutual information (MI). Using Parzen window estimates of the distribution in the joint (input, embedding)-space, we derive a MI-based objective function for dimensionality reduction that can be optimized directly with respect to a set of latent data representatives. Various types of supervision signal can be introduced within the framework by replacing plain MI with several forms of conditional MI. Examples of the semi-(un)supervised algorithms that we obtain this way are a new model for manifold alignment, and a new type of embedding method that performs 'conditional dimensionality reduction'.","PeriodicalId":124011,"journal":{"name":"Proceedings of the 23rd international conference on Machine learning","volume":"90 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123458007","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 18
Optimal kernel selection in Kernel Fisher discriminant analysis 核Fisher判别分析中的最优核选择
Pub Date : 2006-06-25 DOI: 10.1145/1143844.1143903
Seung-Jean Kim, A. Magnani, Stephen P. Boyd
In Kernel Fisher discriminant analysis (KFDA), we carry out Fisher linear discriminant analysis in a high dimensional feature space defined implicitly by a kernel. The performance of KFDA depends on the choice of the kernel; in this paper, we consider the problem of finding the optimal kernel, over a given convex set of kernels. We show that this optimal kernel selection problem can be reformulated as a tractable convex optimization problem which interior-point methods can solve globally and efficiently. The kernel selection method is demonstrated with some UCI machine learning benchmark examples.
在Kernel Fisher判别分析(KFDA)中,我们在由Kernel隐式定义的高维特征空间中进行Fisher线性判别分析。KFDA的性能取决于内核的选择;在本文中,我们考虑在给定核的凸集上寻找最优核的问题。我们证明了这种最优核选择问题可以重新表述为一个可处理的凸优化问题,内点方法可以全局有效地求解。通过一些UCI机器学习的基准示例对核选择方法进行了验证。
{"title":"Optimal kernel selection in Kernel Fisher discriminant analysis","authors":"Seung-Jean Kim, A. Magnani, Stephen P. Boyd","doi":"10.1145/1143844.1143903","DOIUrl":"https://doi.org/10.1145/1143844.1143903","url":null,"abstract":"In Kernel Fisher discriminant analysis (KFDA), we carry out Fisher linear discriminant analysis in a high dimensional feature space defined implicitly by a kernel. The performance of KFDA depends on the choice of the kernel; in this paper, we consider the problem of finding the optimal kernel, over a given convex set of kernels. We show that this optimal kernel selection problem can be reformulated as a tractable convex optimization problem which interior-point methods can solve globally and efficiently. The kernel selection method is demonstrated with some UCI machine learning benchmark examples.","PeriodicalId":124011,"journal":{"name":"Proceedings of the 23rd international conference on Machine learning","volume":"71 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128244721","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 165
Efficient learning of Naive Bayes classifiers under class-conditional classification noise 类条件分类噪声下朴素贝叶斯分类器的高效学习
Pub Date : 2006-06-25 DOI: 10.1145/1143844.1143878
François Denis, C. Magnan, L. Ralaivola
We address the problem of efficiently learning Naive Bayes classifiers under class-conditional classification noise (CCCN). Naive Bayes classifiers rely on the hypothesis that the distributions associated to each class are product distributions. When data is subject to CCC-noise, these conditional distributions are themselves mixtures of product distributions. We give analytical formulas which makes it possible to identify them from data subject to CCCN. Then, we design a learning algorithm based on these formulas able to learn Naive Bayes classifiers under CCCN. We present results on artificial datasets and datasets extracted from the UCI repository database. These results show that CCCN can be efficiently and successfully handled.
研究了在类别条件分类噪声(CCCN)下朴素贝叶斯分类器的高效学习问题。朴素贝叶斯分类器依赖于与每个类相关的分布是乘积分布的假设。当数据受到cc噪声的影响时,这些条件分布本身就是乘积分布的混合物。我们给出了分析公式,使它们能够从CCCN的数据中识别出来。然后,我们基于这些公式设计了一种能够在CCCN下学习朴素贝叶斯分类器的学习算法。我们给出了人工数据集和从UCI存储库数据库中提取的数据集的结果。这些结果表明,CCCN可以有效和成功地处理。
{"title":"Efficient learning of Naive Bayes classifiers under class-conditional classification noise","authors":"François Denis, C. Magnan, L. Ralaivola","doi":"10.1145/1143844.1143878","DOIUrl":"https://doi.org/10.1145/1143844.1143878","url":null,"abstract":"We address the problem of efficiently learning Naive Bayes classifiers under class-conditional classification noise (CCCN). Naive Bayes classifiers rely on the hypothesis that the distributions associated to each class are product distributions. When data is subject to CCC-noise, these conditional distributions are themselves mixtures of product distributions. We give analytical formulas which makes it possible to identify them from data subject to CCCN. Then, we design a learning algorithm based on these formulas able to learn Naive Bayes classifiers under CCCN. We present results on artificial datasets and datasets extracted from the UCI repository database. These results show that CCCN can be efficiently and successfully handled.","PeriodicalId":124011,"journal":{"name":"Proceedings of the 23rd international conference on Machine learning","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128984295","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Nightmare at test time: robust learning by feature deletion 测试时的噩梦:通过特征删除进行稳健学习
Pub Date : 2006-06-25 DOI: 10.1145/1143844.1143889
A. Globerson, S. Roweis
When constructing a classifier from labeled data, it is important not to assign too much weight to any single input feature, in order to increase the robustness of the classifier. This is particularly important in domains with nonstationary feature distributions or with input sensor failures. A common approach to achieving such robustness is to introduce regularization which spreads the weight more evenly between the features. However, this strategy is very generic, and cannot induce robustness specifically tailored to the classification task at hand. In this work, we introduce a new algorithm for avoiding single feature over-weighting by analyzing robustness using a game theoretic formalization. We develop classifiers which are optimally resilient to deletion of features in a minimax sense, and show how to construct such classifiers using quadratic programming. We illustrate the applicability of our methods on spam filtering and handwritten digit recognition tasks, where feature deletion is indeed a realistic noise model.
当从标记数据构建分类器时,重要的是不要给任何单个输入特征分配过多的权重,以增加分类器的鲁棒性。这在具有非平稳特征分布或输入传感器失效的领域中尤为重要。实现这种鲁棒性的常见方法是引入正则化,在特征之间更均匀地分配权重。然而,这种策略是非常通用的,不能产生专门针对手头的分类任务的鲁棒性。在这项工作中,我们引入了一种新的算法,通过使用博弈论形式化分析鲁棒性来避免单个特征的超重。我们开发了在极小极大意义上对特征删除具有最佳弹性的分类器,并展示了如何使用二次规划构建这样的分类器。我们说明了我们的方法在垃圾邮件过滤和手写数字识别任务中的适用性,其中特征删除确实是一个现实的噪声模型。
{"title":"Nightmare at test time: robust learning by feature deletion","authors":"A. Globerson, S. Roweis","doi":"10.1145/1143844.1143889","DOIUrl":"https://doi.org/10.1145/1143844.1143889","url":null,"abstract":"When constructing a classifier from labeled data, it is important not to assign too much weight to any single input feature, in order to increase the robustness of the classifier. This is particularly important in domains with nonstationary feature distributions or with input sensor failures. A common approach to achieving such robustness is to introduce regularization which spreads the weight more evenly between the features. However, this strategy is very generic, and cannot induce robustness specifically tailored to the classification task at hand. In this work, we introduce a new algorithm for avoiding single feature over-weighting by analyzing robustness using a game theoretic formalization. We develop classifiers which are optimally resilient to deletion of features in a minimax sense, and show how to construct such classifiers using quadratic programming. We illustrate the applicability of our methods on spam filtering and handwritten digit recognition tasks, where feature deletion is indeed a realistic noise model.","PeriodicalId":124011,"journal":{"name":"Proceedings of the 23rd international conference on Machine learning","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128556446","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 358
Using inaccurate models in reinforcement learning 在强化学习中使用不准确的模型
Pub Date : 2006-06-25 DOI: 10.1145/1143844.1143845
P. Abbeel, M. Quigley, A. Ng
In the model-based policy search approach to reinforcement learning (RL), policies are found using a model (or "simulator") of the Markov decision process. However, for high-dimensional continuous-state tasks, it can be extremely difficult to build an accurate model, and thus often the algorithm returns a policy that works in simulation but not in real-life. The other extreme, model-free RL, tends to require infeasibly large numbers of real-life trials. In this paper, we present a hybrid algorithm that requires only an approximate model, and only a small number of real-life trials. The key idea is to successively "ground" the policy evaluations using real-life trials, but to rely on the approximate model to suggest local changes. Our theoretical results show that this algorithm achieves near-optimal performance in the real system, even when the model is only approximate. Empirical results also demonstrate that---when given only a crude model and a small number of real-life trials---our algorithm can obtain near-optimal performance in the real system.
在基于模型的策略搜索强化学习(RL)方法中,策略是使用马尔可夫决策过程的模型(或“模拟器”)找到的。然而,对于高维连续状态任务,构建精确的模型可能非常困难,因此算法通常返回的策略在模拟中有效,但在现实生活中无效。另一个极端,无模型强化学习,往往需要不可行的大量现实生活试验。在本文中,我们提出了一种混合算法,它只需要一个近似模型,只需要少量的实际试验。其关键思想是,利用现实生活中的试验,逐步“奠定”政策评估的基础,但依靠近似模型来建议局部变化。我们的理论结果表明,该算法在实际系统中达到了接近最优的性能,即使模型只是近似的。经验结果还表明,当只给出一个粗略的模型和少量的实际试验时,我们的算法可以在实际系统中获得接近最优的性能。
{"title":"Using inaccurate models in reinforcement learning","authors":"P. Abbeel, M. Quigley, A. Ng","doi":"10.1145/1143844.1143845","DOIUrl":"https://doi.org/10.1145/1143844.1143845","url":null,"abstract":"In the model-based policy search approach to reinforcement learning (RL), policies are found using a model (or \"simulator\") of the Markov decision process. However, for high-dimensional continuous-state tasks, it can be extremely difficult to build an accurate model, and thus often the algorithm returns a policy that works in simulation but not in real-life. The other extreme, model-free RL, tends to require infeasibly large numbers of real-life trials. In this paper, we present a hybrid algorithm that requires only an approximate model, and only a small number of real-life trials. The key idea is to successively \"ground\" the policy evaluations using real-life trials, but to rely on the approximate model to suggest local changes. Our theoretical results show that this algorithm achieves near-optimal performance in the real system, even when the model is only approximate. Empirical results also demonstrate that---when given only a crude model and a small number of real-life trials---our algorithm can obtain near-optimal performance in the real system.","PeriodicalId":124011,"journal":{"name":"Proceedings of the 23rd international conference on Machine learning","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129335210","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 251
期刊
Proceedings of the 23rd international conference on Machine learning
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1