From Convex to Nonconvex: A Loss Function Analysis for Binary Classification

Lei Zhao, M. Mammadov, J. Yearwood
{"title":"From Convex to Nonconvex: A Loss Function Analysis for Binary Classification","authors":"Lei Zhao, M. Mammadov, J. Yearwood","doi":"10.1109/ICDMW.2010.57","DOIUrl":null,"url":null,"abstract":"Problems of data classification can be studied in the framework of regularization theory as ill-posed problems. In this framework, loss functions play an important role in the application of regularization theory to classification. In this paper, we review some important convex loss functions, including hinge loss, square loss, modified square loss, exponential loss, logistic regression loss, as well as some non-convex loss functions, such as sigmoid loss, $\\phi$-loss, ramp loss, normalized sigmoid loss, and the loss function of 2 layer neural network. Based on the analysis of these loss functions, we propose a new differentiable non-convex loss function, called smoothed 0-1 loss function, which is a natural approximation of the 0-1 loss function. To compare the performance of different loss functions, we propose two binary classification algorithms for binary classification, one for convex loss functions, the other for non-convex loss functions. A set of experiments are launched on several binary data sets from the UCI repository. The results show that the proposed smoothed 0-1 loss function is robust, especially for those noisy data sets with many outliers.","PeriodicalId":170201,"journal":{"name":"2010 IEEE International Conference on Data Mining Workshops","volume":"21 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2010-12-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"42","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2010 IEEE International Conference on Data Mining Workshops","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICDMW.2010.57","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 42

Abstract

Problems of data classification can be studied in the framework of regularization theory as ill-posed problems. In this framework, loss functions play an important role in the application of regularization theory to classification. In this paper, we review some important convex loss functions, including hinge loss, square loss, modified square loss, exponential loss, logistic regression loss, as well as some non-convex loss functions, such as sigmoid loss, $\phi$-loss, ramp loss, normalized sigmoid loss, and the loss function of 2 layer neural network. Based on the analysis of these loss functions, we propose a new differentiable non-convex loss function, called smoothed 0-1 loss function, which is a natural approximation of the 0-1 loss function. To compare the performance of different loss functions, we propose two binary classification algorithms for binary classification, one for convex loss functions, the other for non-convex loss functions. A set of experiments are launched on several binary data sets from the UCI repository. The results show that the proposed smoothed 0-1 loss function is robust, especially for those noisy data sets with many outliers.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
从凸到非凸:二值分类的损失函数分析
数据分类问题可以在正则化理论的框架下作为不适定问题来研究。在这个框架中,损失函数在正则化理论在分类中的应用中起着重要的作用。本文综述了一些重要的凸损失函数,包括铰链损失、平方损失、修正平方损失、指数损失、逻辑回归损失,以及一些非凸损失函数,如sigmoid损失、$\phi$-损失、斜坡损失、归一化sigmoid损失和2层神经网络的损失函数。在分析这些损失函数的基础上,我们提出了一种新的可微非凸损失函数,称为光滑的0-1损失函数,它是0-1损失函数的自然逼近。为了比较不同损失函数的性能,我们提出了两种二元分类算法,一种用于凸损失函数,另一种用于非凸损失函数。在UCI存储库的几个二进制数据集上启动了一组实验。结果表明,所提出的平滑0-1损失函数具有较强的鲁棒性,尤其适用于含有大量异常值的噪声数据集。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Quantum Path Integral Inspired Query Sequence Suggestion for User Search Task Simplification PTCR-Miner: Progressive Temporal Class Rule Mining for Multivariate Temporal Data Classification Bridging Folksonomies and Domain Ontologies: Getting Out Non-taxonomic Relations SIMPLE: Interactive Analytics on Patent Data Parallel EM-Clustering: Fast Convergence by Asynchronous Model Updates
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1