A. Guerrero-Curieses, R. Alaíz-Rodríguez, Jesús Cid-Sueiro
{"title":"Loss functions to combine learning and decision in multiclass problems","authors":"A. Guerrero-Curieses, R. Alaíz-Rodríguez, Jesús Cid-Sueiro","doi":"10.1109/NNSP.2003.1318031","DOIUrl":null,"url":null,"abstract":"The design of structures and algorithms for non-MAP multiclass decision problems is discussed in this paper. We propose a parametric family of loss functions that provide the most accurate estimates for the posterior class probabilities near the decision regions. Moreover, we discuss learning algorithms based on the stochastic gradient minimization of these loss functions. We show that these algorithms behave like sample selectors: samples near the decision regions are the most relevant during learning. Experimental results on some real datasets are also provided to show the effectiveness of this approach versus the classical cross entropy (based on a global posterior probability estimation).","PeriodicalId":315958,"journal":{"name":"2003 IEEE XIII Workshop on Neural Networks for Signal Processing (IEEE Cat. No.03TH8718)","volume":"570 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2003-09-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"4","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2003 IEEE XIII Workshop on Neural Networks for Signal Processing (IEEE Cat. No.03TH8718)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/NNSP.2003.1318031","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 4
Abstract
The design of structures and algorithms for non-MAP multiclass decision problems is discussed in this paper. We propose a parametric family of loss functions that provide the most accurate estimates for the posterior class probabilities near the decision regions. Moreover, we discuss learning algorithms based on the stochastic gradient minimization of these loss functions. We show that these algorithms behave like sample selectors: samples near the decision regions are the most relevant during learning. Experimental results on some real datasets are also provided to show the effectiveness of this approach versus the classical cross entropy (based on a global posterior probability estimation).