{"title":"Revisiting the ROC curve for diagnostic applications with an unbalanced class distribution","authors":"C. O'Reilly, T. Nielsen","doi":"10.1109/WOSSPA.2013.6602401","DOIUrl":null,"url":null,"abstract":"This communication investigates the impact on classifier evaluation of a high asymmetry between positive and negatives classes. It points out some necessary precautions when reporting classifier performances using threshold-dependent statistics defined with the confusion matrix. It stresses that, in highly unbalanced datasets, reporting the positive predictive value (PPV) is more appropriate than reporting specificity. More elaborate variables such as F-measure and Matthews' correlation coefficient may also provide a reliable portrait. It further concludes that, in many cases, only a small portion of the receiver operating characteristic (ROC) curve is actually useful, unless very low PPV are judged acceptable. Two remedies are proposed to complement the ROC curve: using a positive tradeoff curve (defined herein) or adding iso-PPV lines (i.e., lines of constant PPV) on the ROC graph. The observations reported in this study contribute to understanding of the impact of asymmetry on classifier performances. They also cast some doubt on the pertinence, when dealing with highly asymmetric problems, of using the area under the ROC curve for threshold-independent assessment of classifier performances.","PeriodicalId":417940,"journal":{"name":"2013 8th International Workshop on Systems, Signal Processing and their Applications (WoSSPA)","volume":"71 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2013-05-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"13","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2013 8th International Workshop on Systems, Signal Processing and their Applications (WoSSPA)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/WOSSPA.2013.6602401","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 13
Abstract
This communication investigates the impact on classifier evaluation of a high asymmetry between positive and negatives classes. It points out some necessary precautions when reporting classifier performances using threshold-dependent statistics defined with the confusion matrix. It stresses that, in highly unbalanced datasets, reporting the positive predictive value (PPV) is more appropriate than reporting specificity. More elaborate variables such as F-measure and Matthews' correlation coefficient may also provide a reliable portrait. It further concludes that, in many cases, only a small portion of the receiver operating characteristic (ROC) curve is actually useful, unless very low PPV are judged acceptable. Two remedies are proposed to complement the ROC curve: using a positive tradeoff curve (defined herein) or adding iso-PPV lines (i.e., lines of constant PPV) on the ROC graph. The observations reported in this study contribute to understanding of the impact of asymmetry on classifier performances. They also cast some doubt on the pertinence, when dealing with highly asymmetric problems, of using the area under the ROC curve for threshold-independent assessment of classifier performances.