The study of the hyper-parameter modelling the decision rule of the cautious classifiers based on the Fβ measure

IF 2.3 Q2 COMPUTER SCIENCE, THEORY & METHODS Array Pub Date : 2023-09-01 DOI:10.1016/j.array.2023.100310
Abdelhak Imoussaten
{"title":"The study of the hyper-parameter modelling the decision rule of the cautious classifiers based on the Fβ measure","authors":"Abdelhak Imoussaten","doi":"10.1016/j.array.2023.100310","DOIUrl":null,"url":null,"abstract":"<div><p>In some sensitive domains where data imperfections are present, standard classification techniques reach their limits. To avoid misclassifications that have serious consequences, recent works propose cautious classification algorithms to handle this problem. Despite of the presence of uncertainty and/or imprecision, a point prediction classifier is forced to bet on a single class. While a cautious classifier proposes the appropriate subset of candidate classes that can be assigned to the sample in the presence of imperfect information. On the other hand, cautiousness should not be at the expense of precision and a trade-off has to be made between these two criteria. Among the existing cautious classifiers, two classifiers propose to manage this trade-off in the decision step by the mean of a parametrized objective function. The first one is the non-deterministic classifier (ndc) proposed within the framework of probability theory and the second one is “evidential classifier based on imprecise relabelling” (eclair) proposed within the framework of belief functions. The theoretical aim of the mentioned hyper-parameters is to control the size of predictions for both classifiers. This paper proposes to study this hyper-parameter in order to select the “best” value in a classification task. First the utility for each candidate subset is studied related to the values of the hyper-parameter and some thresholds are proposed to control the size of the predictions. Then two illustrations are proposed where a method to choose this hyper-parameters based on the calibration data is proposed. The first illustration concerns randomly generated data and the second one concerns the images data of fashion mnist. These illustrations show how to control the size of the predictions and give a comparison between the performances of the two classifiers for a tuning based on our proposition and the one based on grid search method.</p></div>","PeriodicalId":8417,"journal":{"name":"Array","volume":null,"pages":null},"PeriodicalIF":2.3000,"publicationDate":"2023-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Array","FirstCategoryId":"1085","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2590005623000358","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"COMPUTER SCIENCE, THEORY & METHODS","Score":null,"Total":0}
引用次数: 0

Abstract

In some sensitive domains where data imperfections are present, standard classification techniques reach their limits. To avoid misclassifications that have serious consequences, recent works propose cautious classification algorithms to handle this problem. Despite of the presence of uncertainty and/or imprecision, a point prediction classifier is forced to bet on a single class. While a cautious classifier proposes the appropriate subset of candidate classes that can be assigned to the sample in the presence of imperfect information. On the other hand, cautiousness should not be at the expense of precision and a trade-off has to be made between these two criteria. Among the existing cautious classifiers, two classifiers propose to manage this trade-off in the decision step by the mean of a parametrized objective function. The first one is the non-deterministic classifier (ndc) proposed within the framework of probability theory and the second one is “evidential classifier based on imprecise relabelling” (eclair) proposed within the framework of belief functions. The theoretical aim of the mentioned hyper-parameters is to control the size of predictions for both classifiers. This paper proposes to study this hyper-parameter in order to select the “best” value in a classification task. First the utility for each candidate subset is studied related to the values of the hyper-parameter and some thresholds are proposed to control the size of the predictions. Then two illustrations are proposed where a method to choose this hyper-parameters based on the calibration data is proposed. The first illustration concerns randomly generated data and the second one concerns the images data of fashion mnist. These illustrations show how to control the size of the predictions and give a comparison between the performances of the two classifiers for a tuning based on our proposition and the one based on grid search method.

查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
基于Fβ测度的谨慎分类器决策规则的超参数建模研究
在一些存在数据缺陷的敏感领域,标准分类技术达到了极限。为了避免产生严重后果的错误分类,最近的工作提出了谨慎的分类算法来处理这个问题。尽管存在不确定性和/或不精确性,点预测分类器还是被迫将赌注押在单个类别上。而谨慎的分类器提出了在存在不完美信息的情况下可以分配给样本的候选类的适当子集。另一方面,谨慎不应以牺牲准确性为代价,必须在这两个标准之间进行权衡。在现有的谨慎分类器中,有两个分类器提出通过参数化的目标函数来管理决策步骤中的这种权衡。第一种是在概率论框架内提出的非确定性分类器(ndc),第二种是在置信函数框架下提出的“基于不精确重新标记的证据分类器”(eclair)。上述超参数的理论目的是控制两个分类器的预测大小。本文提出研究这个超参数,以便在分类任务中选择“最佳”值。首先,研究了每个候选子集与超参数值相关的效用,并提出了一些阈值来控制预测的大小。然后给出了两个例子,其中提出了一种基于校准数据选择该超参数的方法。第一个图示涉及随机生成的数据,第二个图示涉及时尚mnist的图像数据。这些插图展示了如何控制预测的大小,并比较了基于我们的命题和基于网格搜索方法的两个分类器的性能。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
Array
Array Computer Science-General Computer Science
CiteScore
4.40
自引率
0.00%
发文量
93
审稿时长
45 days
期刊最新文献
DART: A Solution for decentralized federated learning model robustness analysis Autonomous UAV navigation using deep learning-based computer vision frameworks: A systematic literature review Threat intelligence named entity recognition techniques based on few-shot learning Reimagining otitis media diagnosis: A fusion of nested U-Net segmentation with graph theory-inspired feature set Modeling and supporting adaptive Complex Data-Intensive Web Systems via XML and the O-O paradigm: The OO-XAHM model
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1