{"title":"基于多平均值的伪近邻分类器","authors":"Dapeng Li, Jing Guo","doi":"10.3233/aic-230312","DOIUrl":null,"url":null,"abstract":"Conventional k nearest neighbor (KNN) rule is a simple yet effective method for classification, but its classification performance is easily degraded in the case of small size training samples with existing outliers. To address this issue, A multi-average based pseudo nearest neighbor classifier (MAPNN) rule is proposed. In the proposed MAPNN rule, k ( k − 1 ) / 2 ( k > 1) local mean vectors of each class are obtained by taking the average of two points randomly from k nearest neighbors in every category, and then k pseudo nearest neighbors are chosen from k ( k − 1 ) / 2 local mean neighbors of every class to determine the category of a query point. The selected k pseudo nearest neighbors can reduce the negative impact of outliers in some degree. Extensive experiments are carried out on twenty-one numerical real data sets and four artificial data sets by comparing MAPNN to other five KNN-based methods. The experimental results demonstrate that the proposed MAPNN is effective for classification task and achieves better classification results in the small-size samples cases comparing to five relative KNN-based classifiers.","PeriodicalId":505412,"journal":{"name":"AI Communications","volume":"44 5","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-03-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"A multi-average based pseudo nearest neighbor classifier\",\"authors\":\"Dapeng Li, Jing Guo\",\"doi\":\"10.3233/aic-230312\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Conventional k nearest neighbor (KNN) rule is a simple yet effective method for classification, but its classification performance is easily degraded in the case of small size training samples with existing outliers. To address this issue, A multi-average based pseudo nearest neighbor classifier (MAPNN) rule is proposed. In the proposed MAPNN rule, k ( k − 1 ) / 2 ( k > 1) local mean vectors of each class are obtained by taking the average of two points randomly from k nearest neighbors in every category, and then k pseudo nearest neighbors are chosen from k ( k − 1 ) / 2 local mean neighbors of every class to determine the category of a query point. The selected k pseudo nearest neighbors can reduce the negative impact of outliers in some degree. Extensive experiments are carried out on twenty-one numerical real data sets and four artificial data sets by comparing MAPNN to other five KNN-based methods. The experimental results demonstrate that the proposed MAPNN is effective for classification task and achieves better classification results in the small-size samples cases comparing to five relative KNN-based classifiers.\",\"PeriodicalId\":505412,\"journal\":{\"name\":\"AI Communications\",\"volume\":\"44 5\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-03-28\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"AI Communications\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.3233/aic-230312\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"AI Communications","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.3233/aic-230312","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
摘要
传统的 k 近邻(KNN)规则是一种简单而有效的分类方法,但在训练样本较小且存在异常值的情况下,其分类性能很容易下降。为了解决这个问题,我们提出了一种基于多平均值的伪近邻分类器(MAPNN)规则。在所提出的 MAPNN 规则中,每个类别的 k ( k - 1 ) / 2 ( k > 1 ) 个局部均值向量是通过从每个类别的 k 个近邻中随机取两个点的平均值得到的,然后从每个类别的 k ( k - 1 ) / 2 个局部均值近邻中选择 k 个伪近邻来确定查询点的类别。选出的 k 个伪近邻可以在一定程度上减少异常值的负面影响。通过将 MAPNN 与其他五种基于 KNN 的方法进行比较,在 21 个数值真实数据集和 4 个人工数据集上进行了广泛的实验。实验结果表明,与基于 KNN 的五种分类器相比,所提出的 MAPNN 能有效地完成分类任务,并在小样本情况下取得更好的分类结果。
A multi-average based pseudo nearest neighbor classifier
Conventional k nearest neighbor (KNN) rule is a simple yet effective method for classification, but its classification performance is easily degraded in the case of small size training samples with existing outliers. To address this issue, A multi-average based pseudo nearest neighbor classifier (MAPNN) rule is proposed. In the proposed MAPNN rule, k ( k − 1 ) / 2 ( k > 1) local mean vectors of each class are obtained by taking the average of two points randomly from k nearest neighbors in every category, and then k pseudo nearest neighbors are chosen from k ( k − 1 ) / 2 local mean neighbors of every class to determine the category of a query point. The selected k pseudo nearest neighbors can reduce the negative impact of outliers in some degree. Extensive experiments are carried out on twenty-one numerical real data sets and four artificial data sets by comparing MAPNN to other five KNN-based methods. The experimental results demonstrate that the proposed MAPNN is effective for classification task and achieves better classification results in the small-size samples cases comparing to five relative KNN-based classifiers.