{"title":"基于并行增量支持向量机的高维图像分类算法","authors":"Thanh-Nghi Doan, Thanh-Nghi Do, F. Poulet","doi":"10.1109/IJCNN.2013.6707121","DOIUrl":null,"url":null,"abstract":"ImageNet dataset [1] with more than 14M images and 21K classes makes the problem of visual classification more difficult to deal with. One of the most difficult tasks is to train a fast and accurate classifier on computers with limited memory resource. In this paper, we address this challenge by extending the state-of-the-art large scale classifier Power Mean SVM (PmSVM) proposed by Jianxin Wu [2] in three ways: (1) An incremental learning for PmSVM, (2) A balanced bagging algorithm for training binary classifiers, (3) Parallelize the training process of classifiers with several multi-core computers. Our approach is evaluated on 1K classes of ImageNet (ILSVRC 1000 [3]). The evaluation shows that our approach can save up to 84.34% memory usage and the training process is 297 times faster than the original implementation and 1508 times faster than the state-of-the-art linear classifier (LIBLINEAR [4]).","PeriodicalId":376975,"journal":{"name":"The 2013 International Joint Conference on Neural Networks (IJCNN)","volume":"61 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2013-08-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"4","resultStr":"{\"title\":\"Parallel incremental SVM for classifying million images with very high-dimensional signatures into thousand classes\",\"authors\":\"Thanh-Nghi Doan, Thanh-Nghi Do, F. Poulet\",\"doi\":\"10.1109/IJCNN.2013.6707121\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"ImageNet dataset [1] with more than 14M images and 21K classes makes the problem of visual classification more difficult to deal with. One of the most difficult tasks is to train a fast and accurate classifier on computers with limited memory resource. In this paper, we address this challenge by extending the state-of-the-art large scale classifier Power Mean SVM (PmSVM) proposed by Jianxin Wu [2] in three ways: (1) An incremental learning for PmSVM, (2) A balanced bagging algorithm for training binary classifiers, (3) Parallelize the training process of classifiers with several multi-core computers. Our approach is evaluated on 1K classes of ImageNet (ILSVRC 1000 [3]). The evaluation shows that our approach can save up to 84.34% memory usage and the training process is 297 times faster than the original implementation and 1508 times faster than the state-of-the-art linear classifier (LIBLINEAR [4]).\",\"PeriodicalId\":376975,\"journal\":{\"name\":\"The 2013 International Joint Conference on Neural Networks (IJCNN)\",\"volume\":\"61 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2013-08-04\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"4\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"The 2013 International Joint Conference on Neural Networks (IJCNN)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/IJCNN.2013.6707121\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"The 2013 International Joint Conference on Neural Networks (IJCNN)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/IJCNN.2013.6707121","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 4
摘要
ImageNet数据集[1]拥有超过14M张图像和21K个类,使得视觉分类问题更加难以处理。在内存有限的计算机上训练快速准确的分类器是最困难的任务之一。在本文中,我们通过三种方式扩展了由Jianxin Wu[2]提出的最先进的大规模分类器Power Mean SVM (PmSVM)来解决这一挑战:(1)PmSVM的增量学习,(2)训练二分类器的平衡bagging算法,(3)在多核计算机上并行化分类器的训练过程。我们的方法在1K个ImageNet类(ILSVRC 1000[3])上进行了评估。评估表明,我们的方法可以节省高达84.34%的内存使用,训练过程比原始实现快297倍,比最先进的线性分类器(LIBLINEAR[4])快1508倍。
Parallel incremental SVM for classifying million images with very high-dimensional signatures into thousand classes
ImageNet dataset [1] with more than 14M images and 21K classes makes the problem of visual classification more difficult to deal with. One of the most difficult tasks is to train a fast and accurate classifier on computers with limited memory resource. In this paper, we address this challenge by extending the state-of-the-art large scale classifier Power Mean SVM (PmSVM) proposed by Jianxin Wu [2] in three ways: (1) An incremental learning for PmSVM, (2) A balanced bagging algorithm for training binary classifiers, (3) Parallelize the training process of classifiers with several multi-core computers. Our approach is evaluated on 1K classes of ImageNet (ILSVRC 1000 [3]). The evaluation shows that our approach can save up to 84.34% memory usage and the training process is 297 times faster than the original implementation and 1508 times faster than the state-of-the-art linear classifier (LIBLINEAR [4]).