{"title":"Error-Driven Triplet-Based Online Fine-Tuning for Cross-Background Image Classification","authors":"Sheng-Luen Chung, Wei-Ting Guo","doi":"10.1109/CCAI57533.2023.10201306","DOIUrl":null,"url":null,"abstract":"Image classification is a fundamental task in computer vision with numerous real-world applications. However, classification models trained on one set of images may not perform well when tested on another set, especially when the two sets differ in terms of the background environment. To address this challenge, we propose an error-driven triplet-based online fine-tuning approach that leverages misclassified samples to refine the classification model at intervals when enough misclassified samples are collected. Our approach builds on the triplet network architecture, which learns to represent images in a low-dimensional space where images with the same label are clustered together. We use a pre-trained classification model that was trained on a collection of 180 types of images from one background scene. However, when we apply the model to a new background scene with additional types of images, its performance is compromised due to the domain shift. Our proposed approach leverages the misclassified samples by contrast them with positive and negative samples as triplet data to fine-tune the model in new background scene. We use a loss function that combines the triplet loss and the classification loss to update the model weights. We evaluate our approach on two challenging image classification datasets with different background environments. The experimental results demonstrate that our approach achieves significant improvements in accuracy compared to the baseline classification model without fine-tuning. Overall, our error-driven triplet-based online fine-tuning approach shows promising results for adapting classification models to changing background with additionally new classification types, where the performance of pre-trained models is limited.","PeriodicalId":285760,"journal":{"name":"2023 IEEE 3rd International Conference on Computer Communication and Artificial Intelligence (CCAI)","volume":"16 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-05-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2023 IEEE 3rd International Conference on Computer Communication and Artificial Intelligence (CCAI)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/CCAI57533.2023.10201306","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Image classification is a fundamental task in computer vision with numerous real-world applications. However, classification models trained on one set of images may not perform well when tested on another set, especially when the two sets differ in terms of the background environment. To address this challenge, we propose an error-driven triplet-based online fine-tuning approach that leverages misclassified samples to refine the classification model at intervals when enough misclassified samples are collected. Our approach builds on the triplet network architecture, which learns to represent images in a low-dimensional space where images with the same label are clustered together. We use a pre-trained classification model that was trained on a collection of 180 types of images from one background scene. However, when we apply the model to a new background scene with additional types of images, its performance is compromised due to the domain shift. Our proposed approach leverages the misclassified samples by contrast them with positive and negative samples as triplet data to fine-tune the model in new background scene. We use a loss function that combines the triplet loss and the classification loss to update the model weights. We evaluate our approach on two challenging image classification datasets with different background environments. The experimental results demonstrate that our approach achieves significant improvements in accuracy compared to the baseline classification model without fine-tuning. Overall, our error-driven triplet-based online fine-tuning approach shows promising results for adapting classification models to changing background with additionally new classification types, where the performance of pre-trained models is limited.