Ali Nuri Şeker, Hüseyin Doğan, Muhammet Üsame Öziç
{"title":"A Novel Approach on Unsupervised Dynamic Background Extraction Using Autoencoders","authors":"Ali Nuri Şeker, Hüseyin Doğan, Muhammet Üsame Öziç","doi":"10.1109/ISMSIT52890.2021.9604737","DOIUrl":null,"url":null,"abstract":"In this Study a novel method has been used for extracting the background from given images. Different from the existing approaches, a Convolutional Neural Network (CNN) Autoencoder (AE) has been trained with frames produced from the same stationary camera source, paired with random frames from the same pool for each sample as a label. A little over 4000 RGB images with the dimensions of 640x480 has been used for training and around 450 of them was used for testing. The mentioned model has 4 convolutional layers each in encoder and decoder sections. The training was conducted for 500 epochs and the value of epoch loss went down to 2.13x10-3 and 2.38x10-3 for training and validation respectively. After the training of the model, generated background samples were subtracted from the input images and was turned into a binary image using two different segmentation methods: HSV Thresholding and OTSU. To use as the ground truth, test images were hand labeled. Mentioned approach had an F1-score of 62.36% for HSV Thresholding and 69.63% for OTSU methods.","PeriodicalId":120997,"journal":{"name":"2021 5th International Symposium on Multidisciplinary Studies and Innovative Technologies (ISMSIT)","volume":"51 2 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-10-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 5th International Symposium on Multidisciplinary Studies and Innovative Technologies (ISMSIT)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ISMSIT52890.2021.9604737","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
In this Study a novel method has been used for extracting the background from given images. Different from the existing approaches, a Convolutional Neural Network (CNN) Autoencoder (AE) has been trained with frames produced from the same stationary camera source, paired with random frames from the same pool for each sample as a label. A little over 4000 RGB images with the dimensions of 640x480 has been used for training and around 450 of them was used for testing. The mentioned model has 4 convolutional layers each in encoder and decoder sections. The training was conducted for 500 epochs and the value of epoch loss went down to 2.13x10-3 and 2.38x10-3 for training and validation respectively. After the training of the model, generated background samples were subtracted from the input images and was turned into a binary image using two different segmentation methods: HSV Thresholding and OTSU. To use as the ground truth, test images were hand labeled. Mentioned approach had an F1-score of 62.36% for HSV Thresholding and 69.63% for OTSU methods.