Kai Su, Huitao Wang, I. M. Chowdhury, Qiangfu Zhao, Yoichi Tomioka
{"title":"You Only Look at Interested Cells: Real-Time Object Detection Based on Cell-Wise Segmentation","authors":"Kai Su, Huitao Wang, I. M. Chowdhury, Qiangfu Zhao, Yoichi Tomioka","doi":"10.1109/iCAST51195.2020.9319469","DOIUrl":null,"url":null,"abstract":"In this paper, we study real-time object detection based on cell-wise segmentation. Existing object detection methods usually focus on detecting interesting object's positions and sizes and demand expensive computing resources. This process makes it difficult to achieve high-speed and high-precision detection with low-cost devices. We propose a method called You Only Look at Interested Cells or in-short YOLIC to solve the problem by focusing on predefined interested cells (i.e., subregions) in an image. A key challenge here is how to predict the object types contained in all interested cells efficiently, all at once. Instead of using multiple predictors for all interested cells, we use only one deep learner to classify all interested cells. In other words, YOLIC applies the concept of multi-label classification for object detection. YOLIC can use exiting classification models without any structural change. The main point is to define a proper loss function for training. Using on-road risk detection as a test case, we confirmed that YOLIC is significantly faster and accurate than YOLO-v3 in terms of FPS and F1-score.","PeriodicalId":212570,"journal":{"name":"2020 11th International Conference on Awareness Science and Technology (iCAST)","volume":"8 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-12-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"2","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2020 11th International Conference on Awareness Science and Technology (iCAST)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/iCAST51195.2020.9319469","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 2
Abstract
In this paper, we study real-time object detection based on cell-wise segmentation. Existing object detection methods usually focus on detecting interesting object's positions and sizes and demand expensive computing resources. This process makes it difficult to achieve high-speed and high-precision detection with low-cost devices. We propose a method called You Only Look at Interested Cells or in-short YOLIC to solve the problem by focusing on predefined interested cells (i.e., subregions) in an image. A key challenge here is how to predict the object types contained in all interested cells efficiently, all at once. Instead of using multiple predictors for all interested cells, we use only one deep learner to classify all interested cells. In other words, YOLIC applies the concept of multi-label classification for object detection. YOLIC can use exiting classification models without any structural change. The main point is to define a proper loss function for training. Using on-road risk detection as a test case, we confirmed that YOLIC is significantly faster and accurate than YOLO-v3 in terms of FPS and F1-score.
本文研究了基于单元分割的实时目标检测。现有的目标检测方法通常集中于检测感兴趣的目标的位置和大小,需要耗费昂贵的计算资源。这一过程使得用低成本的设备实现高速、高精度的检测变得困难。我们提出了一种名为You Only Look at Interested Cells(简称YOLIC)的方法,通过关注图像中预定义的感兴趣的细胞(即子区域)来解决这个问题。这里的一个关键挑战是如何一次有效地预测所有感兴趣的单元格中包含的对象类型。我们只使用一个深度学习器对所有感兴趣的细胞进行分类,而不是对所有感兴趣的细胞使用多个预测器。换句话说,YOLIC将多标签分类的概念应用于目标检测。YOLIC可以在不改变结构的情况下使用现有的分类模型。重点是定义一个合适的训练损失函数。以道路风险检测为例,我们证实YOLIC在FPS和f1分数方面明显比YOLO-v3更快、更准确。