You Only Look at Interested Cells: Real-Time Object Detection Based on Cell-Wise Segmentation

Kai Su, Huitao Wang, I. M. Chowdhury, Qiangfu Zhao, Yoichi Tomioka
{"title":"You Only Look at Interested Cells: Real-Time Object Detection Based on Cell-Wise Segmentation","authors":"Kai Su, Huitao Wang, I. M. Chowdhury, Qiangfu Zhao, Yoichi Tomioka","doi":"10.1109/iCAST51195.2020.9319469","DOIUrl":null,"url":null,"abstract":"In this paper, we study real-time object detection based on cell-wise segmentation. Existing object detection methods usually focus on detecting interesting object's positions and sizes and demand expensive computing resources. This process makes it difficult to achieve high-speed and high-precision detection with low-cost devices. We propose a method called You Only Look at Interested Cells or in-short YOLIC to solve the problem by focusing on predefined interested cells (i.e., subregions) in an image. A key challenge here is how to predict the object types contained in all interested cells efficiently, all at once. Instead of using multiple predictors for all interested cells, we use only one deep learner to classify all interested cells. In other words, YOLIC applies the concept of multi-label classification for object detection. YOLIC can use exiting classification models without any structural change. The main point is to define a proper loss function for training. Using on-road risk detection as a test case, we confirmed that YOLIC is significantly faster and accurate than YOLO-v3 in terms of FPS and F1-score.","PeriodicalId":212570,"journal":{"name":"2020 11th International Conference on Awareness Science and Technology (iCAST)","volume":"8 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-12-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"2","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2020 11th International Conference on Awareness Science and Technology (iCAST)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/iCAST51195.2020.9319469","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 2

Abstract

In this paper, we study real-time object detection based on cell-wise segmentation. Existing object detection methods usually focus on detecting interesting object's positions and sizes and demand expensive computing resources. This process makes it difficult to achieve high-speed and high-precision detection with low-cost devices. We propose a method called You Only Look at Interested Cells or in-short YOLIC to solve the problem by focusing on predefined interested cells (i.e., subregions) in an image. A key challenge here is how to predict the object types contained in all interested cells efficiently, all at once. Instead of using multiple predictors for all interested cells, we use only one deep learner to classify all interested cells. In other words, YOLIC applies the concept of multi-label classification for object detection. YOLIC can use exiting classification models without any structural change. The main point is to define a proper loss function for training. Using on-road risk detection as a test case, we confirmed that YOLIC is significantly faster and accurate than YOLO-v3 in terms of FPS and F1-score.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
你只看感兴趣的细胞:基于细胞明智分割的实时目标检测
本文研究了基于单元分割的实时目标检测。现有的目标检测方法通常集中于检测感兴趣的目标的位置和大小,需要耗费昂贵的计算资源。这一过程使得用低成本的设备实现高速、高精度的检测变得困难。我们提出了一种名为You Only Look at Interested Cells(简称YOLIC)的方法,通过关注图像中预定义的感兴趣的细胞(即子区域)来解决这个问题。这里的一个关键挑战是如何一次有效地预测所有感兴趣的单元格中包含的对象类型。我们只使用一个深度学习器对所有感兴趣的细胞进行分类,而不是对所有感兴趣的细胞使用多个预测器。换句话说,YOLIC将多标签分类的概念应用于目标检测。YOLIC可以在不改变结构的情况下使用现有的分类模型。重点是定义一个合适的训练损失函数。以道路风险检测为例,我们证实YOLIC在FPS和f1分数方面明显比YOLO-v3更快、更准确。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Skeleton Guided Conflict-Free Hand Gesture Recognition for Robot Control Improved Spiking Neural Networks with multiple neurons for digit recognition A Lightweight Transformer with Convolutional Attention Social Media Mining with Dynamic Clustering: A Case Study by COVID-19 Tweets A Visual-SLAM based Line Laser Scanning System using Semantically Segmented Images
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1