TSS-Net: Time-based Semantic Segmentation Neural Network for Road Scene Understanding

Tin Trung Duong, Huy-Hung Nguyen, J. Jeon
{"title":"TSS-Net: Time-based Semantic Segmentation Neural Network for Road Scene Understanding","authors":"Tin Trung Duong, Huy-Hung Nguyen, J. Jeon","doi":"10.1109/IMCOM51814.2021.9377401","DOIUrl":null,"url":null,"abstract":"In this research, a multitask convolutional neural network that can do end-to-end road scene classification and semantic segmentation, which are the two crucial tasks for advanced driver assistance systems (ADAS), is proposed. We name the network TSS which means time-based semantic segmentation. The network contains three main modules: an image encoder, a scene classifier, and two time-based segmentation decoders. For each road scene image, the encoder extracts image features which will be used for classifier and decoders. Next, the image features are fed to the classifier to predict the scene type (in this case a day or a night scene). Then, based on the predicted scene type, the same extracted features are fed to a corresponding segmentation decoder to produce the final semantic segmentation result. By using this classification-driven decoder approach, we can improve the accuracy of the segmentation model, even when the model has been trained excessively earlier. Through the experiment, the validity of our proposed method has been proven. Our approach can be considered as stacking multiple segmentation modules on top of the classification module with all of them share the same image encoder. With this approach, we can utilize the result from classification to gain more accuracy in segmentation in one feed forward only.","PeriodicalId":275121,"journal":{"name":"2021 15th International Conference on Ubiquitous Information Management and Communication (IMCOM)","volume":"2013 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-01-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 15th International Conference on Ubiquitous Information Management and Communication (IMCOM)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/IMCOM51814.2021.9377401","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

In this research, a multitask convolutional neural network that can do end-to-end road scene classification and semantic segmentation, which are the two crucial tasks for advanced driver assistance systems (ADAS), is proposed. We name the network TSS which means time-based semantic segmentation. The network contains three main modules: an image encoder, a scene classifier, and two time-based segmentation decoders. For each road scene image, the encoder extracts image features which will be used for classifier and decoders. Next, the image features are fed to the classifier to predict the scene type (in this case a day or a night scene). Then, based on the predicted scene type, the same extracted features are fed to a corresponding segmentation decoder to produce the final semantic segmentation result. By using this classification-driven decoder approach, we can improve the accuracy of the segmentation model, even when the model has been trained excessively earlier. Through the experiment, the validity of our proposed method has been proven. Our approach can be considered as stacking multiple segmentation modules on top of the classification module with all of them share the same image encoder. With this approach, we can utilize the result from classification to gain more accuracy in segmentation in one feed forward only.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
基于时间的道路场景理解语义分割神经网络
针对先进驾驶辅助系统(ADAS)的关键任务端到端道路场景分类和语义分割,提出了一种多任务卷积神经网络。我们将该网络命名为TSS,即基于时间的语义分割。该网络包含三个主要模块:一个图像编码器、一个场景分类器和两个基于时间的分割解码器。对于每个道路场景图像,编码器提取图像特征,用于分类器和解码器。接下来,将图像特征馈送到分类器以预测场景类型(在本例中是白天或夜间场景)。然后,根据预测的场景类型,将相同提取的特征馈送到相应的分割解码器,从而产生最终的语义分割结果。通过使用这种分类驱动的解码器方法,我们可以提高分割模型的准确性,即使模型已经训练得太早了。通过实验验证了该方法的有效性。我们的方法可以看作是将多个分割模块叠加在分类模块之上,并且所有的分割模块都共享相同的图像编码器。通过这种方法,我们可以利用分类的结果在一个前馈中获得更高的分割精度。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
On a Partially Verifiable Multi-party Multi-argument Zero-knowledge Proof EnvBERT: Multi-Label Text Classification for Imbalanced, Noisy Environmental News Data Method for Changing Users' Attitudes Towards Fashion Styling by Showing Evaluations After Coordinate Selection The Analysis of Web Search Snippets Displaying User's Knowledge An Energy Management System with Edge Computing for Industrial Facility
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1