使用模糊互位置矩阵和基于显著性的谓词排序来描述图像

Marcin Iwanowski, Mateusz Bartosiewicz
{"title":"使用模糊互位置矩阵和基于显著性的谓词排序来描述图像","authors":"Marcin Iwanowski, Mateusz Bartosiewicz","doi":"10.1109/FUZZ45933.2021.9494549","DOIUrl":null,"url":null,"abstract":"Describing the content based on bounding boxes of objects located within the image has recently gained popularity thanks to the fast development of object detection algorithms based on deep learning. Such description, however, does not contain any information on the mutual relations between objects that may be crucial to understand the scene as a whole. In the paper, a method is proposed that extracts, from the set of bounding boxes, a scene description in the form of a list of predicates containing consecutive objects' position, referring them to previously described ones. To estimate bounding boxes' relative position, a fuzzy mutual position matrix is proposed. It contains the complete information on the scene composition stored in fuzzy 2-D position descriptors extracted from fuzzified relative bounding box coordinates by a two-stage fuzzy reasoning process. The descriptors of non-zero membership function values are next considered as potential predicates related to the image content. Their list is ordered using the saliency-based criteria to select the most relevant ones, explaining best the scene composition. From the ordered list, the algorithm extracts the final list of predicates. It contains complete and concise information on the composition of objects within the scene. Some examples of the proposed method illustrate the paper.","PeriodicalId":151289,"journal":{"name":"2021 IEEE International Conference on Fuzzy Systems (FUZZ-IEEE)","volume":"316 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-07-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"2","resultStr":"{\"title\":\"Describing images using fuzzy mutual position matrix and saliency-based ordering of predicates\",\"authors\":\"Marcin Iwanowski, Mateusz Bartosiewicz\",\"doi\":\"10.1109/FUZZ45933.2021.9494549\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Describing the content based on bounding boxes of objects located within the image has recently gained popularity thanks to the fast development of object detection algorithms based on deep learning. Such description, however, does not contain any information on the mutual relations between objects that may be crucial to understand the scene as a whole. In the paper, a method is proposed that extracts, from the set of bounding boxes, a scene description in the form of a list of predicates containing consecutive objects' position, referring them to previously described ones. To estimate bounding boxes' relative position, a fuzzy mutual position matrix is proposed. It contains the complete information on the scene composition stored in fuzzy 2-D position descriptors extracted from fuzzified relative bounding box coordinates by a two-stage fuzzy reasoning process. The descriptors of non-zero membership function values are next considered as potential predicates related to the image content. Their list is ordered using the saliency-based criteria to select the most relevant ones, explaining best the scene composition. From the ordered list, the algorithm extracts the final list of predicates. It contains complete and concise information on the composition of objects within the scene. Some examples of the proposed method illustrate the paper.\",\"PeriodicalId\":151289,\"journal\":{\"name\":\"2021 IEEE International Conference on Fuzzy Systems (FUZZ-IEEE)\",\"volume\":\"316 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2021-07-11\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"2\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2021 IEEE International Conference on Fuzzy Systems (FUZZ-IEEE)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/FUZZ45933.2021.9494549\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 IEEE International Conference on Fuzzy Systems (FUZZ-IEEE)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/FUZZ45933.2021.9494549","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 2

摘要

由于基于深度学习的对象检测算法的快速发展,基于图像内对象的边界框描述内容的方法最近得到了普及。然而,这样的描述不包含任何关于物体之间相互关系的信息,而这些信息对于理解整个场景至关重要。本文提出了一种方法,从边界框集合中提取包含连续对象位置的谓词列表形式的场景描述,并将它们引用到先前描述的对象。为了估计边界框的相对位置,提出了模糊互位置矩阵。通过两阶段模糊推理过程,从模糊化的相对边界框坐标中提取模糊二维位置描述符,存储场景组成的完整信息。非零隶属函数值的描述符接下来被视为与图像内容相关的潜在谓词。他们的列表使用基于显著性的标准来选择最相关的,最好地解释场景构图。从有序列表中,算法提取谓词的最终列表。它包含了场景中物体组成的完整而简洁的信息。文中的一些算例说明了该方法的有效性。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
Describing images using fuzzy mutual position matrix and saliency-based ordering of predicates
Describing the content based on bounding boxes of objects located within the image has recently gained popularity thanks to the fast development of object detection algorithms based on deep learning. Such description, however, does not contain any information on the mutual relations between objects that may be crucial to understand the scene as a whole. In the paper, a method is proposed that extracts, from the set of bounding boxes, a scene description in the form of a list of predicates containing consecutive objects' position, referring them to previously described ones. To estimate bounding boxes' relative position, a fuzzy mutual position matrix is proposed. It contains the complete information on the scene composition stored in fuzzy 2-D position descriptors extracted from fuzzified relative bounding box coordinates by a two-stage fuzzy reasoning process. The descriptors of non-zero membership function values are next considered as potential predicates related to the image content. Their list is ordered using the saliency-based criteria to select the most relevant ones, explaining best the scene composition. From the ordered list, the algorithm extracts the final list of predicates. It contains complete and concise information on the composition of objects within the scene. Some examples of the proposed method illustrate the paper.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
XAI Models for Quality of Experience Prediction in Wireless Networks Application of the Fuzzy Logic to Evaluation and Selection of Attribute Ranges in Machine Learning Kernel-Based k-Representatives Algorithm for Fuzzy Clustering of Categorical Data Necessary and sufficient condition for the existence of Atanassov's Intuitionistic Fuzzy based additive definite integral Identifying and Rectifying Rational Gaps in Fuzzy Rule Based Systems for Regression Problems
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1