Image Generation from Scene Graph with Object Edges

Chenxing Li, Yiping Duan, Qiyuan Du, Chengkang Pan, Guangyi Liu, Xiaoming Tao
{"title":"Image Generation from Scene Graph with Object Edges","authors":"Chenxing Li, Yiping Duan, Qiyuan Du, Chengkang Pan, Guangyi Liu, Xiaoming Tao","doi":"10.1109/VTC2022-Fall57202.2022.10012878","DOIUrl":null,"url":null,"abstract":"Significant progress has been made on methods for generating images from structured semantic descriptions, but the generated images only retain semantic information, and the appearance of objects cannot be constrained and effectively represented. Therefore, we propose a scene graph structure image generation method assisted by object edge information. Our model uses two graph convolution neural networks(GCN) to process scene graphs and obtains object features as well as relation features which aggregate related information. The object bounding boxes are predicted by a method a decoupling the size and position. Where auxiliary models are added to coordinate with segmentation mask network training. Our experiments show that the introduction of object edges provides clearer object appearance information for image generation, which can constrain object shapes and improve image quality greatly. Finally, the cascaded refinement network is used to generate images. Additionally, compared with other appearance features, such as object slices, edge information occupies a smaller quantity of data, which greatly improves the image quality with less increase in the input information. This feature also benefits semantic communication systems. A large number of experiments show that our method is significantly superior to the latest Sg2im method when evaluated on Visual Genome datasets.","PeriodicalId":326047,"journal":{"name":"2022 IEEE 96th Vehicular Technology Conference (VTC2022-Fall)","volume":"11 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"3","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 IEEE 96th Vehicular Technology Conference (VTC2022-Fall)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/VTC2022-Fall57202.2022.10012878","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 3

Abstract

Significant progress has been made on methods for generating images from structured semantic descriptions, but the generated images only retain semantic information, and the appearance of objects cannot be constrained and effectively represented. Therefore, we propose a scene graph structure image generation method assisted by object edge information. Our model uses two graph convolution neural networks(GCN) to process scene graphs and obtains object features as well as relation features which aggregate related information. The object bounding boxes are predicted by a method a decoupling the size and position. Where auxiliary models are added to coordinate with segmentation mask network training. Our experiments show that the introduction of object edges provides clearer object appearance information for image generation, which can constrain object shapes and improve image quality greatly. Finally, the cascaded refinement network is used to generate images. Additionally, compared with other appearance features, such as object slices, edge information occupies a smaller quantity of data, which greatly improves the image quality with less increase in the input information. This feature also benefits semantic communication systems. A large number of experiments show that our method is significantly superior to the latest Sg2im method when evaluated on Visual Genome datasets.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
从带有物体边缘的场景图生成图像
从结构化语义描述生成图像的方法已经取得了重大进展,但生成的图像只保留了语义信息,不能约束和有效地表示对象的外观。为此,我们提出了一种利用物体边缘信息辅助的场景图结构图像生成方法。该模型采用两个图卷积神经网络(GCN)对场景图进行处理,得到对象特征和聚合相关信息的关系特征。通过解耦大小和位置的方法预测对象边界框。其中加入辅助模型配合分割掩码网络训练。我们的实验表明,引入物体边缘为图像生成提供了更清晰的物体外观信息,可以约束物体形状,大大提高图像质量。最后,利用级联细化网络生成图像。此外,与物体切片等其他外观特征相比,边缘信息占用的数据量更小,在输入信息增加较少的情况下,极大地提高了图像质量。这个特性也有利于语义通信系统。大量实验表明,在Visual Genome数据集上,我们的方法明显优于最新的Sg2im方法。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Non-Orthogonal Neighbor Election Random Access for Distributed 6G Wireless Networks Coverage Performance Analysis of Piggyback Mobile IoT in 5G Vehicular Networks Performance Comparison of Error-Control Schemes in Collaborative Multiple-Input Multiple-Output Systems Valuation-Aware Federated Learning: An Auction-Based Approach for User Selection Design of Robust LoS-MIMO Transmission in HAPS Feeder Link
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1