首页 > 最新文献

计算机辅助设计与图形学学报最新文献

英文 中文
Dynamic Obstacle Avoidance Method for Carrier Aircraft Based on Deep Reinforcement Learning 基于深度强化学习的舰载机动态避障方法
Q3 Computer Science Pub Date : 2021-07-01 DOI: 10.3724/sp.j.1089.2021.18637
Junxiao Xue, Xiangya Kong, Yibo Guo, Aiguo Lu, Jian Li, Xi Wan, Mingliang Xu
{"title":"Dynamic Obstacle Avoidance Method for Carrier Aircraft Based on Deep Reinforcement Learning","authors":"Junxiao Xue, Xiangya Kong, Yibo Guo, Aiguo Lu, Jian Li, Xi Wan, Mingliang Xu","doi":"10.3724/sp.j.1089.2021.18637","DOIUrl":"https://doi.org/10.3724/sp.j.1089.2021.18637","url":null,"abstract":"","PeriodicalId":52442,"journal":{"name":"计算机辅助设计与图形学学报","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2021-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"43778362","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Efficient 3D Object Detection of Indoor Scenes Based on RGB-D Video Stream 基于RGB-D视频流的室内场景三维目标高效检测
Q3 Computer Science Pub Date : 2021-07-01 DOI: 10.3724/sp.j.1089.2021.18630
Miao Yongwei, Jiahui Chen, Xinjie Zhang, Ma Wenjuan, S. Sun
: For indoor object detection, the input complex scenes often have some defects such as incomplete RGB-D scanning data or mutual occlusion of its objects. Meanwhile, due to the limitations of frame in the RGB-D video stream. Using SUN RGB-D dataset to train the object detection network of key frame, the detection result of proposed method is accurate, and the overall detection time is greatly reduced if com-paring with the VoteNet based frame-by-frame detection scheme. Experimental results demonstrate that proposed method is effective and efficient.
:对于室内目标检测,输入的复杂场景往往存在RGB-D扫描数据不完整或其目标相互遮挡等缺陷。同时,由于RGB-D视频流中帧数的限制。利用SUN RGB-D数据集对关键帧的目标检测网络进行训练,检测结果准确,与基于VoteNet的逐帧检测方案相比,整体检测时间大大缩短。实验结果表明,该方法是有效的。
{"title":"Efficient 3D Object Detection of Indoor Scenes Based on RGB-D Video Stream","authors":"Miao Yongwei, Jiahui Chen, Xinjie Zhang, Ma Wenjuan, S. Sun","doi":"10.3724/sp.j.1089.2021.18630","DOIUrl":"https://doi.org/10.3724/sp.j.1089.2021.18630","url":null,"abstract":": For indoor object detection, the input complex scenes often have some defects such as incomplete RGB-D scanning data or mutual occlusion of its objects. Meanwhile, due to the limitations of frame in the RGB-D video stream. Using SUN RGB-D dataset to train the object detection network of key frame, the detection result of proposed method is accurate, and the overall detection time is greatly reduced if com-paring with the VoteNet based frame-by-frame detection scheme. Experimental results demonstrate that proposed method is effective and efficient.","PeriodicalId":52442,"journal":{"name":"计算机辅助设计与图形学学报","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2021-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"41602686","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Information Hiding Scheme Based on Quantum Generative Adversarial Network 基于量子生成对抗网络的信息隐藏方案
Q3 Computer Science Pub Date : 2021-07-01 DOI: 10.3724/sp.j.1089.2021.18617
Jia Luo, Rigui Zhou, Yaochong Li, Guangzhong Liu
: Due to the insecurity of quantum image information hiding technology in the face of statisti-cal-based steganalysis algorithm detection, an information hiding scheme based on quantum generative adversarial network (QGAN) is proposed. This scheme first uses the mapping rules to map the secret information into the single qubit gate to prepare for the input state of the parameterized quantum circuit of the gen-erator G . Then the stego quantum image is generated by the generating circuit in QGAN. Finally, the sample data obtained by measuring the stego image and the real data are used as the input of the discriminator D . The iterative optimization is performed so that G can obtain a stego image close to the target image. The ex-perimental results show that proposed scheme can generate stego images that fit the target image distribution well and achieve the non-embedded hiding of information.
针对量子图像信息隐藏技术在面对基于统计量的隐写算法检测时存在的不安全性,提出了一种基于量子生成对抗网络(QGAN)的信息隐藏方案。该方案首先利用映射规则将秘密信息映射到单量子比特门,为发生器G的参数化量子电路的输入状态做准备。然后通过QGAN中的生成电路生成隐存量子图像。最后,将测量隐写图像得到的样本数据和真实数据作为鉴别器D的输入。进行迭代优化,使G能得到接近目标图像的隐进图像。实验结果表明,该方法能够生成符合目标图像分布的隐写图像,实现了信息的非嵌入隐藏。
{"title":"Information Hiding Scheme Based on Quantum Generative Adversarial Network","authors":"Jia Luo, Rigui Zhou, Yaochong Li, Guangzhong Liu","doi":"10.3724/sp.j.1089.2021.18617","DOIUrl":"https://doi.org/10.3724/sp.j.1089.2021.18617","url":null,"abstract":": Due to the insecurity of quantum image information hiding technology in the face of statisti-cal-based steganalysis algorithm detection, an information hiding scheme based on quantum generative adversarial network (QGAN) is proposed. This scheme first uses the mapping rules to map the secret information into the single qubit gate to prepare for the input state of the parameterized quantum circuit of the gen-erator G . Then the stego quantum image is generated by the generating circuit in QGAN. Finally, the sample data obtained by measuring the stego image and the real data are used as the input of the discriminator D . The iterative optimization is performed so that G can obtain a stego image close to the target image. The ex-perimental results show that proposed scheme can generate stego images that fit the target image distribution well and achieve the non-embedded hiding of information.","PeriodicalId":52442,"journal":{"name":"计算机辅助设计与图形学学报","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2021-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"46509410","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Automatic Poetry Generation Based on Ancient Chinese Paintings 基于中国古代绘画的诗歌自动生成
Q3 Computer Science Pub Date : 2021-07-01 DOI: 10.3724/sp.j.1089.2021.18633
Jiazhou Chen, Keyu Huang, Yingchaojie Feng, Wei Zhang, Siwei Tan, Wei Chen
: The Chinese painting poem is a very special art form in the history of world art. It combines ancient Chinese literature and fine arts, complements each other and blends together. In order to obtain com-puter-based painting poetry, an automatic poetry generation is proposed based on ancient Chinese paintings. It extracts multiple sentences from ancient paintings, which improves the literary expression ability of ancient poems in paintings. Firstly, a multi-sentence annotation data set for ancient paintings is established, and then semantic features of ancient paintings are extracted through an improved image captioning method. Finally, these modern text descriptions are converted into a four-character poem through a two-way LSTM encoding and decoding framework. The experiment on the paintings of the Song Dynasty demonstrates that the coherent and prosodic poems generated by our method are consistent with the original content and con-text of the ancient paintings. User study shows that the content consistency and user satisfaction of our method are better than keyword-based methods, which proves the validity of the proposed method
中国画诗歌在世界艺术史上是一种非常特殊的艺术形式。它结合了中国古代文学和美术,相得益彰,相互交融。为了获得基于计算机的绘画诗歌,提出了一种基于中国古代绘画的诗歌自动生成方法。从古画中提取多句,提高了古诗在绘画中的文学表达能力。首先建立多句古画标注数据集,然后通过改进的图像字幕方法提取古画的语义特征。最后,通过双向LSTM编码解码框架将这些现代文本描述转化为四字诗。对宋代绘画的实验表明,我们的方法生成的连贯、韵律的诗歌与古代绘画的原始内容和语境是一致的。用户研究表明,该方法的内容一致性和用户满意度均优于基于关键字的方法,证明了该方法的有效性
{"title":"Automatic Poetry Generation Based on Ancient Chinese Paintings","authors":"Jiazhou Chen, Keyu Huang, Yingchaojie Feng, Wei Zhang, Siwei Tan, Wei Chen","doi":"10.3724/sp.j.1089.2021.18633","DOIUrl":"https://doi.org/10.3724/sp.j.1089.2021.18633","url":null,"abstract":": The Chinese painting poem is a very special art form in the history of world art. It combines ancient Chinese literature and fine arts, complements each other and blends together. In order to obtain com-puter-based painting poetry, an automatic poetry generation is proposed based on ancient Chinese paintings. It extracts multiple sentences from ancient paintings, which improves the literary expression ability of ancient poems in paintings. Firstly, a multi-sentence annotation data set for ancient paintings is established, and then semantic features of ancient paintings are extracted through an improved image captioning method. Finally, these modern text descriptions are converted into a four-character poem through a two-way LSTM encoding and decoding framework. The experiment on the paintings of the Song Dynasty demonstrates that the coherent and prosodic poems generated by our method are consistent with the original content and con-text of the ancient paintings. User study shows that the content consistency and user satisfaction of our method are better than keyword-based methods, which proves the validity of the proposed method","PeriodicalId":52442,"journal":{"name":"计算机辅助设计与图形学学报","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2021-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"69686226","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
L0 Optimization Using Laplacian Operator for Image Smoothing 基于拉普拉斯算子的图像平滑L0优化
Q3 Computer Science Pub Date : 2021-07-01 DOI: 10.3724/sp.j.1089.2021.18627
Menghang Li, Shanshan Gao, Huijian Han, Caiming Zhang
: Image smoothing often leads to the loss of image details and distortion because of over smoothing. An image smoothing method is presented which combines 0 L optimization and the second-order Laplacian operator. Laplacian operator is used to constrain the color change of the image, and 0 L optimization is used to minimize the change of the color gradient, so as to achieve the purpose of smooth color transition of the image. In order to keep the edge features of the image better in the process of smoothing, Sobel operator is introduced as the regular term of energy function, and the alternating solution strategy is adopted to solve the energy function. In the ex-periment, using the classical image in the field of image smoothing and the image obtained through network en-gine, the proposed method is compared qualitatively and quantitatively with 6 smoothing methods and 7 denois-第 ing methods. The experimental results show that the proposed method can reduce the loss of image details while smoothing the image, effectively deal with the phenomenon of stepped edges and color block distribution in the image smoothing, and effectively remove various noises in the image. And the peak signal-to-noise ratio and run-ning time of the proposed method are improved compared with other methods.
:由于过度平滑,图像平滑通常会导致图像细节的丢失和失真。提出了一种将0L优化和二阶拉普拉斯算子相结合的图像平滑方法。拉普拉斯算子用于约束图像的颜色变化,0L优化用于最小化颜色梯度的变化,从而达到图像颜色平滑过渡的目的。为了在平滑过程中更好地保持图像的边缘特征,引入Sobel算子作为能量函数的正则项,并采用交替求解策略求解能量函数。在实验中,利用图像平滑领域的经典图像和通过网络工程获得的图像,将所提出的方法与6种平滑方法和7种去噪方法进行了定性和定量的比较-第 ing方法。实验结果表明,该方法在对图像进行平滑处理的同时,可以减少图像细节的损失,有效地处理图像平滑中的阶梯边缘和色块分布现象,有效地去除图像中的各种噪声。与其他方法相比,该方法的峰值信噪比和运行时间都有所提高。
{"title":"L0 Optimization Using Laplacian Operator for Image Smoothing","authors":"Menghang Li, Shanshan Gao, Huijian Han, Caiming Zhang","doi":"10.3724/sp.j.1089.2021.18627","DOIUrl":"https://doi.org/10.3724/sp.j.1089.2021.18627","url":null,"abstract":": Image smoothing often leads to the loss of image details and distortion because of over smoothing. An image smoothing method is presented which combines 0 L optimization and the second-order Laplacian operator. Laplacian operator is used to constrain the color change of the image, and 0 L optimization is used to minimize the change of the color gradient, so as to achieve the purpose of smooth color transition of the image. In order to keep the edge features of the image better in the process of smoothing, Sobel operator is introduced as the regular term of energy function, and the alternating solution strategy is adopted to solve the energy function. In the ex-periment, using the classical image in the field of image smoothing and the image obtained through network en-gine, the proposed method is compared qualitatively and quantitatively with 6 smoothing methods and 7 denois-第 ing methods. The experimental results show that the proposed method can reduce the loss of image details while smoothing the image, effectively deal with the phenomenon of stepped edges and color block distribution in the image smoothing, and effectively remove various noises in the image. And the peak signal-to-noise ratio and run-ning time of the proposed method are improved compared with other methods.","PeriodicalId":52442,"journal":{"name":"计算机辅助设计与图形学学报","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2021-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"42793534","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
A Real-Time Semantic Segmentation Approach for Autonomous Driving Scenes 一种自动驾驶场景的实时语义分割方法
Q3 Computer Science Pub Date : 2021-07-01 DOI: 10.3724/sp.j.1089.2021.18631
Feiwei Qin, Xiyue Shen, Yong Peng, Yanli Shao, Wenqiang Yuan, Zhongping Ji, Jing Bai
An important part of autonomous driving is the perception of the driving environment of the car, which has created a strong demand for high precision semantic segmentation algorithms that can be run in real time on low-power mobile devices. However, when analyzing the factors that affect the accuracy and speed of the semantic segmentation network, it can be found that in the structure of the previous semantic segmentation algorithm, spatial information and context features are difficult to take into account at the same time, and using two networks to obtain spatial information and context information separately will increase the amount of calculation and storage. Therefore, a new structure is proposed that divides the spatial path and context path from the network based on the residual structure, and a two-path real-time semantic segmentation network is designed based on this structure. The network contains a feature fusion module and an attention refinement module, which are used to realize the function of fusing the multi-scale features of two 第 7 期 秦飞巍, 等: 无人驾驶中的场景实时语义分割方法 1027 paths and optimizing the output results of context path. The network is based on the PyTorch framework and uses NVIDIA 1080Ti graphics cards for experiments. On the road scene data set Cityscapes, mIoU reached 78.8%, and the running speed reached 27.5 fps.
自动驾驶的一个重要部分是对汽车驾驶环境的感知,这对能够在低功耗移动设备上实时运行的高精度语义分割算法产生了强烈的需求。然而,在分析影响语义分割网络准确性和速度的因素时,可以发现,在以前的语义分割算法的结构中,空间信息和上下文特征很难同时考虑,使用两个网络分别获得空间信息和上下文信息将增加计算量和存储量。因此,提出了一种基于残差结构将空间路径和上下文路径从网络中分割出来的新结构,并基于该结构设计了一个双路径实时语义分割网络。该网络包含一个特征融合模块和一个注意力细化模块,用于实现融合两个多尺度特征的功能第 7.期 秦飞巍, 等: 无人驾驶中的场景实时语义分割方法 1027条路径,并优化上下文路径的输出结果。该网络基于PyTorch框架,并使用NVIDIA 1080Ti显卡进行实验。在道路场景数据集Cityscapes上,mIoU达到78.8%,运行速度达到27.5帧/秒。
{"title":"A Real-Time Semantic Segmentation Approach for Autonomous Driving Scenes","authors":"Feiwei Qin, Xiyue Shen, Yong Peng, Yanli Shao, Wenqiang Yuan, Zhongping Ji, Jing Bai","doi":"10.3724/sp.j.1089.2021.18631","DOIUrl":"https://doi.org/10.3724/sp.j.1089.2021.18631","url":null,"abstract":"An important part of autonomous driving is the perception of the driving environment of the car, which has created a strong demand for high precision semantic segmentation algorithms that can be run in real time on low-power mobile devices. However, when analyzing the factors that affect the accuracy and speed of the semantic segmentation network, it can be found that in the structure of the previous semantic segmentation algorithm, spatial information and context features are difficult to take into account at the same time, and using two networks to obtain spatial information and context information separately will increase the amount of calculation and storage. Therefore, a new structure is proposed that divides the spatial path and context path from the network based on the residual structure, and a two-path real-time semantic segmentation network is designed based on this structure. The network contains a feature fusion module and an attention refinement module, which are used to realize the function of fusing the multi-scale features of two 第 7 期 秦飞巍, 等: 无人驾驶中的场景实时语义分割方法 1027 paths and optimizing the output results of context path. The network is based on the PyTorch framework and uses NVIDIA 1080Ti graphics cards for experiments. On the road scene data set Cityscapes, mIoU reached 78.8%, and the running speed reached 27.5 fps.","PeriodicalId":52442,"journal":{"name":"计算机辅助设计与图形学学报","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2021-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49617302","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Dongba Painting Few-Shot Classification Based on Graph Neural Network 基于图神经网络的东巴绘画少镜头分类
Q3 Computer Science Pub Date : 2021-07-01 DOI: 10.3724/sp.j.1089.2021.18618
Ke Li, Wenhua Qian, Chengxue Wang, Dan Xu
{"title":"Dongba Painting Few-Shot Classification Based on Graph Neural Network","authors":"Ke Li, Wenhua Qian, Chengxue Wang, Dan Xu","doi":"10.3724/sp.j.1089.2021.18618","DOIUrl":"https://doi.org/10.3724/sp.j.1089.2021.18618","url":null,"abstract":"","PeriodicalId":52442,"journal":{"name":"计算机辅助设计与图形学学报","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2021-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"48234455","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Semi-Real-Time Bearing Fault Diagnosis Method Combined Image Method 半实时轴承故障诊断方法——组合图像法
Q3 Computer Science Pub Date : 2021-06-01 DOI: 10.3724/sp.j.1089.2021.18579
Pengzhi Wang, Mandun Zhang, Yahong Han, Xu Zhao, Zhengjun Wang
{"title":"Semi-Real-Time Bearing Fault Diagnosis Method Combined Image Method","authors":"Pengzhi Wang, Mandun Zhang, Yahong Han, Xu Zhao, Zhengjun Wang","doi":"10.3724/sp.j.1089.2021.18579","DOIUrl":"https://doi.org/10.3724/sp.j.1089.2021.18579","url":null,"abstract":"","PeriodicalId":52442,"journal":{"name":"计算机辅助设计与图形学学报","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2021-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"46889049","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Spatial Positioning Method of Vehicle in Cross-Camera Traffic Scene 跨摄像头交通场景中车辆空间定位方法
Q3 Computer Science Pub Date : 2021-06-01 DOI: 10.3724/sp.j.1089.2021.18612
Wen Wang, Xinyao Tang, Chaoyang Zhang, Huansheng Song, Hua Cui
{"title":"Spatial Positioning Method of Vehicle in Cross-Camera Traffic Scene","authors":"Wen Wang, Xinyao Tang, Chaoyang Zhang, Huansheng Song, Hua Cui","doi":"10.3724/sp.j.1089.2021.18612","DOIUrl":"https://doi.org/10.3724/sp.j.1089.2021.18612","url":null,"abstract":"","PeriodicalId":52442,"journal":{"name":"计算机辅助设计与图形学学报","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2021-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"47016303","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Pix2Pix-Based Grayscale Image Coloring Method 基于Pix2Pix的灰度图像着色方法
Q3 Computer Science Pub Date : 2021-06-01 DOI: 10.3724/sp.j.1089.2021.18596
Hong Li, Qiaoxue Zheng, Jing Zhang, Zhuo-Ming Du, Zhanli Li, Baosheng Kang
: In this study, a grayscale image coloring method combining the Pix2Pix model is proposed to solve the problem of unclear object boundaries and low image coloring quality in colorization neural net-works. First, an improved U-Net structure, using eight down-sampling and up-sampling layers, is adopted to extract features and predict the image color, which improves the network model’s ability to extract deep image features. Second, the coloring image quality is tested under different loss functions, 1 L loss and smooth 1 L loss, to measure the distance between the generated image and ground truth. Finally, gradient penalty is added to improve the network stability of the training process. The gradient of each input data is penalized by constructing a new data distribution between the generated and real image distribution to limit the dis-criminator gradient. In the same experimental environment, the Pix2Pix model and summer2winter data are utilized for comparative analysis. The experiments demonstrate that the improved U-Net using the smooth 1 L loss as generator loss generates better colored images, whereas the 1 L loss better maintains the structural information of the image. Furthermore, the gradient penalty accelerates the model convergence speed, and improves the model stability and image quality. The proposed image coloring method learns deep image features and reduces the image blurs. The model raises the image quality while effectively maintaining the image structure similarity.
本研究提出了一种结合Pix2Pix模型的灰度图像着色方法,以解决着色神经网络中物体边界不清晰和图像着色质量低的问题。首先,采用改进的U-Net结构,采用8个下采样层和8个上采样层进行特征提取和图像颜色预测,提高了网络模型提取深度图像特征的能力;其次,在不同的损失函数,1 L损失和平滑1 L损失下,测试着色图像的质量,测量生成的图像与地面真值之间的距离。最后,加入梯度惩罚,提高训练过程的网络稳定性。通过在生成的图像和真实图像之间构造一个新的数据分布来限制鉴别器梯度,从而对每个输入数据的梯度进行惩罚。在相同的实验环境下,采用Pix2Pix模型和summer2winter数据进行对比分析。实验表明,采用平滑的1 L损耗作为生成损耗的改进U-Net能生成更好的彩色图像,而1 L损耗能更好地保持图像的结构信息。此外,梯度惩罚加快了模型的收敛速度,提高了模型的稳定性和图像质量。该方法学习了图像的深度特征,降低了图像的模糊程度。该模型在有效保持图像结构相似性的同时,提高了图像质量。
{"title":"Pix2Pix-Based Grayscale Image Coloring Method","authors":"Hong Li, Qiaoxue Zheng, Jing Zhang, Zhuo-Ming Du, Zhanli Li, Baosheng Kang","doi":"10.3724/sp.j.1089.2021.18596","DOIUrl":"https://doi.org/10.3724/sp.j.1089.2021.18596","url":null,"abstract":": In this study, a grayscale image coloring method combining the Pix2Pix model is proposed to solve the problem of unclear object boundaries and low image coloring quality in colorization neural net-works. First, an improved U-Net structure, using eight down-sampling and up-sampling layers, is adopted to extract features and predict the image color, which improves the network model’s ability to extract deep image features. Second, the coloring image quality is tested under different loss functions, 1 L loss and smooth 1 L loss, to measure the distance between the generated image and ground truth. Finally, gradient penalty is added to improve the network stability of the training process. The gradient of each input data is penalized by constructing a new data distribution between the generated and real image distribution to limit the dis-criminator gradient. In the same experimental environment, the Pix2Pix model and summer2winter data are utilized for comparative analysis. The experiments demonstrate that the improved U-Net using the smooth 1 L loss as generator loss generates better colored images, whereas the 1 L loss better maintains the structural information of the image. Furthermore, the gradient penalty accelerates the model convergence speed, and improves the model stability and image quality. The proposed image coloring method learns deep image features and reduces the image blurs. The model raises the image quality while effectively maintaining the image structure similarity.","PeriodicalId":52442,"journal":{"name":"计算机辅助设计与图形学学报","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2021-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"47201506","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
期刊
计算机辅助设计与图形学学报
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1