首页 > 最新文献

IET Image Process.最新文献

英文 中文
Mask-Guided Image Person Removal with Data Synthesis 基于数据合成的面具引导图像人物去除
Pub Date : 2022-09-29 DOI: 10.2139/ssrn.4234905
Yunliang Jiang, Chenyang Gu, Zhenfeng Xue, Xiongtao Zhang, Yong Liu
As a special case of common object removal, image person removal is playing an increasingly important role in social media and criminal investigation domains. Due to the integrity of person area and the complexity of human posture, person removal has its own dilemmas. In this paper, we propose a novel idea to tackle these problems from the perspective of data synthesis. Concerning the lack of dedicated dataset for image person removal, two dataset production methods are proposed to automatically generate images, masks and ground truths respectively. Then, a learning framework similar to local image degradation is proposed so that the masks can be used to guide the feature extraction process and more texture information can be gathered for final prediction. A coarse-to-fine training strategy is further applied to refine the details. The data synthesis and learning framework combine well with each other. Experimental results verify the effectiveness of our method quantitatively and qualitatively, and the trained network proves to have good generalization ability either on real or synthetic images.
图像人移除作为普通物品移除的一种特例,在社交媒体和刑侦领域发挥着越来越重要的作用。由于人体区域的完整性和人体姿态的复杂性,人体移除有其自身的困境。本文从数据综合的角度提出了一种解决这些问题的新思路。针对图像人物去除缺乏专用数据集的问题,提出了两种数据集生成方法,分别自动生成图像、掩模和ground truth。然后,提出了一种类似于局部图像退化的学习框架,利用掩模来指导特征提取过程,收集更多的纹理信息进行最终预测。进一步采用从粗到精的训练策略来细化细节。数据综合和学习框架可以很好地结合在一起。实验结果从定量和定性上验证了该方法的有效性,训练后的网络无论对真实图像还是合成图像都具有良好的泛化能力。
{"title":"Mask-Guided Image Person Removal with Data Synthesis","authors":"Yunliang Jiang, Chenyang Gu, Zhenfeng Xue, Xiongtao Zhang, Yong Liu","doi":"10.2139/ssrn.4234905","DOIUrl":"https://doi.org/10.2139/ssrn.4234905","url":null,"abstract":"As a special case of common object removal, image person removal is playing an increasingly important role in social media and criminal investigation domains. Due to the integrity of person area and the complexity of human posture, person removal has its own dilemmas. In this paper, we propose a novel idea to tackle these problems from the perspective of data synthesis. Concerning the lack of dedicated dataset for image person removal, two dataset production methods are proposed to automatically generate images, masks and ground truths respectively. Then, a learning framework similar to local image degradation is proposed so that the masks can be used to guide the feature extraction process and more texture information can be gathered for final prediction. A coarse-to-fine training strategy is further applied to refine the details. The data synthesis and learning framework combine well with each other. Experimental results verify the effectiveness of our method quantitatively and qualitatively, and the trained network proves to have good generalization ability either on real or synthetic images.","PeriodicalId":13486,"journal":{"name":"IET Image Process.","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2022-09-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"84157705","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
EDAfuse: A encoder-decoder with atrous spatial pyramid network for infrared and visible image fusion EDAfuse:一种用于红外和可见光图像融合的编码器-解码器
Pub Date : 2022-09-01 DOI: 10.2139/ssrn.3982278
Cairen Nie, Dongming Zhou, Rencan Nie
{"title":"EDAfuse: A encoder-decoder with atrous spatial pyramid network for infrared and visible image fusion","authors":"Cairen Nie, Dongming Zhou, Rencan Nie","doi":"10.2139/ssrn.3982278","DOIUrl":"https://doi.org/10.2139/ssrn.3982278","url":null,"abstract":"","PeriodicalId":13486,"journal":{"name":"IET Image Process.","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2022-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"79538121","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Visible part prediction and temporal calibration for pedestrian detection 行人检测的可见部分预测和时间标定
Pub Date : 2022-08-30 DOI: 10.2139/ssrn.4072475
Peiyu Yang, Weixiang Li, Lu Wang, Lisheng Xu, Qingxu Deng
{"title":"Visible part prediction and temporal calibration for pedestrian detection","authors":"Peiyu Yang, Weixiang Li, Lu Wang, Lisheng Xu, Qingxu Deng","doi":"10.2139/ssrn.4072475","DOIUrl":"https://doi.org/10.2139/ssrn.4072475","url":null,"abstract":"","PeriodicalId":13486,"journal":{"name":"IET Image Process.","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2022-08-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82093741","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
STDC-MA Network for Semantic Segmentation 语义分割的STDC-MA网络
Pub Date : 2022-05-10 DOI: 10.48550/arXiv.2205.04639
Xiaochun Lei, Linjun Lu, Zetao Jiang, Zhaoting Gong, Chang Lu, Jiaming Liang
Semantic segmentation is applied extensively in autonomous driving and intelligent transportation with methods that highly demand spatial and semantic information. Here, an STDC-MA network is proposed to meet these demands. First, the STDC-Seg structure is employed in STDC-MA to ensure a lightweight and efficient structure. Subsequently, the feature alignment module (FAM) is applied to understand the offset between high-level and low-level features, solving the problem of pixel offset related to upsampling on the high-level feature map. Our approach implements the effective fusion between high-level features and low-level features. A hierarchical multiscale attention mechanism is adopted to reveal the relationship among attention regions from two different input sizes of one image. Through this relationship, regions receiving much attention are integrated into the segmentation results, thereby reducing the unfocused regions of the input image and improving the effective utilization of multiscale features. STDC- MA maintains the segmentation speed as an STDC-Seg network while improving the segmentation accuracy of small objects. STDC-MA was verified on the verification set of Cityscapes. The segmentation result of STDC-MA attained 76.81% mIOU with the input of 0.5x scale, 3.61% higher than STDC-Seg.
语义分割在自动驾驶和智能交通中有着广泛的应用,其方法对空间信息和语义信息有很高的要求。为此,我们提出了一个STDC-MA网络来满足这些需求。首先,在STDC-MA中采用了STDC-Seg结构,保证了结构的轻量化和高效。随后,利用特征对齐模块FAM (feature alignment module)理解高阶特征与低阶特征之间的偏移量,解决高阶特征映射上采样相关的像素偏移问题。我们的方法实现了高级特征和低级特征的有效融合。采用分层多尺度注意机制揭示了一幅图像两种不同输入尺寸的注意区域之间的关系。通过这种关系,将关注较多的区域整合到分割结果中,从而减少了输入图像的未聚焦区域,提高了多尺度特征的有效利用。STDC- MA在保持STDC- seg网络分割速度的同时,提高了小目标的分割精度。在cityscape验证集上对STDC-MA进行了验证。输入0.5倍尺度时,STDC-MA的分割结果达到76.81% mIOU,比STDC-Seg高3.61%。
{"title":"STDC-MA Network for Semantic Segmentation","authors":"Xiaochun Lei, Linjun Lu, Zetao Jiang, Zhaoting Gong, Chang Lu, Jiaming Liang","doi":"10.48550/arXiv.2205.04639","DOIUrl":"https://doi.org/10.48550/arXiv.2205.04639","url":null,"abstract":"Semantic segmentation is applied extensively in autonomous driving and intelligent transportation with methods that highly demand spatial and semantic information. Here, an STDC-MA network is proposed to meet these demands. First, the STDC-Seg structure is employed in STDC-MA to ensure a lightweight and efficient structure. Subsequently, the feature alignment module (FAM) is applied to understand the offset between high-level and low-level features, solving the problem of pixel offset related to upsampling on the high-level feature map. Our approach implements the effective fusion between high-level features and low-level features. A hierarchical multiscale attention mechanism is adopted to reveal the relationship among attention regions from two different input sizes of one image. Through this relationship, regions receiving much attention are integrated into the segmentation results, thereby reducing the unfocused regions of the input image and improving the effective utilization of multiscale features. STDC- MA maintains the segmentation speed as an STDC-Seg network while improving the segmentation accuracy of small objects. STDC-MA was verified on the verification set of Cityscapes. The segmentation result of STDC-MA attained 76.81% mIOU with the input of 0.5x scale, 3.61% higher than STDC-Seg.","PeriodicalId":13486,"journal":{"name":"IET Image Process.","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2022-05-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"78055664","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multi-similarity based Hyperrelation Network for few-shot segmentation 基于多相似度的小镜头分割超关系网络
Pub Date : 2022-03-17 DOI: 10.48550/arXiv.2203.09550
Xian Shi, Zhe Cui, Shaobing Zhang, Miao Cheng, L. He, Xianghong Tang
Few-shot semantic segmentation aims at recognizing the object regions of unseen categories with only a few annotated examples as supervision. The key to few-shot segmentation is to establish a robust semantic relationship between the support and query images and to prevent overfitting. In this paper, we propose an effective Multi-similarity Hyperrelation Network (MSHNet) to tackle the few-shot semantic segmentation problem. In MSHNet, we propose a new Generative Prototype Similarity (GPS), which together with cosine similarity can establish a strong semantic relation between the support and query images. The locally generated prototype similarity based on global feature is logically complementary to the global cosine similarity based on local feature, and the relationship between the query image and the supported image can be expressed more comprehensively by using the two similarities simultaneously. In addition, we propose a Symmetric Merging Block (SMB) in MSHNet to efficiently merge multi-layer, multi-shot and multi-similarity hyperrelational features. MSHNet is built on the basis of similarity rather than specific category features, which can achieve more general unity and effectively reduce overfitting. On two benchmark semantic segmentation datasets Pascal-5i and COCO-20i, MSHNet achieves new state-of-the-art performances on 1-shot and 5-shot semantic segmentation tasks.
少量语义分割旨在识别未见类别的对象区域,仅使用少量注释示例作为监督。少镜头分割的关键是在支持图像和查询图像之间建立鲁棒的语义关系,防止过拟合。本文提出了一种有效的多相似度超关系网络(MSHNet)来解决少镜头语义分割问题。在MSHNet中,我们提出了一种新的生成原型相似度(GPS),它与余弦相似度一起可以在支持图像和查询图像之间建立强大的语义关系。基于全局特征的局部生成的原型相似度与基于局部特征的全局余弦相似度在逻辑上是互补的,同时使用这两种相似度可以更全面地表达查询图像与支持图像之间的关系。此外,我们在MSHNet中提出了一种对称合并块(SMB)来有效地合并多层、多镜头和多相似的超关系特征。MSHNet是建立在相似度的基础上,而不是基于特定的类别特征,可以达到更一般的统一性,有效地减少过拟合。在Pascal-5i和COCO-20i两个基准语义分割数据集上,MSHNet在1次和5次语义分割任务上取得了新的最先进的性能。
{"title":"Multi-similarity based Hyperrelation Network for few-shot segmentation","authors":"Xian Shi, Zhe Cui, Shaobing Zhang, Miao Cheng, L. He, Xianghong Tang","doi":"10.48550/arXiv.2203.09550","DOIUrl":"https://doi.org/10.48550/arXiv.2203.09550","url":null,"abstract":"Few-shot semantic segmentation aims at recognizing the object regions of unseen categories with only a few annotated examples as supervision. The key to few-shot segmentation is to establish a robust semantic relationship between the support and query images and to prevent overfitting. In this paper, we propose an effective Multi-similarity Hyperrelation Network (MSHNet) to tackle the few-shot semantic segmentation problem. In MSHNet, we propose a new Generative Prototype Similarity (GPS), which together with cosine similarity can establish a strong semantic relation between the support and query images. The locally generated prototype similarity based on global feature is logically complementary to the global cosine similarity based on local feature, and the relationship between the query image and the supported image can be expressed more comprehensively by using the two similarities simultaneously. In addition, we propose a Symmetric Merging Block (SMB) in MSHNet to efficiently merge multi-layer, multi-shot and multi-similarity hyperrelational features. MSHNet is built on the basis of similarity rather than specific category features, which can achieve more general unity and effectively reduce overfitting. On two benchmark semantic segmentation datasets Pascal-5i and COCO-20i, MSHNet achieves new state-of-the-art performances on 1-shot and 5-shot semantic segmentation tasks.","PeriodicalId":13486,"journal":{"name":"IET Image Process.","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2022-03-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"90647288","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
A Screen-Shooting Resilient Document Image Watermarking Scheme using Deep Neural Network 一种基于深度神经网络的截屏弹性文档图像水印方案
Pub Date : 2022-03-10 DOI: 10.48550/arXiv.2203.05198
Sulong Ge, Zhihua Xia, Yao Tong, Jian Weng, Jia-Nan Liu
With the advent of the screen-reading era, the confidential documents displayed on the screen can be easily captured by a camera without leaving any traces. Thus, this paper proposes a novel screen-shooting resilient watermarking scheme for document image using deep neural network. By applying this scheme, when the watermarked image is displayed on the screen and captured by a camera, the watermark can be still extracted from the captured photographs. Specifically, our scheme is an end-to-end neural network with an encoder to embed watermark and a decoder to extract watermark. During the training process, a distortion layer between encoder and decoder is added to simulate the distortions introduced by screen-shooting process in real scenes, such as camera distortion, shooting distortion, light source distortion. Besides, an embedding strength adjustment strategy is designed to improve the visual quality of the watermarked image with little loss of extraction accuracy. The experimental results show that the scheme has higher robustness and visual quality than other three recent state-of-the-arts. Specially, even if the shooting distances and angles are in extreme, our scheme can also obtain high extraction accuracy.
随着屏幕阅读时代的到来,屏幕上显示的机密文件可以很容易地被摄像头捕捉到,不留任何痕迹。为此,本文提出了一种基于深度神经网络的文档图像截屏弹性水印方案。利用该方案,当水印图像显示在屏幕上并被相机捕获时,仍然可以从捕获的照片中提取水印。具体来说,我们的方案是一个端到端的神经网络,用编码器嵌入水印,用解码器提取水印。在训练过程中,在编码器和解码器之间增加一个失真层,模拟真实场景中截屏过程中产生的畸变,如摄像机畸变、拍摄畸变、光源畸变等。此外,设计了一种嵌入强度调整策略,在不影响提取精度的前提下,提高了水印图像的视觉质量。实验结果表明,该方案具有较好的鲁棒性和视觉质量。特别的是,即使在极端的拍摄距离和角度下,我们的方案也能获得很高的提取精度。
{"title":"A Screen-Shooting Resilient Document Image Watermarking Scheme using Deep Neural Network","authors":"Sulong Ge, Zhihua Xia, Yao Tong, Jian Weng, Jia-Nan Liu","doi":"10.48550/arXiv.2203.05198","DOIUrl":"https://doi.org/10.48550/arXiv.2203.05198","url":null,"abstract":"With the advent of the screen-reading era, the confidential documents displayed on the screen can be easily captured by a camera without leaving any traces. Thus, this paper proposes a novel screen-shooting resilient watermarking scheme for document image using deep neural network. By applying this scheme, when the watermarked image is displayed on the screen and captured by a camera, the watermark can be still extracted from the captured photographs. Specifically, our scheme is an end-to-end neural network with an encoder to embed watermark and a decoder to extract watermark. During the training process, a distortion layer between encoder and decoder is added to simulate the distortions introduced by screen-shooting process in real scenes, such as camera distortion, shooting distortion, light source distortion. Besides, an embedding strength adjustment strategy is designed to improve the visual quality of the watermarked image with little loss of extraction accuracy. The experimental results show that the scheme has higher robustness and visual quality than other three recent state-of-the-arts. Specially, even if the shooting distances and angles are in extreme, our scheme can also obtain high extraction accuracy.","PeriodicalId":13486,"journal":{"name":"IET Image Process.","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2022-03-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"75817191","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Geodesic Gramian Denoising Applied to the Images Contaminated With Noise Sampled From Diverse Probability Distributions 测地线格兰曼去噪在不同概率分布中被噪声污染图像中的应用
Pub Date : 2022-03-04 DOI: 10.48550/arXiv.2203.02600
Yonggi Park, K. Gajamannage, Alexey L. Sadovski
As quotidian use of sophisticated cameras surges, people in modern society are more interested in capturing fine-quality images. However, the quality of the images might be inferior to people's expectations due to the noise contamination in the images. Thus, filtering out the noise while preserving vital image features is an essential requirement. Current existing denoising methods have their own assumptions on the probability distribution in which the contaminated noise is sampled for the method to attain its expected denoising performance. In this paper, we utilize our recent Gramian-based filtering scheme to remove noise sampled from five prominent probability distributions from selected images. This method preserves image smoothness by adopting patches partitioned from the image, rather than pixels, and retains vital image features by performing denoising on the manifold underlying the patch space rather than in the image domain. We validate its denoising performance, using three benchmark computer vision test images applied to two state-of-the-art denoising methods, namely BM3D and K-SVD.
随着精密相机的日常使用激增,现代社会的人们对捕捉高质量图像更感兴趣。然而,由于图像中的噪声污染,图像的质量可能会低于人们的期望。因此,在保留重要图像特征的同时滤除噪声是一个基本要求。现有的去噪方法对污染噪声采样的概率分布有自己的假设,以达到预期的去噪效果。在本文中,我们利用我们最近的基于gramian的滤波方案从选定的图像中去除从五个突出的概率分布中采样的噪声。该方法通过采用从图像中分割的小块而不是像素来保持图像的平滑性,并通过在小块空间下的流形上而不是在图像域上进行去噪来保留重要的图像特征。我们验证了它的去噪性能,使用三个基准计算机视觉测试图像应用于两种最先进的去噪方法,即BM3D和K-SVD。
{"title":"Geodesic Gramian Denoising Applied to the Images Contaminated With Noise Sampled From Diverse Probability Distributions","authors":"Yonggi Park, K. Gajamannage, Alexey L. Sadovski","doi":"10.48550/arXiv.2203.02600","DOIUrl":"https://doi.org/10.48550/arXiv.2203.02600","url":null,"abstract":"As quotidian use of sophisticated cameras surges, people in modern society are more interested in capturing fine-quality images. However, the quality of the images might be inferior to people's expectations due to the noise contamination in the images. Thus, filtering out the noise while preserving vital image features is an essential requirement. Current existing denoising methods have their own assumptions on the probability distribution in which the contaminated noise is sampled for the method to attain its expected denoising performance. In this paper, we utilize our recent Gramian-based filtering scheme to remove noise sampled from five prominent probability distributions from selected images. This method preserves image smoothness by adopting patches partitioned from the image, rather than pixels, and retains vital image features by performing denoising on the manifold underlying the patch space rather than in the image domain. We validate its denoising performance, using three benchmark computer vision test images applied to two state-of-the-art denoising methods, namely BM3D and K-SVD.","PeriodicalId":13486,"journal":{"name":"IET Image Process.","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2022-03-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"80212077","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Spectral recovery-guided hyperspectral super-resolution using transfer learning 使用迁移学习的光谱恢复引导的高光谱超分辨率
Pub Date : 2021-05-20 DOI: 10.1049/IPR2.12253
Shaolei Zhang, Guangyuan Fu, Hongqiao Wang, Yuqing Zhao
Single hyperspectral image (HSI) super-resolution (SR) has attracted researcher’s attention; however, most existing methods directly model the mapping between low- and high-resolution images from an external training dataset, which requires large memory and com-puting resources. Moreover, there are few such available training datasets in real cases, which prevent deep-learning-based methods from further improving performance. Here, a novel single HSI SR method based on transfer learning is proposed. The proposed method is composed of two stages: spectral down-sampled image SR reconstruction based on transfer learning and HSI reconstruction via spectral recovery module. Instead of directly applying the learned knowledge from the colour image domain to HSI SR, the spectral down-sampled image is fed into a spatial SR model to obtain a high-resolution image, which acts as a bridge between the colour image and HSI. The spectral recovery network is used to restore the HSI from the bridge image. In addition, pre-training and collaborative fine-tuning are proposed to promote the performance of SR and spectral recovery. Experiments on two public HSI datasets show that the proposed method achieves promising SR performance with a small paired HSI dataset.
单幅高光谱图像(HSI)的超分辨率(SR)越来越受到研究者的关注;然而,大多数现有方法直接对来自外部训练数据集的低分辨率和高分辨率图像之间的映射进行建模,这需要大量的内存和计算资源。此外,在实际案例中,这种可用的训练数据集很少,这阻碍了基于深度学习的方法进一步提高性能。本文提出了一种基于迁移学习的单HSI SR方法。该方法由两个阶段组成:基于迁移学习的频谱下采样图像SR重建和基于频谱恢复模块的HSI重建。不是直接将从彩色图像域学到的知识应用到HSI SR中,而是将光谱下采样图像馈送到空间SR模型中以获得高分辨率图像,该图像充当彩色图像和HSI之间的桥梁。利用光谱恢复网络对桥图像进行HSI恢复。此外,提出了预训练和协同微调来提高SR和光谱恢复的性能。在两个公开的HSI数据集上的实验表明,该方法在一个小的成对HSI数据集上取得了很好的SR性能。
{"title":"Spectral recovery-guided hyperspectral super-resolution using transfer learning","authors":"Shaolei Zhang, Guangyuan Fu, Hongqiao Wang, Yuqing Zhao","doi":"10.1049/IPR2.12253","DOIUrl":"https://doi.org/10.1049/IPR2.12253","url":null,"abstract":"Single hyperspectral image (HSI) super-resolution (SR) has attracted researcher’s attention; however, most existing methods directly model the mapping between low- and high-resolution images from an external training dataset, which requires large memory and com-puting resources. Moreover, there are few such available training datasets in real cases, which prevent deep-learning-based methods from further improving performance. Here, a novel single HSI SR method based on transfer learning is proposed. The proposed method is composed of two stages: spectral down-sampled image SR reconstruction based on transfer learning and HSI reconstruction via spectral recovery module. Instead of directly applying the learned knowledge from the colour image domain to HSI SR, the spectral down-sampled image is fed into a spatial SR model to obtain a high-resolution image, which acts as a bridge between the colour image and HSI. The spectral recovery network is used to restore the HSI from the bridge image. In addition, pre-training and collaborative fine-tuning are proposed to promote the performance of SR and spectral recovery. Experiments on two public HSI datasets show that the proposed method achieves promising SR performance with a small paired HSI dataset.","PeriodicalId":13486,"journal":{"name":"IET Image Process.","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2021-05-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"74797409","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Texture and exposure awareness based refill for HDRI reconstruction of saturated and occluded areas 基于纹理和曝光感知的HDRI饱和和闭塞区域重建
Pub Date : 2021-05-19 DOI: 10.1049/IPR2.12257
Jianming Zhou, Yipeng Deng, Qin Liu, T. Ikenaga
High-dynamic-range image (HDRI) displays scenes as vivid as the real scenes. HDRI can be reconstructed by fusing a set of bracketed-exposure low-dynamic-range images (LDRI). For the reconstruction, many works succeed in removing the ghost artefacts caused by moving objects. The critical issue is reconstructing the areas which are saturated due to bad exposure and occluded due to motion with no ghost artefacts. To overcome this issue, this paper proposes texture and exposure awareness based refill. The proposed work first locates the saturated and occluded areas existing in input image set, then refills background textures or patches containing rough exposure and colour information into located areas. Proposed work can be integrated with multiple existing ghost removal works to improve the reconstruction result. Experimental results show that proposed work removes the ghost artefacts caused by saturated and occluded areas in subjective evaluation. For the objective evaluation, the proposed work improves the HDR-VDP-2 evaluation result for multiple conventional works by 1.33% on average.
高动态范围图像(HDRI)显示的场景与真实场景一样逼真。HDRI可以通过融合一组括号曝光低动态范围图像(LDRI)来重建。对于重建,许多作品成功地去除了移动物体造成的幽灵文物。关键的问题是重建由于曝光不良而饱和的区域和由于运动而遮挡的区域,没有鬼影。为了克服这个问题,本文提出了基于纹理和曝光感知的填充方法。首先定位输入图像集中存在的饱和和遮挡区域,然后将包含粗糙曝光和颜色信息的背景纹理或斑块重新填充到定位区域。建议的工作可以与多个现有的去鬼工作相结合,以改善重建结果。实验结果表明,该方法能够有效地消除主观评价中由于饱和区域和遮挡区域造成的伪影。在客观评价方面,本文对多个常规工程的HDR-VDP-2评价结果平均提高1.33%。
{"title":"Texture and exposure awareness based refill for HDRI reconstruction of saturated and occluded areas","authors":"Jianming Zhou, Yipeng Deng, Qin Liu, T. Ikenaga","doi":"10.1049/IPR2.12257","DOIUrl":"https://doi.org/10.1049/IPR2.12257","url":null,"abstract":"High-dynamic-range image (HDRI) displays scenes as vivid as the real scenes. HDRI can be reconstructed by fusing a set of bracketed-exposure low-dynamic-range images (LDRI). For the reconstruction, many works succeed in removing the ghost artefacts caused by moving objects. The critical issue is reconstructing the areas which are saturated due to bad exposure and occluded due to motion with no ghost artefacts. To overcome this issue, this paper proposes texture and exposure awareness based refill. The proposed work first locates the saturated and occluded areas existing in input image set, then refills background textures or patches containing rough exposure and colour information into located areas. Proposed work can be integrated with multiple existing ghost removal works to improve the reconstruction result. Experimental results show that proposed work removes the ghost artefacts caused by saturated and occluded areas in subjective evaluation. For the objective evaluation, the proposed work improves the HDR-VDP-2 evaluation result for multiple conventional works by 1.33% on average.","PeriodicalId":13486,"journal":{"name":"IET Image Process.","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2021-05-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"91407639","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An improved algorithm using weighted guided coefficient and union self-adaptive image enhancement for single image haze removal 基于加权引导系数和联合自适应图像增强的单幅图像去雾算法
Pub Date : 2021-05-19 DOI: 10.1049/IPR2.12255
Guangbin Zhou, Lifeng He, Yong Qi, Meimei Yang, Xiao Zhao, Y. Chao
The visibility of outdoor images is usually significantly degraded by haze. Existing dehazing algorithms, such as dark channel prior (DCP) and colour attenuation prior (CAP), have made great progress and are highly effective. However, they all suffer from the problems of dark distortion and detailed information loss. This paper proposes an improved algorithm for single-image haze removal based on dark channel prior with weighted guided coefficient and union self-adaptive image enhancement. First, a weighted guided coefficient method with sampling based on guided image filtering is proposed to refine the transmission map efficiently. Second, the k -means clustering method is adopted to calibrate the original image into bright and non-bright colour areas and form a transmission constraint matrix. The constraint matrix is then marked by connected-component labelling, and small bright regions are eliminated to form an atmospheric light constraint matrix, which can suppress the halo effect and optimize the atmospheric light. Finally, an adaptive linear contrast enhancement algorithm with a union score is proposed to optimize restored images. Experimental results demonstrate that the proposed algorithm can overcome the problems of image distortion and detailed information loss and is more efficient than conventional dehazing algorithms.
户外图像的能见度通常会因雾霾而显著降低。现有的去雾算法,如暗通道先验算法(DCP)和颜色衰减先验算法(CAP),已经取得了很大的进步,并且非常有效。然而,它们都存在着暗失真和细节信息丢失的问题。提出了一种基于暗通道先验加权引导系数和联合自适应图像增强的单幅图像去雾算法。首先,提出了一种基于引导图像滤波的加权引导系数采样方法,有效地细化传输图;其次,采用k均值聚类方法将原始图像标定为明亮和非明亮的颜色区域,形成透射约束矩阵;然后对约束矩阵进行连通分量标记,消除小的明亮区域形成大气光约束矩阵,抑制光晕效应,优化大气光。最后,提出了一种带有联合分数的自适应线性对比度增强算法来优化恢复图像。实验结果表明,该算法克服了图像失真和细节信息丢失的问题,比传统的去雾算法效率更高。
{"title":"An improved algorithm using weighted guided coefficient and union self-adaptive image enhancement for single image haze removal","authors":"Guangbin Zhou, Lifeng He, Yong Qi, Meimei Yang, Xiao Zhao, Y. Chao","doi":"10.1049/IPR2.12255","DOIUrl":"https://doi.org/10.1049/IPR2.12255","url":null,"abstract":"The visibility of outdoor images is usually significantly degraded by haze. Existing dehazing algorithms, such as dark channel prior (DCP) and colour attenuation prior (CAP), have made great progress and are highly effective. However, they all suffer from the problems of dark distortion and detailed information loss. This paper proposes an improved algorithm for single-image haze removal based on dark channel prior with weighted guided coefficient and union self-adaptive image enhancement. First, a weighted guided coefficient method with sampling based on guided image filtering is proposed to refine the transmission map efficiently. Second, the k -means clustering method is adopted to calibrate the original image into bright and non-bright colour areas and form a transmission constraint matrix. The constraint matrix is then marked by connected-component labelling, and small bright regions are eliminated to form an atmospheric light constraint matrix, which can suppress the halo effect and optimize the atmospheric light. Finally, an adaptive linear contrast enhancement algorithm with a union score is proposed to optimize restored images. Experimental results demonstrate that the proposed algorithm can overcome the problems of image distortion and detailed information loss and is more efficient than conventional dehazing algorithms.","PeriodicalId":13486,"journal":{"name":"IET Image Process.","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2021-05-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"87041728","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
期刊
IET Image Process.
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1