首页 > 最新文献

Signal Processing-Image Communication最新文献

英文 中文
Determination of Lagrange multipliers for interframe EZBC/JP2K 帧间EZBC/JP2K拉格朗日乘子的确定
IF 3.5 3区 工程技术 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2023-10-01 DOI: 10.1016/j.image.2023.117030
Yuan Liu , John W. Woods

Interframe EZBC/JP2K has been shown to be an effective fine-grain scalable video coding system. However, its Lagrange multiplier values for motion estimation of multiple temporal levels are not specified, and must be specified by the user in the config file in order to run the program. In this paper, we investigate how to select these Lagrange parameters for optimized performance. By designing an iterative mechanism, we make it possible for the encoder to adaptively select Lagrange multipliers based on the feedback of Y-PSNR closed GOP performance. Experimental results regarding both classic test video clips and their concatenations are obtained and discussed. We also present a new analytical model for optimized Lagrange multiplier selection in terms of target Y-PSNR.

帧间EZBC/JP2K已被证明是一种有效的细粒度可扩展视频编码系统。然而,它的拉格朗日乘子值用于多个时间水平的运动估计是没有指定的,并且必须由用户在配置文件中指定,以便运行程序。在本文中,我们研究了如何选择这些拉格朗日参数来优化性能。通过设计迭代机制,使编码器能够基于Y-PSNR闭合GOP性能的反馈自适应选择拉格朗日乘法器。给出了经典测试视频片段及其拼接的实验结果,并对其进行了讨论。我们还提出了一种新的基于目标Y-PSNR的拉格朗日乘子优化选择的解析模型。
{"title":"Determination of Lagrange multipliers for interframe EZBC/JP2K","authors":"Yuan Liu ,&nbsp;John W. Woods","doi":"10.1016/j.image.2023.117030","DOIUrl":"https://doi.org/10.1016/j.image.2023.117030","url":null,"abstract":"<div><p><span>Interframe<span> EZBC/JP2K has been shown to be an effective fine-grain scalable video coding system. However, its </span></span>Lagrange multiplier<span><span> values for motion estimation of multiple temporal levels are not specified, and must be specified by the user in the config file in order to run the program. In this paper, we investigate how to select these </span>Lagrange parameters for optimized performance. By designing an iterative mechanism, we make it possible for the encoder to adaptively select Lagrange multipliers based on the feedback of Y-PSNR closed GOP performance. Experimental results regarding both classic test video clips and their concatenations are obtained and discussed. We also present a new analytical model for optimized Lagrange multiplier selection in terms of target Y-PSNR.</span></p></div>","PeriodicalId":49521,"journal":{"name":"Signal Processing-Image Communication","volume":"118 ","pages":"Article 117030"},"PeriodicalIF":3.5,"publicationDate":"2023-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49896212","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Deep steerable pyramid wavelet network for unified JPEG compression artifact reduction 用于统一JPEG压缩伪影减少的深度可操纵金字塔小波网络
IF 3.5 3区 工程技术 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2023-10-01 DOI: 10.1016/j.image.2023.117011
Yi Zhang , Damon M. Chandler , Xuanqin Mou

Although numerous methods have been proposed to remove blocking artifacts in JPEG-compressed images, one important issue not well addressed so far is the construction of a unified model that requires no prior knowledge of the JPEG encoding parameters to operate effectively on different compression-level images (grayscale/color) while occupying relatively small storage space to save and run. To address this issue, in this paper, we present a unified JPEG compression artifact reduction model called DSPW-Net, which employs (1) the deep steerable pyramid wavelet transform network for Y-channel restoration, and (2) the classic U-Net architecture for CbCr-channel restoration. To enable our model to work effectively on images with a wide range of compression levels, the quality factor (QF) related features extracted by the convolutional layers in the QF-estimation network are incorporated in the two restoration branches. Meanwhile, recursive blocks with shared parameters are utilized to drastically reduce model parameters and shared-source residual learning is employed to avoid the gradient vanishing/explosion problem in training. Extensive quantitative and qualitative results tested on various benchmark datasets demonstrate the effectiveness of our model as compared with other state-of-the-art deblocking methods.

尽管已经提出了许多方法来消除JPEG压缩图像中的阻塞伪像,但迄今为止没有很好地解决的一个重要问题是构建一个统一的模型,该模型不需要事先了解JPEG编码参数,就可以在不同压缩级别的图像(灰度/彩色)上有效地操作,同时占用相对较小的存储空间来保存和运行。为了解决这一问题,本文提出了一种统一的JPEG压缩伪迹减少模型DSPW-Net,该模型采用(1)深度可转向金字塔小波变换网络进行y通道恢复,(2)经典的U-Net结构进行cbcr通道恢复。为了使我们的模型能够有效地处理具有广泛压缩级别的图像,在QF估计网络中,将卷积层提取的质量因子(QF)相关特征合并到两个恢复分支中。同时,利用具有共享参数的递归块来大幅减少模型参数,并利用共享源残差学习来避免训练中的梯度消失/爆炸问题。在各种基准数据集上测试的大量定量和定性结果表明,与其他最先进的块化方法相比,我们的模型是有效的。
{"title":"Deep steerable pyramid wavelet network for unified JPEG compression artifact reduction","authors":"Yi Zhang ,&nbsp;Damon M. Chandler ,&nbsp;Xuanqin Mou","doi":"10.1016/j.image.2023.117011","DOIUrl":"https://doi.org/10.1016/j.image.2023.117011","url":null,"abstract":"<div><p><span>Although numerous methods have been proposed to remove blocking artifacts in JPEG-compressed images, one important issue not well addressed so far is the construction of a unified model that requires no prior knowledge of the JPEG encoding parameters to operate effectively on different compression-level images (grayscale/color) while occupying relatively small storage space to save and run. To address this issue, in this paper, we present a unified JPEG compression artifact<span> reduction model called DSPW-Net, which employs (1) the deep steerable pyramid wavelet transform network for Y-channel restoration, and (2) the classic U-Net architecture for CbCr-channel restoration. To enable our model to work effectively on images with a wide range of compression levels, the quality factor (QF) related features extracted by the </span></span>convolutional layers in the QF-estimation network are incorporated in the two restoration branches. Meanwhile, recursive blocks with shared parameters are utilized to drastically reduce model parameters and shared-source residual learning is employed to avoid the gradient vanishing/explosion problem in training. Extensive quantitative and qualitative results tested on various benchmark datasets demonstrate the effectiveness of our model as compared with other state-of-the-art deblocking methods.</p></div>","PeriodicalId":49521,"journal":{"name":"Signal Processing-Image Communication","volume":"118 ","pages":"Article 117011"},"PeriodicalIF":3.5,"publicationDate":"2023-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49896187","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Soccer line mark segmentation and classification with stochastic watershed transform 基于随机分水岭变换的足球标线分割与分类
IF 3.5 3区 工程技术 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2023-10-01 DOI: 10.1016/j.image.2023.117014
Daniel Berjón, Carlos Cuevas, Narciso García

Augmented reality applications are beginning to change the way sports are broadcast, providing richer experiences and valuable insights to fans. The first step of augmented reality systems is camera calibration, possibly based on detecting the line markings of the playing field. Most existing proposals for line detection rely on edge detection and Hough transform, but radial distortion and extraneous edges cause inaccurate or spurious detections of line markings. We propose a novel strategy to automatically and accurately segment and classify line markings. First, line points are segmented thanks to a stochastic watershed transform that is robust to radial distortions, since it makes no assumptions about line straightness, and is unaffected by the presence of players or the ball. The line points are then linked to primitive structures (straight lines and ellipses) thanks to a very efficient procedure that makes no assumptions about the number of primitives that appear in each image. The strategy has been tested on a new and public database composed by 60 annotated images from matches in five stadiums. The results obtained have proven that the proposed strategy is more robust and accurate than existing approaches, achieving successful line mark detection even under challenging conditions.

增强现实应用程序正在开始改变体育节目的播放方式,为球迷提供更丰富的体验和有价值的见解。增强现实系统的第一步是相机校准,可能是基于检测比赛场地的标线。大多数现有的线检测方案都依赖于边缘检测和霍夫变换,但径向失真和无关边缘会导致对线标记的不准确或虚假检测。我们提出了一种新的策略来自动准确地分割和分类标线。首先,由于随机分水岭变换对径向失真具有鲁棒性,因此线点被分割,因为它不对直线度进行假设,并且不受球员或球的存在的影响。然后,线点被链接到基元结构(直线和椭圆),这要归功于一个非常有效的过程,该过程不对每个图像中出现的基元的数量进行假设。该策略已在一个新的公共数据库中进行了测试,该数据库由五个体育场比赛的60张注释图像组成。所获得的结果证明,所提出的策略比现有方法更稳健、更准确,即使在具有挑战性的条件下也能成功地检测线迹。
{"title":"Soccer line mark segmentation and classification with stochastic watershed transform","authors":"Daniel Berjón,&nbsp;Carlos Cuevas,&nbsp;Narciso García","doi":"10.1016/j.image.2023.117014","DOIUrl":"https://doi.org/10.1016/j.image.2023.117014","url":null,"abstract":"<div><p>Augmented reality applications are beginning to change the way sports are broadcast, providing richer experiences and valuable insights to fans. The first step of augmented reality systems is camera calibration, possibly based on detecting the line markings of the playing field. Most existing proposals for line detection rely on edge detection and Hough transform, but radial distortion and extraneous edges cause inaccurate or spurious detections of line markings. We propose a novel strategy to automatically and accurately segment and classify line markings. First, line points are segmented thanks to a stochastic watershed transform that is robust to radial distortions, since it makes no assumptions about line straightness, and is unaffected by the presence of players or the ball. The line points are then linked to primitive structures (straight lines and ellipses) thanks to a very efficient procedure that makes no assumptions about the number of primitives that appear in each image. The strategy has been tested on a new and public database composed by 60 annotated images from matches in five stadiums. The results obtained have proven that the proposed strategy is more robust and accurate than existing approaches, achieving successful line mark detection even under challenging conditions.</p></div>","PeriodicalId":49521,"journal":{"name":"Signal Processing-Image Communication","volume":"118 ","pages":"Article 117014"},"PeriodicalIF":3.5,"publicationDate":"2023-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49845015","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multi-scale deep feature fusion based sparse dictionary selection for video summarization 基于多尺度深度特征融合的视频摘要稀疏字典选择
IF 3.5 3区 工程技术 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2023-10-01 DOI: 10.1016/j.image.2023.117006
Xiao Wu , Mingyang Ma , Shuai Wan , Xiuxiu Han , Shaohui Mei

The explosive growth of video data constitutes a series of new challenges in computer vision, and the function of video summarization (VS) is becoming more and more prominent. Recent works have shown the effectiveness of sparse dictionary selection (SDS) based VS, which selects a representative frame set to sufficiently reconstruct a given video. Existing SDS based VS methods use conventional handcrafted features or single-scale deep features, which could diminish their summarization performance due to the underutilization of frame feature representation. Deep learning techniques based on convolutional neural networks (CNNs) exhibit powerful capabilities among various vision tasks, as the CNN provides excellent feature representation. Therefore, in this paper, a multi-scale deep feature fusion based sparse dictionary selection (MSDFF-SDS) is proposed for VS. Specifically, multi-scale features include the directly extracted features from the last fully connected layer and the global average pooling (GAP) processed features from intermediate layers, then VS is formulated as a problem of minimizing the reconstruction error using the multi-scale deep feature fusion. In our formulation, the contribution of each scale of features can be adjusted by a balance parameter, and the row-sparsity consistency of the simultaneous reconstruction coefficient is used to select as few keyframes as possible. The resulting MSDFF-SDS model is solved by using an efficient greedy pursuit algorithm. Experimental results on two benchmark datasets demonstrate that the proposed MSDFF-SDS improves the F-score of keyframe based summarization more than 3% compared with the existing SDS methods, and performs better than most deep-learning methods for skimming based summarization.

视频数据的爆炸式增长构成了计算机视觉领域的一系列新挑战,视频摘要(VS)的功能越来越突出。最近的工作已经表明了基于稀疏字典选择(SDS)的VS的有效性,该VS选择具有代表性的帧集来充分重建给定的视频。现有的基于SDS的VS方法使用传统的手工特征或单尺度深度特征,由于框架特征表示的利用不足,这可能会降低其摘要性能。基于卷积神经网络(CNNs)的深度学习技术在各种视觉任务中表现出强大的能力,因为CNN提供了出色的特征表示。因此,本文针对VS提出了一种基于多尺度深度特征融合的稀疏字典选择(MSDFF-SDS)。具体而言,多尺度特征包括来自最后一个完全连接层的直接提取特征和来自中间层的全局平均池(GAP)处理特征,则VS被公式化为使用多尺度深度特征融合来最小化重构误差的问题。在我们的公式中,每个尺度的特征的贡献可以通过平衡参数来调整,同时重建系数的行稀疏性一致性用于选择尽可能少的关键帧。通过使用有效的贪婪追求算法来求解由此产生的MSDFF-SDS模型。在两个基准数据集上的实验结果表明,与现有的SDS方法相比,所提出的MSDFF-SDS将基于关键帧的摘要的F分数提高了3%以上,并且在基于略读的摘要中表现优于大多数深度学习方法。
{"title":"Multi-scale deep feature fusion based sparse dictionary selection for video summarization","authors":"Xiao Wu ,&nbsp;Mingyang Ma ,&nbsp;Shuai Wan ,&nbsp;Xiuxiu Han ,&nbsp;Shaohui Mei","doi":"10.1016/j.image.2023.117006","DOIUrl":"https://doi.org/10.1016/j.image.2023.117006","url":null,"abstract":"<div><p>The explosive growth of video data constitutes a series of new challenges in computer vision<span><span>, and the function of video summarization (VS) is becoming more and more prominent. Recent works have shown the effectiveness of sparse dictionary selection (SDS) based VS, which selects a representative frame set to sufficiently reconstruct a given video. Existing SDS based VS methods use conventional handcrafted features or single-scale deep features, which could diminish their summarization performance due to the underutilization of frame feature representation. Deep learning<span> techniques based on convolutional neural networks<span> (CNNs) exhibit powerful capabilities among various vision tasks, as the CNN provides excellent feature representation. Therefore, in this paper, a multi-scale deep feature fusion<span> based sparse dictionary selection (MSDFF-SDS) is proposed for VS. Specifically, multi-scale features include the directly extracted features from the last fully connected layer and the global average pooling (GAP) processed features from intermediate layers, then VS is formulated as a problem of minimizing the reconstruction error using the multi-scale deep feature fusion. In our formulation, the contribution of each scale of features can be adjusted by a balance parameter, and the row-sparsity consistency of the simultaneous reconstruction coefficient is used to select as few </span></span></span></span>keyframes as possible. The resulting MSDFF-SDS model is solved by using an efficient greedy pursuit algorithm. Experimental results on two benchmark datasets demonstrate that the proposed MSDFF-SDS improves the F-score of keyframe based summarization more than 3% compared with the existing SDS methods, and performs better than most deep-learning methods for skimming based summarization.</span></p></div>","PeriodicalId":49521,"journal":{"name":"Signal Processing-Image Communication","volume":"118 ","pages":"Article 117006"},"PeriodicalIF":3.5,"publicationDate":"2023-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49844963","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Surprise-based JND estimation for perceptual quantization in H.265/HEVC codecs 基于惊喜的H.265/HEVC编解码器感知量化JND估计
IF 3.5 3区 工程技术 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2023-10-01 DOI: 10.1016/j.image.2023.117019
Hongkui Wang , Li Yu , Hailang Yang , Haifeng Xu , Haibing Yin , Guangtao Zhai , Tianzong Li , Zhuo Kuang

Just noticeable distortion (JND), reflecting the perceptual redundancy directly, has been widely used in image and video compression. However, the human visual system (HVS) is extremely complex and the visual signal processing has not been fully understood, which result in existing JND models are not accurate enough and the bitrate saving of JND-based perceptual compression schemes is limited. This paper presents a novel pixel-based JND model for videos and a JND-based perceptual quantization scheme for HEVC codecs. In particular, positive and negative perception effects of the inter-frame difference and the motion information are analyzed and measured with an information-theoretic approach. Then, a surprise-based JND model is developed for perceptual video coding (PVC). In our PVC scheme, the frame-level perceptual quantization parameter (QP) is derived on the premise that the coding distortion is infinitely close to the estimated JND threshold. On the basis of the frame-level perceptual QP, we determine the perceptual QP for each coding unit through a perceptual adjustment function to achieve better perceptual quality. Experimental results indicate that the proposed JND model outperforms existing models significantly, the proposed perceptual quantization scheme improves video compression efficiency with better perceptual quality and lower coding complexity.

刚显失真(JND)是一种直接反映感知冗余的方法,在图像和视频压缩中得到了广泛的应用。然而,人类视觉系统极其复杂,对视觉信号的处理还没有完全了解,这导致现有的JND模型不够精确,并且基于JND的感知压缩方案的比特率节省有限。本文提出了一种新的基于像素的视频JND模型和一种基于JND的HEVC编解码器感知量化方案。特别地,用信息论的方法分析和测量了帧间差分和运动信息的积极和消极感知效应。在此基础上,提出了一种基于惊奇度的感知视频编码JND模型。在我们的PVC方案中,帧级感知量化参数(QP)是在编码失真无限接近估计的JND阈值的前提下导出的。在帧级感知QP的基础上,通过感知调节函数确定每个编码单元的感知QP,以获得更好的感知质量。实验结果表明,所提出的JND模型明显优于现有的模型,所提出的感知量化方案以更好的感知质量和更低的编码复杂度提高了视频压缩效率。
{"title":"Surprise-based JND estimation for perceptual quantization in H.265/HEVC codecs","authors":"Hongkui Wang ,&nbsp;Li Yu ,&nbsp;Hailang Yang ,&nbsp;Haifeng Xu ,&nbsp;Haibing Yin ,&nbsp;Guangtao Zhai ,&nbsp;Tianzong Li ,&nbsp;Zhuo Kuang","doi":"10.1016/j.image.2023.117019","DOIUrl":"https://doi.org/10.1016/j.image.2023.117019","url":null,"abstract":"<div><p><span>Just noticeable distortion (JND), reflecting the perceptual redundancy directly, has been widely used in image and video compression. However, the </span>human visual system<span><span> (HVS) is extremely complex and the visual signal processing has not been fully understood, which result in existing JND models are not accurate enough and the bitrate saving of JND-based perceptual compression schemes<span> is limited. This paper presents a novel pixel-based JND model for videos and a JND-based perceptual quantization scheme for HEVC codecs. In particular, positive and negative perception effects of the inter-frame difference and the motion information are analyzed and measured with an information-theoretic approach. Then, a surprise-based JND model is developed for perceptual video coding (PVC). In our PVC scheme, the frame-level perceptual quantization parameter (QP) is derived on the premise that the coding distortion is infinitely close to the estimated JND threshold. On the basis of the frame-level perceptual QP, we determine the perceptual QP for each coding unit through a perceptual adjustment function to achieve better </span></span>perceptual quality. Experimental results indicate that the proposed JND model outperforms existing models significantly, the proposed perceptual quantization scheme improves video compression efficiency with better perceptual quality and lower coding complexity.</span></p></div>","PeriodicalId":49521,"journal":{"name":"Signal Processing-Image Communication","volume":"118 ","pages":"Article 117019"},"PeriodicalIF":3.5,"publicationDate":"2023-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49881551","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Joint adjustment image steganography networks 联合调整图像隐写网络
IF 3.5 3区 工程技术 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2023-10-01 DOI: 10.1016/j.image.2023.117022
Le Zhang , Yao Lu , Tong Li , Guangming Lu

Image steganography aims to achieve covert communication between two partners utilizing stego images generated by hiding secret images within cover images. Existing deep image steganography methods have been rapidly developed in this area. Such methods, however, usually generate the stego images and reveal the secret images using one-process networks, lacking sufficient refinement in these methods. Thus, the security and quality of stego and revealed secret images still have much room for promotion, especially for large-capacity image steganography. This paper proposes Joint Adjustment Image Steganography Networks (JAIS-Nets), containing a series of coarse-to-fine iterative adjustment processes, for image steganography. Our JAIS-Nets first proposes Cross-Process Contrastive Refinement (CPCR) adjustment method, using the cross-process contrastive information from cover-stego and secret-revealed secret image pairs, to iteratively refine the generated stego and revealed secret images, respectively. In addition, our JAIS-Nets further proposes Cross-Process Multi-Scale (CPMS) adjustment method, using the cross-process multi-scale information from different scales cover-stego and secret-revealed secret image pairs, to directly adjust and enhance the intermediate representations of the proposed JAIS-Nets. Integrating the proposed CPCR with CPMS methods, the proposed JAIS-Nets can jointly adjust the quality of the stego and revealed secret images at both the learning process and image scale levels. Extensive experiments demonstrate that our JAIS-Nets can achieve state-of-the-art performances on the security and quality of the stego and revealed secret images on both the regular and large capacity image steganography.

图像隐写术的目的是利用隐藏在封面图像中的秘密图像生成的隐写图像来实现两个合作伙伴之间的秘密通信。现有的深度图像隐写技术在这一领域得到了迅速发展。然而,这些方法通常使用单过程网络生成隐写图像和揭示秘密图像,缺乏足够的细化。因此,隐写和泄露秘密图像的安全性和质量仍有很大的提升空间,特别是大容量图像隐写。本文提出了一种联合平差图像隐写网络(jis - nets),该网络包含一系列从粗到精的迭代平差过程,用于图像隐写。我们的JAIS-Nets首先提出了跨过程对比细化(CPCR)平差方法,利用覆盖-隐进和秘密-揭示的秘密图像对的跨过程对比信息,分别对生成的隐进和揭示的秘密图像进行迭代细化。此外,我们的JAIS-Nets进一步提出了跨过程多尺度(CPMS)平差方法,利用不同尺度覆盖-隐藏和秘密-揭示的秘密图像对的跨过程多尺度信息,直接调整和增强所提出的JAIS-Nets的中间表示。将所提出的CPCR方法与CPMS方法相结合,所提出的JAIS-Nets可以在学习过程和图像尺度水平上共同调整隐入图像和揭示秘密图像的质量。大量的实验表明,我们的JAIS-Nets在常规和大容量图像隐写上都能达到最先进的安全性和隐写质量,并揭示了秘密图像。
{"title":"Joint adjustment image steganography networks","authors":"Le Zhang ,&nbsp;Yao Lu ,&nbsp;Tong Li ,&nbsp;Guangming Lu","doi":"10.1016/j.image.2023.117022","DOIUrl":"https://doi.org/10.1016/j.image.2023.117022","url":null,"abstract":"<div><p>Image steganography aims to achieve covert communication<span><span> between two partners utilizing stego images generated by hiding </span>secret images<span> within cover images. Existing deep image steganography methods have been rapidly developed in this area. Such methods, however, usually generate the stego images and reveal the secret images using one-process networks, lacking sufficient refinement in these methods. Thus, the security and quality of stego and revealed secret images still have much room for promotion, especially for large-capacity image steganography. This paper proposes Joint Adjustment Image Steganography Networks (JAIS-Nets), containing a series of coarse-to-fine iterative adjustment processes, for image steganography. Our JAIS-Nets first proposes Cross-Process Contrastive Refinement (CPCR) adjustment method, using the cross-process contrastive information from cover-stego and secret-revealed secret image pairs, to iteratively refine the generated stego and revealed secret images, respectively. In addition, our JAIS-Nets further proposes Cross-Process Multi-Scale (CPMS) adjustment method, using the cross-process multi-scale information from different scales cover-stego and secret-revealed secret image pairs, to directly adjust and enhance the intermediate representations of the proposed JAIS-Nets. Integrating the proposed CPCR with CPMS methods, the proposed JAIS-Nets can jointly adjust the quality of the stego and revealed secret images at both the learning process and image scale levels. Extensive experiments demonstrate that our JAIS-Nets can achieve state-of-the-art performances on the security and quality of the stego and revealed secret images on both the regular and large capacity image steganography.</span></span></p></div>","PeriodicalId":49521,"journal":{"name":"Signal Processing-Image Communication","volume":"118 ","pages":"Article 117022"},"PeriodicalIF":3.5,"publicationDate":"2023-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49896214","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Self-embedding reversible color-to-grayscale conversion with watermarking feature 具有水印特征的自嵌入可逆色灰度转换
IF 3.5 3区 工程技术 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2023-09-26 DOI: 10.1016/j.image.2023.117061
Felix S.K. Yu, Yuk-Hee Chan, Kenneth K.M. Lam, Daniel P.K. Lun

This paper presents a self-embedding reversible color-to-grayscale conversion (RCGC) algorithm that makes good use of deep learning, vector quantization, and halftoning techniques to achieve its goals. By decoupling the luminance information of a pixel from its chrominance information, it explicitly controls the luminance error of both the conversion outputs and their corresponding reconstructed color images. It can also alleviate the burden of the deep learning network used to restore the embedded chrominance information during the reconstruction of the color image. Luminance-guided chrominance quantization and checkerboard-based halftoning are introduced in the paper to encode the chrominance information to be embedded while reference-guided inverse halftoning is proposed to restore the color image. Simulation results verify that its performance is remarkably superior to conventional state-of-art RCGC algorithms in various measures. In the aspect of authentication, embedding the watermark and chrominance information is realized with context-based pixel-wise encryption and a key-based watermark bit positioning mechanism, which makes us possible to locate tampered regions and prevent unauthorized use of the chrominance information.

本文提出了一种自嵌入可逆色灰度转换(RCGC)算法,该算法充分利用了深度学习、矢量量化和半色调技术来实现其目标。通过将像素的亮度信息与其色度信息去耦,它明确地控制转换输出及其相应的重建彩色图像的亮度误差。它还可以减轻在彩色图像重建过程中用于恢复嵌入色度信息的深度学习网络的负担。本文引入亮度引导的色度量化和基于棋盘格的半色调来编码要嵌入的色度信息,同时提出参考引导的逆半色调来恢复彩色图像。仿真结果表明,该算法在各种指标上都明显优于传统的RCGC算法。在认证方面,水印和色度信息的嵌入是通过基于上下文的逐像素加密和基于密钥的水印比特定位机制来实现的,这使得我们能够定位篡改区域并防止未经授权使用色度信息。
{"title":"Self-embedding reversible color-to-grayscale conversion with watermarking feature","authors":"Felix S.K. Yu,&nbsp;Yuk-Hee Chan,&nbsp;Kenneth K.M. Lam,&nbsp;Daniel P.K. Lun","doi":"10.1016/j.image.2023.117061","DOIUrl":"https://doi.org/10.1016/j.image.2023.117061","url":null,"abstract":"<div><p>This paper presents a self-embedding reversible color-to-grayscale conversion (RCGC) algorithm that makes good use of deep learning, vector quantization, and halftoning techniques to achieve its goals. By decoupling the luminance information of a pixel from its chrominance information, it explicitly controls the luminance error of both the conversion outputs and their corresponding reconstructed color images. It can also alleviate the burden of the deep learning network used to restore the embedded chrominance information during the reconstruction of the color image. Luminance-guided chrominance quantization and checkerboard-based halftoning are introduced in the paper to encode the chrominance information to be embedded while reference-guided inverse halftoning is proposed to restore the color image. Simulation results verify that its performance is remarkably superior to conventional state-of-art RCGC algorithms in various measures. In the aspect of authentication, embedding the watermark and chrominance information is realized with context-based pixel-wise encryption and a key-based watermark bit positioning mechanism, which makes us possible to locate tampered regions and prevent unauthorized use of the chrominance information.</p></div>","PeriodicalId":49521,"journal":{"name":"Signal Processing-Image Communication","volume":"119 ","pages":"Article 117061"},"PeriodicalIF":3.5,"publicationDate":"2023-09-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49838796","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Blind quality-based pairwise ranking of contrast changed color images using deep networks 利用深度网络对对比度发生变化的彩色图像进行基于质量的成对盲排序
IF 3.5 3区 工程技术 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2023-09-23 DOI: 10.1016/j.image.2023.117059
Aladine Chetouani , Muhammad Ali Qureshi , Mohamed Deriche , Azeddine Beghdadi

Next-generation multimedia networks are expected to provide systems and applications with top Quality of Experience (QoE) to users. To this end, robust quality evaluation metrics are critical. Unfortunately, most current research focuses only on modeling and evaluating mainly distortions across the pipeline of multimedia networks. While distortions are important, it is also as important to consider the effects of enhancement and other manipulations of multimedia content, especially images and videos. In contrast to most existing works dedicated to evaluating image/video quality in its traditional context, very few research efforts have been devoted to Image Quality Enhancement Assessment (IQEA) and more specifically, Contrast Enhancement Evaluation (CEE). Our contribution fills this gap by proposing a pairwise ranking scheme for estimating and evaluating the perceptual quality of image contrast change (contrast enhancement and/or contrast-distorted images) process. We propose a novel Deep Learning-based Blind Quality pairwise Ranking scheme for Contrast-Changed (Deep-BQRCC) images. This method provides an automatic pairwise ranking of a set of contrast-changed images. The proposed framework is based on using a pair of Convolutional Neural Networks (CNN) together with a saliency-based attention model and a color-difference visual map. Extensive experiments were conducted to validate the effectiveness of the proposed workflow through an ablation analysis. Different combinations of CNN models and pooling strategies were analyzed. The proposed Deep-BQRCC approach was evaluated over three dedicated publicly available datasets. The experimental results showed an increase in performance within a range of 310% compared to state-of-the-art IQEA measures.

下一代多媒体网络有望为用户提供最高体验质量(QoE)的系统和应用。为此,稳健的质量评估指标至关重要。遗憾的是,目前的大多数研究都只关注对多媒体网络管道中的主要失真进行建模和评估。失真固然重要,但考虑多媒体内容(尤其是图像和视频)的增强和其他处理效果也同样重要。与大多数致力于在传统背景下评估图像/视频质量的现有作品相比,很少有研究致力于图像质量增强评估(IQEA),更具体地说是对比度增强评估(CEE)。我们的研究填补了这一空白,提出了一种成对排序方案,用于估计和评估图像对比度变化(对比度增强和/或对比度失真图像)过程的感知质量。我们提出了一种新颖的基于深度学习的对比度变化图像质量盲对排序方案(Deep-BQRCC)。该方法可对一组对比度变化的图像进行自动配对排序。所提出的框架基于一对卷积神经网络(CNN),以及基于显著性的注意力模型和色差视觉地图。我们进行了广泛的实验,通过消融分析验证了所提工作流程的有效性。对 CNN 模型的不同组合和池化策略进行了分析。在三个专门的公开数据集上对所提出的 Deep-BQRCC 方法进行了评估。实验结果表明,与最先进的 IQEA 方法相比,该方法的性能提高了 3-10%。
{"title":"Blind quality-based pairwise ranking of contrast changed color images using deep networks","authors":"Aladine Chetouani ,&nbsp;Muhammad Ali Qureshi ,&nbsp;Mohamed Deriche ,&nbsp;Azeddine Beghdadi","doi":"10.1016/j.image.2023.117059","DOIUrl":"10.1016/j.image.2023.117059","url":null,"abstract":"<div><p><span><span><span>Next-generation multimedia networks are expected to provide systems and applications with top </span>Quality of Experience<span><span> (QoE) to users. To this end, robust quality evaluation metrics<span> are critical. Unfortunately, most current research focuses only on modeling and evaluating mainly distortions across the pipeline of multimedia networks. While distortions are important, it is also as important to consider the effects of enhancement and other manipulations of multimedia content, especially images and videos. In contrast to most existing works dedicated to evaluating image/video quality in its traditional context, very few research efforts have been devoted to Image Quality Enhancement Assessment (IQEA) and more specifically, Contrast Enhancement Evaluation (CEE). Our contribution fills this gap by proposing a pairwise ranking scheme for estimating and evaluating the </span></span>perceptual quality of image contrast change (contrast enhancement and/or contrast-distorted images) process. We propose a novel Deep Learning-based Blind Quality pairwise Ranking scheme for Contrast-Changed (Deep-BQRCC) images. This method provides an automatic pairwise ranking of a set of contrast-changed images. The proposed framework is based on using a pair of </span></span>Convolutional Neural Networks (CNN) together with a saliency-based attention model and a color-difference visual map. Extensive experiments were conducted to validate the effectiveness of the proposed workflow through an ablation analysis. Different combinations of CNN models and pooling strategies were analyzed. The proposed Deep-BQRCC approach was evaluated over three dedicated publicly available datasets. The experimental results showed an increase in performance within a range of </span><span><math><mrow><mn>3</mn><mtext>–</mtext><mn>10</mn></mrow></math></span>% compared to state-of-the-art IQEA measures.</p></div>","PeriodicalId":49521,"journal":{"name":"Signal Processing-Image Communication","volume":"121 ","pages":"Article 117059"},"PeriodicalIF":3.5,"publicationDate":"2023-09-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135484422","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Co-occurrence spatial–temporal model for adaptive background initialization in high-dynamic complex scenes 用于高动态复杂场景中自适应背景初始化的共现时空模型
IF 3.5 3区 工程技术 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2023-09-20 DOI: 10.1016/j.image.2023.117056
Wenjun Zhou , Yuheng Deng , Bo Peng , Sheng Xiang , Shun’ichi Kaneko

Background information is an important aspect of pre-processing for advanced applications in computer vision. The literature has made rapid progress in background initialization. However, background initialization still suffers from high-dynamic complex scenes, such as illumination change, background motion, or camera jitter. Therefore, this study presents a novel Co-occurrence Spatial–Temporal (CoST) model for background initialization in high-dynamic complex scenes. CoST achieves a spatial–temporal model through a co-occurrence pixel-block structure. The proposed approach extracts the spatial–temporal information of pixels to self-adaptively generate the background without the influence of high-dynamic complex scenes. The efficiency of CoST is verified through experimental results compared with state-of-the-art algorithms. The source code of CoST is available online at: https://github.com/HelloMrDeng/CoST.git.

背景信息是计算机视觉高级应用预处理的一个重要方面。文献在背景初始化方面取得了快速进展。然而,背景初始化仍然受到高动态复杂场景的影响,例如照明变化、背景运动或相机抖动。因此,本研究提出了一种新的共现时空(CoST)模型,用于高动态复杂场景中的背景初始化。CoST通过共现像素块结构实现了时空模型。该方法提取像素的时空信息,在不受高动态复杂场景影响的情况下自适应生成背景。通过与最先进算法的比较实验结果验证了CoST的有效性。CoST的源代码可在线访问:https://github.com/HelloMrDeng/CoST.git.
{"title":"Co-occurrence spatial–temporal model for adaptive background initialization in high-dynamic complex scenes","authors":"Wenjun Zhou ,&nbsp;Yuheng Deng ,&nbsp;Bo Peng ,&nbsp;Sheng Xiang ,&nbsp;Shun’ichi Kaneko","doi":"10.1016/j.image.2023.117056","DOIUrl":"https://doi.org/10.1016/j.image.2023.117056","url":null,"abstract":"<div><p><span>Background information is an important aspect of pre-processing for advanced applications in computer vision<span>. The literature has made rapid progress in background initialization. However, background initialization still suffers from high-dynamic complex scenes, such as illumination change, background motion, or camera jitter. Therefore, this study presents a novel Co-occurrence Spatial–Temporal (CoST) model for background initialization in high-dynamic complex scenes. CoST achieves a spatial–temporal model through a co-occurrence pixel-block structure. The proposed approach extracts the spatial–temporal information of pixels to self-adaptively generate the background without the influence of high-dynamic complex scenes. The efficiency of CoST is verified through experimental results compared with state-of-the-art algorithms. The source code of CoST is available online at: </span></span><span>https://github.com/HelloMrDeng/CoST.git</span><svg><path></path></svg>.</p></div>","PeriodicalId":49521,"journal":{"name":"Signal Processing-Image Communication","volume":"119 ","pages":"Article 117056"},"PeriodicalIF":3.5,"publicationDate":"2023-09-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49838801","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Multi-operator Image Retargeting based on Saliency Object Ranking and Similarity Evaluation Metric 基于显著性对象排序和相似性评价度量的多算子图像重定目标
IF 3.5 3区 工程技术 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2023-09-19 DOI: 10.1016/j.image.2023.117063
Yingchun Guo, Dan Wang, Ye Zhu, Gang Yan

Image Retargeting (IR) technology is proposed to flexibly display images on various display devices while protecting the important content of the images undistorted. IR methods mainly use Salient Object Detection (SOD) to obtain important content, however, most existing SOD methods treat multiple salient objects with the same saliency degrees, which makes IR methods assign the same retargeting ratios for different objects and leads to producing information-loss retargeted results. Multi-operator IR demonstrates better generalization than single operator by using multiple operators to find the optimal sequence of operators. Meanwhile, the tremendous processing time limits its practical use. To address these problems, we propose a multi-operator IR method based on Salient Object Ranking (SOR) and Similarity Evaluation Metric (SORSEM-IR), which includes two stages: importance map generation and multi-operator IR. In the first stage, a SOR module with Context-aware Semantic Refinement (SORCSR) is proposed, which extracts the salient instances and infers their saliency ranks with a context-aware semantic refinement module, then the SOR map, face map, and gradient map are fused as the importance map. In the second stage, to speed up multiple operations, a similarity evaluation metric is proposed to measure the similarity between the original image and the seam-removal image by Seam Carving (SC) operation, and switch SC to uniform scaling to meet the aspect ratio when distortion caused by SC arrives at a certain extent. Experimental results show that the SORCSR network achieves state-of-the-art performance on the ASSR dataset subjectively and objectively, and the SORSEM-IR guided by SORCSR can not only protect the salient objects with minimum deformation but also meet human aesthetic perception.

图像重定位(IR)技术是为了在各种显示设备上灵活地显示图像,同时保护图像的重要内容不失真而提出的。IR方法主要使用显著目标检测(SOD)来获得重要内容,然而,现有的大多数SOD方法都处理具有相同显著度的多个显著目标,这使得IR方法对不同目标分配相同的重定目标比率,导致产生信息丢失重定目标结果。通过使用多个算子来寻找最优算子序列,多算子IR比单算子表现出更好的泛化能力。同时,巨大的处理时间限制了它的实际应用。为了解决这些问题,我们提出了一种基于显著对象排序(SOR)和相似性评估度量(SORSEM-IR)的多算子IR方法,该方法包括两个阶段:重要性图生成和多算子IR,利用上下文感知语义细化模块提取显著实例并推断其显著性等级,然后将SOR图、人脸图和梯度图融合为重要度图。在第二阶段,为了加快多次操作,提出了一种相似性评估度量,通过接缝雕刻(SC)操作来测量原始图像和接缝去除图像之间的相似性,并在SC引起的失真达到一定程度时将SC切换到均匀缩放以满足宽高比。实验结果表明,SORCSR网络在ASSR数据集上主观和客观上都达到了最先进的性能,SORCSR引导的SORSEM-IR不仅能以最小的变形保护显著物体,还能满足人类的审美感知。
{"title":"Multi-operator Image Retargeting based on Saliency Object Ranking and Similarity Evaluation Metric","authors":"Yingchun Guo,&nbsp;Dan Wang,&nbsp;Ye Zhu,&nbsp;Gang Yan","doi":"10.1016/j.image.2023.117063","DOIUrl":"https://doi.org/10.1016/j.image.2023.117063","url":null,"abstract":"<div><p>Image Retargeting (IR) technology is proposed to flexibly display images on various display devices while protecting the important content of the images undistorted. IR methods mainly use Salient Object Detection (SOD) to obtain important content, however, most existing SOD methods treat multiple salient objects with the same saliency degrees, which makes IR methods assign the same retargeting ratios for different objects and leads to producing information-loss retargeted results. Multi-operator IR demonstrates better generalization than single operator by using multiple operators to find the optimal sequence of operators. Meanwhile, the tremendous processing time limits its practical use. To address these problems, we propose a multi-operator IR method based on Salient Object Ranking (SOR) and Similarity Evaluation Metric<span> (SORSEM-IR), which includes two stages: importance map generation and multi-operator IR. In the first stage, a SOR module with Context-aware Semantic Refinement (SORCSR) is proposed, which extracts the salient instances and infers their saliency ranks with a context-aware semantic refinement module, then the SOR map, face map, and gradient map are fused as the importance map. In the second stage, to speed up multiple operations, a similarity evaluation metric is proposed to measure the similarity between the original image and the seam-removal image by Seam Carving (SC) operation, and switch SC to uniform scaling to meet the aspect ratio when distortion caused by SC arrives at a certain extent. Experimental results show that the SORCSR network achieves state-of-the-art performance on the ASSR dataset subjectively and objectively, and the SORSEM-IR guided by SORCSR can not only protect the salient objects with minimum deformation but also meet human aesthetic perception.</span></p></div>","PeriodicalId":49521,"journal":{"name":"Signal Processing-Image Communication","volume":"119 ","pages":"Article 117063"},"PeriodicalIF":3.5,"publicationDate":"2023-09-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49838794","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Signal Processing-Image Communication
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1