首页 > 最新文献

Journal of Electronic Imaging最新文献

英文 中文
Research on image segmentation effect based on denoising preprocessing 基于去噪预处理的图像分割效果研究
IF 1.1 4区 计算机科学 Q4 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-06-01 DOI: 10.1117/1.jei.33.3.033033
Lu Ronghui, Tzong-Jer Chen
Our study investigates the impact of denoising preprocessing on the accuracy of image segmentation. Specifically, images with Gaussian noise were segmented using the fuzzy c-means method (FCM), local binary fitting (LBF), the adaptive active contour model coupling local and global information (EVOL_LCV), and the U-Net semantic segmentation method. These methods were then quantitatively evaluated. Subsequently, various denoising techniques, such as mean, median, Gaussian, bilateral filtering, and feed-forward denoising convolutional neural network (DnCNN), were applied to the original images, and the segmentation was performed using the methods mentioned above, followed by another round of quantitative evaluations. The two quantitative evaluations revealed that the segmentation results were clearly enhanced after denoising. Specifically, the Dice similarity coefficient of the FCM segmentation improved by 4% to 44%, LBF improved by 16%, and EVOL_LCV presented limited changes. Additionally, the U-Net network trained on denoised images attained a segmentation improvement of over 5%. The accuracy of traditional segmentation and semantic segmentation of Gaussian noise images is improved effectively using DnCNN.
我们的研究探讨了去噪预处理对图像分割准确性的影响。具体来说,我们使用模糊 c-means 法(FCM)、局部二元拟合法(LBF)、耦合局部和全局信息的自适应主动轮廓模型(EVOL_LCV)以及 U-Net 语义分割法对带有高斯噪声的图像进行了分割。然后对这些方法进行了定量评估。随后,对原始图像应用了各种去噪技术,如均值、中值、高斯、双边滤波和前馈去噪卷积神经网络(DnCNN),并使用上述方法进行了分割,然后进行了另一轮定量评估。两次定量评估结果显示,去噪后的分割效果明显提高。具体来说,FCM 分割的 Dice 相似性系数提高了 4% 至 44%,LBF 提高了 16%,EVOL_LCV 的变化有限。此外,在去噪图像上训练的 U-Net 网络的分割效果提高了 5%以上。使用 DnCNN 有效提高了高斯噪声图像的传统分割和语义分割的准确性。
{"title":"Research on image segmentation effect based on denoising preprocessing","authors":"Lu Ronghui, Tzong-Jer Chen","doi":"10.1117/1.jei.33.3.033033","DOIUrl":"https://doi.org/10.1117/1.jei.33.3.033033","url":null,"abstract":"Our study investigates the impact of denoising preprocessing on the accuracy of image segmentation. Specifically, images with Gaussian noise were segmented using the fuzzy c-means method (FCM), local binary fitting (LBF), the adaptive active contour model coupling local and global information (EVOL_LCV), and the U-Net semantic segmentation method. These methods were then quantitatively evaluated. Subsequently, various denoising techniques, such as mean, median, Gaussian, bilateral filtering, and feed-forward denoising convolutional neural network (DnCNN), were applied to the original images, and the segmentation was performed using the methods mentioned above, followed by another round of quantitative evaluations. The two quantitative evaluations revealed that the segmentation results were clearly enhanced after denoising. Specifically, the Dice similarity coefficient of the FCM segmentation improved by 4% to 44%, LBF improved by 16%, and EVOL_LCV presented limited changes. Additionally, the U-Net network trained on denoised images attained a segmentation improvement of over 5%. The accuracy of traditional segmentation and semantic segmentation of Gaussian noise images is improved effectively using DnCNN.","PeriodicalId":54843,"journal":{"name":"Journal of Electronic Imaging","volume":null,"pages":null},"PeriodicalIF":1.1,"publicationDate":"2024-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141518296","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Reconstructing images with attention generative adversarial network against adversarial attacks 利用注意力生成式对抗网络重建图像,抵御对抗性攻击
IF 1.1 4区 计算机科学 Q4 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-06-01 DOI: 10.1117/1.jei.33.3.033029
Xiong Shen, Yiqin Lu, Zhe Cheng, Zhongshu Mao, Zhang Yang, Jiancheng Qin
Deep learning is widely used in the field of computer vision, but the emergence of adversarial examples threatens its application. How to effectively detect adversarial examples and correct their labels has become a problem to be solved in this application field. Generative adversarial networks (GANs) can effectively learn the features from images. Based on GAN, this work proposes a defense method called “Reconstructing images with GAN” (RIG). The adversarial examples are generated by attack algorithms reconstructed by the trained generator of RIG to eliminate the perturbations of the adversarial examples, which disturb the models for classification, so that the models can restore their labels when classifying the reconstructed images. In addition, to improve the defensive performance of RIG, the attention mechanism (AM) is introduced to enhance the defense effect of RIG, which is called reconstructing images with attention GAN (RIAG). Experiments show that RIG and RIAG can effectively eliminate the perturbations of the adversarial examples. The results also show that RIAG has a better defensive performance than RIG in eliminating the perturbations of adversarial examples, which indicates that the introduction of AM can effectively improve the defense effect of RIG.
深度学习在计算机视觉领域得到了广泛应用,但对抗性示例的出现对其应用造成了威胁。如何有效地检测对抗示例并纠正其标签成为该应用领域亟待解决的问题。生成式对抗网络(GAN)可以有效地从图像中学习特征。在 GAN 的基础上,本研究提出了一种名为 "用 GAN 重构图像"(RIG)的防御方法。对抗示例由经过 RIG 训练的生成器重建的攻击算法生成,以消除对抗示例对分类模型的扰动,从而使模型在对重建图像进行分类时能够恢复其标签。此外,为了提高 RIG 的防御性能,还引入了注意力机制(AM)来增强 RIG 的防御效果,这就是注意力 GAN(RIAG)。实验表明,RIG 和 RIAG 能有效消除对抗实例的扰动。实验结果还表明,在消除对抗实例的扰动方面,RIAG 的防御性能优于 RIG,这说明引入 AM 可以有效提高 RIG 的防御效果。
{"title":"Reconstructing images with attention generative adversarial network against adversarial attacks","authors":"Xiong Shen, Yiqin Lu, Zhe Cheng, Zhongshu Mao, Zhang Yang, Jiancheng Qin","doi":"10.1117/1.jei.33.3.033029","DOIUrl":"https://doi.org/10.1117/1.jei.33.3.033029","url":null,"abstract":"Deep learning is widely used in the field of computer vision, but the emergence of adversarial examples threatens its application. How to effectively detect adversarial examples and correct their labels has become a problem to be solved in this application field. Generative adversarial networks (GANs) can effectively learn the features from images. Based on GAN, this work proposes a defense method called “Reconstructing images with GAN” (RIG). The adversarial examples are generated by attack algorithms reconstructed by the trained generator of RIG to eliminate the perturbations of the adversarial examples, which disturb the models for classification, so that the models can restore their labels when classifying the reconstructed images. In addition, to improve the defensive performance of RIG, the attention mechanism (AM) is introduced to enhance the defense effect of RIG, which is called reconstructing images with attention GAN (RIAG). Experiments show that RIG and RIAG can effectively eliminate the perturbations of the adversarial examples. The results also show that RIAG has a better defensive performance than RIG in eliminating the perturbations of adversarial examples, which indicates that the introduction of AM can effectively improve the defense effect of RIG.","PeriodicalId":54843,"journal":{"name":"Journal of Electronic Imaging","volume":null,"pages":null},"PeriodicalIF":1.1,"publicationDate":"2024-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141509998","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Special Section Guest Editorial: Quality Control by Artificial Vision VII 特别栏目特约编辑:人工视觉质量控制 VII
IF 1.1 4区 计算机科学 Q4 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-06-01 DOI: 10.1117/1.jei.33.3.031201
Igor Jovančević, Jean-José Orteu
Guest Editors Igor Jovančević and Jean-José Orteu introduce the Special Section on Quality Control by Artificial Vision VII.
特邀编辑 Igor Jovančević 和 Jean-José Orteu 介绍第七期人工视觉质量控制特辑。
{"title":"Special Section Guest Editorial: Quality Control by Artificial Vision VII","authors":"Igor Jovančević, Jean-José Orteu","doi":"10.1117/1.jei.33.3.031201","DOIUrl":"https://doi.org/10.1117/1.jei.33.3.031201","url":null,"abstract":"Guest Editors Igor Jovančević and Jean-José Orteu introduce the Special Section on Quality Control by Artificial Vision VII.","PeriodicalId":54843,"journal":{"name":"Journal of Electronic Imaging","volume":null,"pages":null},"PeriodicalIF":1.1,"publicationDate":"2024-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141509997","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
AFFNet: adversarial feature fusion network for super-resolution image reconstruction in remote sensing images AFFNet:用于遥感图像超分辨率图像重建的对抗特征融合网络
IF 1.1 4区 计算机科学 Q4 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-06-01 DOI: 10.1117/1.jei.33.3.033032
Qian Zhao, Qianxi Yin
As an important source of Earth surface information, remote sensing image has the problems of rough and fuzzy image details and poor perception quality, which affect further analysis and application of geographic information. To address the above problems, we introduce the adversarial feature fusion network with an attention-based mechanism for super-resolution reconstruction of remote sensing images in this paper. First, residual structures are designed in the generator to enhance the deep feature extraction capability of remote sensing images. The residual structure is composed of the depthwise over-parameterized convolution and self-attention mechanism, which work synergistically to extract deep feature information from remote sensing images. Second, coordinate attention feature fusion module is introduced at the feature fusion stage, which can fuse shallow features and deep high-level features. Therefore, it can enhance the attention of the model to different features and better fuse inconsistent semantic features. Finally, we design the pixel-attention upsampling module in the up-sampling stage. It adaptively focuses on the most information-rich regions of a pixel and restores the image details more accurately. We conducted extensive experiments on several remote sensing image datasets, and the results showed that compared with current advanced models, our method can better restore the details in the image and achieve good subjective visual effects, which also verifies the effectiveness and superiority of the algorithm proposed in this paper.
遥感图像作为地球表面信息的重要来源,存在图像细节粗糙模糊、感知质量差等问题,影响了地理信息的进一步分析和应用。针对上述问题,我们在本文中引入了基于注意力机制的对抗特征融合网络,用于遥感图像的超分辨率重建。首先,在生成器中设计了残差结构,以增强遥感图像的深度特征提取能力。残差结构由深度超参数化卷积和自注意机制组成,两者协同工作,提取遥感图像的深度特征信息。其次,在特征融合阶段引入了坐标注意特征融合模块,该模块可以融合浅层特征和深层高层特征。因此,它可以提高模型对不同特征的关注度,更好地融合不一致的语义特征。最后,我们在上采样阶段设计了像素注意力上采样模块。它能自适应地关注像素中信息最丰富的区域,更准确地还原图像细节。我们在多个遥感图像数据集上进行了大量实验,结果表明,与目前先进的模型相比,我们的方法能更好地还原图像细节,达到良好的主观视觉效果,这也验证了本文提出的算法的有效性和优越性。
{"title":"AFFNet: adversarial feature fusion network for super-resolution image reconstruction in remote sensing images","authors":"Qian Zhao, Qianxi Yin","doi":"10.1117/1.jei.33.3.033032","DOIUrl":"https://doi.org/10.1117/1.jei.33.3.033032","url":null,"abstract":"As an important source of Earth surface information, remote sensing image has the problems of rough and fuzzy image details and poor perception quality, which affect further analysis and application of geographic information. To address the above problems, we introduce the adversarial feature fusion network with an attention-based mechanism for super-resolution reconstruction of remote sensing images in this paper. First, residual structures are designed in the generator to enhance the deep feature extraction capability of remote sensing images. The residual structure is composed of the depthwise over-parameterized convolution and self-attention mechanism, which work synergistically to extract deep feature information from remote sensing images. Second, coordinate attention feature fusion module is introduced at the feature fusion stage, which can fuse shallow features and deep high-level features. Therefore, it can enhance the attention of the model to different features and better fuse inconsistent semantic features. Finally, we design the pixel-attention upsampling module in the up-sampling stage. It adaptively focuses on the most information-rich regions of a pixel and restores the image details more accurately. We conducted extensive experiments on several remote sensing image datasets, and the results showed that compared with current advanced models, our method can better restore the details in the image and achieve good subjective visual effects, which also verifies the effectiveness and superiority of the algorithm proposed in this paper.","PeriodicalId":54843,"journal":{"name":"Journal of Electronic Imaging","volume":null,"pages":null},"PeriodicalIF":1.1,"publicationDate":"2024-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141518299","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Lightweight deep and cross residual skip connection separable CNN for plant leaf diseases classification 用于植物叶片病害分类的轻量级深度和交叉残差跳接可分离 CNN
IF 1.1 4区 计算机科学 Q4 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-06-01 DOI: 10.1117/1.jei.33.3.033035
Naresh Vedhamuru, Ramanathan Malmathanraj, Ponnusamy Palanisamy
Crop diseases have an adverse effect on the yield, productivity, and quality of agricultural produce, which threatens the safety and security of the global feature of food supply. Addressing and controlling plant diseases through implementation of timely disease management strategies to reduce their transmission are essential for ensuring minimal crop loss, and addressing the increasing demand for food worldwide as the population continues to increase in a steadfast manner. Crop disease mitigation measures involve preventive monitoring, resulting in early detection and classification of plant diseases for effective agricultural procedure to improve crop yield. Early detection and accurate diagnosis of plant diseases enables farmers to deploy disease management strategies, such interventions are critical for better management contributing to higher crop output by curbing the spread of infection and limiting the extent of damage caused by diseases. We propose and implement a deep and cross residual skip connection separable convolutional neural network (DCRSCSCNN) for identifying and classifying leaf diseases for crops including apple, corn, cucumber, grape, potato, and guava. The significant feature of DCRSCSCNN includes residual skip connection and cross residual skip connection separable convolution block. The usage of residual skip connections assists in fixing the gradient vanishing issue faced by network architecture. The employment of separable convolution decreases the number of parameters, which leads to a model with a reduced size. So far, there has been limited exploration or investigation of leveraging separable convolution within lightweight neural networks. Extensive evaluation of several training and test sets using distinct datasets demonstrate that the proposed DCRSCSCNN outperforms other state-of-the-art approaches. The DCRSCSCNN achieved exceptional classification and identification accuracy rates of 99.89% for apple, 98.72% for corn, 100% for cucumber, 99.78% for grape, 100% for potato, 99.69% for guava1, and 99.08% for guava2 datasets.
农作物病害对农产品的产量、生产率和质量都有不利影响,威胁着全球粮食供应的安全和保障。通过实施及时的病害管理策略来应对和控制植物病害,减少病害传播,对于确保将作物损失降至最低,以及应对全球人口持续增长带来的粮食需求增长至关重要。作物病害缓解措施包括预防性监测,从而及早发现植物病害并对其进行分类,以采取有效的农业措施提高作物产量。对植物病害的早期检测和准确诊断使农民能够部署病害管理策略,这种干预措施对更好地管理至关重要,可通过遏制感染传播和限制病害造成的损害程度来提高作物产量。我们提出并实施了一种深度和交叉残差跳接可分离卷积神经网络(DCRSCSCNN),用于对苹果、玉米、黄瓜、葡萄、马铃薯和番石榴等作物的叶片病害进行识别和分类。DCRSCSCNN 的重要特征包括残余跳转连接和交叉残余跳转连接可分离卷积块。残差跳转连接的使用有助于解决网络架构所面临的梯度消失问题。可分离卷积的使用减少了参数的数量,从而缩小了模型的规模。迄今为止,在轻量级神经网络中利用可分离卷积的探索或研究还很有限。利用不同的数据集对多个训练集和测试集进行的广泛评估表明,所提出的 DCRSCSCNN 优于其他最先进的方法。DCRSCSCNN 在苹果、玉米、黄瓜、葡萄、马铃薯、番石榴 1 和番石榴 2 数据集上的分类和识别准确率分别达到了 99.89%、98.72%、100%、99.78%、100%、99.69% 和 99.08%。
{"title":"Lightweight deep and cross residual skip connection separable CNN for plant leaf diseases classification","authors":"Naresh Vedhamuru, Ramanathan Malmathanraj, Ponnusamy Palanisamy","doi":"10.1117/1.jei.33.3.033035","DOIUrl":"https://doi.org/10.1117/1.jei.33.3.033035","url":null,"abstract":"Crop diseases have an adverse effect on the yield, productivity, and quality of agricultural produce, which threatens the safety and security of the global feature of food supply. Addressing and controlling plant diseases through implementation of timely disease management strategies to reduce their transmission are essential for ensuring minimal crop loss, and addressing the increasing demand for food worldwide as the population continues to increase in a steadfast manner. Crop disease mitigation measures involve preventive monitoring, resulting in early detection and classification of plant diseases for effective agricultural procedure to improve crop yield. Early detection and accurate diagnosis of plant diseases enables farmers to deploy disease management strategies, such interventions are critical for better management contributing to higher crop output by curbing the spread of infection and limiting the extent of damage caused by diseases. We propose and implement a deep and cross residual skip connection separable convolutional neural network (DCRSCSCNN) for identifying and classifying leaf diseases for crops including apple, corn, cucumber, grape, potato, and guava. The significant feature of DCRSCSCNN includes residual skip connection and cross residual skip connection separable convolution block. The usage of residual skip connections assists in fixing the gradient vanishing issue faced by network architecture. The employment of separable convolution decreases the number of parameters, which leads to a model with a reduced size. So far, there has been limited exploration or investigation of leveraging separable convolution within lightweight neural networks. Extensive evaluation of several training and test sets using distinct datasets demonstrate that the proposed DCRSCSCNN outperforms other state-of-the-art approaches. The DCRSCSCNN achieved exceptional classification and identification accuracy rates of 99.89% for apple, 98.72% for corn, 100% for cucumber, 99.78% for grape, 100% for potato, 99.69% for guava1, and 99.08% for guava2 datasets.","PeriodicalId":54843,"journal":{"name":"Journal of Electronic Imaging","volume":null,"pages":null},"PeriodicalIF":1.1,"publicationDate":"2024-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141518298","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Stega4NeRF: cover selection steganography for neural radiance fields Stega4NeRF:神经辐射场的封面选择隐写术
IF 1.1 4区 计算机科学 Q4 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-06-01 DOI: 10.1117/1.jei.33.3.033031
Weina Dong, Jia Liu, Lifeng Chen, Wenquan Sun, Xiaozhong Pan
The implicit neural representation of visual data (such as images, videos, and 3D models) has become a current hotspot in computer vision research. This work proposes a cover selection steganography scheme for neural radiance fields (NeRFs). The message sender first trains an NeRF model selecting any viewpoint in 3D space as the viewpoint key Kv, to generate a unique secret viewpoint image. Subsequently, a message extractor is trained using overfitting to establish a one-to-one mapping between the secret viewpoint image and the secret message. To address the issue of securely transmitting the message extractor in traditional steganography, the message extractor is concealed within a hybrid model performing standard classification tasks. The receiver possesses a shared extractor key Ke, which is used to recover the message extractor from the hybrid model. Then the secret viewpoint image is obtained by NeRF through the viewpoint key Kv, and the secret message is extracted by inputting it into the message extractor. Experimental results demonstrate that the trained message extractor achieves high-speed steganography with a large capacity and attains a 100% message embedding. Additionally, the vast viewpoint key space of NeRF ensures the concealment of the scheme.
视觉数据(如图像、视频和三维模型)的隐式神经表示已成为当前计算机视觉研究的热点。本研究提出了一种针对神经辐射场(NeRF)的封面选择隐写术方案。信息发送者首先训练神经辐射场模型,选择三维空间中的任意视点作为视点密钥 Kv,生成唯一的秘密视点图像。随后,利用过拟合训练信息提取器,在秘密视点图像和秘密信息之间建立一一对应的映射关系。为了解决传统隐写术中安全传输信息提取器的问题,信息提取器被隐藏在一个执行标准分类任务的混合模型中。接收者拥有一个共享提取器密钥 Ke,用来从混合模型中恢复信息提取器。然后,通过视点密钥 Kv,用 NeRF 获取秘密视点图像,并将其输入信息提取器,提取秘密信息。实验结果表明,训练有素的信息提取器实现了大容量高速隐写,信息嵌入率达到 100%。此外,NeRF 广阔的视角密钥空间确保了该方案的隐蔽性。
{"title":"Stega4NeRF: cover selection steganography for neural radiance fields","authors":"Weina Dong, Jia Liu, Lifeng Chen, Wenquan Sun, Xiaozhong Pan","doi":"10.1117/1.jei.33.3.033031","DOIUrl":"https://doi.org/10.1117/1.jei.33.3.033031","url":null,"abstract":"The implicit neural representation of visual data (such as images, videos, and 3D models) has become a current hotspot in computer vision research. This work proposes a cover selection steganography scheme for neural radiance fields (NeRFs). The message sender first trains an NeRF model selecting any viewpoint in 3D space as the viewpoint key Kv, to generate a unique secret viewpoint image. Subsequently, a message extractor is trained using overfitting to establish a one-to-one mapping between the secret viewpoint image and the secret message. To address the issue of securely transmitting the message extractor in traditional steganography, the message extractor is concealed within a hybrid model performing standard classification tasks. The receiver possesses a shared extractor key Ke, which is used to recover the message extractor from the hybrid model. Then the secret viewpoint image is obtained by NeRF through the viewpoint key Kv, and the secret message is extracted by inputting it into the message extractor. Experimental results demonstrate that the trained message extractor achieves high-speed steganography with a large capacity and attains a 100% message embedding. Additionally, the vast viewpoint key space of NeRF ensures the concealment of the scheme.","PeriodicalId":54843,"journal":{"name":"Journal of Electronic Imaging","volume":null,"pages":null},"PeriodicalIF":1.1,"publicationDate":"2024-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141518301","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Face antispoofing method based on single-modal and lightweight network 基于单模态和轻量级网络的人脸防伪方法
IF 1.1 4区 计算机科学 Q4 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-06-01 DOI: 10.1117/1.jei.33.3.033030
Guoxiang Tong, Xinrong Yan
In the field of face antispoofing, researchers are increasingly focusing their efforts on multimodal and feature fusion. While multimodal approaches are more effective than single-modal ones, they often come with a huge number of parameters, require significant computational resources, and pose challenges for execution on mobile devices. To address the real-time problem, we propose a fast and lightweight framework based on ShuffleNet V2. Our approach takes patch-level images as input, enhances unit performance by introducing an attention module, and addresses dataset sample imbalance issues through the focal loss function. The framework effectively tackles the real-time constraints of the model. We evaluate the performance of our model on CASIA-FASD, Replay-Attack, and MSU-MFSD datasets. The results demonstrate that our method outperforms the current state-of-the-art methods in both intratest and intertest scenarios. Furthermore, our network has only 0.84 M parameters and 0.81 GFlops, making it suitable for deployment in mobile and real-time settings. Our work can serve as a valuable reference for researchers seeking to develop single-modal face antispoofing methods suitable for mobile and real-time applications.
在人脸防欺骗领域,研究人员正越来越多地把精力集中在多模态和特征融合上。虽然多模态方法比单模态方法更有效,但它们往往带有大量参数,需要大量计算资源,并给移动设备的执行带来挑战。为了解决实时性问题,我们提出了一种基于 ShuffleNet V2 的快速轻量级框架。我们的方法将斑块级图像作为输入,通过引入注意力模块增强单元性能,并通过焦点损失函数解决数据集样本不平衡问题。该框架有效地解决了模型的实时性限制。我们在 CASIA-FASD、Replay-Attack 和 MSU-MFSD 数据集上评估了模型的性能。结果表明,我们的方法在测试内和测试间的表现都优于目前最先进的方法。此外,我们的网络只有 0.84 M 参数和 0.81 GFlops,适合在移动和实时环境中部署。我们的工作可以为寻求开发适用于移动和实时应用的单模态人脸反欺骗方法的研究人员提供有价值的参考。
{"title":"Face antispoofing method based on single-modal and lightweight network","authors":"Guoxiang Tong, Xinrong Yan","doi":"10.1117/1.jei.33.3.033030","DOIUrl":"https://doi.org/10.1117/1.jei.33.3.033030","url":null,"abstract":"In the field of face antispoofing, researchers are increasingly focusing their efforts on multimodal and feature fusion. While multimodal approaches are more effective than single-modal ones, they often come with a huge number of parameters, require significant computational resources, and pose challenges for execution on mobile devices. To address the real-time problem, we propose a fast and lightweight framework based on ShuffleNet V2. Our approach takes patch-level images as input, enhances unit performance by introducing an attention module, and addresses dataset sample imbalance issues through the focal loss function. The framework effectively tackles the real-time constraints of the model. We evaluate the performance of our model on CASIA-FASD, Replay-Attack, and MSU-MFSD datasets. The results demonstrate that our method outperforms the current state-of-the-art methods in both intratest and intertest scenarios. Furthermore, our network has only 0.84 M parameters and 0.81 GFlops, making it suitable for deployment in mobile and real-time settings. Our work can serve as a valuable reference for researchers seeking to develop single-modal face antispoofing methods suitable for mobile and real-time applications.","PeriodicalId":54843,"journal":{"name":"Journal of Electronic Imaging","volume":null,"pages":null},"PeriodicalIF":1.1,"publicationDate":"2024-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141518300","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Low-light image enhancement using negative feedback pulse coupled neural network 利用负反馈脉冲耦合神经网络增强弱光图像效果
IF 1.1 4区 计算机科学 Q4 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-06-01 DOI: 10.1117/1.jei.33.3.033037
Ping Gao, Guidong Zhang, Lingling Chen, Xiaoyun Chen
Low-light image enhancement, fundamentally an ill-posed problem, seeks to simultaneously provide superior visual effects and preserve the natural appearance. Current methodologies often exhibit limitations in contrast enhancement, noise reduction, and the mitigation of halo artifacts. Negative feedback pulse coupled neural network (NFPCNN) is proposed to provide a well posed solution based on uniform distribution in contrast enhancement. The negative feedback dynamically adjusts the attenuation amplitude of neuron threshold based on recent neuronal ignited state. Neurons in the concentrated brightness area arrange smaller attenuation amplitude to enhance the local contrast, whereas neurons in the sparse area set larger attenuation amplitude. NFPCNN makes up for the negligence of pulse coupled neural network in the brightness distribution of the input image. Consistent with Weber–Fechner law, gamma correction is employed to adjust the output of NFPCNN. Although contrast enhancement can improve detail expressiveness, it might also introduce artifacts or aggravate noise. To mitigate these issues, the bilateral filter is employed to suppress halo artifacts. Brightness is used as coefficient to refine the Relativity-of-Gaussian noise suppression method. Experimental results show that the proposed method can effectively suppress noise while enhancing image contrast.
低照度图像增强从根本上说是一个难题,需要同时提供卓越的视觉效果和保持自然的外观。目前的方法通常在对比度增强、降噪和减少光晕伪影方面表现出局限性。负反馈脉冲耦合神经网络(NFPCNN)的提出,为对比度增强提供了一种基于均匀分布的合理解决方案。负反馈会根据神经元最近的点燃状态动态调整神经元阈值的衰减幅度。亮度集中区域的神经元设置较小的衰减幅度以增强局部对比度,而稀疏区域的神经元则设置较大的衰减幅度。NFPCNN 弥补了脉冲耦合神经网络对输入图像亮度分布的疏忽。根据韦伯-费希纳定律,NFPCNN 的输出调整采用伽玛校正。虽然对比度增强可以提高细节表现力,但也可能会引入伪影或加重噪声。为了缓解这些问题,我们采用了双边滤波器来抑制光晕伪影。亮度作为系数被用来完善高斯相对噪声抑制方法。实验结果表明,所提出的方法能有效抑制噪声,同时增强图像对比度。
{"title":"Low-light image enhancement using negative feedback pulse coupled neural network","authors":"Ping Gao, Guidong Zhang, Lingling Chen, Xiaoyun Chen","doi":"10.1117/1.jei.33.3.033037","DOIUrl":"https://doi.org/10.1117/1.jei.33.3.033037","url":null,"abstract":"Low-light image enhancement, fundamentally an ill-posed problem, seeks to simultaneously provide superior visual effects and preserve the natural appearance. Current methodologies often exhibit limitations in contrast enhancement, noise reduction, and the mitigation of halo artifacts. Negative feedback pulse coupled neural network (NFPCNN) is proposed to provide a well posed solution based on uniform distribution in contrast enhancement. The negative feedback dynamically adjusts the attenuation amplitude of neuron threshold based on recent neuronal ignited state. Neurons in the concentrated brightness area arrange smaller attenuation amplitude to enhance the local contrast, whereas neurons in the sparse area set larger attenuation amplitude. NFPCNN makes up for the negligence of pulse coupled neural network in the brightness distribution of the input image. Consistent with Weber–Fechner law, gamma correction is employed to adjust the output of NFPCNN. Although contrast enhancement can improve detail expressiveness, it might also introduce artifacts or aggravate noise. To mitigate these issues, the bilateral filter is employed to suppress halo artifacts. Brightness is used as coefficient to refine the Relativity-of-Gaussian noise suppression method. Experimental results show that the proposed method can effectively suppress noise while enhancing image contrast.","PeriodicalId":54843,"journal":{"name":"Journal of Electronic Imaging","volume":null,"pages":null},"PeriodicalIF":1.1,"publicationDate":"2024-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141509899","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Super-resolution reconstruction of images based on residual dual-path interactive fusion combined with attention 基于残差双路径交互融合与注意力相结合的图像超分辨率重建技术
IF 1.1 4区 计算机科学 Q4 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-06-01 DOI: 10.1117/1.jei.33.3.033034
Wang Hao, Peng Taile, Zhou Ying
In recent years, deep learning has made significant progress in the field of single-image super-resolution (SISR) reconstruction, which has greatly improved reconstruction quality. However, most of the SISR networks focus too much on increasing the depth of the network in the process of feature extraction and neglect the connections between different levels of features as well as the full use of low-frequency feature information. To address this problem, this work proposes a network based on residual dual-path interactive fusion combined with attention (RDIFCA). Using the dual interactive fusion strategy, the network achieves the effective fusion and multiplexing of high- and low-frequency information while increasing the depth of the network, which significantly enhances the expressive ability of the network. The experimental results show that the proposed RDIFCA network exhibits certain superiority in terms of objective evaluation indexes and visual effects on the Set5, Set14, BSD100, Urban100, and Manga109 test sets.
近年来,深度学习在单图像超分辨率(SISR)重建领域取得了重大进展,极大地提高了重建质量。然而,大多数 SISR 网络在特征提取过程中过于注重增加网络的深度,而忽视了不同层次特征之间的联系以及低频特征信息的充分利用。针对这一问题,本研究提出了一种基于残差双路径交互融合结合注意力(RDIFCA)的网络。利用双交互融合策略,该网络在增加网络深度的同时,实现了高频和低频信息的有效融合和复用,显著增强了网络的表达能力。实验结果表明,所提出的 RDIFCA 网络在 Set5、Set14、BSD100、Urban100 和 Manga109 测试集上的客观评价指标和视觉效果方面都表现出了一定的优越性。
{"title":"Super-resolution reconstruction of images based on residual dual-path interactive fusion combined with attention","authors":"Wang Hao, Peng Taile, Zhou Ying","doi":"10.1117/1.jei.33.3.033034","DOIUrl":"https://doi.org/10.1117/1.jei.33.3.033034","url":null,"abstract":"In recent years, deep learning has made significant progress in the field of single-image super-resolution (SISR) reconstruction, which has greatly improved reconstruction quality. However, most of the SISR networks focus too much on increasing the depth of the network in the process of feature extraction and neglect the connections between different levels of features as well as the full use of low-frequency feature information. To address this problem, this work proposes a network based on residual dual-path interactive fusion combined with attention (RDIFCA). Using the dual interactive fusion strategy, the network achieves the effective fusion and multiplexing of high- and low-frequency information while increasing the depth of the network, which significantly enhances the expressive ability of the network. The experimental results show that the proposed RDIFCA network exhibits certain superiority in terms of objective evaluation indexes and visual effects on the Set5, Set14, BSD100, Urban100, and Manga109 test sets.","PeriodicalId":54843,"journal":{"name":"Journal of Electronic Imaging","volume":null,"pages":null},"PeriodicalIF":1.1,"publicationDate":"2024-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141518297","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Adaptive sparse attention module based on reciprocal nearest neighbors 基于互惠近邻的自适应稀疏注意力模块
IF 1.1 4区 计算机科学 Q4 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-06-01 DOI: 10.1117/1.jei.33.3.033038
Zhonggui Sun, Can Zhang, Mingzhu Zhang
The attention mechanism has become a crucial technique in deep feature representation for computer vision tasks. Using a similarity matrix, it enhances the current feature point with global context from the feature map of the network. However, the indiscriminate utilization of all information can easily introduce some irrelevant contents, inevitably hampering performance. In response to this challenge, sparsing, a common information filtering strategy, has been applied in many related studies. Regrettably, their filtering processes often lack reliability and adaptability. To address this issue, we first define an adaptive-reciprocal nearest neighbors (A-RNN) relationship. In identifying neighbors, it gains flexibility through learning adaptive thresholds. In addition, by introducing a reciprocity mechanism, the reliability of neighbors is ensured. Then, we use A-RNN to rectify the similarity matrix in the conventional attention module. In the specific implementation, to distinctly consider non-local and local information, we introduce two blocks: the non-local sparse constraint block and the local sparse constraint block. The former utilizes A-RNN to sparsify non-local information, whereas the latter uses adaptive thresholds to sparsify local information. As a result, an adaptive sparse attention (ASA) module is achieved, inheriting the advantages of flexibility and reliability from A-RNN. In the validation for the proposed ASA module, we use it to replace the attention module in NLNet and conduct experiments on semantic segmentation benchmarks including Cityscapes, ADE20K and PASCAL VOC 2012. With the same backbone (ResNet101), our ASA module outperforms the conventional attention module and its some state-of-the-art variants.
注意力机制已成为计算机视觉任务中深度特征表示的关键技术。它利用相似性矩阵,通过网络特征图中的全局上下文来增强当前特征点。然而,不加区分地利用所有信息很容易引入一些不相关的内容,从而不可避免地影响性能。为了应对这一挑战,稀疏化作为一种常见的信息过滤策略,已在许多相关研究中得到应用。遗憾的是,它们的过滤过程往往缺乏可靠性和适应性。为了解决这个问题,我们首先定义了一种自适应互惠近邻(A-RNN)关系。在识别邻居时,它通过学习自适应阈值获得灵活性。此外,通过引入互惠机制,确保了邻居的可靠性。然后,我们利用 A-RNN 修正传统注意力模块中的相似性矩阵。在具体实现中,为了区别考虑非本地信息和本地信息,我们引入了两个区块:非本地稀疏约束区块和本地稀疏约束区块。前者利用 A-RNN 来稀疏非本地信息,而后者则利用自适应阈值来稀疏本地信息。因此,自适应稀疏注意(ASA)模块继承了 A-RNN 的灵活性和可靠性优势。在验证所提出的 ASA 模块时,我们用它取代了 NLNet 中的注意力模块,并在包括 Cityscapes、ADE20K 和 PASCAL VOC 2012 在内的语义分割基准上进行了实验。在相同的骨干网(ResNet101)上,我们的 ASA 模块优于传统的注意力模块及其一些最先进的变体。
{"title":"Adaptive sparse attention module based on reciprocal nearest neighbors","authors":"Zhonggui Sun, Can Zhang, Mingzhu Zhang","doi":"10.1117/1.jei.33.3.033038","DOIUrl":"https://doi.org/10.1117/1.jei.33.3.033038","url":null,"abstract":"The attention mechanism has become a crucial technique in deep feature representation for computer vision tasks. Using a similarity matrix, it enhances the current feature point with global context from the feature map of the network. However, the indiscriminate utilization of all information can easily introduce some irrelevant contents, inevitably hampering performance. In response to this challenge, sparsing, a common information filtering strategy, has been applied in many related studies. Regrettably, their filtering processes often lack reliability and adaptability. To address this issue, we first define an adaptive-reciprocal nearest neighbors (A-RNN) relationship. In identifying neighbors, it gains flexibility through learning adaptive thresholds. In addition, by introducing a reciprocity mechanism, the reliability of neighbors is ensured. Then, we use A-RNN to rectify the similarity matrix in the conventional attention module. In the specific implementation, to distinctly consider non-local and local information, we introduce two blocks: the non-local sparse constraint block and the local sparse constraint block. The former utilizes A-RNN to sparsify non-local information, whereas the latter uses adaptive thresholds to sparsify local information. As a result, an adaptive sparse attention (ASA) module is achieved, inheriting the advantages of flexibility and reliability from A-RNN. In the validation for the proposed ASA module, we use it to replace the attention module in NLNet and conduct experiments on semantic segmentation benchmarks including Cityscapes, ADE20K and PASCAL VOC 2012. With the same backbone (ResNet101), our ASA module outperforms the conventional attention module and its some state-of-the-art variants.","PeriodicalId":54843,"journal":{"name":"Journal of Electronic Imaging","volume":null,"pages":null},"PeriodicalIF":1.1,"publicationDate":"2024-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141509996","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Journal of Electronic Imaging
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1