首页 > 最新文献

Journal of Electronic Imaging最新文献

英文 中文
Yarn hairiness measurement based on multi-camera system and perspective maximization model 基于多摄像头系统和透视最大化模型的纱线毛羽测量
IF 1.1 4区 计算机科学 Q4 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-08-01 DOI: 10.1117/1.jei.33.4.043043
Hongyan Cao, Zhenze Chen, Haihua Hu, Xiangbing Huai, Hao Zhu, Zhongjian Li
Accurate measurement and identification of the number and length of yarn hairiness is crucial for spinning process optimization and product quality control. However, the existing methods have problems, such as low detection accuracy and efficiency, and incomplete detection. In order to overcome the above defects, an image acquisition device based on a multi-camera system is established to accurately obtain multiple perspectives of hairiness images. An automatic threshold segmentation method based on the local bimodal is proposed based on image difference, convolution kernel enhancement, and histogram equalization. Then, the clear and unbroken yarn hairiness segmentation images are obtained according to the hairiness edge extraction method. Finally, a perspective maximization model is proposed to realize the calculation of the hairiness H value and the number of hairiness in interval. Six kinds of cotton ring-spun yarn with different linear densities are tested using the proposed method, YG133B/M instrument, manual method, and single perspective method. The results show that the proposed multi-camera method can realize the index measurement of the yarn hairiness.
精确测量和识别纱线毛羽的数量和长度对于纺纱工艺优化和产品质量控制至关重要。然而,现有方法存在检测精度和效率低、检测不全面等问题。为了克服上述缺陷,建立了一种基于多摄像头系统的图像采集装置,以精确获取多角度的毛羽图像。在图像差分、卷积核增强和直方图均衡化的基础上,提出了一种基于局部双模态的自动阈值分割方法。然后,根据毛羽边缘提取方法,得到清晰、完整的纱线毛羽分割图像。最后,提出了透视最大化模型,实现了毛羽 H 值和区间毛羽数的计算。使用提出的方法、YG133B/M 仪器、手动方法和单透视法测试了六种不同线性密度的棉环锭纺纱线。结果表明,所提出的多摄像头方法可以实现纱线毛羽的指标测量。
{"title":"Yarn hairiness measurement based on multi-camera system and perspective maximization model","authors":"Hongyan Cao, Zhenze Chen, Haihua Hu, Xiangbing Huai, Hao Zhu, Zhongjian Li","doi":"10.1117/1.jei.33.4.043043","DOIUrl":"https://doi.org/10.1117/1.jei.33.4.043043","url":null,"abstract":"Accurate measurement and identification of the number and length of yarn hairiness is crucial for spinning process optimization and product quality control. However, the existing methods have problems, such as low detection accuracy and efficiency, and incomplete detection. In order to overcome the above defects, an image acquisition device based on a multi-camera system is established to accurately obtain multiple perspectives of hairiness images. An automatic threshold segmentation method based on the local bimodal is proposed based on image difference, convolution kernel enhancement, and histogram equalization. Then, the clear and unbroken yarn hairiness segmentation images are obtained according to the hairiness edge extraction method. Finally, a perspective maximization model is proposed to realize the calculation of the hairiness H value and the number of hairiness in interval. Six kinds of cotton ring-spun yarn with different linear densities are tested using the proposed method, YG133B/M instrument, manual method, and single perspective method. The results show that the proposed multi-camera method can realize the index measurement of the yarn hairiness.","PeriodicalId":54843,"journal":{"name":"Journal of Electronic Imaging","volume":null,"pages":null},"PeriodicalIF":1.1,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142202464","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Video frame interpolation based on depthwise over-parameterized recurrent residual convolution 基于深度过参数化递归残差卷积的视频帧插值技术
IF 1.1 4区 计算机科学 Q4 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-08-01 DOI: 10.1117/1.jei.33.4.043036
Xiaohui Yang, Weijing Liu, Shaowen Wang
To effectively address the challenges of large motions, complex backgrounds and large occlusions in videos, we introduce an end-to-end method for video frame interpolation based on recurrent residual convolution and depthwise over-parameterized convolution in this paper. Specifically, we devise a U-Net architecture utilizing recurrent residual convolution to enhance the quality of interpolated frame. First, the recurrent residual U-Net feature extractor is employed to extract features from input frames, yielding the kernel for each pixel. Subsequently, an adaptive collaboration of flows is utilized to warp the input frames, which are then fed into the frame synthesis network to generate initial interpolated frames. Finally, the proposed network incorporates depthwise over-parameterized convolution to further enhance the quality of interpolated frame. Experimental results on various datasets demonstrate the superiority of our method over state-of-the-art techniques in both objective and subjective evaluations.
为了有效应对视频中的大运动、复杂背景和大遮挡等挑战,我们在本文中介绍了一种基于递归残差卷积和深度过参数化卷积的端到端视频帧插值方法。具体来说,我们设计了一种利用递归残差卷积的 U-Net 架构,以提高插值帧的质量。首先,利用递归残差 U-Net 特征提取器从输入帧中提取特征,为每个像素生成内核。随后,利用自适应协作流对输入帧进行翘曲,然后将其输入帧合成网络,生成初始插值帧。最后,建议的网络结合了深度过参数化卷积,以进一步提高插值帧的质量。在各种数据集上的实验结果表明,无论是客观评价还是主观评价,我们的方法都优于最先进的技术。
{"title":"Video frame interpolation based on depthwise over-parameterized recurrent residual convolution","authors":"Xiaohui Yang, Weijing Liu, Shaowen Wang","doi":"10.1117/1.jei.33.4.043036","DOIUrl":"https://doi.org/10.1117/1.jei.33.4.043036","url":null,"abstract":"To effectively address the challenges of large motions, complex backgrounds and large occlusions in videos, we introduce an end-to-end method for video frame interpolation based on recurrent residual convolution and depthwise over-parameterized convolution in this paper. Specifically, we devise a U-Net architecture utilizing recurrent residual convolution to enhance the quality of interpolated frame. First, the recurrent residual U-Net feature extractor is employed to extract features from input frames, yielding the kernel for each pixel. Subsequently, an adaptive collaboration of flows is utilized to warp the input frames, which are then fed into the frame synthesis network to generate initial interpolated frames. Finally, the proposed network incorporates depthwise over-parameterized convolution to further enhance the quality of interpolated frame. Experimental results on various datasets demonstrate the superiority of our method over state-of-the-art techniques in both objective and subjective evaluations.","PeriodicalId":54843,"journal":{"name":"Journal of Electronic Imaging","volume":null,"pages":null},"PeriodicalIF":1.1,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141933510","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Additive cosine margin for unsupervised softmax embedding 无监督软最大嵌入的加余弦余量
IF 1.1 4区 计算机科学 Q4 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-08-01 DOI: 10.1117/1.jei.33.4.040501
Dan Wang, Jianwei Yang, Cailing Wang
Unsupervised embedding learning aims to learn highly discriminative features of images without using class labels. Existing instance-wise softmax embedding methods treat each instance as a distinct class and explore the underlying instance-to-instance visual similarity relationships. However, overfitting the instance features leads to insufficient discriminability and poor generalizability of networks. To tackle this issue, we introduce an instance-wise softmax embedding with cosine margin (SEwCM), which for the first time adds margin in the unsupervised instance softmax classification function from the cosine perspective. The cosine margin is used to separate the classification decision boundaries between instances. SEwCM explicitly optimizes the feature mapping of networks by maximizing the cosine similarity between instances, thus learning a highly discriminative model. Exhaustive experiments on three fine-grained image datasets demonstrate the effectiveness of our proposed method over existing methods.
无监督嵌入学习的目的是在不使用类别标签的情况下,学习图像的高区分度特征。现有的实例明智软最大嵌入方法将每个实例视为一个不同的类别,并探索实例与实例之间潜在的视觉相似性关系。然而,过度拟合实例特征会导致网络的可区分性不足和泛化能力差。为了解决这个问题,我们引入了带余弦余量的实例软最大嵌入(SEwCM),首次从余弦角度在无监督实例软最大分类函数中增加了余量。余弦余量用于区分实例之间的分类决策边界。SEwCM 通过最大化实例之间的余弦相似度,明确优化了网络的特征映射,从而学习出一个高辨别度的模型。在三个细粒度图像数据集上进行的详尽实验证明了我们提出的方法比现有方法更有效。
{"title":"Additive cosine margin for unsupervised softmax embedding","authors":"Dan Wang, Jianwei Yang, Cailing Wang","doi":"10.1117/1.jei.33.4.040501","DOIUrl":"https://doi.org/10.1117/1.jei.33.4.040501","url":null,"abstract":"Unsupervised embedding learning aims to learn highly discriminative features of images without using class labels. Existing instance-wise softmax embedding methods treat each instance as a distinct class and explore the underlying instance-to-instance visual similarity relationships. However, overfitting the instance features leads to insufficient discriminability and poor generalizability of networks. To tackle this issue, we introduce an instance-wise softmax embedding with cosine margin (SEwCM), which for the first time adds margin in the unsupervised instance softmax classification function from the cosine perspective. The cosine margin is used to separate the classification decision boundaries between instances. SEwCM explicitly optimizes the feature mapping of networks by maximizing the cosine similarity between instances, thus learning a highly discriminative model. Exhaustive experiments on three fine-grained image datasets demonstrate the effectiveness of our proposed method over existing methods.","PeriodicalId":54843,"journal":{"name":"Journal of Electronic Imaging","volume":null,"pages":null},"PeriodicalIF":1.1,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142202439","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multi-scale adaptive low-light image enhancement based on deep learning 基于深度学习的多尺度自适应弱光图像增强技术
IF 1.1 4区 计算机科学 Q4 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-08-01 DOI: 10.1117/1.jei.33.4.043033
Taotao Cao, Taile Peng, Hao Wang, Xiaotong Zhu, Jia Guo, Zhen Zhang
Existing low-light image enhancement (LLIE) technologies have difficulty balancing image quality and computational efficiency. In addition, they amplify the noise and artifacts of the original image when enhancing deep dark images. Therefore, this study proposes a multi-scale adaptive low-light image enhancement method based on deep learning. Specifically, feature extraction and noise reduction modules are designed. First, a more effective low-light enhancement effect is achieved by extracting the details of the dark area of an image. Depth extraction of the details of dark areas is realized through the design of a residual attention mechanism and nonlocal neural network in the UNet model to obtain a visual-attention map of the dark area. Second, the designed noise network obtains the real noise map of the low-light image. Subsequently, the enhanced network uses the dark area visual-attention and noise maps in conjunction with the original low-light image as inputs to adaptively realize LLIE. The LLIE results using the proposed network achieve excellent performance in terms of color, tone, contrast, and detail. Finally, quantitative and visual experiments on multiple test benchmark datasets demonstrate that the proposed method is superior to current state-of-the-art methods in terms of dark area details, image quality enhancement, and image noise reduction. The results of this study can help to address the real world challenges of low-light image quality, such as low contrast, poor visibility, and high noise levels.
现有的低照度图像增强(LLIE)技术难以在图像质量和计算效率之间取得平衡。此外,在增强深暗图像时,它们会放大原始图像的噪声和伪影。因此,本研究提出了一种基于深度学习的多尺度自适应低照度图像增强方法。具体来说,设计了特征提取和降噪模块。首先,通过提取图像暗部细节,实现更有效的弱光增强效果。暗部细节的深度提取是通过在 UNet 模型中设计残留注意力机制和非局部神经网络来实现的,从而获得暗部的视觉注意力图谱。其次,设计的噪声网络可获得弱光图像的真实噪声图。随后,增强型网络将暗区视觉注意力图和噪声图与原始弱光图像一起作为输入,自适应地实现 LLIE。使用所提出的网络实现的 LLIE 结果在色彩、色调、对比度和细节方面都表现出色。最后,在多个测试基准数据集上进行的定量和视觉实验表明,所提出的方法在暗区细节、图像质量增强和图像降噪方面都优于目前最先进的方法。这项研究的结果有助于解决低对比度、可视性差和高噪声水平等低照度图像质量方面的现实挑战。
{"title":"Multi-scale adaptive low-light image enhancement based on deep learning","authors":"Taotao Cao, Taile Peng, Hao Wang, Xiaotong Zhu, Jia Guo, Zhen Zhang","doi":"10.1117/1.jei.33.4.043033","DOIUrl":"https://doi.org/10.1117/1.jei.33.4.043033","url":null,"abstract":"Existing low-light image enhancement (LLIE) technologies have difficulty balancing image quality and computational efficiency. In addition, they amplify the noise and artifacts of the original image when enhancing deep dark images. Therefore, this study proposes a multi-scale adaptive low-light image enhancement method based on deep learning. Specifically, feature extraction and noise reduction modules are designed. First, a more effective low-light enhancement effect is achieved by extracting the details of the dark area of an image. Depth extraction of the details of dark areas is realized through the design of a residual attention mechanism and nonlocal neural network in the UNet model to obtain a visual-attention map of the dark area. Second, the designed noise network obtains the real noise map of the low-light image. Subsequently, the enhanced network uses the dark area visual-attention and noise maps in conjunction with the original low-light image as inputs to adaptively realize LLIE. The LLIE results using the proposed network achieve excellent performance in terms of color, tone, contrast, and detail. Finally, quantitative and visual experiments on multiple test benchmark datasets demonstrate that the proposed method is superior to current state-of-the-art methods in terms of dark area details, image quality enhancement, and image noise reduction. The results of this study can help to address the real world challenges of low-light image quality, such as low contrast, poor visibility, and high noise levels.","PeriodicalId":54843,"journal":{"name":"Journal of Electronic Imaging","volume":null,"pages":null},"PeriodicalIF":1.1,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141885486","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Alternative evaluation of industrial surface defect synthesis data based on analytic hierarchy process 基于层次分析法的工业表面缺陷合成数据替代评估
IF 1.1 4区 计算机科学 Q4 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-08-01 DOI: 10.1117/1.jei.33.4.043055
Yang Lu, Hang Hao, Linhui Chen, Longfei Yang, Xiaoheng Jiang
Deep learning based defect detection methods require a large amount of high-quality defect data. However, the defect samples obtained in practical production are relatively expensive and lack diversity. The data generation method based on generative adversarial networks (GANs) can address the issue of insufficient defect samples at a lower cost. However, the training process of data generation algorithms based on GANs may be affected by various factors, making it challenging to ensure the stability of the quality of the synthesized defect data. Since high-quality defect data determine the performance and representation of the detection model, it is necessary to conduct alternative evaluations on the synthesized defect data. We comprehensively consider the evaluation indicators that affect the generated defect data and propose an alternative evaluation method for comprehensive data on surface defects of industrial products. First, an evaluation index system is constructed based on the attributes of defect data. Then, a substitution evaluation model for surface defect data is built using a multi-level quantitative analytic hierarchy process. Finally, to verify the effectiveness of the evaluation model, we use three advanced defect detection networks and validate the effectiveness of the evaluation model through comparative experiments. We provide an effective solution for screening high-quality defect data generation and improve the performance of downstream task defect detection models.
基于深度学习的缺陷检测方法需要大量高质量的缺陷数据。然而,实际生产中获得的缺陷样本相对昂贵且缺乏多样性。基于生成式对抗网络(GAN)的数据生成方法能以较低的成本解决缺陷样本不足的问题。然而,基于生成式对抗网络的数据生成算法在训练过程中可能会受到各种因素的影响,因此要确保合成缺陷数据质量的稳定性具有挑战性。由于高质量的缺陷数据决定了检测模型的性能和代表性,因此有必要对合成的缺陷数据进行替代评估。我们综合考虑了影响生成缺陷数据的评价指标,提出了工业产品表面缺陷综合数据的替代评价方法。首先,根据缺陷数据的属性构建评价指标体系。然后,利用多层次定量分析层次过程建立了表面缺陷数据的替代评价模型。最后,为了验证评价模型的有效性,我们使用了三种先进的缺陷检测网络,并通过对比实验验证了评价模型的有效性。我们为筛选高质量缺陷数据生成提供了有效的解决方案,并提高了下游任务缺陷检测模型的性能。
{"title":"Alternative evaluation of industrial surface defect synthesis data based on analytic hierarchy process","authors":"Yang Lu, Hang Hao, Linhui Chen, Longfei Yang, Xiaoheng Jiang","doi":"10.1117/1.jei.33.4.043055","DOIUrl":"https://doi.org/10.1117/1.jei.33.4.043055","url":null,"abstract":"Deep learning based defect detection methods require a large amount of high-quality defect data. However, the defect samples obtained in practical production are relatively expensive and lack diversity. The data generation method based on generative adversarial networks (GANs) can address the issue of insufficient defect samples at a lower cost. However, the training process of data generation algorithms based on GANs may be affected by various factors, making it challenging to ensure the stability of the quality of the synthesized defect data. Since high-quality defect data determine the performance and representation of the detection model, it is necessary to conduct alternative evaluations on the synthesized defect data. We comprehensively consider the evaluation indicators that affect the generated defect data and propose an alternative evaluation method for comprehensive data on surface defects of industrial products. First, an evaluation index system is constructed based on the attributes of defect data. Then, a substitution evaluation model for surface defect data is built using a multi-level quantitative analytic hierarchy process. Finally, to verify the effectiveness of the evaluation model, we use three advanced defect detection networks and validate the effectiveness of the evaluation model through comparative experiments. We provide an effective solution for screening high-quality defect data generation and improve the performance of downstream task defect detection models.","PeriodicalId":54843,"journal":{"name":"Journal of Electronic Imaging","volume":null,"pages":null},"PeriodicalIF":1.1,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142202446","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Enhancing hyperspectral image classification with graph attention neural network 利用图注意神经网络加强高光谱图像分类
IF 1.1 4区 计算机科学 Q4 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-08-01 DOI: 10.1117/1.jei.33.4.043052
Niruban Rathakrishnan, Deepa Raja
Due to the rapid advancement of hyperspectral remote sensing technology, classification methods based on hyperspectral images (HSIs) have gained increasing significance in the processes of target identification, mineral mapping, and environmental management. This importance arises from the fact that HSIs offer a more comprehensive understanding of a target’s composition. However, addressing the challenges posed by the high dimensionality and redundancy of HSI sets, coupled with potential class imbalances in hyperspectral datasets, remains a complex task. Both convolutional neural networks (CNNs) and graph convolutional networks (GCNs) have demonstrated promising results in HSI classification in recent years. Nonetheless, CNNs struggle to attain high accuracy with limited sample sizes, whereas GCNs demand substantial computational resources. Oversmoothing remains a persistent challenge with conventional GCNs. In response to these issues, an approach known as the graph attention neural network for remote target classification (GATN-RTC) has been proposed. GATN-RTC employs a spectral filter and an autoregressive moving average filter to classify distant targets, addressing datasets both with and without labeled samples. To evaluate the performance of GATN-RTC, we conducted a comparative analysis against state-of-the-art methodologies using key performance metrics, such as overall accuracy (OA), per-class accuracy, and the Cohen’s Kappa statistic (KC). The findings reveal that GATN-RTC outperforms existing approaches, achieving improvements of 5.95% in OA, 5.33% in per-class accuracy, and 8.28% in the Cohen’s KC for the Salinas dataset. Furthermore, it demonstrates enhancements of 6.05% and 6.4% in OA, 6.56% and 5.89% in per-class accuracy, and 6.71% and 6.23% in the Cohen’s KC for the Pavia University dataset.
由于高光谱遥感技术的飞速发展,基于高光谱图像(HSI)的分类方法在目标识别、矿物测绘和环境管理过程中的重要性与日俱增。之所以如此重要,是因为高光谱图像能够更全面地了解目标的构成。然而,解决高光谱数据集的高维度和冗余性以及潜在的类别不平衡所带来的挑战仍然是一项复杂的任务。近年来,卷积神经网络(CNN)和图卷积网络(GCN)在 HSI 分类方面都取得了可喜的成果。然而,CNN 在样本量有限的情况下难以达到高准确度,而 GCN 则需要大量的计算资源。过平滑仍然是传统 GCN 面临的长期挑战。针对这些问题,有人提出了一种被称为远程目标分类图注意力神经网络(GATN-RTC)的方法。GATN-RTC 采用频谱滤波器和自回归移动平均滤波器对远距离目标进行分类,可处理有标签样本和无标签样本的数据集。为了评估 GATN-RTC 的性能,我们使用关键性能指标(如总体准确率 (OA)、每类准确率和 Cohen's Kappa 统计量 (KC))与最先进的方法进行了比较分析。研究结果表明,GATN-RTC 的性能优于现有方法,在萨利纳斯数据集上,其 OA 提高了 5.95%,每类准确率提高了 5.33%,Cohen's KC 提高了 8.28%。此外,在帕维亚大学数据集上,它的 OA 分别提高了 6.05% 和 6.4%,每类准确率分别提高了 6.56% 和 5.89%,Cohen's KC 分别提高了 6.71% 和 6.23%。
{"title":"Enhancing hyperspectral image classification with graph attention neural network","authors":"Niruban Rathakrishnan, Deepa Raja","doi":"10.1117/1.jei.33.4.043052","DOIUrl":"https://doi.org/10.1117/1.jei.33.4.043052","url":null,"abstract":"Due to the rapid advancement of hyperspectral remote sensing technology, classification methods based on hyperspectral images (HSIs) have gained increasing significance in the processes of target identification, mineral mapping, and environmental management. This importance arises from the fact that HSIs offer a more comprehensive understanding of a target’s composition. However, addressing the challenges posed by the high dimensionality and redundancy of HSI sets, coupled with potential class imbalances in hyperspectral datasets, remains a complex task. Both convolutional neural networks (CNNs) and graph convolutional networks (GCNs) have demonstrated promising results in HSI classification in recent years. Nonetheless, CNNs struggle to attain high accuracy with limited sample sizes, whereas GCNs demand substantial computational resources. Oversmoothing remains a persistent challenge with conventional GCNs. In response to these issues, an approach known as the graph attention neural network for remote target classification (GATN-RTC) has been proposed. GATN-RTC employs a spectral filter and an autoregressive moving average filter to classify distant targets, addressing datasets both with and without labeled samples. To evaluate the performance of GATN-RTC, we conducted a comparative analysis against state-of-the-art methodologies using key performance metrics, such as overall accuracy (OA), per-class accuracy, and the Cohen’s Kappa statistic (KC). The findings reveal that GATN-RTC outperforms existing approaches, achieving improvements of 5.95% in OA, 5.33% in per-class accuracy, and 8.28% in the Cohen’s KC for the Salinas dataset. Furthermore, it demonstrates enhancements of 6.05% and 6.4% in OA, 6.56% and 5.89% in per-class accuracy, and 6.71% and 6.23% in the Cohen’s KC for the Pavia University dataset.","PeriodicalId":54843,"journal":{"name":"Journal of Electronic Imaging","volume":null,"pages":null},"PeriodicalIF":1.1,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142202441","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
D2Net: discriminative feature extraction and details preservation network for salient object detection D2Net:用于突出物体检测的判别特征提取和细节保存网络
IF 1.1 4区 计算机科学 Q4 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-08-01 DOI: 10.1117/1.jei.33.4.043047
Qianqian Guo, Yanjiao Shi, Jin Zhang, Jinyu Yang, Qing Zhang
Convolutional neural networks (CNNs) with a powerful feature extraction ability have raised the performance of salient object detection (SOD) to a unique level, and how to effectively decode the rich features from CNN is the key to improving the performance of the SOD model. Some previous works ignored the differences between the high-level and low-level features and neglected the information loss during feature processing, making them fail in some challenging scenes. To solve this problem, we propose a discriminative feature extraction and details preservation network (D2Net) for SOD. According to the different characteristics of high-level and low-level features, we design a residual optimization module for filtering complex background noise in shallow features and a pyramid feature extraction module to eliminate the information loss caused by atrous convolution in high-level features. Furthermore, we design a features aggregation module to aggregate the elaborately processed high-level and low-level features, which fully considers the performance of different level features and preserves the delicate boundary of salient object. The comparisons with 17 existing state-of-the-art SOD methods on five popular datasets demonstrate the superiority of the proposed D2Net, and the effectiveness of each proposed module is verified through numerous ablation experiments.
具有强大特征提取能力的卷积神经网络(CNN)将突出物体检测(SOD)的性能提升到了一个独特的高度,而如何有效地解码 CNN 中的丰富特征是提高 SOD 模型性能的关键。以往的一些研究忽略了高层特征和低层特征之间的差异,也忽视了特征处理过程中的信息损失,导致在一些具有挑战性的场景中无法成功。为了解决这一问题,我们提出了一种用于 SOD 的判别特征提取和细节保存网络(D2Net)。根据高层和低层特征的不同特点,我们设计了一个残差优化模块来过滤浅层特征中的复杂背景噪声,并设计了一个金字塔特征提取模块来消除高层特征中的无规则卷积所造成的信息损失。此外,我们还设计了一个特征聚合模块,将精心处理过的高层和低层特征进行聚合,既充分考虑了不同层次特征的性能,又保留了突出对象的精细边界。在五个流行数据集上与现有的 17 种最先进的 SOD 方法进行的比较证明了所提出的 D2Net 的优越性,并且通过大量的消融实验验证了所提出的每个模块的有效性。
{"title":"D2Net: discriminative feature extraction and details preservation network for salient object detection","authors":"Qianqian Guo, Yanjiao Shi, Jin Zhang, Jinyu Yang, Qing Zhang","doi":"10.1117/1.jei.33.4.043047","DOIUrl":"https://doi.org/10.1117/1.jei.33.4.043047","url":null,"abstract":"Convolutional neural networks (CNNs) with a powerful feature extraction ability have raised the performance of salient object detection (SOD) to a unique level, and how to effectively decode the rich features from CNN is the key to improving the performance of the SOD model. Some previous works ignored the differences between the high-level and low-level features and neglected the information loss during feature processing, making them fail in some challenging scenes. To solve this problem, we propose a discriminative feature extraction and details preservation network (D2Net) for SOD. According to the different characteristics of high-level and low-level features, we design a residual optimization module for filtering complex background noise in shallow features and a pyramid feature extraction module to eliminate the information loss caused by atrous convolution in high-level features. Furthermore, we design a features aggregation module to aggregate the elaborately processed high-level and low-level features, which fully considers the performance of different level features and preserves the delicate boundary of salient object. The comparisons with 17 existing state-of-the-art SOD methods on five popular datasets demonstrate the superiority of the proposed D2Net, and the effectiveness of each proposed module is verified through numerous ablation experiments.","PeriodicalId":54843,"journal":{"name":"Journal of Electronic Imaging","volume":null,"pages":null},"PeriodicalIF":1.1,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142202445","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Edge-oriented unrolling network for infrared and visible image fusion 用于红外和可见光图像融合的面向边缘的展开网络
IF 1.1 4区 计算机科学 Q4 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-08-01 DOI: 10.1117/1.jei.33.4.043051
Tianhui Yuan, Zongliang Gan, Changhong Chen, Ziguan Cui
Under unfavorable conditions, fusion images of infrared and visible images often lack edge contrast and details. To address this issue, we propose an edge-oriented unrolling network, which comprises a feature extraction network and a feature fusion network. In our approach, after respective enhancement processes, the original infrared/visible image pair with their enhancement version is combined as the input to get more prior information acquisition. First, the feature extraction network consists of four independent iterative edge-oriented unrolling feature extraction networks based on the edge-oriented deep unrolling residual module (EURM), in which the convolutions in the EURM modules are replaced with edge-oriented convolution blocks to enhance the edge features. Then, the convolutional feature fusion network with differential structure is proposed to obtain the final fusion result, through utilizing the concatenate operation to map multidimensional features. In addition, the loss function in the fusion network is optimized to balance multiple features with significant differences in order to achieve better visual effect. Experimental results on multiple datasets demonstrate that the proposed method produces competitive fusion images as evaluated subjectively and objectively, with balanced luminance, sharper edge, and better details.
在不利条件下,红外图像和可见光图像的融合图像往往缺乏边缘对比度和细节。针对这一问题,我们提出了一种面向边缘的展开网络,它由特征提取网络和特征融合网络组成。在我们的方法中,经过各自的增强处理后,原始红外/可见光图像对与它们的增强版本相结合作为输入,以获取更多的先验信息。首先,特征提取网络由四个独立的面向边缘的迭代开卷特征提取网络组成,这些网络基于面向边缘的深度开卷残差模块(EURM),其中,EURM 模块中的卷积被替换为面向边缘的卷积块,以增强边缘特征。然后,通过利用连接操作映射多维特征,提出了具有差分结构的卷积特征融合网络,以获得最终的融合结果。此外,还优化了融合网络中的损失函数,以平衡具有显著差异的多个特征,从而获得更好的视觉效果。在多个数据集上的实验结果表明,经过主观和客观评估,所提出的方法能生成具有竞争力的融合图像,图像亮度均衡,边缘更清晰,细节更完美。
{"title":"Edge-oriented unrolling network for infrared and visible image fusion","authors":"Tianhui Yuan, Zongliang Gan, Changhong Chen, Ziguan Cui","doi":"10.1117/1.jei.33.4.043051","DOIUrl":"https://doi.org/10.1117/1.jei.33.4.043051","url":null,"abstract":"Under unfavorable conditions, fusion images of infrared and visible images often lack edge contrast and details. To address this issue, we propose an edge-oriented unrolling network, which comprises a feature extraction network and a feature fusion network. In our approach, after respective enhancement processes, the original infrared/visible image pair with their enhancement version is combined as the input to get more prior information acquisition. First, the feature extraction network consists of four independent iterative edge-oriented unrolling feature extraction networks based on the edge-oriented deep unrolling residual module (EURM), in which the convolutions in the EURM modules are replaced with edge-oriented convolution blocks to enhance the edge features. Then, the convolutional feature fusion network with differential structure is proposed to obtain the final fusion result, through utilizing the concatenate operation to map multidimensional features. In addition, the loss function in the fusion network is optimized to balance multiple features with significant differences in order to achieve better visual effect. Experimental results on multiple datasets demonstrate that the proposed method produces competitive fusion images as evaluated subjectively and objectively, with balanced luminance, sharper edge, and better details.","PeriodicalId":54843,"journal":{"name":"Journal of Electronic Imaging","volume":null,"pages":null},"PeriodicalIF":1.1,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142202447","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Precipitation nowcasting based on ConvLSTM-UNet deep spatiotemporal network 基于 ConvLSTM-UNet 深度时空网络的降水预报
IF 1.1 4区 计算机科学 Q4 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-08-01 DOI: 10.1117/1.jei.33.4.043053
Xiangming Zheng, Huawang Qin, Haoran Chen, Weixi Wang, Piao Shi
The primary objective of precipitation forecasting is to accurately predict short-term precipitation at a high resolution within a specific area, which is a significant and intricate challenge. Traditional models often struggle to capture the multidimensional characteristics of precipitation clouds in both time and space, leading to imprecise predictions due to their expansion, dissipation, and deformation. Recognizing this limitation, we introduce ConvLSTM-UNet, which leverages spatiotemporal feature extraction from meteorological images. ConvLSTM-UNet is an efficient convolutional neural network (CNN) based on the classical UNet architecture, equipped with ConvLSTM and improved deep separable convolutions. We evaluate our approach on the generic time series dataset Moving MNIST and the regional precipitation dataset of the Netherlands. The experimental results show that the proposed method has better spatiotemporal prediction skills than other tested models, and the mean squared error is reduced by more than 7.2%. In addition, the visualization results of the precipitation forecast show that the approach has a better ability to capture heavy precipitation, and the texture details of the precipitation forecast are closer to the ground truth.
降水预报的主要目标是以高分辨率准确预测特定区域内的短期降水,这是一项重大而复杂的挑战。传统模型往往难以捕捉降水云在时间和空间上的多维特征,从而导致因降水云的膨胀、消散和变形而产生的不精确预测。认识到这一局限性,我们引入了 ConvLSTM-UNet,它利用了气象图像的时空特征提取。ConvLSTM-UNet 是一种基于经典 UNet 架构的高效卷积神经网络(CNN),配备了 ConvLSTM 和改进的深度可分离卷积。我们对通用时间序列数据集 Moving MNIST 和荷兰地区降水数据集进行了评估。实验结果表明,与其他测试模型相比,所提出的方法具有更好的时空预测能力,均方误差降低了 7.2% 以上。此外,降水预报的可视化结果表明,该方法捕捉强降水的能力更强,降水预报的纹理细节更接近地面实况。
{"title":"Precipitation nowcasting based on ConvLSTM-UNet deep spatiotemporal network","authors":"Xiangming Zheng, Huawang Qin, Haoran Chen, Weixi Wang, Piao Shi","doi":"10.1117/1.jei.33.4.043053","DOIUrl":"https://doi.org/10.1117/1.jei.33.4.043053","url":null,"abstract":"The primary objective of precipitation forecasting is to accurately predict short-term precipitation at a high resolution within a specific area, which is a significant and intricate challenge. Traditional models often struggle to capture the multidimensional characteristics of precipitation clouds in both time and space, leading to imprecise predictions due to their expansion, dissipation, and deformation. Recognizing this limitation, we introduce ConvLSTM-UNet, which leverages spatiotemporal feature extraction from meteorological images. ConvLSTM-UNet is an efficient convolutional neural network (CNN) based on the classical UNet architecture, equipped with ConvLSTM and improved deep separable convolutions. We evaluate our approach on the generic time series dataset Moving MNIST and the regional precipitation dataset of the Netherlands. The experimental results show that the proposed method has better spatiotemporal prediction skills than other tested models, and the mean squared error is reduced by more than 7.2%. In addition, the visualization results of the precipitation forecast show that the approach has a better ability to capture heavy precipitation, and the texture details of the precipitation forecast are closer to the ground truth.","PeriodicalId":54843,"journal":{"name":"Journal of Electronic Imaging","volume":null,"pages":null},"PeriodicalIF":1.1,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142202440","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Squeeze-and-excitation attention and bi-directional feature pyramid network for filter screens surface detection 用于滤网表面检测的挤压-激发注意和双向特征金字塔网络
IF 1.1 4区 计算机科学 Q4 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-08-01 DOI: 10.1117/1.jei.33.4.043044
Junpeng Xu, Xiangbo Zhu, Lei Shi, Jin Li, Ziman Guo
Based on the enhanced YOLOv5, a deep learning defect detection technique is presented to deal with the problem of inadequate effectiveness in manually detecting problems on the surface of filter screens. In the last layer of the backbone network, the method combines the squeeze-and-excitation attention mechanism module, the method assigns weights to image locations based on the channel domain perspective to obtain more feature information. It also compares the results with a simple, parameter-free attention model (SimAM), which is an attention mechanism without the channel domain, and the results are higher than SimAM 0.7%. In addition, the neck network replaces the basic PANet structure with the bi-directional feature pyramid network module, which introduces multi-scale feature fusion. The experimental results show that the improved YOLOv5 algorithm has an average defect detection accuracy of 97.7% on the dataset, which is 11.3%, 12.8%, 2%, 7.8%, 5.1%, and 1.3% higher than YOLOv3, faster R-CNN, YOLOv5, SSD, YOLOv7, and YOLOv8, respectively. It can quickly and accurately identify various defects on the surface of the filter, which has an outstanding contribution to the filter manufacturing industry.
基于增强型 YOLOv5,提出了一种深度学习缺陷检测技术,以解决人工检测滤网表面问题效果不佳的问题。在骨干网络的最后一层,该方法结合了挤压-激发注意机制模块,基于通道域视角为图像位置分配权重,以获取更多特征信息。该方法还将结果与简单的无参数注意力模型(SimAM)进行了比较,后者是一种没有通道域的注意力机制,结果比 SimAM 高出 0.7%。此外,颈部网络用双向特征金字塔网络模块取代了基本的 PANet 结构,引入了多尺度特征融合。实验结果表明,改进后的 YOLOv5 算法在数据集上的平均缺陷检测准确率为 97.7%,分别比 YOLOv3、更快的 R-CNN、YOLOv5、SSD、YOLOv7 和 YOLOv8 高 11.3%、12.8%、2%、7.8%、5.1% 和 1.3%。它能快速准确地识别过滤器表面的各种缺陷,为过滤器制造业做出了突出贡献。
{"title":"Squeeze-and-excitation attention and bi-directional feature pyramid network for filter screens surface detection","authors":"Junpeng Xu, Xiangbo Zhu, Lei Shi, Jin Li, Ziman Guo","doi":"10.1117/1.jei.33.4.043044","DOIUrl":"https://doi.org/10.1117/1.jei.33.4.043044","url":null,"abstract":"Based on the enhanced YOLOv5, a deep learning defect detection technique is presented to deal with the problem of inadequate effectiveness in manually detecting problems on the surface of filter screens. In the last layer of the backbone network, the method combines the squeeze-and-excitation attention mechanism module, the method assigns weights to image locations based on the channel domain perspective to obtain more feature information. It also compares the results with a simple, parameter-free attention model (SimAM), which is an attention mechanism without the channel domain, and the results are higher than SimAM 0.7%. In addition, the neck network replaces the basic PANet structure with the bi-directional feature pyramid network module, which introduces multi-scale feature fusion. The experimental results show that the improved YOLOv5 algorithm has an average defect detection accuracy of 97.7% on the dataset, which is 11.3%, 12.8%, 2%, 7.8%, 5.1%, and 1.3% higher than YOLOv3, faster R-CNN, YOLOv5, SSD, YOLOv7, and YOLOv8, respectively. It can quickly and accurately identify various defects on the surface of the filter, which has an outstanding contribution to the filter manufacturing industry.","PeriodicalId":54843,"journal":{"name":"Journal of Electronic Imaging","volume":null,"pages":null},"PeriodicalIF":1.1,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142202451","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Journal of Electronic Imaging
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1