首页 > 最新文献

IEEE transactions on image processing : a publication of the IEEE Signal Processing Society最新文献

英文 中文
DREAM-PCD: Deep Reconstruction and Enhancement of mmWave Radar Pointcloud DREAM-PCD:毫米波雷达点云的深度重建与增强
Ruixu Geng;Yadong Li;Dongheng Zhang;Jincheng Wu;Yating Gao;Yang Hu;Yan Chen
Millimeter-wave (mmWave) radar pointcloud offers attractive potential for 3D sensing, thanks to its robustness in challenging conditions such as smoke and low illumination. However, existing methods failed to simultaneously address the three main challenges in mmWave radar pointcloud reconstruction: specular information lost, low angular resolution, and severe interference. In this paper, we propose DREAM-PCD, a novel framework specifically designed for real-time 3D environment sensing that combines signal processing and deep learning methods into three well-designed components to tackle all three challenges: Non-Coherent Accumulation for dense points, Synthetic Aperture Accumulation for improved angular resolution, and Real-Denoise Multiframe network for interference removal. By leveraging causal multiple viewpoints accumulation and the “real-denoise” mechanism, DREAM-PCD significantly enhances the generalization performance and real-time capability. We also introduce RadarEyes, the largest mmWave indoor dataset with over 1,000,000 frames, featuring a unique design incorporating two orthogonal single-chip radars, Lidar, and camera, enriching dataset diversity and applications. Experimental results demonstrate that DREAM-PCD surpasses existing methods in reconstruction quality, and exhibits superior generalization and real-time capabilities, enabling high-quality real-time reconstruction of radar pointcloud under various parameters and scenarios. We believe that DREAM-PCD, along with the RadarEyes dataset, will significantly advance mmWave radar perception in future real-world applications.
由于毫米波(mmWave)雷达点云在烟雾和低照度等具有挑战性的条件下的鲁棒性,它为3D传感提供了诱人的潜力。然而,现有的方法无法同时解决毫米波雷达点云重建中的三个主要挑战:镜面信息丢失、低角分辨率和严重干扰。在本文中,我们提出了DREAM-PCD,这是一个专门为实时3D环境传感设计的新框架,它将信号处理和深度学习方法结合到三个精心设计的组件中,以解决所有三个挑战:密集点的非相干积累,提高角分辨率的合成孔径积累,以及用于去除干扰的real-降噪多帧网络。DREAM-PCD利用因果多视点积累和“实噪”机制,显著提高了泛化性能和实时性。我们还推出了RadarEyes,这是最大的毫米波室内数据集,拥有超过1,000,000帧,具有独特的设计,结合了两个正交的单芯片雷达,激光雷达和摄像头,丰富了数据集的多样性和应用。实验结果表明,DREAM-PCD在重建质量上优于现有方法,具有较好的泛化能力和实时性,能够在各种参数和场景下实现高质量的雷达点云实时重建。我们相信DREAM-PCD与RadarEyes数据集将在未来的实际应用中显著推进毫米波雷达感知。
{"title":"DREAM-PCD: Deep Reconstruction and Enhancement of mmWave Radar Pointcloud","authors":"Ruixu Geng;Yadong Li;Dongheng Zhang;Jincheng Wu;Yating Gao;Yang Hu;Yan Chen","doi":"10.1109/TIP.2024.3512356","DOIUrl":"10.1109/TIP.2024.3512356","url":null,"abstract":"Millimeter-wave (mmWave) radar pointcloud offers attractive potential for 3D sensing, thanks to its robustness in challenging conditions such as smoke and low illumination. However, existing methods failed to simultaneously address the three main challenges in mmWave radar pointcloud reconstruction: specular information lost, low angular resolution, and severe interference. In this paper, we propose DREAM-PCD, a novel framework specifically designed for real-time 3D environment sensing that combines signal processing and deep learning methods into three well-designed components to tackle all three challenges: Non-Coherent Accumulation for dense points, Synthetic Aperture Accumulation for improved angular resolution, and Real-Denoise Multiframe network for interference removal. By leveraging causal multiple viewpoints accumulation and the “real-denoise” mechanism, DREAM-PCD significantly enhances the generalization performance and real-time capability. We also introduce RadarEyes, the largest mmWave indoor dataset with over 1,000,000 frames, featuring a unique design incorporating two orthogonal single-chip radars, Lidar, and camera, enriching dataset diversity and applications. Experimental results demonstrate that DREAM-PCD surpasses existing methods in reconstruction quality, and exhibits superior generalization and real-time capabilities, enabling high-quality real-time reconstruction of radar pointcloud under various parameters and scenarios. We believe that DREAM-PCD, along with the RadarEyes dataset, will significantly advance mmWave radar perception in future real-world applications.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":"33 ","pages":"6774-6789"},"PeriodicalIF":0.0,"publicationDate":"2024-12-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142809286","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Learning Frame-Event Fusion for Motion Deblurring 学习帧-事件融合运动去模糊
Wen Yang;Jinjian Wu;Jupo Ma;Leida Li;Weisheng Dong;Guangming Shi
Motion deblurring is a highly ill-posed problem due to the significant loss of motion information in the blurring process. Complementary informative features from auxiliary sensors such as event cameras can be explored for guiding motion deblurring. The event camera can capture rich motion information asynchronously with microsecond accuracy. In this paper, a novel frame-event fusion framework is proposed for event-driven motion deblurring (FEF-Deblur), which can sufficiently explore long-range cross-modal information interactions. Firstly, different modalities are usually complementary and also redundant. Cross-modal fusion is modeled as complementary-unique features separation-and-aggregation, avoiding the modality redundancy. Unique features and complementary features are first inferred with parallel intra-modal self-attention and inter-modal cross-attention respectively. After that, a correlation-based constraint is designed to act between unique and complementary features to facilitate their differentiation, which assists in cross-modal redundancy suppression. Additionally, spatio-temporal dependencies among neighboring inputs are crucial for motion deblurring. A recurrent cross attention is introduced to preserve inter-input attention information, in which the current spatial features and aggregated temporal features are attending to each other by establishing the long-range interaction between them. Extensive experiments on both synthetic and real-world motion deblurring datasets demonstrate our method outperforms state-of-the-art event-based and image/video-based methods.
运动去模糊是一个高度不适定的问题,因为在模糊过程中运动信息丢失严重。来自辅助传感器(如事件相机)的补充信息特征可以用于指导运动去模糊。该事件相机能够以微秒级精度异步捕捉丰富的运动信息。针对事件驱动运动去模糊(FEF-Deblur),提出了一种新的帧-事件融合框架,能够充分探索远距离跨模态信息交互。首先,不同的模态通常是互补的,也是冗余的。跨模态融合建模为互补-唯一特征分离-聚合,避免了模态冗余。首先用平行的模态内自注意和模态间交叉注意分别推断出独特特征和互补特征。之后,设计了一个基于相关性的约束,在唯一特征和互补特征之间起作用,以促进它们的区分,这有助于抑制跨模态冗余。此外,相邻输入之间的时空依赖关系对于运动去模糊至关重要。为了保存输入间的注意信息,引入了循环交叉注意,其中当前空间特征和聚合时间特征通过建立它们之间的远程交互作用而相互关注。在合成和真实世界的运动去模糊数据集上进行的大量实验表明,我们的方法优于最先进的基于事件和基于图像/视频的方法。
{"title":"Learning Frame-Event Fusion for Motion Deblurring","authors":"Wen Yang;Jinjian Wu;Jupo Ma;Leida Li;Weisheng Dong;Guangming Shi","doi":"10.1109/TIP.2024.3512362","DOIUrl":"10.1109/TIP.2024.3512362","url":null,"abstract":"Motion deblurring is a highly ill-posed problem due to the significant loss of motion information in the blurring process. Complementary informative features from auxiliary sensors such as event cameras can be explored for guiding motion deblurring. The event camera can capture rich motion information asynchronously with microsecond accuracy. In this paper, a novel frame-event fusion framework is proposed for event-driven motion deblurring (FEF-Deblur), which can sufficiently explore long-range cross-modal information interactions. Firstly, different modalities are usually complementary and also redundant. Cross-modal fusion is modeled as complementary-unique features separation-and-aggregation, avoiding the modality redundancy. Unique features and complementary features are first inferred with parallel intra-modal self-attention and inter-modal cross-attention respectively. After that, a correlation-based constraint is designed to act between unique and complementary features to facilitate their differentiation, which assists in cross-modal redundancy suppression. Additionally, spatio-temporal dependencies among neighboring inputs are crucial for motion deblurring. A recurrent cross attention is introduced to preserve inter-input attention information, in which the current spatial features and aggregated temporal features are attending to each other by establishing the long-range interaction between them. Extensive experiments on both synthetic and real-world motion deblurring datasets demonstrate our method outperforms state-of-the-art event-based and image/video-based methods.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":"33 ","pages":"6836-6849"},"PeriodicalIF":0.0,"publicationDate":"2024-12-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142809160","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Portrait Shadow Removal Using Context-Aware Illumination Restoration Network 使用上下文感知照明恢复网络的肖像阴影去除
Jiangjian Yu;Ling Zhang;Qing Zhang;Qifei Zhang;Daiguo Zhou;Chao Liang;Chunxia Xiao
Portrait shadow removal is a challenging task due to the complex surface of the face. Although existing work in this field makes substantial progress, these methods tend to overlook information in the background areas. However, this background information not only contains some important illumination cues but also plays a pivotal role in achieving lighting harmony between the face and the background after shadow elimination. In this paper, we propose a Context-aware Illumination Restoration Network (CIRNet) for portrait shadow removal. Our CIRNet consists of three stages. First, the Coarse Shadow Removal Network (CSRNet) mitigates the illumination discrepancies between shadow and non-shadow areas. Next, the Area-aware Shadow Restoration Network (ASRNet) predicts the illumination characteristics of shadowed areas by utilizing background context and non-shadow portrait context as references. Lastly, we introduce a Global Fusion Network to adaptively merge contextual information from different areas and generate the final shadow removal result. This approach leverages the illumination information from the background region while ensuring a more consistent overall illumination in the generated images. Our approach can also be extended to high-resolution portrait shadow removal and portrait specular highlight removal. Besides, we construct the first real facial shadow dataset for portrait shadow removal, consisting of 6200 pairs of facial images. Qualitative and quantitative comparisons demonstrate the advantages of our proposed dataset as well as our method.
由于脸部复杂的表面,人像阴影去除是一项具有挑战性的任务。虽然该领域的现有工作取得了实质性进展,但这些方法往往忽略了背景领域的信息。然而,这些背景信息不仅包含一些重要的照明线索,而且在消除阴影后实现面部与背景之间的照明和谐方面起着关键作用。在本文中,我们提出了一个上下文感知照明恢复网络(CIRNet)用于肖像阴影去除。我们的CIRNet由三个阶段组成。首先,粗糙阴影去除网络(CSRNet)减轻了阴影和非阴影区域之间的光照差异。其次,区域感知阴影恢复网络(ASRNet)利用背景环境和非阴影人像环境作为参考,预测阴影区域的照明特征。最后,我们引入了一个全局融合网络,自适应地融合来自不同区域的上下文信息,并生成最终的阴影去除结果。这种方法利用来自背景区域的照明信息,同时确保生成的图像中更一致的整体照明。我们的方法也可以扩展到高分辨率的肖像阴影去除和肖像镜面高光去除。此外,我们构建了第一个用于人像阴影去除的真实人脸阴影数据集,该数据集包含6200对人脸图像。定性和定量比较证明了我们提出的数据集以及我们的方法的优势。
{"title":"Portrait Shadow Removal Using Context-Aware Illumination Restoration Network","authors":"Jiangjian Yu;Ling Zhang;Qing Zhang;Qifei Zhang;Daiguo Zhou;Chao Liang;Chunxia Xiao","doi":"10.1109/TIP.2024.3497802","DOIUrl":"10.1109/TIP.2024.3497802","url":null,"abstract":"Portrait shadow removal is a challenging task due to the complex surface of the face. Although existing work in this field makes substantial progress, these methods tend to overlook information in the background areas. However, this background information not only contains some important illumination cues but also plays a pivotal role in achieving lighting harmony between the face and the background after shadow elimination. In this paper, we propose a Context-aware Illumination Restoration Network (CIRNet) for portrait shadow removal. Our CIRNet consists of three stages. First, the Coarse Shadow Removal Network (CSRNet) mitigates the illumination discrepancies between shadow and non-shadow areas. Next, the Area-aware Shadow Restoration Network (ASRNet) predicts the illumination characteristics of shadowed areas by utilizing background context and non-shadow portrait context as references. Lastly, we introduce a Global Fusion Network to adaptively merge contextual information from different areas and generate the final shadow removal result. This approach leverages the illumination information from the background region while ensuring a more consistent overall illumination in the generated images. Our approach can also be extended to high-resolution portrait shadow removal and portrait specular highlight removal. Besides, we construct the first real facial shadow dataset for portrait shadow removal, consisting of 6200 pairs of facial images. Qualitative and quantitative comparisons demonstrate the advantages of our proposed dataset as well as our method.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":"34 ","pages":"1-15"},"PeriodicalIF":0.0,"publicationDate":"2024-12-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142782546","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
CrossEI: Boosting Motion-Oriented Object Tracking With an Event Camera 交叉:用事件相机增强面向运动的对象跟踪
Zhiwen Chen;Jinjian Wu;Weisheng Dong;Leida Li;Guangming Shi
With the differential sensitivity and high time resolution, event cameras can record detailed motion clues, which form a complementary advantage with frame-based cameras to enhance the object tracking, especially in challenging dynamic scenes. However, how to better match heterogeneous event-image data and exploit rich complementary cues from them still remains an open issue. In this paper, we align event-image modalities by proposing a motion adaptive event sampling method, and we revisit the cross-complementarities of event-image data to design a bidirectional-enhanced fusion framework. Specifically, this sampling strategy can adapt to different dynamic scenes and integrate aligned event-image pairs. Besides, we design an image-guided motion estimation unit for extracting explicit instance-level motions, aiming at refining the uncertain event clues to distinguish primary objects and background. Then, a semantic modulation module is devised to utilize the enhanced object motion to modify the image features. Coupled with these two modules, this framework learns both the high motion sensitivity of events and the full texture of images to achieve more accurate and robust tracking. The proposed method is easily embedded in existing tracking pipelines, and trained end-to-end. We evaluate it on four large benchmarks, i.e. FE108, VisEvent, FE240hz and CoeSot. Extensive experiments demonstrate our method achieves state-of-the-art performance, and large improvements are pointed as contributions by our sampling strategy and fusion concept.
由于差分灵敏度和高时间分辨率,事件相机可以记录详细的运动线索,这与基于帧的相机形成互补优势,以增强目标跟踪,特别是在具有挑战性的动态场景中。然而,如何更好地匹配异构事件图像数据并从中挖掘丰富的互补线索仍然是一个悬而未决的问题。在本文中,我们通过提出一种运动自适应事件采样方法来对齐事件图像模式,并重新审视事件图像数据的交叉互补性来设计一个双向增强的融合框架。具体来说,该采样策略可以适应不同的动态场景,并整合对齐的事件-图像对。此外,我们设计了一个图像引导的运动估计单元,用于提取显式的实例级运动,旨在提炼不确定事件线索,以区分主要目标和背景。然后,设计了语义调制模块,利用增强的物体运动对图像特征进行修改。结合这两个模块,该框架学习了事件的高运动灵敏度和图像的完整纹理,以实现更准确和鲁棒的跟踪。该方法易于嵌入到现有的跟踪管道中,并且可以进行端到端训练。我们在四个大型基准测试上进行了评估,即FE108, VisEvent, FE240hz和CoeSot。大量的实验表明,我们的方法达到了最先进的性能,并指出了我们的采样策略和融合概念的巨大改进。
{"title":"CrossEI: Boosting Motion-Oriented Object Tracking With an Event Camera","authors":"Zhiwen Chen;Jinjian Wu;Weisheng Dong;Leida Li;Guangming Shi","doi":"10.1109/TIP.2024.3505672","DOIUrl":"10.1109/TIP.2024.3505672","url":null,"abstract":"With the differential sensitivity and high time resolution, event cameras can record detailed motion clues, which form a complementary advantage with frame-based cameras to enhance the object tracking, especially in challenging dynamic scenes. However, how to better match heterogeneous event-image data and exploit rich complementary cues from them still remains an open issue. In this paper, we align event-image modalities by proposing a motion adaptive event sampling method, and we revisit the cross-complementarities of event-image data to design a bidirectional-enhanced fusion framework. Specifically, this sampling strategy can adapt to different dynamic scenes and integrate aligned event-image pairs. Besides, we design an image-guided motion estimation unit for extracting explicit instance-level motions, aiming at refining the uncertain event clues to distinguish primary objects and background. Then, a semantic modulation module is devised to utilize the enhanced object motion to modify the image features. Coupled with these two modules, this framework learns both the high motion sensitivity of events and the full texture of images to achieve more accurate and robust tracking. The proposed method is easily embedded in existing tracking pipelines, and trained end-to-end. We evaluate it on four large benchmarks, i.e. FE108, VisEvent, FE240hz and CoeSot. Extensive experiments demonstrate our method achieves state-of-the-art performance, and large improvements are pointed as contributions by our sampling strategy and fusion concept.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":"34 ","pages":"73-84"},"PeriodicalIF":0.0,"publicationDate":"2024-12-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142776604","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
PTH-Net: Dynamic Facial Expression Recognition Without Face Detection and Alignment PTH-Net:无人脸检测和对齐的动态面部表情识别
Min Li;Xiaoqin Zhang;Tangfei Liao;Sheng Lin;Guobao Xiao
Pyramid Temporal Hierarchy Network (PTH-Net) is a new paradigm for dynamic facial expression recognition, applied directly to raw videos, without face detection and alignment. Unlike the traditional paradigm, which focus only on facial areas and often overlooks valuable information like body movements, PTH-Net preserves more critical information. It does this by distinguishing between backgrounds and human bodies at the feature level, offering greater flexibility as an end-to-end network. Specifically, PTH-Net utilizes a pre-trained backbone to extract multiple general features of video understanding at various temporal frequencies, forming a temporal feature pyramid. It then further expands this temporal hierarchy through differentiated parameter sharing and downsampling, ultimately refining emotional information under the supervision of expression temporal-frequency invariance. Additionally, PTH-Net features an efficient Scalable Semantic Distinction layer that enhances feature discrimination, helping to better identify target expressions versus non-target ones in the video. Finally, extensive experiments demonstrate that PTH-Net performs excellently in eight challenging benchmarks, with lower computational costs compared to previous methods. The source code is available at https://github.com/lm495455/PTH-Net.
金字塔时态层次网络(PTH-Net)是一种新的动态面部表情识别范式,直接应用于原始视频,不需要人脸检测和对齐。与传统模式不同,PTH-Net只关注面部区域,往往忽略了身体动作等有价值的信息,而PTH-Net保留了更多关键信息。它通过在特征级别上区分背景和人体来实现这一点,作为端到端网络提供了更大的灵活性。具体来说,PTH-Net利用预训练的主干提取不同时间频率下视频理解的多个一般特征,形成一个时间特征金字塔。然后,它通过差异化参数共享和下采样进一步扩展这种时间层次,最终在表达时间-频率不变性的监督下提炼情感信息。此外,PTH-Net具有高效的可扩展语义区分层,增强了特征识别,有助于更好地识别视频中的目标表达式与非目标表达式。最后,大量的实验表明,PTH-Net在8个具有挑战性的基准测试中表现出色,与以前的方法相比,计算成本更低。源代码可从https://github.com/lm495455/PTH-Net获得。
{"title":"PTH-Net: Dynamic Facial Expression Recognition Without Face Detection and Alignment","authors":"Min Li;Xiaoqin Zhang;Tangfei Liao;Sheng Lin;Guobao Xiao","doi":"10.1109/TIP.2024.3504298","DOIUrl":"10.1109/TIP.2024.3504298","url":null,"abstract":"Pyramid Temporal Hierarchy Network (PTH-Net) is a new paradigm for dynamic facial expression recognition, applied directly to raw videos, without face detection and alignment. Unlike the traditional paradigm, which focus only on facial areas and often overlooks valuable information like body movements, PTH-Net preserves more critical information. It does this by distinguishing between backgrounds and human bodies at the feature level, offering greater flexibility as an end-to-end network. Specifically, PTH-Net utilizes a pre-trained backbone to extract multiple general features of video understanding at various temporal frequencies, forming a temporal feature pyramid. It then further expands this temporal hierarchy through differentiated parameter sharing and downsampling, ultimately refining emotional information under the supervision of expression temporal-frequency invariance. Additionally, PTH-Net features an efficient Scalable Semantic Distinction layer that enhances feature discrimination, helping to better identify target expressions versus non-target ones in the video. Finally, extensive experiments demonstrate that PTH-Net performs excellently in eight challenging benchmarks, with lower computational costs compared to previous methods. The source code is available at \u0000<uri>https://github.com/lm495455/PTH-Net</uri>\u0000.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":"34 ","pages":"30-43"},"PeriodicalIF":0.0,"publicationDate":"2024-11-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142752984","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
RSB-Pose: Robust Short-Baseline Binocular 3D Human Pose Estimation With Occlusion Handling RSB-Pose:鲁棒短基线双目三维人体姿态估计与遮挡处理
Xiaoyue Wan;Zhuo Chen;Xu Zhao
In the domain of 3D Human Pose Estimation, which finds widespread daily applications, the requirement for convenient acquisition equipment continues to grow. To satisfy this demand, we focus on a short-baseline binocular setup that offers both portability and a geometric measurement capability that significantly reduces depth ambiguity. However, as the binocular baseline shortens, two serious challenges emerge: first, the robustness of 3D reconstruction against 2D errors deteriorates; second, occlusion reoccurs frequently due to the limited visual differences between two views. To address the first challenge, we propose the Stereo Co-Keypoints Estimation module to improve the view consistency of 2D keypoints and enhance the 3D robustness. In this module, the disparity is utilized to represent the correspondence of binocular 2D points, and the Stereo Volume Feature (SVF) is introduced to contain binocular features across different disparities. Through the regression of SVF, two-view 2D keypoints are simultaneously estimated in a collaborative way which restricts their view consistency. Furthermore, to deal with occlusions, a Pre-trained Pose Transformer module is introduced. Through this module, 3D poses are refined by perceiving pose coherence, a representation of joint correlations. This perception is injected by the Pose Transformer network and learned through a pre-training task that recovers iterative masked joints. Comprehensive experiments on H36M and MHAD datasets validate the effectiveness of our approach in the short-baseline binocular 3D Human Pose Estimation and occlusion handling.
在日常应用广泛的三维人体姿态估计领域,对方便的采集设备的需求不断增长。为了满足这一需求,我们专注于短基线双目设置,提供便携性和几何测量能力,显着减少深度模糊。然而,随着双眼基线的缩短,出现了两个严峻的挑战:一是三维重建对二维误差的鲁棒性下降;其次,由于两个视图之间的视觉差异有限,遮挡经常再次发生。为了解决第一个问题,我们提出了立体协同关键点估计模块,以提高二维关键点的视图一致性和增强三维鲁棒性。该模块利用视差来表示双目二维点的对应关系,并引入立体体积特征(SVF)来包含不同视差间的双目特征。通过SVF的回归,以一种协同的方式同时估计两视图二维关键点,但限制了它们的视图一致性。此外,为了处理遮挡,引入了预训练的Pose Transformer模块。通过该模块,通过感知姿态一致性(关节相关性的一种表示)来改进3D姿态。这种感知是由Pose Transformer网络注入的,并通过恢复迭代掩蔽关节的预训练任务来学习。在H36M和MHAD数据集上的综合实验验证了我们的方法在短基线双目三维人体姿态估计和遮挡处理中的有效性。
{"title":"RSB-Pose: Robust Short-Baseline Binocular 3D Human Pose Estimation With Occlusion Handling","authors":"Xiaoyue Wan;Zhuo Chen;Xu Zhao","doi":"10.1109/TIP.2024.3490401","DOIUrl":"10.1109/TIP.2024.3490401","url":null,"abstract":"In the domain of 3D Human Pose Estimation, which finds widespread daily applications, the requirement for convenient acquisition equipment continues to grow. To satisfy this demand, we focus on a short-baseline binocular setup that offers both portability and a geometric measurement capability that significantly reduces depth ambiguity. However, as the binocular baseline shortens, two serious challenges emerge: first, the robustness of 3D reconstruction against 2D errors deteriorates; second, occlusion reoccurs frequently due to the limited visual differences between two views. To address the first challenge, we propose the Stereo Co-Keypoints Estimation module to improve the view consistency of 2D keypoints and enhance the 3D robustness. In this module, the disparity is utilized to represent the correspondence of binocular 2D points, and the Stereo Volume Feature (SVF) is introduced to contain binocular features across different disparities. Through the regression of SVF, two-view 2D keypoints are simultaneously estimated in a collaborative way which restricts their view consistency. Furthermore, to deal with occlusions, a Pre-trained Pose Transformer module is introduced. Through this module, 3D poses are refined by perceiving pose coherence, a representation of joint correlations. This perception is injected by the Pose Transformer network and learned through a pre-training task that recovers iterative masked joints. Comprehensive experiments on H36M and MHAD datasets validate the effectiveness of our approach in the short-baseline binocular 3D Human Pose Estimation and occlusion handling.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":"34 ","pages":"60-72"},"PeriodicalIF":0.0,"publicationDate":"2024-11-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142752985","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Unified Video Reconstruction for Rolling Shutter and Global Shutter Cameras 卷帘式和全局快门相机的统一视频重建
Bin Fan;Zhexiong Wan;Boxin Shi;Chao Xu;Yuchao Dai
Currently, the general domain of video reconstruction (VR) is fragmented into different shutters spanning global shutter and rolling shutter cameras. Despite rapid progress in the state-of-the-art, existing methods overwhelmingly follow shutter-specific paradigms and cannot conceptually generalize to other shutter types, hindering the uniformity of VR models. In this paper, we propose UniVR, a versatile framework to handle various shutters through unified modeling and shared parameters. Specifically, UniVR encodes diverse shutter types into a unified space via a tractable shutter adapter, which is parameter-free and thus can be seamlessly delivered to current well-established VR architectures for cross-shutter transfer. To demonstrate its effectiveness, we conceptualize UniVR as three shutter-generic VR methods, namely Uni-SoftSplat, Uni-SuperSloMo, and Uni-RIFE. Extensive experimental results demonstrate that the pre-trained model without any fine-tuning can achieve reasonable performance even on novel shutters. After fine-tuning, new state-of-the-art performances are established that go beyond shutter-specific methods and enjoy strong generalization. The code is available at https://github.com/GitCVfb/UniVR.
目前,视频重建(VR)的一般领域被分割成不同的快门,包括全局快门和滚动快门相机。尽管最新技术进步迅速,但现有方法绝大多数遵循快门特定范式,无法在概念上推广到其他快门类型,从而阻碍了VR模型的统一性。在本文中,我们提出了一个通用框架UniVR,通过统一建模和共享参数来处理各种快门。具体来说,UniVR通过一个易于处理的快门适配器将不同的快门类型编码到一个统一的空间中,这是无参数的,因此可以无缝地交付到当前完善的VR架构中进行跨快门传输。为了证明其有效性,我们将UniVR概念化为三种快门通用VR方法,即Uni-SoftSplat, Uni-SuperSloMo和Uni-RIFE。大量的实验结果表明,无需任何微调的预训练模型即使在新型百叶窗上也能获得合理的性能。经过微调,建立了新的最先进的性能,超越了特定的快门方法,并具有很强的通用性。代码可在https://github.com/GitCVfb/UniVR上获得。
{"title":"Unified Video Reconstruction for Rolling Shutter and Global Shutter Cameras","authors":"Bin Fan;Zhexiong Wan;Boxin Shi;Chao Xu;Yuchao Dai","doi":"10.1109/TIP.2024.3504275","DOIUrl":"10.1109/TIP.2024.3504275","url":null,"abstract":"Currently, the general domain of video reconstruction (VR) is fragmented into different shutters spanning global shutter and rolling shutter cameras. Despite rapid progress in the state-of-the-art, existing methods overwhelmingly follow shutter-specific paradigms and cannot conceptually generalize to other shutter types, hindering the uniformity of VR models. In this paper, we propose UniVR, a versatile framework to handle various shutters through unified modeling and shared parameters. Specifically, UniVR encodes diverse shutter types into a unified space via a tractable shutter adapter, which is parameter-free and thus can be seamlessly delivered to current well-established VR architectures for cross-shutter transfer. To demonstrate its effectiveness, we conceptualize UniVR as three shutter-generic VR methods, namely Uni-SoftSplat, Uni-SuperSloMo, and Uni-RIFE. Extensive experimental results demonstrate that the pre-trained model without any fine-tuning can achieve reasonable performance even on novel shutters. After fine-tuning, new state-of-the-art performances are established that go beyond shutter-specific methods and enjoy strong generalization. The code is available at \u0000<uri>https://github.com/GitCVfb/UniVR</uri>\u0000.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":"33 ","pages":"6821-6835"},"PeriodicalIF":0.0,"publicationDate":"2024-11-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142752986","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
HAda: Hyper-Adaptive Parameter-Efficient Learning for Multi-View ConvNets HAda:多视图卷积神经网络的超自适应参数高效学习
Shiye Wang;Changsheng Li;Zeyu Yan;Wanjun Liang;Ye Yuan;Guoren Wang
Recent years have witnessed a great success of multi-view learning empowered by deep ConvNets, leveraging a large number of network parameters. Nevertheless, there is an ongoing consideration regarding the essentiality of all these parameters in multi-view ConvNets. As we know, hypernetworks offer a promising solution to reduce the number of parameters by learning a concise network to generate weights for the larger target network, illustrating the presence of redundant information within network parameters. However, how to leverage hypernetworks for learning parameter-efficient multi-view ConvNets remains underexplored. In this paper, we present a lightweight multi-layer shared Hyper-Adaptive network (HAda), aiming to simultaneously generate adaptive weights for different views and convolutional layers of deep multi-view ConvNets. The adaptability inherent in HAda not only contributes to a substantial reduction in parameter redundancy but also enables the modeling of intricate view-aware and layer-wise information. This capability ensures the maintenance of high performance, ultimately achieving parameter-efficient learning. Specifically, we design a multi-view shared module in HAda to capture information common across views. This module incorporates a shared global gated interpolation strategy, which generates layer-wise gating factors. These factors facilitate adaptive interpolation of global contextual information into the weights. Meanwhile, we put forward a tailored weight-calibrated adapter for each view that facilitates the conveyance of view-specific information. These adapters generate view-adaptive weight scaling calibrators, allowing the selective emphasis of personalized information for each view without introducing excessive parameters. Extensive experiments on six publicly available datasets demonstrate the effectiveness of the proposed method. In particular, HAda can serve as a flexible plug-in strategy to work well with existing multi-view methods for both image classification and image clustering tasks.
近年来,利用大量网络参数的深度卷积神经网络支持的多视图学习取得了巨大成功。然而,关于多视图卷积神经网络中所有这些参数的重要性,人们一直在考虑。正如我们所知,超网络提供了一个很有前途的解决方案,通过学习一个简洁的网络来为更大的目标网络生成权值,从而减少参数的数量,这说明了网络参数中存在冗余信息。然而,如何利用超网络来学习参数高效的多视图卷积神经网络仍未得到充分探索。本文提出了一种轻量级的多层共享超自适应网络(HAda),旨在同时为深度多视图卷积神经网络的不同视图和卷积层生成自适应权重。HAda固有的适应性不仅有助于大幅减少参数冗余,而且还可以对复杂的视图感知和分层信息进行建模。这种能力确保了高性能的维护,最终实现了参数高效学习。具体来说,我们在HAda中设计了一个多视图共享模块来捕获视图之间的公共信息。该模块采用共享的全局门控插值策略,生成分层门控因子。这些因素有助于将全局上下文信息自适应地插值到权重中。同时,我们为每个视图提供了量身定制的权重校准适配器,以方便视图特定信息的传输。这些适配器生成自适应视图的权重缩放校准器,允许在不引入过多参数的情况下为每个视图选择性地强调个性化信息。在六个公开可用的数据集上进行的大量实验证明了所提出方法的有效性。特别是,HAda可以作为一种灵活的插件策略,与现有的多视图方法一起很好地用于图像分类和图像聚类任务。
{"title":"HAda: Hyper-Adaptive Parameter-Efficient Learning for Multi-View ConvNets","authors":"Shiye Wang;Changsheng Li;Zeyu Yan;Wanjun Liang;Ye Yuan;Guoren Wang","doi":"10.1109/TIP.2024.3504252","DOIUrl":"10.1109/TIP.2024.3504252","url":null,"abstract":"Recent years have witnessed a great success of multi-view learning empowered by deep ConvNets, leveraging a large number of network parameters. Nevertheless, there is an ongoing consideration regarding the essentiality of all these parameters in multi-view ConvNets. As we know, hypernetworks offer a promising solution to reduce the number of parameters by learning a concise network to generate weights for the larger target network, illustrating the presence of redundant information within network parameters. However, how to leverage hypernetworks for learning parameter-efficient multi-view ConvNets remains underexplored. In this paper, we present a lightweight multi-layer shared Hyper-Adaptive network (HAda), aiming to simultaneously generate adaptive weights for different views and convolutional layers of deep multi-view ConvNets. The adaptability inherent in HAda not only contributes to a substantial reduction in parameter redundancy but also enables the modeling of intricate view-aware and layer-wise information. This capability ensures the maintenance of high performance, ultimately achieving parameter-efficient learning. Specifically, we design a multi-view shared module in HAda to capture information common across views. This module incorporates a shared global gated interpolation strategy, which generates layer-wise gating factors. These factors facilitate adaptive interpolation of global contextual information into the weights. Meanwhile, we put forward a tailored weight-calibrated adapter for each view that facilitates the conveyance of view-specific information. These adapters generate view-adaptive weight scaling calibrators, allowing the selective emphasis of personalized information for each view without introducing excessive parameters. Extensive experiments on six publicly available datasets demonstrate the effectiveness of the proposed method. In particular, HAda can serve as a flexible plug-in strategy to work well with existing multi-view methods for both image classification and image clustering tasks.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":"34 ","pages":"85-99"},"PeriodicalIF":0.0,"publicationDate":"2024-11-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142752987","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Saliency Segmentation Oriented Deep Image Compression With Novel Bit Allocation 面向显著性分割的新型位分配深度图像压缩
Yuan Li;Wei Gao;Ge Li;Siwei Ma
Image compression distortion can cause performance degradation of machine analysis tasks, therefore recent years have witnessed fast progress in developing deep image compression methods optimized for machine perception. However, the investigation still lacks for saliency segmentation. First, in this paper we propose a deep compression network increasing local signal fidelity of important image pixels for saliency segmentation, which is different from existing methods utilizing the analysis network loss for backward propagation. By this means, these two types of networks can be decoupled to improve the compatibility of proposed compression method for diverse saliency segmentation networks. Second, pixel-level bit weights are modeled with probability distribution in the proposed bit allocation method. The ascending cosine roll-down (ACRD) function allocates bits to those important pixels, which fits the essence that saliency segmentation can be regarded as the pixel-level bi-classification task. Third, the compression network is trained without the help of saliency segmentation, where latent representations are decomposed into base and enhancement channels. Base channels are retained in the whole image, while enhancement channels are utilized only for important pixels, and therefore more bits can benefit saliency segmentation via enhancement channels. Extensive experimental results demonstrate that the proposed method can save an average of 10.34% bitrate compared with the state-of-the-art deep image compression method, where the rate-accuracy (R-A) performances are evaluated on sixteen downstream saliency segmentation networks with five conventional SOD datasets. The code will be available at: https://openi.pcl.ac.cn/OpenAICoding/SaliencyIC and https://github.com/AkeLiLi/SaliencyIC.
图像压缩失真会导致机器分析任务的性能下降,因此近年来在开发针对机器感知优化的深度图像压缩方法方面取得了快速进展。然而,对于显著性分割的研究仍然缺乏。首先,本文提出了一种深度压缩网络,提高重要图像像素的局部信号保真度,用于显著性分割,这与利用分析网络损失进行反向传播的现有方法不同。通过这种方法,这两种类型的网络可以解耦,以提高所提出的压缩方法对不同显著性分割网络的兼容性。其次,在所提出的比特分配方法中,采用概率分布对像素级比特权进行建模。上升余弦下滚(ACRD)函数为重要的像素分配比特,这符合显著性分割可视为像素级双分类任务的本质。第三,在没有显著性分割的情况下训练压缩网络,在显著性分割中,潜在表示被分解为基本通道和增强通道。基本通道保留在整个图像中,而增强通道仅用于重要像素,因此通过增强通道可以获得更多的显着性分割。大量的实验结果表明,与目前最先进的深度图像压缩方法相比,该方法可以平均节省10.34%的比特率,其中在5个传统SOD数据集的16个下游显著性分割网络上评估了率精度(R-A)性能。代码将在https://openi.pcl.ac.cn/OpenAICoding/SaliencyIC和https://github.com/AkeLiLi/SaliencyIC上提供。
{"title":"Saliency Segmentation Oriented Deep Image Compression With Novel Bit Allocation","authors":"Yuan Li;Wei Gao;Ge Li;Siwei Ma","doi":"10.1109/TIP.2024.3504282","DOIUrl":"10.1109/TIP.2024.3504282","url":null,"abstract":"Image compression distortion can cause performance degradation of machine analysis tasks, therefore recent years have witnessed fast progress in developing deep image compression methods optimized for machine perception. However, the investigation still lacks for saliency segmentation. First, in this paper we propose a deep compression network increasing local signal fidelity of important image pixels for saliency segmentation, which is different from existing methods utilizing the analysis network loss for backward propagation. By this means, these two types of networks can be decoupled to improve the compatibility of proposed compression method for diverse saliency segmentation networks. Second, pixel-level bit weights are modeled with probability distribution in the proposed bit allocation method. The ascending cosine roll-down (ACRD) function allocates bits to those important pixels, which fits the essence that saliency segmentation can be regarded as the pixel-level bi-classification task. Third, the compression network is trained without the help of saliency segmentation, where latent representations are decomposed into base and enhancement channels. Base channels are retained in the whole image, while enhancement channels are utilized only for important pixels, and therefore more bits can benefit saliency segmentation via enhancement channels. Extensive experimental results demonstrate that the proposed method can save an average of 10.34% bitrate compared with the state-of-the-art deep image compression method, where the rate-accuracy (R-A) performances are evaluated on sixteen downstream saliency segmentation networks with five conventional SOD datasets. The code will be available at: \u0000<uri>https://openi.pcl.ac.cn/OpenAICoding/SaliencyIC</uri>\u0000 and \u0000<uri>https://github.com/AkeLiLi/SaliencyIC</uri>\u0000.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":"34 ","pages":"16-29"},"PeriodicalIF":0.0,"publicationDate":"2024-11-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142752988","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MWFormer: Multi-Weather Image Restoration Using Degradation-Aware Transformers MWFormer:利用感知退化的变换器修复多天气图像
Ruoxi Zhu;Zhengzhong Tu;Jiaming Liu;Alan C. Bovik;Yibo Fan
Restoring images captured under adverse weather conditions is a fundamental task for many computer vision applications. However, most existing weather restoration approaches are only capable of handling a specific type of degradation, which is often insufficient in real-world scenarios, such as rainy-snowy or rainy-hazy weather. Towards being able to address these situations, we propose a multi-weather Transformer, or MWFormer for short, which is a holistic vision Transformer that aims to solve multiple weather-induced degradations using a single, unified architecture. MWFormer uses hyper-networks and feature-wise linear modulation blocks to restore images degraded by various weather types using the same set of learned parameters. We first employ contrastive learning to train an auxiliary network that extracts content-independent, distortion-aware feature embeddings that efficiently represent predicted weather types, of which more than one may occur. Guided by these weather-informed predictions, the image restoration Transformer adaptively modulates its parameters to conduct both local and global feature processing, in response to multiple possible weather. Moreover, MWFormer allows for a novel way of tuning, during application, to either a single type of weather restoration or to hybrid weather restoration without any retraining, offering greater controllability than existing methods. Our experimental results on multi-weather restoration benchmarks show that MWFormer achieves significant performance improvements compared to existing state-of-the-art methods, without requiring much computational cost. Moreover, we demonstrate that our methodology of using hyper-networks can be integrated into various network architectures to further boost their performance. The code is available at: https://github.com/taco-group/MWFormer.
恢复在恶劣天气条件下捕获的图像是许多计算机视觉应用的基本任务。然而,大多数现有的天气恢复方法只能处理特定类型的退化,这在现实场景中往往是不够的,例如雨雪天气或雨雾天气。为了能够解决这些情况,我们提出了一个多天气变压器,或简称MWFormer,这是一个整体的愿景变压器,旨在使用一个单一的、统一的体系结构来解决多种天气导致的退化。MWFormer使用超网络和特征线性调制块,使用相同的学习参数集来恢复因各种天气类型而退化的图像。我们首先使用对比学习来训练一个辅助网络,该网络提取与内容无关的、扭曲感知的特征嵌入,这些特征嵌入有效地表示预测的天气类型,其中可能出现不止一种。在这些天气预报的指导下,图像恢复变压器自适应地调节其参数,以响应多种可能的天气,进行局部和全局特征处理。此外,MWFormer允许在应用过程中以一种新颖的方式进行调整,既可以进行单一类型的天气恢复,也可以进行混合天气恢复,而无需进行任何重新训练,比现有方法提供更大的可控性。我们在多天气恢复基准上的实验结果表明,与现有的最先进的方法相比,MWFormer在不需要太多计算成本的情况下实现了显著的性能改进。此外,我们证明了我们使用超网络的方法可以集成到各种网络架构中,以进一步提高其性能。代码可从https://github.com/taco-group/MWFormer获得。
{"title":"MWFormer: Multi-Weather Image Restoration Using Degradation-Aware Transformers","authors":"Ruoxi Zhu;Zhengzhong Tu;Jiaming Liu;Alan C. Bovik;Yibo Fan","doi":"10.1109/TIP.2024.3501855","DOIUrl":"10.1109/TIP.2024.3501855","url":null,"abstract":"Restoring images captured under adverse weather conditions is a fundamental task for many computer vision applications. However, most existing weather restoration approaches are only capable of handling a specific type of degradation, which is often insufficient in real-world scenarios, such as rainy-snowy or rainy-hazy weather. Towards being able to address these situations, we propose a multi-weather Transformer, or MWFormer for short, which is a holistic vision Transformer that aims to solve multiple weather-induced degradations using a single, unified architecture. MWFormer uses hyper-networks and feature-wise linear modulation blocks to restore images degraded by various weather types using the same set of learned parameters. We first employ contrastive learning to train an auxiliary network that extracts content-independent, distortion-aware feature embeddings that efficiently represent predicted weather types, of which more than one may occur. Guided by these weather-informed predictions, the image restoration Transformer adaptively modulates its parameters to conduct both local and global feature processing, in response to multiple possible weather. Moreover, MWFormer allows for a novel way of tuning, during application, to either a single type of weather restoration or to hybrid weather restoration without any retraining, offering greater controllability than existing methods. Our experimental results on multi-weather restoration benchmarks show that MWFormer achieves significant performance improvements compared to existing state-of-the-art methods, without requiring much computational cost. Moreover, we demonstrate that our methodology of using hyper-networks can be integrated into various network architectures to further boost their performance. The code is available at: \u0000<uri>https://github.com/taco-group/MWFormer</uri>\u0000.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":"33 ","pages":"6790-6805"},"PeriodicalIF":0.0,"publicationDate":"2024-11-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142712465","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE transactions on image processing : a publication of the IEEE Signal Processing Society
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1