首页 > 最新文献

IEEE transactions on image processing : a publication of the IEEE Signal Processing Society最新文献

英文 中文
Dark-EvGS: Event Camera as an Eye for Radiance Field in the Dark. Dark- evgs:事件相机在黑暗中作为发光场的眼睛。
IF 13.7 Pub Date : 2026-03-20 DOI: 10.1109/TIP.2026.3674360
Jingqian Wu, Peiqi Duan, Zongqiang Wang, Changwei Wang, Boxin Shi, Edmund Y Lam

In low-light environments, conventional cameras often struggle to capture clear multi-view images of objects due to dynamic range limitations and motion blur caused by long exposure. Event cameras, with their high-dynamic range and high-speed properties, have the potential to mitigate these issues. Additionally, 3D Gaussian Splatting (GS) enables radiance field reconstruction, facilitating bright frame synthesis from multiple viewpoints in low-light conditions. However, naively using an event-assisted 3D GS approach still faced challenges because, in low lights, events are noisy, frames lack quality, and the color tone may be inconsistent. To address these issues, we propose Dark-EvGS, the first event-assisted 3D GS framework that enables the reconstruction of bright frames from arbitrary viewpoints along the camera trajectory. Triplet-level supervision is proposed to gain holistic knowledge, granular details, and sharp scene rendering. The color tone matching block is proposed to guarantee the color consistency of the rendered frames. Furthermore, we introduce the first real-captured dataset for the event-guided bright frame synthesis task via 3D GS-based radiance field reconstruction. Experiments demonstrate that our method achieves better results than existing methods, conquering radiance field reconstruction under challenging low-light conditions. The code and sample data are included in the supplementary material.

在低光环境下,由于动态范围的限制和长时间曝光造成的运动模糊,传统相机往往难以捕捉清晰的多视图图像。事件摄像机具有高动态范围和高速特性,有可能缓解这些问题。此外,3D高斯溅射(GS)可以实现辐射场重建,促进在低光条件下从多个视点合成明亮的帧。然而,天真地使用事件辅助3D GS方法仍然面临挑战,因为在低光下,事件是嘈杂的,帧缺乏质量,色调可能不一致。为了解决这些问题,我们提出了Dark-EvGS,这是第一个事件辅助3D GS框架,可以沿着相机轨迹从任意视点重建明亮帧。提出了三重级监督,以获得整体知识,颗粒细节和清晰的场景渲染。为了保证渲染帧的颜色一致性,提出了色调匹配块。此外,我们引入了第一个真实捕获的数据集,用于通过基于3D gs的亮度场重建进行事件引导的明亮帧合成任务。实验结果表明,该方法在克服具有挑战性的低光条件下的辐射场重建方面取得了比现有方法更好的效果。代码和样本数据包含在补充资料中。
{"title":"Dark-EvGS: Event Camera as an Eye for Radiance Field in the Dark.","authors":"Jingqian Wu, Peiqi Duan, Zongqiang Wang, Changwei Wang, Boxin Shi, Edmund Y Lam","doi":"10.1109/TIP.2026.3674360","DOIUrl":"https://doi.org/10.1109/TIP.2026.3674360","url":null,"abstract":"<p><p>In low-light environments, conventional cameras often struggle to capture clear multi-view images of objects due to dynamic range limitations and motion blur caused by long exposure. Event cameras, with their high-dynamic range and high-speed properties, have the potential to mitigate these issues. Additionally, 3D Gaussian Splatting (GS) enables radiance field reconstruction, facilitating bright frame synthesis from multiple viewpoints in low-light conditions. However, naively using an event-assisted 3D GS approach still faced challenges because, in low lights, events are noisy, frames lack quality, and the color tone may be inconsistent. To address these issues, we propose Dark-EvGS, the first event-assisted 3D GS framework that enables the reconstruction of bright frames from arbitrary viewpoints along the camera trajectory. Triplet-level supervision is proposed to gain holistic knowledge, granular details, and sharp scene rendering. The color tone matching block is proposed to guarantee the color consistency of the rendered frames. Furthermore, we introduce the first real-captured dataset for the event-guided bright frame synthesis task via 3D GS-based radiance field reconstruction. Experiments demonstrate that our method achieves better results than existing methods, conquering radiance field reconstruction under challenging low-light conditions. The code and sample data are included in the supplementary material.</p>","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":"PP ","pages":""},"PeriodicalIF":13.7,"publicationDate":"2026-03-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147492384","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Long-Tailed and Inter-Class Homogeneity Matters in Multi-Class Weakly Supervised Tissue Segmentation of Histopathology Images 组织病理学图像的多类弱监督组织分割中的长尾和类间同质性问题
IF 13.7 Pub Date : 2026-03-12 DOI: 10.1109/TIP.2026.3671622
Siyang Feng;Xipeng Pan;Huadeng Wang;Zhenbing Liu;Weidong Zhang;Rushi Lan
Using image-level weakly supervised semantic segmentation (WSSS) techniques to segment tissue regions in giga-pixel histopathological whole slide images (WSI) has garnered widespread attention, as it can reduce many annotation workloads for pathologists. Most recent studies are based on class activation mapping (CAM) to generate pseudo masks, which are then used to train segmentation model in a fully supervised manner. However, it is still a challenge to accurately segment non-predominant tissue categories due to the existence of long-tailed and inter-class homogeneity matters. For these matters, we propose three designs to solve them: 1) Diffusion-based Data Generation to synthesis new images of tail class to expand data distribution; 2) Feature Recalibration to reassign the logits in CAM to narrow the feature-level prediction gap between predominant and non-predominant classes; 3) Grade-skip Learning to correct the under-fitting tendency of hard samples during the segmentation phase. Moreover, we also design a powerful pipeline LoHo for histopathology tissue segmentation. Extensive experiments demonstrate that our method not only achieves new state-of-the-art performances but also significantly improves segmentation of tail classes. In addition, our methods are plug-and-play, making it easily integrable into many mainstream WSSS frameworks.
利用图像级弱监督语义分割(WSSS)技术对千兆像素组织病理学全切片图像(WSI)中的组织区域进行分割,可以减少病理学家的许多注释工作量,因此受到了广泛的关注。最近的研究大多是基于类激活映射(CAM)来生成伪掩码,然后用伪掩码以完全监督的方式训练分割模型。然而,由于长尾和类间同质性问题的存在,对非优势组织类别的准确划分仍然是一个挑战。针对这些问题,我们提出了三种解决方案:1)基于扩散的数据生成(Diffusion-based Data Generation),合成新的尾类图像,扩大数据分布;2) Feature Recalibration,重新分配CAM中的logits,以缩小优势类与非优势类之间的特征级预测差距;3)分段学习,纠正硬样本在分割阶段的欠拟合倾向。此外,我们还设计了一个强大的流水线LoHo用于组织病理学组织分割。大量的实验表明,我们的方法不仅达到了新的最先进的性能,而且显著改善了尾类的分割。此外,我们的方法是即插即用的,这使得它很容易集成到许多主流的WSSS框架中。
{"title":"Long-Tailed and Inter-Class Homogeneity Matters in Multi-Class Weakly Supervised Tissue Segmentation of Histopathology Images","authors":"Siyang Feng;Xipeng Pan;Huadeng Wang;Zhenbing Liu;Weidong Zhang;Rushi Lan","doi":"10.1109/TIP.2026.3671622","DOIUrl":"10.1109/TIP.2026.3671622","url":null,"abstract":"Using image-level weakly supervised semantic segmentation (WSSS) techniques to segment tissue regions in giga-pixel histopathological whole slide images (WSI) has garnered widespread attention, as it can reduce many annotation workloads for pathologists. Most recent studies are based on class activation mapping (CAM) to generate pseudo masks, which are then used to train segmentation model in a fully supervised manner. However, it is still a challenge to accurately segment non-predominant tissue categories due to the existence of long-tailed and inter-class homogeneity matters. For these matters, we propose three designs to solve them: 1) Diffusion-based Data Generation to synthesis new images of tail class to expand data distribution; 2) Feature Recalibration to reassign the logits in CAM to narrow the feature-level prediction gap between predominant and non-predominant classes; 3) Grade-skip Learning to correct the under-fitting tendency of hard samples during the segmentation phase. Moreover, we also design a powerful pipeline LoHo for histopathology tissue segmentation. Extensive experiments demonstrate that our method not only achieves new state-of-the-art performances but also significantly improves segmentation of tail classes. In addition, our methods are plug-and-play, making it easily integrable into many mainstream WSSS frameworks.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":"35 ","pages":"2513-2528"},"PeriodicalIF":13.7,"publicationDate":"2026-03-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147439808","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
DiffLLFace: Learning Alternate Illumination-Diffusion Adaptation for Low-Light Face Super-Resolution and Beyond DiffLLFace:学习交替照明扩散适应低光面超分辨率和超越
IF 13.7 Pub Date : 2026-03-12 DOI: 10.1109/TIP.2026.3671638
Runmin Cong;Kaisheng Pang;Feng Li;Hua Li;Huihui Bai;Sam Kwong;Wei Zhang
Facial image acquisition under constrained illumination and with limited-resolution imaging devices often results in coupled photometric and geometric degradations, manifesting as low-light and low-resolution (LLR) conditions. Prevailing research predominantly follows fragmented optimization paradigms that address low-light image enhancement (LLIE) and face super-resolution (FSR) as isolated tasks. This approach overlooks the compound nature of the degradations, thereby significantly limiting their applicability in practical scenarios. To bridge this gap, we present DiffLLFace, a unified framework that harnesses diffusive generative capabilities with illumination-aware trajectories to achieve robust FSR from LLR observations. The core of our method lies in its alternate illumination-diffusion adaptation, which operates throughout the generation process. This mechanism not only captures degradation patterns in both brightness and structure to harmonize latent representations but also dynamically calibrates the illumination prior with the generative knowledge inherent to diffusion models. As such, DiffLLFace attains precise control over conditional adaptation and illumination rectification. We further devise a simple yet effective non-parametric Fourier enhancement strategy, which provides structural appearance clues that work in concert with the alternate adaptation to ensure texture and color consistency. Extensive experiments demonstrate the superiority of DiffLLFace over existing methods and remarkable generalizability on complex natural scenes. Code is available at https://github.com/KaishengPang/DiffLLFace
在受限照明和有限分辨率成像设备下的面部图像采集通常会导致光度和几何图形的耦合退化,表现为低光和低分辨率(LLR)条件。当前的研究主要遵循碎片化的优化范式,将低光图像增强(LLIE)和人脸超分辨率(FSR)作为孤立的任务来解决。这种方法忽略了降解的复合性质,因此大大限制了它们在实际场景中的适用性。为了弥补这一差距,我们提出了DiffLLFace,这是一个统一的框架,利用具有照明感知轨迹的扩散生成能力,从LLR观测中实现鲁棒的FSR。该方法的核心在于其在整个生成过程中运行的交替光照-扩散适应。该机制不仅捕获亮度和结构的退化模式以协调潜在表示,而且还动态地校准具有扩散模型固有的生成知识的照明先验。因此,DiffLLFace实现了对条件适应和光照校正的精确控制。我们进一步设计了一种简单而有效的非参数傅里叶增强策略,该策略提供了结构外观线索,这些线索与替代适应协同工作,以确保纹理和颜色的一致性。大量的实验证明了DiffLLFace相对于现有方法的优越性,在复杂的自然场景上具有显著的泛化能力。代码可从https://github.com/KaishengPang/DiffLLFace获得。
{"title":"DiffLLFace: Learning Alternate Illumination-Diffusion Adaptation for Low-Light Face Super-Resolution and Beyond","authors":"Runmin Cong;Kaisheng Pang;Feng Li;Hua Li;Huihui Bai;Sam Kwong;Wei Zhang","doi":"10.1109/TIP.2026.3671638","DOIUrl":"10.1109/TIP.2026.3671638","url":null,"abstract":"Facial image acquisition under constrained illumination and with limited-resolution imaging devices often results in coupled photometric and geometric degradations, manifesting as low-light and low-resolution (LLR) conditions. Prevailing research predominantly follows fragmented optimization paradigms that address low-light image enhancement (LLIE) and face super-resolution (FSR) as isolated tasks. This approach overlooks the compound nature of the degradations, thereby significantly limiting their applicability in practical scenarios. To bridge this gap, we present DiffLLFace, a unified framework that harnesses diffusive generative capabilities with illumination-aware trajectories to achieve robust FSR from LLR observations. The core of our method lies in its alternate illumination-diffusion adaptation, which operates throughout the generation process. This mechanism not only captures degradation patterns in both brightness and structure to harmonize latent representations but also dynamically calibrates the illumination prior with the generative knowledge inherent to diffusion models. As such, DiffLLFace attains precise control over conditional adaptation and illumination rectification. We further devise a simple yet effective non-parametric Fourier enhancement strategy, which provides structural appearance clues that work in concert with the alternate adaptation to ensure texture and color consistency. Extensive experiments demonstrate the superiority of DiffLLFace over existing methods and remarkable generalizability on complex natural scenes. Code is available at <uri>https://github.com/KaishengPang/DiffLLFace</uri>","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":"35 ","pages":"2499-2512"},"PeriodicalIF":13.7,"publicationDate":"2026-03-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147439809","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
JDPNet: A Network Based on Joint Degradation Processing for Underwater Image Enhancement JDPNet:基于联合退化处理的水下图像增强网络
IF 13.7 Pub Date : 2026-03-12 DOI: 10.1109/TIP.2025.3641833
Tao Ye;Hongbin Ren;Chongbing Zhang;Haoran Chen;Xiaosong Li
Given the complexity of underwater environments and the variability of water as a medium, underwater images are inevitably subject to various types of degradation. The degradations present nonlinear coupling rather than simple superposition, which renders the effective processing of such coupled degradations particularly challenging. Most existing methods focus on designing specific branches, modules, or strategies for specific degradations, with little attention paid to the potential information embedded in their coupling. Consequently, they struggle to effectively capture and process the nonlinear interactions of multiple degradations from a bottom-up perspective. To address this issue, we propose JDPNet, a joint degradation processing network, that mines and unifies the potential information inherent in coupled degradations within a unified framework. Specifically, we introduce a joint feature-mining module, along with a probabilistic bootstrap distribution strategy, to facilitate effective mining and unified adjustment of coupled degradation features. Furthermore, to balance color, clarity, and contrast, we design a novel AquaBalanceLoss to guide the network in learning from multiple coupled degradation losses. Experiments on six publicly available underwater datasets, as well as two new datasets constructed in this study, show that JDPNet exhibits state-of-the-art performance while offering a better tradeoff between performance, parameter size, and computational cost.
鉴于水下环境的复杂性和水作为介质的可变性,水下图像不可避免地会受到各种类型的退化。退化表现为非线性耦合而不是简单的叠加,这使得对这种耦合退化的有效处理特别具有挑战性。大多数现有的方法侧重于为特定的降级设计特定的分支、模块或策略,很少关注它们耦合中嵌入的潜在信息。因此,他们努力从自下而上的角度有效地捕获和处理多重退化的非线性相互作用。为了解决这个问题,我们提出了JDPNet,一个联合退化处理网络,在一个统一的框架内挖掘和统一耦合退化中固有的潜在信息。具体来说,我们引入了一个联合特征挖掘模块,以及一个概率自举分布策略,以促进耦合退化特征的有效挖掘和统一调整。此外,为了平衡颜色、清晰度和对比度,我们设计了一种新的AquaBalanceLoss来指导网络从多重耦合退化损失中学习。在六个公开可用的水下数据集以及本研究中构建的两个新数据集上进行的实验表明,jppnet具有最先进的性能,同时在性能、参数大小和计算成本之间提供了更好的权衡。
{"title":"JDPNet: A Network Based on Joint Degradation Processing for Underwater Image Enhancement","authors":"Tao Ye;Hongbin Ren;Chongbing Zhang;Haoran Chen;Xiaosong Li","doi":"10.1109/TIP.2025.3641833","DOIUrl":"10.1109/TIP.2025.3641833","url":null,"abstract":"Given the complexity of underwater environments and the variability of water as a medium, underwater images are inevitably subject to various types of degradation. The degradations present nonlinear coupling rather than simple superposition, which renders the effective processing of such coupled degradations particularly challenging. Most existing methods focus on designing specific branches, modules, or strategies for specific degradations, with little attention paid to the potential information embedded in their coupling. Consequently, they struggle to effectively capture and process the nonlinear interactions of multiple degradations from a bottom-up perspective. To address this issue, we propose JDPNet, a joint degradation processing network, that mines and unifies the potential information inherent in coupled degradations within a unified framework. Specifically, we introduce a joint feature-mining module, along with a probabilistic bootstrap distribution strategy, to facilitate effective mining and unified adjustment of coupled degradation features. Furthermore, to balance color, clarity, and contrast, we design a novel AquaBalanceLoss to guide the network in learning from multiple coupled degradation losses. Experiments on six publicly available underwater datasets, as well as two new datasets constructed in this study, show that JDPNet exhibits state-of-the-art performance while offering a better tradeoff between performance, parameter size, and computational cost.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":"35 ","pages":"2423-2437"},"PeriodicalIF":13.7,"publicationDate":"2026-03-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147439783","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Nonlinear Transformed Low-Rank Quaternion Tensor Total Variation for Multidimensional Color Image Completion 非线性变换低秩四元数张量全变分的多维彩色图像补全。
IF 13.7 Pub Date : 2026-03-05 DOI: 10.1109/TIP.2026.3666728
Liqiao Yang;Yexun Hu;Tai-Xiang Jiang;Yimin Wei;Guisong Liu;Michael K. Ng
Completing multidimensional color images is a fundamental challenge in image processing and computer vision. However, some tensor-based methods often treat RGB channels as independent modes, thereby neglecting their intrinsic correlations. To address this limitation, we represent RGB values as pure quaternions and organize them into a quaternion tensor for holistic modeling that preserves chromatic relationships. To better capture the nonlinear characteristics inherent in visual data and to improve the compactness of low-rank representations, we propose a nonlinear transformation within the quaternion domain. This design enables more expressive modeling compared to conventional linear approaches. In addition, we introduce two novel regularization terms that jointly encode global low-rankness and local smoothness, with the nonlinear transformation further enhancing the exploitation of structural priors. The overall model is optimized via a nonlinear alternating direction method of multipliers (ADMM), with theoretical guarantees of convergence. Extensive experiments on several datasets demonstrate that the proposed method significantly outperforms state-of-the-art low-rank tensor and quaternion tensor recovery techniques in multidimensional color image completion tasks.
完成多维彩色图像是图像处理和计算机视觉领域的一个基本挑战。然而,一些基于张量的方法通常将RGB通道视为独立模式,从而忽略了它们的内在相关性。为了解决这一限制,我们将RGB值表示为纯四元数,并将它们组织成一个四元数张量,用于保留颜色关系的整体建模。为了更好地捕捉视觉数据固有的非线性特征,并提高低秩表示的紧凑性,我们提出了一种四元数域内的非线性变换。与传统的线性方法相比,这种设计使建模更具表现力。此外,我们引入了两个新的正则化项,它们共同编码全局低秩和局部平滑,通过非线性变换进一步增强了对结构先验的利用。通过非线性交替方向乘法器(ADMM)对整个模型进行了优化,并从理论上保证了模型的收敛性。在多个数据集上进行的大量实验表明,该方法在多维彩色图像补全任务中显著优于最先进的低秩张量和四元数张量恢复技术。
{"title":"Nonlinear Transformed Low-Rank Quaternion Tensor Total Variation for Multidimensional Color Image Completion","authors":"Liqiao Yang;Yexun Hu;Tai-Xiang Jiang;Yimin Wei;Guisong Liu;Michael K. Ng","doi":"10.1109/TIP.2026.3666728","DOIUrl":"10.1109/TIP.2026.3666728","url":null,"abstract":"Completing multidimensional color images is a fundamental challenge in image processing and computer vision. However, some tensor-based methods often treat RGB channels as independent modes, thereby neglecting their intrinsic correlations. To address this limitation, we represent RGB values as pure quaternions and organize them into a quaternion tensor for holistic modeling that preserves chromatic relationships. To better capture the nonlinear characteristics inherent in visual data and to improve the compactness of low-rank representations, we propose a nonlinear transformation within the quaternion domain. This design enables more expressive modeling compared to conventional linear approaches. In addition, we introduce two novel regularization terms that jointly encode global low-rankness and local smoothness, with the nonlinear transformation further enhancing the exploitation of structural priors. The overall model is optimized via a nonlinear alternating direction method of multipliers (ADMM), with theoretical guarantees of convergence. Extensive experiments on several datasets demonstrate that the proposed method significantly outperforms state-of-the-art low-rank tensor and quaternion tensor recovery techniques in multidimensional color image completion tasks.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":"35 ","pages":"2470-2483"},"PeriodicalIF":13.7,"publicationDate":"2026-03-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147359258","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Collaborated With Hallucination: Enhancing Egocentric Grounded Question Answering via Error Demonstrations 与幻觉合作:通过错误演示加强以自我为中心的扎根问题回答。
IF 13.7 Pub Date : 2026-03-04 DOI: 10.1109/TIP.2026.3666732
Shenshen Li;Xing Xu;Fumin Shen;Zhe Sun;Andrzej Cichocki;Heng Tao Shen
The grounded question answering in egocentric videos (Ego-GQA) aims to identify the relevant temporal window and generate corresponding responses in natural language given a textual question. Compared with third-person videos, egocentric video understanding requires more advanced human-centric thinking capability. However, existing Ego-GQA approaches often fail to distinguish the inherent limitations of dynamic egocentric context understanding, treating both first-person and third-person perspectives equally. This oversight leads to hallucinations and a lack of proper egocentric reasoning in first-person video understanding. To address this issue, we propose a novel Collaborated with Hallucination (CoHa) framework for the Ego-GQA, which quantifies the hallucinations generated by an Ego-GQA model and further leverages them as error demonstrations to constrain the model’s reasoning process, encouraging it to ground predictions in egocentric visual cues instead of relying on biased pretraining priors. Specifically, we first employ Subjective Logic to quantify the degree of uncertainty in unreliable answers. We then generate diffusion-based noisy visual inputs to amplify the hallucinations as error demonstrations, which are used to append appropriate constraints to the model according to the uncertainty. These constraints effectively steer predictions away from the unreliable semantics induced by inherent drawbacks in egocentric thinking. Additionally, we incorporate an interactive refinement module to facilitate the model to explore more fine-grained cues observed from the first-person view. Extensive experiments on two widely used benchmarks demonstrate that our CoHa method outperforms recent state-of-the-art methods. Our code is available at https://github.com/Mrshenshen/CoHa
自我中心视频中的基础问答(Ego-GQA)旨在识别相关的时间窗口,并在给定文本问题的情况下用自然语言生成相应的回答。与第三人称视频相比,以自我为中心的视频理解需要更先进的以人为中心的思维能力。然而,现有的Ego-GQA方法往往无法区分动态自我中心上下文理解的固有局限性,即平等地对待第一人称和第三人称视角。在第一人称视频理解中,这种疏忽导致了幻觉和缺乏适当的自我中心推理。为了解决这个问题,我们为Ego-GQA提出了一个新的CoHa框架,该框架量化了Ego-GQA模型产生的幻觉,并进一步利用它们作为错误演示来约束模型的推理过程,鼓励它在以自我为中心的视觉线索中进行预测,而不是依赖于有偏见的预训练先验。具体来说,我们首先采用主观逻辑来量化不可靠答案的不确定性程度。然后,我们生成基于扩散的噪声视觉输入,以放大幻觉作为误差演示,用于根据不确定性向模型附加适当的约束。这些约束有效地引导预测远离由自我中心思维固有缺陷引起的不可靠语义。此外,我们还合并了一个交互式细化模块,以促进模型探索从第一人称视角观察到的更细粒度的线索。在两个广泛使用的基准测试上进行的大量实验表明,我们的CoHa方法优于最近最先进的方法。我们的代码可在https://github.com/Mrshenshen/CoHa上获得。
{"title":"Collaborated With Hallucination: Enhancing Egocentric Grounded Question Answering via Error Demonstrations","authors":"Shenshen Li;Xing Xu;Fumin Shen;Zhe Sun;Andrzej Cichocki;Heng Tao Shen","doi":"10.1109/TIP.2026.3666732","DOIUrl":"10.1109/TIP.2026.3666732","url":null,"abstract":"The grounded question answering in egocentric videos (Ego-GQA) aims to identify the relevant temporal window and generate corresponding responses in natural language given a textual question. Compared with third-person videos, egocentric video understanding requires more advanced human-centric thinking capability. However, existing Ego-GQA approaches often fail to distinguish the inherent limitations of dynamic egocentric context understanding, treating both first-person and third-person perspectives equally. This oversight leads to hallucinations and a lack of proper egocentric reasoning in first-person video understanding. To address this issue, we propose a novel Collaborated with Hallucination (CoHa) framework for the Ego-GQA, which quantifies the hallucinations generated by an Ego-GQA model and further leverages them as error demonstrations to constrain the model’s reasoning process, encouraging it to ground predictions in egocentric visual cues instead of relying on biased pretraining priors. Specifically, we first employ Subjective Logic to quantify the degree of uncertainty in unreliable answers. We then generate diffusion-based noisy visual inputs to amplify the hallucinations as error demonstrations, which are used to append appropriate constraints to the model according to the uncertainty. These constraints effectively steer predictions away from the unreliable semantics induced by inherent drawbacks in egocentric thinking. Additionally, we incorporate an interactive refinement module to facilitate the model to explore more fine-grained cues observed from the first-person view. Extensive experiments on two widely used benchmarks demonstrate that our CoHa method outperforms recent state-of-the-art methods. Our code is available at <uri>https://github.com/Mrshenshen/CoHa</uri>","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":"35 ","pages":"2438-2453"},"PeriodicalIF":13.7,"publicationDate":"2026-03-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147350712","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Unified Framework for Backdoor Trigger Segmentation 后门触发分割的统一框架。
IF 13.7 Pub Date : 2026-03-02 DOI: 10.1109/TIP.2026.3666796
Dizhan Xue;Shengsheng Qian;Changsheng Xu
Recently, backdoor attacks on Deep Neural Networks (DNNs) have raised urgent security threats, which can manipulate the behavior of an attacked model by embedding the backdoor trigger into the input. Since triggers can be designed to be stealthy and hard to recognize by the naked eye, segmenting these triggers in backdoor samples becomes a significant challenge. However, finding triggers embedded by the attacker can be crucial for analyzing the attacks and formulating a defense strategy. Therefore, in this paper, we propose the Backdoor Trigger Segmentation (BTS) task with a comprehensive benchmark consisting of 8 attack methods, 8 unique triggers, and 179 attack settings for image or text data. Moreover, we construct a mathematical system for BTS, abstracting various backdoor triggers into a unified theoretical framework. Based on the theoretical guarantees, we propose a unified Trigger Locator (TriLoc) algorithm to segment various triggers in backdoor samples of both image and text modalities, without prior knowledge of triggers. Extensive experimental results on our benchmark demonstrate the superior performance of our algorithm compared to state-of-the-art methods. Our benchmark and code are available at https://github.com/LivXue/Backdoor-Trigger-Segmentation
最近,针对深度神经网络(dnn)的后门攻击引发了紧迫的安全威胁,这些攻击可以通过将后门触发器嵌入到输入中来操纵被攻击模型的行为。由于触发器可以被设计成隐形的,难以被肉眼识别,因此在后门样本中分割这些触发器成为一个重大挑战。然而,找到攻击者嵌入的触发器对于分析攻击和制定防御策略至关重要。因此,在本文中,我们提出了后门触发分割(BTS)任务,并对图像或文本数据进行了综合基准测试,包括8种攻击方法,8种唯一触发器和179种攻击设置。此外,我们构建了BTS的数学系统,将各种后门触发器抽象成一个统一的理论框架。在理论保证的基础上,我们提出了一种统一的触发器定位器(TriLoc)算法来分割图像和文本模态后门样本中的各种触发器,而无需事先了解触发器。在我们的基准上进行的大量实验结果表明,与最先进的方法相比,我们的算法具有优越的性能。我们的基准测试和代码可在https://github.com/LivXue/Backdoor-Trigger-Segmentation上获得。
{"title":"A Unified Framework for Backdoor Trigger Segmentation","authors":"Dizhan Xue;Shengsheng Qian;Changsheng Xu","doi":"10.1109/TIP.2026.3666796","DOIUrl":"10.1109/TIP.2026.3666796","url":null,"abstract":"Recently, backdoor attacks on Deep Neural Networks (DNNs) have raised urgent security threats, which can manipulate the behavior of an attacked model by embedding the backdoor trigger into the input. Since triggers can be designed to be stealthy and hard to recognize by the naked eye, segmenting these triggers in backdoor samples becomes a significant challenge. However, finding triggers embedded by the attacker can be crucial for analyzing the attacks and formulating a defense strategy. Therefore, in this paper, we propose the Backdoor Trigger Segmentation (BTS) task with a comprehensive benchmark consisting of 8 attack methods, 8 unique triggers, and 179 attack settings for image or text data. Moreover, we construct a mathematical system for BTS, abstracting various backdoor triggers into a unified theoretical framework. Based on the theoretical guarantees, we propose a unified Trigger Locator (TriLoc) algorithm to segment various triggers in backdoor samples of both image and text modalities, without prior knowledge of triggers. Extensive experimental results on our benchmark demonstrate the superior performance of our algorithm compared to state-of-the-art methods. Our benchmark and code are available at <uri>https://github.com/LivXue/Backdoor-Trigger-Segmentation</uri>","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":"35 ","pages":"2364-2379"},"PeriodicalIF":13.7,"publicationDate":"2026-03-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147329218","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Zero-Pose-Prior NeRF: Recursive Radiance Field Reconstruction From Unposed and Unordered Images 零姿态先验NeRF:递归辐射场重建从未定位和无序图像。
IF 13.7 Pub Date : 2026-02-27 DOI: 10.1109/TIP.2026.3666734
Xinxin Liu;Qi Zhang;Xue Wang;Guoqing Zhou;Qing Wang
The dependence of neural radiance fields (NeRF) on accurate camera poses has emerged as a critical obstacle to their widespread real-world applications. While recent advances have demonstrated the potential for simultaneously addressing camera registration and scene reconstruction, these methods inherently rely on reasonable initialization derived from pose or scene priors and struggle with complex scenes involving large camera motions, particularly in unordered 360-degree scenes. In this work, we propose Zero-Pose-Prior NeRF to recover radiance fields from unposed and unordered image collections without any prior knowledge. Our key insight is to decompose this complex problem into smaller sub-problems, wherein the sub-problems’ camera poses are initially estimated to provide self-bootstrapping priors for the global pose estimation, followed by a recursive registration and reconstruction. To achieve this, we first perform scene partitioning to establish a hierarchical structure that describes registration order from local to global. Thereafter, we devise a conditionally-decoupled positional encoding for NeRFs, which serves as the basic model for camera pose estimation and scene representation. Following this, we develop a recursive registration to recursively estimate the poses of local scenes and register them into a unified global pose space, ultimately enabling the reconstruction of the entire scene. Experiments on real-world scenes show that our approach outperforms the state-of-the-art pose-free methods in terms of accurate camera poses and robust radiance field reconstruction, resulting in high-fidelity view synthesis.
神经辐射场(NeRF)对精确相机姿态的依赖已成为其广泛应用于现实世界的关键障碍。虽然最近的进展已经证明了同时解决摄像机注册和场景重建的潜力,但这些方法本质上依赖于从姿势或场景先验中获得的合理初始化,并且难以处理涉及大型摄像机运动的复杂场景,特别是无序的360度场景。在这项工作中,我们提出了零姿态先验NeRF,在没有任何先验知识的情况下从未放置和无序的图像集合中恢复辐射场。我们的关键见解是将这个复杂问题分解成更小的子问题,其中子问题的相机姿态最初估计为全局姿态估计提供自引导先验,然后递归配准和重建。为了实现这一点,我们首先执行场景划分以建立描述从本地到全局注册顺序的分层结构。然后,我们为nerf设计了一种条件解耦的位置编码,作为相机姿态估计和场景表示的基本模型。随后,我们开发了递归配准,递归估计局部场景的姿态,并将其注册到统一的全局姿态空间中,最终实现整个场景的重建。在真实场景的实验表明,我们的方法在准确的相机姿态和鲁棒的辐射场重建方面优于目前最先进的无姿态方法,从而实现高保真的视图合成。
{"title":"Zero-Pose-Prior NeRF: Recursive Radiance Field Reconstruction From Unposed and Unordered Images","authors":"Xinxin Liu;Qi Zhang;Xue Wang;Guoqing Zhou;Qing Wang","doi":"10.1109/TIP.2026.3666734","DOIUrl":"10.1109/TIP.2026.3666734","url":null,"abstract":"The dependence of neural radiance fields (NeRF) on accurate camera poses has emerged as a critical obstacle to their widespread real-world applications. While recent advances have demonstrated the potential for simultaneously addressing camera registration and scene reconstruction, these methods inherently rely on reasonable initialization derived from pose or scene priors and struggle with complex scenes involving large camera motions, particularly in unordered 360-degree scenes. In this work, we propose Zero-Pose-Prior NeRF to recover radiance fields from unposed and unordered image collections without any prior knowledge. Our key insight is to decompose this complex problem into smaller sub-problems, wherein the sub-problems’ camera poses are initially estimated to provide self-bootstrapping priors for the global pose estimation, followed by a recursive registration and reconstruction. To achieve this, we first perform scene partitioning to establish a hierarchical structure that describes registration order from local to global. Thereafter, we devise a conditionally-decoupled positional encoding for NeRFs, which serves as the basic model for camera pose estimation and scene representation. Following this, we develop a recursive registration to recursively estimate the poses of local scenes and register them into a unified global pose space, ultimately enabling the reconstruction of the entire scene. Experiments on real-world scenes show that our approach outperforms the state-of-the-art pose-free methods in terms of accurate camera poses and robust radiance field reconstruction, resulting in high-fidelity view synthesis.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":"35 ","pages":"2320-2334"},"PeriodicalIF":13.7,"publicationDate":"2026-02-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147319102","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
ReconX: Reconstruct Any Scene From Sparse Views With Video Diffusion Model ReconX:用视频扩散模型从稀疏视图重建任何场景。
IF 13.7 Pub Date : 2026-02-27 DOI: 10.1109/TIP.2026.3666733
Fangfu Liu;Wenqiang Sun;Hanyang Wang;Yikai Wang;Haowen Sun;Junliang Ye;Jun Zhang;Yueqi Duan
Advancements in 3D scene reconstruction have transformed 2D images from the real world into 3D models, producing realistic 3D results from hundreds of input photos. Despite great success in dense-view reconstruction scenarios, rendering a detailed scene from sparse views is still an ill-posed optimization problem, often resulting in artifacts and distortions in unseen areas. In this paper, we propose ReconX, a novel 3D scene reconstruction paradigm that reframes the ambiguous reconstruction problem as a temporal generation task. The key insight is to unleash the strong generative prior of large pre-trained video diffusion models for sparse-view reconstruction. Nevertheless, it is challenging to preserve 3D view consistency when directly generating video frames from pre-trained models. To address this issue, given limited input views, the proposed ReconX first constructs a global point cloud and encodes it into a contextual space as the 3D structure condition. Guided by the condition, the video diffusion model then synthesizes video frames that are detail-preserved and exhibit a high degree of 3D consistency, ensuring the coherence of the scene from various perspectives. Finally, we recover the 3D scene from the generated video through a confidence-aware 3D Gaussian Splatting optimization scheme. Extensive experiments on various real-world datasets show the superiority of ReconX over state-of-the-art methods in terms of quality and generalizability.
3D场景重建的进步已经将现实世界的2D图像转换为3D模型,从数百张输入照片中产生逼真的3D结果。尽管在密集视图重建场景中取得了巨大的成功,但从稀疏视图中绘制详细场景仍然是一个不适定优化问题,经常导致未见区域的伪影和扭曲。在本文中,我们提出了一种新的三维场景重建范式ReconX,它将模糊重建问题重新定义为一个时间生成任务。关键的见解是释放出用于稀疏视图重建的大型预训练视频扩散模型的强生成先验。然而,当直接从预训练模型生成视频帧时,保持3D视图一致性是具有挑战性的。为了解决这个问题,在输入视图有限的情况下,提出的ReconX首先构建一个全局点云,并将其编码为上下文空间作为3D结构条件。在条件的引导下,视频扩散模型合成的视频帧保留了细节,呈现出高度的3D一致性,保证了场景从各个角度的连贯性。最后,我们通过置信度感知的三维高斯飞溅优化方案从生成的视频中恢复三维场景。在各种真实数据集上进行的大量实验表明,ReconX在质量和通用性方面优于最先进的方法。
{"title":"ReconX: Reconstruct Any Scene From Sparse Views With Video Diffusion Model","authors":"Fangfu Liu;Wenqiang Sun;Hanyang Wang;Yikai Wang;Haowen Sun;Junliang Ye;Jun Zhang;Yueqi Duan","doi":"10.1109/TIP.2026.3666733","DOIUrl":"10.1109/TIP.2026.3666733","url":null,"abstract":"Advancements in 3D scene reconstruction have transformed 2D images from the real world into 3D models, producing realistic 3D results from hundreds of input photos. Despite great success in dense-view reconstruction scenarios, rendering a detailed scene from sparse views is still an ill-posed optimization problem, often resulting in artifacts and distortions in unseen areas. In this paper, we propose ReconX, a novel 3D scene reconstruction paradigm that reframes the ambiguous reconstruction problem as a temporal generation task. The key insight is to unleash the strong generative prior of large pre-trained video diffusion models for sparse-view reconstruction. Nevertheless, it is challenging to preserve 3D view consistency when directly generating video frames from pre-trained models. To address this issue, given limited input views, the proposed ReconX first constructs a global point cloud and encodes it into a contextual space as the 3D structure condition. Guided by the condition, the video diffusion model then synthesizes video frames that are detail-preserved and exhibit a high degree of 3D consistency, ensuring the coherence of the scene from various perspectives. Finally, we recover the 3D scene from the generated video through a confidence-aware 3D Gaussian Splatting optimization scheme. Extensive experiments on various real-world datasets show the superiority of ReconX over state-of-the-art methods in terms of quality and generalizability.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":"35 ","pages":"2305-2319"},"PeriodicalIF":13.7,"publicationDate":"2026-02-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147319129","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Multi-Level Self-Distillation-Based Unified Tracker for Efficient RGB-T Tracking 基于多级自蒸馏的RGB-T高效跟踪统一跟踪器。
IF 13.7 Pub Date : 2026-02-27 DOI: 10.1109/TIP.2026.3666737
Mohamed Awad;Ahmed Elliethy;M. Omair Ahmad;M. N. S. Swamy
RGB-Thermal (RGB-T) tracking enhances visual tracking robustness by combining RGB and thermal infrared (TIR) modalities, addressing limitations of RGB-only trackers under challenging conditions such as low light and appearance variations. However, most existing RGB-T trackers rely on complex fusion modules or modality-specific architectures, sacrificing efficiency for performance. In this paper, we propose a novel Multi-level Self-Distillation (MSD) framework that adapts a one-stream RGB tracker to the RGB-T setting without modifying the network architecture or adding any extra parameters. RGB and TIR inputs are jointly processed through a shared backbone, and training is guided by a combination of self-supervised and supervised objectives to enhance cross-modal feature representation. The self-supervised component includes a contrastive loss that aligns semantically consistent regions across template-search pairs, as well as a modality-gap alignment loss that reduces discrepancies between RGB and TIR features. These internal signals complement task-driven supervision, including an intermediate focal loss that strengthens early localization by enhancing shallow and mid-level features, modality-specific losses that preserve distinctive cues under partial modality degradation, and a fused tracking loss that drives final bounding box prediction. Comprehensive evaluations on LasHeR, RGBT234, and GTOT benchmarks demonstrate that MSD achieves state-of-the-art tracking accuracy while maintaining the computational efficiency of the original RGB tracker. Our work establishes a new paradigm in multi-modal tracking by demonstrating that optimized training strategies can outperform complex architectural modifications, offering significant practical advantages for real-world deployment.
RGB-热(RGB- t)跟踪通过结合RGB和热红外(TIR)模式增强了视觉跟踪的鲁棒性,解决了仅RGB跟踪器在低光和外观变化等具有挑战性的条件下的局限性。然而,大多数现有的RGB-T跟踪器依赖于复杂的融合模块或特定于模态的架构,从而牺牲了性能的效率。在本文中,我们提出了一种新的多层自蒸馏(MSD)框架,该框架在不修改网络架构或添加任何额外参数的情况下,使单流RGB跟踪器适应RGB- t设置。RGB和TIR输入通过共享主干联合处理,训练由自监督目标和监督目标相结合指导,以增强跨模态特征表示。自监督组件包括在模板搜索对中对齐语义一致区域的对比损失,以及减少RGB和TIR特征之间差异的模态间隙对齐损失。这些内部信号补充了任务驱动的监督,包括通过增强浅层和中层特征来加强早期定位的中间焦点丢失,在部分模态退化下保留独特线索的模态特异性丢失,以及驱动最终边界盒预测的融合跟踪丢失。对LasHeR, RGBT234和GTOT基准的综合评估表明,MSD在保持原始RGB跟踪器的计算效率的同时实现了最先进的跟踪精度。我们的工作通过证明优化的训练策略可以胜过复杂的架构修改,为现实世界的部署提供了显著的实际优势,从而建立了多模态跟踪的新范例。
{"title":"A Multi-Level Self-Distillation-Based Unified Tracker for Efficient RGB-T Tracking","authors":"Mohamed Awad;Ahmed Elliethy;M. Omair Ahmad;M. N. S. Swamy","doi":"10.1109/TIP.2026.3666737","DOIUrl":"10.1109/TIP.2026.3666737","url":null,"abstract":"RGB-Thermal (RGB-T) tracking enhances visual tracking robustness by combining RGB and thermal infrared (TIR) modalities, addressing limitations of RGB-only trackers under challenging conditions such as low light and appearance variations. However, most existing RGB-T trackers rely on complex fusion modules or modality-specific architectures, sacrificing efficiency for performance. In this paper, we propose a novel Multi-level Self-Distillation (MSD) framework that adapts a one-stream RGB tracker to the RGB-T setting without modifying the network architecture or adding any extra parameters. RGB and TIR inputs are jointly processed through a shared backbone, and training is guided by a combination of self-supervised and supervised objectives to enhance cross-modal feature representation. The self-supervised component includes a contrastive loss that aligns semantically consistent regions across template-search pairs, as well as a modality-gap alignment loss that reduces discrepancies between RGB and TIR features. These internal signals complement task-driven supervision, including an intermediate focal loss that strengthens early localization by enhancing shallow and mid-level features, modality-specific losses that preserve distinctive cues under partial modality degradation, and a fused tracking loss that drives final bounding box prediction. Comprehensive evaluations on LasHeR, RGBT234, and GTOT benchmarks demonstrate that MSD achieves state-of-the-art tracking accuracy while maintaining the computational efficiency of the original RGB tracker. Our work establishes a new paradigm in multi-modal tracking by demonstrating that optimized training strategies can outperform complex architectural modifications, offering significant practical advantages for real-world deployment.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":"35 ","pages":"2407-2422"},"PeriodicalIF":13.7,"publicationDate":"2026-02-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147319116","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE transactions on image processing : a publication of the IEEE Signal Processing Society
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1