首页 > 最新文献

IEEE transactions on image processing : a publication of the IEEE Signal Processing Society最新文献

英文 中文
MTRAG: Multi-Target Referring and Grounding via Hybrid Semantic-Spatial Integration 基于混合语义-空间集成的多目标引用与接地
IF 13.7 Pub Date : 2026-02-16 DOI: 10.1109/TIP.2026.3663039
Yili Ren;Jinyang Du;Xi Liu;Qianxiao Su;Yue Deng;Hongjue Li
Fine-grained visual referring and grounding are critical for enhancing scene understanding and enabling various real-world vision-language applications. Although recent studies have extended multimodal large language models (MLLMs) to these tasks, they still face significant challenges in fine-grained multi-target scenarios. To address this, we propose MTRAG, a pixel-level multi-target referring and grounding framework that leverages semantic-spatial collaboration. Specifically, we introduce a Channel Extension Mechanism (CEM) that enables a global image encoder to extract global semantics and multi-region representations while retaining background context, without extra region feature extractors. Moreover, we introduce a grounding branch for pixel-level grounding and design a Hybrid Adapter (HA) to fuse semantic features from the MLLM branch with spatial information from the grounding branch, thereby enhancing the semantic-spatial alignment. For training, we meticulously curate MTRAG-D, a dataset comprising single- and multi-target referring and grounding samples derived from existing datasets and newly synthesized free-form multi-target referring instruction-following data. We also present MTR-Bench, a benchmark for systematic evaluation of multi-target referring. Extensive experiments across five core tasks, including single- and multi-target referring and grounding as well as image-level captioning, show that MTRAG consistently outperforms strong baselines on both multi- and single-target tasks, while maintaining competitive image-level understanding. The code is available at https://github.com/deng-ai-lab/MTRAG
细粒度的视觉参考和基础对于增强场景理解和实现各种现实世界的视觉语言应用至关重要。尽管最近的研究已经将多模态大语言模型(mllm)扩展到这些任务中,但它们在细粒度多目标场景中仍然面临重大挑战。为了解决这个问题,我们提出了MTRAG,这是一个利用语义空间协作的像素级多目标参考和基础框架。具体来说,我们引入了通道扩展机制(CEM),该机制使全局图像编码器能够在保留背景上下文的同时提取全局语义和多区域表示,而无需额外的区域特征提取器。此外,我们引入了一个用于像素级接地的接地支路,并设计了一个混合适配器(Hybrid Adapter, HA),将MLLM支路的语义特征与接地支路的空间信息融合在一起,从而增强了语义-空间对齐。对于训练,我们精心策划了MTRAG-D,这是一个数据集,包括来自现有数据集的单目标和多目标参考和接地样本,以及新合成的自由形式多目标参考指令遵循数据。我们还提出了MTR-Bench,一个系统评价多目标参考的基准。在五个核心任务(包括单目标和多目标参考和基础以及图像级字幕)上进行的广泛实验表明,MTRAG在多目标和单目标任务上的表现始终优于强基线,同时保持了具有竞争力的图像级理解。代码可在https://github.com/deng-ai-lab/MTRAG上获得
{"title":"MTRAG: Multi-Target Referring and Grounding via Hybrid Semantic-Spatial Integration","authors":"Yili Ren;Jinyang Du;Xi Liu;Qianxiao Su;Yue Deng;Hongjue Li","doi":"10.1109/TIP.2026.3663039","DOIUrl":"10.1109/TIP.2026.3663039","url":null,"abstract":"Fine-grained visual referring and grounding are critical for enhancing scene understanding and enabling various real-world vision-language applications. Although recent studies have extended multimodal large language models (MLLMs) to these tasks, they still face significant challenges in fine-grained multi-target scenarios. To address this, we propose MTRAG, a pixel-level multi-target referring and grounding framework that leverages semantic-spatial collaboration. Specifically, we introduce a Channel Extension Mechanism (CEM) that enables a global image encoder to extract global semantics and multi-region representations while retaining background context, without extra region feature extractors. Moreover, we introduce a grounding branch for pixel-level grounding and design a Hybrid Adapter (HA) to fuse semantic features from the MLLM branch with spatial information from the grounding branch, thereby enhancing the semantic-spatial alignment. For training, we meticulously curate MTRAG-D, a dataset comprising single- and multi-target referring and grounding samples derived from existing datasets and newly synthesized free-form multi-target referring instruction-following data. We also present MTR-Bench, a benchmark for systematic evaluation of multi-target referring. Extensive experiments across five core tasks, including single- and multi-target referring and grounding as well as image-level captioning, show that MTRAG consistently outperforms strong baselines on both multi- and single-target tasks, while maintaining competitive image-level understanding. The code is available at <uri>https://github.com/deng-ai-lab/MTRAG</uri>","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":"35 ","pages":"2167-2181"},"PeriodicalIF":13.7,"publicationDate":"2026-02-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146204870","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Interactive Spatial-Frequency Fusion Mamba for Multi-Modal Image Fusion 用于多模态图像融合的交互式空频融合曼巴
IF 13.7 Pub Date : 2026-02-16 DOI: 10.1109/TIP.2026.3662596
Yixin Zhu;Long Lv;Pingping Zhang;Xuehu Liu;Tongdan Tang;Feng Tian;Weibing Sun;Huchuan Lu
Multi-Modal Image Fusion (MMIF) aims to combine images from different modalities to produce fused images, retaining texture details and preserving significant information. Recently, some MMIF methods incorporate frequency domain information to enhance spatial features. However, these methods typically rely on simple serial or parallel spatial-frequency fusion without interaction. In this paper, we propose a novel Interactive Spatial-Frequency Fusion Mamba (ISFM) framework for MMIF. Specifically, we begin with a Modality-Specific Extractor (MSE) to extract features from different modalities. It models long-range dependencies across the image with linear computational complexity. To effectively leverage frequency information, we then propose a Multi-scale Frequency Fusion (MFF). It adaptively integrates low-frequency and high-frequency components across multiple scales, enabling robust representations of frequency features. More importantly, we further propose an Interactive Spatial-Frequency Fusion (ISF). It incorporates frequency features to guide spatial features across modalities, enhancing complementary representations. Extensive experiments are conducted on six MMIF datasets. The experimental results demonstrate that our ISFM can achieve better performances than other state-of-the-art methods. The source code is available at https://github.com/Namn23/ISFM.
多模态图像融合(MMIF)旨在将来自不同模态的图像组合在一起产生融合图像,同时保留纹理细节和重要信息。近年来,一些MMIF方法引入了频域信息来增强空间特征。然而,这些方法通常依赖于简单的串行或并行空间频率融合,而没有相互作用。在本文中,我们提出了一种新的交互式空频融合曼巴(ISFM)框架。具体来说,我们从一个模态特定提取器(MSE)开始,从不同的模态中提取特征。它以线性计算复杂度对图像上的长期依赖关系进行建模。为了有效地利用频率信息,我们提出了一种多尺度频率融合(MFF)。它自适应地集成了多个尺度的低频和高频组件,实现了频率特征的鲁棒表示。更重要的是,我们进一步提出了一种交互式空间-频率融合(ISF)。它结合了频率特征来指导跨模态的空间特征,增强了互补表示。在六个MMIF数据集上进行了大量实验。实验结果表明,我们的ISFM比其他先进的方法具有更好的性能。源代码可从https://github.com/Namn23/ISFM获得。
{"title":"Interactive Spatial-Frequency Fusion Mamba for Multi-Modal Image Fusion","authors":"Yixin Zhu;Long Lv;Pingping Zhang;Xuehu Liu;Tongdan Tang;Feng Tian;Weibing Sun;Huchuan Lu","doi":"10.1109/TIP.2026.3662596","DOIUrl":"10.1109/TIP.2026.3662596","url":null,"abstract":"Multi-Modal Image Fusion (MMIF) aims to combine images from different modalities to produce fused images, retaining texture details and preserving significant information. Recently, some MMIF methods incorporate frequency domain information to enhance spatial features. However, these methods typically rely on simple serial or parallel spatial-frequency fusion without interaction. In this paper, we propose a novel Interactive Spatial-Frequency Fusion Mamba (ISFM) framework for MMIF. Specifically, we begin with a Modality-Specific Extractor (MSE) to extract features from different modalities. It models long-range dependencies across the image with linear computational complexity. To effectively leverage frequency information, we then propose a Multi-scale Frequency Fusion (MFF). It adaptively integrates low-frequency and high-frequency components across multiple scales, enabling robust representations of frequency features. More importantly, we further propose an Interactive Spatial-Frequency Fusion (ISF). It incorporates frequency features to guide spatial features across modalities, enhancing complementary representations. Extensive experiments are conducted on six MMIF datasets. The experimental results demonstrate that our ISFM can achieve better performances than other state-of-the-art methods. The source code is available at <uri>https://github.com/Namn23/ISFM</uri>.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":"35 ","pages":"2380-2392"},"PeriodicalIF":13.7,"publicationDate":"2026-02-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146204871","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
DeepGSR: Deep Group-Based Sparse Representation Network for Solving Image Inverse Problems DeepGSR:用于解决图像逆问题的基于深度组的稀疏表示网络。
IF 13.7 Pub Date : 2026-02-13 DOI: 10.1109/TIP.2026.3662583
Ke Jiang;Xinya Ji;Baoshun Shi
In the past few years, group-based sparse representation (GSR) has emerged as a powerful paradigm for image inverse problems by synergizing model-driven interpretability with nonlocal self-similarity priors. Nevertheless, its practical utility is hindered by computationally expensive iterative processes. Deep learning (DL) methods can avoid this deficiency, but they often lack of model interpretability. To bridge this gap, we propose a novel deep group-based sparse representation framework, termed DeepGSR, which brings the GSR method and the DL approach together. DeepGSR not only circumvents the iterative bottlenecks of conventional GSR but also preserves its model interpretability through a learnable parameterization. Specifically, the network is built upon a GSR model that leverages nonlocal self-similarity, and it integrates adaptive patch matching and aggregation mechanisms to model complex intra-group relationships in the latent space. To reduce the computational complexity associated with traditional SVD-based rank shrinkage, we introduce a learnable low-rank shrinkage module that incorporates low-rank constraints while enhancing the interpretability and adaptability of the model. To better exploit frequency-specific structures, the network incorporates a shifting wavelet-domain patch partitioning strategy, which separately models high- and low-frequency components to further enhance the representation ability of the network. Extensive experiments demonstrate that DeepGSR, when applied as a drop-in replacement module to various image inverse problems such as image denoising, single-image deraining, metal artifact reduction, sparse-view computed tomography reconstruction, phase retrieval, and all-in-one image restoration consistently delivers effective performance, validating the effectiveness of the proposed framework. The source code and datasets have been made publicly available at https://github.com/shibaoshun/DeepGSR
在过去的几年中,基于组的稀疏表示(GSR)通过将模型驱动的可解释性与非局部自相似先验相结合,成为图像反演问题的一个强大范例。然而,它的实际应用受到计算昂贵的迭代过程的阻碍。深度学习(DL)方法可以避免这一缺陷,但它们往往缺乏模型可解释性。为了弥补这一差距,我们提出了一种新的基于深度组的稀疏表示框架,称为DeepGSR,它将GSR方法和深度学习方法结合在一起。DeepGSR不仅克服了传统GSR的迭代瓶颈,而且通过可学习的参数化保持了模型的可解释性。具体而言,该网络建立在利用非局部自相似的GSR模型之上,并集成了自适应补丁匹配和聚集机制来模拟潜在空间中复杂的组内关系。为了降低传统基于奇异值分解的秩缩模型的计算复杂度,我们引入了一个可学习的低秩缩模型,该模型结合了低秩约束,同时增强了模型的可解释性和适应性。为了更好地利用频率特异性结构,该网络采用了移动小波域补丁划分策略,分别对高频和低频分量建模,进一步增强了网络的表示能力。大量的实验表明,当DeepGSR作为一个替代模块应用于各种图像逆问题时,如图像去噪、单图像脱噪、金属伪影还原、稀疏视图计算机断层扫描重建、相位检索和一体化图像恢复,始终提供有效的性能,验证了所提出框架的有效性。源代码和数据集已在https://github.com/shibaoshun/DeepGSR上公开提供。
{"title":"DeepGSR: Deep Group-Based Sparse Representation Network for Solving Image Inverse Problems","authors":"Ke Jiang;Xinya Ji;Baoshun Shi","doi":"10.1109/TIP.2026.3662583","DOIUrl":"10.1109/TIP.2026.3662583","url":null,"abstract":"In the past few years, group-based sparse representation (GSR) has emerged as a powerful paradigm for image inverse problems by synergizing model-driven interpretability with nonlocal self-similarity priors. Nevertheless, its practical utility is hindered by computationally expensive iterative processes. Deep learning (DL) methods can avoid this deficiency, but they often lack of model interpretability. To bridge this gap, we propose a novel deep group-based sparse representation framework, termed DeepGSR, which brings the GSR method and the DL approach together. DeepGSR not only circumvents the iterative bottlenecks of conventional GSR but also preserves its model interpretability through a learnable parameterization. Specifically, the network is built upon a GSR model that leverages nonlocal self-similarity, and it integrates adaptive patch matching and aggregation mechanisms to model complex intra-group relationships in the latent space. To reduce the computational complexity associated with traditional SVD-based rank shrinkage, we introduce a learnable low-rank shrinkage module that incorporates low-rank constraints while enhancing the interpretability and adaptability of the model. To better exploit frequency-specific structures, the network incorporates a shifting wavelet-domain patch partitioning strategy, which separately models high- and low-frequency components to further enhance the representation ability of the network. Extensive experiments demonstrate that DeepGSR, when applied as a drop-in replacement module to various image inverse problems such as image denoising, single-image deraining, metal artifact reduction, sparse-view computed tomography reconstruction, phase retrieval, and all-in-one image restoration consistently delivers effective performance, validating the effectiveness of the proposed framework. The source code and datasets have been made publicly available at <uri>https://github.com/shibaoshun/DeepGSR</uri>","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":"35 ","pages":"2454-2469"},"PeriodicalIF":13.7,"publicationDate":"2026-02-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146195774","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
ASDTracker: Adaptively Sparse Detection With Attention-Guided Refinement for Efficient Multi-Object Tracking ASDTracker:基于注意力引导的自适应稀疏检测,用于高效的多目标跟踪。
IF 13.7 Pub Date : 2026-02-13 DOI: 10.1109/TIP.2026.3662594
Yueying Wang;Chenyang Yan;Cairong Zhao;Weidong Zhang;Dan Zeng
Tracking-by-Detection paradigms shine in generic multi-object tracking (MOT), while their compact construction hinders the real-time applications. In this work, we attribute the substantial computational burden to two expensive components, i.e. detection and re-identification. Building upon the principle of adaptively maintaining acceptable inference efficiency, we present Adaptively Sparse Detection with attention-guided refinement (ASDTracker) for efficient tracking. In specific, our ASDTracker rapidly assess the short-term and long-term occlusion, dynamically determining the usage of the expensive detector. For non-key frames, we efficiently refine small-size crops out of Kalman Filter predictions and introduce the noisy shadow labels to robustly train this refinement network. Additionally, we substitute the lightweight appearance representation for the heavy ReID network, which efficiently extracts sufficient appearance cues in the coarsely quantized color spaces. Extensive experiments on four benchmarks demonstrate that ASDTracker achieves competitive performance in generalization and robustness under favorable inference speed. Moreover, the efficient tracking deployment is further implemented to an unmanned surface vehicle with high accuracy and low latency in real-world scenarios.
基于检测的跟踪模式在多目标跟踪(MOT)中发挥着重要作用,但其紧凑的结构阻碍了实时应用。在这项工作中,我们将大量的计算负担归因于两个昂贵的组件,即检测和重新识别。基于自适应保持可接受的推理效率的原则,我们提出了带有注意引导改进的自适应稀疏检测(ASDTracker)来进行有效的跟踪。具体而言,我们的ASDTracker快速评估短期和长期遮挡,动态确定昂贵检测器的使用情况。对于非关键帧,我们有效地从卡尔曼滤波预测中细化小尺寸的作物,并引入噪声阴影标签来鲁棒训练该细化网络。此外,我们用轻量的外观表示代替重量的ReID网络,有效地在粗量化的颜色空间中提取足够的外观线索。在四个基准测试上的大量实验表明,在良好的推理速度下,ASDTracker在泛化和鲁棒性方面具有竞争力。在实际场景中,进一步实现了对无人水面车辆高精度、低时延的高效跟踪部署。
{"title":"ASDTracker: Adaptively Sparse Detection With Attention-Guided Refinement for Efficient Multi-Object Tracking","authors":"Yueying Wang;Chenyang Yan;Cairong Zhao;Weidong Zhang;Dan Zeng","doi":"10.1109/TIP.2026.3662594","DOIUrl":"10.1109/TIP.2026.3662594","url":null,"abstract":"Tracking-by-Detection paradigms shine in generic multi-object tracking (MOT), while their compact construction hinders the real-time applications. In this work, we attribute the substantial computational burden to two expensive components, i.e. detection and re-identification. Building upon the principle of adaptively maintaining acceptable inference efficiency, we present Adaptively Sparse Detection with attention-guided refinement (ASDTracker) for efficient tracking. In specific, our ASDTracker rapidly assess the short-term and long-term occlusion, dynamically determining the usage of the expensive detector. For non-key frames, we efficiently refine small-size crops out of Kalman Filter predictions and introduce the noisy shadow labels to robustly train this refinement network. Additionally, we substitute the lightweight appearance representation for the heavy ReID network, which efficiently extracts sufficient appearance cues in the coarsely quantized color spaces. Extensive experiments on four benchmarks demonstrate that ASDTracker achieves competitive performance in generalization and robustness under favorable inference speed. Moreover, the efficient tracking deployment is further implemented to an unmanned surface vehicle with high accuracy and low latency in real-world scenarios.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":"35 ","pages":"1993-2006"},"PeriodicalIF":13.7,"publicationDate":"2026-02-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146195486","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Robust 2.5D Feature Matching in Light Fields via a Learnable Parameterized Depth-Degraded Projection 基于可学习参数化深度退化投影的光场鲁棒2.5D特征匹配。
IF 13.7 Pub Date : 2026-02-13 DOI: 10.1109/TIP.2026.3662579
Meng Zhang;Haiyan Jin;Zhaolin Xiao;Jinglei Shi;Xiaoran Jiang
Due to the loss of 3D information, accurate and robust 2D image feature matching remains challenging for many computer vision applications. This paper introduces a 2.5D feature that uses the disparity value from the light field Fourier disparity layer (FDL) as a rough proxy of scene depth. Without explicit depth estimation, a parameterized depth-degraded projection is proposed to construct the geometric transformation of paired features between two light fields. Then, we propose a parameterized learning solution to calculate the depth-degraded projection. This solution estimates a global constant fundamental matrix, a variable disparity-guided translation vector, and a depth compensation term using a very simple network. Although the 0.5D relative disparity provided by the FDL does not represent precise depth, it can also significantly reduce the depth ambiguity in feature matching. Therefore, the proposed solution achieves accurate feature-matching results by minimizing the sum of reprojection errors across all matching candidates. On the public light field feature-matching dataset, the proposed solution outperforms existing 2D image feature-matching solutions and light field feature-matching algorithms in terms of matching accuracy and robustness. The code is available online.
由于三维信息的丢失,精确和鲁棒的二维图像特征匹配仍然是许多计算机视觉应用的挑战。本文介绍了一种利用光场傅立叶视差层(FDL)的视差值作为场景深度的粗略代理的2.5D特征。在没有显式深度估计的情况下,提出了一种参数化深度退化投影来构造两个光场之间成对特征的几何变换。然后,我们提出了一种参数化学习方法来计算深度退化投影。该解决方案估计一个全局常数基本矩阵,一个变量差分引导平移向量,并使用一个非常简单的网络深度补偿项。FDL提供的0.5D相对视差虽然不能代表精确的深度,但也能显著降低特征匹配中的深度模糊度。因此,所提出的解决方案通过最小化所有匹配候选者的重投影误差之和来获得准确的特征匹配结果。在公共光场特征匹配数据集上,该方案在匹配精度和鲁棒性方面优于现有二维图像特征匹配方案和光场特征匹配算法。代码可以在网上找到。
{"title":"Robust 2.5D Feature Matching in Light Fields via a Learnable Parameterized Depth-Degraded Projection","authors":"Meng Zhang;Haiyan Jin;Zhaolin Xiao;Jinglei Shi;Xiaoran Jiang","doi":"10.1109/TIP.2026.3662579","DOIUrl":"10.1109/TIP.2026.3662579","url":null,"abstract":"Due to the loss of 3D information, accurate and robust 2D image feature matching remains challenging for many computer vision applications. This paper introduces a 2.5D feature that uses the disparity value from the light field Fourier disparity layer (FDL) as a rough proxy of scene depth. Without explicit depth estimation, a parameterized depth-degraded projection is proposed to construct the geometric transformation of paired features between two light fields. Then, we propose a parameterized learning solution to calculate the depth-degraded projection. This solution estimates a global constant fundamental matrix, a variable disparity-guided translation vector, and a depth compensation term using a very simple network. Although the 0.5D relative disparity provided by the FDL does not represent precise depth, it can also significantly reduce the depth ambiguity in feature matching. Therefore, the proposed solution achieves accurate feature-matching results by minimizing the sum of reprojection errors across all matching candidates. On the public light field feature-matching dataset, the proposed solution outperforms existing 2D image feature-matching solutions and light field feature-matching algorithms in terms of matching accuracy and robustness. The code is available online.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":"35 ","pages":"2235-2248"},"PeriodicalIF":13.7,"publicationDate":"2026-02-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146195721","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
C-WOE: Clustering for Out-of-Distribution Detection Learning With Wild Outlier Exposure C-WOE:具有野离群值暴露的分布外检测学习聚类。
IF 13.7 Pub Date : 2026-02-13 DOI: 10.1109/TIP.2026.3662593
Long Lan;Zhaohui Hu;He Li;Tongliang Liu;Xinwang Liu
Out-of-distribution (OOD) detection plays a crucial role as a mechanism for handling anomalies in computer vision systems. Among existing approaches, outlier exposure (OE), which trains the model with an additional auxiliary OOD dataset, has demonstrated strong effectiveness. However, acquiring clean and well-curated auxiliary OOD data is often infeasible, particularly within large and complex systems. Alternatively, wild outliers, i.e., unlabeled samples collected directly in deployment environments, are abundant and easy to obtain, and recent studies have shown that they can substantially benefit OOD detection learning. Nevertheless, wild outliers typically contain a mixture of in-distribution (ID) and OOD samples. Directly using them as auxiliary OOD data unavoidably exposes the model to adverse supervision signals arising from the contained ID samples. Yet existing methods still lack an effective strategy that can fully leverage wild outliers while suppressing the negative influence introduced by their ID subset. To this end, we propose a simple yet effective method named Clustering for Wild Outlier Exposure (C-WOE), which alleviates the adverse effect of the ID samples contained within wild outliers by reweighting them. Specifically, C-WOE assigns higher weights to real OOD samples and lower weights to ID samples and dynamically updates these weights during training. Theoretically, we establish solid guarantees for the proposed method. Empirically, extensive experiments conducted on various real-world benchmarks and simulated datasets demonstrate that C-WOE notably achieves superior performance compared with state-of-the-art methods, validating its reliability in image processing applications.
在计算机视觉系统中,超分布(OOD)检测作为一种异常处理机制起着至关重要的作用。在现有的方法中,使用额外的辅助OOD数据集训练模型的离群值暴露(OE)已经证明了很强的有效性。然而,获取干净和精心策划的辅助OOD数据通常是不可行的,特别是在大型和复杂的系统中。另外,野生异常值,即在部署环境中直接收集的未标记样本,丰富且易于获得,最近的研究表明,它们可以极大地促进OOD检测学习。然而,野生异常值通常包含分布内(ID)和良好样本的混合物。直接使用它们作为辅助OOD数据,不可避免地会使模型暴露于由所含ID样本产生的不利监督信号。然而,现有的方法仍然缺乏一种有效的策略,可以充分利用野生异常值,同时抑制其ID子集引入的负面影响。为此,我们提出了一种简单而有效的方法,即聚类野生离群暴露(C-WOE),该方法通过重新加权来减轻野生离群中包含的ID样本的不利影响。具体来说,C-WOE为真实OOD样本分配更高的权重,为ID样本分配更低的权重,并在训练过程中动态更新这些权重。理论上,我们为所提出的方法建立了坚实的保证。从经验上看,在各种现实世界基准和模拟数据集上进行的大量实验表明,与最先进的方法相比,C-WOE显着实现了卓越的性能,验证了其在图像处理应用中的可靠性。
{"title":"C-WOE: Clustering for Out-of-Distribution Detection Learning With Wild Outlier Exposure","authors":"Long Lan;Zhaohui Hu;He Li;Tongliang Liu;Xinwang Liu","doi":"10.1109/TIP.2026.3662593","DOIUrl":"10.1109/TIP.2026.3662593","url":null,"abstract":"Out-of-distribution (OOD) detection plays a crucial role as a mechanism for handling anomalies in computer vision systems. Among existing approaches, outlier exposure (OE), which trains the model with an additional auxiliary OOD dataset, has demonstrated strong effectiveness. However, acquiring clean and well-curated auxiliary OOD data is often infeasible, particularly within large and complex systems. Alternatively, wild outliers, i.e., unlabeled samples collected directly in deployment environments, are abundant and easy to obtain, and recent studies have shown that they can substantially benefit OOD detection learning. Nevertheless, wild outliers typically contain a mixture of in-distribution (ID) and OOD samples. Directly using them as auxiliary OOD data unavoidably exposes the model to adverse supervision signals arising from the contained ID samples. Yet existing methods still lack an effective strategy that can fully leverage wild outliers while suppressing the negative influence introduced by their ID subset. To this end, we propose a simple yet effective method named Clustering for Wild Outlier Exposure (C-WOE), which alleviates the adverse effect of the ID samples contained within wild outliers by reweighting them. Specifically, C-WOE assigns higher weights to real OOD samples and lower weights to ID samples and dynamically updates these weights during training. Theoretically, we establish solid guarantees for the proposed method. Empirically, extensive experiments conducted on various real-world benchmarks and simulated datasets demonstrate that C-WOE notably achieves superior performance compared with state-of-the-art methods, validating its reliability in image processing applications.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":"35 ","pages":"2066-2079"},"PeriodicalIF":13.7,"publicationDate":"2026-02-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146195549","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
IMPRESS: Incomplete Human Motion Prediction via Motion Recovery and Structural-Semantic Fusion 基于运动恢复和结构语义融合的不完全人类运动预测。
IF 13.7 Pub Date : 2026-02-13 DOI: 10.1109/TIP.2026.3662597
Hao Deng;Jinkai Li;Jinxing Li;Jie Wen;Yong Xu
Human motion prediction is a key task in computer vision and human-robot interaction, which has received much attention in recent years. However, existing approaches suffer from two issues: 1) They typically rely only on complete data and overlook real-world challenges such as missing observations. 2) Recent works fail to capture the diverse relations among body parts in different action categories, which limits their prediction performance. To address the above problems, we propose a novel Incomplete human Motion Prediction method through motion Re covery and Structure-Semantic fusion (IMPRESS). Specifically, for motion recovery, we introduce a wavelet-based self-attention module. It captures motion details from high-frequency features and extracts global trends from low-frequency components. To enhance the relations among different body parts, we design a structure-semantic fusion graph convolutional network. Moreover, we employ a dual-channel sliding window attention mechanism to capture motion periodicity, enabling smoother predictions. Extensive experiments on two benchmark datasets (Human3.6M, CMU-MoCap) demonstrate that IMPRESS achieves state-of-the-art average prediction performance under both complete and incomplete observations.
人体运动预测是计算机视觉和人机交互领域的一项关键任务,近年来受到广泛关注。然而,现有的方法存在两个问题:(1)它们通常只依赖于完整的数据,而忽略了现实世界的挑战,如缺失的观测值。(2)最近的研究未能捕捉到不同动作类别中身体部位之间的不同关系,这限制了它们的预测性能。针对上述问题,我们提出了一种新的基于运动恢复和结构语义融合的不完全人体运动预测方法(IMPRESS)。具体来说,对于运动恢复,我们引入了一个基于小波的自注意模块。它从高频特征中捕捉运动细节,并从低频成分中提取全球趋势。为了增强人体不同部位之间的联系,我们设计了一个结构-语义融合图卷积网络。此外,我们采用双通道滑动窗口注意机制来捕获运动周期性,从而实现更平滑的预测。在两个基准数据集(Human3.6M, mu - mocap)上进行的大量实验表明,IMPRESS在完整和不完整观察下都能达到最先进的平均预测性能。
{"title":"IMPRESS: Incomplete Human Motion Prediction via Motion Recovery and Structural-Semantic Fusion","authors":"Hao Deng;Jinkai Li;Jinxing Li;Jie Wen;Yong Xu","doi":"10.1109/TIP.2026.3662597","DOIUrl":"10.1109/TIP.2026.3662597","url":null,"abstract":"Human motion prediction is a key task in computer vision and human-robot interaction, which has received much attention in recent years. However, existing approaches suffer from two issues: 1) They typically rely only on complete data and overlook real-world challenges such as missing observations. 2) Recent works fail to capture the diverse relations among body parts in different action categories, which limits their prediction performance. To address the above problems, we propose a novel Incomplete human Motion Prediction method through motion R<sc>e</small> covery and Structure-Semantic fusion (IMPRESS). Specifically, for motion recovery, we introduce a wavelet-based self-attention module. It captures motion details from high-frequency features and extracts global trends from low-frequency components. To enhance the relations among different body parts, we design a structure-semantic fusion graph convolutional network. Moreover, we employ a dual-channel sliding window attention mechanism to capture motion periodicity, enabling smoother predictions. Extensive experiments on two benchmark datasets (Human3.6M, CMU-MoCap) demonstrate that IMPRESS achieves state-of-the-art average prediction performance under both complete and incomplete observations.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":"35 ","pages":"2393-2406"},"PeriodicalIF":13.7,"publicationDate":"2026-02-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146195737","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Hybrid Granularity Distribution Estimation for Few-Shot Learning: Statistics Transfer From Categories and Instances 少射学习的混合粒度分布估计:类别和实例的统计迁移
IF 13.7 Pub Date : 2026-02-11 DOI: 10.1109/TIP.2026.3661814
Shuo Wang;Tianyu Qi;Xingyu Zhu;Yanbin Hao;Beier Zhu;Hanwang Zhang;Meng Wang
Distribution estimation is a pivotal strategy in few-shot learning (FSL) to mitigate data scarcity by sampling from estimated distributions, utilizing statistical properties (mean and variance) transferred from related base categories. However, category-level estimation alone often fails to generate representative samples due to significant dissimilarities between base and novel categories, leading to suboptimal performance. To address this limitation, we propose Hybrid Granularity Distribution Estimation (HGDE), which integrates both coarse-grained category-level statistics and fine-grained instance-level statistics. By leveraging instance statistics from the nearest base samples, HGDE enhances the characterization of novel categories, capturing subtle features that category-level estimation overlooks. These statistics are fused through linear interpolation to form a robust distribution for novel categories, ensuring both diversity and representativeness in generated samples. Additionally, HGDE employs refined estimation techniques, such as weighted summation for mean calculation and principal component retention for covariance, to further improve accuracy. Empirical evaluations on four FSL benchmarks, including Mini-ImageNet, Tiered-ImageNet, CUB and CIFAR-FS, demonstrate that HGDE offers effective distribution estimation capabilities and leads to notable accuracy gains, with improvements of more than 1.8% in 1-shot tasks on CUB. These results highlight HGDE’s ability to balance mean precision and variance diversity, making it a versatile and effective solution for FSL.
分布估计是少射学习(FSL)的关键策略,通过从估计的分布中采样,利用从相关基本类别转移的统计属性(均值和方差)来缓解数据稀缺性。然而,由于基本类别和新类别之间存在显著差异,单独的类别水平估计往往无法生成具有代表性的样本,从而导致性能次优。为了解决这一限制,我们提出了混合粒度分布估计(HGDE),它集成了粗粒度的类别级统计和细粒度的实例级统计。通过利用来自最近的基本样本的实例统计信息,HGDE增强了新类别的特征,捕获了类别级别估计忽略的细微特征。这些统计数据通过线性插值融合,形成新类别的鲁棒分布,确保生成样本的多样性和代表性。此外,HGDE采用了精细的估计技术,如加权求和的平均值计算和主成分保留的协方差,以进一步提高准确性。在四个FSL基准测试(包括Mini-ImageNet、Tiered-ImageNet、CUB和CIFAR-FS)上的经验评估表明,HGDE提供了有效的分布估计能力,并带来了显著的准确性提高,在CUB上的单次任务中提高了1.8%以上。这些结果突出了HGDE平衡平均精度和方差多样性的能力,使其成为FSL的通用有效解决方案。
{"title":"Hybrid Granularity Distribution Estimation for Few-Shot Learning: Statistics Transfer From Categories and Instances","authors":"Shuo Wang;Tianyu Qi;Xingyu Zhu;Yanbin Hao;Beier Zhu;Hanwang Zhang;Meng Wang","doi":"10.1109/TIP.2026.3661814","DOIUrl":"10.1109/TIP.2026.3661814","url":null,"abstract":"Distribution estimation is a pivotal strategy in few-shot learning (FSL) to mitigate data scarcity by sampling from estimated distributions, utilizing statistical properties (mean and variance) transferred from related base categories. However, category-level estimation alone often fails to generate representative samples due to significant dissimilarities between base and novel categories, leading to suboptimal performance. To address this limitation, we propose Hybrid Granularity Distribution Estimation (HGDE), which integrates both coarse-grained category-level statistics and fine-grained instance-level statistics. By leveraging instance statistics from the nearest base samples, HGDE enhances the characterization of novel categories, capturing subtle features that category-level estimation overlooks. These statistics are fused through linear interpolation to form a robust distribution for novel categories, ensuring both diversity and representativeness in generated samples. Additionally, HGDE employs refined estimation techniques, such as weighted summation for mean calculation and principal component retention for covariance, to further improve accuracy. Empirical evaluations on four FSL benchmarks, including Mini-ImageNet, Tiered-ImageNet, CUB and CIFAR-FS, demonstrate that HGDE offers effective distribution estimation capabilities and leads to notable accuracy gains, with improvements of more than 1.8% in 1-shot tasks on CUB. These results highlight HGDE’s ability to balance mean precision and variance diversity, making it a versatile and effective solution for FSL.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":"35 ","pages":"2080-2093"},"PeriodicalIF":13.7,"publicationDate":"2026-02-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146161244","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Adaptive Fine-Grained Fusion Network for Multimodal UAV Object Detection 多模态无人机目标检测的自适应细粒度融合网络
IF 13.7 Pub Date : 2026-02-11 DOI: 10.1109/TIP.2026.3661868
Zhanyan Tang;Zhihao Wu;Mu Li;Jie Wen;Bob Zhang;Yong Xu;Jianqiang Li
Multimodal perception and fusion play a vital role in uncrewed aerial vehicle (UAV) object detection. Existing methods typically adopt global fusion strategies across modalities. However, due to illumination variation, the effectiveness of RGB and infrared modalities may differ across local regions within the same image, particularly in UAV perspectives where occlusions and dense small objects are prevalent, leading to suboptimal performance of global fusion methods. To address this issue, we propose an adaptive fine-grained fusion network for multimodal UAV object detection. First, we design a local feature consistency-based modality fusion module, which adaptively assigns local fusion weights according to the structural consistency of high-response regions across modalities, thereby enabling more effective aggregation of object-relevant features. Second, we introduce a mutual information-guided feature contrastive loss to encourage the preservation of modality-specific information during the early training phase. Experimental results demonstrate that the proposed method effectively addresses the issue of object occlusion in UAV perspectives, achieving state-of-the-art performance on multimodal UAV object detection benchmarks. Code will be available at https://github.com/lingf5877/AFFNet
多模态感知与融合在无人机目标检测中起着至关重要的作用。现有方法通常采用跨模式的全局融合策略。然而,由于光照的变化,RGB和红外模式的有效性可能在同一图像的局部区域有所不同,特别是在无人机视角中,遮挡和密集的小物体普遍存在,导致全局融合方法的性能不理想。为了解决这一问题,我们提出了一种用于多模式无人机目标检测的自适应细粒度融合网络。首先,设计了基于局部特征一致性的模态融合模块,根据模态高响应区域的结构一致性自适应分配局部融合权值,从而实现更有效的目标相关特征聚合。其次,我们引入了一个相互信息引导的特征对比损失,以鼓励在早期训练阶段保留特定于模态的信息。实验结果表明,该方法有效地解决了无人机视角中的目标遮挡问题,在多模态无人机目标检测基准上取得了最先进的性能。代码将在https://github.com/lingf5877/AFFNet上提供
{"title":"Adaptive Fine-Grained Fusion Network for Multimodal UAV Object Detection","authors":"Zhanyan Tang;Zhihao Wu;Mu Li;Jie Wen;Bob Zhang;Yong Xu;Jianqiang Li","doi":"10.1109/TIP.2026.3661868","DOIUrl":"10.1109/TIP.2026.3661868","url":null,"abstract":"Multimodal perception and fusion play a vital role in uncrewed aerial vehicle (UAV) object detection. Existing methods typically adopt global fusion strategies across modalities. However, due to illumination variation, the effectiveness of RGB and infrared modalities may differ across local regions within the same image, particularly in UAV perspectives where occlusions and dense small objects are prevalent, leading to suboptimal performance of global fusion methods. To address this issue, we propose an adaptive fine-grained fusion network for multimodal UAV object detection. First, we design a local feature consistency-based modality fusion module, which adaptively assigns local fusion weights according to the structural consistency of high-response regions across modalities, thereby enabling more effective aggregation of object-relevant features. Second, we introduce a mutual information-guided feature contrastive loss to encourage the preservation of modality-specific information during the early training phase. Experimental results demonstrate that the proposed method effectively addresses the issue of object occlusion in UAV perspectives, achieving state-of-the-art performance on multimodal UAV object detection benchmarks. Code will be available at <uri>https://github.com/lingf5877/AFFNet</uri>","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":"35 ","pages":"1870-1882"},"PeriodicalIF":13.7,"publicationDate":"2026-02-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146161243","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
CGMNet: A Center-Pixel and Gated Mechanism-Based Attention Network for Hyperspectral Change Detection CGMNet:一个基于中心像素和门控机制的高光谱变化检测注意网络
IF 13.7 Pub Date : 2026-02-11 DOI: 10.1109/TIP.2026.3661851
Lanxin Wu;Jiangtao Peng;Bing Yang;Weiwei Sun;Mingzhu Huang
Change detection (CD) in hyperspectral images (HSIs) has become an increasingly vital research field in remote sensing. Over the past few years, the adoption of deep learning approaches, particularly convolutional neural network (CNN) and transformer-based architectures have significantly advanced performance in this field. While these models effectively capture spectral-spatial features, they may also introduce redundant or irrelevant spatial information, potentially degrading the accuracy of HSI CD. To address this challenge, a center-pixel and gated mechanism-based attention network (CGMNet) is proposed for HSI CD, leveraging the central pixel’s significance to enhance accuracy and robustness. First, a gated-based center spatial attention (GCSA) module is designed to emphasize spatial relationships surrounding the central pixel. By incorporating gating mechanisms, GCSA selectively enhances relevant spatial features while suppressing irrelevant information. Second, a gated-based spectral attention (GSA) module is proposed to dynamically highlight the most significant spectral features, ensuring an effective spectral representation. Finally, a global transform fusion (GTF) module is proposed to capture global contextual information and to fuse it with the extracted spatial and spectral features. Moreover, we introduce a novel benchmark dataset, named the Hangzhou Bay (HZB), specifically designed to advance coastal remote sensing research. Experimental evaluations conducted on three publicly available datasets, as well as the HZB dataset, show that our CGMNet consistently outperforms some state-of-the-art methods in the HSI CD task. The source code of the proposed CGMNet, along with the HZB dataset, will be made publicly available at https://github.com/creativeXin/CGMNet
高光谱图像的变化检测已成为遥感领域一个日益重要的研究领域。在过去的几年里,深度学习方法的采用,特别是卷积神经网络(CNN)和基于变压器的架构在这一领域取得了显著的进步。虽然这些模型有效地捕获了光谱空间特征,但它们也可能引入冗余或不相关的空间信息,从而可能降低HSI CD的精度。为了解决这一挑战,我们提出了一种基于中心像素和门控机制的HSI CD注意网络(CGMNet),利用中心像素的重要性来提高精度和鲁棒性。首先,设计了一个基于栅格的中心空间注意(GCSA)模块来强调中心像素周围的空间关系。通过引入门控机制,GCSA可以选择性地增强相关空间特征,同时抑制无关信息。其次,提出了一种基于门控的光谱注意(GSA)模块,用于动态突出最重要的光谱特征,确保有效的光谱表示。最后,提出了一个全局变换融合(GTF)模块,用于捕获全局上下文信息,并将其与提取的空间和光谱特征融合。此外,我们还介绍了一个新的基准数据集,命名为杭州湾(HZB),专门用于推进沿海遥感研究。在三个公开可用的数据集以及HZB数据集上进行的实验评估表明,我们的CGMNet在HSI CD任务中始终优于一些最先进的方法。拟议的CGMNet的源代码以及HZB数据集将在https://github.com/creativeXin/CGMNet上公开提供
{"title":"CGMNet: A Center-Pixel and Gated Mechanism-Based Attention Network for Hyperspectral Change Detection","authors":"Lanxin Wu;Jiangtao Peng;Bing Yang;Weiwei Sun;Mingzhu Huang","doi":"10.1109/TIP.2026.3661851","DOIUrl":"10.1109/TIP.2026.3661851","url":null,"abstract":"Change detection (CD) in hyperspectral images (HSIs) has become an increasingly vital research field in remote sensing. Over the past few years, the adoption of deep learning approaches, particularly convolutional neural network (CNN) and transformer-based architectures have significantly advanced performance in this field. While these models effectively capture spectral-spatial features, they may also introduce redundant or irrelevant spatial information, potentially degrading the accuracy of HSI CD. To address this challenge, a center-pixel and gated mechanism-based attention network (CGMNet) is proposed for HSI CD, leveraging the central pixel’s significance to enhance accuracy and robustness. First, a gated-based center spatial attention (GCSA) module is designed to emphasize spatial relationships surrounding the central pixel. By incorporating gating mechanisms, GCSA selectively enhances relevant spatial features while suppressing irrelevant information. Second, a gated-based spectral attention (GSA) module is proposed to dynamically highlight the most significant spectral features, ensuring an effective spectral representation. Finally, a global transform fusion (GTF) module is proposed to capture global contextual information and to fuse it with the extracted spatial and spectral features. Moreover, we introduce a novel benchmark dataset, named the Hangzhou Bay (HZB), specifically designed to advance coastal remote sensing research. Experimental evaluations conducted on three publicly available datasets, as well as the HZB dataset, show that our CGMNet consistently outperforms some state-of-the-art methods in the HSI CD task. The source code of the proposed CGMNet, along with the HZB dataset, will be made publicly available at <uri>https://github.com/creativeXin/CGMNet</uri>","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":"35 ","pages":"1951-1965"},"PeriodicalIF":13.7,"publicationDate":"2026-02-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146161245","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE transactions on image processing : a publication of the IEEE Signal Processing Society
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1