首页 > 最新文献

IET Computer Vision最新文献

英文 中文
SIM-Net: A Multimodal Fusion Network Using Inferred 3D Object Shape Point Clouds From RGB Images for 2D Classification SIM-Net:利用RGB图像中推断的3D物体形状点云进行2D分类的多模态融合网络
IF 1.3 4区 计算机科学 Q4 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-07-09 DOI: 10.1049/cvi2.70036
Youcef Sklab, Hanane Ariouat, Eric Chenin, Edi Prifti, Jean-Daniel Zucker

We introduce the shape-image multimodal network (SIM-Net), a novel 2D image classification architecture that integrates 3D point cloud representations inferred directly from RGB images. Our key contribution lies in a pixel-to-point transformation that converts 2D object masks into 3D point clouds, enabling the fusion of texture-based and geometric features for enhanced classification performance. SIM-Net is particularly well-suited for the classification of digitised herbarium specimens—a task made challenging by heterogeneous backgrounds, nonplant elements, and occlusions that compromise conventional image-based models. To address these issues, SIM-Net employs a segmentation-based preprocessing step to extract object masks prior to 3D point cloud generation. The architecture comprises a CNN encoder for 2D image features and a PointNet-based encoder for geometric features, which are fused into a unified latent space. Experimental evaluations on herbarium datasets demonstrate that SIM-Net consistently outperforms ResNet101, achieving gains of up to 9.9% in accuracy and 12.3% in F-score. It also surpasses several transformer-based state-of-the-art architectures, highlighting the benefits of incorporating 3D structural reasoning into 2D image classification tasks.

我们介绍了形状-图像多模态网络(SIM-Net),这是一种新的2D图像分类架构,集成了直接从RGB图像推断的3D点云表示。我们的关键贡献在于将2D对象蒙版转换为3D点云的像素到点转换,从而实现基于纹理和几何特征的融合,从而增强分类性能。SIM-Net特别适合于数字化植物标本的分类,这是一项具有挑战性的任务,因为不同的背景、非植物元素和遮挡会损害传统的基于图像的模型。为了解决这些问题,SIM-Net采用基于分段的预处理步骤,在生成3D点云之前提取对象掩模。该架构包括一个用于二维图像特征的CNN编码器和一个用于几何特征的基于pointnet的编码器,它们融合到一个统一的潜在空间中。对植物标本馆数据集的实验评估表明,SIM-Net始终优于ResNet101,准确率提高了9.9%,F-score提高了12.3%。它还超越了几种基于变压器的最先进架构,突出了将3D结构推理纳入2D图像分类任务的好处。
{"title":"SIM-Net: A Multimodal Fusion Network Using Inferred 3D Object Shape Point Clouds From RGB Images for 2D Classification","authors":"Youcef Sklab,&nbsp;Hanane Ariouat,&nbsp;Eric Chenin,&nbsp;Edi Prifti,&nbsp;Jean-Daniel Zucker","doi":"10.1049/cvi2.70036","DOIUrl":"10.1049/cvi2.70036","url":null,"abstract":"<p>We introduce the shape-image multimodal network (SIM-Net), a novel 2D image classification architecture that integrates 3D point cloud representations inferred directly from RGB images. Our key contribution lies in a pixel-to-point transformation that converts 2D object masks into 3D point clouds, enabling the fusion of texture-based and geometric features for enhanced classification performance. SIM-Net is particularly well-suited for the classification of digitised herbarium specimens—a task made challenging by heterogeneous backgrounds, nonplant elements, and occlusions that compromise conventional image-based models. To address these issues, SIM-Net employs a segmentation-based preprocessing step to extract object masks prior to 3D point cloud generation. The architecture comprises a CNN encoder for 2D image features and a PointNet-based encoder for geometric features, which are fused into a unified latent space. Experimental evaluations on herbarium datasets demonstrate that SIM-Net consistently outperforms ResNet101, achieving gains of up to 9.9% in accuracy and 12.3% in F-score. It also surpasses several transformer-based state-of-the-art architectures, highlighting the benefits of incorporating 3D structural reasoning into 2D image classification tasks.</p>","PeriodicalId":56304,"journal":{"name":"IET Computer Vision","volume":"19 1","pages":""},"PeriodicalIF":1.3,"publicationDate":"2025-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1049/cvi2.70036","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144589534","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Improved SAR Aircraft Detection Algorithm Based on Visual State Space Models 基于视觉状态空间模型的SAR飞机检测改进算法
IF 1.3 4区 计算机科学 Q4 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-07-07 DOI: 10.1049/cvi2.70032
Yaqiong Wang, Jing Zhang, Yipei Wang, Shiyu Hu, Baoguo Shen, Zhenhua Hou, Wanting Zhou

In recent years, the development of deep learning algorithms has significantly advanced the application of synthetic aperture radar (SAR) aircraft detection in remote sensing and military fields. However, existing methods face a dual dilemma: CNN-based models suffer from insufficient detection accuracy due to limitations in local receptive fields, whereas Transformer-based models improve accuracy by leveraging attention mechanisms but incur significant computational overhead due to their quadratic complexity. This imbalance between accuracy and efficiency severely limits the development of SAR aircraft detection. To address this problem, this paper propose a novel neural network based on state space models (SSM), termed the Mamba SAR detection network (MSAD). Specifically, we design a feature encoding module, MEBlock, that integrates CNN with SSM to enhance global feature modelling capabilities. Meanwhile, the linear computational complexity brought by SSM is superior to that of Transformer architectures, achieving a reduction in computational overhead. Additionally, we propose a context-aware feature fusion module (CAFF) that combines attention mechanisms to achieve adaptive fusion of multi-scale features. Lastly, a lightweight parameter-shared detection head (PSHead) is utilised to effectively reduce redundant parameters through implicit feature interaction. Experiments on the SAR-AirCraft-v1.0 and SADD datasets show that MSAD achieves higher accuracy than existing algorithms, whereas its GFLOPs are 2.7 times smaller than those of the Transformer architecture RT-DETR. These results validate the core role of SSM as an accuracy-efficiency balancer, reflecting MSAD's perceptual capability and performance in SAR aircraft detection in complex environments.

近年来,深度学习算法的发展极大地推动了合成孔径雷达(SAR)飞机探测在遥感和军事领域的应用。然而,现有的方法面临着双重困境:基于cnn的模型由于局部接受域的限制而导致检测精度不足,而基于transformer的模型通过利用注意机制来提高准确性,但由于其二次复杂度而导致大量的计算开销。这种精度与效率的不平衡严重限制了SAR飞机探测的发展。为了解决这一问题,本文提出了一种新的基于状态空间模型(SSM)的神经网络,称为曼巴SAR检测网络(MSAD)。具体来说,我们设计了一个特征编码模块MEBlock,该模块集成了CNN和SSM,以增强全局特征建模能力。同时,SSM带来的线性计算复杂度优于Transformer架构,实现了计算开销的降低。此外,我们提出了一种结合注意机制的上下文感知特征融合模块(CAFF),以实现多尺度特征的自适应融合。最后,利用轻量级参数共享检测头(PSHead),通过隐式特征交互有效减少冗余参数。在SAR-AirCraft-v1.0和SADD数据集上的实验表明,MSAD的精度高于现有算法,而GFLOPs比Transformer架构RT-DETR的GFLOPs小2.7倍。这些结果验证了SSM作为精度-效率平衡器的核心作用,反映了MSAD在复杂环境下的SAR飞机检测中的感知能力和性能。
{"title":"Improved SAR Aircraft Detection Algorithm Based on Visual State Space Models","authors":"Yaqiong Wang,&nbsp;Jing Zhang,&nbsp;Yipei Wang,&nbsp;Shiyu Hu,&nbsp;Baoguo Shen,&nbsp;Zhenhua Hou,&nbsp;Wanting Zhou","doi":"10.1049/cvi2.70032","DOIUrl":"10.1049/cvi2.70032","url":null,"abstract":"<p>In recent years, the development of deep learning algorithms has significantly advanced the application of synthetic aperture radar (SAR) aircraft detection in remote sensing and military fields. However, existing methods face a dual dilemma: CNN-based models suffer from insufficient detection accuracy due to limitations in local receptive fields, whereas Transformer-based models improve accuracy by leveraging attention mechanisms but incur significant computational overhead due to their quadratic complexity. This imbalance between accuracy and efficiency severely limits the development of SAR aircraft detection. To address this problem, this paper propose a novel neural network based on state space models (SSM), termed the Mamba SAR detection network (MSAD). Specifically, we design a feature encoding module, MEBlock, that integrates CNN with SSM to enhance global feature modelling capabilities. Meanwhile, the linear computational complexity brought by SSM is superior to that of Transformer architectures, achieving a reduction in computational overhead. Additionally, we propose a context-aware feature fusion module (CAFF) that combines attention mechanisms to achieve adaptive fusion of multi-scale features. Lastly, a lightweight parameter-shared detection head (PSHead) is utilised to effectively reduce redundant parameters through implicit feature interaction. Experiments on the SAR-AirCraft-v1.0 and SADD datasets show that MSAD achieves higher accuracy than existing algorithms, whereas its GFLOPs are 2.7 times smaller than those of the Transformer architecture RT-DETR. These results validate the core role of SSM as an accuracy-efficiency balancer, reflecting MSAD's perceptual capability and performance in SAR aircraft detection in complex environments.</p>","PeriodicalId":56304,"journal":{"name":"IET Computer Vision","volume":"19 1","pages":""},"PeriodicalIF":1.3,"publicationDate":"2025-07-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1049/cvi2.70032","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144574132","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multi-Attention Fusion Artistic Radiance Fields and Beyond 多关注融合艺术光芒领域和超越
IF 1.3 4区 计算机科学 Q4 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-06-21 DOI: 10.1049/cvi2.70017
Qianru Chen, Yufan Zhou, Xintong Hou, Kunze Jiang, Jincheng Li, Chao Wu

We present MRF (multi-attention fusion artistic radiance fields), a novel approach to 3D scene stylisation that synthesises artistic rendering by integrating stylised 2D images with neural radiance fields. Our method effectively incorporates high-frequency stylistic elements from 2D artistic representations while maintaining geometric consistency across multiple viewpoints. To address the challenges of view-dependent stylisation coherence and semantic fidelity, we introduce two key components: (1) a multi-scale attention module (MAM) that facilitates hierarchical feature extraction and fusion across different spatial resolutions and (2) a CLIP-guided semantic consistency module that preserves the underlying scene structure during style transfer. Through extensive experimentation, we demonstrate that MRF achieves superior stylisation quality and detail preservation compared to state-of-the-art methods, particularly in capturing fine artistic details while maintaining view consistency. Our approach represents a significant advancement in neural rendering-based artistic stylisation of 3D scenes.

我们提出了MRF(多注意力融合艺术辐射场),这是一种3D场景风格化的新方法,通过将风格化的2D图像与神经辐射场集成来合成艺术渲染。我们的方法有效地结合了来自2D艺术表现的高频风格元素,同时保持了多个视点的几何一致性。为了解决依赖于视图的风格一致性和语义保真度的挑战,我们引入了两个关键组件:(1)促进不同空间分辨率的分层特征提取和融合的多尺度注意力模块(MAM)和(2)在风格转移过程中保留底层场景结构的clip引导的语义一致性模块。通过广泛的实验,我们证明,与最先进的方法相比,MRF实现了卓越的风格化质量和细节保存,特别是在捕捉精美的艺术细节的同时保持视图一致性。我们的方法代表了基于神经渲染的3D场景艺术风格的重大进步。
{"title":"Multi-Attention Fusion Artistic Radiance Fields and Beyond","authors":"Qianru Chen,&nbsp;Yufan Zhou,&nbsp;Xintong Hou,&nbsp;Kunze Jiang,&nbsp;Jincheng Li,&nbsp;Chao Wu","doi":"10.1049/cvi2.70017","DOIUrl":"10.1049/cvi2.70017","url":null,"abstract":"<p>We present MRF (multi-attention fusion artistic radiance fields), a novel approach to 3D scene stylisation that synthesises artistic rendering by integrating stylised 2D images with neural radiance fields. Our method effectively incorporates high-frequency stylistic elements from 2D artistic representations while maintaining geometric consistency across multiple viewpoints. To address the challenges of view-dependent stylisation coherence and semantic fidelity, we introduce two key components: (1) a multi-scale attention module (MAM) that facilitates hierarchical feature extraction and fusion across different spatial resolutions and (2) a CLIP-guided semantic consistency module that preserves the underlying scene structure during style transfer. Through extensive experimentation, we demonstrate that MRF achieves superior stylisation quality and detail preservation compared to state-of-the-art methods, particularly in capturing fine artistic details while maintaining view consistency. Our approach represents a significant advancement in neural rendering-based artistic stylisation of 3D scenes.</p>","PeriodicalId":56304,"journal":{"name":"IET Computer Vision","volume":"19 1","pages":""},"PeriodicalIF":1.3,"publicationDate":"2025-06-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1049/cvi2.70017","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144331868","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
PAD: Detail-Preserving Point Cloud Reconstruction and Generation via Autodecoders PAD:保留细节的点云重建和生成通过自动解码器
IF 1.3 4区 计算机科学 Q4 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-06-18 DOI: 10.1049/cvi2.70031
Yakai Zhang, Ping Yang, Haoran Wang, Zizhao Wu, Xiaoling Gu, Alexandru Telea, Jiri Kosinka

High-accuracy point cloud (self-) reconstruction is crucial for point cloud editing, translation, and unsupervised representation learning. However, existing point cloud reconstruction methods often sacrifice many geometric details. Altough many techniques have proposed how to construct better point cloud decoders, only a few have designed point cloud encoders from a reconstruction perspective. We propose an autodecoder architecture to achieve detail-preserving point cloud reconstruction while bypassing the performance bottleneck of the encoder. Our architecture is theoretically applicable to any existing point cloud decoder. For training, both the weights of the decoder and the pre-initialised latent codes, corresponding to the input points, are updated simultaneously. Experimental results demonstrate that our autodecoder achieves an average reduction of 24.62% in Chamfer Distance compared to existing methods, significantly improving reconstruction quality on the ShapeNet dataset. Furthermore, we verify the effectiveness of our autodecoder in point cloud generation, upsampling, and unsupervised representation learning to demonstrate its performance on downstream tasks, which is comparable to the state-of-the-art methods. We will make our code publicly available after peer review.

高精度的点云(自)重建对于点云编辑、翻译和无监督表示学习至关重要。然而,现有的点云重建方法往往牺牲了许多几何细节。虽然已有许多技术提出如何构建更好的点云解码器,但从重构的角度设计点云编码器的技术很少。我们提出了一种自动解码器架构,以实现保留细节的点云重建,同时绕过编码器的性能瓶颈。我们的架构理论上适用于任何现有的点云解码器。对于训练,同时更新译码器的权值和与输入点对应的预初始化潜在码的权值。实验结果表明,与现有方法相比,我们的自解码器平均减少了24.62%的倒角距离,显著提高了ShapeNet数据集的重建质量。此外,我们验证了我们的自动解码器在点云生成、上采样和无监督表示学习方面的有效性,以展示其在下游任务上的性能,这与最先进的方法相当。我们将在同行评审后公开我们的代码。
{"title":"PAD: Detail-Preserving Point Cloud Reconstruction and Generation via Autodecoders","authors":"Yakai Zhang,&nbsp;Ping Yang,&nbsp;Haoran Wang,&nbsp;Zizhao Wu,&nbsp;Xiaoling Gu,&nbsp;Alexandru Telea,&nbsp;Jiri Kosinka","doi":"10.1049/cvi2.70031","DOIUrl":"10.1049/cvi2.70031","url":null,"abstract":"<p>High-accuracy point cloud (self-) reconstruction is crucial for point cloud editing, translation, and unsupervised representation learning. However, existing point cloud reconstruction methods often sacrifice many geometric details. Altough many techniques have proposed how to construct better point cloud decoders, only a few have designed point cloud encoders from a reconstruction perspective. We propose an autodecoder architecture to achieve detail-preserving point cloud reconstruction while bypassing the performance bottleneck of the encoder. Our architecture is theoretically applicable to any existing point cloud decoder. For training, both the weights of the decoder and the pre-initialised latent codes, corresponding to the input points, are updated simultaneously. Experimental results demonstrate that our autodecoder achieves an average reduction of 24.62% in Chamfer Distance compared to existing methods, significantly improving reconstruction quality on the ShapeNet dataset. Furthermore, we verify the effectiveness of our autodecoder in point cloud generation, upsampling, and unsupervised representation learning to demonstrate its performance on downstream tasks, which is comparable to the state-of-the-art methods. We will make our code publicly available after peer review.</p>","PeriodicalId":56304,"journal":{"name":"IET Computer Vision","volume":"19 1","pages":""},"PeriodicalIF":1.3,"publicationDate":"2025-06-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1049/cvi2.70031","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144315358","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
GRVT: Improving the Transferability of Adversarial Attacks Through Gradient Related Variance and Input Transformation GRVT:通过梯度相关方差和输入变换提高对抗性攻击的可转移性
IF 1.3 4区 计算机科学 Q4 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-06-11 DOI: 10.1049/cvi2.70034
Yanlei Wei, Xiaolin Zhang, Yongping Wang, Jingyu Wang, Lixin Liu

As we all know, the emergence of a large number of adversarial samples reveals the vulnerability of deep neural networks. Attackers seriously affect the performance of models by adding imperceptible perturbations. Although adversarial samples have a high transferability success rate in white-box models, they are less effective in black-box models. To address this problem, this paper proposes a new transferability attack strategy, Gradient Related Variance and Input Transformation Attack (GRVT). First, the image is divided into small blocks, and random transformations are applied to each block to generate diversified images; then, in the gradient update process, the gradient of the neighbourhood area is introduced, and the current gradient is associated with the neighbourhood average gradient through Cosine Similarity. The current gradient direction is adjusted using the associated gradient combined with the previous gradient variance, and a step size reducer adjusts the gradient step size. Experiments on the ILSVRC 2012 dataset show that the transferability success rate of adversarial samples between convolutional neural network (CNN) and vision transformer (ViT) models is higher than that of currently advanced methods. Additionally, the adversarial samples generated on the ensemble model are practical against nine defence strategies. GRVT shows excellent transferability and broad applicability.

众所周知,大量对抗性样本的出现暴露了深度神经网络的脆弱性。攻击者通过添加难以察觉的扰动严重影响模型的性能。尽管对抗性样本在白盒模型中具有较高的可转移成功率,但在黑盒模型中效果较差。针对这一问题,本文提出了一种新的可转移性攻击策略——梯度相关方差和输入变换攻击(GRVT)。首先,将图像分成小块,对每个小块进行随机变换,生成多样化的图像;然后,在梯度更新过程中引入邻域梯度,通过余弦相似度将当前梯度与邻域平均梯度相关联;使用关联的梯度结合之前的梯度方差来调整当前的梯度方向,并使用步长减速器来调整梯度步长。在ILSVRC 2012数据集上的实验表明,卷积神经网络(CNN)和视觉变压器(ViT)模型之间的对抗性样本转移成功率高于目前先进的方法。此外,在集成模型上生成的对抗样本对九种防御策略都是实用的。GRVT具有优良的可移植性和广泛的适用性。
{"title":"GRVT: Improving the Transferability of Adversarial Attacks Through Gradient Related Variance and Input Transformation","authors":"Yanlei Wei,&nbsp;Xiaolin Zhang,&nbsp;Yongping Wang,&nbsp;Jingyu Wang,&nbsp;Lixin Liu","doi":"10.1049/cvi2.70034","DOIUrl":"10.1049/cvi2.70034","url":null,"abstract":"<p>As we all know, the emergence of a large number of adversarial samples reveals the vulnerability of deep neural networks. Attackers seriously affect the performance of models by adding imperceptible perturbations. Although adversarial samples have a high transferability success rate in white-box models, they are less effective in black-box models. To address this problem, this paper proposes a new transferability attack strategy, Gradient Related Variance and Input Transformation Attack (GRVT). First, the image is divided into small blocks, and random transformations are applied to each block to generate diversified images; then, in the gradient update process, the gradient of the neighbourhood area is introduced, and the current gradient is associated with the neighbourhood average gradient through Cosine Similarity. The current gradient direction is adjusted using the associated gradient combined with the previous gradient variance, and a step size reducer adjusts the gradient step size. Experiments on the ILSVRC 2012 dataset show that the transferability success rate of adversarial samples between convolutional neural network (CNN) and vision transformer (ViT) models is higher than that of currently advanced methods. Additionally, the adversarial samples generated on the ensemble model are practical against nine defence strategies. GRVT shows excellent transferability and broad applicability.</p>","PeriodicalId":56304,"journal":{"name":"IET Computer Vision","volume":"19 1","pages":""},"PeriodicalIF":1.3,"publicationDate":"2025-06-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1049/cvi2.70034","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144264598","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Enhanced Foreground–Background Discrimination for Weakly Supervised Semantic Segmentation 弱监督语义分割的增强前景背景判别
IF 1.3 4区 计算机科学 Q4 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-06-09 DOI: 10.1049/cvi2.70029
Zhoufeng Liu, Bingrui Li, Miao Yu, Guangshuai Gao, Chunlei Li

Weakly supervised semantic segmentation (WSSS) methods are extensively studied due to the availability of image-level annotations. Relying on class activation maps (CAMs) derived from original classification networks often suffers from issues such as inaccurate object localization, incomplete object regions, and the inclusion of confusing background pixels. To address these issues, we propose a two-stage method that enhances the foreground–background discriminative ability in a global context (FB-DGC). Specifically, a cross-domain feature calibration module (CFCM) is first proposed to calibrate foreground and background salient features using global spatial location information, thereby expanding foreground features while mitigating the impact of inaccurate localization in class activation regions. A class-specific distance module (CSDM) is further adopted to facilitate the separation of foreground–background features, thereby enhancing the activation of target regions, which alleviates the over-smoothing of features produced by the network and mitigates issues associated with confused features. In addition, an adaptive edge feature extraction (AEFE) strategy is proposed to identify target features in candidate boundary regions and capture missed features, compensating for drawbacks in recognising the co-occurrence of multiple targets. The proposed method is extensively evaluated on the challenging PASCAL VOC 2012 and MS COCO 2014 datasets, demonstrating its feasibility and superiority.

由于图像级标注的可用性,弱监督语义分割方法得到了广泛的研究。依赖于原始分类网络衍生的类激活图(CAMs)经常会遇到诸如不准确的对象定位、不完整的对象区域以及包含令人困惑的背景像素等问题。为了解决这些问题,我们提出了一种两阶段方法来增强全球背景下的前景-背景区分能力(FB-DGC)。具体而言,首先提出了一种跨域特征校准模块(CFCM),利用全局空间位置信息对前景和背景显著特征进行校准,从而在扩展前景特征的同时减轻类激活区域定位不准确的影响。进一步采用类特定距离模块(class-specific distance module, CSDM),实现前景与背景特征的分离,从而增强目标区域的激活,缓解了网络产生的特征的过度平滑,缓解了特征混淆的问题。此外,提出了一种自适应边缘特征提取(AEFE)策略,用于识别候选边界区域中的目标特征并捕获缺失特征,弥补了识别多目标共现的不足。该方法在具有挑战性的PASCAL VOC 2012和MS COCO 2014数据集上进行了广泛的评估,证明了其可行性和优越性。
{"title":"Enhanced Foreground–Background Discrimination for Weakly Supervised Semantic Segmentation","authors":"Zhoufeng Liu,&nbsp;Bingrui Li,&nbsp;Miao Yu,&nbsp;Guangshuai Gao,&nbsp;Chunlei Li","doi":"10.1049/cvi2.70029","DOIUrl":"10.1049/cvi2.70029","url":null,"abstract":"<p>Weakly supervised semantic segmentation (WSSS) methods are extensively studied due to the availability of image-level annotations. Relying on class activation maps (CAMs) derived from original classification networks often suffers from issues such as inaccurate object localization, incomplete object regions, and the inclusion of confusing background pixels. To address these issues, we propose a two-stage method that enhances the foreground–background discriminative ability in a global context (FB-DGC). Specifically, a cross-domain feature calibration module (CFCM) is first proposed to calibrate foreground and background salient features using global spatial location information, thereby expanding foreground features while mitigating the impact of inaccurate localization in class activation regions. A class-specific distance module (CSDM) is further adopted to facilitate the separation of foreground–background features, thereby enhancing the activation of target regions, which alleviates the over-smoothing of features produced by the network and mitigates issues associated with confused features. In addition, an adaptive edge feature extraction (AEFE) strategy is proposed to identify target features in candidate boundary regions and capture missed features, compensating for drawbacks in recognising the co-occurrence of multiple targets. The proposed method is extensively evaluated on the challenging PASCAL VOC 2012 and MS COCO 2014 datasets, demonstrating its feasibility and superiority.</p>","PeriodicalId":56304,"journal":{"name":"IET Computer Vision","volume":"19 1","pages":""},"PeriodicalIF":1.3,"publicationDate":"2025-06-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1049/cvi2.70029","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144244550","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Mamba4SOD: RGB-T Salient Object Detection Using Mamba-Based Fusion Module Mamba4SOD:基于mamba融合模块的RGB-T显著目标检测
IF 1.3 4区 计算机科学 Q4 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-06-05 DOI: 10.1049/cvi2.70033
Yi Xu, Ruichao Hou, Ziheng Qi, Tongwei Ren

RGB and thermal salient object detection (RGB-T SOD) aims to accurately locate and segment salient objects in aligned visible and thermal image pairs. However, existing methods often struggle to produce complete masks and sharp boundaries in challenging scenarios due to insufficient exploration of complementary features from the dual modalities. In this paper, we propose a novel mamba-based fusion network for RGB-T SOD task, named Mamba4SOD, which integrates the strengths of Swin Transformer and Mamba to construct robust multi-modal representations, effectively reducing pixel misclassification. Specifically, we leverage Swin Transformer V2 to establish long-range contextual dependencies and thoroughly analyse the impact of features at various levels on detection performance. Additionally, we develop a novel Mamba-based fusion module with linear complexity, boosting multi-modal enhancement and fusion. Experimental results on VT5000, VT1000 and VT821 datasets demonstrate that our method outperforms the state-of-the-art RGB-T SOD methods.

RGB和热显著目标检测(RGB- t SOD)旨在准确定位和分割对齐的可见光和热图像对中的显著目标。然而,由于对双模态互补特征的探索不足,现有的方法往往难以在具有挑战性的场景中产生完整的掩模和清晰的边界。在本文中,我们提出了一种新的基于Mamba4SOD的RGB-T SOD融合网络,该网络融合了Swin Transformer和Mamba的优点,构建了鲁棒的多模态表示,有效地减少了像素的误分类。具体来说,我们利用Swin Transformer V2来建立远程上下文依赖关系,并彻底分析各个级别的特性对检测性能的影响。此外,我们开发了一种新颖的基于mamba的融合模块,具有线性复杂性,促进了多模态增强和融合。在VT5000, VT1000和VT821数据集上的实验结果表明,我们的方法优于最先进的RGB-T SOD方法。
{"title":"Mamba4SOD: RGB-T Salient Object Detection Using Mamba-Based Fusion Module","authors":"Yi Xu,&nbsp;Ruichao Hou,&nbsp;Ziheng Qi,&nbsp;Tongwei Ren","doi":"10.1049/cvi2.70033","DOIUrl":"10.1049/cvi2.70033","url":null,"abstract":"<p>RGB and thermal salient object detection (RGB-T SOD) aims to accurately locate and segment salient objects in aligned visible and thermal image pairs. However, existing methods often struggle to produce complete masks and sharp boundaries in challenging scenarios due to insufficient exploration of complementary features from the dual modalities. In this paper, we propose a novel mamba-based fusion network for RGB-T SOD task, named Mamba4SOD, which integrates the strengths of Swin Transformer and Mamba to construct robust multi-modal representations, effectively reducing pixel misclassification. Specifically, we leverage Swin Transformer V2 to establish long-range contextual dependencies and thoroughly analyse the impact of features at various levels on detection performance. Additionally, we develop a novel Mamba-based fusion module with linear complexity, boosting multi-modal enhancement and fusion. Experimental results on VT5000, VT1000 and VT821 datasets demonstrate that our method outperforms the state-of-the-art RGB-T SOD methods.</p>","PeriodicalId":56304,"journal":{"name":"IET Computer Vision","volume":"19 1","pages":""},"PeriodicalIF":1.3,"publicationDate":"2025-06-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1049/cvi2.70033","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144220255","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Object Detection Based on CNN and Vision-Transformer: A Survey 基于CNN和视觉变换的目标检测研究进展
IF 1.3 4区 计算机科学 Q4 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-05-31 DOI: 10.1049/cvi2.70028
Jinfeng Cao, Bo Peng, Mingzhong Gao, Haichun Hao, Xinfang Li, Hongwei Mou

Object detection is the most crucial and challenging task of computer vision and has been used in various fields in recent years, such as autonomous driving and industrial inspection. Traditional object detection methods are mainly based on the sliding windows and the handcrafted features, which have problems such as insufficient understanding of image features and low accuracy of detection. With the rapid advancements in deep learning, convolutional neural networks (CNNs) and vision transformers have become fundamental components in object detection models. These components are capable of learning more advanced and deeper image properties, leading to a transformational breakthrough in the performance of object detection. In this review, we comprehensively review the representative object detection models from deep learning periods, tracing their architectural shifts and technological breakthroughs. Furthermore, we discuss key challenges and promising research directions in the object detection. This review aims to provide a comprehensive foundation for practitioners to enhance their understanding of object detection technologies.

目标检测是计算机视觉中最关键和最具挑战性的任务,近年来在自动驾驶和工业检测等各个领域得到了应用。传统的目标检测方法主要基于滑动窗口和手工特征,存在对图像特征理解不足、检测精度低等问题。随着深度学习的快速发展,卷积神经网络(cnn)和视觉变压器已经成为目标检测模型的基本组成部分。这些组件能够学习更高级和更深层次的图像属性,从而在目标检测性能方面实现转型突破。在这篇综述中,我们全面回顾了深度学习时期具有代表性的目标检测模型,追踪了它们的架构转变和技术突破。此外,我们还讨论了目标检测的关键挑战和有前景的研究方向。本综述旨在为从业者提供一个全面的基础,以提高他们对目标检测技术的理解。
{"title":"Object Detection Based on CNN and Vision-Transformer: A Survey","authors":"Jinfeng Cao,&nbsp;Bo Peng,&nbsp;Mingzhong Gao,&nbsp;Haichun Hao,&nbsp;Xinfang Li,&nbsp;Hongwei Mou","doi":"10.1049/cvi2.70028","DOIUrl":"10.1049/cvi2.70028","url":null,"abstract":"<p>Object detection is the most crucial and challenging task of computer vision and has been used in various fields in recent years, such as autonomous driving and industrial inspection. Traditional object detection methods are mainly based on the sliding windows and the handcrafted features, which have problems such as insufficient understanding of image features and low accuracy of detection. With the rapid advancements in deep learning, convolutional neural networks (CNNs) and vision transformers have become fundamental components in object detection models. These components are capable of learning more advanced and deeper image properties, leading to a transformational breakthrough in the performance of object detection. In this review, we comprehensively review the representative object detection models from deep learning periods, tracing their architectural shifts and technological breakthroughs. Furthermore, we discuss key challenges and promising research directions in the object detection. This review aims to provide a comprehensive foundation for practitioners to enhance their understanding of object detection technologies.</p>","PeriodicalId":56304,"journal":{"name":"IET Computer Vision","volume":"19 1","pages":""},"PeriodicalIF":1.3,"publicationDate":"2025-05-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1049/cvi2.70028","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144179248","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
FastVDT: Fast Transformer With Optimised Attention Masks and Positional Encoding for Visual Dialogue FastVDT:快速变压器与优化的注意力面具和位置编码的视觉对话
IF 1.3 4区 计算机科学 Q4 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-05-29 DOI: 10.1049/cvi2.70022
Qiangqiang He, Shuwei Qian, Chongjun Wang

The visual dialogue task requires computers to comprehend image content and preceding question-and-answer history to accurately answer related questions, with each round of dialogue providing the necessary historical context for subsequent interactions. Existing research typically processes multiple questions related to a single image as independent samples, which results in redundant modelling of the images and their captions and substantially increases computational costs. To address the challenges above, we introduce a fast transformer for visual dialogue, termed FastVDT, which utilises novel attention masks and continuous positional encoding. FastVDT models multiple image-related questions as an integrated entity, accurately processing prior conversation history in each dialogue round while predicting answers to multiple questions. Our method effectively captures the interrelations among questions and significantly reduces computational overhead. Experimental results demonstrate that our method delivers outstanding performance on the VisDial v0.9 and v1.0 datasets. FastVDT achieves comparable performance to VD-BERT and VU-BERT while reducing computational costs by 80% and 56%, respectively.

视觉对话任务需要计算机理解图像内容和之前的问答历史,以准确回答相关问题,每一轮对话都为后续互动提供必要的历史背景。现有的研究通常将与一张图像相关的多个问题作为独立的样本进行处理,这导致了图像及其标题的冗余建模,大大增加了计算成本。为了解决上述挑战,我们引入了一种用于视觉对话的快速转换器,称为FastVDT,它利用了新颖的注意力掩模和连续位置编码。FastVDT将多个与图像相关的问题建模为一个完整的实体,在每个对话轮中准确地处理之前的对话历史,同时预测多个问题的答案。我们的方法有效地捕获了问题之间的相互关系,大大减少了计算开销。实验结果表明,该方法在VisDial v0.9和v1.0数据集上具有出色的性能。FastVDT实现了与VD-BERT和VU-BERT相当的性能,同时将计算成本分别降低了80%和56%。
{"title":"FastVDT: Fast Transformer With Optimised Attention Masks and Positional Encoding for Visual Dialogue","authors":"Qiangqiang He,&nbsp;Shuwei Qian,&nbsp;Chongjun Wang","doi":"10.1049/cvi2.70022","DOIUrl":"10.1049/cvi2.70022","url":null,"abstract":"<p>The visual dialogue task requires computers to comprehend image content and preceding question-and-answer history to accurately answer related questions, with each round of dialogue providing the necessary historical context for subsequent interactions. Existing research typically processes multiple questions related to a single image as independent samples, which results in redundant modelling of the images and their captions and substantially increases computational costs. To address the challenges above, we introduce a fast transformer for visual dialogue, termed FastVDT, which utilises novel attention masks and continuous positional encoding. FastVDT models multiple image-related questions as an integrated entity, accurately processing prior conversation history in each dialogue round while predicting answers to multiple questions. Our method effectively captures the interrelations among questions and significantly reduces computational overhead. Experimental results demonstrate that our method delivers outstanding performance on the VisDial v0.9 and v1.0 datasets. FastVDT achieves comparable performance to VD-BERT and VU-BERT while reducing computational costs by 80% and 56%, respectively.</p>","PeriodicalId":56304,"journal":{"name":"IET Computer Vision","volume":"19 1","pages":""},"PeriodicalIF":1.3,"publicationDate":"2025-05-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1049/cvi2.70022","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144171519","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
End-to-End Cascaded Image Restoration and Object Detection for Rain and Fog Conditions 雨和雾条件下的端到端级联图像恢复和目标检测
IF 1.3 4区 计算机科学 Q4 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-05-19 DOI: 10.1049/cvi2.70021
Peng Li, Jun Ni, Dapeng Tao

Adverse weather conditions in real-world scenarios can degrade the performance of deep learning-based object detection models. A commonly used approach is to apply image restoration before object detection to improve degraded images. However, there is no direct correlation between the visual quality of image restoration and the object detection accuracy. Furthermore, image restoration and object detection have potential conflicting objectives, making joint optimisation difficult. To address this, we propose an end-to-end object detection network specifically designed for rainy and foggy conditions. Our approach cascades an image restoration subnetwork with a detection subnetwork and optimises them jointly through a shared objective. Specifically, we introduce an expanded dilated convolution block and a weather attention block to enhance the effectiveness and robustness of the restoration network under various weather degradations. Additionally, we incorporate an auxiliary alignment branch with feature alignment loss to align the features of restored and clean images within the detection backbone, enabling joint optimisation of both subnetworks. A novel training strategy is also proposed to further improve object detection performance under rainy and foggy conditions. Extensive experiments on the vehicle-rain-fog, VOC-fog and real-world fog datasets demonstrate that our method outperforms recent state-of-the-art approaches in image restoration quality and detection accuracy. The code is available at https://github.com/HappyPessimism/RainFog-Restoration-Detection.

现实场景中的恶劣天气条件会降低基于深度学习的目标检测模型的性能。一种常用的方法是在目标检测之前进行图像恢复,以改善退化图像。然而,图像恢复的视觉质量与目标检测精度之间没有直接的相关性。此外,图像恢复和目标检测具有潜在的冲突目标,使得联合优化变得困难。为了解决这个问题,我们提出了一个端到端的目标检测网络,专门为雨天和雾天条件设计。我们的方法将图像恢复子网与检测子网级联,并通过共享目标共同优化它们。具体来说,我们引入了一个扩展的扩张卷积块和一个天气关注块,以提高在各种天气退化下恢复网络的有效性和鲁棒性。此外,我们结合了一个具有特征对齐损失的辅助对齐分支,以在检测骨干内对齐恢复和干净图像的特征,从而实现两个子网的联合优化。提出了一种新的训练策略,以进一步提高在雨雾条件下的目标检测性能。在车辆雨雾、voc雾和真实世界雾数据集上进行的大量实验表明,我们的方法在图像恢复质量和检测精度方面优于最近最先进的方法。代码可在https://github.com/HappyPessimism/RainFog-Restoration-Detection上获得。
{"title":"End-to-End Cascaded Image Restoration and Object Detection for Rain and Fog Conditions","authors":"Peng Li,&nbsp;Jun Ni,&nbsp;Dapeng Tao","doi":"10.1049/cvi2.70021","DOIUrl":"10.1049/cvi2.70021","url":null,"abstract":"<p>Adverse weather conditions in real-world scenarios can degrade the performance of deep learning-based object detection models. A commonly used approach is to apply image restoration before object detection to improve degraded images. However, there is no direct correlation between the visual quality of image restoration and the object detection accuracy. Furthermore, image restoration and object detection have potential conflicting objectives, making joint optimisation difficult. To address this, we propose an end-to-end object detection network specifically designed for rainy and foggy conditions. Our approach cascades an image restoration subnetwork with a detection subnetwork and optimises them jointly through a shared objective. Specifically, we introduce an expanded dilated convolution block and a weather attention block to enhance the effectiveness and robustness of the restoration network under various weather degradations. Additionally, we incorporate an auxiliary alignment branch with feature alignment loss to align the features of restored and clean images within the detection backbone, enabling joint optimisation of both subnetworks. A novel training strategy is also proposed to further improve object detection performance under rainy and foggy conditions. Extensive experiments on the vehicle-rain-fog, VOC-fog and real-world fog datasets demonstrate that our method outperforms recent state-of-the-art approaches in image restoration quality and detection accuracy. The code is available at https://github.com/HappyPessimism/RainFog-Restoration-Detection.</p>","PeriodicalId":56304,"journal":{"name":"IET Computer Vision","volume":"19 1","pages":""},"PeriodicalIF":1.3,"publicationDate":"2025-05-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1049/cvi2.70021","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144091874","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IET Computer Vision
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1