首页 > 最新文献

Image and Vision Computing最新文献

英文 中文
Efficient 6DoF pose estimation for multi-instance objects from a single image 单幅图像中多实例对象的高效6DoF姿态估计
IF 4.2 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-02-01 Epub Date: 2025-12-16 DOI: 10.1016/j.imavis.2025.105882
Wen-Nung Lie , Lee Aing
Estimating 6 degrees of freedom poses for multiple objects from a single image and making it practical in industry is difficult since several metrics, like accuracy, speed and complexity must be traded. This study adopts a fast bottom-up approach to estimate poses for multi-instance objects in an image simultaneously. We design a convolutional neural network with simple end-to-end training to output 4 feature maps: error mask, semantic mask, center vector map and 6D coordinate map (6DCM). Specifically, 6DCM is capable of providing the rear-side 3D object point clouds information that are originally invisible from the camera's viewpoint. This procedure enriches shape information about target objects which can be used to construct each instance's 2D-3D correspondences for pose parameter estimation. Experimental results show that our proposed bottom-up approach is fast and can process a single image containing 7 objects at 25 frames per second with competitive accuracy to other top-down methods.
从一张图像中估计多个物体的6个自由度姿势并使其在工业中实用是很困难的,因为必须权衡精度、速度和复杂性等几个指标。本研究采用快速自底向上的方法同时估计图像中多实例物体的姿态。我们设计了一个简单的端到端训练卷积神经网络,输出4个特征图:错误掩码、语义掩码、中心向量图和6D坐标图(6DCM)。具体来说,6DCM能够提供最初从相机视点看不见的后侧3D物体点云信息。该过程丰富了目标物体的形状信息,可用于构建每个实例的2D-3D对应关系,用于姿态参数估计。实验结果表明,我们提出的自底向上方法速度快,可以以每秒25帧的速度处理包含7个物体的单幅图像,并且精度与其他自顶向下方法相当。
{"title":"Efficient 6DoF pose estimation for multi-instance objects from a single image","authors":"Wen-Nung Lie ,&nbsp;Lee Aing","doi":"10.1016/j.imavis.2025.105882","DOIUrl":"10.1016/j.imavis.2025.105882","url":null,"abstract":"<div><div>Estimating 6 degrees of freedom poses for multiple objects from a single image and making it practical in industry is difficult since several metrics, like accuracy, speed and complexity must be traded. This study adopts a fast bottom-up approach to estimate poses for multi-instance objects in an image simultaneously. We design a convolutional neural network with simple end-to-end training to output 4 feature maps: error mask, semantic mask, center vector map and 6D coordinate map (6DCM). Specifically, 6DCM is capable of providing the rear-side 3D object point clouds information that are originally invisible from the camera's viewpoint. This procedure enriches shape information about target objects which can be used to construct each instance's 2D-3D correspondences for pose parameter estimation. Experimental results show that our proposed bottom-up approach is fast and can process a single image containing 7 objects at 25 frames per second with competitive accuracy to other top-down methods.</div></div>","PeriodicalId":50374,"journal":{"name":"Image and Vision Computing","volume":"166 ","pages":"Article 105882"},"PeriodicalIF":4.2,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145790256","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
HEL-Net: Heterogeneous Ensemble Learning for comprehensive diabetic retinopathy multi-lesion segmentation via Mamba-UNet hell - net:基于Mamba-UNet的糖尿病视网膜病变多病灶分割的异构集成学习
IF 4.2 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-02-01 Epub Date: 2025-12-16 DOI: 10.1016/j.imavis.2025.105879
Lingyu Wu , Haiying Xia , Shuxiang Song , Yang Lan
Diabetic Retinopathy (DR) is the leading cause of blindness in adults with diabetes. Early automated detection of DR lesions is crucial for preventing vision loss and assisting ophthalmologists in treatment. However, accurately segmenting multiple types of DR lesions poses significant challenges due to their large diversity in size, shape, and location, as well as the conflict in feature modeling between local details and long-range dependencies. To address these issues, we propose a novel Heterogeneous Ensemble Learning Network (HEL-Net) specifically designed for four-lesion segmentation. HEL-Net comprises two ensemble stages: the first stage utilizes Mamba-UNet to generate coarse multi-lesion prediction results, which serve as contextual priors for the second stage, forming a multi-perspective lesion navigation strategy. The second stage employs a heterogeneous structure, integrating specialized networks (Mamba-UNet and U-Net) tailored to different lesion characteristics. Mamba-UNet excels in capturing large lesions by modeling long-range dependencies, while U-Net focuses on small lesions with significant local features. The heterogeneous ensemble framework leverages their complementary strengths to promote comprehensive lesion feature learning. Extensive quantitative and qualitative evaluations on two public datasets (IDRiD and DDR) demonstrate that our HEL-Net achieves competitive performance compared to state-of-the-art methods, achieving an mAUPR of 69.52%, mDice of 67.40%, and mIoU of 51.99% on the IDRiD dataset.
糖尿病视网膜病变(DR)是导致成人糖尿病患者失明的主要原因。早期自动检测DR病变对于预防视力丧失和协助眼科医生治疗至关重要。然而,由于多种类型的DR病变在大小、形状和位置上的巨大差异,以及局部细节和远程依赖关系之间的特征建模冲突,准确分割多种类型的DR病变带来了巨大的挑战。为了解决这些问题,我们提出了一种新的异构集成学习网络(hell - net),专门用于四病灶分割。HEL-Net包括两个集成阶段:第一阶段利用Mamba-UNet生成粗糙的多病变预测结果,作为第二阶段的上下文先验,形成多视角病变导航策略。第二阶段采用异质结构,整合针对不同病变特征定制的专用网络(Mamba-UNet和U-Net)。Mamba-UNet擅长通过建模远程依赖关系来捕获大型病变,而U-Net则专注于具有重要局部特征的小病变。异构集成框架利用它们的互补优势,促进全面的病变特征学习。在两个公共数据集(IDRiD和DDR)上进行的大量定量和定性评估表明,与最先进的方法相比,我们的hell - net的性能具有竞争力,在IDRiD数据集上实现了69.52%的mAUPR, 67.40%的mdevice和51.99%的mIoU。
{"title":"HEL-Net: Heterogeneous Ensemble Learning for comprehensive diabetic retinopathy multi-lesion segmentation via Mamba-UNet","authors":"Lingyu Wu ,&nbsp;Haiying Xia ,&nbsp;Shuxiang Song ,&nbsp;Yang Lan","doi":"10.1016/j.imavis.2025.105879","DOIUrl":"10.1016/j.imavis.2025.105879","url":null,"abstract":"<div><div>Diabetic Retinopathy (DR) is the leading cause of blindness in adults with diabetes. Early automated detection of DR lesions is crucial for preventing vision loss and assisting ophthalmologists in treatment. However, accurately segmenting multiple types of DR lesions poses significant challenges due to their large diversity in size, shape, and location, as well as the conflict in feature modeling between local details and long-range dependencies. To address these issues, we propose a novel Heterogeneous Ensemble Learning Network (<em>HEL-Net</em>) specifically designed for four-lesion segmentation. HEL-Net comprises two ensemble stages: the first stage utilizes Mamba-UNet to generate coarse multi-lesion prediction results, which serve as contextual priors for the second stage, forming a multi-perspective lesion navigation strategy. The second stage employs a heterogeneous structure, integrating specialized networks (Mamba-UNet and U-Net) tailored to different lesion characteristics. Mamba-UNet excels in capturing large lesions by modeling long-range dependencies, while U-Net focuses on small lesions with significant local features. The heterogeneous ensemble framework leverages their complementary strengths to promote comprehensive lesion feature learning. Extensive quantitative and qualitative evaluations on two public datasets (IDRiD and DDR) demonstrate that our HEL-Net achieves competitive performance compared to state-of-the-art methods, achieving an mAUPR of 69.52%, mDice of 67.40%, and mIoU of 51.99% on the IDRiD dataset.</div></div>","PeriodicalId":50374,"journal":{"name":"Image and Vision Computing","volume":"166 ","pages":"Article 105879"},"PeriodicalIF":4.2,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145790365","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
LCFusion: Infrared and visible image fusion network based on local contour enhancement LCFusion:基于局部轮廓增强的红外与可见光图像融合网络
IF 4.2 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-02-01 Epub Date: 2025-12-03 DOI: 10.1016/j.imavis.2025.105856
Yitong Yang , Lei Zhu , Xinyang Yao , Hua Wang , Yang Pan , Bo Zhang
Infrared and visible light image fusion aims to generate integrated representations that synergistically preserve salient thermal targets in the infrared modality and high-resolution textural details in the visible light modality. However, existing methods face two core challenges: First, high-frequency noise in visible images, such as sensor noise and nonuniform illumination artifacts, is often highly coupled with effective textures. Traditional fusion paradigms readily amplify noise interference while enhancing details, leading to structural distortion and visual graininess in fusion results. Second, mainstream approaches predominantly rely on simple aggregation operations like feature stitching or linear weighting, lacking deep modeling of cross-modal semantic correlations. This prevents adaptive interaction and collaborative enhancement of complementary information between modalities, creating a significant trade-off between target saliency and detail preservation. To address these challenges, we propose a dual-branch fusion network based on local contour enhancement. Specifically, it distinguishes and enhances meaningful contour details in a learnable manner while suppressing meaningless noise, thereby purifying the detail information used for fusion at its source. Cross-attention weights are computed based on feature representations extracted from different modal branches, enabling a feature selection mechanism that facilitates dynamic cross-modal interaction between infrared and visible light information. We evaluate our method against 11 state-of-the-art deep learning-based fusion approaches across four benchmark datasets using both subjective assessments and objective metrics. The experimental results demonstrate superior performance on public datasets. Furthermore, YOLOv12-based detection tests reveal that our method achieves higher confidence scores and better overall detection performance compared to other fusion techniques.
红外和可见光图像融合的目的是生成集成的表示,在红外模态中协同保存显著的热目标,在可见光模态中协同保存高分辨率的纹理细节。然而,现有方法面临两个核心挑战:首先,可见光图像中的高频噪声,如传感器噪声和非均匀照明伪影,通常与有效纹理高度耦合。传统的融合范式容易在增强细节的同时放大噪声干扰,导致融合结果的结构失真和视觉颗粒化。其次,主流方法主要依赖于简单的聚合操作,如特征拼接或线性加权,缺乏跨模态语义相关性的深度建模。这阻碍了模式之间互补信息的自适应交互和协作增强,在目标显著性和细节保存之间产生了重大权衡。为了解决这些问题,我们提出了一种基于局部轮廓增强的双分支融合网络。具体而言,它以可学习的方式区分和增强有意义的轮廓细节,同时抑制无意义的噪声,从而从源头净化用于融合的细节信息。交叉关注权重基于从不同模态分支提取的特征表示计算,实现了一种特征选择机制,促进了红外和可见光信息之间的动态跨模态交互。我们使用主观评估和客观指标对四个基准数据集上11种最先进的基于深度学习的融合方法进行了评估。实验结果表明,该方法在公共数据集上具有优异的性能。此外,基于yolov12的检测测试表明,与其他融合技术相比,我们的方法获得了更高的置信度分数和更好的整体检测性能。
{"title":"LCFusion: Infrared and visible image fusion network based on local contour enhancement","authors":"Yitong Yang ,&nbsp;Lei Zhu ,&nbsp;Xinyang Yao ,&nbsp;Hua Wang ,&nbsp;Yang Pan ,&nbsp;Bo Zhang","doi":"10.1016/j.imavis.2025.105856","DOIUrl":"10.1016/j.imavis.2025.105856","url":null,"abstract":"<div><div>Infrared and visible light image fusion aims to generate integrated representations that synergistically preserve salient thermal targets in the infrared modality and high-resolution textural details in the visible light modality. However, existing methods face two core challenges: First, high-frequency noise in visible images, such as sensor noise and nonuniform illumination artifacts, is often highly coupled with effective textures. Traditional fusion paradigms readily amplify noise interference while enhancing details, leading to structural distortion and visual graininess in fusion results. Second, mainstream approaches predominantly rely on simple aggregation operations like feature stitching or linear weighting, lacking deep modeling of cross-modal semantic correlations. This prevents adaptive interaction and collaborative enhancement of complementary information between modalities, creating a significant trade-off between target saliency and detail preservation. To address these challenges, we propose a dual-branch fusion network based on local contour enhancement. Specifically, it distinguishes and enhances meaningful contour details in a learnable manner while suppressing meaningless noise, thereby purifying the detail information used for fusion at its source. Cross-attention weights are computed based on feature representations extracted from different modal branches, enabling a feature selection mechanism that facilitates dynamic cross-modal interaction between infrared and visible light information. We evaluate our method against 11 state-of-the-art deep learning-based fusion approaches across four benchmark datasets using both subjective assessments and objective metrics. The experimental results demonstrate superior performance on public datasets. Furthermore, YOLOv12-based detection tests reveal that our method achieves higher confidence scores and better overall detection performance compared to other fusion techniques.</div></div>","PeriodicalId":50374,"journal":{"name":"Image and Vision Computing","volume":"166 ","pages":"Article 105856"},"PeriodicalIF":4.2,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145685263","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Privacy-aware knowledge distillation for retinal scans de-identification through adversarial perturbations 通过对抗性扰动进行视网膜扫描去识别的隐私感知知识蒸馏
IF 4.2 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-02-01 Epub Date: 2025-11-26 DOI: 10.1016/j.imavis.2025.105849
Keramat Allah Ghaffary, Mohsen Ebrahimi Moghaddam
The rapid increase in retinal scan datasets over the past several years has played a pivotal role in advancing the deep learning approaches for retinopathy detection. However, given the retinal vessel pattern as an accurate biometric identifier, sharing this sensitive data raises critical ethical and legal considerations pertaining to individual re-identification and privacy violations. To mitigate these concerns, we propose a knowledge distillation-based approach that aims to disrupt identity recognition models by applying imperceptible adversarial perturbations to retinal scans while simultaneously preserving their utility for medical purposes. A multi-objective loss function is utilized for this approach, consisting of several terms that effectively guide the generator part of the student network to learn the effective perturbation needed to balance these conflicting goals. Extensive experiments and explainable analysis are conducted on three public fundus datasets using five deep architectures for identity recognition and retinopathy detection. With an average identity fooling rate of 94.67% and average retinopathy detection accuracy of 97.52% for multiple unseen models, the proposed approach outperforms state-of-the-art methods in balancing between medical utility and enhancing patient privacy.
在过去几年中,视网膜扫描数据集的快速增长在推进视网膜病变检测的深度学习方法方面发挥了关键作用。然而,鉴于视网膜血管模式作为准确的生物识别标识符,共享这些敏感数据会引起与个人重新识别和侵犯隐私有关的关键伦理和法律考虑。为了减轻这些担忧,我们提出了一种基于知识蒸馏的方法,旨在通过对视网膜扫描应用难以察觉的对抗性扰动来破坏身份识别模型,同时保留其在医疗目的上的效用。这种方法使用了一个多目标损失函数,它由几个项组成,这些项有效地指导学生网络的生成器部分学习平衡这些冲突目标所需的有效扰动。在三个公共眼底数据集上进行了广泛的实验和可解释的分析,使用五种深度架构进行身份识别和视网膜病变检测。对于多个未见模型,该方法的平均识别欺骗率为94.67%,平均视网膜病变检测准确率为97.52%,在平衡医疗效用和增强患者隐私方面优于现有方法。
{"title":"Privacy-aware knowledge distillation for retinal scans de-identification through adversarial perturbations","authors":"Keramat Allah Ghaffary,&nbsp;Mohsen Ebrahimi Moghaddam","doi":"10.1016/j.imavis.2025.105849","DOIUrl":"10.1016/j.imavis.2025.105849","url":null,"abstract":"<div><div>The rapid increase in retinal scan datasets over the past several years has played a pivotal role in advancing the deep learning approaches for retinopathy detection. However, given the retinal vessel pattern as an accurate biometric identifier, sharing this sensitive data raises critical ethical and legal considerations pertaining to individual re-identification and privacy violations. To mitigate these concerns, we propose a knowledge distillation-based approach that aims to disrupt identity recognition models by applying imperceptible adversarial perturbations to retinal scans while simultaneously preserving their utility for medical purposes. A multi-objective loss function is utilized for this approach, consisting of several terms that effectively guide the generator part of the student network to learn the effective perturbation needed to balance these conflicting goals. Extensive experiments and explainable analysis are conducted on three public fundus datasets using five deep architectures for identity recognition and retinopathy detection. With an average identity fooling rate of 94.67% and average retinopathy detection accuracy of 97.52% for multiple unseen models, the proposed approach outperforms state-of-the-art methods in balancing between medical utility and enhancing patient privacy.</div></div>","PeriodicalId":50374,"journal":{"name":"Image and Vision Computing","volume":"166 ","pages":"Article 105849"},"PeriodicalIF":4.2,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145685262","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Gaussian landmarks tracking-based real-time splatting reconstruction model 基于高斯地标跟踪的飞溅实时重建模型
IF 4.2 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-02-01 Epub Date: 2025-12-08 DOI: 10.1016/j.imavis.2025.105869
Donglin Zhu, Zhongli Wang, Xiaoyang Fan, Miao Chen, Jiuyu Chen
Real-time and high-quality scene reconstruction remains a critical challenge for robotics applications. 3D Gaussian Splatting (3DGS) demonstrates remarkable capabilities in scene rendering. However, its integration with SLAM systems confronts two critical limitations: (1) slow pose tracking caused by full-frame rendering multiple times, and (2) susceptibility to environmental variations such as illumination variations and motion blur. To alleviate these issues, this paper proposes gaussian landmarks-based real-time reconstruction framework — GLT-SLAM, which composes of a ray casting-driven tracking module, a multi-modal keyframe selector, and an incremental geometric–photometric mapping module. To avoid redundant rendering computations, the tracking module achieves efficient 3D-2D correspondence by encoding Gaussian landmark-emitted rays and fusing attention scores. Furthermore, to enhance the framework’s robustness against complex environmental conditions, the keyframe selector balances multiple influencing factors including image quality, tracking uncertainty, information entropy, and feature overlap ratios. Finally, to achieve a compact map representation, the mapping module adds only Gaussian primitives of points, lines, and planes, and performs global map optimization through joint photometric–geometric constraints. Experimental results on the Replica, TUM RGB-D, and BJTU datasets demonstrate that the proposed method achieves a real-time processing rate of over 30 Hz on a platform with an NVIDIA RTX 3090, demonstrating a 19% higher efficiency than the fastest Photo-SLAM method while significantly outperforming other baseline methods in both localization and mapping accuracy. The source code will be available on GitHub.1
实时和高质量的场景重建仍然是机器人应用的关键挑战。三维高斯溅射(3DGS)在场景渲染中表现出非凡的能力。然而,它与SLAM系统的集成面临两个关键限制:(1)多次绘制全帧导致的姿态跟踪缓慢;(2)易受光照变化和运动模糊等环境变化的影响。为了解决这些问题,本文提出了基于高斯地标的实时重建框架GLT-SLAM,该框架由光线投射驱动的跟踪模块、多模态关键帧选择器和增量几何光度映射模块组成。为了避免冗余的渲染计算,跟踪模块通过对高斯地标发射射线进行编码并融合注意分数来实现高效的3D-2D对应。此外,为了增强框架对复杂环境条件的鲁棒性,关键帧选择器平衡了多个影响因素,包括图像质量、跟踪不确定性、信息熵和特征重叠率。最后,为了实现紧凑的地图表示,映射模块仅添加点、线、面高斯原语,并通过光度-几何联合约束进行全局地图优化。在Replica, TUM RGB-D和BJTU数据集上的实验结果表明,该方法在NVIDIA RTX 3090平台上实现了超过30 Hz的实时处理速率,比最快的Photo-SLAM方法效率高出19%,同时在定位和制图精度方面显着优于其他基准方法。源代码可以在GitHub.1上获得
{"title":"Gaussian landmarks tracking-based real-time splatting reconstruction model","authors":"Donglin Zhu,&nbsp;Zhongli Wang,&nbsp;Xiaoyang Fan,&nbsp;Miao Chen,&nbsp;Jiuyu Chen","doi":"10.1016/j.imavis.2025.105869","DOIUrl":"10.1016/j.imavis.2025.105869","url":null,"abstract":"<div><div>Real-time and high-quality scene reconstruction remains a critical challenge for robotics applications. 3D Gaussian Splatting (3DGS) demonstrates remarkable capabilities in scene rendering. However, its integration with SLAM systems confronts two critical limitations: (1) slow pose tracking caused by full-frame rendering multiple times, and (2) susceptibility to environmental variations such as illumination variations and motion blur. To alleviate these issues, this paper proposes gaussian landmarks-based real-time reconstruction framework — GLT-SLAM, which composes of a ray casting-driven tracking module, a multi-modal keyframe selector, and an incremental geometric–photometric mapping module. To avoid redundant rendering computations, the tracking module achieves efficient 3D-2D correspondence by encoding Gaussian landmark-emitted rays and fusing attention scores. Furthermore, to enhance the framework’s robustness against complex environmental conditions, the keyframe selector balances multiple influencing factors including image quality, tracking uncertainty, information entropy, and feature overlap ratios. Finally, to achieve a compact map representation, the mapping module adds only Gaussian primitives of points, lines, and planes, and performs global map optimization through joint photometric–geometric constraints. Experimental results on the Replica, TUM RGB-D, and BJTU datasets demonstrate that the proposed method achieves a real-time processing rate of over 30 Hz on a platform with an NVIDIA RTX 3090, demonstrating a 19% higher efficiency than the fastest Photo-SLAM method while significantly outperforming other baseline methods in both localization and mapping accuracy. The source code will be available on GitHub.<span><span><sup>1</sup></span></span></div></div>","PeriodicalId":50374,"journal":{"name":"Image and Vision Computing","volume":"166 ","pages":"Article 105869"},"PeriodicalIF":4.2,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145737142","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
PCNet3D++: A pillar-based cascaded 3D object detection model with an enhanced 2D backbone pcnet3d++:一个基于柱的级联3D目标检测模型,具有增强的2D骨干
IF 4.2 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-02-01 Epub Date: 2025-11-29 DOI: 10.1016/j.imavis.2025.105854
Thurimerla Prasanth , Ram Prasad Padhy , B. Sivaselvan
Autonomous Vehicles (AVs) depend on sophisticated perception systems to serve as the vital component of intelligent transportation to ensure secure and smooth navigation. Perception is an essential component of AVs and enables real-time analysis and understanding of the environment for effective decision-making. 3D object detection (3D-OD) is crucial among perception tasks as it accurately determines the 3D geometry and spatial positioning of surrounding objects. The commonly used modalities for 3D-OD are camera, LiDAR, and sensor fusion. In this work, we propose a LiDAR-based 3D-OD approach using point cloud data. The proposed model achieves superior performance while maintaining computational efficiency. This approach utilizes Pillar-based LiDAR processing and uses only 2D convolutions. The model pipeline becomes simple and more efficient by employing only 2D convolutions. We propose a Cascaded Convolutional Backbone (CCB) integrated with 1 × 1 convolutions to improve detection accuracy. We combined the fast Pillar-based encoding with our lightweight backbone. The proposed model reduces complexity to make it well-suited for real-time navigation of an AV. We evaluated our model on the official KITTI test server. The model results are decent in 3D and Bird’s Eye View (BEV) detection benchmarks for the car and cyclist classes. The results of our proposed model are featured on the official KITTI leaderboard.
自动驾驶汽车(AVs)依靠复杂的感知系统作为智能交通的重要组成部分,以确保安全顺畅的导航。感知是自动驾驶汽车的重要组成部分,能够实时分析和理解环境,从而做出有效的决策。3D物体检测(3D- od)在感知任务中至关重要,因为它可以准确地确定周围物体的三维几何形状和空间定位。3D-OD常用的模式是摄像头、激光雷达和传感器融合。在这项工作中,我们提出了一种基于激光雷达的3D-OD方法,使用点云数据。该模型在保持计算效率的同时,取得了较好的性能。这种方法利用基于柱的激光雷达处理,只使用二维卷积。通过只使用二维卷积,模型管道变得简单和高效。为了提高检测精度,我们提出了一种集成了1 × 1卷积的级联卷积主干(CCB)。我们将基于支柱的快速编码与轻量级主干结合起来。提出的模型降低了复杂性,使其非常适合自动驾驶汽车的实时导航。我们在官方KITTI测试服务器上评估了我们的模型。该模型在汽车和自行车类的3D和鸟瞰(BEV)检测基准中效果良好。我们提出的模型的结果在官方KITTI排行榜上有特色。
{"title":"PCNet3D++: A pillar-based cascaded 3D object detection model with an enhanced 2D backbone","authors":"Thurimerla Prasanth ,&nbsp;Ram Prasad Padhy ,&nbsp;B. Sivaselvan","doi":"10.1016/j.imavis.2025.105854","DOIUrl":"10.1016/j.imavis.2025.105854","url":null,"abstract":"<div><div>Autonomous Vehicles (AVs) depend on sophisticated perception systems to serve as the vital component of intelligent transportation to ensure secure and smooth navigation. Perception is an essential component of AVs and enables real-time analysis and understanding of the environment for effective decision-making. 3D object detection (3D-OD) is crucial among perception tasks as it accurately determines the 3D geometry and spatial positioning of surrounding objects. The commonly used modalities for 3D-OD are camera, LiDAR, and sensor fusion. In this work, we propose a LiDAR-based 3D-OD approach using point cloud data. The proposed model achieves superior performance while maintaining computational efficiency. This approach utilizes Pillar-based LiDAR processing and uses only 2D convolutions. The model pipeline becomes simple and more efficient by employing only 2D convolutions. We propose a Cascaded Convolutional Backbone (CCB) integrated with 1 × 1 convolutions to improve detection accuracy. We combined the fast Pillar-based encoding with our lightweight backbone. The proposed model reduces complexity to make it well-suited for real-time navigation of an AV. We evaluated our model on the official KITTI test server. The model results are decent in 3D and Bird’s Eye View (BEV) detection benchmarks for the car and cyclist classes. The results of our proposed model are featured on the official KITTI leaderboard.</div></div>","PeriodicalId":50374,"journal":{"name":"Image and Vision Computing","volume":"166 ","pages":"Article 105854"},"PeriodicalIF":4.2,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145685261","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Distinct Polyp Generator Network for polyp segmentation 用于息肉分割的独特息肉生成器网络
IF 4.2 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-02-01 Epub Date: 2025-11-29 DOI: 10.1016/j.imavis.2025.105847
Huan Wan , Jing Ai , Jing Liu , Xin Wei , Jinshan Zeng , Jianyi Wan
Accurate polyp segmentation from the colonoscopy images is crucial for diagnosing and treating colorectal diseases. Although many automatic polyp segmentation models have been proposed and achieved good progress, they still suffer from under-segmentation or over-segmentation problems caused by the characteristics of colonoscopy images: blurred boundaries and widely varied polyp sizes. To address these problems, we propose a novel model, the Distinct Polyp Generator Network (DPG-Net), for polyp segmentation. In DPG-Net, a Feature Progressive Enhancement Module (FPEM) and a Dynamical Aggregation Module (DAM) are developed. The proposed FPEM is responsible for enhancing the polyps and polyp boundaries by jointly utilizing the boundary information and global prior information. Simultaneously, a DAM is developed to integrate all decoding features based on their own traits and detect polyps with various sizes. Finally, accurate segmentation results are obtained. Extensive experiments on five widely used datasets demonstrate that the proposed DPG-Net model is superior to the state-of-the-art models. To evaluate the cross-domain generalization ability, we adopt the proposed DPG-Net for the skin lesion segmentation task. Again, experimental results show that our DPG-Net achieves advanced performance in this task, which verifies the strong generalizability of DPG-Net.
从结肠镜图像中准确分割息肉是诊断和治疗结直肠疾病的关键。虽然已经提出了许多自动息肉分割模型并取得了良好的进展,但由于结肠镜图像边界模糊、息肉大小变化较大等特点,仍然存在分割不足或分割过度的问题。为了解决这些问题,我们提出了一个新的模型,独特的息肉生成器网络(DPG-Net),用于息肉分割。在DPG-Net中,开发了特征递进增强模块(FPEM)和动态聚合模块(DAM)。该算法利用边界信息和全局先验信息对息肉和息肉边界进行增强。同时,开发了一种基于自身特征整合所有解码特征的DAM,用于检测不同大小的息肉。最后得到准确的分割结果。在五个广泛使用的数据集上进行的大量实验表明,所提出的DPG-Net模型优于最先进的模型。为了评估该算法的跨域泛化能力,我们将提出的DPG-Net用于皮肤病变分割任务。实验结果再次表明,我们的DPG-Net在该任务中取得了先进的性能,验证了DPG-Net的强泛化性。
{"title":"Distinct Polyp Generator Network for polyp segmentation","authors":"Huan Wan ,&nbsp;Jing Ai ,&nbsp;Jing Liu ,&nbsp;Xin Wei ,&nbsp;Jinshan Zeng ,&nbsp;Jianyi Wan","doi":"10.1016/j.imavis.2025.105847","DOIUrl":"10.1016/j.imavis.2025.105847","url":null,"abstract":"<div><div>Accurate polyp segmentation from the colonoscopy images is crucial for diagnosing and treating colorectal diseases. Although many automatic polyp segmentation models have been proposed and achieved good progress, they still suffer from under-segmentation or over-segmentation problems caused by the characteristics of colonoscopy images: blurred boundaries and widely varied polyp sizes. To address these problems, we propose a novel model, the Distinct Polyp Generator Network (DPG-Net), for polyp segmentation. In DPG-Net, a Feature Progressive Enhancement Module (FPEM) and a Dynamical Aggregation Module (DAM) are developed. The proposed FPEM is responsible for enhancing the polyps and polyp boundaries by jointly utilizing the boundary information and global prior information. Simultaneously, a DAM is developed to integrate all decoding features based on their own traits and detect polyps with various sizes. Finally, accurate segmentation results are obtained. Extensive experiments on five widely used datasets demonstrate that the proposed DPG-Net model is superior to the state-of-the-art models. To evaluate the cross-domain generalization ability, we adopt the proposed DPG-Net for the skin lesion segmentation task. Again, experimental results show that our DPG-Net achieves advanced performance in this task, which verifies the strong generalizability of DPG-Net.</div></div>","PeriodicalId":50374,"journal":{"name":"Image and Vision Computing","volume":"166 ","pages":"Article 105847"},"PeriodicalIF":4.2,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145685260","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Combining short-term and long-term memory for robust visual tracking 结合短期和长期记忆,实现强大的视觉跟踪
IF 4.2 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-02-01 Epub Date: 2025-11-29 DOI: 10.1016/j.imavis.2025.105850
Zifan Rui , Xiaoxiao Wang , Yiteng Yang , Guang Han
In visual object tracking, addressing challenges such as target appearance deformation and occlusion has attracted increasing attention. To this end, this paper proposes CSLMTrack, a multiple memory tracking model that more comprehensively reflects the human memory mechanism. It contains short-term and long-term memory modules, as well as a novel feed-forward network TFFN for temporal information aggregation. A dynamic memory update strategy including memory, information transfer, recall, and forgetting processes is also designed, which can effectively avoid memory explosion while integrating memory elements into the tracking network. Extensive experiments conducted on multiple challenging benchmarks demonstrate that CSLMTrack achieves impressive performance, reaching SOTA-level performance compared to state-of-the-art trackers.
在视觉目标跟踪中,如何解决目标外观变形和遮挡等问题越来越受到人们的关注。为此,本文提出了一种更全面反映人类记忆机制的多重记忆跟踪模型CSLMTrack。它包括短期和长期记忆模块,以及一种新颖的前馈网络TFFN,用于时间信息聚合。设计了包含记忆、信息传递、回忆和遗忘过程的动态记忆更新策略,在将记忆元素整合到跟踪网络的同时,有效避免了记忆爆炸。在多个具有挑战性的基准测试中进行的大量实验表明,CSLMTrack取得了令人印象深刻的性能,与最先进的跟踪器相比,达到了sota级别的性能。
{"title":"Combining short-term and long-term memory for robust visual tracking","authors":"Zifan Rui ,&nbsp;Xiaoxiao Wang ,&nbsp;Yiteng Yang ,&nbsp;Guang Han","doi":"10.1016/j.imavis.2025.105850","DOIUrl":"10.1016/j.imavis.2025.105850","url":null,"abstract":"<div><div>In visual object tracking, addressing challenges such as target appearance deformation and occlusion has attracted increasing attention. To this end, this paper proposes CSLMTrack, a multiple memory tracking model that more comprehensively reflects the human memory mechanism. It contains short-term and long-term memory modules, as well as a novel feed-forward network TFFN for temporal information aggregation. A dynamic memory update strategy including memory, information transfer, recall, and forgetting processes is also designed, which can effectively avoid memory explosion while integrating memory elements into the tracking network. Extensive experiments conducted on multiple challenging benchmarks demonstrate that CSLMTrack achieves impressive performance, reaching SOTA-level performance compared to state-of-the-art trackers.</div></div>","PeriodicalId":50374,"journal":{"name":"Image and Vision Computing","volume":"166 ","pages":"Article 105850"},"PeriodicalIF":4.2,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145685266","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A deformable registration framework for brain MR images based on a dual-channel fusion strategy using GMamba 一种基于双通道融合策略的脑磁共振图像形变配准框架
IF 4.2 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-02-01 Epub Date: 2025-12-07 DOI: 10.1016/j.imavis.2025.105868
Liwei Deng , Songyu Chen , Xin Yang , Sijuan Huang , Jing Wang
Medical image registration has important applications in medical image analysis. Although deep learning-based registration methods are widely recognized, there is still performance improvement space for existing algorithms due to the complex physiological structure of brain images. In this paper, we aim to propose a deformable medical image registration method that is highly accurate and capable of handling complex physiological structures. Therefore, we propose DFMNet, a dual-channel fusion method based on GMamba, to achieve accurate brain MRI image registration. Compared with state-of-the-art networks like TransMorph, DFMNet has a dual-channel network structure with different fusion strategies. We propose the GMamba block to efficiently capture the remote dependencies in moving and fixed image features. Meanwhile, we propose a context extraction channel to enhance the texture structure of the image content. In addition, we designed a weighted fusion block to help the features of the two channels can be fused efficiently. Extensive experiments on three public brain datasets demonstrate the effectiveness of DFMNet. The experimental results demonstrate that DFMNet outperforms multiple current state-of-the-art deformable registration methods in structural registration of brain images.
医学图像配准在医学图像分析中有着重要的应用。尽管基于深度学习的配准方法得到了广泛的认可,但由于脑图像的生理结构复杂,现有的配准算法仍有性能提升的空间。在本文中,我们旨在提出一种高精度且能够处理复杂生理结构的可变形医学图像配准方法。为此,我们提出了一种基于GMamba的双通道融合方法DFMNet来实现脑MRI图像的精确配准。与TransMorph等最先进的网络相比,DFMNet采用双通道网络结构,融合策略不同。我们提出了GMamba块来有效地捕获移动和固定图像特征中的远程依赖关系。同时,我们提出了一种上下文提取通道来增强图像内容的纹理结构。此外,我们还设计了一个加权融合块,以帮助有效地融合两个通道的特征。在三个公共脑数据集上的大量实验证明了DFMNet的有效性。实验结果表明,DFMNet在脑图像结构配准方面优于当前多种可变形配准方法。
{"title":"A deformable registration framework for brain MR images based on a dual-channel fusion strategy using GMamba","authors":"Liwei Deng ,&nbsp;Songyu Chen ,&nbsp;Xin Yang ,&nbsp;Sijuan Huang ,&nbsp;Jing Wang","doi":"10.1016/j.imavis.2025.105868","DOIUrl":"10.1016/j.imavis.2025.105868","url":null,"abstract":"<div><div>Medical image registration has important applications in medical image analysis. Although deep learning-based registration methods are widely recognized, there is still performance improvement space for existing algorithms due to the complex physiological structure of brain images. In this paper, we aim to propose a deformable medical image registration method that is highly accurate and capable of handling complex physiological structures. Therefore, we propose DFMNet, a dual-channel fusion method based on GMamba, to achieve accurate brain MRI image registration. Compared with state-of-the-art networks like TransMorph, DFMNet has a dual-channel network structure with different fusion strategies. We propose the GMamba block to efficiently capture the remote dependencies in moving and fixed image features. Meanwhile, we propose a context extraction channel to enhance the texture structure of the image content. In addition, we designed a weighted fusion block to help the features of the two channels can be fused efficiently. Extensive experiments on three public brain datasets demonstrate the effectiveness of DFMNet. The experimental results demonstrate that DFMNet outperforms multiple current state-of-the-art deformable registration methods in structural registration of brain images.</div></div>","PeriodicalId":50374,"journal":{"name":"Image and Vision Computing","volume":"166 ","pages":"Article 105868"},"PeriodicalIF":4.2,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145737137","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
PASS: Peer-agreement based sample selection for training with instance dependent noisy labels PASS:基于同行协议的样本选择,用于与实例相关的噪声标签的训练
IF 4.2 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-02-01 Epub Date: 2025-12-16 DOI: 10.1016/j.imavis.2025.105877
Arpit Garg , Cuong Nguyen , Rafael Felix , Thanh-Toan Do , Gustavo Carneiro
Deep learning encounters significant challenges in the form of noisy-label samples, which can cause the overfitting of trained models. A primary challenge in learning with noisy-label (LNL) techniques is their ability to differentiate between hard samples (clean-label samples near the decision boundary) and instance-dependent noisy (IDN) label samples to allow these samples to be treated differently during training. Existing methodologies to identify IDN samples, including the small-loss hypothesis and feature-based selection, have demonstrated limited efficacy, thus impeding their effectiveness in dealing with real-world label noise. We present Peer-Agreement-based Sample Selection (PASS), a novel approach that utilises three classifiers, where a consensus-driven agreement between two models accurately differentiates between clean and noisy-label IDN samples to train the third model. In contrast to current techniques, PASS is specifically designed to address the complexities of IDN, where noise patterns are correlated with instance features. Our approach seamlessly integrates with existing LNL algorithms to enhance the accuracy of detecting both noisy and clean samples. Comprehensive experiments conducted on simulated benchmarks (CIFAR-100 and Red mini-ImageNet) and real-world datasets (Animal-10N, CIFAR-N, Clothing1M, and mini-WebVision) demonstrated that PASS substantially improved the performance of multiple state-of-the-art methods. This technique achieves superior classification accuracy, particularly in scenarios with high noise levels.1
深度学习在噪声标签样本方面遇到了重大挑战,这可能导致训练模型的过拟合。使用噪声标签(LNL)技术学习的主要挑战是区分硬样本(靠近决策边界的干净标签样本)和实例相关噪声(IDN)标签样本的能力,以便在训练期间对这些样本进行不同的处理。现有的识别IDN样本的方法,包括小损失假设和基于特征的选择,已经证明了有限的有效性,从而阻碍了它们在处理现实世界标签噪声方面的有效性。我们提出了基于同行协议的样本选择(PASS),这是一种利用三个分类器的新方法,其中两个模型之间的共识驱动协议准确区分干净和噪声标签的IDN样本,以训练第三个模型。与目前的技术相比,PASS是专门为解决IDN的复杂性而设计的,其中噪声模式与实例特征相关。我们的方法与现有的LNL算法无缝集成,以提高检测噪声和干净样本的准确性。在模拟基准测试(CIFAR-100和Red mini-ImageNet)和真实数据集(Animal-10N、CIFAR-N、Clothing1M和mini-WebVision)上进行的综合实验表明,PASS大大提高了多种最先进方法的性能。该技术实现了更高的分类精度,特别是在高噪声水平的情况下
{"title":"PASS: Peer-agreement based sample selection for training with instance dependent noisy labels","authors":"Arpit Garg ,&nbsp;Cuong Nguyen ,&nbsp;Rafael Felix ,&nbsp;Thanh-Toan Do ,&nbsp;Gustavo Carneiro","doi":"10.1016/j.imavis.2025.105877","DOIUrl":"10.1016/j.imavis.2025.105877","url":null,"abstract":"<div><div>Deep learning encounters significant challenges in the form of noisy-label samples, which can cause the overfitting of trained models. A primary challenge in learning with noisy-label (LNL) techniques is their ability to differentiate between hard samples (clean-label samples near the decision boundary) and instance-dependent noisy (IDN) label samples to allow these samples to be treated differently during training. Existing methodologies to identify IDN samples, including the small-loss hypothesis and feature-based selection, have demonstrated limited efficacy, thus impeding their effectiveness in dealing with real-world label noise. We present Peer-Agreement-based Sample Selection (PASS), a novel approach that utilises three classifiers, where a consensus-driven agreement between two models accurately differentiates between clean and noisy-label IDN samples to train the third model. In contrast to current techniques, PASS is specifically designed to address the complexities of IDN, where noise patterns are correlated with instance features. Our approach seamlessly integrates with existing LNL algorithms to enhance the accuracy of detecting both noisy and clean samples. Comprehensive experiments conducted on simulated benchmarks (CIFAR-100 and Red mini-ImageNet) and real-world datasets (Animal-10N, CIFAR-N, Clothing1M, and mini-WebVision) demonstrated that PASS substantially improved the performance of multiple state-of-the-art methods. This technique achieves superior classification accuracy, particularly in scenarios with high noise levels.<span><span><sup>1</sup></span></span></div></div>","PeriodicalId":50374,"journal":{"name":"Image and Vision Computing","volume":"166 ","pages":"Article 105877"},"PeriodicalIF":4.2,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145790213","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Image and Vision Computing
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1