首页 > 最新文献

IEEE transactions on pattern analysis and machine intelligence最新文献

英文 中文
Revealing the Dark Side of Non-Local Attention in Single Image Super-Resolution 揭示单像超分辨率中非局部注意力的阴暗面
Pub Date : 2024-09-10 DOI: 10.1109/TPAMI.2024.3457790
Jian-Nan Su;Guodong Fan;Min Gan;Guang-Yong Chen;Wenzhong Guo;C. L. Philip Chen
Single Image Super-Resolution (SISR) aims to reconstruct a high-resolution image from its corresponding low-resolution input. A common technique to enhance the reconstruction quality is Non-Local Attention (NLA), which leverages self-similar texture patterns in images. However, we have made a novel finding that challenges the prevailing wisdom. Our research reveals that NLA can be detrimental to SISR and even produce severely distorted textures. For example, when dealing with severely degrade textures, NLA may generate unrealistic results due to the inconsistency of non-local texture patterns. This problem is overlooked by existing works, which only measure the average reconstruction quality of the whole image, without considering the potential risks of using NLA. To address this issue, we propose a new perspective for evaluating the reconstruction quality of NLA, by focusing on the sub-pixel level that matches the pixel-wise fusion manner of NLA. From this perspective, we provide the approximate reconstruction performance upper bound of NLA, which guides us to design a concise yet effective Texture-Fidelity Strategy (TFS) to mitigate the degradation caused by NLA. Moreover, the proposed TFS can be conveniently integrated into existing NLA-based SISR models as a general building block. Based on the TFS, we develop a Deep Texture-Fidelity Network (DTFN), which achieves state-of-the-art performance for SISR. Our code and a pre-trained DTFN are available on GitHub for verification.
单图像超分辨率(SISR)旨在从相应的低分辨率输入图像重建高分辨率图像。提高重建质量的常用技术是非局部关注(NLA),它利用图像中的自相似纹理模式。然而,我们发现了一项新发现,对流行的观点提出了挑战。我们的研究发现,NLA 可能对 SISR 不利,甚至会产生严重失真的纹理。例如,在处理严重退化的纹理时,由于非局部纹理模式的不一致性,NLA 可能会产生不切实际的结果。现有的工作忽略了这一问题,它们只测量整个图像的平均重建质量,而没有考虑使用 NLA 的潜在风险。为了解决这个问题,我们提出了评估 NLA 重建质量的新视角,即关注与 NLA 的像素融合方式相匹配的子像素级别。从这个角度出发,我们提供了 NLA 的近似重建性能上限,从而指导我们设计出一种简洁而有效的纹理保真策略(TFS),以减轻 NLA 带来的性能下降。此外,所提出的 TFS 作为一个通用构件,可以方便地集成到现有的基于 NLA 的 SISR 模型中。在 TFS 的基础上,我们开发了深度纹理保真网络(DTFN),使 SISR 达到了最先进的性能。我们的代码和预训练的 DTFN 可在 GitHub† 上进行验证。
{"title":"Revealing the Dark Side of Non-Local Attention in Single Image Super-Resolution","authors":"Jian-Nan Su;Guodong Fan;Min Gan;Guang-Yong Chen;Wenzhong Guo;C. L. Philip Chen","doi":"10.1109/TPAMI.2024.3457790","DOIUrl":"10.1109/TPAMI.2024.3457790","url":null,"abstract":"Single Image Super-Resolution (SISR) aims to reconstruct a high-resolution image from its corresponding low-resolution input. A common technique to enhance the reconstruction quality is Non-Local Attention (NLA), which leverages self-similar texture patterns in images. However, we have made a novel finding that challenges the prevailing wisdom. Our research reveals that NLA can be detrimental to SISR and even produce severely distorted textures. For example, when dealing with severely degrade textures, NLA may generate unrealistic results due to the inconsistency of non-local texture patterns. This problem is overlooked by existing works, which only measure the average reconstruction quality of the whole image, without considering the potential risks of using NLA. To address this issue, we propose a new perspective for evaluating the reconstruction quality of NLA, by focusing on the sub-pixel level that matches the pixel-wise fusion manner of NLA. From this perspective, we provide the approximate reconstruction performance upper bound of NLA, which guides us to design a concise yet effective Texture-Fidelity Strategy (TFS) to mitigate the degradation caused by NLA. Moreover, the proposed TFS can be conveniently integrated into existing NLA-based SISR models as a general building block. Based on the TFS, we develop a Deep Texture-Fidelity Network (DTFN), which achieves state-of-the-art performance for SISR. Our code and a pre-trained DTFN are available on GitHub\u0000<sup>†</sup>\u0000 for verification.","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":"46 12","pages":"11476-11490"},"PeriodicalIF":0.0,"publicationDate":"2024-09-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142166412","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
HiSC4D: Human-Centered Interaction and 4D Scene Capture in Large-Scale Space Using Wearable IMUs and LiDAR HiSC4D:使用可穿戴式 IMU 和激光雷达在大规模空间进行以人为本的交互和 4D 场景捕捉
Pub Date : 2024-09-10 DOI: 10.1109/TPAMI.2024.3457229
Yudi Dai;Zhiyong Wang;Xiping Lin;Chenglu Wen;Lan Xu;Siqi Shen;Yuexin Ma;Cheng Wang
We introduce HiSC4D, a novel Human-centered interaction and 4D Scene Capture method, aimed at accurately and efficiently creating a dynamic digital world, containing large-scale indoor-outdoor scenes, diverse human motions, rich human-human interactions, and human-environment interactions. By utilizing body-mounted IMUs and a head-mounted LiDAR, HiSC4D can capture egocentric human motions in unconstrained space without the need for external devices and pre-built maps. This affords great flexibility and accessibility for human-centered interaction and 4D scene capturing in various environments. Taking into account that IMUs can capture human spatially unrestricted poses but are prone to drifting for long-period using, and while LiDAR is stable for global localization but rough for local positions and orientations, HiSC4D employs a joint optimization method, harmonizing all sensors and utilizing environment cues, yielding promising results for long-term capture in large scenes. To promote research of egocentric human interaction in large scenes and facilitate downstream tasks, we also present a dataset, containing 8 sequences in 4 large scenes (200 to 5,000 $text{m}^{2}$), providing 36 k frames of accurate 4D human motions with SMPL annotations and dynamic scenes, 31k frames of cropped human point clouds, and scene mesh of the environment. A variety of scenarios, such as the basketball gym and commercial street, alongside challenging human motions, such as daily greeting, one-on-one basketball playing, and tour guiding, demonstrate the effectiveness and the generalization ability of HiSC4D. The dataset and code will be publicly available for research purposes.
我们介绍的 HiSC4D 是一种新颖的以人为本的交互和 4D 场景捕捉方法,旨在准确高效地创建一个动态数字世界,其中包含大规模室内外场景、多样化的人体运动、丰富的人与人之间的交互以及人与环境之间的交互。通过利用安装在身体上的 IMU 和头戴式激光雷达,HiSC4D 可以捕捉无约束空间中以自我为中心的人体运动,而无需外部设备和预制地图。这为在各种环境中进行以人为中心的交互和 4D 场景捕捉提供了极大的灵活性和便利性。考虑到 IMUs 可以捕捉人类不受空间限制的姿势,但在长时间使用时容易发生漂移,而 LiDAR 对于全局定位是稳定的,但对于局部位置和方向却很粗糙,HiSC4D 采用了一种联合优化方法,协调所有传感器并利用环境线索,在大型场景的长期捕捉方面取得了可喜的成果。为了促进大场景中以自我为中心的人机交互研究并推动下游任务,我们还提出了一个数据集,其中包含4个大场景(200到5000美元text{m}^{2}$)中的8个序列,提供了36k帧带有SMPL注释和动态场景的精确4D人体运动、31k帧裁剪人体点云和环境场景网格。篮球馆和商业街等各种场景,以及日常问候、一对一篮球比赛和导游等具有挑战性的人体动作,都证明了 HiSC4D 的有效性和泛化能力。数据集和代码将公开用于研究目的。
{"title":"HiSC4D: Human-Centered Interaction and 4D Scene Capture in Large-Scale Space Using Wearable IMUs and LiDAR","authors":"Yudi Dai;Zhiyong Wang;Xiping Lin;Chenglu Wen;Lan Xu;Siqi Shen;Yuexin Ma;Cheng Wang","doi":"10.1109/TPAMI.2024.3457229","DOIUrl":"10.1109/TPAMI.2024.3457229","url":null,"abstract":"We introduce HiSC4D, a novel \u0000<b>H</b>\u0000uman-centered \u0000<b>i</b>\u0000nteraction and \u0000<b>4D</b>\u0000 \u0000<b>S</b>\u0000cene \u0000<b>C</b>\u0000apture method, aimed at accurately and efficiently creating a dynamic digital world, containing large-scale indoor-outdoor scenes, diverse human motions, rich human-human interactions, and human-environment interactions. By utilizing body-mounted IMUs and a head-mounted LiDAR, HiSC4D can capture egocentric human motions in unconstrained space without the need for external devices and pre-built maps. This affords great flexibility and accessibility for human-centered interaction and 4D scene capturing in various environments. Taking into account that IMUs can capture human spatially unrestricted poses but are prone to drifting for long-period using, and while LiDAR is stable for global localization but rough for local positions and orientations, HiSC4D employs a joint optimization method, harmonizing all sensors and utilizing environment cues, yielding promising results for long-term capture in large scenes. To promote research of egocentric human interaction in large scenes and facilitate downstream tasks, we also present a dataset, containing 8 sequences in 4 large scenes (200 to 5,000 \u0000<inline-formula><tex-math>$text{m}^{2}$</tex-math></inline-formula>\u0000), providing 36 k frames of accurate 4D human motions with SMPL annotations and dynamic scenes, 31k frames of cropped human point clouds, and scene mesh of the environment. A variety of scenarios, such as the basketball gym and commercial street, alongside challenging human motions, such as daily greeting, one-on-one basketball playing, and tour guiding, demonstrate the effectiveness and the generalization ability of HiSC4D. The dataset and code will be publicly available for research purposes.","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":"46 12","pages":"11236-11253"},"PeriodicalIF":0.0,"publicationDate":"2024-09-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142166411","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
One-Stage Anchor-Free Online Multiple Target Tracking With Deformable Local Attention and Task-Aware Prediction 利用可变形局部注意力和任务感知预测实现单级无锚在线多目标跟踪
Pub Date : 2024-09-10 DOI: 10.1109/TPAMI.2024.3457886
Weiming Hu;Shaoru Wang;Zongwei Zhou;Jin Gao;Yangxi Li;Stephen Maybank
The tracking-by-detection paradigm currently dominates multiple target tracking algorithms. It usually includes three tasks: target detection, appearance feature embedding, and data association. Carrying out these three tasks successively usually leads to lower tracking efficiency. In this paper, we propose a one-stage anchor-free multiple task learning framework which carries out target detection and appearance feature embedding in parallel to substantially increase the tracking speed. This framework simultaneously predicts a target detection and produces a feature embedding for each location, by sharing a pyramid of feature maps. We propose a deformable local attention module which utilizes the correlations between features at different locations within a target to obtain more discriminative features. We further propose a task-aware prediction module which utilizes deformable convolutions to select the most suitable locations for the different tasks. At the selected locations, classification of samples into foreground or background, appearance feature embedding, and target box regression are carried out. Two effective training strategies, regression range overlapping and sample reweighting, are proposed to reduce missed detections in dense scenes. Ambiguous samples whose identities are difficult to determine are effectively dealt with to obtain more accurate feature embedding of target appearance. An appearance-enhanced non-maximum suppression is proposed to reduce over-suppression of true targets in crowded scenes. Based on the one-stage anchor-free network with the deformable local attention module and the task-aware prediction module, we implement a new online multiple target tracker. Experimental results show that our tracker achieves a very fast speed while maintaining a high tracking accuracy.
通过检测进行跟踪的模式目前在多种目标跟踪算法中占主导地位。它通常包括三项任务:目标检测、外观特征嵌入和数据关联。连续执行这三个任务通常会降低跟踪效率。在本文中,我们提出了一种单阶段无锚多任务学习框架,它可以并行执行目标检测和外观特征嵌入,从而大幅提高跟踪速度。该框架通过共享一个金字塔形的特征图,同时预测目标检测并为每个位置生成特征嵌入。我们提出了一个可变形的局部关注模块,该模块利用目标内不同位置的特征之间的相关性来获取更具区分性的特征。我们还提出了任务感知预测模块,利用可变形卷积为不同任务选择最合适的位置。在选定的位置,将样本分类为前景或背景、外观特征嵌入和目标盒回归。我们提出了两种有效的训练策略,即回归范围重叠和样本重新加权,以减少密集场景中的漏检。对难以确定身份的模糊样本进行有效处理,以获得更准确的目标外观特征嵌入。提出了一种外观增强型非最大抑制方法,以减少拥挤场景中对真实目标的过度抑制。基于带有可变形局部注意力模块和任务感知预测模块的单级无锚网络,我们实现了一种新的在线多目标跟踪器。实验结果表明,我们的跟踪器在保持较高跟踪精度的同时,实现了极快的跟踪速度。
{"title":"One-Stage Anchor-Free Online Multiple Target Tracking With Deformable Local Attention and Task-Aware Prediction","authors":"Weiming Hu;Shaoru Wang;Zongwei Zhou;Jin Gao;Yangxi Li;Stephen Maybank","doi":"10.1109/TPAMI.2024.3457886","DOIUrl":"10.1109/TPAMI.2024.3457886","url":null,"abstract":"The tracking-by-detection paradigm currently dominates multiple target tracking algorithms. It usually includes three tasks: target detection, appearance feature embedding, and data association. Carrying out these three tasks successively usually leads to lower tracking efficiency. In this paper, we propose a one-stage anchor-free multiple task learning framework which carries out target detection and appearance feature embedding in parallel to substantially increase the tracking speed. This framework simultaneously predicts a target detection and produces a feature embedding for each location, by sharing a pyramid of feature maps. We propose a deformable local attention module which utilizes the correlations between features at different locations within a target to obtain more discriminative features. We further propose a task-aware prediction module which utilizes deformable convolutions to select the most suitable locations for the different tasks. At the selected locations, classification of samples into foreground or background, appearance feature embedding, and target box regression are carried out. Two effective training strategies, regression range overlapping and sample reweighting, are proposed to reduce missed detections in dense scenes. Ambiguous samples whose identities are difficult to determine are effectively dealt with to obtain more accurate feature embedding of target appearance. An appearance-enhanced non-maximum suppression is proposed to reduce over-suppression of true targets in crowded scenes. Based on the one-stage anchor-free network with the deformable local attention module and the task-aware prediction module, we implement a new online multiple target tracker. Experimental results show that our tracker achieves a very fast speed while maintaining a high tracking accuracy.","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":"46 12","pages":"11446-11463"},"PeriodicalIF":0.0,"publicationDate":"2024-09-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142166410","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
R$^{3}$3LIVE++: A Robust, Real-Time, Radiance Reconstruction Package With a Tightly-Coupled LiDAR-Inertial-Visual State Estimator r $^{3}$ live++:带有紧密耦合的激光雷达-惯性-视觉状态估计器的鲁棒、实时、辐射重构软件包
Pub Date : 2024-09-09 DOI: 10.1109/TPAMI.2024.3456473
Jiarong Lin;Fu Zhang
This work proposed a LiDAR-inertial-visual fusion framework termed R$^{3}$LIVE++ to achieve robust and accurate state estimation while simultaneously reconstructing the radiance map on the fly. R$^{3}$LIVE++ consists of a LiDAR-inertial odometry (LIO) and a visual-inertial odometry (VIO), both running in real-time. The LIO subsystem utilizes the measurements from a LiDAR for reconstructing the geometric structure, while the VIO subsystem simultaneously recovers the radiance information of the geometric structure from the input images. R$^{3}$LIVE++ is developed based on R$^{3}$LIVE and further improves the accuracy in localization and mapping by accounting for the camera photometric calibration and the online estimation of camera exposure time. We conduct more extensive experiments on public and self-collected datasets to compare our proposed system against other state-of-the-art SLAM systems. Quantitative and qualitative results show that R$^{3}$LIVE++ has significant improvements over others in both accuracy and robustness. Moreover, to demonstrate the extendability of R$^{3}$LIVE++, we developed several applications based on our reconstructed maps, such as high dynamic range (HDR) imaging, virtual environment exploration, and 3D video gaming.
这项研究提出了一个称为R$^{3}$LIVE++的激光雷达-惯性-视觉融合框架,以实现稳健而准确的状态估计,同时在飞行中重建辐射图。R$^{3}$LIVE++ 由实时运行的激光雷达-惯性里程计(LIO)和视觉-惯性里程计(VIO)组成。LIO 子系统利用激光雷达的测量数据重建几何结构,而 VIO 子系统则同时从输入图像中恢复几何结构的辐射信息。R$^{3}$LIVE++ 是在 R$^{3}$LIVE 的基础上开发的,通过考虑相机光度校准和相机曝光时间的在线估计,进一步提高了定位和绘图的精度。我们在公共数据集和自收集数据集上进行了更广泛的实验,将我们提出的系统与其他最先进的 SLAM 系统进行比较。定量和定性结果表明,R$^{3}$LIVE++ 在准确性和鲁棒性方面都比其他系统有显著提高。此外,为了证明R$^{3}$LIVE++的可扩展性,我们基于重建的地图开发了多个应用,如高动态范围(HDR)成像、虚拟环境探索和三维视频游戏。
{"title":"R$^{3}$3LIVE++: A Robust, Real-Time, Radiance Reconstruction Package With a Tightly-Coupled LiDAR-Inertial-Visual State Estimator","authors":"Jiarong Lin;Fu Zhang","doi":"10.1109/TPAMI.2024.3456473","DOIUrl":"10.1109/TPAMI.2024.3456473","url":null,"abstract":"This work proposed a LiDAR-inertial-visual fusion framework termed R\u0000<inline-formula><tex-math>$^{3}$</tex-math></inline-formula>\u0000LIVE++ to achieve robust and accurate state estimation while simultaneously reconstructing the radiance map on the fly. R\u0000<inline-formula><tex-math>$^{3}$</tex-math></inline-formula>\u0000LIVE++ consists of a LiDAR-inertial odometry (LIO) and a visual-inertial odometry (VIO), both running in real-time. The LIO subsystem utilizes the measurements from a LiDAR for reconstructing the geometric structure, while the VIO subsystem simultaneously recovers the radiance information of the geometric structure from the input images. R\u0000<inline-formula><tex-math>$^{3}$</tex-math></inline-formula>\u0000LIVE++ is developed based on R\u0000<inline-formula><tex-math>$^{3}$</tex-math></inline-formula>\u0000LIVE and further improves the accuracy in localization and mapping by accounting for the camera photometric calibration and the online estimation of camera exposure time. We conduct more extensive experiments on public and self-collected datasets to compare our proposed system against other state-of-the-art SLAM systems. Quantitative and qualitative results show that R\u0000<inline-formula><tex-math>$^{3}$</tex-math></inline-formula>\u0000LIVE++ has significant improvements over others in both accuracy and robustness. Moreover, to demonstrate the extendability of R\u0000<inline-formula><tex-math>$^{3}$</tex-math></inline-formula>\u0000LIVE++, we developed several applications based on our reconstructed maps, such as high dynamic range (HDR) imaging, virtual environment exploration, and 3D video gaming.","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":"46 12","pages":"11168-11185"},"PeriodicalIF":0.0,"publicationDate":"2024-09-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142160560","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Vision-Centric BEV Perception: A Survey 以视觉为中心的电动汽车感知:一项调查
Pub Date : 2024-09-09 DOI: 10.1109/TPAMI.2024.3449912
Yuexin Ma;Tai Wang;Xuyang Bai;Huitong Yang;Yuenan Hou;Yaming Wang;Yu Qiao;Ruigang Yang;Xinge Zhu
In recent years, vision-centric Bird’s Eye View (BEV) perception has garnered significant interest from both industry and academia due to its inherent advantages, such as providing an intuitive representation of the world and being conducive to data fusion. The rapid advancements in deep learning have led to the proposal of numerous methods for addressing vision-centric BEV perception challenges. However, there has been no recent survey encompassing this novel and burgeoning research field. To catalyze future research, this paper presents a comprehensive survey of the latest developments in vision-centric BEV perception and its extensions. It compiles and organizes up-to-date knowledge, offering a systematic review and summary of prevalent algorithms. Additionally, the paper provides in-depth analyses and comparative results on various BEV perception tasks, facilitating the evaluation of future works and sparking new research directions. Furthermore, the paper discusses and shares valuable empirical implementation details to aid in the advancement of related algorithms.
近年来,以视觉为中心的鸟瞰(BEV)感知因其固有的优势,如对世界的直观呈现和有利于数据融合等,引起了业界和学术界的极大兴趣。随着深度学习的快速发展,人们提出了许多方法来应对以视觉为中心的 BEV 感知挑战。然而,最近还没有一项调查涵盖了这一新兴的研究领域。为了促进未来的研究,本文全面介绍了以视觉为中心的 BEV 感知及其扩展的最新进展。本文汇编并整理了最新知识,对流行算法进行了系统回顾和总结。此外,本文还对各种 BEV 感知任务进行了深入分析并提供了比较结果,有助于对未来工作进行评估,并激发新的研究方向。此外,论文还讨论并分享了有价值的经验实施细节,以帮助相关算法的发展。
{"title":"Vision-Centric BEV Perception: A Survey","authors":"Yuexin Ma;Tai Wang;Xuyang Bai;Huitong Yang;Yuenan Hou;Yaming Wang;Yu Qiao;Ruigang Yang;Xinge Zhu","doi":"10.1109/TPAMI.2024.3449912","DOIUrl":"10.1109/TPAMI.2024.3449912","url":null,"abstract":"In recent years, vision-centric Bird’s Eye View (BEV) perception has garnered significant interest from both industry and academia due to its inherent advantages, such as providing an intuitive representation of the world and being conducive to data fusion. The rapid advancements in deep learning have led to the proposal of numerous methods for addressing vision-centric BEV perception challenges. However, there has been no recent survey encompassing this novel and burgeoning research field. To catalyze future research, this paper presents a comprehensive survey of the latest developments in vision-centric BEV perception and its extensions. It compiles and organizes up-to-date knowledge, offering a systematic review and summary of prevalent algorithms. Additionally, the paper provides in-depth analyses and comparative results on various BEV perception tasks, facilitating the evaluation of future works and sparking new research directions. Furthermore, the paper discusses and shares valuable empirical implementation details to aid in the advancement of related algorithms.","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":"46 12","pages":"10978-10997"},"PeriodicalIF":0.0,"publicationDate":"2024-09-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142160561","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MINN: Learning the Dynamics of Differential-Algebraic Equations and Application to Battery Modeling MINN:学习微分代数方程的动力学并将其应用于电池建模
Pub Date : 2024-09-09 DOI: 10.1109/TPAMI.2024.3456475
Yicun Huang;Changfu Zou;Yang Li;Torsten Wik
The concept of integrating physics-based and data-driven approaches has become popular for modeling sustainable energy systems. However, the existing literature mainly focuses on the data-driven surrogates generated to replace physics-based models. These models often trade accuracy for speed but lack the generalizability, adaptability, and interpretability inherent in physics-based models, which are often indispensable in modeling real-world dynamic systems for optimization and control purposes. We propose a novel machine learning architecture, termed model-integrated neural networks (MINN), that can learn the physics-based dynamics of general autonomous or non-autonomous systems consisting of partial differential-algebraic equations (PDAEs). The obtained architecture systematically solves an unsettled research problem in control-oriented modeling, i.e., how to obtain optimally simplified models that are physically insightful, numerically accurate, and computationally tractable simultaneously. We apply the proposed neural network architecture to model the electrochemical dynamics of lithium-ion batteries and show that MINN is extremely data-efficient to train while being sufficiently generalizable to previously unseen input data, owing to its underlying physical invariants. The MINN battery model has an accuracy comparable to the first principle-based model in predicting both the system outputs and any locally distributed electrochemical behaviors but achieves two orders of magnitude reduction in the solution time.
在可持续能源系统建模中,将基于物理的方法与数据驱动的方法相结合的概念已变得十分流行。然而,现有文献主要关注为取代基于物理的模型而生成的数据驱动代用模型。这些模型通常以准确性换取速度,但缺乏基于物理的模型所固有的通用性、适应性和可解释性,而这些往往是为优化和控制目的模拟真实世界动态系统所不可或缺的。我们提出了一种新颖的机器学习架构,即模型集成神经网络(MINN),它可以学习由偏微分代数方程(PDAE)组成的一般自主或非自主系统的物理动态。所获得的架构系统地解决了面向控制的建模中一个悬而未决的研究问题,即如何同时获得具有物理洞察力、数值精确性和计算可操作性的最佳简化模型。我们将提出的神经网络架构应用于锂离子电池的电化学动力学建模,结果表明 MINN 具有极高的数据训练效率,同时由于其潜在的物理不变性,对以前未见过的输入数据具有足够的通用性。MINN 电池模型在预测系统输出和任何局部分布式电化学行为方面的准确性与基于第一原理的模型相当,但求解时间却缩短了两个数量级。
{"title":"MINN: Learning the Dynamics of Differential-Algebraic Equations and Application to Battery Modeling","authors":"Yicun Huang;Changfu Zou;Yang Li;Torsten Wik","doi":"10.1109/TPAMI.2024.3456475","DOIUrl":"10.1109/TPAMI.2024.3456475","url":null,"abstract":"The concept of integrating physics-based and data-driven approaches has become popular for modeling sustainable energy systems. However, the existing literature mainly focuses on the data-driven surrogates generated to replace physics-based models. These models often trade accuracy for speed but lack the generalizability, adaptability, and interpretability inherent in physics-based models, which are often indispensable in modeling real-world dynamic systems for optimization and control purposes. We propose a novel machine learning architecture, termed model-integrated neural networks (MINN), that can learn the physics-based dynamics of general autonomous or non-autonomous systems consisting of partial differential-algebraic equations (PDAEs). The obtained architecture systematically solves an unsettled research problem in control-oriented modeling, i.e., how to obtain optimally simplified models that are physically insightful, numerically accurate, and computationally tractable simultaneously. We apply the proposed neural network architecture to model the electrochemical dynamics of lithium-ion batteries and show that MINN is extremely data-efficient to train while being sufficiently generalizable to previously unseen input data, owing to its underlying physical invariants. The MINN battery model has an accuracy comparable to the first principle-based model in predicting both the system outputs and any locally distributed electrochemical behaviors but achieves two orders of magnitude reduction in the solution time.","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":"46 12","pages":"11331-11344"},"PeriodicalIF":0.0,"publicationDate":"2024-09-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142160562","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Diversifying Policies With Non-Markov Dispersion to Expand the Solution Space 利用非马尔可夫离散性的多样化政策来扩展求解空间
Pub Date : 2024-09-06 DOI: 10.1109/TPAMI.2024.3455257
Bohao Qu;Xiaofeng Cao;Yi Chang;Ivor W. Tsang;Yew-Soon Ong
Policy diversity, encompassing the variety of policies an agent can adopt, enhances reinforcement learning (RL) success by fostering more robust, adaptable, and innovative problem-solving in the environment. The environment in which standard RL operates is usually modeled with a Markov Decision Process (MDP) as the theoretical foundation. However, in many real-world scenarios, the rewards depend on an agent's history of states and actions leading to a non-MDP. Under the premise of policy diffusion initialization, non-MDPs may have unstructured expanding solution space due to varying historical information and temporal dependencies. This results in solutions having non-equivalent closed forms in non-MDPs. In this paper, deriving diverse solutions for non-MDPs requires policies to break through the boundaries of the current solution space through gradual dispersion. The goal is to expand the solution space, thereby obtaining more diverse policies. Specifically, we first model the sequences of states and actions by a transformer-based method to learn policy embeddings for dispersion in the solution space, since the transformer has advantages in handling sequential data and capturing long-range dependencies for non-MDP. Then, we stack the policy embeddings to construct a dispersion matrix as the policy diversity measure to induce the policy dispersion in the solution space and obtain a set of diverse policies. Finally, we prove that if the dispersion matrix is positive definite, the dispersed embeddings can effectively enlarge the disagreements across policies, yielding a diverse expression for the original policy embedding distribution. Experimental results of both non-MDP and MDP environments show that this dispersion scheme can obtain more expressive diverse policies via expanding the solution space, showing more robust performance than the recent learning baselines.
策略多样性包括一个代理可以采用的各种策略,它通过在环境中促进更稳健、适应性更强和更具创新性的问题解决,来提高强化学习(RL)的成功率。标准强化学习的运行环境通常以马尔可夫决策过程(MDP)作为理论基础。然而,在现实世界的许多场景中,奖励取决于代理的历史状态和行动,从而导致非马尔可夫决策过程。在策略扩散初始化的前提下,由于历史信息和时间依赖性的不同,非 MDP 可能会有非结构化的扩展解空间。这就导致非 MDP 中的解具有非等价的封闭形式。在本文中,要推导出非 MDPs 的多样化解,需要政策通过逐步分散来突破当前解空间的边界。我们的目标是扩大解空间,从而获得更多样化的策略。具体来说,由于变换器在处理顺序数据和捕捉非 MDP 的长程依赖性方面具有优势,因此我们首先通过基于变换器的方法对状态和行动序列进行建模,以学习解空间中分散的策略嵌入。然后,我们通过堆叠策略嵌入来构建分散矩阵作为策略多样性度量,从而诱导解空间中的策略分散,并得到一组多样性策略。最后,我们证明了如果分散矩阵是正定的,分散的嵌入可以有效地扩大政策间的分歧,从而得到原始政策嵌入分布的多样性表达式。非 MDP 和 MDP 环境的实验结果表明,这种分散方案可以通过扩大解空间获得更具表现力的多样化策略,与最近的学习基线相比,表现出更强的鲁棒性。
{"title":"Diversifying Policies With Non-Markov Dispersion to Expand the Solution Space","authors":"Bohao Qu;Xiaofeng Cao;Yi Chang;Ivor W. Tsang;Yew-Soon Ong","doi":"10.1109/TPAMI.2024.3455257","DOIUrl":"10.1109/TPAMI.2024.3455257","url":null,"abstract":"Policy diversity, encompassing the variety of policies an agent can adopt, enhances reinforcement learning (RL) success by fostering more robust, adaptable, and innovative problem-solving in the environment. The environment in which standard RL operates is usually modeled with a Markov Decision Process (MDP) as the theoretical foundation. However, in many real-world scenarios, the rewards depend on an agent's history of states and actions leading to a non-MDP. Under the premise of policy diffusion initialization, non-MDPs may have unstructured expanding solution space due to varying historical information and temporal dependencies. This results in solutions having non-equivalent closed forms in non-MDPs. In this paper, deriving diverse solutions for non-MDPs requires policies to break through the boundaries of the current solution space through gradual dispersion. The goal is to expand the solution space, thereby obtaining more diverse policies. Specifically, we first model the sequences of states and actions by a transformer-based method to learn policy embeddings for dispersion in the solution space, since the transformer has advantages in handling sequential data and capturing long-range dependencies for non-MDP. Then, we stack the policy embeddings to construct a dispersion matrix as the policy diversity measure to induce the policy dispersion in the solution space and obtain a set of diverse policies. Finally, we prove that if the dispersion matrix is positive definite, the dispersed embeddings can effectively enlarge the disagreements across policies, yielding a diverse expression for the original policy embedding distribution. Experimental results of both non-MDP and MDP environments show that this dispersion scheme can obtain more expressive diverse policies via expanding the solution space, showing more robust performance than the recent learning baselines.","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":"46 12","pages":"11392-11408"},"PeriodicalIF":0.0,"publicationDate":"2024-09-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142143505","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Integrating Neural Radiance Fields End-to-End for Cognitive Visuomotor Navigation 从头到尾整合神经辐射场,实现认知视觉运动导航。
Pub Date : 2024-09-06 DOI: 10.1109/TPAMI.2024.3455252
Qiming Liu;Haoran Xin;Zhe Liu;Hesheng Wang
We propose an end-to-end visuomotor navigation framework that leverages Neural Radiance Fields (NeRF) for spatial cognition. To the best of our knowledge, this is the first effort to integrate such implicit spatial representation with embodied policy end-to-end for cognitive decision-making. Consequently, our system does not necessitate modularized designs nor transformations into explicit scene representations for downstream control. The NeRF-based memory is constructed online during navigation, without relying on any environmental priors. To enhance the extraction of decision-critical historical insights from the rigid and implicit structure of NeRF, we introduce a spatial information extraction mechanism named Structural Radiance Attention (SRA). SRA empowers the agent to grasp complex scene structures and task objectives, thus paving the way for the development of intelligent behavioral patterns. Our comprehensive testing in image-goal navigation tasks demonstrates that our approach significantly outperforms existing navigation models. We demonstrate that SRA markedly improves the agent's understanding of both the scene and the task by retrieving historical information stored in NeRF memory. The agent also learns exploratory awareness from our pipeline to better adapt to low signal-to-noise memory signals in unknown scenes. We deploy our navigation system on a mobile robot in real-world scenarios, where it exhibits evident cognitive capabilities while ensuring real-time performance.
我们提出了一个端到端的视觉运动导航框架,利用神经辐射场(NeRF)进行空间认知。据我们所知,这是首次将这种隐式空间表征与用于认知决策的端到端嵌入式策略整合在一起。因此,我们的系统既不需要模块化设计,也不需要转换成显式场景表示来进行下游控制。基于 NeRF 的记忆是在导航过程中在线构建的,无需依赖任何环境先验。为了加强从 NeRF 的刚性和隐式结构中提取对决策至关重要的历史洞察力,我们引入了一种名为 "结构辐射注意"(SRA)的空间信息提取机制。SRA 可使代理掌握复杂的场景结构和任务目标,从而为开发智能行为模式铺平道路。我们在图像目标导航任务中进行的全面测试表明,我们的方法明显优于现有的导航模型。我们证明,通过检索存储在 NeRF 内存中的历史信息,SRA 显著提高了代理对场景和任务的理解。代理还能从我们的管道中学习探索意识,从而更好地适应未知场景中的低信噪比记忆信号。我们在实际场景中的移动机器人上部署了我们的导航系统,该系统在确保实时性能的同时,还展现了明显的认知能力。
{"title":"Integrating Neural Radiance Fields End-to-End for Cognitive Visuomotor Navigation","authors":"Qiming Liu;Haoran Xin;Zhe Liu;Hesheng Wang","doi":"10.1109/TPAMI.2024.3455252","DOIUrl":"10.1109/TPAMI.2024.3455252","url":null,"abstract":"We propose an end-to-end visuomotor navigation framework that leverages Neural Radiance Fields (NeRF) for spatial cognition. To the best of our knowledge, this is the first effort to integrate such implicit spatial representation with embodied policy end-to-end for cognitive decision-making. Consequently, our system does not necessitate modularized designs nor transformations into explicit scene representations for downstream control. The NeRF-based memory is constructed online during navigation, without relying on any environmental priors. To enhance the extraction of decision-critical historical insights from the rigid and implicit structure of NeRF, we introduce a spatial information extraction mechanism named Structural Radiance Attention (SRA). SRA empowers the agent to grasp complex scene structures and task objectives, thus paving the way for the development of intelligent behavioral patterns. Our comprehensive testing in image-goal navigation tasks demonstrates that our approach significantly outperforms existing navigation models. We demonstrate that SRA markedly improves the agent's understanding of both the scene and the task by retrieving historical information stored in NeRF memory. The agent also learns exploratory awareness from our pipeline to better adapt to low signal-to-noise memory signals in unknown scenes. We deploy our navigation system on a mobile robot in real-world scenarios, where it exhibits evident cognitive capabilities while ensuring real-time performance.","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":"46 12","pages":"11200-11215"},"PeriodicalIF":0.0,"publicationDate":"2024-09-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142143506","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Variational Label Enhancement for Instance-Dependent Partial Label Learning 针对依赖于实例的部分标签学习的变量标签增强。
Pub Date : 2024-09-06 DOI: 10.1109/TPAMI.2024.3455260
Ning Xu;Congyu Qiao;Yuchen Zhao;Xin Geng;Min-Ling Zhang
Partial label learning (PLL) is a form of weakly supervised learning, where each training example is linked to a set of candidate labels, among which only one label is correct. Most existing PLL approaches assume that the incorrect labels in each training example are randomly picked as the candidate labels. However, in practice, this assumption may not hold true, as the candidate labels are often instance-dependent. In this paper, we address the instance-dependent PLL problem and assume that each example is associated with a latent label distribution where the incorrect label with a high degree is more likely to be annotated as a candidate label. Motivated by this consideration, we propose two methods Valen and Milen, which train the predictive model via utilizing the latent label distributions recovered by the label enhancement process. Specifically, Valen recovers the latent label distributions via inferring the variational posterior density parameterized by an inference model with the deduced evidence lower bound. Milen recovers the latent label distribution by adopting the variational approximation to bound the mutual information among the latent label distribution, observed labels and augmented instances. Experiments on benchmark and real-world datasets validate the effectiveness of the proposed methods.
部分标签学习(PLL)是弱监督学习的一种形式,每个训练实例都与一组候选标签相关联,其中只有一个标签是正确的。大多数现有的 PLL 方法都假定,每个训练示例中的错误标签都是随机选取的候选标签。然而,在实践中,这一假设可能并不成立,因为候选标签往往取决于实例。在本文中,我们将解决与实例相关的 PLL 问题,并假设每个实例都与潜在标签分布相关联,其中错误标签的度数高的标签更有可能被注释为候选标签。基于这一考虑,我们提出了 VALEN 和 MILEN 两种方法,它们通过利用标签增强过程恢复的潜在标签分布来训练预测模型。具体来说,VALEN 通过推断由推理模型参数化的变分后验密度和推导出的证据下限来恢复潜在标签分布。MILEN 通过采用变分近似来约束潜在标签分布、观察标签和增强实例之间的互信息,从而恢复潜在标签分布。在基准数据集和实际数据集上进行的实验验证了所提方法的有效性。
{"title":"Variational Label Enhancement for Instance-Dependent Partial Label Learning","authors":"Ning Xu;Congyu Qiao;Yuchen Zhao;Xin Geng;Min-Ling Zhang","doi":"10.1109/TPAMI.2024.3455260","DOIUrl":"10.1109/TPAMI.2024.3455260","url":null,"abstract":"Partial label learning (PLL) is a form of weakly supervised learning, where each training example is linked to a set of candidate labels, among which only one label is correct. Most existing PLL approaches assume that the incorrect labels in each training example are randomly picked as the candidate labels. However, in practice, this assumption may not hold true, as the candidate labels are often instance-dependent. In this paper, we address the instance-dependent PLL problem and assume that each example is associated with a latent \u0000<italic>label distribution</i>\u0000 where the incorrect label with a high degree is more likely to be annotated as a candidate label. Motivated by this consideration, we propose two methods \u0000<sc>Valen</small>\u0000 and \u0000<sc>Milen</small>\u0000, which train the predictive model via utilizing the latent label distributions recovered by the label enhancement process. Specifically, \u0000<sc>Valen</small>\u0000 recovers the latent label distributions via inferring the variational posterior density parameterized by an inference model with the deduced evidence lower bound. \u0000<sc>Milen</small>\u0000 recovers the latent label distribution by adopting the variational approximation to bound the mutual information among the latent label distribution, observed labels and augmented instances. Experiments on benchmark and real-world datasets validate the effectiveness of the proposed methods.","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":"46 12","pages":"11298-11313"},"PeriodicalIF":0.0,"publicationDate":"2024-09-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142143507","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
TagCLIP: Improving Discrimination Ability of Zero-Shot Semantic Segmentation TagCLIP:提高零镜头语义分割的识别能力。
Pub Date : 2024-09-04 DOI: 10.1109/TPAMI.2024.3454647
Jingyao Li;Pengguang Chen;Shengju Qian;Shu Liu;Jiaya Jia
Contrastive Language-Image Pre-training (CLIP) has recently shown great promise in pixel-level zero-shot learning tasks. However, existing approaches utilizing CLIP's text and patch embeddings to generate semantic masks often misidentify input pixels from unseen classes, leading to confusion between novel classes and semantically similar ones. In this work, we propose a novel approach, TagCLIP (Trusty-aware guided CLIP), to address this issue. We disentangle the ill-posed optimization problem into two parallel processes: semantic matching performed individually and reliability judgment for improving discrimination ability. Building on the idea of special tokens in language modeling representing sentence-level embeddings, we introduce a trusty token that enables distinguishing novel classes from known ones in prediction. To evaluate our approach, we conduct experiments on two benchmark datasets, PASCAL VOC 2012 and COCO-Stuff 164 K. Our results show that TagCLIP improves the Intersection over Union (IoU) of unseen classes by 7.4% and 1.7%, respectively, with negligible overheads. The code is available at here.
对比语言图像预训练(CLIP)最近在像素级零点学习任务中显示出了巨大的潜力。然而,利用 CLIP 的文本和补丁嵌入生成语义掩码的现有方法经常会错误识别来自未见类别的输入像素,从而导致新类别和语义相似类别之间的混淆。在这项工作中,我们提出了一种新方法 TagCLIP(信任感知引导式 CLIP)来解决这个问题。我们将难以解决的优化问题分解为两个并行过程:单独进行的语义匹配和提高辨别能力的可靠性判断。基于语言建模中代表句子级嵌入的特殊标记的想法,我们引入了一种可信标记,它能在预测中将新类别与已知类别区分开来。为了评估我们的方法,我们在 PASCAL VOC 2012 和 COCO-Stuff 164K 这两个基准数据集上进行了实验。结果表明,TagCLIP 将未见类别的 "交集大于联合"(Intersection over Union,IoU)分别提高了 7.4% 和 1.7%,而开销几乎可以忽略不计。
{"title":"TagCLIP: Improving Discrimination Ability of Zero-Shot Semantic Segmentation","authors":"Jingyao Li;Pengguang Chen;Shengju Qian;Shu Liu;Jiaya Jia","doi":"10.1109/TPAMI.2024.3454647","DOIUrl":"10.1109/TPAMI.2024.3454647","url":null,"abstract":"Contrastive Language-Image Pre-training (CLIP) has recently shown great promise in pixel-level zero-shot learning tasks. However, existing approaches utilizing CLIP's text and patch embeddings to generate semantic masks often misidentify input pixels from unseen classes, leading to confusion between novel classes and semantically similar ones. In this work, we propose a novel approach, \u0000<bold>TagCLIP</b>\u0000 (\u0000<bold>T</b>\u0000rusty-\u0000<bold>a</b>\u0000ware \u0000<bold>g</b>\u0000uided CLIP), to address this issue. We disentangle the ill-posed optimization problem into two parallel processes: semantic matching performed individually and reliability judgment for improving discrimination ability. Building on the idea of special tokens in language modeling representing sentence-level embeddings, we introduce a trusty token that enables distinguishing novel classes from known ones in prediction. To evaluate our approach, we conduct experiments on two benchmark datasets, PASCAL VOC 2012 and COCO-Stuff 164 K. Our results show that TagCLIP improves the Intersection over Union (IoU) of unseen classes by 7.4% and 1.7%, respectively, with negligible overheads. The code is available at here.","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":"46 12","pages":"11287-11297"},"PeriodicalIF":0.0,"publicationDate":"2024-09-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142134894","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE transactions on pattern analysis and machine intelligence
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1