首页 > 最新文献

IEEE Robotics and Automation Letters最新文献

英文 中文
Human-Inspired Adaptive Gait Learning for Humanoids Locomotion 仿人自适应步态学习
IF 5.3 2区 计算机科学 Q2 ROBOTICS Pub Date : 2026-01-12 DOI: 10.1109/LRA.2026.3653300
Lequn Fu;Xiao Li;Yibin Liu;Xiangan Zeng;Yibo Peng;Youjun Xiong;Shiqi Li
Achieving natural, robust, and energy-efficient locomotion remains a central challenge for humanoid control. While imitation learning enables robots to reproduce human-like behaviors, differences in morphology, actuation, and partial observability often limit direct motion replication. This work proposes a human-inspired reinforcement learning framework that integrates both implicit and explicit guidance. Implicit human motion priors, obtained through adversarial learning, provide style alignment with human data, while explicit biomechanical rewards encode characteristic gait principles to promote symmetry, stability, and adaptability. In addition, a history-based state estimator explicitly reconstructs base velocities from partial observations, mitigating observability gaps and enhancing robustness in real-world settings. To assess human-likeness, we introduce a tri-metric evaluation protocol covering gait symmetry, human–robot similarity, and energy efficiency. Extensive experiments demonstrate that the proposed approach produces locomotion that is not only robust and transferable across diverse terrains but also energy-efficient and recognizably human-like.
实现自然、稳健和节能的运动仍然是人形控制的核心挑战。虽然模仿学习使机器人能够复制类似人类的行为,但形态、驱动和部分可观察性的差异往往限制了直接运动复制。这项工作提出了一个由人类启发的强化学习框架,该框架集成了隐式和显式指导。通过对抗学习获得的内隐人类运动先验提供了与人类数据的风格对齐,而显式生物力学奖励编码了特征步态原则,以促进对称性、稳定性和适应性。此外,基于历史的状态估计器明确地从部分观测中重建基本速度,减轻了可观测性差距,增强了现实环境中的鲁棒性。为了评估人类的相似性,我们引入了一种三度量评估方案,包括步态对称性、人-机器人相似性和能量效率。大量的实验表明,所提出的方法产生的运动不仅是鲁棒性和可转移在不同的地形,而且节能和可识别的人类。
{"title":"Human-Inspired Adaptive Gait Learning for Humanoids Locomotion","authors":"Lequn Fu;Xiao Li;Yibin Liu;Xiangan Zeng;Yibo Peng;Youjun Xiong;Shiqi Li","doi":"10.1109/LRA.2026.3653300","DOIUrl":"https://doi.org/10.1109/LRA.2026.3653300","url":null,"abstract":"Achieving natural, robust, and energy-efficient locomotion remains a central challenge for humanoid control. While imitation learning enables robots to reproduce human-like behaviors, differences in morphology, actuation, and partial observability often limit direct motion replication. This work proposes a human-inspired reinforcement learning framework that integrates both implicit and explicit guidance. Implicit human motion priors, obtained through adversarial learning, provide style alignment with human data, while explicit biomechanical rewards encode characteristic gait principles to promote symmetry, stability, and adaptability. In addition, a history-based state estimator explicitly reconstructs base velocities from partial observations, mitigating observability gaps and enhancing robustness in real-world settings. To assess human-likeness, we introduce a tri-metric evaluation protocol covering gait symmetry, human–robot similarity, and energy efficiency. Extensive experiments demonstrate that the proposed approach produces locomotion that is not only robust and transferable across diverse terrains but also energy-efficient and recognizably human-like.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"11 3","pages":"2458-2465"},"PeriodicalIF":5.3,"publicationDate":"2026-01-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146001865","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
PointVLA: Injecting the 3D World Into Vision-Language-Action Models pointla:将3D世界注入视觉语言动作模型
IF 5.3 2区 计算机科学 Q2 ROBOTICS Pub Date : 2026-01-12 DOI: 10.1109/LRA.2026.3653303
Chengmeng Li;Junjie Wen;Yaxin Peng;Yan Peng;Yichen Zhu
Vision-Language-Action (VLA) models excel at robotic tasks by leveraging large-scale 2D vision-language pretraining, but their reliance on RGB images limits spatial reasoning critical for real-world interaction. Retraining these models with 3D data is computationally prohibitive, while discarding existing 2D datasets wastes valuable resources. To bridge this gap, we propose PointVLA, a framework that enhances pre-trained VLAs with point cloud inputs without requiring retraining. Our method freezes the vanilla action expert and injects 3D features via a lightweight modular block. To identify the most effective way of integrating point cloud representations, we conduct a skip-block analysis to pinpoint less useful blocks in the vanilla action expert, ensuring that 3D features are injected only into these blocks—minimizing disruption to pre-trained representations. Extensive experiments demonstrate that PointVLA outperforms state-of-the-art 2D imitation learning methods, such as OpenVLA, Diffusion Policy and DexVLA, across both simulated and real-world robotic tasks. Specifically, we highlight several key advantages of PointVLA enabled by point cloud integration: (1) Few-shot multi-tasking, where PointVLA successfully performs four different tasks using only 20 demonstrations each; (2) Real-vs-photo discrimination, where PointVLA distinguishes real objects from their images, leveraging 3D world knowledge to improve safety and reliability; (3) Height adaptability, where unlike conventional 2D imitation learning methods, PointVLA enables robots to adapt to objects at varying table heights that were unseen in training data. Furthermore, PointVLA achieves strong performance in long-horizon tasks, such as picking and packing objects from a moving conveyor belt, showcasing its ability to generalize across complex, dynamic environments.
视觉-语言-行动(VLA)模型通过利用大规模的2D视觉语言预训练在机器人任务中表现出色,但它们对RGB图像的依赖限制了对现实世界交互至关重要的空间推理。用3D数据重新训练这些模型在计算上是禁止的,而丢弃现有的2D数据集浪费了宝贵的资源。为了弥补这一差距,我们提出了PointVLA,这是一个使用点云输入增强预训练vla而无需再训练的框架。我们的方法是冻结普通的动作专家,并通过轻量级模块块注入3D功能。为了确定整合点云表示的最有效方法,我们进行了跳过块分析,以确定vanilla动作专家中不太有用的块,确保3D特征仅注入到这些块中,从而最大限度地减少对预训练表示的干扰。大量实验表明,PointVLA在模拟和现实机器人任务中都优于最先进的2D模仿学习方法,如OpenVLA、Diffusion Policy和DexVLA。具体来说,我们强调了点云集成支持的PointVLA的几个关键优势:(1)少镜头多任务,其中PointVLA成功执行四个不同的任务,每个任务仅使用20个演示;(2) real -vs-photo - discrimination, PointVLA将真实物体与其图像区分开来,利用3D世界知识提高安全性和可靠性;(3)高度适应性,与传统的2D模仿学习方法不同,PointVLA使机器人能够适应不同桌子高度的物体,这些物体在训练数据中是看不见的。此外,PointVLA在长期任务中表现出色,例如从移动的传送带中挑选和包装物体,展示了其在复杂动态环境中的泛化能力。
{"title":"PointVLA: Injecting the 3D World Into Vision-Language-Action Models","authors":"Chengmeng Li;Junjie Wen;Yaxin Peng;Yan Peng;Yichen Zhu","doi":"10.1109/LRA.2026.3653303","DOIUrl":"https://doi.org/10.1109/LRA.2026.3653303","url":null,"abstract":"Vision-Language-Action (VLA) models excel at robotic tasks by leveraging large-scale 2D vision-language pretraining, but their reliance on RGB images limits spatial reasoning critical for real-world interaction. Retraining these models with 3D data is computationally prohibitive, while discarding existing 2D datasets wastes valuable resources. To bridge this gap, we propose PointVLA, a framework that enhances pre-trained VLAs with point cloud inputs without requiring retraining. Our method freezes the vanilla action expert and injects 3D features via a lightweight modular block. To identify the most effective way of integrating point cloud representations, we conduct a skip-block analysis to pinpoint less useful blocks in the vanilla action expert, ensuring that 3D features are injected only into these blocks—minimizing disruption to pre-trained representations. Extensive experiments demonstrate that PointVLA outperforms state-of-the-art 2D imitation learning methods, such as OpenVLA, Diffusion Policy and DexVLA, across both simulated and real-world robotic tasks. Specifically, we highlight several key advantages of PointVLA enabled by point cloud integration: (1) <bold>Few-shot multi-tasking</b>, where PointVLA successfully performs four different tasks using only 20 demonstrations each; (2) <bold>Real-vs-photo discrimination</b>, where PointVLA distinguishes real objects from their images, leveraging 3D world knowledge to improve safety and reliability; (3) <bold>Height adaptability</b>, where unlike conventional 2D imitation learning methods, PointVLA enables robots to adapt to objects at varying table heights that were unseen in training data. Furthermore, PointVLA achieves strong performance in long-horizon tasks, such as picking and packing objects from a moving conveyor belt, showcasing its ability to generalize across complex, dynamic environments.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"11 3","pages":"2506-2513"},"PeriodicalIF":5.3,"publicationDate":"2026-01-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146001872","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Distributed 3-D Multi-Robot Cooperative Localization: An Efficient and Consistent Approach 分布式三维多机器人协同定位:一种高效一致的方法
IF 5.3 2区 计算机科学 Q2 ROBOTICS Pub Date : 2026-01-12 DOI: 10.1109/LRA.2026.3653287
Yizhi Zhou;Yufan Liu;Xuan Wang
This letter studies the problem of Cooperative Localization (CL) for multi-robot systems in 3-D environments, where a group of mobile robots jointly localize themselves by using measurements from onboard sensors and shared information from other robots. To ensure the efficiency of information fusion and observability consistency in a distributed CL system, we propose a distributed multi-robot CL method based on Lie groups, well-suited for 3-D scenarios with full 3-D rotational dynamics and generic nonlinear inter-robot measurement models. Unlike most existing distributed CL algorithms that operate in vector space and are only applicable to simple 2-D environments, the proposed algorithm performs distributed information fusion directly on the manifold that inherently accounts for the non-Euclidean nature of 3-D rotations and translations. By leveraging the nice property of invariant errors, we analytically prove that the proposed algorithm naturally preserves the observability consistency of the CL system. This ensures that the system maintains the correct structure of unobservable directions throughout the estimation process. The effectiveness of the proposed algorithm is validated by several numerical experiments conducted to rigorously investigate the effects of relative information fusion in the distributed CL system.
本文研究了三维环境下多机器人系统的协同定位(CL)问题,其中一组移动机器人通过使用机载传感器的测量数据和其他机器人的共享信息共同定位自己。为了保证分布式CL系统中信息融合的效率和可观测性一致性,我们提出了一种基于李群的分布式多机器人CL方法,该方法适用于具有全三维旋转动力学和通用非线性机器人间测量模型的三维场景。与大多数在矢量空间中操作且仅适用于简单的二维环境的现有分布式CL算法不同,该算法直接在流形上执行分布式信息融合,该流形固有地解释了三维旋转和平移的非欧几里德性质。利用误差不变的良好性质,我们分析证明了所提出的算法自然地保持了CL系统的可观察一致性。这确保了系统在整个估计过程中保持不可观察方向的正确结构。通过多个数值实验验证了该算法的有效性,并对分布式CL系统中相对信息融合的效果进行了严格的研究。
{"title":"Distributed 3-D Multi-Robot Cooperative Localization: An Efficient and Consistent Approach","authors":"Yizhi Zhou;Yufan Liu;Xuan Wang","doi":"10.1109/LRA.2026.3653287","DOIUrl":"https://doi.org/10.1109/LRA.2026.3653287","url":null,"abstract":"This letter studies the problem of Cooperative Localization (CL) for multi-robot systems in 3-D environments, where a group of mobile robots jointly localize themselves by using measurements from onboard sensors and shared information from other robots. To ensure the efficiency of information fusion and observability consistency in a distributed CL system, we propose a distributed multi-robot CL method based on Lie groups, well-suited for 3-D scenarios with full 3-D rotational dynamics and generic nonlinear inter-robot measurement models. Unlike most existing distributed CL algorithms that operate in vector space and are only applicable to simple 2-D environments, the proposed algorithm performs distributed information fusion directly on the manifold that inherently accounts for the non-Euclidean nature of 3-D rotations and translations. By leveraging the nice property of invariant errors, we analytically prove that the proposed algorithm naturally preserves the observability consistency of the CL system. This ensures that the system maintains the correct structure of unobservable directions throughout the estimation process. The effectiveness of the proposed algorithm is validated by several numerical experiments conducted to rigorously investigate the effects of relative information fusion in the distributed CL system.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"11 2","pages":"2306-2313"},"PeriodicalIF":5.3,"publicationDate":"2026-01-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146026344","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An Anatomy-Aware Shared Control Approach for Assisted Teleoperation of Lung Ultrasound Examinations 肺超声检查辅助远程操作的解剖感知共享控制方法
IF 5.3 2区 计算机科学 Q2 ROBOTICS Pub Date : 2026-01-12 DOI: 10.1109/LRA.2026.3653292
Davide Nardi;Edoardo Lamon;Daniele Fontanelli;Matteo Saveriano;Luigi Palopoli
Although fully autonomous systems still face challenges due to patients' anatomical variability, teleoperated systems appear to be more practical in current healthcare settings. This paper presents an anatomy-aware control framework for teleoperated lung ultrasound. Leveraging biomechanically accurate 3D modelling, the system applies virtual constraints on the ultrasound probe pose and provides real-time visual feedback to assist in precise probe placement tasks. A twofold evaluation, one with 5 naïve operators on a single volunteer and the second with a single experienced operator on 6 volunteers, compared our method with a standard teleoperation baseline. The results of the first one characterised the accuracy of the anatomical model and the improved perceived performance by the naïve operators, while the second one focused on the efficiency of the system in improving probe placement and reducing procedure time compared to traditional teleoperation. The results demonstrate that the proposed framework enhances the physician's capabilities in executing remote lung ultrasound, reducing more than 20% of execution time on 4-point acquisitions, towards faster, more objective and repeatable exams.
尽管完全自主的系统仍然面临着挑战,由于患者的解剖变异,远程操作系统似乎在当前的医疗环境中更实用。提出了一种基于解剖感知的肺超声遥控控制框架。利用生物力学精确的3D建模,该系统对超声探头姿势施加虚拟约束,并提供实时视觉反馈,以协助精确的探头放置任务。一个双重评估,一个是5名naïve操作员对一个志愿者,第二个是一个有经验的操作员对6个志愿者,将我们的方法与标准的远程操作基线进行比较。第一项研究的结果体现了解剖模型的准确性和naïve操作人员感知性能的提高,而第二项研究的重点是与传统远程操作相比,该系统在改善探针放置和减少手术时间方面的效率。结果表明,所提出的框架提高了医生执行远程肺部超声的能力,减少了20%以上的4点采集执行时间,朝着更快、更客观和可重复的方向发展。
{"title":"An Anatomy-Aware Shared Control Approach for Assisted Teleoperation of Lung Ultrasound Examinations","authors":"Davide Nardi;Edoardo Lamon;Daniele Fontanelli;Matteo Saveriano;Luigi Palopoli","doi":"10.1109/LRA.2026.3653292","DOIUrl":"https://doi.org/10.1109/LRA.2026.3653292","url":null,"abstract":"Although fully autonomous systems still face challenges due to patients' anatomical variability, teleoperated systems appear to be more practical in current healthcare settings. This paper presents an anatomy-aware control framework for teleoperated lung ultrasound. Leveraging biomechanically accurate 3D modelling, the system applies virtual constraints on the ultrasound probe pose and provides real-time visual feedback to assist in precise probe placement tasks. A twofold evaluation, one with 5 naïve operators on a single volunteer and the second with a single experienced operator on 6 volunteers, compared our method with a standard teleoperation baseline. The results of the first one characterised the accuracy of the anatomical model and the improved perceived performance by the naïve operators, while the second one focused on the efficiency of the system in improving probe placement and reducing procedure time compared to traditional teleoperation. The results demonstrate that the proposed framework enhances the physician's capabilities in executing remote lung ultrasound, reducing more than 20% of execution time on 4-point acquisitions, towards faster, more objective and repeatable exams.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"11 3","pages":"2570-2577"},"PeriodicalIF":5.3,"publicationDate":"2026-01-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11346947","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146026562","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
TacFinRay: Soft Tactile Fin-Ray Finger With Indirect Tactile Sensing for Robust Grasping TacFinRay:具有间接触觉的软触觉鳍状射线手指,用于稳健抓取
IF 5.3 2区 计算机科学 Q2 ROBOTICS Pub Date : 2026-01-12 DOI: 10.1109/LRA.2026.3653299
Saekwang Nam;Bowen Deng;Loong Yi Lee;Jonathan M. Rossiter;Nathan F. Lepora
We present a tactile-sensorized Fin-Ray finger that enables simultaneous detection of contact location and indentation depth through an indirect sensing approach. A hinge mechanism is integrated between the soft Fin-Ray structure and a rigid sensing module, allowing deformation and translation information to be transferred to a bottom crossbeam upon which are an array of marker-tipped pins based on the biomimetic structure of the TacTip vision-based tactile sensor. Deformation patterns captured by an internal camera are processed using a convolutional neural network to infer contact conditions without directly sensing the finger surface. The finger design was optimized by varying pin configurations and hinge orientations, achieving 0.1 mm depth and 2 mm location-sensing accuracies. The perception demonstrated robust generalization to various indenter shapes and sizes, which was applied to a pick-and-place task under uncertain picking positions, where the tactile feedback significantly improved placement accuracy. Overall, this work provides a lightweight, flexible, and scalable tactile sensing solution suitable for soft robotic structures where the sensing needs situating away from the contact interface.
我们提出了一种触觉感应鳍射线手指,可以通过间接感应方法同时检测接触位置和压痕深度。在软鳍-射线结构和刚性传感模块之间集成了一个铰链机构,允许变形和平移信息传递到底部横梁上,横梁上是一系列基于基于tactical视觉的触觉传感器的仿生结构的标记针脚。由内部相机捕获的变形模式使用卷积神经网络进行处理,以推断接触条件,而无需直接感知手指表面。通过改变销钉配置和铰链方向,优化了手指设计,实现了0.1 mm的深度和2mm的位置传感精度。该感知对不同形状和尺寸的压头具有强大的泛化能力,并应用于不确定拾取位置下的拾取和放置任务,其中触觉反馈显著提高了放置精度。总的来说,这项工作提供了一种轻量级、灵活和可扩展的触觉传感解决方案,适用于需要远离接触界面的软机器人结构。
{"title":"TacFinRay: Soft Tactile Fin-Ray Finger With Indirect Tactile Sensing for Robust Grasping","authors":"Saekwang Nam;Bowen Deng;Loong Yi Lee;Jonathan M. Rossiter;Nathan F. Lepora","doi":"10.1109/LRA.2026.3653299","DOIUrl":"https://doi.org/10.1109/LRA.2026.3653299","url":null,"abstract":"We present a tactile-sensorized Fin-Ray finger that enables simultaneous detection of contact location and indentation depth through an indirect sensing approach. A hinge mechanism is integrated between the soft Fin-Ray structure and a rigid sensing module, allowing deformation and translation information to be transferred to a bottom crossbeam upon which are an array of marker-tipped pins based on the biomimetic structure of the TacTip vision-based tactile sensor. Deformation patterns captured by an internal camera are processed using a convolutional neural network to infer contact conditions without directly sensing the finger surface. The finger design was optimized by varying pin configurations and hinge orientations, achieving 0.1 mm depth and 2 mm location-sensing accuracies. The perception demonstrated robust generalization to various indenter shapes and sizes, which was applied to a pick-and-place task under uncertain picking positions, where the tactile feedback significantly improved placement accuracy. Overall, this work provides a lightweight, flexible, and scalable tactile sensing solution suitable for soft robotic structures where the sensing needs situating away from the contact interface.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"11 3","pages":"2722-2729"},"PeriodicalIF":5.3,"publicationDate":"2026-01-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146026583","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Christoffel-Consistent Coriolis Factorization and Its Effect on the Control of a Robot Christoffel-Consistent Coriolis分解及其对机器人控制的影响
IF 5.3 2区 计算机科学 Q2 ROBOTICS Pub Date : 2026-01-12 DOI: 10.1109/LRA.2026.3653394
Jonghyeok Kim;Wan Kyun Chung
Among the many choices in the matrix-vector factorization of the Coriolis and centripetal terms satisfying the skew-symmetry condition in system dynamics, the unique factorization, called Christoffel-consistent (CC) factorization, has been proposed. We derived the unique CC factorization in the Lie group context and examined the impact of Christoffel inconsistency in Coriolis matrix factorization on the dynamic behavior of robot systems during both free motion and interaction with humans, particularly in the context of passivity-based controllers and augmented PD controllers. Specifically, the question is: What are the advantages of using the CC factorization, and what is the effect of non-CC factorization on the robot’s dynamic behavior, which has been rarely explored? We showed that Christoffel inconsistency generates unwanted torsion, causing the system to deviate from the desired trajectory, and this results in undesirable dynamic behavior when controlling the system, especially when the dynamics of the robot is described by twist and wrench. Through simulation and a real-world robot experiment, this phenomenon is verified for the first time.
在系统动力学中满足偏对称条件的科里奥利项和向心项的矩阵-向量分解的众多选择中,提出了一种独特的分解方法,称为Christoffel-consistent (CC)分解。我们在李群环境中推导了独特的CC分解,并研究了科里奥利矩阵分解中Christoffel不一致性对机器人系统在自由运动和与人交互过程中的动态行为的影响,特别是在基于被动的控制器和增强PD控制器的环境中。具体来说,问题是:使用CC分解的优势是什么,以及非CC分解对机器人动态行为的影响是什么,这一点很少被探索。我们发现Christoffel不一致性会产生不需要的扭转,导致系统偏离期望的轨迹,从而导致控制系统时不需要的动态行为,特别是当机器人的动力学用扭转和扳手来描述时。通过仿真和真实机器人实验,首次验证了这一现象。
{"title":"Christoffel-Consistent Coriolis Factorization and Its Effect on the Control of a Robot","authors":"Jonghyeok Kim;Wan Kyun Chung","doi":"10.1109/LRA.2026.3653394","DOIUrl":"https://doi.org/10.1109/LRA.2026.3653394","url":null,"abstract":"Among the many choices in the matrix-vector factorization of the Coriolis and centripetal terms satisfying the skew-symmetry condition in system dynamics, the unique factorization, called Christoffel-consistent (CC) factorization, has been proposed. We derived the unique CC factorization in the Lie group context and examined the impact of Christoffel inconsistency in Coriolis matrix factorization on the dynamic behavior of robot systems during both free motion and interaction with humans, particularly in the context of passivity-based controllers and augmented PD controllers. Specifically, the question is: What are the advantages of using the CC factorization, and what is the effect of non-CC factorization on the robot’s dynamic behavior, which has been rarely explored? We showed that Christoffel inconsistency generates unwanted torsion, causing the system to deviate from the desired trajectory, and this results in undesirable dynamic behavior when controlling the system, especially when the dynamics of the robot is described by twist and wrench. Through simulation and a real-world robot experiment, this phenomenon is verified for the first time.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"11 3","pages":"2682-2689"},"PeriodicalIF":5.3,"publicationDate":"2026-01-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146026592","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Development and Control of Supernumerary Robotic Limbs for Overhead Tube Manipulation Task 架空管操作任务中多余机械臂的研制与控制
IF 5.3 2区 计算机科学 Q2 ROBOTICS Pub Date : 2026-01-12 DOI: 10.1109/LRA.2026.3652069
Jianxi Zhang;Jingtian Zhang;Hong Zeng;Dapeng Chen;Huijun Li;Aiguo Song
The foreign objects on utility poles may damage power lines and cause significant disruptions in electricity supply. A widely used approach to address this issue is for qualified personnel to climb on the pole and remove the foreign objects in a timely manner using an insulating tube. However, prolonged overhead manipulation of the insulating tube in the constrained environment not only leads to considerable upper-limb fatigue but also makes accurate tube positioning increasingly challenging. To address these challenges, wearable robotic limbs with an active control strategy have the potential to effectively reduce upper-limb fatigue and assist in tube positioning. This work presents supernumerary robotic limbs (SRLs) designed to assist electrical workers in a simulated overhead foreign objects removal task. We further propose a shared control method based on finite-horizon non-zero-sum game theory. This method models the cooperation between the SRL and the worker to adaptively modulate the input of the SRL, thereby providing rapid and accurate assistance in tube positioning. Experimental results show that the proposed SRL can reduce primary upper-limb muscle activity (deltoid, biceps brachii, brachioradialis and flexor carpi radialis) by up to 59.73% compared with performing the task without the SRL. Moreover, compared with a method that ignores human input, the proposed control strategy achieves more accurate positioning during human-SRLs cooperation. These results demonstrate the potential of both the SRL and the control strategy for the live-line overhead foreign objects removal task.
电线杆上的异物可能会损坏电线,造成电力供应的严重中断。解决该问题的常用方法是由有资质的人员爬上杆子,利用绝缘管及时清除异物。然而,在受限的环境中,长时间的架空操作不仅会导致上肢疲劳,而且使准确的管定位越来越具有挑战性。为了应对这些挑战,具有主动控制策略的可穿戴机器人四肢有可能有效地减少上肢疲劳并协助管道定位。这项工作提出了多余的机械肢体(srl),旨在协助电气工人在模拟的头顶异物清除任务。我们进一步提出了一种基于有限视界非零和博弈理论的共享控制方法。该方法模拟了SRL和工作人员之间的合作,以自适应调节SRL的输入,从而为管道定位提供快速准确的帮助。实验结果表明,与不进行SRL相比,SRL可使上肢主要肌肉活动(三角肌、肱二头肌、肱桡肌和桡腕屈肌)减少59.73%。此外,与忽略人工输入的方法相比,该控制策略在人- srl合作过程中实现了更精确的定位。这些结果证明了SRL和控制策略对于在线架空异物去除任务的潜力。
{"title":"Development and Control of Supernumerary Robotic Limbs for Overhead Tube Manipulation Task","authors":"Jianxi Zhang;Jingtian Zhang;Hong Zeng;Dapeng Chen;Huijun Li;Aiguo Song","doi":"10.1109/LRA.2026.3652069","DOIUrl":"https://doi.org/10.1109/LRA.2026.3652069","url":null,"abstract":"The foreign objects on utility poles may damage power lines and cause significant disruptions in electricity supply. A widely used approach to address this issue is for qualified personnel to climb on the pole and remove the foreign objects in a timely manner using an insulating tube. However, prolonged overhead manipulation of the insulating tube in the constrained environment not only leads to considerable upper-limb fatigue but also makes accurate tube positioning increasingly challenging. To address these challenges, wearable robotic limbs with an active control strategy have the potential to effectively reduce upper-limb fatigue and assist in tube positioning. This work presents supernumerary robotic limbs (SRLs) designed to assist electrical workers in a simulated overhead foreign objects removal task. We further propose a shared control method based on finite-horizon non-zero-sum game theory. This method models the cooperation between the SRL and the worker to adaptively modulate the input of the SRL, thereby providing rapid and accurate assistance in tube positioning. Experimental results show that the proposed SRL can reduce primary upper-limb muscle activity (deltoid, biceps brachii, brachioradialis and flexor carpi radialis) by up to 59.73% compared with performing the task without the SRL. Moreover, compared with a method that ignores human input, the proposed control strategy achieves more accurate positioning during human-SRLs cooperation. These results demonstrate the potential of both the SRL and the control strategy for the live-line overhead foreign objects removal task.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"11 3","pages":"2634-2641"},"PeriodicalIF":5.3,"publicationDate":"2026-01-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146026609","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
EiGS: Event-Informed 3D Deblur Reconstruction With Gaussian Splatting eig:事件通知的3D去模糊重建与高斯飞溅
IF 5.3 2区 计算机科学 Q2 ROBOTICS Pub Date : 2026-01-12 DOI: 10.1109/LRA.2026.3653290
Yuchen Weng;Nuo Li;Peng Yu;Qi Wang;Yongqiang Qi;Shaoze You;Jun Wang
Neural Radiance Fields (NeRF) have significantly advanced photorealistic novel view synthesis. Recently, 3D Gaussian Splatting has emerged as a promising technique with faster training and rendering speeds. However, both methods rely heavily on clear images and precise camera poses, limiting performance under motion blur. To address this, we introduce Event-Informed 3D Deblur Reconstruction with Gaussian Splatting(EiGS), a novel approach leveraging event camera data to enhance 3D Gaussian Splatting, improving sharpness and clarity in scenes affected by motion blur. Our method employs an Adaptive Deviation Estimator to learn Gaussian center shifts as the inverse of complex camera jitter, enabling simulation of motion blur during training. A motion consistency loss ensures global coherence in Gaussian displacements, while Blurriness and Event Integration Losses guide the model toward precise 3D representations. Extensive experiments demonstrate superior sharpness and real-time rendering capabilities compared to existing methods, with ablation studies validating the effectiveness of our components in robust, high-quality reconstruction for complex static scenes.
神经辐射场(Neural Radiance Fields, NeRF)在真实感新视角合成方面具有显著的进步。最近,3D高斯飞溅已经成为一种有前途的技术,具有更快的训练和渲染速度。然而,这两种方法都严重依赖于清晰的图像和精确的相机姿势,限制了运动模糊下的性能。为了解决这个问题,我们引入了带有高斯飞溅的事件通知3D去模糊重建(EiGS),这是一种利用事件相机数据增强3D高斯飞溅的新方法,提高了受运动模糊影响场景的清晰度和清晰度。我们的方法采用自适应偏差估计器来学习高斯中心偏移作为复杂相机抖动的逆,从而在训练过程中模拟运动模糊。运动一致性损失确保了高斯位移的全局一致性,而模糊和事件集成损失指导模型走向精确的3D表示。与现有方法相比,大量的实验证明了优越的清晰度和实时渲染能力,烧蚀研究验证了我们的组件在复杂静态场景的鲁棒性,高质量重建中的有效性。
{"title":"EiGS: Event-Informed 3D Deblur Reconstruction With Gaussian Splatting","authors":"Yuchen Weng;Nuo Li;Peng Yu;Qi Wang;Yongqiang Qi;Shaoze You;Jun Wang","doi":"10.1109/LRA.2026.3653290","DOIUrl":"https://doi.org/10.1109/LRA.2026.3653290","url":null,"abstract":"Neural Radiance Fields (NeRF) have significantly advanced photorealistic novel view synthesis. Recently, 3D Gaussian Splatting has emerged as a promising technique with faster training and rendering speeds. However, both methods rely heavily on clear images and precise camera poses, limiting performance under motion blur. To address this, we introduce Event-Informed 3D Deblur Reconstruction with Gaussian Splatting(EiGS), a novel approach leveraging event camera data to enhance 3D Gaussian Splatting, improving sharpness and clarity in scenes affected by motion blur. Our method employs an Adaptive Deviation Estimator to learn Gaussian center shifts as the inverse of complex camera jitter, enabling simulation of motion blur during training. A motion consistency loss ensures global coherence in Gaussian displacements, while Blurriness and Event Integration Losses guide the model toward precise 3D representations. Extensive experiments demonstrate superior sharpness and real-time rendering capabilities compared to existing methods, with ablation studies validating the effectiveness of our components in robust, high-quality reconstruction for complex static scenes.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"11 3","pages":"2474-2481"},"PeriodicalIF":5.3,"publicationDate":"2026-01-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146001871","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
LSV-Loc: LiDAR to StreetView Image Cross-Modal Localization lv - loc:激光雷达到街景图像的跨模态定位
IF 5.3 2区 计算机科学 Q2 ROBOTICS Pub Date : 2026-01-12 DOI: 10.1109/LRA.2026.3653282
Sangmin Lee;Donghyun Choi;Jee-Hwan Ryu
Accurate global localization remains a fundamental challenge in autonomous vehicle navigation. Traditional methods typically rely on high-definition (HD) maps generated through prior traverses or utilize auxiliary sensors, such as a global positioning system (GPS). However, the above approaches are often limited by high costs, scalability issues, and decreased reliability where GPS is unavailable. Moreover, prior methods require route-specific sensor calibration and impose modality-specific constraints, which restrict generalization across different sensor types. The proposed framework addresses this limitation by leveraging a shared embedding space, learned via a weight-sharing Vision Transformer (ViT) encoder, that aligns heterogeneous sensor modalities, Light Detection and Ranging (LiDAR) images, and geo-tagged StreetView panoramas. The proposed alignment enables reliable cross-modal retrieval and coarse-level localization without HD-map priors or route-specific calibration. Further, to address the heading inconsistency between query LiDAR and StreetView, an equirectangular perspective-n-point (PnP) solver is proposed to refine the relative pose through patch-level feature correspondences. As a result, the framework achieves coarse 3-degree-of-freedom (DoF) localization from a single LiDAR scan and publicly available StreetView imagery, bridging the gap between place recognition and metric localization. Experiments demonstrate that the proposed method achieves high recall and heading accuracy, offering scalability in urban settings covered by public Street View without reliance on HD maps.
准确的全球定位仍然是自动驾驶汽车导航的一个基本挑战。传统的方法通常依赖于通过预先穿越生成的高清(HD)地图或利用辅助传感器,如全球定位系统(GPS)。然而,上述方法通常受到高成本、可伸缩性问题以及GPS不可用时可靠性降低的限制。此外,先前的方法需要特定路线的传感器校准并施加特定模态的约束,这限制了不同传感器类型的泛化。提出的框架通过利用共享嵌入空间来解决这一限制,通过权重共享视觉变压器(ViT)编码器学习,该编码器可以对齐异构传感器模式、光探测和测距(LiDAR)图像以及地理标记街景全景图。所提出的校准能够实现可靠的跨模态检索和粗略的定位,而无需hd地图先验或特定路线的校准。此外,为了解决查询LiDAR和StreetView之间的航向不一致问题,提出了一种等矩形视角-n-点(PnP)求解器,通过斑块级特征对应来细化相对姿态。因此,该框架通过单一激光雷达扫描和公开的街景图像实现了粗略的3自由度(DoF)定位,弥合了位置识别和度量定位之间的差距。实验表明,该方法具有较高的查全率和航向精度,在城市街景环境中具有可扩展性,无需依赖高清地图。
{"title":"LSV-Loc: LiDAR to StreetView Image Cross-Modal Localization","authors":"Sangmin Lee;Donghyun Choi;Jee-Hwan Ryu","doi":"10.1109/LRA.2026.3653282","DOIUrl":"https://doi.org/10.1109/LRA.2026.3653282","url":null,"abstract":"Accurate global localization remains a fundamental challenge in autonomous vehicle navigation. Traditional methods typically rely on high-definition (HD) maps generated through prior traverses or utilize auxiliary sensors, such as a global positioning system (GPS). However, the above approaches are often limited by high costs, scalability issues, and decreased reliability where GPS is unavailable. Moreover, prior methods require route-specific sensor calibration and impose modality-specific constraints, which restrict generalization across different sensor types. The proposed framework addresses this limitation by leveraging a shared embedding space, learned via a weight-sharing Vision Transformer (ViT) encoder, that aligns heterogeneous sensor modalities, Light Detection and Ranging (LiDAR) images, and geo-tagged StreetView panoramas. The proposed alignment enables reliable cross-modal retrieval and coarse-level localization without HD-map priors or route-specific calibration. Further, to address the heading inconsistency between query LiDAR and StreetView, an equirectangular perspective-n-point (PnP) solver is proposed to refine the relative pose through patch-level feature correspondences. As a result, the framework achieves coarse 3-degree-of-freedom (DoF) localization from a single LiDAR scan and publicly available StreetView imagery, bridging the gap between place recognition and metric localization. Experiments demonstrate that the proposed method achieves high recall and heading accuracy, offering scalability in urban settings covered by public Street View without reliance on HD maps.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"11 3","pages":"2514-2521"},"PeriodicalIF":5.3,"publicationDate":"2026-01-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146001877","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Lie Group Implicit Kinematics for Redundant Parallel Manipulators: Left-Trivialized Extended Jacobians and Gradient-Based Online Redundancy Flows for Singularity Avoidance 冗余并联机器人的李群隐式运动学:左平凡化扩展雅可比矩阵和基于梯度的在线冗余流奇异避免
IF 5.3 2区 计算机科学 Q2 ROBOTICS Pub Date : 2026-01-12 DOI: 10.1109/LRA.2026.3653387
Yifei Liu;Kefei Wen
We present a Lie group implicit formulation for kinematically redundant parallel manipulators that yields left-trivialized extended Jacobians for the extended task variable $x=(g,rho)in text{SE}(3)times mathcal {R}$. On top of this model we design a gradient-based redundancy flow on the redundancy manifold that empirically maintains a positive manipulability margin along prescribed $text{SE}(3)$ trajectories. The framework uses right-multiplicative state updates, remains compatible with automatic differentiation, and avoids mechanism-specific analytic Jacobians; it works with either direct inverse kinematics or a numeric solver. A specialization to $text{SO}(2)^{3}$ provides computation-friendly first- and second-order steps. We validate the approach on two representative mechanisms: a (6+3)-degree-of-freedom (DoF) Stewart platform and a Spherical–Revolute platform. Across dense-coverage orientation trajectories and interactive gamepad commands, the extended Jacobian remained well conditioned while the redundancy planner ran at approximately 2 kHz in software-in-the-loop on a laptop-class CPU. The method integrates cleanly with existing kinematic stacks and is suitable for real-time deployment.
我们提出了一个运动冗余并联机器人的李群隐式公式,该公式对扩展任务变量$x=(g,rho)in text{SE}(3) mathcal {R}$产生左平凡化扩展雅可比矩阵。在此模型之上,我们在冗余流形上设计了一个基于梯度的冗余流,该冗余流形沿着规定的$text{SE}(3)$轨迹经验地保持正可操作性裕度。该框架使用右乘状态更新,与自动微分保持兼容,并避免了特定于机制的解析雅可比矩阵;它适用于直接逆运动学或数值解。对$text{SO}(2)^{3}$的专门化提供了计算友好的一阶和二阶步骤。我们在两个代表性机构上验证了该方法:一个(6+3)自由度(DoF) Stewart平台和一个球面-转动平台。在密集覆盖的方向轨迹和交互式手柄命令中,扩展的雅可比矩阵保持良好的条件,而冗余规划器在笔记本电脑级CPU的环内软件中以大约2 kHz的频率运行。该方法与现有的运动学堆栈清晰地集成在一起,适合实时部署。
{"title":"Lie Group Implicit Kinematics for Redundant Parallel Manipulators: Left-Trivialized Extended Jacobians and Gradient-Based Online Redundancy Flows for Singularity Avoidance","authors":"Yifei Liu;Kefei Wen","doi":"10.1109/LRA.2026.3653387","DOIUrl":"https://doi.org/10.1109/LRA.2026.3653387","url":null,"abstract":"We present a Lie group implicit formulation for kinematically redundant parallel manipulators that yields left-trivialized extended Jacobians for the extended task variable <inline-formula><tex-math>$x=(g,rho)in text{SE}(3)times mathcal {R}$</tex-math></inline-formula>. On top of this model we design a gradient-based redundancy flow on the redundancy manifold that empirically maintains a positive manipulability margin along prescribed <inline-formula><tex-math>$text{SE}(3)$</tex-math></inline-formula> trajectories. The framework uses right-multiplicative state updates, remains compatible with automatic differentiation, and avoids mechanism-specific analytic Jacobians; it works with either direct inverse kinematics or a numeric solver. A specialization to <inline-formula><tex-math>$text{SO}(2)^{3}$</tex-math></inline-formula> provides computation-friendly first- and second-order steps. We validate the approach on two representative mechanisms: a (6+3)-degree-of-freedom (DoF) Stewart platform and a Spherical–Revolute platform. Across dense-coverage orientation trajectories and interactive gamepad commands, the extended Jacobian remained well conditioned while the redundancy planner ran at approximately 2 kHz in software-in-the-loop on a laptop-class CPU. The method integrates cleanly with existing kinematic stacks and is suitable for real-time deployment.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"11 2","pages":"2322-2329"},"PeriodicalIF":5.3,"publicationDate":"2026-01-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146026476","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE Robotics and Automation Letters
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1