Pub Date : 2026-01-08eCollection Date: 2025-01-01DOI: 10.3389/frobt.2025.1706910
Beril Yalcinkaya, Micael S Couceiro, Salviano Soares, António Valente
Robotic fleet management systems are increasingly vital for sustainable operations in agriculture, forestry, and other field domains where labor shortages, efficiency, and environmental concerns intersect. We present FORMIGA, a fleet management framework that integrates human operators and autonomous robots into a collaborative ecosystem. FORMIGA combines standardised communication through the Robot Operating System with a user-centered interface for monitoring and intervention, while also leveraging large language models to generate executable task code from natural language prompts. The framework was deployed and validated within the FEROX project, a European initiative addressing sustainable berry harvesting in remote environments. In simulation-based trials, FORMIGA demonstrated adaptive task allocation, reduced operator workload, and faster task completion compared to semi-autonomous control, enabling dynamic labor division between humans and robots. By enhancing productivity, supporting worker safety, and promoting resource-efficient operations, FORMIGA contributes to the economic, and environmental dimensions of sustainability, offering a transferable tool for advancing human-robot collaboration in field robotics.
{"title":"FORMIGA: a fleet management framework for sustainable human-robot collaboration in field robotics.","authors":"Beril Yalcinkaya, Micael S Couceiro, Salviano Soares, António Valente","doi":"10.3389/frobt.2025.1706910","DOIUrl":"10.3389/frobt.2025.1706910","url":null,"abstract":"<p><p>Robotic fleet management systems are increasingly vital for sustainable operations in agriculture, forestry, and other field domains where labor shortages, efficiency, and environmental concerns intersect. We present FORMIGA, a fleet management framework that integrates human operators and autonomous robots into a collaborative ecosystem. FORMIGA combines standardised communication through the Robot Operating System with a user-centered interface for monitoring and intervention, while also leveraging large language models to generate executable task code from natural language prompts. The framework was deployed and validated within the FEROX project, a European initiative addressing sustainable berry harvesting in remote environments. In simulation-based trials, FORMIGA demonstrated adaptive task allocation, reduced operator workload, and faster task completion compared to semi-autonomous control, enabling dynamic labor division between humans and robots. By enhancing productivity, supporting worker safety, and promoting resource-efficient operations, FORMIGA contributes to the economic, and environmental dimensions of sustainability, offering a transferable tool for advancing human-robot collaboration in field robotics.</p>","PeriodicalId":47597,"journal":{"name":"Frontiers in Robotics and AI","volume":"12 ","pages":"1706910"},"PeriodicalIF":3.0,"publicationDate":"2026-01-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12823971/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146054659","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-01-08eCollection Date: 2025-01-01DOI: 10.3389/frobt.2025.1656564
Xuan Lin
This paper presents a comparative study of data-driven acceleration techniques for mixed-integer bilinear programs (MIBLPs) applied to robot motion planning. MIBLPs combine discrete decision variables and nonlinear constraints, making them computationally challenging for real-time robotics applications. We investigate two reformulation strategies: (1) converting binary variables into continuous variables with complementarity constraints (MPCC), and (2) converting bilinear constraints into mixed-integer linear constraints using McCormick envelopes (MICP). Using offline computed solutions as datasets, we apply K-nearest neighbor methods to warm-start both reformulations. We experimented with the proposed data-driven MIBLP formulation for motion planning on a linear inverted pendulum with contacts, and planning motion using a single rigid body model with mode transitions and contacts. Our results demonstrate that when sufficient data is available, MICP achieves consistently fast solving speeds that are suitable for real-time computation, while MPCC achieves higher success rates with limited amount of data. Our approach is capable of planning motions for the SCALER robot platform to transition between bipedal and quadrupedal configurations to navigate around obstacles without pre-specified gaits. Code for reproducing our results is available at https://github.com/XuanLin/MIBLP_benchmark.
{"title":"Data-driven acceleration of mixed-integer bilinear programs: a comparative study for robot motion planning.","authors":"Xuan Lin","doi":"10.3389/frobt.2025.1656564","DOIUrl":"https://doi.org/10.3389/frobt.2025.1656564","url":null,"abstract":"<p><p>This paper presents a comparative study of data-driven acceleration techniques for mixed-integer bilinear programs (MIBLPs) applied to robot motion planning. MIBLPs combine discrete decision variables and nonlinear constraints, making them computationally challenging for real-time robotics applications. We investigate two reformulation strategies: (1) converting binary variables into continuous variables with complementarity constraints (MPCC), and (2) converting bilinear constraints into mixed-integer linear constraints using McCormick envelopes (MICP). Using offline computed solutions as datasets, we apply K-nearest neighbor methods to warm-start both reformulations. We experimented with the proposed data-driven MIBLP formulation for motion planning on a linear inverted pendulum with contacts, and planning motion using a single rigid body model with mode transitions and contacts. Our results demonstrate that when sufficient data is available, MICP achieves consistently fast solving speeds that are suitable for real-time computation, while MPCC achieves higher success rates with limited amount of data. Our approach is capable of planning motions for the SCALER robot platform to transition between bipedal and quadrupedal configurations to navigate around obstacles without pre-specified gaits. Code for reproducing our results is available at https://github.com/XuanLin/MIBLP_benchmark.</p>","PeriodicalId":47597,"journal":{"name":"Frontiers in Robotics and AI","volume":"12 ","pages":"1656564"},"PeriodicalIF":3.0,"publicationDate":"2026-01-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12824877/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146047146","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-01-07eCollection Date: 2025-01-01DOI: 10.3389/frobt.2025.1737238
Xiang Han
The path planning capability of autonomous robots in complex environments is crucial for their widespread application in the real world. However, long-term decision-making and sparse reward signals pose significant challenges to traditional reinforcement learning (RL) algorithms. Offline hierarchical reinforcement learning offers an effective approach by decomposing tasks into two stages: high-level subgoal generation and low-level subgoal attainment. Advanced Offline HRL methods, such as Guider and HIQL, typically introduce latent spaces in high-level policies to represent subgoals, thereby handling high-dimensional states and enhancing generalization. However, these approaches require the high-level policy to search and generate sub-objectives within a continuous latent space. This remains a complex and sample-inefficient challenge for policy optimization algorithms-particularly policy gradient-based PPO-often leading to unstable training and slow convergence. To address this core limitation, this paper proposes a novel offline hierarchical PPO framework-LG-H-PPO (Latent Graph-based Hierarchical PPO). The core innovation of LG-H-PPO lies in discretizing the continuous latent space into a structured "latent graph." By transforming high-level planning from challenging "continuous creation" to simple "discrete selection," LG-H-PPO substantially reduces the learning difficulty for the high-level policy. Preliminary experiments on standard D4RL offline navigation benchmarks demonstrate that LG-H-PPO achieves significant advantages over advanced baselines like Guider and HIQL in both convergence speed and final task success rates. The main contribution of this paper is introducing graph structures into latent variable HRL planning. This effectively simplifies the action space for high-level policies, enhancing the training efficiency and stability of offline HRL algorithms for long-sequence navigation tasks. It lays the foundation for future offline HRL research combining latent variable representations with explicit graph planning.
自主机器人在复杂环境中的路径规划能力对其在现实世界中的广泛应用至关重要。然而,长期决策和稀疏奖励信号对传统的强化学习(RL)算法提出了重大挑战。离线分层强化学习提供了一种有效的方法,它将任务分解为两个阶段:高级子目标生成和低级子目标实现。高级离线HRL方法,如Guider和HIQL,通常在高级策略中引入潜在空间来表示子目标,从而处理高维状态并增强泛化。然而,这些方法需要高层策略在连续的潜在空间中搜索和生成子目标。对于策略优化算法(尤其是基于策略梯度的ppo)来说,这仍然是一个复杂且样本效率低下的挑战,通常会导致训练不稳定和收敛缓慢。为了解决这一核心限制,本文提出了一种新的离线分层PPO框架- lg - h- PPO (Latent Graph-based hierarchical PPO)。LG-H-PPO的核心创新在于将连续潜空间离散成结构化的“潜图”。通过将高层次规划从具有挑战性的“连续创造”转变为简单的“离散选择”,LG-H-PPO大大降低了高层次政策的学习难度。在标准D4RL离线导航基准上的初步实验表明,与Guider和HIQL等先进基线相比,LG-H-PPO在收敛速度和最终任务成功率方面都具有显著优势。本文的主要贡献是将图结构引入到潜在变量HRL规划中。这有效地简化了高层策略的动作空间,提高了离线HRL算法对长序列导航任务的训练效率和稳定性。将潜在变量表示与显式图规划相结合,为未来的离线HRL研究奠定了基础。
{"title":"LG-H-PPO: offline hierarchical PPO for robot path planning on a latent graph.","authors":"Xiang Han","doi":"10.3389/frobt.2025.1737238","DOIUrl":"https://doi.org/10.3389/frobt.2025.1737238","url":null,"abstract":"<p><p>The path planning capability of autonomous robots in complex environments is crucial for their widespread application in the real world. However, long-term decision-making and sparse reward signals pose significant challenges to traditional reinforcement learning (RL) algorithms. Offline hierarchical reinforcement learning offers an effective approach by decomposing tasks into two stages: high-level subgoal generation and low-level subgoal attainment. Advanced Offline HRL methods, such as Guider and HIQL, typically introduce latent spaces in high-level policies to represent subgoals, thereby handling high-dimensional states and enhancing generalization. However, these approaches require the high-level policy to search and generate sub-objectives within a continuous latent space. This remains a complex and sample-inefficient challenge for policy optimization algorithms-particularly policy gradient-based PPO-often leading to unstable training and slow convergence. To address this core limitation, this paper proposes a novel offline hierarchical PPO framework-LG-H-PPO (Latent Graph-based Hierarchical PPO). The core innovation of LG-H-PPO lies in discretizing the continuous latent space into a structured \"latent graph.\" By transforming high-level planning from challenging \"continuous creation\" to simple \"discrete selection,\" LG-H-PPO substantially reduces the learning difficulty for the high-level policy. Preliminary experiments on standard D4RL offline navigation benchmarks demonstrate that LG-H-PPO achieves significant advantages over advanced baselines like Guider and HIQL in both convergence speed and final task success rates. The main contribution of this paper is introducing graph structures into latent variable HRL planning. This effectively simplifies the action space for high-level policies, enhancing the training efficiency and stability of offline HRL algorithms for long-sequence navigation tasks. It lays the foundation for future offline HRL research combining latent variable representations with explicit graph planning.</p>","PeriodicalId":47597,"journal":{"name":"Frontiers in Robotics and AI","volume":"12 ","pages":"1737238"},"PeriodicalIF":3.0,"publicationDate":"2026-01-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12819167/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146031202","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-01-07eCollection Date: 2025-01-01DOI: 10.3389/frobt.2025.1686350
Lin Cong, Xiaowei Sun, Xiaolu Xi, Ke Yuan, Yajing Cao, Qiang Xie, Yue Zhu
Study design: Prospective study.
Objectives: This study aimed to evaluate the accuracy and safety of robot-assisted anterior transpedicular screw (ATPS) fixation in human cervical spine specimens.
Methods: A spine robotic system was used to implant thirty-six 1.2 mm Kirschner wires (K-wires) into the cervical pedicles (C4-C7) of five human specimens. Accuracy was assessed by comparing the planned trajectories with the actual K-wire positions. The Gertzbein-Robbins classification system (GRS), adapted for cervical pedicles, was used to evaluate accuracy; Grades A and B (<2 mm pedicle breach) were considered clinically acceptable. Secondary metrics included entry point and angle offsets.
Results: Of the 36 K-wires implanted, nine were placed in C4 and C6, 10 in C5, and eight in C7. According to the adapted GRS, 25 placements (69.4%) were Grade A, 10 (27.8%) were Grade B, and one was Grade C, resulting in a 97.2% clinically acceptable placement rate. The mean target offset was 2.29 ± 1.72 mm, the entry offset was 2.47 ± 1.57 mm, and the angle offset was 5.67° ± 3.72°. No significant differences were observed between the left and right sides (p > 0.05).
Conclusion: Robot-assisted ATPS fixation in cervical specimens achieved high accuracy with 97.2% of placements rated clinically acceptable, indicating its technical feasibility and potential utility in anterior cervical procedures.
{"title":"Accuracy of robot-assisted anterior transpedicular screws in the subaxial cervical spine: an experimental study on human specimens.","authors":"Lin Cong, Xiaowei Sun, Xiaolu Xi, Ke Yuan, Yajing Cao, Qiang Xie, Yue Zhu","doi":"10.3389/frobt.2025.1686350","DOIUrl":"10.3389/frobt.2025.1686350","url":null,"abstract":"<p><strong>Study design: </strong>Prospective study.</p><p><strong>Objectives: </strong>This study aimed to evaluate the accuracy and safety of robot-assisted anterior transpedicular screw (ATPS) fixation in human cervical spine specimens.</p><p><strong>Methods: </strong>A spine robotic system was used to implant thirty-six 1.2 mm Kirschner wires (K-wires) into the cervical pedicles (C4-C7) of five human specimens. Accuracy was assessed by comparing the planned trajectories with the actual K-wire positions. The Gertzbein-Robbins classification system (GRS), adapted for cervical pedicles, was used to evaluate accuracy; Grades A and B (<2 mm pedicle breach) were considered clinically acceptable. Secondary metrics included entry point and angle offsets.</p><p><strong>Results: </strong>Of the 36 K-wires implanted, nine were placed in C4 and C6, 10 in C5, and eight in C7. According to the adapted GRS, 25 placements (69.4%) were Grade A, 10 (27.8%) were Grade B, and one was Grade C, resulting in a 97.2% clinically acceptable placement rate. The mean target offset was 2.29 ± 1.72 mm, the entry offset was 2.47 ± 1.57 mm, and the angle offset was 5.67° ± 3.72°. No significant differences were observed between the left and right sides (p > 0.05).</p><p><strong>Conclusion: </strong>Robot-assisted ATPS fixation in cervical specimens achieved high accuracy with 97.2% of placements rated clinically acceptable, indicating its technical feasibility and potential utility in anterior cervical procedures.</p>","PeriodicalId":47597,"journal":{"name":"Frontiers in Robotics and AI","volume":"12 ","pages":"1686350"},"PeriodicalIF":3.0,"publicationDate":"2026-01-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12819262/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146031243","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-01-07eCollection Date: 2025-01-01DOI: 10.3389/frobt.2025.1708564
Mikayla Schneider, Zane A Colvin, Alena M Grabowski, Cara Gonzalez Welker
For individuals with unilateral transtibial amputation, powered ankle-foot prostheses have the potential to reduce the metabolic rate of walking, which could contribute to improvements in mobility and quality of life; however, physiological improvements have not been consistently demonstrated in experimental studies. To improve our understanding of the biomechanical mechanisms that drive metabolic rate outcomes, we used a machine learning approach to model the relationship between multimodal biomechanical factors and the metabolic rate of walking with a powered ankle-foot prosthesis. Our model included 50 features describing spatiotemporal parameters, step-to-step transition work, joint kinematics, muscle activity, ground reaction forces, prosthesis settings, and subject characteristics, and resulted in a pseudo-R2 of 0.986. Accumulated local effects plots were used to visualize the direction and magnitude of the relationship between each feature and the metabolic rate of walking. The features with the largest effect on metabolic rate were peak unaffected side ankle inversion angle, leading affected leg positive work during the step-to-step transition, and peak affected knee extension angle. This work furthers our knowledge about the biomechanical and physiological response to powered ankle-foot prosthesis use and could assist in developing new strategies to drive reductions in metabolic rate.
{"title":"Modeling the biomechanical features affecting the metabolic rate of walking with a powered ankle-foot prosthesis.","authors":"Mikayla Schneider, Zane A Colvin, Alena M Grabowski, Cara Gonzalez Welker","doi":"10.3389/frobt.2025.1708564","DOIUrl":"10.3389/frobt.2025.1708564","url":null,"abstract":"<p><p>For individuals with unilateral transtibial amputation, powered ankle-foot prostheses have the potential to reduce the metabolic rate of walking, which could contribute to improvements in mobility and quality of life; however, physiological improvements have not been consistently demonstrated in experimental studies. To improve our understanding of the biomechanical mechanisms that drive metabolic rate outcomes, we used a machine learning approach to model the relationship between multimodal biomechanical factors and the metabolic rate of walking with a powered ankle-foot prosthesis. Our model included 50 features describing spatiotemporal parameters, step-to-step transition work, joint kinematics, muscle activity, ground reaction forces, prosthesis settings, and subject characteristics, and resulted in a pseudo-R<sup>2</sup> of 0.986. Accumulated local effects plots were used to visualize the direction and magnitude of the relationship between each feature and the metabolic rate of walking. The features with the largest effect on metabolic rate were peak unaffected side ankle inversion angle, leading affected leg positive work during the step-to-step transition, and peak affected knee extension angle. This work furthers our knowledge about the biomechanical and physiological response to powered ankle-foot prosthesis use and could assist in developing new strategies to drive reductions in metabolic rate.</p>","PeriodicalId":47597,"journal":{"name":"Frontiers in Robotics and AI","volume":"12 ","pages":"1708564"},"PeriodicalIF":3.0,"publicationDate":"2026-01-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12819243/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146031265","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-01-06eCollection Date: 2025-01-01DOI: 10.3389/frobt.2025.1616462
Kazuho Kawashima, Shadi Ghali, Justin W Collins, Ali Esmaeili
Background: Head-mounted virtual reality (VR) simulations are increasingly explored in healthcare, particularly in patient education, stroke rehabilitation, and surgical training. While VR-based simulation plays a growing role in robotic-assisted surgery (RAS) training, the implications of head-mounted VR in this context remain underexamined.
Method: This prospective, randomised, controlled trial with a single-arm crossover compared two training modalities: a head-mounted VR simulation and a conventional console-based simulation. Participants in the experimental group used head-mounted VR as their primary training method, while the control group trained on a conventional console. Both groups completed a running suture task at baseline, midterm, and final assessments on the surgical console. The primary outcome was the composite score from the final assessment.
Results: Fourteen participants were equally distributed between the two arms. Baseline scores showed no significant differences. A two-way repeated measures ANOVA demonstrated significant overall improvement across assessments (F (1.688, 20.26) = 48.34, p < 0.001, partial η2 = 0.80). No statistical difference was found in final composite scores (mean difference: 8.4 ± 9.45, p = 0.391, Cohen's d = -0.48), midterm scores, or granular kinematic data. However, non-inferiority could not be established as the confidence interval fell outside our pre-set margin. The crossover group required less time (mean difference: 39 ± 9.01 min, p = 0.004) and fewer attempts (mean difference: 8 ± 2.2, p = 0.009) to reach benchmark performance compared to the control group.
Conclusion: Both head-mounted VR and console-based training significantly improved fundamental RAS skills in novices. While our study showed that the VR training shortened the time and attempts required to reach proficiency benchmarks, the small scale of this trial and the breadth of the confidence intervals mean the results should be viewed as preliminary observations. These results provide an initial signal of feasibility that warrants confirmation in larger studies.
背景:头戴式虚拟现实(VR)模拟越来越多地应用于医疗保健领域,特别是在患者教育、中风康复和外科培训方面。虽然基于VR的模拟在机器人辅助手术(RAS)训练中发挥着越来越大的作用,但在这种情况下,头戴式VR的影响仍未得到充分研究。方法:这项前瞻性、随机、对照试验采用单臂交叉试验,比较了两种训练方式:头戴式VR模拟和传统的基于控制台的模拟。实验组的参与者使用头戴式VR作为他们的主要训练方法,而对照组则在传统的控制台进行训练。两组均在手术控制台上完成了基线、中期和最终评估的连续缝合任务。主要结果是最终评估的综合得分。结果:14名参与者平均分布在两组之间。基线评分无显著差异。双向重复测量方差分析显示各评估的总体改善显著(F (1.688, 20.26) = 48.34, p < 0.001,部分η2 = 0.80)。最终综合评分(平均差值:8.4±9.45,p = 0.391, Cohen’s d = -0.48)、中期评分或颗粒运动数据均无统计学差异。然而,由于置信区间超出了我们预先设定的范围,因此无法建立非劣效性。与对照组相比,交叉组达到基准性能所需的时间更短(平均差值:39±9.01 min, p = 0.004),尝试次数更少(平均差值:8±2.2,p = 0.009)。结论:头戴式VR和基于控制台的训练均能显著提高新手的RAS基础技能。虽然我们的研究表明虚拟现实训练缩短了达到熟练程度基准所需的时间和尝试,但该试验的小规模和置信区间的广度意味着结果应被视为初步观察结果。这些结果提供了可行性的初步信号,值得在更大规模的研究中得到证实。
{"title":"A feasibility study: a non-inferiority study comparing head-mounted and console-based virtual reality for robotic surgery training.","authors":"Kazuho Kawashima, Shadi Ghali, Justin W Collins, Ali Esmaeili","doi":"10.3389/frobt.2025.1616462","DOIUrl":"10.3389/frobt.2025.1616462","url":null,"abstract":"<p><strong>Background: </strong>Head-mounted virtual reality (VR) simulations are increasingly explored in healthcare, particularly in patient education, stroke rehabilitation, and surgical training. While VR-based simulation plays a growing role in robotic-assisted surgery (RAS) training, the implications of head-mounted VR in this context remain underexamined.</p><p><strong>Method: </strong>This prospective, randomised, controlled trial with a single-arm crossover compared two training modalities: a head-mounted VR simulation and a conventional console-based simulation. Participants in the experimental group used head-mounted VR as their primary training method, while the control group trained on a conventional console. Both groups completed a running suture task at baseline, midterm, and final assessments on the surgical console. The primary outcome was the composite score from the final assessment.</p><p><strong>Results: </strong>Fourteen participants were equally distributed between the two arms. Baseline scores showed no significant differences. A two-way repeated measures ANOVA demonstrated significant overall improvement across assessments (F (1.688, 20.26) = 48.34, p < 0.001, partial η<sup>2</sup> = 0.80). No statistical difference was found in final composite scores (mean difference: 8.4 ± 9.45, p = 0.391, Cohen's d = -0.48), midterm scores, or granular kinematic data. However, non-inferiority could not be established as the confidence interval fell outside our pre-set margin. The crossover group required less time (mean difference: 39 ± 9.01 min, p = 0.004) and fewer attempts (mean difference: 8 ± 2.2, p = 0.009) to reach benchmark performance compared to the control group.</p><p><strong>Conclusion: </strong>Both head-mounted VR and console-based training significantly improved fundamental RAS skills in novices. While our study showed that the VR training shortened the time and attempts required to reach proficiency benchmarks, the small scale of this trial and the breadth of the confidence intervals mean the results should be viewed as preliminary observations. These results provide an initial signal of feasibility that warrants confirmation in larger studies.</p>","PeriodicalId":47597,"journal":{"name":"Frontiers in Robotics and AI","volume":"12 ","pages":"1616462"},"PeriodicalIF":3.0,"publicationDate":"2026-01-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12816980/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146020069","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-01-05eCollection Date: 2025-01-01DOI: 10.3389/frobt.2025.1724149
Ryuma Shineha
A cybernetic avatar (CA) is a concept that encompasses not only avatars representing virtual bodies in cyberspace but also information and communication technology (ICT) and robotic technologies that enhance the physical, cognitive, and perceptual capabilities of humans. CAs can enable multiple people to remotely operate numerous avatars and robots together to perform complex tasks on a large scale and create the necessary infrastructure for their operation and other related activities. However, due to the novelty of this concept, the ethical, legal, and social implications (ELSI) of CAs have not been discussed sufficiently. Therefore, the objective of this paper is to provide an overview of ELSI in the context of a CA, taking into account the implications from fields similar to that of CAs, such as robotic avatars, virtual avatars, metaverses, virtual reality, extended reality, social robots, human-robot interaction, telepresence, telexistence, embodied technology, and exoskeletons. In our review of ELSI in related fields, we found common themes: safety and security, data privacy, identity theft and identity loss, manipulation, intellectual property management, user addiction and overdependence, cyber abuse, risk management in a specific domain (e.g., medical applications), regulatory gaps, balance between free expression and harmful content, accountability, transparency, distributive justice, prevention of inequalities, dual use, and conceptual changes of familiarity. These issues should not be ignored when considering the social implementation of CAs.
{"title":"Exploring the ethical, legal, and social implications of cybernetic avatars.","authors":"Ryuma Shineha","doi":"10.3389/frobt.2025.1724149","DOIUrl":"10.3389/frobt.2025.1724149","url":null,"abstract":"<p><p>A cybernetic avatar (CA) is a concept that encompasses not only avatars representing virtual bodies in cyberspace but also information and communication technology (ICT) and robotic technologies that enhance the physical, cognitive, and perceptual capabilities of humans. CAs can enable multiple people to remotely operate numerous avatars and robots together to perform complex tasks on a large scale and create the necessary infrastructure for their operation and other related activities. However, due to the novelty of this concept, the ethical, legal, and social implications (ELSI) of CAs have not been discussed sufficiently. Therefore, the objective of this paper is to provide an overview of ELSI in the context of a CA, taking into account the implications from fields similar to that of CAs, such as robotic avatars, virtual avatars, metaverses, virtual reality, extended reality, social robots, human-robot interaction, telepresence, telexistence, embodied technology, and exoskeletons. In our review of ELSI in related fields, we found common themes: safety and security, data privacy, identity theft and identity loss, manipulation, intellectual property management, user addiction and overdependence, cyber abuse, risk management in a specific domain (e.g., medical applications), regulatory gaps, balance between free expression and harmful content, accountability, transparency, distributive justice, prevention of inequalities, dual use, and conceptual changes of familiarity. These issues should not be ignored when considering the social implementation of CAs.</p>","PeriodicalId":47597,"journal":{"name":"Frontiers in Robotics and AI","volume":"12 ","pages":"1724149"},"PeriodicalIF":3.0,"publicationDate":"2026-01-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12812702/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146012845","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-01-05eCollection Date: 2025-01-01DOI: 10.3389/frobt.2025.1694338
Luisa Damiano, Antonio Fleres, Andrea Roli, Pasquale Stano
Wetware Network-Based Artificial Intelligence (WNAI) introduces a new approach to robotic cognition and artificial intelligence: autonomous cognitive agents built from synthetic chemical networks. Rooted in Wetware Neuromorphic Engineering, WNAI shifts the focus of this emerging field from disembodied computation and biological mimicry to reticular chemical self-organization as a substrate for cognition. At the theoretical level, WNAI integrates insights from network cybernetics, autopoietic theory and enaction to frame cognition as a materially grounded, emergent phenomenon. At the heuristic level, WNAI defines its role as complementary to existing leading approaches. On the one hand, it complements embodied AI and xenobotics by expanding the design space of artificial embodied cognition into fully synthetic domains. On the other hand, it engages in mutual exchange with neural network architectures, advancing cross-substrate principles of network-based cognition. At the technological level, WNAI offers a roadmap for implementing chemical neural networks and protocellular agents, with potential applications in robotic systems requiring minimal, adaptive, and substrate-sensitive intelligence. By situating wetware neuromorphic engineering within the broader landscape of robotics and AI, this article outlines a programmatic framework that highlights its potential to expand artificial cognition beyond silicon and biohybrid systems.
{"title":"Wetware network-based AI: a chemical approach to embodied cognition for robotics and artificial intelligence.","authors":"Luisa Damiano, Antonio Fleres, Andrea Roli, Pasquale Stano","doi":"10.3389/frobt.2025.1694338","DOIUrl":"10.3389/frobt.2025.1694338","url":null,"abstract":"<p><p><i>Wetware Network-Based Artificial Intelligence</i> (WNAI) introduces a new approach to robotic cognition and artificial intelligence: autonomous cognitive agents built from synthetic chemical networks. Rooted in <i>Wetware Neuromorphic Engineering</i>, WNAI shifts the focus of this emerging field from disembodied computation and biological mimicry to reticular chemical self-organization as a substrate for cognition. At the <i>theoretical level</i>, WNAI integrates insights from network cybernetics, autopoietic theory and enaction to frame cognition as a materially grounded, emergent phenomenon. At the <i>heuristic level</i>, WNAI defines its role as complementary to existing leading approaches. On the one hand, it complements embodied AI and xenobotics by expanding the design space of artificial embodied cognition into fully synthetic domains. On the other hand, it engages in mutual exchange with neural network architectures, advancing cross-substrate principles of network-based cognition. At the <i>technological level</i>, WNAI offers a roadmap for implementing chemical neural networks and protocellular agents, with potential applications in robotic systems requiring minimal, adaptive, and substrate-sensitive intelligence. By situating wetware neuromorphic engineering within the broader landscape of robotics and AI, this article outlines a programmatic framework that highlights its potential to expand artificial cognition beyond silicon and biohybrid systems.</p>","PeriodicalId":47597,"journal":{"name":"Frontiers in Robotics and AI","volume":"12 ","pages":"1694338"},"PeriodicalIF":3.0,"publicationDate":"2026-01-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12812610/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146012858","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-01-02eCollection Date: 2025-01-01DOI: 10.3389/frobt.2025.1569040
Ahmed Salem, Kaoru Sumi
As robots became increasingly integrated into daily life, their ability to influence human emotions through verbal and nonverbal expressions is gaining attention. While robots have been explored for their role in emotional expression, their potential in emotion regulation particularly in mitigating or amplifying embarrassment remains under-explored in human-robot interaction. To address this gap, this study investigates whether and how robots can regulate the embarrassment emotion through their responses. A between-subjects experiment was conducted with 96 participants (48 males and 48 females) using the social robot Furhat. Participants experienced an embarrassing situation induced by a failure of meshing scenario, followed by the robot adopting one of three response attitudes: neutral, empathic, or ridiculing. Additionally, the robot's social agency was manipulated by varying its facial appearance between a human-like and an anime-like appearances. The findings indicate that embarrassment was effectively induced, as evidenced by physiological data, body movements, facial expressions, and participants' verbal responses. The anime-faced robot elicited lower embarrassment and arousal due to its lower perceived social agency and anthropomorphism. The robot's attitude was the dominant factor shaping participants' emotional responses and perceptions. The neutral and empathic attitudes with an anime face were found to be the most effective in eliciting mild emotions and mitigating embarrassment. Interestingly, an empathic attitude is suspected to be favored over a neutral one as it elicited the lowest embarrassment. However, an empathic attitude risks shaming the participant due to its indirect confrontation that inherently acknowledges the embarrassing incident which is undesirable in Japanese culture. Nevertheless, in terms of the robot's perceived evaluation by participants, a neutral attitude was the most favored. This study highlights the role of robot responses in emotion regulation, particularly in embarrassment control, and provides insights into designing socially intelligent robots that can modulate human emotions effectively.
{"title":"Embarrassment in HRI: remediation and the role of robot responses in emotion control.","authors":"Ahmed Salem, Kaoru Sumi","doi":"10.3389/frobt.2025.1569040","DOIUrl":"10.3389/frobt.2025.1569040","url":null,"abstract":"<p><p>As robots became increasingly integrated into daily life, their ability to influence human emotions through verbal and nonverbal expressions is gaining attention. While robots have been explored for their role in emotional expression, their potential in emotion regulation particularly in mitigating or amplifying embarrassment remains under-explored in human-robot interaction. To address this gap, this study investigates whether and how robots can regulate the embarrassment emotion through their responses. A between-subjects experiment was conducted with 96 participants (48 males and 48 females) using the social robot Furhat. Participants experienced an embarrassing situation induced by a failure of meshing scenario, followed by the robot adopting one of three response attitudes: neutral, empathic, or ridiculing. Additionally, the robot's social agency was manipulated by varying its facial appearance between a human-like and an anime-like appearances. The findings indicate that embarrassment was effectively induced, as evidenced by physiological data, body movements, facial expressions, and participants' verbal responses. The anime-faced robot elicited lower embarrassment and arousal due to its lower perceived social agency and anthropomorphism. The robot's attitude was the dominant factor shaping participants' emotional responses and perceptions. The neutral and empathic attitudes with an anime face were found to be the most effective in eliciting mild emotions and mitigating embarrassment. Interestingly, an empathic attitude is suspected to be favored over a neutral one as it elicited the lowest embarrassment. However, an empathic attitude risks shaming the participant due to its indirect confrontation that inherently acknowledges the embarrassing incident which is undesirable in Japanese culture. Nevertheless, in terms of the robot's perceived evaluation by participants, a neutral attitude was the most favored. This study highlights the role of robot responses in emotion regulation, particularly in embarrassment control, and provides insights into designing socially intelligent robots that can modulate human emotions effectively.</p>","PeriodicalId":47597,"journal":{"name":"Frontiers in Robotics and AI","volume":"12 ","pages":"1569040"},"PeriodicalIF":3.0,"publicationDate":"2026-01-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12807912/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145999440","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-19eCollection Date: 2025-01-01DOI: 10.3389/frobt.2025.1708987
Matija Mavsar, Mihael Simonič, Aleš Ude
Collaboration between humans and robots is essential for optimizing the performance of complex tasks in industrial environments, reducing worker strain, and improving safety. This paper presents an integrated human-robot collaboration (HRC) system that leverages advanced intention recognition for real-time task sharing and interaction. By utilizing state-of-the-art human pose estimation combined with deep learning models, we developed a robust framework for detecting and predicting worker intentions. Specifically, we employed LSTM-based and transformer-based neural networks with convolutional and pooling layers to classify human hand trajectories, achieving higher accuracy compared to previous approaches. Additionally, our system integrates dynamic movement primitives (DMPs) for smooth robot motion transitions, collision prevention, and automatic motion onset/cessation detection. We validated the system in a real-world industrial assembly task, demonstrating its effectiveness in enhancing the fluency, safety, and efficiency of human-robot collaboration. The proposed method shows promise in improving real-time decision-making in collaborative environments, offering a safer and more intuitive interaction between humans and robots.
{"title":"Human intention recognition by deep LSTM and transformer networks for real-time human-robot collaboration.","authors":"Matija Mavsar, Mihael Simonič, Aleš Ude","doi":"10.3389/frobt.2025.1708987","DOIUrl":"10.3389/frobt.2025.1708987","url":null,"abstract":"<p><p>Collaboration between humans and robots is essential for optimizing the performance of complex tasks in industrial environments, reducing worker strain, and improving safety. This paper presents an integrated human-robot collaboration (HRC) system that leverages advanced intention recognition for real-time task sharing and interaction. By utilizing state-of-the-art human pose estimation combined with deep learning models, we developed a robust framework for detecting and predicting worker intentions. Specifically, we employed LSTM-based and transformer-based neural networks with convolutional and pooling layers to classify human hand trajectories, achieving higher accuracy compared to previous approaches. Additionally, our system integrates dynamic movement primitives (DMPs) for smooth robot motion transitions, collision prevention, and automatic motion onset/cessation detection. We validated the system in a real-world industrial assembly task, demonstrating its effectiveness in enhancing the fluency, safety, and efficiency of human-robot collaboration. The proposed method shows promise in improving real-time decision-making in collaborative environments, offering a safer and more intuitive interaction between humans and robots.</p>","PeriodicalId":47597,"journal":{"name":"Frontiers in Robotics and AI","volume":"12 ","pages":"1708987"},"PeriodicalIF":3.0,"publicationDate":"2025-12-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12757248/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145900997","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}