Pub Date : 2025-11-03eCollection Date: 2025-01-01DOI: 10.3389/frobt.2025.1674421
Kamilya Smagulova, Ahmed Elsheikh, Diego A Silva, Mohammed E Fouda, Ahmed M Eltawil
Autonomous driving has the potential to enhance driving comfort and accessibility, reduce accidents, and improve road safety, with vision sensors playing a key role in enabling vehicle autonomy. Among existing sensors, event-based cameras offer advantages such as a high dynamic range, low power consumption, and enhanced motion detection capabilities compared to traditional frame-based cameras. However, their sparse and asynchronous data present unique processing challenges that require specialized algorithms and hardware. While some models originally developed for frame-based inputs have been adapted to handle event data, they often fail to fully exploit the distinct properties of this novel data format, primarily due to its fundamental structural differences. As a result, new algorithms, including neuromorphic, have been developed specifically for event data. Many of these models are still in the early stages and often lack the maturity and accuracy of traditional approaches. This survey paper focuses on end-to-end event-based object detection for autonomous driving, covering key aspects such as sensing and processing hardware designs, datasets, and algorithms, including dense, spiking, and graph-based neural networks, along with relevant encoding and pre-processing techniques. In addition, this work highlights the shortcomings in the evaluation practices to ensure fair and meaningful comparisons across different event data processing approaches and hardware platforms. Within the scope of this survey, system-level throughput was evaluated from raw event data to model output on an RTX 4090 24GB GPU for several state-of-the-art models using the GEN1 and 1MP datasets. The study also includes a discussion and outlines potential directions for future research.
{"title":"Efficient and real-time perception: a survey on end-to-end event-based object detection in autonomous driving.","authors":"Kamilya Smagulova, Ahmed Elsheikh, Diego A Silva, Mohammed E Fouda, Ahmed M Eltawil","doi":"10.3389/frobt.2025.1674421","DOIUrl":"10.3389/frobt.2025.1674421","url":null,"abstract":"<p><p>Autonomous driving has the potential to enhance driving comfort and accessibility, reduce accidents, and improve road safety, with vision sensors playing a key role in enabling vehicle autonomy. Among existing sensors, event-based cameras offer advantages such as a high dynamic range, low power consumption, and enhanced motion detection capabilities compared to traditional frame-based cameras. However, their sparse and asynchronous data present unique processing challenges that require specialized algorithms and hardware. While some models originally developed for frame-based inputs have been adapted to handle event data, they often fail to fully exploit the distinct properties of this novel data format, primarily due to its fundamental structural differences. As a result, new algorithms, including neuromorphic, have been developed specifically for event data. Many of these models are still in the early stages and often lack the maturity and accuracy of traditional approaches. This survey paper focuses on end-to-end event-based object detection for autonomous driving, covering key aspects such as sensing and processing hardware designs, datasets, and algorithms, including dense, spiking, and graph-based neural networks, along with relevant encoding and pre-processing techniques. In addition, this work highlights the shortcomings in the evaluation practices to ensure fair and meaningful comparisons across different event data processing approaches and hardware platforms. Within the scope of this survey, system-level throughput was evaluated from raw event data to model output on an RTX 4090 24GB GPU for several state-of-the-art models using the GEN1 and 1MP datasets. The study also includes a discussion and outlines potential directions for future research.</p>","PeriodicalId":47597,"journal":{"name":"Frontiers in Robotics and AI","volume":"12 ","pages":"1674421"},"PeriodicalIF":3.0,"publicationDate":"2025-11-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12620194/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145551676","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Cleaning PV (photovoltaic) panels is essential for a PV station, as dirt or dust reduces the effective irradiation of solar energy and weakens the efficiency of converting solar energy into free electrons. The inconsistent (cleaning efficacy) and unsafe (summarized voltage and current) manual method is a challenge for a PV station. Therefore, this paper develops a cleaning robot with PV detection, path planning, and action control. Firstly, a lightweight Mobile-VIT (Mobile Vision Transformer) model with a Self-Attention mechanism was used to improve YOLOv8 (You Only Look Once v8), resulting in an accuracy of 91.08% and a processing speed of 215 fps (frames per second). Secondly, an A* and a DWA (Dynamic Window Approach) path planning algorithm were improved. The simulation result shows that the time consumption decreased from 1.19 to 0.66 s and the Turn Number decreased from 23 to 10 p (places). Finally, the robot was evaluated and calibrated in both indoor and outdoor environments. The results showed that the algorithm can successfully clean PV arrays without manual control, with the rate increasing by 23% after its implementation. This study supports the maintenance of PV stations and serves as a reference for technical applications of deep learning, computer vision, and robot navigation.
清洁光伏板对于光伏电站来说是必不可少的,因为污垢或灰尘会降低太阳能的有效辐射,并降低将太阳能转化为自由电子的效率。人工方法的不一致(清洁效果)和不安全(汇总电压和电流)是光伏电站面临的挑战。因此,本文开发了一种具有PV检测、路径规划和动作控制的清洁机器人。首先,采用具有自注意机制的轻量级Mobile- vit (Mobile Vision Transformer)模型对YOLOv8 (You Only Look Once v8)进行改进,使准确率达到91.08%,处理速度达到215 fps(帧/秒)。其次,对A*和DWA (Dynamic Window Approach)路径规划算法进行了改进。仿真结果表明,该算法耗时从1.19 s减少到0.66 s,转数从23位减少到10位。最后,在室内和室外环境下对机器人进行了评估和校准。结果表明,该算法可以在不需要人工控制的情况下成功清洗光伏阵列,实现后的清洗率提高了23%。本研究为光伏电站维护提供支持,为深度学习、计算机视觉、机器人导航等技术应用提供参考。
{"title":"A photovoltaic panel cleaning robot with a lightweight YOLO v8.","authors":"Jidong Luo, Guoyi Wang, Yanjiao Lei, Dong Wang, Yayong Chen, Hongzhou Zhang","doi":"10.3389/frobt.2025.1606774","DOIUrl":"10.3389/frobt.2025.1606774","url":null,"abstract":"<p><p>Cleaning PV (photovoltaic) panels is essential for a PV station, as dirt or dust reduces the effective irradiation of solar energy and weakens the efficiency of converting solar energy into free electrons. The inconsistent (cleaning efficacy) and unsafe (summarized voltage and current) manual method is a challenge for a PV station. Therefore, this paper develops a cleaning robot with PV detection, path planning, and action control. Firstly, a lightweight Mobile-VIT (Mobile Vision Transformer) model with a Self-Attention mechanism was used to improve YOLOv8 (You Only Look Once v8), resulting in an accuracy of 91.08% and a processing speed of 215 fps (frames per second). Secondly, an A* and a DWA (Dynamic Window Approach) path planning algorithm were improved. The simulation result shows that the time consumption decreased from 1.19 to 0.66 s and the Turn Number decreased from 23 to 10 p (places). Finally, the robot was evaluated and calibrated in both indoor and outdoor environments. The results showed that the algorithm can successfully clean PV arrays without manual control, with the rate increasing by 23% after its implementation. This study supports the maintenance of PV stations and serves as a reference for technical applications of deep learning, computer vision, and robot navigation.</p>","PeriodicalId":47597,"journal":{"name":"Frontiers in Robotics and AI","volume":"12 ","pages":"1606774"},"PeriodicalIF":3.0,"publicationDate":"2025-10-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12615241/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145543298","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-10-31eCollection Date: 2025-01-01DOI: 10.3389/frobt.2025.1671336
Sofia Miranda, Carlos Renato Vázquez, Manuel Navarro-Gutiérrez
Energy consumption is a key concern in modern industrial facilities. Power peak is also a relevant feature in industrial energy analysis and managment, since the electrical infrastructure must be implemented to provide not only the total consumed energy, but the power peaks. Collaborative robots are gaining popularity due to its flexible use and convenient set up. In this context, a power and energy consumption study of the popular UR10 collaborative robot of Universal Robots is reported in this work. For this, an experiment was conducted to obtain current consumption data from the UR10 API, when performing movements with different loads and parameters. Next, the dependency of the trajectory programming parameters on the power peak, total consumed energy, and time spent per trajectory was analyzed. The results show that the higher the speed limit and acceleration limit, the lower the total energy consumed per trajectory, but the higher the power peak. This behavior represents a trade-off: reducing the consumed energy involves increasing the peak power. Based on the captured data, artificial neural network models were trained to predict the power peak and the total energy consumed by the robot when performing a movement under certain parameters. These models were later used by a genetic optimization algorithm to obtain the best parameters for a given target position, providing the most efficient performance while fulfilling a power peak bound.
{"title":"Energy consumption analysis and optimization in collaborative robots.","authors":"Sofia Miranda, Carlos Renato Vázquez, Manuel Navarro-Gutiérrez","doi":"10.3389/frobt.2025.1671336","DOIUrl":"10.3389/frobt.2025.1671336","url":null,"abstract":"<p><p>Energy consumption is a key concern in modern industrial facilities. Power peak is also a relevant feature in industrial energy analysis and managment, since the electrical infrastructure must be implemented to provide not only the total consumed energy, but the power peaks. Collaborative robots are gaining popularity due to its flexible use and convenient set up. In this context, a power and energy consumption study of the popular UR10 collaborative robot of Universal Robots is reported in this work. For this, an experiment was conducted to obtain current consumption data from the UR10 API, when performing movements with different loads and parameters. Next, the dependency of the trajectory programming parameters on the power peak, total consumed energy, and time spent per trajectory was analyzed. The results show that the higher the speed limit and acceleration limit, the lower the total energy consumed per trajectory, but the higher the power peak. This behavior represents a trade-off: reducing the consumed energy involves increasing the peak power. Based on the captured data, artificial neural network models were trained to predict the power peak and the total energy consumed by the robot when performing a movement under certain parameters. These models were later used by a genetic optimization algorithm to obtain the best parameters for a given target position, providing the most efficient performance while fulfilling a power peak bound.</p>","PeriodicalId":47597,"journal":{"name":"Frontiers in Robotics and AI","volume":"12 ","pages":"1671336"},"PeriodicalIF":3.0,"publicationDate":"2025-10-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12616253/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145543305","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-10-30eCollection Date: 2025-01-01DOI: 10.3389/frobt.2025.1426676
Samantha Chapin, William Chapin, Erik Komendera
This article proposes a method that uses information about modules and desired assembly locations within a large truss structure to create a semantic and fiducial aided graph simultaneous localization and mapping (SF-GraphSLAM) algorithm that is better tailored for use during robotic in-space assembly and servicing operations. This is achieved by first reducing the number of modules using a mixed assembly method vs. a strut-by-strut method. Then, each module is correlated to a visual tag (in this article, an AprilTag) to reduce the number of elements being observed further from the number of sub-struts in that module to a single AprilTag marker. Two tags are required to ensure proper deployment of most deployable modules. Subsequently, we are able to use semantic information about the desired transformation matrix between any two adjacent module AprilTags within the desired assembly structure. For our experimentation, we expanded a factor graph smoothing and mapping model and added the semantic information, looking at the smaller number of landmark AprilTags, with a camera representing the robot for simplicity. The mathematical approach to arrive at this new method is included in this article, as are simulations to test it against the state of the art (SOA) using no structural knowledge. Overall, this research contributes to the SOA for both general SLAM work and, more specifically, to the underdeveloped field of SLAM for in-space assembly and servicing of large truss structures. It is critical to ensure that as a robot is assembling the modules, each module is within the desired tolerances to ensure the final structure is within the design requirements. Being able to build a virtual twin of the truss structure as it is being assembled is a key tent pole in achieving large space structures.
{"title":"Semantic and fiducial-aided graph simultaneous localization and mapping (SF-GraphSLAM) for robotic in-space assembly and servicing of large truss structures.","authors":"Samantha Chapin, William Chapin, Erik Komendera","doi":"10.3389/frobt.2025.1426676","DOIUrl":"10.3389/frobt.2025.1426676","url":null,"abstract":"<p><p>This article proposes a method that uses information about modules and desired assembly locations within a large truss structure to create a semantic and fiducial aided graph simultaneous localization and mapping (SF-GraphSLAM) algorithm that is better tailored for use during robotic in-space assembly and servicing operations. This is achieved by first reducing the number of modules using a mixed assembly method vs. a strut-by-strut method. Then, each module is correlated to a visual tag (in this article, an AprilTag) to reduce the number of elements being observed further from the number of sub-struts in that module to a single AprilTag marker. Two tags are required to ensure proper deployment of most deployable modules. Subsequently, we are able to use semantic information about the desired transformation matrix between any two adjacent module AprilTags within the desired assembly structure. For our experimentation, we expanded a factor graph smoothing and mapping model and added the semantic information, looking at the smaller number of landmark AprilTags, with a camera representing the robot for simplicity. The mathematical approach to arrive at this new method is included in this article, as are simulations to test it against the state of the art (SOA) using no structural knowledge. Overall, this research contributes to the SOA for both general SLAM work and, more specifically, to the underdeveloped field of SLAM for in-space assembly and servicing of large truss structures. It is critical to ensure that as a robot is assembling the modules, each module is within the desired tolerances to ensure the final structure is within the design requirements. Being able to build a virtual twin of the truss structure as it is being assembled is a key tent pole in achieving large space structures.</p>","PeriodicalId":47597,"journal":{"name":"Frontiers in Robotics and AI","volume":"12 ","pages":"1426676"},"PeriodicalIF":3.0,"publicationDate":"2025-10-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12613163/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145543228","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-10-30eCollection Date: 2025-01-01DOI: 10.3389/frobt.2025.1614444
Mary Ellen Foster, Jennifer N Stinson, Lauren Harris, Kate Kyuri Kim, Sasha Litwin, Patricia Candelaria, Summer Hudson, Julie Leung, Ronald P A Petrick, Alan Lindsay, Andrés Ramírez-Duque, David Harris Smith, Frauke Zeller, Samina Ali
Introduction: Children undergoing medical procedures in paediatric Emergency Departments (EDs) often experience significant pain and distress. Socially Assistive Robots (SARs) offer a promising avenue for delivering distraction and emotional support in these high-pressure environments. This study presents the design, development, and formative evaluation of an AI-enhanced SAR to support children during intravenous insertion (IVI) procedures.
Methods: The robot system was developed through a participatory design process involving healthcare professionals, patients, caregivers, and interdisciplinary research teams. The SAR was designed to autonomously adapt its behaviour to the child's affective state using AI planning and social signal processing. A two-cycle usability study was conducted across two Canadian paediatric EDs, involving 25 children and their caregivers. Feedback was collected through observations, interviews, and system logs.
Results: The SAR was successfully integrated into clinical workflows, with positive responses from children, caregivers, and healthcare providers. Usability testing identified key technical and interaction challenges, which were addressed through iterative refinement. The final system demonstrated robust performance and was deemed ready for a formal randomised controlled trial.
Discussion: This work highlights the importance of co-design, operator control, and environmental adaptability in deploying SARs in clinical settings. Lessons learned from the development and deployment process informed six concrete design guidelines for future SAR implementations in healthcare.
{"title":"Bringing a socially assistive robot to the paediatric emergency department: design, development, and usability testing.","authors":"Mary Ellen Foster, Jennifer N Stinson, Lauren Harris, Kate Kyuri Kim, Sasha Litwin, Patricia Candelaria, Summer Hudson, Julie Leung, Ronald P A Petrick, Alan Lindsay, Andrés Ramírez-Duque, David Harris Smith, Frauke Zeller, Samina Ali","doi":"10.3389/frobt.2025.1614444","DOIUrl":"10.3389/frobt.2025.1614444","url":null,"abstract":"<p><strong>Introduction: </strong>Children undergoing medical procedures in paediatric Emergency Departments (EDs) often experience significant pain and distress. Socially Assistive Robots (SARs) offer a promising avenue for delivering distraction and emotional support in these high-pressure environments. This study presents the design, development, and formative evaluation of an AI-enhanced SAR to support children during intravenous insertion (IVI) procedures.</p><p><strong>Methods: </strong>The robot system was developed through a participatory design process involving healthcare professionals, patients, caregivers, and interdisciplinary research teams. The SAR was designed to autonomously adapt its behaviour to the child's affective state using AI planning and social signal processing. A two-cycle usability study was conducted across two Canadian paediatric EDs, involving 25 children and their caregivers. Feedback was collected through observations, interviews, and system logs.</p><p><strong>Results: </strong>The SAR was successfully integrated into clinical workflows, with positive responses from children, caregivers, and healthcare providers. Usability testing identified key technical and interaction challenges, which were addressed through iterative refinement. The final system demonstrated robust performance and was deemed ready for a formal randomised controlled trial.</p><p><strong>Discussion: </strong>This work highlights the importance of co-design, operator control, and environmental adaptability in deploying SARs in clinical settings. Lessons learned from the development and deployment process informed six concrete design guidelines for future SAR implementations in healthcare.</p>","PeriodicalId":47597,"journal":{"name":"Frontiers in Robotics and AI","volume":"12 ","pages":"1614444"},"PeriodicalIF":3.0,"publicationDate":"2025-10-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12612628/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145543235","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-10-30eCollection Date: 2025-01-01DOI: 10.3389/frobt.2025.1589025
Davide Picchi, Sigrid Brell-Çokcan
Mini cranes play a pivotal role in construction due to their versatility across numerous scenarios. Recent advancements in Reinforcement Learning (RL) have enabled agents to operate cranes in virtual environments for predetermined tasks, paving the way for future real-world deployment. Traditionally, most RL agents use a squashed Gaussian distribution to select actions. In this study, we investigate a mini-crane scenario that could potentially be fully automated by AI and explore replacing the Gaussian distribution with the Kumaraswamy distribution, a close relative of the Beta distribution, for action stochastic selection. Our results indicate that the Kumaraswamy distribution offers computational advantages while maintaining robust performance, making it an attractive alternative for RL applications in continuous control applications.
{"title":"Exploiting the Kumaraswamy distribution in a reinforcement learning context.","authors":"Davide Picchi, Sigrid Brell-Çokcan","doi":"10.3389/frobt.2025.1589025","DOIUrl":"10.3389/frobt.2025.1589025","url":null,"abstract":"<p><p>Mini cranes play a pivotal role in construction due to their versatility across numerous scenarios. Recent advancements in Reinforcement Learning (RL) have enabled agents to operate cranes in virtual environments for predetermined tasks, paving the way for future real-world deployment. Traditionally, most RL agents use a squashed Gaussian distribution to select actions. In this study, we investigate a mini-crane scenario that could potentially be fully automated by AI and explore replacing the Gaussian distribution with the Kumaraswamy distribution, a close relative of the Beta distribution, for action stochastic selection. Our results indicate that the Kumaraswamy distribution offers computational advantages while maintaining robust performance, making it an attractive alternative for RL applications in continuous control applications.</p>","PeriodicalId":47597,"journal":{"name":"Frontiers in Robotics and AI","volume":"12 ","pages":"1589025"},"PeriodicalIF":3.0,"publicationDate":"2025-10-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12611641/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145543215","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-10-30eCollection Date: 2025-01-01DOI: 10.3389/frobt.2025.1660691
Junya Yamamoto, Kenji Tahara, Takahiro Wada
In response to the growing need for flexibility in handling complex tasks, research on human-robot collaboration (HRC) has garnered considerable attention. Recent studies on HRC have achieved smooth handover tasks between humans and robots by adaptively responding to human states. Collaboration was further improved by conveying the state of the robot to humans via robotic interactive motion cues. However, in scenarios such as collaborative assembly tasks that require precise positioning, methods relying on motion or forces caused by interactions through the shared object compromise both task accuracy and smoothness, and are therefore not directly applicable. To address this, the present study proposes a method to convey the stiffness of the robot to a human arm during collaborative human-robot assembly tasks in a manner that does not affect the shared object or task, aiming to enhance efficiency and reduce human workload. Sixteen participants performed a collaborative assembly task with a robot, which involved unscrewing, repositioning, and reattaching a part while the robot held and adjusted the position of the part. The experiment examined the effectiveness of the proposed method, in which the robot's stiffness was communicated to a participant's forearm. The independent variable, tested within-subjects, was the stiffness presentation method, with three levels: without the proposed method (no presentation) and with the proposed method (real-time and predictive presentations). The results demonstrated that the proposed method enhanced task efficiency by shortening task completion time, which was associated with lower subjective workload scores.
{"title":"Effect of presenting robot hand stiffness to human arm on human-robot collaborative assembly tasks.","authors":"Junya Yamamoto, Kenji Tahara, Takahiro Wada","doi":"10.3389/frobt.2025.1660691","DOIUrl":"10.3389/frobt.2025.1660691","url":null,"abstract":"<p><p>In response to the growing need for flexibility in handling complex tasks, research on human-robot collaboration (HRC) has garnered considerable attention. Recent studies on HRC have achieved smooth handover tasks between humans and robots by adaptively responding to human states. Collaboration was further improved by conveying the state of the robot to humans via robotic interactive motion cues. However, in scenarios such as collaborative assembly tasks that require precise positioning, methods relying on motion or forces caused by interactions through the shared object compromise both task accuracy and smoothness, and are therefore not directly applicable. To address this, the present study proposes a method to convey the stiffness of the robot to a human arm during collaborative human-robot assembly tasks in a manner that does not affect the shared object or task, aiming to enhance efficiency and reduce human workload. Sixteen participants performed a collaborative assembly task with a robot, which involved unscrewing, repositioning, and reattaching a part while the robot held and adjusted the position of the part. The experiment examined the effectiveness of the proposed method, in which the robot's stiffness was communicated to a participant's forearm. The independent variable, tested within-subjects, was the stiffness presentation method, with three levels: without the proposed method (no presentation) and with the proposed method (real-time and predictive presentations). The results demonstrated that the proposed method enhanced task efficiency by shortening task completion time, which was associated with lower subjective workload scores.</p>","PeriodicalId":47597,"journal":{"name":"Frontiers in Robotics and AI","volume":"12 ","pages":"1660691"},"PeriodicalIF":3.0,"publicationDate":"2025-10-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12611644/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145543261","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-10-30eCollection Date: 2025-01-01DOI: 10.3389/frobt.2025.1594529
Silvia Filogna, Giovanni Arras, Tommaso Turchi, Giuseppe Prencipe, Elena Beani, Clara Bombonato, Francesca Fedeli, Gemma D'Alessandro, Antea Scrocco, Giuseppina Sgandurra
Despite the growing interest in Artificial Intelligence (AI) for pediatric rehabilitation, family engagement in the technologies design remains limited. Understanding how AI-driven tools align with family needs, caregiving routines, and ethical concerns is crucial for their successful adoption. In this study, we actively involved nine families of children with Cerebral Palsy (CP) in an online participatory design workshop, underscoring both the feasibility and the need of integrating family's perspectives into AI development. Families enthusiastically participated, not only sharing insights but also appreciating the opportunity to contribute to shaping future technologies. Their active engagement challenges the assumption that co-design with families is complex or impractical, highlighting how structured yet flexible methodologies can make such crucial initiatives highly effective. The online format further facilitated participation, allowing families to join the discussion and ensuring a diverse range of perspectives. The workshop's key findings reveal three core priorities for families: 1. AI should adapt to daily caregiving routines rather than impose rigid structures; 2. digital tools should enhance communication and collaboration between families and clinicians, rather than replace human interaction; and 3. AI-driven systems could empower children's autonomy while maintaining parental oversight. Additionally, families raised critical concerns about data privacy, transparency, and the need to preserve empathy in AI-mediated care. Our findings reinforce the urgent need to shift toward family-centered AI design, moving beyond purely technological solutions toward ethically responsible, inclusive innovations. This research not only demonstrates the possibility and success of engaging families in co-design processes but also provides a model for future AI development that genuinely reflects the lived experiences of children and caregivers.
{"title":"Pathways to family-centered healthcare: co-designing AI solutions with families in pediatric rehabilitation.","authors":"Silvia Filogna, Giovanni Arras, Tommaso Turchi, Giuseppe Prencipe, Elena Beani, Clara Bombonato, Francesca Fedeli, Gemma D'Alessandro, Antea Scrocco, Giuseppina Sgandurra","doi":"10.3389/frobt.2025.1594529","DOIUrl":"10.3389/frobt.2025.1594529","url":null,"abstract":"<p><p>Despite the growing interest in Artificial Intelligence (AI) for pediatric rehabilitation, family engagement in the technologies design remains limited. Understanding how AI-driven tools align with family needs, caregiving routines, and ethical concerns is crucial for their successful adoption. In this study, we actively involved nine families of children with Cerebral Palsy (CP) in an online participatory design workshop, underscoring both the feasibility and the need of integrating family's perspectives into AI development. Families enthusiastically participated, not only sharing insights but also appreciating the opportunity to contribute to shaping future technologies. Their active engagement challenges the assumption that co-design with families is complex or impractical, highlighting how structured yet flexible methodologies can make such crucial initiatives highly effective. The online format further facilitated participation, allowing families to join the discussion and ensuring a diverse range of perspectives. The workshop's key findings reveal three core priorities for families: 1. AI should adapt to daily caregiving routines rather than impose rigid structures; 2. digital tools should enhance communication and collaboration between families and clinicians, rather than replace human interaction; and 3. AI-driven systems could empower children's autonomy while maintaining parental oversight. Additionally, families raised critical concerns about data privacy, transparency, and the need to preserve empathy in AI-mediated care. Our findings reinforce the urgent need to shift toward family-centered AI design, moving beyond purely technological solutions toward ethically responsible, inclusive innovations. This research not only demonstrates the possibility and success of engaging families in co-design processes but also provides a model for future AI development that genuinely reflects the lived experiences of children and caregivers.</p>","PeriodicalId":47597,"journal":{"name":"Frontiers in Robotics and AI","volume":"12 ","pages":"1594529"},"PeriodicalIF":3.0,"publicationDate":"2025-10-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12611681/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145543293","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Editorial: NeuroDesign in human-robot interaction: the making of engaging HRI technology your brain can't resist.","authors":"Ker-Jiun Wang, Ramana Vinjamuri, Maryam Alimardani, Tharun Kumar Reddy, Zhi-Hong Mao","doi":"10.3389/frobt.2025.1699371","DOIUrl":"https://doi.org/10.3389/frobt.2025.1699371","url":null,"abstract":"","PeriodicalId":47597,"journal":{"name":"Frontiers in Robotics and AI","volume":"12 ","pages":"1699371"},"PeriodicalIF":3.0,"publicationDate":"2025-10-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12611685/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145543249","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-10-29eCollection Date: 2025-01-01DOI: 10.3389/frobt.2025.1682031
Michaela Kümpel, Manuel Scheibl, Jan-Philipp Töberg, Vanessa Hassouna, Philipp Cimiano, Britta Wrede, Michael Beetz
This paper addresses the challenge of enabling robots to autonomously prepare meals by bridging natural language recipe instructions and robotic action execution. We propose a novel methodology leveraging Actionable Knowledge Graphs to map recipe instructions into six core categories of robotic manipulation tasks, termed Action Cores cutting, pouring, mixing, preparing, pick and place, and cook and cool. Each AC is subdivided into Action Groups which represent a specific motion parameterization required for task execution. Using the Recipe1M + dataset (Marín et al., IEEE Transactions on Pattern Analysis and Machine Intelligence, 2021, 43, 187-203), encompassing over one million recipes, we systematically analysed action verbs and matched them to ACs by using direct matching and cosine similarity, achieving a coverage of 76.5%. For the unmatched verbs, we employ a neuro-symbolic approach, matching verbs to existing AGs or generating new action cores utilizing a Large Language Model Our findings highlight the versatility of AKGs in adapting general plans to specific robotic tasks, validated through an experimental application in a meal preparation scenario. This work sets a foundation for adaptive robotic systems capable of performing a wide array of complex culinary tasks with minimal human intervention.
本文通过连接自然语言食谱指令和机器人动作执行,解决了使机器人能够自主准备饭菜的挑战。我们提出了一种新的方法,利用可操作知识图将配方指令映射到机器人操作任务的六个核心类别,称为动作核心切割,浇注,混合,准备,采摘和放置以及烹饪和冷却。每个AC被细分为动作组,动作组代表任务执行所需的特定动作参数化。使用Recipe1M +数据集(Marín et al., IEEE Transactions on Pattern Analysis and Machine Intelligence, 2021, 43, 187-203),包含超过100万个食谱,我们系统地分析了动作动词,并通过直接匹配和余cosine相似度将它们与ACs进行匹配,达到了76.5%的覆盖率。对于不匹配的动词,我们采用了一种神经符号方法,将动词与现有的AGs进行匹配,或者利用大型语言模型生成新的动作核心。我们的研究结果强调了akg在适应特定机器人任务的总体计划方面的多功能性,并通过在饭菜准备场景中的实验应用得到了验证。这项工作为自适应机器人系统奠定了基础,该系统能够在最少的人为干预下执行各种复杂的烹饪任务。
{"title":"Everything robots need to know about cooking actions: creating actionable knowledge graphs to support robotic meal preparation.","authors":"Michaela Kümpel, Manuel Scheibl, Jan-Philipp Töberg, Vanessa Hassouna, Philipp Cimiano, Britta Wrede, Michael Beetz","doi":"10.3389/frobt.2025.1682031","DOIUrl":"10.3389/frobt.2025.1682031","url":null,"abstract":"<p><p>This paper addresses the challenge of enabling robots to autonomously prepare meals by bridging natural language recipe instructions and robotic action execution. We propose a novel methodology leveraging Actionable Knowledge Graphs to map recipe instructions into six core categories of robotic manipulation tasks, termed Action Cores cutting, pouring, mixing, preparing, pick and place, and cook and cool. Each AC is subdivided into Action Groups which represent a specific motion parameterization required for task execution. Using the Recipe1M + dataset (Marín et al., IEEE Transactions on Pattern Analysis and Machine Intelligence, 2021, 43, 187-203), encompassing over one million recipes, we systematically analysed action verbs and matched them to ACs by using direct matching and cosine similarity, achieving a coverage of 76.5%. For the unmatched verbs, we employ a neuro-symbolic approach, matching verbs to existing AGs or generating new action cores utilizing a Large Language Model Our findings highlight the versatility of AKGs in adapting general plans to specific robotic tasks, validated through an experimental application in a meal preparation scenario. This work sets a foundation for adaptive robotic systems capable of performing a wide array of complex culinary tasks with minimal human intervention.</p>","PeriodicalId":47597,"journal":{"name":"Frontiers in Robotics and AI","volume":"12 ","pages":"1682031"},"PeriodicalIF":3.0,"publicationDate":"2025-10-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12605030/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145514757","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}