Pub Date : 2025-11-04eCollection Date: 2025-01-01DOI: 10.3389/frobt.2025.1680285
Sufola Das Chagas Silva Araujo, Goh Kah Ong Michael, Uttam U Deshpande, Sudhindra Deshpande, Manjunath G Avalappa, Yash Amasi, Sumit Patil, Swathi Bhat, Sudarshan Karigoudar
Current industrial robots deployed in small and medium-sized businesses (SMEs) are too complex, expensive, or dependent on external computing resources. In order to bridge this gap, we introduce an autonomous logistics robot that combines adaptive control and visual perception on a small edge computing platform. The NVIDIA Jetson Nano was equipped with a modified ResNet-18 model that allowed it to concurrently execute three tasks: object-handling zone recognition, obstacle detection, and path tracking. A lightweight rack-and-pinion mechanism enables payload lifting of up to 2 kg without external assistance. Experimental evaluation in semi-structured warehouse settings demonstrated a path tracking accuracy of 92%, obstacle avoidance success of 88%, and object handling success of 90%, with a maximum perception-to-action latency of 150 m. The system maintains stable operation for up to 3 hours on a single charge. Unlike other approaches that focus on single functions or require cloud support, our design integrates navigation, perception, and mechanical handling into a low-power, standalone solution. This highlights its potential as a practical and cost-effective automation platform for SMEs.
{"title":"ResNet-18 based multi-task visual inference and adaptive control for an edge-deployed autonomous robot.","authors":"Sufola Das Chagas Silva Araujo, Goh Kah Ong Michael, Uttam U Deshpande, Sudhindra Deshpande, Manjunath G Avalappa, Yash Amasi, Sumit Patil, Swathi Bhat, Sudarshan Karigoudar","doi":"10.3389/frobt.2025.1680285","DOIUrl":"10.3389/frobt.2025.1680285","url":null,"abstract":"<p><p>Current industrial robots deployed in small and medium-sized businesses (SMEs) are too complex, expensive, or dependent on external computing resources. In order to bridge this gap, we introduce an autonomous logistics robot that combines adaptive control and visual perception on a small edge computing platform. The NVIDIA Jetson Nano was equipped with a modified ResNet-18 model that allowed it to concurrently execute three tasks: object-handling zone recognition, obstacle detection, and path tracking. A lightweight rack-and-pinion mechanism enables payload lifting of up to 2 kg without external assistance. Experimental evaluation in semi-structured warehouse settings demonstrated a path tracking accuracy of 92%, obstacle avoidance success of 88%, and object handling success of 90%, with a maximum perception-to-action latency of 150 m. The system maintains stable operation for up to 3 hours on a single charge. Unlike other approaches that focus on single functions or require cloud support, our design integrates navigation, perception, and mechanical handling into a low-power, standalone solution. This highlights its potential as a practical and cost-effective automation platform for SMEs.</p>","PeriodicalId":47597,"journal":{"name":"Frontiers in Robotics and AI","volume":"12 ","pages":"1680285"},"PeriodicalIF":3.0,"publicationDate":"2025-11-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12624282/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145557698","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-11-03eCollection Date: 2025-01-01DOI: 10.3389/frobt.2025.1681187
Mau Adachi, Masayuki Kakio
As service robots become increasingly integrated into public spaces, effective communication between robots and humans is essential. Elevators, being common shared spaces, present unique challenges and opportunities for such interactions. In this study, we developed a Human-Facility Interaction (HFI) system to facilitate communication between service robots and passengers in elevator environments. The system provided both verbal (voice announcements) and non-verbal (light signals) information to passengers waiting for an elevator alongside a service robot. We installed the system in a hotel and conducted two experiments involving 31 participants to evaluate its impact on passengers' impressions of the elevator and the robot. Our findings revealed that voice-based information significantly improved passengers' impressions and reduced perceived waiting time. However, light-based information had minimal impact on impressions and unexpectedly increased perceived waiting time. These results offer valuable insights for designing future HFI systems to support the integration of service robots in buildings.
{"title":"Human-facility interaction improving people's understanding of service robots and elevators - system design and evaluation.","authors":"Mau Adachi, Masayuki Kakio","doi":"10.3389/frobt.2025.1681187","DOIUrl":"10.3389/frobt.2025.1681187","url":null,"abstract":"<p><p>As service robots become increasingly integrated into public spaces, effective communication between robots and humans is essential. Elevators, being common shared spaces, present unique challenges and opportunities for such interactions. In this study, we developed a Human-Facility Interaction (HFI) system to facilitate communication between service robots and passengers in elevator environments. The system provided both verbal (voice announcements) and non-verbal (light signals) information to passengers waiting for an elevator alongside a service robot. We installed the system in a hotel and conducted two experiments involving 31 participants to evaluate its impact on passengers' impressions of the elevator and the robot. Our findings revealed that voice-based information significantly improved passengers' impressions and reduced perceived waiting time. However, light-based information had minimal impact on impressions and unexpectedly increased perceived waiting time. These results offer valuable insights for designing future HFI systems to support the integration of service robots in buildings.</p>","PeriodicalId":47597,"journal":{"name":"Frontiers in Robotics and AI","volume":"12 ","pages":"1681187"},"PeriodicalIF":3.0,"publicationDate":"2025-11-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12620198/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145551645","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-11-03eCollection Date: 2025-01-01DOI: 10.3389/frobt.2025.1674421
Kamilya Smagulova, Ahmed Elsheikh, Diego A Silva, Mohammed E Fouda, Ahmed M Eltawil
Autonomous driving has the potential to enhance driving comfort and accessibility, reduce accidents, and improve road safety, with vision sensors playing a key role in enabling vehicle autonomy. Among existing sensors, event-based cameras offer advantages such as a high dynamic range, low power consumption, and enhanced motion detection capabilities compared to traditional frame-based cameras. However, their sparse and asynchronous data present unique processing challenges that require specialized algorithms and hardware. While some models originally developed for frame-based inputs have been adapted to handle event data, they often fail to fully exploit the distinct properties of this novel data format, primarily due to its fundamental structural differences. As a result, new algorithms, including neuromorphic, have been developed specifically for event data. Many of these models are still in the early stages and often lack the maturity and accuracy of traditional approaches. This survey paper focuses on end-to-end event-based object detection for autonomous driving, covering key aspects such as sensing and processing hardware designs, datasets, and algorithms, including dense, spiking, and graph-based neural networks, along with relevant encoding and pre-processing techniques. In addition, this work highlights the shortcomings in the evaluation practices to ensure fair and meaningful comparisons across different event data processing approaches and hardware platforms. Within the scope of this survey, system-level throughput was evaluated from raw event data to model output on an RTX 4090 24GB GPU for several state-of-the-art models using the GEN1 and 1MP datasets. The study also includes a discussion and outlines potential directions for future research.
{"title":"Efficient and real-time perception: a survey on end-to-end event-based object detection in autonomous driving.","authors":"Kamilya Smagulova, Ahmed Elsheikh, Diego A Silva, Mohammed E Fouda, Ahmed M Eltawil","doi":"10.3389/frobt.2025.1674421","DOIUrl":"10.3389/frobt.2025.1674421","url":null,"abstract":"<p><p>Autonomous driving has the potential to enhance driving comfort and accessibility, reduce accidents, and improve road safety, with vision sensors playing a key role in enabling vehicle autonomy. Among existing sensors, event-based cameras offer advantages such as a high dynamic range, low power consumption, and enhanced motion detection capabilities compared to traditional frame-based cameras. However, their sparse and asynchronous data present unique processing challenges that require specialized algorithms and hardware. While some models originally developed for frame-based inputs have been adapted to handle event data, they often fail to fully exploit the distinct properties of this novel data format, primarily due to its fundamental structural differences. As a result, new algorithms, including neuromorphic, have been developed specifically for event data. Many of these models are still in the early stages and often lack the maturity and accuracy of traditional approaches. This survey paper focuses on end-to-end event-based object detection for autonomous driving, covering key aspects such as sensing and processing hardware designs, datasets, and algorithms, including dense, spiking, and graph-based neural networks, along with relevant encoding and pre-processing techniques. In addition, this work highlights the shortcomings in the evaluation practices to ensure fair and meaningful comparisons across different event data processing approaches and hardware platforms. Within the scope of this survey, system-level throughput was evaluated from raw event data to model output on an RTX 4090 24GB GPU for several state-of-the-art models using the GEN1 and 1MP datasets. The study also includes a discussion and outlines potential directions for future research.</p>","PeriodicalId":47597,"journal":{"name":"Frontiers in Robotics and AI","volume":"12 ","pages":"1674421"},"PeriodicalIF":3.0,"publicationDate":"2025-11-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12620194/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145551676","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Cleaning PV (photovoltaic) panels is essential for a PV station, as dirt or dust reduces the effective irradiation of solar energy and weakens the efficiency of converting solar energy into free electrons. The inconsistent (cleaning efficacy) and unsafe (summarized voltage and current) manual method is a challenge for a PV station. Therefore, this paper develops a cleaning robot with PV detection, path planning, and action control. Firstly, a lightweight Mobile-VIT (Mobile Vision Transformer) model with a Self-Attention mechanism was used to improve YOLOv8 (You Only Look Once v8), resulting in an accuracy of 91.08% and a processing speed of 215 fps (frames per second). Secondly, an A* and a DWA (Dynamic Window Approach) path planning algorithm were improved. The simulation result shows that the time consumption decreased from 1.19 to 0.66 s and the Turn Number decreased from 23 to 10 p (places). Finally, the robot was evaluated and calibrated in both indoor and outdoor environments. The results showed that the algorithm can successfully clean PV arrays without manual control, with the rate increasing by 23% after its implementation. This study supports the maintenance of PV stations and serves as a reference for technical applications of deep learning, computer vision, and robot navigation.
清洁光伏板对于光伏电站来说是必不可少的,因为污垢或灰尘会降低太阳能的有效辐射,并降低将太阳能转化为自由电子的效率。人工方法的不一致(清洁效果)和不安全(汇总电压和电流)是光伏电站面临的挑战。因此,本文开发了一种具有PV检测、路径规划和动作控制的清洁机器人。首先,采用具有自注意机制的轻量级Mobile- vit (Mobile Vision Transformer)模型对YOLOv8 (You Only Look Once v8)进行改进,使准确率达到91.08%,处理速度达到215 fps(帧/秒)。其次,对A*和DWA (Dynamic Window Approach)路径规划算法进行了改进。仿真结果表明,该算法耗时从1.19 s减少到0.66 s,转数从23位减少到10位。最后,在室内和室外环境下对机器人进行了评估和校准。结果表明,该算法可以在不需要人工控制的情况下成功清洗光伏阵列,实现后的清洗率提高了23%。本研究为光伏电站维护提供支持,为深度学习、计算机视觉、机器人导航等技术应用提供参考。
{"title":"A photovoltaic panel cleaning robot with a lightweight YOLO v8.","authors":"Jidong Luo, Guoyi Wang, Yanjiao Lei, Dong Wang, Yayong Chen, Hongzhou Zhang","doi":"10.3389/frobt.2025.1606774","DOIUrl":"10.3389/frobt.2025.1606774","url":null,"abstract":"<p><p>Cleaning PV (photovoltaic) panels is essential for a PV station, as dirt or dust reduces the effective irradiation of solar energy and weakens the efficiency of converting solar energy into free electrons. The inconsistent (cleaning efficacy) and unsafe (summarized voltage and current) manual method is a challenge for a PV station. Therefore, this paper develops a cleaning robot with PV detection, path planning, and action control. Firstly, a lightweight Mobile-VIT (Mobile Vision Transformer) model with a Self-Attention mechanism was used to improve YOLOv8 (You Only Look Once v8), resulting in an accuracy of 91.08% and a processing speed of 215 fps (frames per second). Secondly, an A* and a DWA (Dynamic Window Approach) path planning algorithm were improved. The simulation result shows that the time consumption decreased from 1.19 to 0.66 s and the Turn Number decreased from 23 to 10 p (places). Finally, the robot was evaluated and calibrated in both indoor and outdoor environments. The results showed that the algorithm can successfully clean PV arrays without manual control, with the rate increasing by 23% after its implementation. This study supports the maintenance of PV stations and serves as a reference for technical applications of deep learning, computer vision, and robot navigation.</p>","PeriodicalId":47597,"journal":{"name":"Frontiers in Robotics and AI","volume":"12 ","pages":"1606774"},"PeriodicalIF":3.0,"publicationDate":"2025-10-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12615241/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145543298","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-10-31eCollection Date: 2025-01-01DOI: 10.3389/frobt.2025.1671336
Sofia Miranda, Carlos Renato Vázquez, Manuel Navarro-Gutiérrez
Energy consumption is a key concern in modern industrial facilities. Power peak is also a relevant feature in industrial energy analysis and managment, since the electrical infrastructure must be implemented to provide not only the total consumed energy, but the power peaks. Collaborative robots are gaining popularity due to its flexible use and convenient set up. In this context, a power and energy consumption study of the popular UR10 collaborative robot of Universal Robots is reported in this work. For this, an experiment was conducted to obtain current consumption data from the UR10 API, when performing movements with different loads and parameters. Next, the dependency of the trajectory programming parameters on the power peak, total consumed energy, and time spent per trajectory was analyzed. The results show that the higher the speed limit and acceleration limit, the lower the total energy consumed per trajectory, but the higher the power peak. This behavior represents a trade-off: reducing the consumed energy involves increasing the peak power. Based on the captured data, artificial neural network models were trained to predict the power peak and the total energy consumed by the robot when performing a movement under certain parameters. These models were later used by a genetic optimization algorithm to obtain the best parameters for a given target position, providing the most efficient performance while fulfilling a power peak bound.
{"title":"Energy consumption analysis and optimization in collaborative robots.","authors":"Sofia Miranda, Carlos Renato Vázquez, Manuel Navarro-Gutiérrez","doi":"10.3389/frobt.2025.1671336","DOIUrl":"10.3389/frobt.2025.1671336","url":null,"abstract":"<p><p>Energy consumption is a key concern in modern industrial facilities. Power peak is also a relevant feature in industrial energy analysis and managment, since the electrical infrastructure must be implemented to provide not only the total consumed energy, but the power peaks. Collaborative robots are gaining popularity due to its flexible use and convenient set up. In this context, a power and energy consumption study of the popular UR10 collaborative robot of Universal Robots is reported in this work. For this, an experiment was conducted to obtain current consumption data from the UR10 API, when performing movements with different loads and parameters. Next, the dependency of the trajectory programming parameters on the power peak, total consumed energy, and time spent per trajectory was analyzed. The results show that the higher the speed limit and acceleration limit, the lower the total energy consumed per trajectory, but the higher the power peak. This behavior represents a trade-off: reducing the consumed energy involves increasing the peak power. Based on the captured data, artificial neural network models were trained to predict the power peak and the total energy consumed by the robot when performing a movement under certain parameters. These models were later used by a genetic optimization algorithm to obtain the best parameters for a given target position, providing the most efficient performance while fulfilling a power peak bound.</p>","PeriodicalId":47597,"journal":{"name":"Frontiers in Robotics and AI","volume":"12 ","pages":"1671336"},"PeriodicalIF":3.0,"publicationDate":"2025-10-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12616253/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145543305","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-10-30eCollection Date: 2025-01-01DOI: 10.3389/frobt.2025.1426676
Samantha Chapin, William Chapin, Erik Komendera
This article proposes a method that uses information about modules and desired assembly locations within a large truss structure to create a semantic and fiducial aided graph simultaneous localization and mapping (SF-GraphSLAM) algorithm that is better tailored for use during robotic in-space assembly and servicing operations. This is achieved by first reducing the number of modules using a mixed assembly method vs. a strut-by-strut method. Then, each module is correlated to a visual tag (in this article, an AprilTag) to reduce the number of elements being observed further from the number of sub-struts in that module to a single AprilTag marker. Two tags are required to ensure proper deployment of most deployable modules. Subsequently, we are able to use semantic information about the desired transformation matrix between any two adjacent module AprilTags within the desired assembly structure. For our experimentation, we expanded a factor graph smoothing and mapping model and added the semantic information, looking at the smaller number of landmark AprilTags, with a camera representing the robot for simplicity. The mathematical approach to arrive at this new method is included in this article, as are simulations to test it against the state of the art (SOA) using no structural knowledge. Overall, this research contributes to the SOA for both general SLAM work and, more specifically, to the underdeveloped field of SLAM for in-space assembly and servicing of large truss structures. It is critical to ensure that as a robot is assembling the modules, each module is within the desired tolerances to ensure the final structure is within the design requirements. Being able to build a virtual twin of the truss structure as it is being assembled is a key tent pole in achieving large space structures.
{"title":"Semantic and fiducial-aided graph simultaneous localization and mapping (SF-GraphSLAM) for robotic in-space assembly and servicing of large truss structures.","authors":"Samantha Chapin, William Chapin, Erik Komendera","doi":"10.3389/frobt.2025.1426676","DOIUrl":"10.3389/frobt.2025.1426676","url":null,"abstract":"<p><p>This article proposes a method that uses information about modules and desired assembly locations within a large truss structure to create a semantic and fiducial aided graph simultaneous localization and mapping (SF-GraphSLAM) algorithm that is better tailored for use during robotic in-space assembly and servicing operations. This is achieved by first reducing the number of modules using a mixed assembly method vs. a strut-by-strut method. Then, each module is correlated to a visual tag (in this article, an AprilTag) to reduce the number of elements being observed further from the number of sub-struts in that module to a single AprilTag marker. Two tags are required to ensure proper deployment of most deployable modules. Subsequently, we are able to use semantic information about the desired transformation matrix between any two adjacent module AprilTags within the desired assembly structure. For our experimentation, we expanded a factor graph smoothing and mapping model and added the semantic information, looking at the smaller number of landmark AprilTags, with a camera representing the robot for simplicity. The mathematical approach to arrive at this new method is included in this article, as are simulations to test it against the state of the art (SOA) using no structural knowledge. Overall, this research contributes to the SOA for both general SLAM work and, more specifically, to the underdeveloped field of SLAM for in-space assembly and servicing of large truss structures. It is critical to ensure that as a robot is assembling the modules, each module is within the desired tolerances to ensure the final structure is within the design requirements. Being able to build a virtual twin of the truss structure as it is being assembled is a key tent pole in achieving large space structures.</p>","PeriodicalId":47597,"journal":{"name":"Frontiers in Robotics and AI","volume":"12 ","pages":"1426676"},"PeriodicalIF":3.0,"publicationDate":"2025-10-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12613163/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145543228","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-10-30eCollection Date: 2025-01-01DOI: 10.3389/frobt.2025.1614444
Mary Ellen Foster, Jennifer N Stinson, Lauren Harris, Kate Kyuri Kim, Sasha Litwin, Patricia Candelaria, Summer Hudson, Julie Leung, Ronald P A Petrick, Alan Lindsay, Andrés Ramírez-Duque, David Harris Smith, Frauke Zeller, Samina Ali
Introduction: Children undergoing medical procedures in paediatric Emergency Departments (EDs) often experience significant pain and distress. Socially Assistive Robots (SARs) offer a promising avenue for delivering distraction and emotional support in these high-pressure environments. This study presents the design, development, and formative evaluation of an AI-enhanced SAR to support children during intravenous insertion (IVI) procedures.
Methods: The robot system was developed through a participatory design process involving healthcare professionals, patients, caregivers, and interdisciplinary research teams. The SAR was designed to autonomously adapt its behaviour to the child's affective state using AI planning and social signal processing. A two-cycle usability study was conducted across two Canadian paediatric EDs, involving 25 children and their caregivers. Feedback was collected through observations, interviews, and system logs.
Results: The SAR was successfully integrated into clinical workflows, with positive responses from children, caregivers, and healthcare providers. Usability testing identified key technical and interaction challenges, which were addressed through iterative refinement. The final system demonstrated robust performance and was deemed ready for a formal randomised controlled trial.
Discussion: This work highlights the importance of co-design, operator control, and environmental adaptability in deploying SARs in clinical settings. Lessons learned from the development and deployment process informed six concrete design guidelines for future SAR implementations in healthcare.
{"title":"Bringing a socially assistive robot to the paediatric emergency department: design, development, and usability testing.","authors":"Mary Ellen Foster, Jennifer N Stinson, Lauren Harris, Kate Kyuri Kim, Sasha Litwin, Patricia Candelaria, Summer Hudson, Julie Leung, Ronald P A Petrick, Alan Lindsay, Andrés Ramírez-Duque, David Harris Smith, Frauke Zeller, Samina Ali","doi":"10.3389/frobt.2025.1614444","DOIUrl":"10.3389/frobt.2025.1614444","url":null,"abstract":"<p><strong>Introduction: </strong>Children undergoing medical procedures in paediatric Emergency Departments (EDs) often experience significant pain and distress. Socially Assistive Robots (SARs) offer a promising avenue for delivering distraction and emotional support in these high-pressure environments. This study presents the design, development, and formative evaluation of an AI-enhanced SAR to support children during intravenous insertion (IVI) procedures.</p><p><strong>Methods: </strong>The robot system was developed through a participatory design process involving healthcare professionals, patients, caregivers, and interdisciplinary research teams. The SAR was designed to autonomously adapt its behaviour to the child's affective state using AI planning and social signal processing. A two-cycle usability study was conducted across two Canadian paediatric EDs, involving 25 children and their caregivers. Feedback was collected through observations, interviews, and system logs.</p><p><strong>Results: </strong>The SAR was successfully integrated into clinical workflows, with positive responses from children, caregivers, and healthcare providers. Usability testing identified key technical and interaction challenges, which were addressed through iterative refinement. The final system demonstrated robust performance and was deemed ready for a formal randomised controlled trial.</p><p><strong>Discussion: </strong>This work highlights the importance of co-design, operator control, and environmental adaptability in deploying SARs in clinical settings. Lessons learned from the development and deployment process informed six concrete design guidelines for future SAR implementations in healthcare.</p>","PeriodicalId":47597,"journal":{"name":"Frontiers in Robotics and AI","volume":"12 ","pages":"1614444"},"PeriodicalIF":3.0,"publicationDate":"2025-10-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12612628/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145543235","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-10-30eCollection Date: 2025-01-01DOI: 10.3389/frobt.2025.1589025
Davide Picchi, Sigrid Brell-Çokcan
Mini cranes play a pivotal role in construction due to their versatility across numerous scenarios. Recent advancements in Reinforcement Learning (RL) have enabled agents to operate cranes in virtual environments for predetermined tasks, paving the way for future real-world deployment. Traditionally, most RL agents use a squashed Gaussian distribution to select actions. In this study, we investigate a mini-crane scenario that could potentially be fully automated by AI and explore replacing the Gaussian distribution with the Kumaraswamy distribution, a close relative of the Beta distribution, for action stochastic selection. Our results indicate that the Kumaraswamy distribution offers computational advantages while maintaining robust performance, making it an attractive alternative for RL applications in continuous control applications.
{"title":"Exploiting the Kumaraswamy distribution in a reinforcement learning context.","authors":"Davide Picchi, Sigrid Brell-Çokcan","doi":"10.3389/frobt.2025.1589025","DOIUrl":"10.3389/frobt.2025.1589025","url":null,"abstract":"<p><p>Mini cranes play a pivotal role in construction due to their versatility across numerous scenarios. Recent advancements in Reinforcement Learning (RL) have enabled agents to operate cranes in virtual environments for predetermined tasks, paving the way for future real-world deployment. Traditionally, most RL agents use a squashed Gaussian distribution to select actions. In this study, we investigate a mini-crane scenario that could potentially be fully automated by AI and explore replacing the Gaussian distribution with the Kumaraswamy distribution, a close relative of the Beta distribution, for action stochastic selection. Our results indicate that the Kumaraswamy distribution offers computational advantages while maintaining robust performance, making it an attractive alternative for RL applications in continuous control applications.</p>","PeriodicalId":47597,"journal":{"name":"Frontiers in Robotics and AI","volume":"12 ","pages":"1589025"},"PeriodicalIF":3.0,"publicationDate":"2025-10-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12611641/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145543215","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-10-30eCollection Date: 2025-01-01DOI: 10.3389/frobt.2025.1660691
Junya Yamamoto, Kenji Tahara, Takahiro Wada
In response to the growing need for flexibility in handling complex tasks, research on human-robot collaboration (HRC) has garnered considerable attention. Recent studies on HRC have achieved smooth handover tasks between humans and robots by adaptively responding to human states. Collaboration was further improved by conveying the state of the robot to humans via robotic interactive motion cues. However, in scenarios such as collaborative assembly tasks that require precise positioning, methods relying on motion or forces caused by interactions through the shared object compromise both task accuracy and smoothness, and are therefore not directly applicable. To address this, the present study proposes a method to convey the stiffness of the robot to a human arm during collaborative human-robot assembly tasks in a manner that does not affect the shared object or task, aiming to enhance efficiency and reduce human workload. Sixteen participants performed a collaborative assembly task with a robot, which involved unscrewing, repositioning, and reattaching a part while the robot held and adjusted the position of the part. The experiment examined the effectiveness of the proposed method, in which the robot's stiffness was communicated to a participant's forearm. The independent variable, tested within-subjects, was the stiffness presentation method, with three levels: without the proposed method (no presentation) and with the proposed method (real-time and predictive presentations). The results demonstrated that the proposed method enhanced task efficiency by shortening task completion time, which was associated with lower subjective workload scores.
{"title":"Effect of presenting robot hand stiffness to human arm on human-robot collaborative assembly tasks.","authors":"Junya Yamamoto, Kenji Tahara, Takahiro Wada","doi":"10.3389/frobt.2025.1660691","DOIUrl":"10.3389/frobt.2025.1660691","url":null,"abstract":"<p><p>In response to the growing need for flexibility in handling complex tasks, research on human-robot collaboration (HRC) has garnered considerable attention. Recent studies on HRC have achieved smooth handover tasks between humans and robots by adaptively responding to human states. Collaboration was further improved by conveying the state of the robot to humans via robotic interactive motion cues. However, in scenarios such as collaborative assembly tasks that require precise positioning, methods relying on motion or forces caused by interactions through the shared object compromise both task accuracy and smoothness, and are therefore not directly applicable. To address this, the present study proposes a method to convey the stiffness of the robot to a human arm during collaborative human-robot assembly tasks in a manner that does not affect the shared object or task, aiming to enhance efficiency and reduce human workload. Sixteen participants performed a collaborative assembly task with a robot, which involved unscrewing, repositioning, and reattaching a part while the robot held and adjusted the position of the part. The experiment examined the effectiveness of the proposed method, in which the robot's stiffness was communicated to a participant's forearm. The independent variable, tested within-subjects, was the stiffness presentation method, with three levels: without the proposed method (no presentation) and with the proposed method (real-time and predictive presentations). The results demonstrated that the proposed method enhanced task efficiency by shortening task completion time, which was associated with lower subjective workload scores.</p>","PeriodicalId":47597,"journal":{"name":"Frontiers in Robotics and AI","volume":"12 ","pages":"1660691"},"PeriodicalIF":3.0,"publicationDate":"2025-10-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12611644/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145543261","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-10-30eCollection Date: 2025-01-01DOI: 10.3389/frobt.2025.1594529
Silvia Filogna, Giovanni Arras, Tommaso Turchi, Giuseppe Prencipe, Elena Beani, Clara Bombonato, Francesca Fedeli, Gemma D'Alessandro, Antea Scrocco, Giuseppina Sgandurra
Despite the growing interest in Artificial Intelligence (AI) for pediatric rehabilitation, family engagement in the technologies design remains limited. Understanding how AI-driven tools align with family needs, caregiving routines, and ethical concerns is crucial for their successful adoption. In this study, we actively involved nine families of children with Cerebral Palsy (CP) in an online participatory design workshop, underscoring both the feasibility and the need of integrating family's perspectives into AI development. Families enthusiastically participated, not only sharing insights but also appreciating the opportunity to contribute to shaping future technologies. Their active engagement challenges the assumption that co-design with families is complex or impractical, highlighting how structured yet flexible methodologies can make such crucial initiatives highly effective. The online format further facilitated participation, allowing families to join the discussion and ensuring a diverse range of perspectives. The workshop's key findings reveal three core priorities for families: 1. AI should adapt to daily caregiving routines rather than impose rigid structures; 2. digital tools should enhance communication and collaboration between families and clinicians, rather than replace human interaction; and 3. AI-driven systems could empower children's autonomy while maintaining parental oversight. Additionally, families raised critical concerns about data privacy, transparency, and the need to preserve empathy in AI-mediated care. Our findings reinforce the urgent need to shift toward family-centered AI design, moving beyond purely technological solutions toward ethically responsible, inclusive innovations. This research not only demonstrates the possibility and success of engaging families in co-design processes but also provides a model for future AI development that genuinely reflects the lived experiences of children and caregivers.
{"title":"Pathways to family-centered healthcare: co-designing AI solutions with families in pediatric rehabilitation.","authors":"Silvia Filogna, Giovanni Arras, Tommaso Turchi, Giuseppe Prencipe, Elena Beani, Clara Bombonato, Francesca Fedeli, Gemma D'Alessandro, Antea Scrocco, Giuseppina Sgandurra","doi":"10.3389/frobt.2025.1594529","DOIUrl":"10.3389/frobt.2025.1594529","url":null,"abstract":"<p><p>Despite the growing interest in Artificial Intelligence (AI) for pediatric rehabilitation, family engagement in the technologies design remains limited. Understanding how AI-driven tools align with family needs, caregiving routines, and ethical concerns is crucial for their successful adoption. In this study, we actively involved nine families of children with Cerebral Palsy (CP) in an online participatory design workshop, underscoring both the feasibility and the need of integrating family's perspectives into AI development. Families enthusiastically participated, not only sharing insights but also appreciating the opportunity to contribute to shaping future technologies. Their active engagement challenges the assumption that co-design with families is complex or impractical, highlighting how structured yet flexible methodologies can make such crucial initiatives highly effective. The online format further facilitated participation, allowing families to join the discussion and ensuring a diverse range of perspectives. The workshop's key findings reveal three core priorities for families: 1. AI should adapt to daily caregiving routines rather than impose rigid structures; 2. digital tools should enhance communication and collaboration between families and clinicians, rather than replace human interaction; and 3. AI-driven systems could empower children's autonomy while maintaining parental oversight. Additionally, families raised critical concerns about data privacy, transparency, and the need to preserve empathy in AI-mediated care. Our findings reinforce the urgent need to shift toward family-centered AI design, moving beyond purely technological solutions toward ethically responsible, inclusive innovations. This research not only demonstrates the possibility and success of engaging families in co-design processes but also provides a model for future AI development that genuinely reflects the lived experiences of children and caregivers.</p>","PeriodicalId":47597,"journal":{"name":"Frontiers in Robotics and AI","volume":"12 ","pages":"1594529"},"PeriodicalIF":3.0,"publicationDate":"2025-10-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12611681/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145543293","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}