Pub Date : 2025-01-31eCollection Date: 2025-01-01DOI: 10.3389/frobt.2025.1503622
Wenchao Liu, Xin Xin, Chenyu Zheng
The concept of anthropomorphism is crucial in enhancing interactions between service robots and humans, serving as a key consideration in the design of these robots. Nevertheless, the specific mechanisms by which the anthropomorphic traits of service robots influence customer behavioral responses remain inadequately understood. Furthermore, the incorporation of anthropomorphic robotic technology into customer service operational strategies presents a significant challenge for businesses. To explore the underlying mechanisms through which the anthropomorphic characteristics of service robots impact customer acceptance, this study conducted a series of six experiments to empirically test the proposed hypotheses. The empirical findings indicate notable differences in customer switching behaviors and selection quantity metrics, which can be linked to service contexts characterized by varying degrees of behavioral anthropomorphism. Additionally, social presence has been identified as a mediating variable that affects the relationship between the anthropomorphism of service robot behavior and its influence on customer variety-seeking behavior. The situational context of customer decision-making is also found to moderate the relationship between social presence and variety-seeking behavior. Consequently, it is recommended that service organizations implement service robots with diverse anthropomorphic features to enhance customer acquisition, cultivate loyalty, and improve overall marketing effectiveness.
{"title":"The effect of behaviorally anthropomorphic service robots on customers' variety-seeking behavior: an analytical examination of social presence and decision-making context.","authors":"Wenchao Liu, Xin Xin, Chenyu Zheng","doi":"10.3389/frobt.2025.1503622","DOIUrl":"10.3389/frobt.2025.1503622","url":null,"abstract":"<p><p>The concept of anthropomorphism is crucial in enhancing interactions between service robots and humans, serving as a key consideration in the design of these robots. Nevertheless, the specific mechanisms by which the anthropomorphic traits of service robots influence customer behavioral responses remain inadequately understood. Furthermore, the incorporation of anthropomorphic robotic technology into customer service operational strategies presents a significant challenge for businesses. To explore the underlying mechanisms through which the anthropomorphic characteristics of service robots impact customer acceptance, this study conducted a series of six experiments to empirically test the proposed hypotheses. The empirical findings indicate notable differences in customer switching behaviors and selection quantity metrics, which can be linked to service contexts characterized by varying degrees of behavioral anthropomorphism. Additionally, social presence has been identified as a mediating variable that affects the relationship between the anthropomorphism of service robot behavior and its influence on customer variety-seeking behavior. The situational context of customer decision-making is also found to moderate the relationship between social presence and variety-seeking behavior. Consequently, it is recommended that service organizations implement service robots with diverse anthropomorphic features to enhance customer acquisition, cultivate loyalty, and improve overall marketing effectiveness.</p>","PeriodicalId":47597,"journal":{"name":"Frontiers in Robotics and AI","volume":"12 ","pages":"1503622"},"PeriodicalIF":2.9,"publicationDate":"2025-01-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11825330/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143434102","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-01-30eCollection Date: 2024-01-01DOI: 10.3389/frobt.2024.1442813
Wilhelm J Marais, Oscar Pizarro, Stefan B Williams
This paper presents methods for finding optimal configurations and actuator forces/torques to maximise contact wrenches in a desired direction for underwater vehicle manipulator systems (UVMS). The wrench maximisation problem is formulated as a bi-level optimisation problem, with upper-level variables in a low-dimensional parameterised redundancy space, and a linear lower-level problem. We additionally consider the cases of one or more manipulators with multiple contact forces, maximising the wrench capability while tracking a trajectory and generating large wrench impulses using dynamic motions. The specific cases of maximising force to lift a heavy load and maximising torque during a valve-turning operation are considered. Extensive experimental results are presented using a 6 degree of freedom (DOF) underwater robotic platform equipped with a 4DOF manipulator and show significant increases in the wrench capability compared to existing methods for mobile manipulators.
{"title":"Maximising the wrench capability of mobile manipulators with experiments on a UVMS.","authors":"Wilhelm J Marais, Oscar Pizarro, Stefan B Williams","doi":"10.3389/frobt.2024.1442813","DOIUrl":"https://doi.org/10.3389/frobt.2024.1442813","url":null,"abstract":"<p><p>This paper presents methods for finding optimal configurations and actuator forces/torques to maximise contact wrenches in a desired direction for underwater vehicle manipulator systems (UVMS). The wrench maximisation problem is formulated as a bi-level optimisation problem, with upper-level variables in a low-dimensional parameterised redundancy space, and a linear lower-level problem. We additionally consider the cases of one or more manipulators with multiple contact forces, maximising the wrench capability while tracking a trajectory and generating large wrench impulses using dynamic motions. The specific cases of maximising force to lift a heavy load and maximising torque during a valve-turning operation are considered. Extensive experimental results are presented using a 6 degree of freedom (DOF) underwater robotic platform equipped with a 4DOF manipulator and show significant increases in the wrench capability compared to existing methods for mobile manipulators.</p>","PeriodicalId":47597,"journal":{"name":"Frontiers in Robotics and AI","volume":"11 ","pages":"1442813"},"PeriodicalIF":2.9,"publicationDate":"2025-01-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11822218/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143415495","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-01-29eCollection Date: 2024-01-01DOI: 10.3389/frobt.2024.1363150
Sven Schneider, Nico Hochgeschwender, Herman Bruyninckx
This article introduces a model-based design, implementation, deployment, and execution methodology, with tools supporting the systematic composition of algorithms from generic and domain-specific computational building blocks that prevent code duplication and enable robots to adapt their software themselves. The envisaged algorithms are numerical solvers based on graph structures. In this article, we focus on kinematics and dynamics algorithms, but examples such as message passing on probabilistic networks and factor graphs or cascade control diagrams fall under the same pattern. The tools rely on mature standards from the Semantic Web. They first synthesize algorithms symbolically, from which they then generate efficient code. The use case is an overactuated mobile robot with two redundant arms.
{"title":"Semantic composition of robotic solver algorithms on graph structures.","authors":"Sven Schneider, Nico Hochgeschwender, Herman Bruyninckx","doi":"10.3389/frobt.2024.1363150","DOIUrl":"10.3389/frobt.2024.1363150","url":null,"abstract":"<p><p>This article introduces a model-based design, implementation, deployment, and execution methodology, with tools supporting the systematic composition of algorithms from generic and domain-specific computational building blocks that prevent code duplication and enable robots to adapt their software themselves. The envisaged algorithms are numerical solvers based on graph structures. In this article, we focus on kinematics and dynamics algorithms, but examples such as message passing on probabilistic networks and factor graphs or cascade control diagrams fall under the same pattern. The tools rely on mature standards from the Semantic Web. They first synthesize algorithms symbolically, from which they then generate efficient code. The use case is an overactuated mobile robot with two redundant arms.</p>","PeriodicalId":47597,"journal":{"name":"Frontiers in Robotics and AI","volume":"11 ","pages":"1363150"},"PeriodicalIF":2.9,"publicationDate":"2025-01-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11813742/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143411261","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-01-28eCollection Date: 2024-01-01DOI: 10.3389/frobt.2024.1418677
Fabrizio Nunnari, Dimitra Tsovaltzi, Matteo Lavit Nicora, Sebastian Beyrodt, Pooja Prajod, Lara Chehayeb, Ingrid Brdar, Antonella Delle Fave, Luca Negri, Elisabeth André, Patrick Gebhard, Matteo Malosio
This article presents the development of a socially interactive industrial robot. An Avatar is used to embody a cobot for collaborative industrial assembly tasks. The embodied covatar (cobot plus its avatar) is introduced to support Flow experiences through co-regulation, interactive emotion regulation guidance. A real-time continuous emotional modeling method and an aligned transparent behavioral model, BASSF (Boredom, Anxiety, Self-efficacy, Self-compassion, Flow) is developed. The BASSF model anticipates and co-regulates counterproductive emotional experiences of operators working under stress with cobots on tedious industrial tasks. The targeted Flow experience is represented in the three-dimensional Pleasure, Arousal, and Dominance (PAD) space. We present how, despite their noisy nature, PAD signals can be used to drive the BASSF model with its theory-based interventions. The empirical results and analysis provides empirical support for the theoretically defined model, and clearly points to the need for data pre-filtering and per-user calibration. The proposed post-processing method helps quantify the parameters needed to control the frequency of intervention of the agent; still leaving the experimenter with a run-time adjustable global control of its sensitivity. A controlled empirical study (Study 1, N = 20), tested the model's main theoretical assumptions about Flow, Dominance, Self-Efficacy, and boredom, to legitimate its implementation in this context. Participants worked on a task for an hour, assembling pieces in collaboration with the covatar. After the task, participants completed questionnaires on Flow, their affective experience, and Self-Efficacy, and they were interviewed to understand their emotions and regulation during the task. The results from Study 1 suggest that the Dominance dimension plays a vital role in task-related settings as it predicts the participants' Self-Efficacy and Flow. However, the relationship between Flow, pleasure, and arousal requires further investigation. Qualitative interview analysis revealed that participants regulated negative emotions, like boredom, also without support, but some strategies could negatively impact wellbeing and productivity, which aligns with theory. Additional results from a first evaluation of the overall system (Study 2, N = 12) align with these findings and provide support for the use of socially interactive industrial robots to support wellbeing, job satisfaction, and involvement, while reducing unproductive emotional experiences and their regulation.
{"title":"Socially interactive industrial robots: a PAD model of flow for emotional co-regulation.","authors":"Fabrizio Nunnari, Dimitra Tsovaltzi, Matteo Lavit Nicora, Sebastian Beyrodt, Pooja Prajod, Lara Chehayeb, Ingrid Brdar, Antonella Delle Fave, Luca Negri, Elisabeth André, Patrick Gebhard, Matteo Malosio","doi":"10.3389/frobt.2024.1418677","DOIUrl":"10.3389/frobt.2024.1418677","url":null,"abstract":"<p><p>This article presents the development of a socially interactive industrial robot. An Avatar is used to embody a cobot for collaborative industrial assembly tasks. The embodied covatar (cobot plus its avatar) is introduced to support Flow experiences through co-regulation, interactive emotion regulation guidance. A real-time continuous emotional modeling method and an aligned transparent behavioral model, BASSF (Boredom, Anxiety, Self-efficacy, Self-compassion, Flow) is developed. The BASSF model anticipates and co-regulates counterproductive emotional experiences of operators working under stress with cobots on tedious industrial tasks. The targeted Flow experience is represented in the three-dimensional Pleasure, Arousal, and Dominance (PAD) space. We present how, despite their noisy nature, PAD signals can be used to drive the BASSF model with its theory-based interventions. The empirical results and analysis provides empirical support for the theoretically defined model, and clearly points to the need for data pre-filtering and per-user calibration. The proposed post-processing method helps quantify the parameters needed to control the frequency of intervention of the agent; still leaving the experimenter with a run-time adjustable global control of its sensitivity. A controlled empirical study (Study 1, N = 20), tested the model's main theoretical assumptions about Flow, Dominance, Self-Efficacy, and boredom, to legitimate its implementation in this context. Participants worked on a task for an hour, assembling pieces in collaboration with the covatar. After the task, participants completed questionnaires on Flow, their affective experience, and Self-Efficacy, and they were interviewed to understand their emotions and regulation during the task. The results from Study 1 suggest that the Dominance dimension plays a vital role in task-related settings as it predicts the participants' Self-Efficacy and Flow. However, the relationship between Flow, pleasure, and arousal requires further investigation. Qualitative interview analysis revealed that participants regulated negative emotions, like boredom, also without support, but some strategies could negatively impact wellbeing and productivity, which aligns with theory. Additional results from a first evaluation of the overall system (Study 2, N = 12) align with these findings and provide support for the use of socially interactive industrial robots to support wellbeing, job satisfaction, and involvement, while reducing unproductive emotional experiences and their regulation.</p>","PeriodicalId":47597,"journal":{"name":"Frontiers in Robotics and AI","volume":"11 ","pages":"1418677"},"PeriodicalIF":2.9,"publicationDate":"2025-01-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11811909/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143400411","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Editorial: Latest trends in bio-inspired medical robotics: structural design, manufacturing, sensing, actuation and control.","authors":"Yilun Sun, Houde Dai, Shuang Song, Angela Faragasso, Sara-Adela Abad Guaman","doi":"10.3389/frobt.2025.1544097","DOIUrl":"10.3389/frobt.2025.1544097","url":null,"abstract":"","PeriodicalId":47597,"journal":{"name":"Frontiers in Robotics and AI","volume":"12 ","pages":"1544097"},"PeriodicalIF":2.9,"publicationDate":"2025-01-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11810720/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143400413","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-01-27eCollection Date: 2025-01-01DOI: 10.3389/frobt.2025.1506290
F Adetunji, A Karukayil, P Samant, S Shabana, F Varghese, U Upadhyay, R A Yadav, A Partridge, E Pendleton, R Plant, Y R Petillot, M Koskinopoulou
Introduction: This paper addresses the challenges of vision-based manipulation for autonomous cutting and unpacking of transparent plastic bags in industrial setups, contributing to the Industry 4.0 paradigm. Industry 4.0, emphasizing data-driven processes, connectivity, and robotics, enhances accessibility and sustainability across the value chain. Integrating autonomous systems, including collaborative robots (cobots), into industrial workflows is crucial for improving efficiency and safety.
Methods: The proposed system employs advanced Machine Learning algorithms, particularly Convolutional Neural Networks (CNNs), for identifying transparent plastic bags under diverse lighting and background conditions. Tracking algorithms and depth-sensing technologies are integrated to enable 3D spatial awareness during pick-and-place operations. The system incorporates vacuum gripping technology with compliance control for optimal grasping and manipulation points, using a Franka Emika robot arm.
Results: The system successfully demonstrates its capability to automate the unpacking and cutting of transparent plastic bags for an 8-stack bulk-loader. Rigorous lab testing showed high accuracy in bag detection and manipulation under varying environmental conditions, as well as reliable performance in handling and processing tasks. The approach effectively addressed challenges related to transparency, plastic bag manipulation and industrial automation.
Discussion: The results indicate that the proposed solution is highly effective for industrial applications requiring precision and adaptability, aligning with the principles of Industry 4.0. By combining advanced vision algorithms, depth sensing, and compliance control, the system offers a robust method for automating challenging tasks. The integration of cobots into such workflows demonstrates significant potential for enhancing efficiency, safety, and sustainability in industrial settings.
{"title":"Vision-based manipulation of transparent plastic bags in industrial setups.","authors":"F Adetunji, A Karukayil, P Samant, S Shabana, F Varghese, U Upadhyay, R A Yadav, A Partridge, E Pendleton, R Plant, Y R Petillot, M Koskinopoulou","doi":"10.3389/frobt.2025.1506290","DOIUrl":"10.3389/frobt.2025.1506290","url":null,"abstract":"<p><strong>Introduction: </strong>This paper addresses the challenges of vision-based manipulation for autonomous cutting and unpacking of transparent plastic bags in industrial setups, contributing to the Industry 4.0 paradigm. Industry 4.0, emphasizing data-driven processes, connectivity, and robotics, enhances accessibility and sustainability across the value chain. Integrating autonomous systems, including collaborative robots (cobots), into industrial workflows is crucial for improving efficiency and safety.</p><p><strong>Methods: </strong>The proposed system employs advanced Machine Learning algorithms, particularly Convolutional Neural Networks (CNNs), for identifying transparent plastic bags under diverse lighting and background conditions. Tracking algorithms and depth-sensing technologies are integrated to enable 3D spatial awareness during pick-and-place operations. The system incorporates vacuum gripping technology with compliance control for optimal grasping and manipulation points, using a Franka Emika robot arm.</p><p><strong>Results: </strong>The system successfully demonstrates its capability to automate the unpacking and cutting of transparent plastic bags for an 8-stack bulk-loader. Rigorous lab testing showed high accuracy in bag detection and manipulation under varying environmental conditions, as well as reliable performance in handling and processing tasks. The approach effectively addressed challenges related to transparency, plastic bag manipulation and industrial automation.</p><p><strong>Discussion: </strong>The results indicate that the proposed solution is highly effective for industrial applications requiring precision and adaptability, aligning with the principles of Industry 4.0. By combining advanced vision algorithms, depth sensing, and compliance control, the system offers a robust method for automating challenging tasks. The integration of cobots into such workflows demonstrates significant potential for enhancing efficiency, safety, and sustainability in industrial settings.</p>","PeriodicalId":47597,"journal":{"name":"Frontiers in Robotics and AI","volume":"12 ","pages":"1506290"},"PeriodicalIF":2.9,"publicationDate":"2025-01-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11807811/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143400414","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-01-21eCollection Date: 2024-01-01DOI: 10.3389/frobt.2024.1470886
Stefano Nolfi, Paolo Pagliuca
The usage of broad sets of training data is paramount to evolve adaptive agents. In this respect, competitive co-evolution is a widespread technique in which the coexistence of different learning agents fosters adaptation, which in turn makes agents experience continuously varying environmental conditions. However, a major pitfall is related to the emergence of endless limit cycles where agents discover, forget and rediscover similar strategies during evolution. In this work, we investigate the use of competitive co-evolution for synthesizing progressively better solutions. Specifically, we introduce a set of methods to measure historical and global progress. We discuss the factors that facilitate genuine progress. Finally, we compare the efficacy of four qualitatively different algorithms, including two newly introduced methods. The selected algorithms promote genuine progress by creating an archive of opponents used to evaluate evolving individuals, generating archives that include high-performing and well-differentiated opponents, identifying and discarding variations that lead to local progress only (i.e., progress against the opponents experienced and retrogressing against others). The results obtained in a predator-prey scenario, commonly used to study competitive evolution, demonstrate that all the considered methods lead to global progress in the long term. However, the rate of progress and the ratio of progress versus retrogressions vary significantly among algorithms. In particular, our outcomes indicate that the Generalist method introduced in this work outperforms the other three considered methods and represents the only algorithm capable of producing global progress during evolution.
{"title":"Global progress in competitive co-evolution: a systematic comparison of alternative methods.","authors":"Stefano Nolfi, Paolo Pagliuca","doi":"10.3389/frobt.2024.1470886","DOIUrl":"10.3389/frobt.2024.1470886","url":null,"abstract":"<p><p>The usage of broad sets of training data is paramount to evolve adaptive agents. In this respect, competitive co-evolution is a widespread technique in which the coexistence of different learning agents fosters adaptation, which in turn makes agents experience continuously varying environmental conditions. However, a major pitfall is related to the emergence of endless limit cycles where agents discover, forget and rediscover similar strategies during evolution. In this work, we investigate the use of competitive co-evolution for synthesizing progressively better solutions. Specifically, we introduce a set of methods to measure historical and global progress. We discuss the factors that facilitate genuine progress. Finally, we compare the efficacy of four qualitatively different algorithms, including two newly introduced methods. The selected algorithms promote genuine progress by creating an archive of opponents used to evaluate evolving individuals, generating archives that include high-performing and well-differentiated opponents, identifying and discarding variations that lead to local progress only (i.e., progress against the opponents experienced and retrogressing against others). The results obtained in a predator-prey scenario, commonly used to study competitive evolution, demonstrate that all the considered methods lead to global progress in the long term. However, the rate of progress and the ratio of progress versus retrogressions vary significantly among algorithms. In particular, our outcomes indicate that the Generalist method introduced in this work outperforms the other three considered methods and represents the only algorithm capable of producing global progress during evolution.</p>","PeriodicalId":47597,"journal":{"name":"Frontiers in Robotics and AI","volume":"11 ","pages":"1470886"},"PeriodicalIF":2.9,"publicationDate":"2025-01-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11791907/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143190915","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-01-20eCollection Date: 2024-01-01DOI: 10.3389/frobt.2024.1426282
Gal A Kaminka, Yinon Douchan
Introduction: Inspired by natural phenomena, generations of researchers have been investigating how a swarm of robots can act coherently and purposefully, when individual robots can only sense and communicate with nearby peers, with no means of global communications and coordination. In this paper, we will show that swarms can perform better, when they self-adapt to admit heterogeneous behavior roles.
Methods: We model a foraging swarm task as an extensive-form fully-cooperative game, in which the swarm reward is an additive function of individual contributions (the sum of collected items). To maximize the swarm reward, previous work proposed using distributed reinforcement learning, where each robot adapts its own collision-avoidance decisions based on the Effectiveness Index reward (EI). EI uses information about the time between their own collisions (information readily available even to simple physical robots). While promising, the use of EI is brittle (as we show), since robots that selfishly seek to optimize their own EI (minimizing time spent on collisions) can actually cause swarm-wide performance to degrade.
Results: To address this, we derive a reward function from a game-theoretic view of swarm foraging as a fully-cooperative, unknown horizon repeating game. We demonstrate analytically that the total coordination overhead of the swarm (total time spent on collision-avoidance, rather than foraging per-se) is directly tied to the total utility of the swarm: less overhead, more items collected. Treating every collision as a stage in the repeating game, the overhead is bounded by the total EI of all robots. We then use a marginal-contribution (difference-reward) formulation to derive individual rewards from the total EI. The resulting Aligned Effective Index reward has the property that each individual can estimate the impact of its decisions on the swarm: individual improvements translate to swarm improvements. We show that provably generalizes previous work, adding a component that computes the effect of counterfactual robot absence. Different assumptions on this counterfactual lead to bounds on from above and below.
Discussion: While the theoretical analysis clarifies both assumptions and gaps with respect to the reality of robots, experiments with real and simulated robots empirically demonstrate the efficacy of the approach in practice, and the importance of behavioral (decision-making) diversity in optimizing swarm goals.
{"title":"Heterogeneous foraging swarms can be better.","authors":"Gal A Kaminka, Yinon Douchan","doi":"10.3389/frobt.2024.1426282","DOIUrl":"10.3389/frobt.2024.1426282","url":null,"abstract":"<p><strong>Introduction: </strong>Inspired by natural phenomena, generations of researchers have been investigating how a swarm of robots can act coherently and purposefully, when individual robots can only sense and communicate with nearby peers, with no means of global communications and coordination. In this paper, we will show that swarms can perform better, when they self-adapt to admit heterogeneous behavior roles.</p><p><strong>Methods: </strong>We model a foraging swarm task as an extensive-form fully-cooperative game, in which the swarm reward is an additive function of individual contributions (the sum of collected items). To maximize the swarm reward, previous work proposed using distributed reinforcement learning, where each robot adapts its own collision-avoidance decisions based on the Effectiveness Index reward (<i>EI</i>). <i>EI</i> uses information about the time between their own collisions (information readily available even to simple physical robots). While promising, the use of <i>EI</i> is brittle (as we show), since robots that selfishly seek to optimize their own <i>EI</i> (minimizing time spent on collisions) can actually cause swarm-wide performance to degrade.</p><p><strong>Results: </strong>To address this, we derive a reward function from a game-theoretic view of swarm foraging as a fully-cooperative, unknown horizon repeating game. We demonstrate analytically that the total coordination overhead of the swarm (total time spent on collision-avoidance, rather than foraging per-se) is directly tied to the total utility of the swarm: less overhead, more items collected. Treating every collision as a stage in the repeating game, the overhead is bounded by the total <i>EI</i> of all robots. We then use a marginal-contribution (difference-reward) formulation to derive individual rewards from the total <i>EI</i>. The resulting Aligned Effective Index <math><mrow><mo>(</mo> <mrow><mi>A</mi> <mi>E</mi> <mi>I</mi></mrow> <mo>)</mo></mrow> </math> reward has the property that each individual can estimate the impact of its decisions on the swarm: individual improvements translate to swarm improvements. We show that <math><mrow><mi>A</mi> <mi>E</mi> <mi>I</mi></mrow> </math> provably generalizes previous work, adding a component that computes the effect of counterfactual robot absence. Different assumptions on this counterfactual lead to bounds on <math><mrow><mi>A</mi> <mi>E</mi> <mi>I</mi></mrow> </math> from above and below.</p><p><strong>Discussion: </strong>While the theoretical analysis clarifies both assumptions and gaps with respect to the reality of robots, experiments with real and simulated robots empirically demonstrate the efficacy of the approach in practice, and the importance of behavioral (decision-making) diversity in optimizing swarm goals.</p>","PeriodicalId":47597,"journal":{"name":"Frontiers in Robotics and AI","volume":"11 ","pages":"1426282"},"PeriodicalIF":2.9,"publicationDate":"2025-01-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11788533/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143123752","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-01-15eCollection Date: 2024-01-01DOI: 10.3389/frobt.2024.1435197
Thomas Barbier, Céline Teulière, Jochen Triesch
Biological vision systems simultaneously learn to efficiently encode their visual inputs and to control the movements of their eyes based on the visual input they sample. This autonomous joint learning of visual representations and actions has previously been modeled in the Active Efficient Coding (AEC) framework and implemented using traditional frame-based cameras. However, modern event-based cameras are inspired by the retina and offer advantages in terms of acquisition rate, dynamic range, and power consumption. Here, we propose a first AEC system that is fully implemented as a Spiking Neural Network (SNN) driven by inputs from an event-based camera. This input is efficiently encoded by a two-layer SNN, which in turn feeds into a spiking reinforcement learner that learns motor commands to maximize an intrinsic reward signal. This reward signal is computed directly from the activity levels of the first two layers. We test our approach on two different behaviors: visual tracking of a translating target and stabilizing the orientation of a rotating target. To the best of our knowledge, our work represents the first ever fully spiking AEC model.
{"title":"A spiking neural network for active efficient coding.","authors":"Thomas Barbier, Céline Teulière, Jochen Triesch","doi":"10.3389/frobt.2024.1435197","DOIUrl":"https://doi.org/10.3389/frobt.2024.1435197","url":null,"abstract":"<p><p>Biological vision systems simultaneously learn to efficiently encode their visual inputs and to control the movements of their eyes based on the visual input they sample. This autonomous joint learning of visual representations and actions has previously been modeled in the Active Efficient Coding (AEC) framework and implemented using traditional frame-based cameras. However, modern event-based cameras are inspired by the retina and offer advantages in terms of acquisition rate, dynamic range, and power consumption. Here, we propose a first AEC system that is fully implemented as a Spiking Neural Network (SNN) driven by inputs from an event-based camera. This input is efficiently encoded by a two-layer SNN, which in turn feeds into a spiking reinforcement learner that learns motor commands to maximize an intrinsic reward signal. This reward signal is computed directly from the activity levels of the first two layers. We test our approach on two different behaviors: visual tracking of a translating target and stabilizing the orientation of a rotating target. To the best of our knowledge, our work represents the first ever fully spiking AEC model.</p>","PeriodicalId":47597,"journal":{"name":"Frontiers in Robotics and AI","volume":"11 ","pages":"1435197"},"PeriodicalIF":2.9,"publicationDate":"2025-01-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11775837/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143068827","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}