Pub Date : 2024-08-14DOI: 10.1016/j.cogsys.2024.101275
Jiale Wang , Zhen Liu , Tingting Liu , Yuanyi Wang
Simulating crowd motion in emergency scenarios remains a challenge in computer graphics due to crowd heterogeneity and environmental complexity. However, existing crowd simulation methods homogenize the agent model and simplify target selection and motion navigation of emergency crowds. To address these problems, we propose a multi-agent motion simulation method for emergency scenario deduction. First, we propose a multi-agent model to simulate crowd heterogeneity. This model includes a personality-based heterogeneous agent model and an agent perception model that considers vision, hearing, and familiarity with the environment. Second, we propose a target selection strategy based on the motion patterns of actual pedestrians. This strategy employs mathematical models and our agent perception model to guide agents in selecting appropriate targets. Finally, we propose a global navigation algorithm that combines random sampling with heuristic search methods. Concurrently, we use our multi-agent model to adjust the agent’s local motion planning to deduce the motion states of emergency crowds naturally. Experimental results validate that our method can realistically and reasonably simulate crowd motion in emergency scenarios.
{"title":"A multi-agent motion simulation method for emergency scenario deduction","authors":"Jiale Wang , Zhen Liu , Tingting Liu , Yuanyi Wang","doi":"10.1016/j.cogsys.2024.101275","DOIUrl":"10.1016/j.cogsys.2024.101275","url":null,"abstract":"<div><p>Simulating crowd motion in emergency scenarios remains a challenge in computer graphics due to crowd heterogeneity and environmental complexity. However, existing crowd simulation methods homogenize the agent model and simplify target selection and motion navigation of emergency crowds. To address these problems, we propose a multi-agent motion simulation method for emergency scenario deduction. First, we propose a multi-agent model to simulate crowd heterogeneity. This model includes a personality-based heterogeneous agent model and an agent perception model that considers vision, hearing, and familiarity with the environment. Second, we propose a target selection strategy based on the motion patterns of actual pedestrians. This strategy employs mathematical models and our agent perception model to guide agents in selecting appropriate targets. Finally, we propose a global navigation algorithm that combines random sampling with heuristic search methods. Concurrently, we use our multi-agent model to adjust the agent’s local motion planning to deduce the motion states of emergency crowds naturally. Experimental results validate that our method can realistically and reasonably simulate crowd motion in emergency scenarios.</p></div>","PeriodicalId":55242,"journal":{"name":"Cognitive Systems Research","volume":null,"pages":null},"PeriodicalIF":2.1,"publicationDate":"2024-08-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142164456","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-07-31DOI: 10.1016/j.cogsys.2024.101266
Rocco Ballester , Yanis Labeyrie , Mehmet Oguz Mulayim , Jose Luis Fernandez-Marquez , Jesus Cerquides
In emergency situations, social media platforms produce a vast amount of real-time data that holds immense value, particularly in the first 72 h following a disaster event. Despite previous efforts, efficiently determining the geographical location of images related to a new disaster remains an unresolved operational challenge. Currently, the state-of-the-art approach for dealing with these first response mapping is first filtering and then submitting the images to be geolocated to a volunteer crowd, assigning the images randomly to the volunteers. In this work, we extend our previous paper (Ballester et al., 2023) to explore the potential of artificial intelligence (AI) in aiding emergency responders and disaster relief organizations in geolocating social media images from a zone recently hit by a disaster. Our contributions include building two different models in which we try to (i) be able to learn volunteers’ error profiles and (ii) intelligently assign tasks to those volunteers who exhibit higher proficiency. Moreover, we present methods that outperform random allocation of tasks, analyze the effect on the models’ performance when varying numerous parameters, and show that for a given set of tasks and volunteers, we are able to process them with a significantly lower annotation budget, that is, we are able to make fewer volunteer solicitations without losing any quality on the final consensus.
{"title":"Crowdsourced geolocation: Detailed exploration of mathematical and computational modeling approaches","authors":"Rocco Ballester , Yanis Labeyrie , Mehmet Oguz Mulayim , Jose Luis Fernandez-Marquez , Jesus Cerquides","doi":"10.1016/j.cogsys.2024.101266","DOIUrl":"10.1016/j.cogsys.2024.101266","url":null,"abstract":"<div><p>In emergency situations, social media platforms produce a vast amount of real-time data that holds immense value, particularly in the first 72 h following a disaster event. Despite previous efforts, efficiently determining the geographical location of images related to a new disaster remains an unresolved operational challenge. Currently, the state-of-the-art approach for dealing with these first response mapping is first filtering and then submitting the images to be geolocated to a volunteer crowd, assigning the images randomly to the volunteers. In this work, we extend our previous paper (Ballester et al., 2023) to explore the potential of artificial intelligence (AI) in aiding emergency responders and disaster relief organizations in geolocating social media images from a zone recently hit by a disaster. Our contributions include building two different models in which we try to (i) be able to learn volunteers’ error profiles and (ii) intelligently assign tasks to those volunteers who exhibit higher proficiency. Moreover, we present methods that outperform random allocation of tasks, analyze the effect on the models’ performance when varying numerous parameters, and show that for a given set of tasks and volunteers, we are able to process them with a significantly lower annotation budget, that is, we are able to make fewer volunteer solicitations without losing any quality on the final consensus.</p></div>","PeriodicalId":55242,"journal":{"name":"Cognitive Systems Research","volume":null,"pages":null},"PeriodicalIF":2.1,"publicationDate":"2024-07-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141931768","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-07-25DOI: 10.1016/j.cogsys.2024.101267
Xun Wang , Tingting Liu , Zhen Liu , Zheng Fang
Empathy plays an important role in human conversations as an ability that enables individuals to understand the emotions and situations of others. Integrating empathy into dialogue systems is a crucial step in making them humanized. Relevant psychological studies have shown that a complete, high-quality empathetic dialogue should consist of the following two stages: (1) Empathetic Perception: the listener needs to perceive the emotional state of the speaker from both cognitive and affective aspects; (2) Empathetic Expression: the appropriate expression is chosen to respond to the perceived information. However, many existing studies on empathetic response generation only focus on one of these stages, resulting in incomplete and insufficiently empathetic responses. To this end, we propose the EmpCI, a two-stage empathetic response generation model that utilizes commonsense knowledge and mixed empathetic intent, respectively. Specifically, we use commonsense knowledge in the first stage to enhance the model’s perception of the user’s emotion and introduce mixed empathetic intent in the second stage to generate responses with appropriate expressions for the perceived information. Finally, we evaluated the EmpCI on the EmpatheticDialogues dataset, and extensive experiment results show that the proposed model outperforms the baselines in both perceiving users’ emotions and generating empathetic responses.
{"title":"EmpCI: Empathetic response generation with common sense and empathetic intent","authors":"Xun Wang , Tingting Liu , Zhen Liu , Zheng Fang","doi":"10.1016/j.cogsys.2024.101267","DOIUrl":"10.1016/j.cogsys.2024.101267","url":null,"abstract":"<div><p>Empathy plays an important role in human conversations as an ability that enables individuals to understand the emotions and situations of others. Integrating empathy into dialogue systems is a crucial step in making them humanized. Relevant psychological studies have shown that a complete, high-quality empathetic dialogue should consist of the following two stages: (1) Empathetic Perception: the listener needs to perceive the emotional state of the speaker from both cognitive and affective aspects; (2) Empathetic Expression: the appropriate expression is chosen to respond to the perceived information. However, many existing studies on empathetic response generation only focus on one of these stages, resulting in incomplete and insufficiently empathetic responses. To this end, we propose the EmpCI, a two-stage empathetic response generation model that utilizes commonsense knowledge and mixed empathetic intent, respectively. Specifically, we use commonsense knowledge in the first stage to enhance the model’s perception of the user’s emotion and introduce mixed empathetic intent in the second stage to generate responses with appropriate expressions for the perceived information. Finally, we evaluated the EmpCI on the EmpatheticDialogues dataset, and extensive experiment results show that the proposed model outperforms the baselines in both perceiving users’ emotions and generating empathetic responses.</p></div>","PeriodicalId":55242,"journal":{"name":"Cognitive Systems Research","volume":null,"pages":null},"PeriodicalIF":2.1,"publicationDate":"2024-07-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141848637","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
In the digital age, people prefer digital content, but screen-related health concerns like eye strain and blue light emerge. Legibility gains importance in digital text, especially in fields like optometry and for those with low vision. Therefore, having good letter recognition ensures better readability of words and written language in general. This work focuses on defining three typeface legibility indices from the judgements of a group of 31 observers. Those indices are based on statistics, confusion matrices, and power indices from game theory. As far as we know, this is the first time that typeface legibility indices have been defined using game theory. These indices help us to globally assess how legible is a typeface. We apply them to three commonly used typefaces (Roboto, Helvetica and Georgia), and to a new one developed for the authors (Optotipica 5 v2022). This comparison helps us understand which typefaces are more legible according to the defined indices on digital screens. The major conclusions are: (1) The three indices are highly consistent pairwise; (2) Helvetica is the most legible typeface for uppercase letters, whilst Optotipica is the most legible for lowercase; (3) the two cases of Helvetica exhibit uniform high legibility metrics, ensuring optimal recognition regardless of letter case.
{"title":"Typeface recognition and legibility metrics","authors":"Xavier Molinero , Montserrat Tàpias , Andreu Balius , Francesc Salvadó","doi":"10.1016/j.cogsys.2024.101263","DOIUrl":"10.1016/j.cogsys.2024.101263","url":null,"abstract":"<div><p>In the digital age, people prefer digital content, but screen-related health concerns like eye strain and blue light emerge. Legibility gains importance in digital text, especially in fields like optometry and for those with low vision. Therefore, having good letter recognition ensures better readability of words and written language in general. This work focuses on defining three typeface legibility indices from the judgements of a group of 31 observers. Those indices are based on statistics, confusion matrices, and power indices from game theory. As far as we know, this is the first time that typeface legibility indices have been defined using game theory. These indices help us to globally assess how legible is a typeface. We apply them to three commonly used typefaces (Roboto, Helvetica and Georgia), and to a new one developed for the authors (Optotipica 5 v2022). This comparison helps us understand which typefaces are more legible according to the defined indices on digital screens. The major conclusions are: (1) The three indices are highly consistent pairwise; (2) Helvetica is the most legible typeface for uppercase letters, whilst Optotipica is the most legible for lowercase; (3) the two cases of Helvetica exhibit uniform high legibility metrics, ensuring optimal recognition regardless of letter case.</p></div>","PeriodicalId":55242,"journal":{"name":"Cognitive Systems Research","volume":null,"pages":null},"PeriodicalIF":2.1,"publicationDate":"2024-07-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141779529","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-07-10DOI: 10.1016/j.cogsys.2024.101265
Carlos Johnnatan Sandoval-Arrayga, Gustavo Palacios-Ramirez, Felix Francisco Ramos-Corchado
In 2020, Mc Fadden published an article in which he discusses how algorithms can be encoded in time and space. By analyzing the topology of the cytoarchitecture of the brain, cognitive architectures can understand the underlying mechanisms that have led to the development of human intelligence in space. In this study, our focus lies in investigating temporal heterogeneity as a mechanism that the brain could have developed not solely as a biological constraint, but also as an evolutionary advantage. To accomplish this, we employed virtual agents within a virtual environment and constructed a prototype cognitive architecture. Subsequently, we compared the benefits and drawbacks of having this cognitive architecture operate under a model of temporal heterogeneity versus one characterized by temporal homogeneity. At the conclusion of the article, we present the results obtained from two perspectives. From a quantitative standpoint, we contrast the agents’ adaptation to the environment based on the cognitive architecture model employed by each agent. On this front, we found evidence that temporal heterogeneity might be useful in finding parameter optimizations faster, amongst other benefits. From a qualitative perspective, we examine the potential of this model to explore the cognitive processes of the virtual agents, concluding that a different representation of percepts is needed, which we further discuss.
{"title":"Temporal heterogeneity in cognitive architectures","authors":"Carlos Johnnatan Sandoval-Arrayga, Gustavo Palacios-Ramirez, Felix Francisco Ramos-Corchado","doi":"10.1016/j.cogsys.2024.101265","DOIUrl":"10.1016/j.cogsys.2024.101265","url":null,"abstract":"<div><p>In 2020, Mc Fadden published an article in which he discusses how algorithms can be encoded in time and space. By analyzing the topology of the cytoarchitecture of the brain, cognitive architectures can understand the underlying mechanisms that have led to the development of human intelligence in space. In this study, our focus lies in investigating temporal heterogeneity as a mechanism that the brain could have developed not solely as a biological constraint, but also as an evolutionary advantage. To accomplish this, we employed virtual agents within a virtual environment and constructed a prototype cognitive architecture. Subsequently, we compared the benefits and drawbacks of having this cognitive architecture operate under a model of temporal heterogeneity versus one characterized by temporal homogeneity. At the conclusion of the article, we present the results obtained from two perspectives. From a quantitative standpoint, we contrast the agents’ adaptation to the environment based on the cognitive architecture model employed by each agent. On this front, we found evidence that temporal heterogeneity might be useful in finding parameter optimizations faster, amongst other benefits. From a qualitative perspective, we examine the potential of this model to explore the cognitive processes of the virtual agents, concluding that a different representation of percepts is needed, which we further discuss.</p></div>","PeriodicalId":55242,"journal":{"name":"Cognitive Systems Research","volume":null,"pages":null},"PeriodicalIF":2.1,"publicationDate":"2024-07-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S1389041724000597/pdfft?md5=556b6d9aa1318731fad683858307103a&pid=1-s2.0-S1389041724000597-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141713901","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-07-10DOI: 10.1016/j.cogsys.2024.101264
Vikas Sharma , Akshi Kumar , Kapil Sharma
Digital Twin (DT) will transform digital healthcare and push it far beyond expectations. DT creates a virtual representation of a physical object reflecting its current state using real-time converted data. Nowadays, Women’s health is more frequently impacted by cervical cancer, but early detection and rapid treatment are critical factors in the cure of cervical cancer. This paper proposes and implements an automated cervical cancer detection DT framework in healthcare. This framework is a valuable approach to enhance digital healthcare operations. In this proposed work, the SIPaKMeD dataset was used for multi-cell classification. There were 1013 images (Input size 224 × 224 × 3) in the collection, from which 4103 cells could be extracted. As a result, the CervixNet classifier model is developed using machine learning to detect cervical problems and diagnose cervical disease. Using pre-trained recurrent neural networks (RNNs), CervixNet extracted 1172 features, and after that, 792 features were selected using an independent principal component analysis (PCA) algorithm. The implemented models achieved the highest accuracy for predicting cervical cancer using different algorithms. The collected information has shown that integrating DT with the healthcare industry will enhance healthcare procedures by integrating patients and medical staff in a scalable, intelligent, and comprehensive health ecosystem. Finally, the suggested method produces an impressive 98.91 % classification accuracy in all classes, especially for support vector machines (SVM).
{"title":"Digital twin application in women’s health: Cervical cancer diagnosis with CervixNet","authors":"Vikas Sharma , Akshi Kumar , Kapil Sharma","doi":"10.1016/j.cogsys.2024.101264","DOIUrl":"https://doi.org/10.1016/j.cogsys.2024.101264","url":null,"abstract":"<div><p>Digital Twin (DT) will transform digital healthcare and push it far beyond expectations. DT creates a virtual representation of a physical object reflecting its current state using real-time converted data. Nowadays, Women’s health is more frequently impacted by cervical cancer, but early detection and rapid treatment are critical factors in the cure of cervical cancer. This paper proposes and implements an automated cervical cancer detection DT framework in healthcare. This framework is a valuable approach to enhance digital healthcare operations. In this proposed work, the SIPaKMeD dataset was used for multi-cell classification. There were 1013 images (Input size 224 × 224 × 3) in the collection, from which 4103 cells could be extracted. As a result, the CervixNet classifier model is developed using machine learning to detect cervical problems and diagnose cervical disease. Using pre-trained recurrent neural networks (RNNs), CervixNet extracted 1172 features, and after that, 792 features were selected using an independent principal component analysis (PCA) algorithm. The implemented models achieved the highest accuracy for predicting cervical cancer using different algorithms. The collected information has shown that integrating DT with the healthcare industry will enhance healthcare procedures by integrating patients and medical staff in a scalable, intelligent, and comprehensive health ecosystem. Finally, the suggested method produces an impressive 98.91 % classification accuracy in all classes, especially for support vector machines (SVM).</p></div>","PeriodicalId":55242,"journal":{"name":"Cognitive Systems Research","volume":null,"pages":null},"PeriodicalIF":2.1,"publicationDate":"2024-07-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141606012","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-07-01DOI: 10.1016/j.cogsys.2024.101262
Ivan Axel Dounce, Félix Ramos
As humans, we have an excellent performance when perceiving the environment. In the artificial world, it is important for machines to perceive their environment so they can make correct decisions and act accordingly. An essential process to accomplish perception is to identify objects in a scene, but, as in reality, these objects can appear as ambiguous, and additionally, those objects are embedded into a particular scene. For our proposal, we created an architecture to identify ambiguous objects by using scene information to guide the identification process. The design is based on the human cortical systems that participate in object and scene recognition. In our study, we validate this proposal by analyzing a prior human experiment that demonstrates and quantifies the impact of scene information on ambiguous objects. Our findings demonstrate that employing the presented architecture on an object recognition task results in superior machine performance with familiar scenes, as opposed to unfamiliar or absent ones, consistent with human behavior.
{"title":"Biologically inspired architecture for the identification of ambiguous objects using scene associations","authors":"Ivan Axel Dounce, Félix Ramos","doi":"10.1016/j.cogsys.2024.101262","DOIUrl":"https://doi.org/10.1016/j.cogsys.2024.101262","url":null,"abstract":"<div><p>As humans, we have an excellent performance when perceiving the environment. In the artificial world, it is important for machines to perceive their environment so they can make correct decisions and act accordingly. An essential process to accomplish perception is to identify objects in a scene, but, as in reality, these objects can appear as ambiguous, and additionally, those objects are embedded into a particular scene. For our proposal, we created an architecture to identify ambiguous objects by using scene information to guide the identification process. The design is based on the human cortical systems that participate in object and scene recognition. In our study, we validate this proposal by analyzing a prior human experiment that demonstrates and quantifies the impact of scene information on ambiguous objects. Our findings demonstrate that employing the presented architecture on an object recognition task results in superior machine performance with familiar scenes, as opposed to unfamiliar or absent ones, consistent with human behavior.</p></div>","PeriodicalId":55242,"journal":{"name":"Cognitive Systems Research","volume":null,"pages":null},"PeriodicalIF":2.1,"publicationDate":"2024-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141543102","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-06-26DOI: 10.1016/j.cogsys.2024.101261
Oleg Sychev
Automation of teaching people new skills requires modeling of human reasoning because human cognition involves active reasoning over the new subject domain to acquire skills that will later become automatic. The article presents Thought Process Trees — a language for modeling human reasoning that was created to facilitate the development of intelligent tutoring systems, which can perform the same reasoning that is expected of a student and find deficiencies in their line of thinking, providing explanatory messages and allowing them to learn from performance errors. The methodology of building trees which better reflect human learning is discussed, with examples of design choices during the modeling process and their consequences. The characteristics of educational modeling that impact building subject-domain models for intelligent tutoring systems are discussed. The trees were formalized and served as a basis for developing a framework for constructing intelligent tutoring systems. This significantly lowered the time required to build and debug a constraint-based subject-domain model. The framework has already been used to develop five intelligent tutoring systems and their prototypes and is being used to develop more of them.
{"title":"Educational models for cognition: Methodology of modeling intellectual skills for intelligent tutoring systems","authors":"Oleg Sychev","doi":"10.1016/j.cogsys.2024.101261","DOIUrl":"https://doi.org/10.1016/j.cogsys.2024.101261","url":null,"abstract":"<div><p>Automation of teaching people new skills requires modeling of human reasoning because human cognition involves active reasoning over the new subject domain to acquire skills that will later become automatic. The article presents Thought Process Trees — a language for modeling human reasoning that was created to facilitate the development of intelligent tutoring systems, which can perform the same reasoning that is expected of a student and find deficiencies in their line of thinking, providing explanatory messages and allowing them to learn from performance errors. The methodology of building trees which better reflect human learning is discussed, with examples of design choices during the modeling process and their consequences. The characteristics of educational modeling that impact building subject-domain models for intelligent tutoring systems are discussed. The trees were formalized and served as a basis for developing a framework for constructing intelligent tutoring systems. This significantly lowered the time required to build and debug a constraint-based subject-domain model. The framework has already been used to develop five intelligent tutoring systems and their prototypes and is being used to develop more of them.</p></div>","PeriodicalId":55242,"journal":{"name":"Cognitive Systems Research","volume":null,"pages":null},"PeriodicalIF":2.1,"publicationDate":"2024-06-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141543101","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-06-18DOI: 10.1016/j.cogsys.2024.101257
Massimo Cossentino, Giovanni Pilato
In the last years, the System 1/System 2 cognitive architecture, proposed by psychologist Daniel Kahneman, raised the interest of many researchers in the field. System 1 is an intuitive, automatic, and fast-thinking system working effortlessly, without conscious effort. System 2 is a deliberate, analytical, and slower-thinking system employing conscious effort and attention. This work proposes an innovative approach that exploits techniques typical of information retrieval (the trie data structure) to efficiently encode the solutions’ repository at the border between System 2 and System 1. This repository stores the solutions (successful plans) the agent has already used and can re-enact to achieve the goals. System 2 conceives new plans and delegates System 1 to execute them. If the plan is successful (and so it becomes a solution), System 1 stores that in the repository to quickly retrieve any solution that may help fulfil the goals deliberated by System 2 in the future.
{"title":"Using a trie-based approach for storage and retrieval of goal-oriented plans in an S1/S2 cognitive architecture","authors":"Massimo Cossentino, Giovanni Pilato","doi":"10.1016/j.cogsys.2024.101257","DOIUrl":"https://doi.org/10.1016/j.cogsys.2024.101257","url":null,"abstract":"<div><p>In the last years, the System 1/System 2 cognitive architecture, proposed by psychologist Daniel Kahneman, raised the interest of many researchers in the field. <em>System 1</em> is an intuitive, automatic, and fast-thinking system working effortlessly, without conscious effort. <em>System 2</em> is a deliberate, analytical, and slower-thinking system employing conscious effort and attention. This work proposes an innovative approach that exploits techniques typical of information retrieval (the trie data structure) to efficiently encode the solutions’ repository at the border between System 2 and System 1. This repository stores the solutions (successful plans) the agent has already used and can re-enact to achieve the goals. System 2 conceives new plans and delegates System 1 to execute them. If the plan is successful (and so it becomes a solution), System 1 stores that in the repository to quickly retrieve any solution that may help fulfil the goals deliberated by System 2 in the future.</p></div>","PeriodicalId":55242,"journal":{"name":"Cognitive Systems Research","volume":null,"pages":null},"PeriodicalIF":2.1,"publicationDate":"2024-06-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S1389041724000512/pdfft?md5=90857dbbd82c5a305bde8800fccf4b48&pid=1-s2.0-S1389041724000512-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141543100","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-06-16DOI: 10.1016/j.cogsys.2024.101260
Christ Devia , Camilo Jara Do Nascimento , Samuel Madariaga , Pedro.E. Maldonado , Catalina Murúa , Rodrigo C. Vergara
This article presents a transdisciplinary analysis of the challenges in fusing neuroscience concepts with artificial intelligence (AI) to create AI systems inspired by biological cognition. We explore the structural and functional disparities between the neocortex’s canonical microcircuits and existing AI models, focusing on architectural differences, learning mechanisms, and energy efficiency. The discussion extends to adapting non-goal-oriented learning and dynamic neuronal connections from biological brains to enhance AI’s flexibility and efficiency. This work underscores the potential of neuroscientific insights to revolutionize AI development, advocating for a paradigm shift towards more adaptable and brain-like AI systems. We conclude that there is major room for bioinspiration by focusing on developing architecture, objective functions, and learning rules using a local instead of a global approach.
{"title":"Exploring biological challenges in building a thinking machine","authors":"Christ Devia , Camilo Jara Do Nascimento , Samuel Madariaga , Pedro.E. Maldonado , Catalina Murúa , Rodrigo C. Vergara","doi":"10.1016/j.cogsys.2024.101260","DOIUrl":"10.1016/j.cogsys.2024.101260","url":null,"abstract":"<div><p>This article presents a transdisciplinary analysis of the challenges in fusing neuroscience concepts with artificial intelligence (AI) to create AI systems inspired by biological cognition. We explore the structural and functional disparities between the neocortex’s canonical microcircuits and existing AI models, focusing on architectural differences, learning mechanisms, and energy efficiency. The discussion extends to adapting non-goal-oriented learning and dynamic neuronal connections from biological brains to enhance AI’s flexibility and efficiency. This work underscores the potential of neuroscientific insights to revolutionize AI development, advocating for a paradigm shift towards more adaptable and brain-like AI systems. We conclude that there is major room for bioinspiration by focusing on developing architecture, objective functions, and learning rules using a local instead of a global approach.</p></div>","PeriodicalId":55242,"journal":{"name":"Cognitive Systems Research","volume":null,"pages":null},"PeriodicalIF":2.1,"publicationDate":"2024-06-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141412078","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}