Pub Date : 2024-02-15DOI: 10.1016/j.cogsys.2024.101216
Toni Taipalus
Vector database management systems have emerged as an important component in modern data management, driven by the growing importance for the need to computationally describe rich data such as texts, images and video in various domains such as recommender systems, similarity search, and chatbots. These data descriptions are captured as numerical vectors that are computationally inexpensive to store and compare. However, the unique characteristics of vectorized data, including high dimensionality and sparsity, demand specialized solutions for efficient storage, retrieval, and processing. This narrative literature review provides an accessible introduction to the fundamental concepts, use-cases, and current challenges associated with vector database management systems, offering an overview for researchers and practitioners seeking to facilitate effective vector data management.
{"title":"Vector database management systems: Fundamental concepts, use-cases, and current challenges","authors":"Toni Taipalus","doi":"10.1016/j.cogsys.2024.101216","DOIUrl":"https://doi.org/10.1016/j.cogsys.2024.101216","url":null,"abstract":"<div><p>Vector database management systems have emerged as an important component in modern data management, driven by the growing importance for the need to computationally describe rich data such as texts, images and video in various domains such as recommender systems, similarity search, and chatbots. These data descriptions are captured as numerical vectors that are computationally inexpensive to store and compare. However, the unique characteristics of vectorized data, including high dimensionality and sparsity, demand specialized solutions for efficient storage, retrieval, and processing. This narrative literature review provides an accessible introduction to the fundamental concepts, use-cases, and current challenges associated with vector database management systems, offering an overview for researchers and practitioners seeking to facilitate effective vector data management.</p></div>","PeriodicalId":55242,"journal":{"name":"Cognitive Systems Research","volume":"85 ","pages":"Article 101216"},"PeriodicalIF":3.9,"publicationDate":"2024-02-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S1389041724000093/pdfft?md5=c470854e4bee590cea4c3fc43ca20924&pid=1-s2.0-S1389041724000093-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139744343","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-02-10DOI: 10.1016/j.cogsys.2024.101214
Maxim Sharaev , Maxim Nekrashevich , Daria Kostanian , Victoria Voinova , Olga Sysoeva
Rett Syndrome (RTT) is a rare neurodevelopmental disorder caused by mutation in the MECP2 gene. No cures are still available, but several clinical trials are ongoing. Here we examine neurophysiological correlates of auditory processing for ability to differentiate patients with RTT from typically developing (TD) peers applying standard machine learning (ML) methods and pipelines. Capitalized on the available event-related potential (ERP) data recorded in response to tone presented at different rates (stimulus onset asynchrony 900, 1800 and 3600 ms) from 24 patients with RTT and 27 their TD peer. We considered the most common ML models that are widely used for classification tasks. These include both linear models (logistic regression, support-vector machine with linear kernel) and tree-based nonlinear models (random forest, gradient boosting). Based on these methods we were able to differentiate RTT from TD children with high accuracy (with up to 0.94 ROC-AUC score), which was evidently higher at the fastest presentation rate. Importance analysis and perturbation importance pointed out that the most important feature for classification is P2-N2 peak-to-peak amplitude, consistently across the approaches and blocks with different presentation rate. The results suggest the unique pattern of ERP characteristics for RTT and points to features of importance. The results might be relevant for establishing outcome measures for clinical trials.
{"title":"Auditory event-related potential differentiates girls with Rett syndrome from their typically-developing peers with high accuracy: Machine learning study","authors":"Maxim Sharaev , Maxim Nekrashevich , Daria Kostanian , Victoria Voinova , Olga Sysoeva","doi":"10.1016/j.cogsys.2024.101214","DOIUrl":"https://doi.org/10.1016/j.cogsys.2024.101214","url":null,"abstract":"<div><p>Rett Syndrome (RTT) is a rare neurodevelopmental disorder caused by mutation in the <em>MECP2</em> gene. No cures are still available, but several clinical trials are ongoing. Here we examine neurophysiological correlates of auditory processing for ability to differentiate patients with RTT from typically developing (TD) peers applying standard machine learning (ML) methods and pipelines. Capitalized on the available event-related potential (ERP) data recorded in response to tone presented at different rates (stimulus onset asynchrony 900, 1800 and 3600 ms) from 24 patients with RTT and 27 their TD peer. We considered the most common ML models that are widely used for classification tasks. These include both linear models (logistic regression, support-vector machine with linear kernel) and tree-based nonlinear models (random forest, gradient boosting). Based on these methods we were able to differentiate RTT from TD children with high accuracy (with up to 0.94 ROC-AUC score), which was evidently higher at the fastest presentation rate. Importance analysis and perturbation importance pointed out that the most important feature for classification is P2-N2 peak-to-peak amplitude, consistently across the approaches and blocks with different presentation rate. The results suggest the unique pattern of ERP characteristics for RTT and points to features of importance. The results might be relevant for establishing outcome measures for clinical trials.</p></div>","PeriodicalId":55242,"journal":{"name":"Cognitive Systems Research","volume":"85 ","pages":"Article 101214"},"PeriodicalIF":3.9,"publicationDate":"2024-02-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139738798","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-01-03DOI: 10.1016/j.cogsys.2023.101204
Luis A. Pineda
The Turing Machine is the paradigmatic case of computing machines, but there are others such as analogical, connectionist, quantum and diverse forms of unconventional computing, each based on a particular intuition of the phenomenon of computing. This variety can be captured in terms of system levels, re-interpreting and generalizing Newell’s hierarchy, which includes the knowledge level at the top and the symbol level immediately below it. In this re-interpretation the knowledge level consists of human knowledge and the symbol level is generalized into a new level that here is called The Mode of Computing. Mental processes performed by natural brains are often thought of informally as computing processes and that the brain is alike to computing machinery. However, if natural computing does exist it should be characterized on its own. A proposal to such an effect is that natural computing appeared when interpretations were first made by biological entities, so natural computing and interpreting are two aspects of the same phenomenon, or that consciousness and experience are the manifestations of computing/interpreting. By analogy with computing machinery, there must be a system level at the top of the neural circuitry and directly below the knowledge level that is named here The mode of Natural Computing. If it turns out that such putative object does not exist the proposition that the mind is a computing process should be dropped; but characterizing it would come with solving the hard problem of consciousness.
{"title":"The mode of computing","authors":"Luis A. Pineda","doi":"10.1016/j.cogsys.2023.101204","DOIUrl":"10.1016/j.cogsys.2023.101204","url":null,"abstract":"<div><p><span>The Turing Machine is the paradigmatic case of computing machines, but there are others such as analogical, connectionist, quantum and diverse forms of unconventional computing, each based on a particular intuition of the phenomenon of computing. This variety can be captured in terms of system levels, re-interpreting and generalizing Newell’s hierarchy, which includes the knowledge level at the top and the symbol level immediately below it. In this re-interpretation the knowledge level consists of human knowledge and the symbol level is generalized into a new level that here is called </span><em>The Mode of Computing</em><span>. Mental processes performed by natural brains are often thought of informally as computing processes and that the brain is alike to computing machinery<span>. However, if natural computing does exist it should be characterized on its own. A proposal to such an effect is that natural computing appeared when interpretations were first made by biological entities, so natural computing and interpreting are two aspects of the same phenomenon, or that consciousness and experience are the manifestations of computing/interpreting. By analogy with computing machinery, there must be a system level at the top of the neural circuitry and directly below the knowledge level that is named here </span></span><em>The mode of Natural Computing</em>. If it turns out that such putative object does not exist the proposition that the mind is a computing process should be dropped; but characterizing it would come with solving the hard problem of consciousness.</p></div>","PeriodicalId":55242,"journal":{"name":"Cognitive Systems Research","volume":"84 ","pages":"Article 101204"},"PeriodicalIF":3.9,"publicationDate":"2024-01-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139103957","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-01-02DOI: 10.1016/j.cogsys.2023.101202
Paul R. Smart
The terms “extended cognition” and the “extended mind” identify two strands of philosophical argument that are commonly subsumed under the general heading of active externalism. The present paper describes an integrated approach to understanding extended cognition and the extended mind—one that papers over the differences between these two, ostensibly distinct, forms of cognitive extension. As an added bonus, the paper describes how active externalism might be applied to the realm of non-cognitive phenomena, thereby yielding an expansion in the theoretical and empirical scope of the active externalist enterprise. Both these points of progress stem from what is called the dispositional hypothesis. According to the dispositional hypothesis, extended cognition occurs when the mechanisms responsible for the manifestation of dispositional properties include components that lie beyond the borders of the thing to which the dispositional properties are ascribed.
{"title":"Extended X: Extending the reach of active externalism","authors":"Paul R. Smart","doi":"10.1016/j.cogsys.2023.101202","DOIUrl":"10.1016/j.cogsys.2023.101202","url":null,"abstract":"<div><p>The terms “extended cognition” and the “extended mind” identify two strands of philosophical argument that are commonly subsumed under the general heading of active externalism. The present paper describes an integrated approach to understanding extended cognition and the extended mind—one that papers over the differences between these two, ostensibly distinct, forms of cognitive extension. As an added bonus, the paper describes how active externalism might be applied to the realm of non-cognitive phenomena, thereby yielding an expansion in the theoretical and empirical scope of the active externalist enterprise. Both these points of progress stem from what is called the <em>dispositional hypothesis</em>. According to the dispositional hypothesis, extended cognition occurs when the mechanisms responsible for the manifestation of dispositional properties include components that lie beyond the borders of the thing to which the dispositional properties are ascribed.</p></div>","PeriodicalId":55242,"journal":{"name":"Cognitive Systems Research","volume":"84 ","pages":"Article 101202"},"PeriodicalIF":3.9,"publicationDate":"2024-01-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139078443","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-12-30DOI: 10.1016/j.cogsys.2023.101206
Xuyao Dai , Zhen Liu , Tingting Liu , Guokun Zuo , Jialin Xu , Changcheng Shi , Yuanyi Wang
Empathy mechanism in communication is the cornerstone for effective and meaningful interaction. Establishing an empathy mechanism in conversation agent (CA) requires accurate recognition of users’ emotions to facilitate generates appropriate empathetic responses. Therefore, we proposed a Multimodal Emotion Recognition Model (MERM) to recognizes a user’s emotional state from multimodal data (audio, facial expressions, and conversation text) during conversation, and an Interactive Empathetic Conversation Model (IECM) to generate empathetic responses based on the MERM. Comparative and ablation study results indicated that the proposed models outperform existing methods in recognizing the user’s emotions and generating appropriate empathetic responses. We also conducted an experiment study, the results indicated that the CA significantly enhances the user’s emotional experience.
对话代理(CA)在人机交互研究中发挥着举足轻重的作用。以往的研究主要集中在会话生成、非语言行为表现和共情能力等方面。然而,由于用户生成的多模态信息中存在各种关系,CA 的行为与上下文信息之间可能会出现不一致。为了应对这一挑战,我们开展了跨学科研究,旨在克服以往研究中观察到的有关 CA 中移情机制和情感交互的局限性。在这方面,我们开发了一个多模态人机情感交互综合框架,使 CA 能够识别人类情感并做出适当回应。该框架包括一个虚拟现实环境中的人形 CA,以及一个交互式多模态情感识别-移情对话生成循环架构。CA通过利用多模态信号(包括音频、面部表情和对话文本)来推断用户的情绪。随后,它通过交互式移情对话模型展示语言行为。已进行的几项与情感相关的实验表明,CA 的多模态识别和表达能力以及行为一致性提高了多模态人机交互的自然度和可信度。总之,这项研究有助于开发多模态人机情感交互的综合框架,提高 CA 的质量、可信度和移情能力。
{"title":"Modelling conversational agent with empathy mechanism","authors":"Xuyao Dai , Zhen Liu , Tingting Liu , Guokun Zuo , Jialin Xu , Changcheng Shi , Yuanyi Wang","doi":"10.1016/j.cogsys.2023.101206","DOIUrl":"10.1016/j.cogsys.2023.101206","url":null,"abstract":"<div><p>Empathy mechanism in communication is the cornerstone for effective and meaningful interaction. Establishing an empathy mechanism in conversation agent (CA) requires accurate recognition of users’ emotions to facilitate generates appropriate empathetic responses. Therefore, we proposed a Multimodal Emotion Recognition Model (MERM) to recognizes a user’s emotional state from multimodal data (audio, facial expressions, and conversation text) during conversation, and an Interactive Empathetic Conversation Model (IECM) to generate empathetic responses based on the MERM. Comparative and ablation study results indicated that the proposed models outperform existing methods in recognizing the user’s emotions and generating appropriate empathetic responses. We also conducted an experiment study, the results indicated that the CA significantly enhances the user’s emotional experience.</p></div>","PeriodicalId":55242,"journal":{"name":"Cognitive Systems Research","volume":"84 ","pages":"Article 101206"},"PeriodicalIF":3.9,"publicationDate":"2023-12-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139070732","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-12-29DOI: 10.1016/j.cogsys.2023.101207
Kazuteru Miyazaki , Hitomi Miyazaki
In recent years, damage caused by negative tweets has become a social problem. In this paper, we consider a method of suppressing negative tweets by using reinforcement learning. In particular, we consider the case where tweet writing is modeled as a multi-agent environment. Numerical experiments verify the effects of suppression using various reinforcement learning methods. We will also verify robustness to environmental changes. We compared the results of Profit Sharing (PS) and Q-learning (QL) as reinforcement learning methods to confirm the effectiveness of PS, and confirmed the behavior of the rationality theorem in a multi-agent environment. Furthermore, in experiments regarding the ability to follow environmental changes, it was confirmed that PS is more robust than QL. If machines can appropriately intervene and interact with posts made by humans, we can expect that negative tweets and even blow-ups can be suppressed automatically without the need for costly human eye monitoring.
{"title":"Suppression of negative tweets using reinforcement learning systems","authors":"Kazuteru Miyazaki , Hitomi Miyazaki","doi":"10.1016/j.cogsys.2023.101207","DOIUrl":"10.1016/j.cogsys.2023.101207","url":null,"abstract":"<div><p>In recent years, damage caused by negative tweets has become a social problem. In this paper, we consider a method of suppressing negative tweets by using reinforcement learning<span>. In particular, we consider the case where tweet writing is modeled as a multi-agent environment. Numerical experiments verify the effects of suppression using various reinforcement learning methods. We will also verify robustness to environmental changes. We compared the results of Profit Sharing (PS) and Q-learning (QL) as reinforcement learning methods to confirm the effectiveness of PS, and confirmed the behavior of the rationality theorem in a multi-agent environment. Furthermore, in experiments regarding the ability to follow environmental changes, it was confirmed that PS is more robust than QL. If machines can appropriately intervene and interact with posts made by humans, we can expect that negative tweets and even blow-ups can be suppressed automatically without the need for costly human eye monitoring.</span></p></div>","PeriodicalId":55242,"journal":{"name":"Cognitive Systems Research","volume":"84 ","pages":"Article 101207"},"PeriodicalIF":3.9,"publicationDate":"2023-12-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139070792","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-12-29DOI: 10.1016/j.cogsys.2023.101203
Andreas Brännström , Joel Wester , Juan Carlos Nieves
Interactive software agents, such as chatbots, are progressively being used in the area of health and well-being. In such applications, where agents engage with users in interpersonal conversations for, e.g., coaching, comfort or behavior-change interventions, there is an increased need for understanding agents’ empathic capabilities. In the current state-of-the-art, there are no tools to do that. In order to understand empathic capabilities in interactive software agents, we need a precise notion of empathy. The literature discusses a variety of definitions of empathy, but there is no consensus of a formal definition. Based on a systematic literature review and a qualitative analysis of recent approaches to empathy in interactive agents for health and well-being, a formal definition—an ontology—of empathy is developed. We present the potential of the formal definition in a controlled user-study by applying it as a tool for assessing empathy in two state-of-the-art health and well-being chatbots; Replika and Wysa. Our findings suggest that our definition captures necessary conditions for assessing empathy in interactive agents, and how it can uncover and explain trends in changing perceptions of empathy over time. The definition, implemented in Web Ontology Language (OWL), may serve as an automated tool, enabling systems to recognize empathy in interactions—be it an interactive agent evaluating its own empathic performance or an intelligent system assessing the empathic capability of its interlocutors.
{"title":"A formal understanding of computational empathy in interactive agents","authors":"Andreas Brännström , Joel Wester , Juan Carlos Nieves","doi":"10.1016/j.cogsys.2023.101203","DOIUrl":"10.1016/j.cogsys.2023.101203","url":null,"abstract":"<div><p>Interactive software agents, such as chatbots<span>, are progressively being used in the area of health and well-being. In such applications, where agents engage with users in interpersonal conversations for, e.g., coaching, comfort or behavior-change interventions, there is an increased need for understanding agents’ empathic capabilities. In the current state-of-the-art, there are no tools to do that. In order to understand empathic capabilities in interactive software agents, we need a precise notion of empathy. The literature discusses a variety of definitions of empathy, but there is no consensus of a formal definition. Based on a systematic literature review and a qualitative analysis of recent approaches to empathy in interactive agents for health and well-being, a formal definition—an ontology—of empathy is developed. We present the potential of the formal definition in a controlled user-study by applying it as a tool for assessing empathy in two state-of-the-art health and well-being chatbots; Replika and Wysa. Our findings suggest that our definition captures necessary conditions for assessing empathy in interactive agents, and how it can uncover and explain trends in changing perceptions of empathy over time. The definition, implemented in Web Ontology Language (OWL), may serve as an automated tool, enabling systems to recognize empathy in interactions—be it an interactive agent evaluating its own empathic performance or an intelligent system assessing the empathic capability of its interlocutors.</span></p></div>","PeriodicalId":55242,"journal":{"name":"Cognitive Systems Research","volume":"85 ","pages":"Article 101203"},"PeriodicalIF":3.9,"publicationDate":"2023-12-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139070782","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-12-28DOI: 10.1016/j.cogsys.2023.101200
Rosa Cao , Daniel Yamins
Computational modeling plays an increasingly important role in neuroscience, highlighting the philosophical question of how computational models explain. In the particular case of neural network models, concerns have been raised about their intelligibility, and how these models relate (if at all) to what is found in the brain. We claim that what makes a system intelligible is an understanding of the dependencies between its behavior and the factors that are responsible for that behavior. In biology, many of these dependencies are naturally “top-down”, as ethological imperatives interact with evolutionary and developmental constraints under natural selection to produce systems with capabilities and behaviors appropriate to their evolutionary needs. We describe how the optimization techniques used to construct neural network models capture some key aspects of these dependencies, and thus help explain why brain systems are as they are — because when a challenging ecologically-relevant goal is shared by a neural network and the brain, it places constraints on the possible mechanisms exhibited in both kinds of systems. The presence and strength of these constraints explain why some outcomes are more likely than others. By combining two familiar modes of explanation — one based on bottom-up mechanistic description (whose relation to neural network models we address in a companion paper) and the other based on top-down constraints, these models have the potential to illuminate brain function.
{"title":"Explanatory models in neuroscience, Part 2: Functional intelligibility and the contravariance principle","authors":"Rosa Cao , Daniel Yamins","doi":"10.1016/j.cogsys.2023.101200","DOIUrl":"10.1016/j.cogsys.2023.101200","url":null,"abstract":"<div><p>Computational modeling plays an increasingly important role in neuroscience, highlighting the philosophical question of how computational models explain. In the particular case of neural network models, concerns have been raised about their intelligibility, and how these models relate (if at all) to what is found in the brain. We claim that what makes a system intelligible is an understanding of the dependencies between its behavior and the factors that are responsible for that behavior. In biology, many of these dependencies are naturally “top-down”, as ethological imperatives interact with evolutionary and developmental constraints under natural selection to produce systems with capabilities and behaviors appropriate to their evolutionary needs. We describe how the optimization techniques used to construct neural network models capture some key aspects of these dependencies, and thus help explain <em>why</em> brain systems are as they are — because when a challenging ecologically-relevant goal is shared by a neural network and the brain, it places constraints on the possible mechanisms exhibited in both kinds of systems. The presence and strength of these constraints explain why some outcomes are more likely than others. By combining two familiar modes of explanation — one based on bottom-up mechanistic description (whose relation to neural network models we address in a companion paper) and the other based on top-down constraints, these models have the potential to illuminate brain function.</p></div>","PeriodicalId":55242,"journal":{"name":"Cognitive Systems Research","volume":"85 ","pages":"Article 101200"},"PeriodicalIF":3.9,"publicationDate":"2023-12-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139070619","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-12-22DOI: 10.1016/j.cogsys.2023.101201
Mostafa Haghir Chehreghani
From early days, a key and controversial question inside the artificial intelligence community was whether Artificial General Intelligence (AGI) is achievable. AGI is the ability of machines and computer programs to achieve human-level intelligence and do all tasks that a human being can. While there exist a number of systems in the literature claiming they realize AGI, several other researchers argue that it is impossible to achieve it.
In this paper, we take a different view to the problem. First, we discuss that in order to realize AGI, along with building intelligent machines and programs, an intelligent world should also be constructed which is on the one hand, an accurate approximation of our world and on the other hand, a significant part of reasoning of intelligent machines is already embedded in this world. Then we discuss that AGI is not a product or algorithm, rather it is a continuous process which will become more and more mature over time (like human civilization and wisdom). Then, we argue that pre-trained embeddings play a key role in building this intelligent world and as a result, realizing AGI. We discuss how pre-trained embeddings facilitate achieving several characteristics of human-level intelligence, such as embodiment, common sense knowledge, unconscious knowledge and continuality of learning, by machines.
{"title":"The embeddings world and Artificial General Intelligence","authors":"Mostafa Haghir Chehreghani","doi":"10.1016/j.cogsys.2023.101201","DOIUrl":"10.1016/j.cogsys.2023.101201","url":null,"abstract":"<div><p><span>From early days, a key and controversial question inside the artificial intelligence community was whether </span>Artificial General Intelligence (AGI) is achievable. AGI is the ability of machines and computer programs to achieve human-level intelligence and do all tasks that a human being can. While there exist a number of systems in the literature claiming they realize AGI, several other researchers argue that it is impossible to achieve it.</p><p><span>In this paper, we take a different view to the problem. First, we discuss that in order to realize AGI, along with building intelligent machines and programs, an intelligent world should also be constructed which is on the one hand, an accurate approximation of our world and on the other hand, a significant part of reasoning of intelligent machines is already embedded in this world. Then we discuss that AGI is not a product or algorithm, rather it is a continuous process which will become more and more mature over time (like human civilization and wisdom). Then, we argue that pre-trained embeddings play a key role in building this intelligent world and as a result, realizing AGI. We discuss how pre-trained embeddings facilitate achieving several characteristics of human-level intelligence, such as embodiment, </span>common sense knowledge, unconscious knowledge and continuality of learning, by machines.</p></div>","PeriodicalId":55242,"journal":{"name":"Cognitive Systems Research","volume":"84 ","pages":"Article 101201"},"PeriodicalIF":3.9,"publicationDate":"2023-12-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139022529","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-12-16DOI: 10.1016/j.cogsys.2023.101205
Michael Sharwood Smith
Much attention has been paid to ways in which different categories of individual combine different modalities to communicate meanings to others. One major challenge that remains is to gain a deeper understanding of the cognitive processing responsible for the simultaneous deployment and integration of various different resources during multimodal communication. In response to this challenge, a Modular Cognition Framework analysis will be applied to the question of how people communicate using all the resources at their disposal. The discussion will include relevant features of this framework. Then research into similarities between humans and chimpanzee communication will be discussed followed by a comparison between verbal and non- verbal communication and a consideration of a distinct but similar approach. Emphasis will be placed on the role of meaning and on synergies between the conceptual system and the two systems responsible for linguistic structure.
{"title":"The place of language in multimodal communication in humans and other primates","authors":"Michael Sharwood Smith","doi":"10.1016/j.cogsys.2023.101205","DOIUrl":"10.1016/j.cogsys.2023.101205","url":null,"abstract":"<div><p>Much attention has been paid to ways in which different categories of individual combine different modalities to communicate meanings to others. One major challenge that remains is to gain a deeper understanding of the cognitive processing responsible for the simultaneous deployment and integration of various different resources during multimodal communication. In response to this challenge, a Modular Cognition Framework analysis will be applied to the question of how people communicate using all the resources at their disposal. The discussion will include relevant features of this framework. Then research into similarities between humans and chimpanzee communication will be discussed followed by a comparison between verbal and non- verbal communication and a consideration of a distinct but similar approach. Emphasis will be placed on the role of meaning and on synergies between the conceptual system and the two systems responsible for linguistic structure.</p></div>","PeriodicalId":55242,"journal":{"name":"Cognitive Systems Research","volume":"84 ","pages":"Article 101205"},"PeriodicalIF":3.9,"publicationDate":"2023-12-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138692610","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}