Pub Date : 2024-05-20DOI: 10.1609/aaaiss.v3i1.31211
Yuxin Zi, Kaushik Roy, Vignesh Narayanan, Amit Sheth
Despite their extensive application in language understanding tasks, large language models (LLMs) still encounter challenges including hallucinations - occasional fabrication of information - and alignment issues - lack of associations with human-curated world models (e.g., intuitive physics or common-sense knowledge). Moreover, the black-box nature of LLMs presents significant obstacles in training them effectively to achieve desired behaviors. In particular, modifying the concept embedding spaces of LLMs can be highly intractable. This process involves analyzing the implicit impact of such adjustments on the myriad parameters within LLMs and the resulting inductive biases. We propose a novel architecture that wraps powerful function approximation architectures within an outer, interpretable read-out layer. This read-out layer can be scrutinized to explicitly observe the effects of concept modeling during the training of the LLM. Our method stands in contrast with gradient-based implicit mechanisms, which depend solely on adjustments to the LLM parameters and thus evade scrutiny. By conducting extensive experiments across both generative and discriminative language modeling tasks, we evaluate the capabilities of our proposed architecture relative to state-of-the-art LLMs of similar sizes. Additionally, we offer a qualitative examination of the interpretable read-out layer and visualize the concepts it captures. The results demonstrate the potential of our approach for effectively controlling LLM hallucinations and enhancing the alignment with human expectations.
{"title":"Exploring Alternative Approaches to Language Modeling for Learning from Data and Knowledge","authors":"Yuxin Zi, Kaushik Roy, Vignesh Narayanan, Amit Sheth","doi":"10.1609/aaaiss.v3i1.31211","DOIUrl":"https://doi.org/10.1609/aaaiss.v3i1.31211","url":null,"abstract":"Despite their extensive application in language understanding tasks, large language models (LLMs) still encounter challenges including hallucinations - occasional fabrication of information - and alignment issues - lack of associations with human-curated world models (e.g., intuitive physics or common-sense knowledge). Moreover, the black-box nature of LLMs presents significant obstacles in training them effectively to achieve desired behaviors. In particular, modifying the concept embedding spaces of LLMs can be highly intractable. This process involves analyzing the implicit impact of such adjustments on the myriad parameters within LLMs and the resulting inductive biases. We propose a novel architecture that wraps powerful function approximation architectures within an outer, interpretable read-out layer. This read-out layer can be scrutinized to explicitly observe the effects of concept modeling during the training of the LLM. Our method stands in contrast with gradient-based implicit mechanisms, which depend solely on adjustments to the LLM parameters and thus evade scrutiny. By conducting extensive experiments across both generative and discriminative language modeling tasks, we evaluate the capabilities of our proposed architecture relative to state-of-the-art LLMs of similar sizes. Additionally, we offer a qualitative examination of the interpretable read-out layer and visualize the concepts it captures. The results demonstrate the potential of our approach for effectively controlling LLM hallucinations and enhancing the alignment with human expectations.","PeriodicalId":516827,"journal":{"name":"Proceedings of the AAAI Symposium Series","volume":"90 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-05-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141122776","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-05-20DOI: 10.1609/aaaiss.v3i1.31281
Viraj Parimi, Sungkweon Hong, Brian Williams
Deep Reinforcement Learning (DRL) has garnered substantial acclaim for its versatility and widespread applications across diverse domains. Aligned with human-like learning, DRL is grounded in the fundamental principle of learning from interaction, wherein agents dynamically adjust behavior based on environmental feedback in the form of rewards. This iterative trial-and-error process, mirroring human learning, underscores the importance of observation, experimentation, and feedback in shaping understanding and behavior. DRL agents, trained to navigate complex surroundings, refine their knowledge through hierarchical and abstract representations, empowered by deep neural networks. These representations enable efficient handling of long-horizon tasks and flexible adaptation to novel situations, akin to the human ability to construct mental models for comprehending complex concepts and predicting outcomes. Hence, abstract representation building emerges as a critical aspect in the learning processes of both artificial agents and human learners, particularly in long-horizon tasks. Furthermore, human decision-making, deeply rooted in evolutionary history, exhibits a remarkable capacity to balance the tradeoff between risk and cost across various domains. This cognitive process involves assessing potential negative consequences, evaluating factors such as the likelihood of adverse outcomes, severity of potential harm, and overall uncertainty. Humans intuitively gauge inherent risks and adeptly weigh associated costs, extending beyond monetary expenses to include time, effort, and opportunity costs. The nuanced ability of humans to consider the tradeoff between risk and cost highlights the complexity and adaptability of human decision-making, a skill lacking in typical DRL agents. Principles like these derived from human-like learning present an avenue for inspiring advancements in DRL, fostering the development of more adaptive and intelligent artificial agents. Motivated by these observations and focusing on practical challenges in robotics, our efforts target risk-aware stochastic sequential decision-making problem which is crucial for tasks with extended time frames and varied strategies. A novel integration of model-based conditional planning with DRL is proposed, inspired by hierarchical techniques. This approach breaks down complex tasks into manageable subtasks(motion primitives), ensuring safety constraints and informed decision-making. Unlike existing methods, our approach addresses motion primitive improvement iteratively, employing diverse prioritization functions to guide the search process effectively. This risk-bounded planning algorithm seamlessly integrates conditional planning and motion primitive learning, prioritizing computational efforts for enhanced efficiency within specified time limits.
{"title":"Task-driven Risk-bounded Hierarchical Reinforcement Learning Based on Iterative Refinement","authors":"Viraj Parimi, Sungkweon Hong, Brian Williams","doi":"10.1609/aaaiss.v3i1.31281","DOIUrl":"https://doi.org/10.1609/aaaiss.v3i1.31281","url":null,"abstract":"Deep Reinforcement Learning (DRL) has garnered substantial acclaim for its versatility and widespread applications across diverse domains. Aligned with human-like learning, DRL is grounded in the fundamental principle of learning from interaction, wherein agents dynamically adjust behavior based on environmental feedback in the form of rewards. This iterative trial-and-error process, mirroring human learning, underscores the importance of observation, experimentation, and feedback in shaping understanding and behavior. DRL agents, trained to navigate complex surroundings, refine their knowledge through hierarchical and abstract representations, empowered by deep neural networks. These representations enable efficient handling of long-horizon tasks and flexible adaptation to novel situations, akin to the human ability to construct mental models for comprehending complex concepts and predicting outcomes. Hence, abstract representation building emerges as a critical aspect in the learning processes of both artificial agents and human learners, particularly in long-horizon tasks.\u0000\u0000Furthermore, human decision-making, deeply rooted in evolutionary history, exhibits a remarkable capacity to balance the tradeoff between risk and cost across various domains. This cognitive process involves assessing potential negative consequences, evaluating factors such as the likelihood of adverse outcomes, severity of potential harm, and overall uncertainty. Humans intuitively gauge inherent risks and adeptly weigh associated costs, extending beyond monetary expenses to include time, effort, and opportunity costs. The nuanced ability of humans to consider the tradeoff between risk and cost highlights the complexity and adaptability of human decision-making, a skill lacking in typical DRL agents. Principles like these derived from human-like learning present an avenue for inspiring advancements in DRL, fostering the development of more adaptive and intelligent artificial agents.\u0000\u0000Motivated by these observations and focusing on practical challenges in robotics, our efforts target risk-aware stochastic sequential decision-making problem which is crucial for tasks with extended time frames and varied strategies. A novel integration of model-based conditional planning with DRL is proposed, inspired by hierarchical techniques. This approach breaks down complex tasks into manageable subtasks(motion primitives), ensuring safety constraints and informed decision-making. Unlike existing methods, our approach addresses motion primitive improvement iteratively, employing diverse prioritization functions to guide the search process effectively. This risk-bounded planning algorithm seamlessly integrates conditional planning and motion primitive learning, prioritizing computational efforts for enhanced efficiency within specified time limits.","PeriodicalId":516827,"journal":{"name":"Proceedings of the AAAI Symposium Series","volume":"13 18","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-05-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141119776","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Large language models (LLM) have gathered attention with the advent of ChatGPT. However, developing personalized LLM models faces challenges in real-world applications due to data scarcity and privacy concerns. Federated learning addresses these issues, providing collaborative training while preserving the client’s data. Although it has made significant progress, federated learning still faces ongoing challenges, such as communication efficiency, heterogeneous data, and privacy-preserving methods. This paper presents a novel, fully decentralized federated learning framework for LLMs to address these challenges. We utilize different blockchain-federated LLM (BC-FL) algorithms, effectively balancing the trade-off between latency and accuracy in a decentralized-federated learning environment. Additionally, we address the challenge of communication overhead in peer-to-peer networks by optimizing the path for weight transfer and mitigating node anomalies. We conducted experiments to evaluate memory usage and latency in server and serverless environments. Our results demonstrate a decrease in latency by 5X and a 13% increase in accuracy for serverless cases. Comparisons between synchronous and asynchronous scenarios revealed a 76% reduction in information passing time for the latter. The PageRank method is most efficient in eliminating anomalous nodes for better performance of the global federated LLM model. The code is available on GitHub (https://github.com/Sreebhargavibalijaa/Federated_finetuning_LLM-s_p2p_environment)
{"title":"Building Communication Efficient Asynchronous Peer-to-Peer Federated LLMs with Blockchain","authors":"Sree Bhargavi Balija, Amitash Nanda, Debashis Sahoo","doi":"10.1609/aaaiss.v3i1.31212","DOIUrl":"https://doi.org/10.1609/aaaiss.v3i1.31212","url":null,"abstract":"Large language models (LLM) have gathered attention with the advent of ChatGPT. However, developing personalized LLM models faces challenges in real-world applications due to data scarcity and privacy concerns. Federated learning addresses these issues, providing collaborative training while preserving the client’s data. Although it has made significant progress, federated learning still faces ongoing challenges, such as communication efficiency, heterogeneous data, and privacy-preserving methods. This paper presents a novel, fully decentralized federated learning framework for LLMs to address these challenges. We utilize different blockchain-federated LLM (BC-FL) algorithms, effectively balancing the trade-off between latency and accuracy in a decentralized-federated learning environment. Additionally, we address the challenge of communication overhead in peer-to-peer networks by optimizing the path for weight transfer and mitigating node anomalies. We conducted experiments to evaluate memory usage and latency in server and serverless environments. Our results demonstrate a decrease in latency by 5X and a 13% increase in accuracy for serverless cases. Comparisons between synchronous and asynchronous scenarios revealed a 76% reduction in information passing time for the latter. The PageRank method is most efficient in eliminating anomalous nodes for better performance of the global federated LLM model. The code is available on GitHub (https://github.com/Sreebhargavibalijaa/Federated_finetuning_LLM-s_p2p_environment)","PeriodicalId":516827,"journal":{"name":"Proceedings of the AAAI Symposium Series","volume":"7 2","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-05-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141119356","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-05-20DOI: 10.1609/aaaiss.v3i1.31216
Seyyedali Hosseinalipour
In this work, we aim to unveil a new class of intermediate FL architectures between centralized and decentralized schemes called “FedCrawl.” FedCrawl takes advantage of benefits of D2D communications similar to decentralized schemes; however, it uses them in a nuanced way. FedCrawl is inspired by web crawlers, which effectively explore the websites to find updated/new content posted on the internet. The cornerstone of FedCrawl is its innovative conceptualization of neural networks (NNs) or other used ML models as autonomous entities, called random walkers, with the capability to move or jump across nodes in the network through peer-to-peer (P2P) or device-to-device (D2D) connections. We introduce five research aspects to study the nuanced intricacies governing random walker behavior in these environments. The first research aspect addresses the interplay between network topology and data distribution, emphasizing the importance of considering both factors for designing efficient random walks in FedCrawl. The second research aspect explores the applicability of node importance metrics in optimizing random walker paths for FedCrawl. We propose a dynamic perception-aware design, discussed in the third research aspect, where transition matrices adapt to the evolving state of random walkers, balancing exploration and exploitation. The fourth research aspect introduces innovative features like skipping, memory look-back, and caching/trailing to enhance random walker performance. Lastly, the fifth research aspect delves into the dynamics of multiple random walkers in networked environments, introducing the concept of multi-pole random walkers. Complementing these five research aspects, we present five conjectures, each introducing novel perspectives and methodologies in the domain of decentralized learning. These conjectures encompass areas such as temperature-based characterization of random walkers and network nodes, dynamic transition matrices, non-Markovian processes, and an evolutionary framework for random walker patterns.
{"title":"Confluence of Random Walks, Interacting Particle Systems, and Distributed Machine Learning: Federated Learning through Crawling over Networks","authors":"Seyyedali Hosseinalipour","doi":"10.1609/aaaiss.v3i1.31216","DOIUrl":"https://doi.org/10.1609/aaaiss.v3i1.31216","url":null,"abstract":"In this work, we aim to unveil a new class of intermediate FL architectures between centralized and decentralized schemes called “FedCrawl.” FedCrawl takes advantage of benefits of D2D communications similar to decentralized schemes; however, it uses them in a nuanced way. FedCrawl is inspired by web crawlers, which effectively explore the websites to find updated/new content posted on the internet. The cornerstone of FedCrawl is its innovative conceptualization of neural networks (NNs) or other used ML models as autonomous entities, called random walkers, with the capability to move or jump across nodes in the network through peer-to-peer (P2P) or device-to-device (D2D) connections. We introduce five research aspects to study the nuanced intricacies governing random walker behavior in these environments. The first research aspect addresses the interplay between network topology and data distribution, emphasizing the importance of considering both factors for designing efficient random walks in FedCrawl. The second research aspect explores the applicability of node importance metrics in\u0000optimizing random walker paths for FedCrawl. We propose a dynamic perception-aware design, discussed in the third research aspect, where transition matrices adapt to the evolving state of random walkers, balancing exploration and exploitation. The fourth research aspect introduces innovative features like skipping, memory look-back, and caching/trailing to enhance random walker performance. Lastly, the fifth research aspect delves into the dynamics of multiple random walkers in networked environments, introducing the concept of multi-pole random walkers. Complementing these five research aspects, we present five conjectures, each introducing novel perspectives and methodologies in the domain of decentralized learning. These conjectures encompass areas such as temperature-based characterization of random walkers and network nodes, dynamic transition matrices, non-Markovian processes, and an evolutionary framework for random walker patterns.","PeriodicalId":516827,"journal":{"name":"Proceedings of the AAAI Symposium Series","volume":"3 6","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-05-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141121054","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-05-20DOI: 10.1609/aaaiss.v3i1.31223
J. A. Sanchez Viloria, Dimitris Stripelis, Panos P. Markopoulos, G. Sklivanitis, D. Pados
The ability to rapidly understand and label the radio spectrum in an autonomous way is key for monitoring spectrum interference, spectrum utilization efficiency, protecting passive users, monitoring and enforcing compliance with regulations, detecting faulty radios, dynamic spectrum access, opportunistic mesh networking, and numerous NextG regulatory and defense applications. We consider the problem of automatic modulation classification (AMC) by a distributed network of wireless sensors that monitor the spectrum for signal transmissions of interest over a large deployment area. Each sensor receives signals under a specific channel condition depending on its location and trains an individual model of a deep neural network (DNN) accordingly to classify signals. To improve modulation classification accuracy, we consider federated learning (FL) where each individual sensor shares its trained model with a centralized controller, which, after aggregation, initializes its model for the next round of training. Without exchanging any spectrum data (such as in cooperative spectrum sensing), this process is repeated over time. A common DNN is built across the net- work while preserving the privacy associated with signals collected at different locations. Given their distributed nature, the statistics of the data across these sensors are likely to differ significantly. We propose the use of adaptive federated learning for AMC. Specifically, we use FEDADAM -an algorithm using Adam for server optimization – and ex- amine how it compares to the FEDAVG algorithm -one of the standard FL algorithms, which averages client parameters after some local iterations, in particular in challenging scenarios that include class imbalance and/or noise-level imbalance across the network. Our extensive numerical studies over 11 standard modulation classes corroborate the merit of adaptive FL, outperforming its standard alternatives in various challenging cases and for various network sizes.
{"title":"Adaptive Federated Learning for Automatic Modulation Classification Under Class and Noise Imbalance","authors":"J. A. Sanchez Viloria, Dimitris Stripelis, Panos P. Markopoulos, G. Sklivanitis, D. Pados","doi":"10.1609/aaaiss.v3i1.31223","DOIUrl":"https://doi.org/10.1609/aaaiss.v3i1.31223","url":null,"abstract":"The ability to rapidly understand and label the radio spectrum in an autonomous way is key for monitoring spectrum interference, spectrum utilization efficiency, protecting passive users, monitoring and enforcing compliance with regulations, detecting faulty radios, dynamic spectrum access, opportunistic mesh networking, and numerous NextG regulatory and defense applications. We consider the problem of automatic modulation classification (AMC) by a distributed network of wireless sensors that monitor the spectrum for signal transmissions of interest over a large deployment area. Each sensor receives signals under a specific channel condition depending on its location and trains an individual model of a deep neural network (DNN) accordingly to classify signals. To improve modulation classification accuracy, we consider federated learning (FL) where each individual sensor shares its trained model with a centralized controller, which, after aggregation, initializes its model for the next round of training. Without exchanging any spectrum data (such as in cooperative spectrum sensing), this process is repeated over time. A common DNN is built across the net- work while preserving the privacy associated with signals collected at different locations. Given their distributed nature, the statistics of the data across these sensors are likely to differ significantly. We propose the use of adaptive federated learning for AMC. Specifically, we use FEDADAM -an algorithm using Adam for server optimization – and ex- amine how it compares to the FEDAVG algorithm -one of the standard FL algorithms, which averages client parameters after some local iterations, in particular in challenging scenarios that include class imbalance and/or noise-level imbalance across the network. Our extensive numerical studies over 11 standard modulation classes corroborate the merit of adaptive FL, outperforming its standard alternatives in various challenging cases and for various network sizes.","PeriodicalId":516827,"journal":{"name":"Proceedings of the AAAI Symposium Series","volume":"46 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-05-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141121939","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-05-20DOI: 10.1609/aaaiss.v3i1.31268
Sri Yash Tadimalla, Mary Lou Maher
The field of Artificial Intelligence (AI) is rapidly advancing, with significant potential to transform society. However, it faces a notable challenge: lack of diversity, a longstanding issue in STEM fields. In this context, this position paper examines the intersection of AI and identity as a pathway to understanding biases, inequalities, and ethical considerations in AI development and deployment. We present a multifaceted definition of AI identity, which encompasses its creators, applications, and their broader impacts. Understanding AI's identity involves analyzing the diverse individuals involved in AI's development, the technologies produced, and the social, ethical, and psychological implications. After exploring the AI identity ecosystem and its societal dynamics, We propose a framework that highlights the need for diversity in AI across three dimensions: Creators, Creations, and Consequences through the lens of identity. This paper presents a research framework for examining the implications and changes needed to foster a more inclusive and responsible AI ecosystem through the lens of identity.
{"title":"Implications of Identity in AI: Creators, Creations, and Consequences","authors":"Sri Yash Tadimalla, Mary Lou Maher","doi":"10.1609/aaaiss.v3i1.31268","DOIUrl":"https://doi.org/10.1609/aaaiss.v3i1.31268","url":null,"abstract":"The field of Artificial Intelligence (AI) is rapidly advancing, with significant potential to transform society. However, it faces a notable challenge: lack of diversity, a longstanding issue in STEM fields. In this context, this position paper examines the intersection of AI and identity as a pathway to understanding biases, inequalities, and ethical considerations in AI development and deployment. We present a multifaceted definition of AI identity, which encompasses its creators, applications, and their broader impacts. Understanding AI's identity involves analyzing the diverse individuals involved in AI's development, the technologies produced, and the social, ethical, and psychological implications. After exploring the AI identity ecosystem and its societal dynamics, We propose a framework that highlights the need for diversity in AI across three dimensions: Creators, Creations, and Consequences through the lens of identity. This paper presents a research framework for examining the implications and changes needed to foster a more inclusive and responsible AI ecosystem through the lens of identity.","PeriodicalId":516827,"journal":{"name":"Proceedings of the AAAI Symposium Series","volume":"32 11","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-05-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141118904","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-05-20DOI: 10.1609/aaaiss.v3i1.31207
Ben Schuering, Thomas Schmid
In recent years, artificial intelligence (AI) has seen significant advances that have in fact exceeded even optimistic prognoses. Using data-driven AI, namely deep learning techniques, it has been demonstrated that computers may now be equipped with abilities of remarkable scope and quality, such as solving image and text processing tasks at human level. Large language models, in particular, have sparked debates regarding opportunities and challenges of this rapidly developing area. Will remaining fundamental challenges of data-driven AI, such as factual or logical mistakes, be overcome for good if complemented and hybridized with symbolic AI techniques, such as knowledge representation and reasoning? Will systems of artificial general intelligence (AGI) emerge from this, possessing common sense and in fact completing the decades-old quest for AI that motivated the raise of the field in the 1950s? In the light of these questions, we review the likewise, decades-old philosophical debate about capabilities and limitations of computers from a hybrid AI point of view. Here, we discuss how hybrid AI is coming closer to disproving Hubert Dreyfus’ famous statements regarding what computers can not do. At the same time, we shed light on a lesser discussed challenge for hybrid AI: the possibility that its developers might be its biggest limiters.
{"title":"What Can Computers Do Now? Dreyfus Revisited for the Third Wave of Artificial Intelligence","authors":"Ben Schuering, Thomas Schmid","doi":"10.1609/aaaiss.v3i1.31207","DOIUrl":"https://doi.org/10.1609/aaaiss.v3i1.31207","url":null,"abstract":"In recent years, artificial intelligence (AI) has seen significant advances that have in fact exceeded even optimistic prognoses. Using data-driven AI, namely deep learning techniques, it has been demonstrated that computers may now be equipped with abilities of remarkable scope and quality, such as solving image and text processing tasks at human level. Large language models, in particular, have sparked debates regarding opportunities and challenges of this rapidly developing area. Will remaining fundamental challenges of data-driven AI, such as factual or logical mistakes, be overcome for good if complemented and hybridized with symbolic AI techniques, such as knowledge representation and reasoning? Will systems of artificial general intelligence (AGI) emerge from this, possessing common sense and in fact completing the decades-old quest for AI that motivated the raise of the field in the 1950s? In the light of these questions, we review the likewise, decades-old philosophical debate about capabilities and limitations of computers from a hybrid AI point of view. Here, we discuss how hybrid AI is coming closer to disproving Hubert Dreyfus’ famous statements regarding what computers can not do. At the same time, we shed light on a lesser discussed challenge for hybrid AI: the possibility that its developers might be its biggest limiters.","PeriodicalId":516827,"journal":{"name":"Proceedings of the AAAI Symposium Series","volume":"80 23","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-05-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141122914","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-05-20DOI: 10.1609/aaaiss.v3i1.31233
Sara Darwish, Alison Bragaw-Butler, Paul Marcelli, Kaylee Gassner
Artificial Intelligence (AI) adoption has seen substantial growth across industries. This paper explores the escalating use of AI within the United States Department of Defense (DoD) and the implications that diversity, equity, and inclusion (DEI) have on Service members and Civilians across the Department. More specifically, this paper explores the DEI considerations within AI technologies on individual, team, and Department readiness. The DoD's AI usage spans various strategic and operational capabilities, however this paper explores two critical domains: healthcare and recruitment. In healthcare, AI offers the promise of early disease detection, enhanced diagnostic capabilities, and streamlined administrative processes. However, potential biases stemming from homogenous training data threaten the accuracy and reliability of these systems, jeopardizing Service member health and eroding trust in AI-assisted medical decision-making and potentially the DoD at large. In recruitment, while AI promises efficiency in identifying ideal candidates, its deployment can perpetuate biases, especially when the training data used is not representative of all demographics. Despite efforts to design "unbiased" systems by excluding demographic data, such strategies may inadvertently overlook the unique challenges faced by marginalized communities, further entrenching existing disparities. Both case studies underscore the importance of considering DEI in the development and deployment of AI systems. As the DoD continues to integrate AI into its operations, this paper’s recommendations stress the necessity of continuous DEI assessment to ensure that AI serves as an asset rather than a liability. The authors recommend the following: 1. Data diversity & review 2. Continuous monitoring and calibration 3. Stakeholder engagement 4. Adoption of DEI requirements within Ethical AI Frameworks 5. Further research
人工智能(AI)的应用在各行各业都有大幅增长。本文探讨了人工智能在美国国防部(DoD)中不断升级的应用,以及多样性、公平性和包容性(DEI)对整个国防部的军人和文职人员的影响。更具体地说,本文探讨了人工智能技术对个人、团队和国防部战备状态的 DEI 影响。国防部的人工智能应用涵盖各种战略和作战能力,但本文探讨的是两个关键领域:医疗保健和征兵。在医疗保健领域,人工智能为早期疾病检测、增强诊断能力和简化管理流程带来了希望。然而,由同质化训练数据产生的潜在偏见威胁着这些系统的准确性和可靠性,危害着军人的健康,削弱了对人工智能辅助医疗决策的信任,并有可能影响整个国防部。在征兵方面,虽然人工智能有望提高识别理想候选人的效率,但其部署可能会使偏见长期存在,特别是当所使用的训练数据不能代表所有人口统计数据时。尽管努力通过排除人口数据来设计 "无偏见 "的系统,但这种策略可能会无意中忽视边缘化群体所面临的独特挑战,从而进一步巩固现有的差距。随着国防部继续将人工智能整合到其行动中,本文的建议强调有必要持续进行发展指数评估,以确保人工智能成为一种资产而非负债。作者建议如下:1. 数据多样性与审查2.持续监控和校准3.利益相关者的参与4.在人工智能道德框架内采用 DEI 要求5.进一步研究
{"title":"Diversity, Equity, and Inclusion, and the Deployment of Artificial Intelligence Within the Department of Defense","authors":"Sara Darwish, Alison Bragaw-Butler, Paul Marcelli, Kaylee Gassner","doi":"10.1609/aaaiss.v3i1.31233","DOIUrl":"https://doi.org/10.1609/aaaiss.v3i1.31233","url":null,"abstract":"Artificial Intelligence (AI) adoption has seen substantial growth across industries. This paper explores the escalating use of AI within the United States Department of Defense (DoD) and the implications that diversity, equity, and inclusion (DEI) have on Service members and Civilians across the Department. More specifically, this paper explores the DEI considerations within AI technologies on individual, team, and Department readiness. The DoD's AI usage spans various strategic and operational capabilities, however this paper explores two critical domains: healthcare and recruitment.\u0000In healthcare, AI offers the promise of early disease detection, enhanced diagnostic capabilities, and streamlined administrative processes. However, potential biases stemming from homogenous training data threaten the accuracy and reliability of these systems, jeopardizing Service member health and eroding trust in AI-assisted medical decision-making and potentially the DoD at large.\u0000In recruitment, while AI promises efficiency in identifying ideal candidates, its deployment can perpetuate biases, especially when the training data used is not representative of all demographics. Despite efforts to design \"unbiased\" systems by excluding demographic data, such strategies may inadvertently overlook the unique challenges faced by marginalized communities, further entrenching existing disparities.\u0000Both case studies underscore the importance of considering DEI in the development and deployment of AI systems. As the DoD continues to integrate AI into its operations, this paper’s recommendations stress the necessity of continuous DEI assessment to ensure that AI serves as an asset rather than a liability. The authors recommend the following:\u00001. Data diversity & review\u00002. Continuous monitoring and calibration\u00003. Stakeholder engagement\u00004. Adoption of DEI requirements within Ethical AI Frameworks\u00005. Further research","PeriodicalId":516827,"journal":{"name":"Proceedings of the AAAI Symposium Series","volume":"74 20","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-05-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141123179","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-05-20DOI: 10.1609/aaaiss.v3i1.31271
Yinuo Du, Baptiste Prébot, Cleotilde Gonzalez
During the past decade, researchers of behavioral cyber security have created cognitive agents that are able to learn and make decisions in dynamic environments in ways that assimilate human decision processes. However, many of these efforts have been limited to simple detection tasks and represent basic cognitive functions rather than a whole set of cognitive capabilities required in dynamic cyber defense scenarios. Our current work aims at advancing the development of cognitive agents that learn and make defense-dynamic decisions during cyber attacks by intelligent attack agents. We also aim to evaluate the capability of these cognitive models in ``Turing-like'' experiments, comparing the decisions and performance of these agents against human cyber defenders. In this paper, we present an initial demonstration of a cognitive model of the defender that relies on a cognitive theory of dynamic decision-making, Instance-Based Learning Theory (IBLT); we also demonstrate the execution of the same defense task by human defenders. We rely on OpenAI Gym and CybORG and adapt an existing CAGE scenario to generate a simulation experiment using an IBL defender. We also offer a new Interactive Defense Game (IDG), where textit{human} defenders can perform the same CAGE scenario simulated with the IBL model. Our results suggest that the IBL model makes decisions against two intelligent attack agents that are similar to those observed in a subsequent human experiment. We conclude with a description of the cognitive foundations required to build autonomous intelligent cyber defense agents that can collaborate with humans in autonomous cyber defense teams.
{"title":"Turing-like Experiment in a Cyber Defense Game","authors":"Yinuo Du, Baptiste Prébot, Cleotilde Gonzalez","doi":"10.1609/aaaiss.v3i1.31271","DOIUrl":"https://doi.org/10.1609/aaaiss.v3i1.31271","url":null,"abstract":"During the past decade, researchers of behavioral cyber security have created cognitive agents that are able to learn and make decisions in dynamic environments in ways that assimilate human decision processes. However, many of these efforts have been limited to simple detection tasks and represent basic cognitive functions rather than a whole set of cognitive capabilities required in dynamic cyber defense scenarios. Our current work aims at advancing the development of cognitive agents that learn and make defense-dynamic decisions during cyber attacks by intelligent attack agents. We also aim to evaluate the capability of these cognitive models in ``Turing-like'' experiments, comparing the decisions and performance of these agents against human cyber defenders. In this paper, we present an initial demonstration of a cognitive model of the defender that relies on a cognitive theory of dynamic decision-making, Instance-Based Learning Theory (IBLT); we also demonstrate the execution of the same defense task by human defenders. We rely on OpenAI Gym and CybORG and adapt an existing CAGE scenario to generate a simulation experiment using an IBL defender. We also offer a new Interactive Defense Game (IDG), where textit{human} defenders can perform the same CAGE scenario simulated with the IBL model. Our results suggest that the IBL model makes decisions against two intelligent attack agents that are similar to those observed in a subsequent human experiment. We conclude with a description of the cognitive foundations required to build autonomous intelligent cyber defense agents that can collaborate with humans in autonomous cyber defense teams.","PeriodicalId":516827,"journal":{"name":"Proceedings of the AAAI Symposium Series","volume":"39 17","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-05-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141122173","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-05-20DOI: 10.1609/aaaiss.v3i1.31187
Charles Dickens, Connor Pryor, Lise Getoor
Neural-symbolic (NeSy) AI strives to empower machine learning and large language models with fast, reliable predictions that exhibit commonsense and trustworthy reasoning by seamlessly integrating neural and symbolic methods. With such a broad scope, several taxonomies have been proposed to categorize this integration, emphasizing knowledge representation, reasoning algorithms, and applications. We introduce a knowledge representation-agnostic taxonomy focusing on the neural-symbolic interface capturing methods that reason with probability, logic, and arithmetic constraints. Moreover, we derive expressions for gradients of a prominent class of learning losses and a formalization of reasoning and learning. Through a rigorous empirical analysis spanning three tasks, we show NeSy approaches reach up to a 37% improvement over neural baselines in a semi-supervised setting and a 19% improvement over GPT-4 on question-answering.
{"title":"Modeling Patterns for Neural-Symbolic Reasoning Using Energy-based Models","authors":"Charles Dickens, Connor Pryor, Lise Getoor","doi":"10.1609/aaaiss.v3i1.31187","DOIUrl":"https://doi.org/10.1609/aaaiss.v3i1.31187","url":null,"abstract":"Neural-symbolic (NeSy) AI strives to empower machine learning and large language models with fast, reliable predictions that exhibit commonsense and trustworthy reasoning by seamlessly integrating neural and symbolic methods. With such a broad scope, several taxonomies have been proposed to categorize this integration, emphasizing knowledge representation, reasoning algorithms, and applications. We introduce a knowledge representation-agnostic taxonomy focusing on the neural-symbolic interface capturing methods that reason with probability, logic, and arithmetic constraints. Moreover, we derive expressions for gradients of a prominent class of learning losses and a formalization of reasoning and learning. Through a rigorous empirical analysis spanning three tasks, we show NeSy approaches reach up to a 37% improvement over neural baselines in a semi-supervised setting and a 19% improvement over GPT-4 on question-answering.","PeriodicalId":516827,"journal":{"name":"Proceedings of the AAAI Symposium Series","volume":"24 6","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-05-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141120864","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}