Decoupled and Explainable Associative Memory for Effective Knowledge Propagation

IF 8.9 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE IEEE transactions on neural networks and learning systems Pub Date : 2024-11-12 DOI:10.1109/TNNLS.2024.3492133
Tharindu Fernando;Darshana Priyasad;Sridha Sridharan;Clinton Fookes
{"title":"Decoupled and Explainable Associative Memory for Effective Knowledge Propagation","authors":"Tharindu Fernando;Darshana Priyasad;Sridha Sridharan;Clinton Fookes","doi":"10.1109/TNNLS.2024.3492133","DOIUrl":null,"url":null,"abstract":"Long-term memory often plays a pivotal role in human cognition through the analysis of contextual information. Machine learning researchers have attempted to emulate this process through the development of memory-augmented neural networks (MANNs) to leverage indirectly related but resourceful historical observations during learning and inference. The area of MANN, however, is still in its infancy and significant research effort is required to enable machines to achieve performance close to the human cognition process. This article presents an innovative MANN framework for the advanced incorporation of historical knowledge into a predictive framework. Within the key-value memory structure, we propose to decouple the key representations from the learned value memory embeddings to offer improved associations between the inputs and latent memory embeddings. We argue that the keys should be static, sparse, and unique representations of a particular observation to offer robust input to memory associations, while the value embeddings could be trainable, dense latent vectors such that they can better capture historical knowledge. Moreover, we introduce a novel memory update procedure that preserves the explainability of the historical knowledge extraction process, which would enable the human end-users to interpret the deep machine learning model decisions, fostering their trust. With extensive experiments conducted on three different datasets using audio, text, and image modalities, we demonstrate that our proposed innovations collectively allow this framework to outperform the current state-of-the-art methods by significant margins, irrespective of the modalities or the downstream tasks. The code is available at <uri>https://github.com/tha725/DE-KVMN/tree/main</uri>.","PeriodicalId":13303,"journal":{"name":"IEEE transactions on neural networks and learning systems","volume":"36 7","pages":"13204-13218"},"PeriodicalIF":8.9000,"publicationDate":"2024-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE transactions on neural networks and learning systems","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/10750896/","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0

Abstract

Long-term memory often plays a pivotal role in human cognition through the analysis of contextual information. Machine learning researchers have attempted to emulate this process through the development of memory-augmented neural networks (MANNs) to leverage indirectly related but resourceful historical observations during learning and inference. The area of MANN, however, is still in its infancy and significant research effort is required to enable machines to achieve performance close to the human cognition process. This article presents an innovative MANN framework for the advanced incorporation of historical knowledge into a predictive framework. Within the key-value memory structure, we propose to decouple the key representations from the learned value memory embeddings to offer improved associations between the inputs and latent memory embeddings. We argue that the keys should be static, sparse, and unique representations of a particular observation to offer robust input to memory associations, while the value embeddings could be trainable, dense latent vectors such that they can better capture historical knowledge. Moreover, we introduce a novel memory update procedure that preserves the explainability of the historical knowledge extraction process, which would enable the human end-users to interpret the deep machine learning model decisions, fostering their trust. With extensive experiments conducted on three different datasets using audio, text, and image modalities, we demonstrate that our proposed innovations collectively allow this framework to outperform the current state-of-the-art methods by significant margins, irrespective of the modalities or the downstream tasks. The code is available at https://github.com/tha725/DE-KVMN/tree/main.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
去耦合和可解释关联记忆,实现有效的知识传播
长期记忆在人类的认知过程中起着重要的作用。机器学习研究人员试图通过开发记忆增强神经网络(MANNs)来模拟这一过程,以利用学习和推理过程中间接相关但资源丰富的历史观察。然而,MANN领域仍处于起步阶段,需要大量的研究工作才能使机器实现接近人类认知过程的性能。本文提出了一个创新的MANN框架,用于将历史知识先进地纳入预测框架。在键-值记忆结构中,我们建议将键表示从学习值记忆嵌入中解耦,以改善输入和潜在记忆嵌入之间的关联。我们认为,键应该是静态的,稀疏的,并且是特定观察的唯一表示,以便为记忆关联提供鲁棒输入,而值嵌入可以是可训练的,密集的潜在向量,以便它们可以更好地捕获历史知识。此外,我们引入了一种新的记忆更新过程,该过程保留了历史知识提取过程的可解释性,这将使人类最终用户能够解释深度机器学习模型的决策,从而培养他们的信任。通过使用音频、文本和图像模式在三种不同的数据集上进行的广泛实验,我们证明了我们提出的创新共同允许该框架在很大程度上优于当前最先进的方法,无论模式或下游任务如何。代码可在https://github.com/tha725/DE-KVMN/tree/main上获得。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
IEEE transactions on neural networks and learning systems
IEEE transactions on neural networks and learning systems COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE-COMPUTER SCIENCE, HARDWARE & ARCHITECTURE
CiteScore
23.80
自引率
9.60%
发文量
2102
审稿时长
3-8 weeks
期刊介绍: The focus of IEEE Transactions on Neural Networks and Learning Systems is to present scholarly articles discussing the theory, design, and applications of neural networks as well as other learning systems. The journal primarily highlights technical and scientific research in this domain.
期刊最新文献
Neural-Adaptive Finite-Time Consensus Tracking of Singular Multiagent Systems With Actuator Attack Class-Incremental Cloud-Device Collaborative Adaptation With Contrastive Learning in Dynamic Changing Environments. ATTE: An Innovative Ensemble Learning Method With Age-Group-Guided Fine-Tuning for Improving Personalized Blood Pressure Estimation Resource-Efficient and Layer Interdependence-Aware CNN Pruning Leveraging Filter Replacement. Logits-Level Balanced Machine Unlearning for LLM-Based Recommendation System.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1