Hyperdimensional computing with holographic and adaptive encoder

Alejandro Hernández-Cano, Yang Ni, Zhuowen Zou, Ali Zakeri, Mohsen Imani
{"title":"Hyperdimensional computing with holographic and adaptive encoder","authors":"Alejandro Hernández-Cano, Yang Ni, Zhuowen Zou, Ali Zakeri, Mohsen Imani","doi":"10.3389/frai.2024.1371988","DOIUrl":null,"url":null,"abstract":"Introduction Brain-inspired computing has become an emerging field, where a growing number of works focus on developing algorithms that bring machine learning closer to human brains at the functional level. As one of the promising directions, Hyperdimensional Computing (HDC) is centered around the idea of having holographic and high-dimensional representation as the neural activities in our brains. Such representation is the fundamental enabler for the efficiency and robustness of HDC. However, existing HDC-based algorithms suffer from limitations within the encoder. To some extent, they all rely on manually selected encoders, meaning that the resulting representation is never adapted to the tasks at hand. Methods In this paper, we propose FLASH, a novel hyperdimensional learning method that incorporates an adaptive and learnable encoder design, aiming at better overall learning performance while maintaining good properties of HDC representation. Current HDC encoders leverage Random Fourier Features (RFF) for kernel correspondence and enable locality-preserving encoding. We propose to learn the encoder matrix distribution via gradient descent and effectively adapt the kernel for a more suitable HDC encoding. Results Our experiments on various regression datasets show that tuning the HDC encoder can significantly boost the accuracy, surpassing the current HDC-based algorithm and providing faster inference than other baselines, including RFF-based kernel ridge regression. Discussion The results indicate the importance of an adaptive encoder and customized high-dimensional representation in HDC.","PeriodicalId":508738,"journal":{"name":"Frontiers in Artificial Intelligence","volume":"55 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-04-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Frontiers in Artificial Intelligence","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.3389/frai.2024.1371988","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Introduction Brain-inspired computing has become an emerging field, where a growing number of works focus on developing algorithms that bring machine learning closer to human brains at the functional level. As one of the promising directions, Hyperdimensional Computing (HDC) is centered around the idea of having holographic and high-dimensional representation as the neural activities in our brains. Such representation is the fundamental enabler for the efficiency and robustness of HDC. However, existing HDC-based algorithms suffer from limitations within the encoder. To some extent, they all rely on manually selected encoders, meaning that the resulting representation is never adapted to the tasks at hand. Methods In this paper, we propose FLASH, a novel hyperdimensional learning method that incorporates an adaptive and learnable encoder design, aiming at better overall learning performance while maintaining good properties of HDC representation. Current HDC encoders leverage Random Fourier Features (RFF) for kernel correspondence and enable locality-preserving encoding. We propose to learn the encoder matrix distribution via gradient descent and effectively adapt the kernel for a more suitable HDC encoding. Results Our experiments on various regression datasets show that tuning the HDC encoder can significantly boost the accuracy, surpassing the current HDC-based algorithm and providing faster inference than other baselines, including RFF-based kernel ridge regression. Discussion The results indicate the importance of an adaptive encoder and customized high-dimensional representation in HDC.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
利用全息和自适应编码器进行超维计算
引言 脑启发计算已成为一个新兴领域,越来越多的研究工作侧重于开发在功能层面使机器学习更接近人类大脑的算法。超维计算(Hyperdimensional Computing,HDC)是其中一个前景广阔的方向,其核心思想是将我们大脑中的神经活动进行全息和高维表示。这种表征是提高 HDC 效率和鲁棒性的基础。然而,现有的基于 HDC 的算法受到编码器内部的限制。在某种程度上,它们都依赖于人工选择的编码器,这意味着生成的表示从未适应过手头的任务。方法 在本文中,我们提出了一种新颖的超维度学习方法 FLASH,它结合了自适应和可学习的编码器设计,旨在提高整体学习性能,同时保持 HDC 表示法的良好特性。当前的 HDC 编码器利用随机傅里叶特征(RFF)进行内核对应,并实现了位置保护编码。我们建议通过梯度下降来学习编码器矩阵分布,并有效地调整内核以获得更合适的 HDC 编码。结果 我们在各种回归数据集上的实验表明,调整 HDC 编码器可以显著提高准确率,超越当前基于 HDC 的算法,并提供比其他基线(包括基于 RFF 的核脊回归)更快的推理速度。讨论 结果表明了自适应编码器和定制高维表示在 HDC 中的重要性。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Using synthetic dataset for semantic segmentation of the human body in the problem of extracting anthropometric data Enhancing educational Q&A systems using a Chaotic Fuzzy Logic-Augmented large language model AI can empower agriculture for global food security: challenges and prospects in developing nations Examining the impact of green technological specialization and the integration of AI technologies on green innovation performance: evidence from China Expandable-RCNN: toward high-efficiency incremental few-shot object detection
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1