首页 > 最新文献

arXiv - CS - Neural and Evolutionary Computing最新文献

英文 中文
Active Learning for Neural PDE Solvers 神经 PDE 求解器的主动学习
Pub Date : 2024-08-02 DOI: arxiv-2408.01536
Daniel Musekamp, Marimuthu Kalimuthu, David Holzmüller, Makoto Takamoto, Mathias Niepert
Solving partial differential equations (PDEs) is a fundamental problem inengineering and science. While neural PDE solvers can be more efficient thanestablished numerical solvers, they often require large amounts of trainingdata that is costly to obtain. Active Learning (AL) could help surrogate modelsreach the same accuracy with smaller training sets by querying classicalsolvers with more informative initial conditions and PDE parameters. While ALis more common in other domains, it has yet to be studied extensively forneural PDE solvers. To bridge this gap, we introduce AL4PDE, a modular andextensible active learning benchmark. It provides multiple parametric PDEs andstate-of-the-art surrogate models for the solver-in-the-loop setting, enablingthe evaluation of existing and the development of new AL methods for PDEsolving. We use the benchmark to evaluate batch active learning algorithms suchas uncertainty- and feature-based methods. We show that AL reduces the averageerror by up to 71% compared to random sampling and significantly reducesworst-case errors. Moreover, AL generates similar datasets across repeatedruns, with consistent distributions over the PDE parameters and initialconditions. The acquired datasets are reusable, providing benefits forsurrogate models not involved in the data generation.
求解偏微分方程(PDE)是工程和科学领域的一个基本问题。虽然神经偏微分方程求解器比现有的数值求解器更高效,但它们通常需要大量的训练数据,而获取这些数据的成本很高。主动学习(Active Learning,AL)可以帮助代用模型以更小的训练集达到相同的精度,方法是用更多信息的初始条件和 PDE 参数查询经典求解器。虽然主动学习在其他领域更为常见,但在神经 PDE 求解器方面还没有广泛的研究。为了弥补这一差距,我们引入了 AL4PDE,这是一种模块化、可扩展的主动学习基准。它为解算器在环设置提供了多个参数化 PDE 和最先进的代理模型,使我们能够评估现有的 PDE 求解方法并开发新的 AL 方法。我们使用该基准来评估批量主动学习算法,如基于不确定性和特征的方法。我们的研究表明,与随机抽样相比,AL 将平均误差降低了 71%,并显著降低了最坏情况下的误差。此外,AL 还能在重复运行中生成相似的数据集,并且在 PDE 参数和初始条件上具有一致的分布。获得的数据集可重复使用,为未参与数据生成的代用模型带来了好处。
{"title":"Active Learning for Neural PDE Solvers","authors":"Daniel Musekamp, Marimuthu Kalimuthu, David Holzmüller, Makoto Takamoto, Mathias Niepert","doi":"arxiv-2408.01536","DOIUrl":"https://doi.org/arxiv-2408.01536","url":null,"abstract":"Solving partial differential equations (PDEs) is a fundamental problem in\u0000engineering and science. While neural PDE solvers can be more efficient than\u0000established numerical solvers, they often require large amounts of training\u0000data that is costly to obtain. Active Learning (AL) could help surrogate models\u0000reach the same accuracy with smaller training sets by querying classical\u0000solvers with more informative initial conditions and PDE parameters. While AL\u0000is more common in other domains, it has yet to be studied extensively for\u0000neural PDE solvers. To bridge this gap, we introduce AL4PDE, a modular and\u0000extensible active learning benchmark. It provides multiple parametric PDEs and\u0000state-of-the-art surrogate models for the solver-in-the-loop setting, enabling\u0000the evaluation of existing and the development of new AL methods for PDE\u0000solving. We use the benchmark to evaluate batch active learning algorithms such\u0000as uncertainty- and feature-based methods. We show that AL reduces the average\u0000error by up to 71% compared to random sampling and significantly reduces\u0000worst-case errors. Moreover, AL generates similar datasets across repeated\u0000runs, with consistent distributions over the PDE parameters and initial\u0000conditions. The acquired datasets are reusable, providing benefits for\u0000surrogate models not involved in the data generation.","PeriodicalId":501347,"journal":{"name":"arXiv - CS - Neural and Evolutionary Computing","volume":"8 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-08-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141945656","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
General-purpose Dataflow Model with Neuromorphic Primitives 具有神经形态基元的通用数据流模型
Pub Date : 2024-08-02 DOI: arxiv-2408.01090
Weihao Zhang, Yu Du, Hongyi Li, Songchen Ma, Rong Zhao
Neuromorphic computing exhibits great potential to provide high-performancebenefits in various applications beyond neural networks. However, ageneral-purpose program execution model that aligns with the features ofneuromorphic computing is required to bridge the gap between programversatility and neuromorphic hardware efficiency. The dataflow model offers apotential solution, but it faces high graph complexity and incompatibility withneuromorphic hardware when dealing with control flow programs, which decreasesthe programmability and performance. Here, we present a dataflow model tailoredfor neuromorphic hardware, called neuromorphic dataflow, which provides acompact, concise, and neuromorphic-compatible program representation forcontrol logic. The neuromorphic dataflow introduces "when" and "where"primitives, which restructure the view of control. The neuromorphic dataflowembeds these primitives in the dataflow schema with the plasticity inheritedfrom the spiking algorithms. Our method enables the deployment ofgeneral-purpose programs on neuromorphic hardware with both programmability andplasticity, while fully utilizing the hardware's potential.
神经形态计算在神经网络以外的各种应用中展现出提供高性能优势的巨大潜力。然而,要弥合程序通用性与神经形态硬件效率之间的差距,需要一种符合神经形态计算特点的通用程序执行模型。数据流模型提供了一种潜在的解决方案,但它在处理控制流程序时面临着高图形复杂性和与神经形态硬件不兼容的问题,从而降低了可编程性和性能。在这里,我们提出了一种为神经形态硬件量身定制的数据流模型,称为神经形态数据流,它为控制逻辑提供了一种紧凑、简洁、与神经形态兼容的程序表示法。神经形态数据流引入了 "when"(何时)和 "where"(何地)原语,重新构建了控制视图。神经形态数据流在数据流模式中嵌入了这些基元,并继承了尖峰算法的可塑性。我们的方法可以在神经形态硬件上部署通用程序,同时兼具可编程性和可塑性,充分发挥硬件的潜力。
{"title":"General-purpose Dataflow Model with Neuromorphic Primitives","authors":"Weihao Zhang, Yu Du, Hongyi Li, Songchen Ma, Rong Zhao","doi":"arxiv-2408.01090","DOIUrl":"https://doi.org/arxiv-2408.01090","url":null,"abstract":"Neuromorphic computing exhibits great potential to provide high-performance\u0000benefits in various applications beyond neural networks. However, a\u0000general-purpose program execution model that aligns with the features of\u0000neuromorphic computing is required to bridge the gap between program\u0000versatility and neuromorphic hardware efficiency. The dataflow model offers a\u0000potential solution, but it faces high graph complexity and incompatibility with\u0000neuromorphic hardware when dealing with control flow programs, which decreases\u0000the programmability and performance. Here, we present a dataflow model tailored\u0000for neuromorphic hardware, called neuromorphic dataflow, which provides a\u0000compact, concise, and neuromorphic-compatible program representation for\u0000control logic. The neuromorphic dataflow introduces \"when\" and \"where\"\u0000primitives, which restructure the view of control. The neuromorphic dataflow\u0000embeds these primitives in the dataflow schema with the plasticity inherited\u0000from the spiking algorithms. Our method enables the deployment of\u0000general-purpose programs on neuromorphic hardware with both programmability and\u0000plasticity, while fully utilizing the hardware's potential.","PeriodicalId":501347,"journal":{"name":"arXiv - CS - Neural and Evolutionary Computing","volume":"44 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-08-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141945583","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Continuous-Time Neural Networks Can Stably Memorize Random Spike Trains 连续时间神经网络能稳定记忆随机尖峰列车
Pub Date : 2024-08-02 DOI: arxiv-2408.01166
Hugo Aguettaz, Hans-Andrea Loeliger
The paper explores the capability of continuous-time recurrent neuralnetworks to store and recall precisely timed spike patterns. We show (bynumerical experiments) that this is indeed possible: within some range ofparameters, any random score of spike trains (for all neurons in the network)can be robustly memorized and autonomously reproduced with stable accuraterelative timing of all spikes, with probability close to one. We alsodemonstrate associative recall under noisy conditions. In these experiments, the required synaptic weights are computed offline, tosatisfy a template that encourages temporal stability.
本文探讨了连续时间递归神经网络存储和调用精确定时的尖峰模式的能力。我们通过数值实验证明了这确实是可能的:在一定参数范围内,(网络中所有神经元的)尖峰列车的任何随机分数都能被稳健地记忆下来,并以接近于 1 的概率自主地复制出所有尖峰的稳定精确的相对定时。我们还演示了噪声条件下的联想记忆。在这些实验中,所需的突触权重是离线计算的,以满足鼓励时间稳定性的模板。
{"title":"Continuous-Time Neural Networks Can Stably Memorize Random Spike Trains","authors":"Hugo Aguettaz, Hans-Andrea Loeliger","doi":"arxiv-2408.01166","DOIUrl":"https://doi.org/arxiv-2408.01166","url":null,"abstract":"The paper explores the capability of continuous-time recurrent neural\u0000networks to store and recall precisely timed spike patterns. We show (by\u0000numerical experiments) that this is indeed possible: within some range of\u0000parameters, any random score of spike trains (for all neurons in the network)\u0000can be robustly memorized and autonomously reproduced with stable accurate\u0000relative timing of all spikes, with probability close to one. We also\u0000demonstrate associative recall under noisy conditions. In these experiments, the required synaptic weights are computed offline, to\u0000satisfy a template that encourages temporal stability.","PeriodicalId":501347,"journal":{"name":"arXiv - CS - Neural and Evolutionary Computing","volume":"113 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-08-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141945581","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Using CSNNs to Perform Event-based Data Processing & Classification on ASL-DVS 使用 CSNN 在 ASL-DVS 上执行基于事件的数据处理和分类
Pub Date : 2024-08-01 DOI: arxiv-2408.00611
Ria Patel, Sujit Tripathy, Zachary Sublett, Seoyoung An, Riya Patel
Recent advancements in bio-inspired visual sensing and neuromorphic computinghave led to the development of various highly efficient bio-inspired solutionswith real-world applications. One notable application integrates event-basedcameras with spiking neural networks (SNNs) to process event-based sequencesthat are asynchronous and sparse, making them difficult to handle. In thisproject, we develop a convolutional spiking neural network (CSNN) architecturethat leverages convolutional operations and recurrent properties of a spikingneuron to learn the spatial and temporal relations in the ASL-DVS gesturedataset. The ASL-DVS gesture dataset is a neuromorphic dataset containing handgestures when displaying 24 letters (A to Y, excluding J and Z due to thenature of their symbols) from the American Sign Language (ASL). We performedclassification on a pre-processed subset of the full ASL-DVS dataset toidentify letter signs and achieved 100% training accuracy. Specifically, thiswas achieved by training in the Google Cloud compute platform while using alearning rate of 0.0005, batch size of 25 (total of 20 batches), 200iterations, and 10 epochs.
生物启发视觉传感和神经形态计算领域的最新进展开发出了各种具有实际应用价值的高效生物启发解决方案。其中一个值得注意的应用是将基于事件的摄像头与尖峰神经网络(SNN)集成在一起,以处理基于事件的序列,这些序列具有异步性和稀疏性,因此难以处理。在本项目中,我们开发了一种卷积尖峰神经网络(CSNN)架构,利用尖峰神经元的卷积操作和递归特性来学习 ASL-DVS 手势集中的空间和时间关系。ASL-DVS 手势数据集是一个神经形态数据集,包含显示 24 个美国手语(ASL)字母(从 A 到 Y,不包括 J 和 Z,因为它们的符号性质不同)时的手势。我们对完整 ASL-DVS 数据集的预处理子集进行了分类,以识别字母符号,训练准确率达到 100%。具体来说,这是通过在谷歌云计算平台上使用 0.0005 的学习率、25 个批次(共 20 个批次)、200 次iterations 和 10 个 epochs 进行训练实现的。
{"title":"Using CSNNs to Perform Event-based Data Processing & Classification on ASL-DVS","authors":"Ria Patel, Sujit Tripathy, Zachary Sublett, Seoyoung An, Riya Patel","doi":"arxiv-2408.00611","DOIUrl":"https://doi.org/arxiv-2408.00611","url":null,"abstract":"Recent advancements in bio-inspired visual sensing and neuromorphic computing\u0000have led to the development of various highly efficient bio-inspired solutions\u0000with real-world applications. One notable application integrates event-based\u0000cameras with spiking neural networks (SNNs) to process event-based sequences\u0000that are asynchronous and sparse, making them difficult to handle. In this\u0000project, we develop a convolutional spiking neural network (CSNN) architecture\u0000that leverages convolutional operations and recurrent properties of a spiking\u0000neuron to learn the spatial and temporal relations in the ASL-DVS gesture\u0000dataset. The ASL-DVS gesture dataset is a neuromorphic dataset containing hand\u0000gestures when displaying 24 letters (A to Y, excluding J and Z due to the\u0000nature of their symbols) from the American Sign Language (ASL). We performed\u0000classification on a pre-processed subset of the full ASL-DVS dataset to\u0000identify letter signs and achieved 100% training accuracy. Specifically, this\u0000was achieved by training in the Google Cloud compute platform while using a\u0000learning rate of 0.0005, batch size of 25 (total of 20 batches), 200\u0000iterations, and 10 epochs.","PeriodicalId":501347,"journal":{"name":"arXiv - CS - Neural and Evolutionary Computing","volume":"82 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141886829","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
High Performance Im2win and Direct Convolutions using Three Tensor Layouts on SIMD Architectures 在 SIMD 架构上使用三种张量布局实现高性能 Im2win 和直接卷积
Pub Date : 2024-08-01 DOI: arxiv-2408.00278
Xiang Fu, Xinpeng Zhang, Jixiang Ma, Peng Zhao, Shuai Lu, Xu T. Liu
Convolution is the core component within deep neural networks and it iscomputationally intensive and time consuming. Tensor data layouts significantlyimpact convolution operations in terms of memory access and computationalefficiency. Yet, there is still a lack of comprehensive performancecharacterization on data layouts on SIMD architectures concerning convolutionmethods. This paper proposes three novel data layouts for im2win convolution:NHWC, CHWN, and CHWN8, and introduces a set of general optimization techniquesfor both direct and im2win convolutions. We compare the optimized im2winconvolution with the direct convolution and PyTorch's im2col-based convolutionacross the aforementioned layouts on SIMD machines. The experimentsdemonstrated that the im2win convolution with the new NHWC layout achieved upto 355% performance speedup over NCHW layout. Our optimizations alsosignificantly improve the performance of both im2win and direct convolutions.Our optimized im2win and direct convolutions achieved up to 95% and 94% ofmachine's theoretical peak performance, respectively.
卷积是深度神经网络的核心组件,其计算密集且耗时。张量数据布局在内存访问和计算效率方面极大地影响了卷积操作。然而,关于卷积方法的 SIMD 架构上的数据布局,仍然缺乏全面的性能描述。本文提出了三种新颖的 im2win 卷积数据布局:NHWC、CHWN 和 CHWN8,并介绍了一套针对直接卷积和 im2win 卷积的通用优化技术。我们比较了经过优化的 im2win 卷积与直接卷积以及 PyTorch 基于 im2col 的卷积在 SIMD 机器上的上述布局。实验证明,采用新的 NHWC 布局的 im2win 卷积比 NCHW 布局的性能提高了 355%。我们的优化还显著提高了im2win卷积和直接卷积的性能,优化后的im2win卷积和直接卷积分别达到了机器理论峰值性能的95%和94%。
{"title":"High Performance Im2win and Direct Convolutions using Three Tensor Layouts on SIMD Architectures","authors":"Xiang Fu, Xinpeng Zhang, Jixiang Ma, Peng Zhao, Shuai Lu, Xu T. Liu","doi":"arxiv-2408.00278","DOIUrl":"https://doi.org/arxiv-2408.00278","url":null,"abstract":"Convolution is the core component within deep neural networks and it is\u0000computationally intensive and time consuming. Tensor data layouts significantly\u0000impact convolution operations in terms of memory access and computational\u0000efficiency. Yet, there is still a lack of comprehensive performance\u0000characterization on data layouts on SIMD architectures concerning convolution\u0000methods. This paper proposes three novel data layouts for im2win convolution:\u0000NHWC, CHWN, and CHWN8, and introduces a set of general optimization techniques\u0000for both direct and im2win convolutions. We compare the optimized im2win\u0000convolution with the direct convolution and PyTorch's im2col-based convolution\u0000across the aforementioned layouts on SIMD machines. The experiments\u0000demonstrated that the im2win convolution with the new NHWC layout achieved up\u0000to 355% performance speedup over NCHW layout. Our optimizations also\u0000significantly improve the performance of both im2win and direct convolutions.\u0000Our optimized im2win and direct convolutions achieved up to 95% and 94% of\u0000machine's theoretical peak performance, respectively.","PeriodicalId":501347,"journal":{"name":"arXiv - CS - Neural and Evolutionary Computing","volume":"98 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141886831","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Hilbert curves for efficient exploratory landscape analysis neighbourhood sampling 用于高效探索性景观分析邻域采样的希尔伯特曲线
Pub Date : 2024-08-01 DOI: arxiv-2408.00526
Johannes J. Pienaar, Anna S. Bosman, Katherine M. Malan
Landscape analysis aims to characterise optimisation problems based on theirobjective (or fitness) function landscape properties. The problem search spaceis typically sampled, and various landscape features are estimated based on thesamples. One particularly salient set of features is information content, whichrequires the samples to be sequences of neighbouring solutions, such that thelocal relationships between consecutive sample points are preserved. Generatingsuch spatially correlated samples that also provide good search space coverageis challenging. It is therefore common to first obtain an unordered sample withgood search space coverage, and then apply an ordering algorithm such as thenearest neighbour to minimise the distance between consecutive points in thesample. However, the nearest neighbour algorithm becomes computationallyprohibitive in higher dimensions, thus there is a need for more efficientalternatives. In this study, Hilbert space-filling curves are proposed as amethod to efficiently obtain high-quality ordered samples. Hilbert curves are aspecial case of fractal curves, and guarantee uniform coverage of a boundedsearch space while providing a spatially correlated sample. We study theeffectiveness of Hilbert curves as samplers, and discover that they are capableof extracting salient features at a fraction of the computational cost comparedto Latin hypercube sampling with post-factum ordering. Further, we investigatethe use of Hilbert curves as an ordering strategy, and find that they order thesample significantly faster than the nearest neighbour ordering, withoutsacrificing the saliency of the extracted features.
景观分析旨在根据目标(或适合度)函数的景观特性来描述优化问题。通常会对问题搜索空间进行采样,并根据样本估计各种景观特征。其中一组特别突出的特征是信息含量,它要求样本是相邻解决方案的序列,这样连续样本点之间的局部关系才能得到保留。生成这种空间相关的样本,同时还能提供良好的搜索空间覆盖率是一项挑战。因此,通常的做法是,首先获得一个具有良好搜索空间覆盖率的无序样本,然后应用近邻等排序算法,最小化样本中连续点之间的距离。然而,近邻算法在高维度下计算量过大,因此需要更高效的替代算法。在本研究中,提出了希尔伯特空间填充曲线作为高效获取高质量有序样本的方法。希尔伯特曲线是分形曲线的特例,它能保证均匀地覆盖有界搜索空间,同时提供空间相关的样本。我们研究了希尔伯特曲线作为采样器的有效性,发现与使用后事实有序的拉丁超立方采样相比,希尔伯特曲线能够以极低的计算成本提取突出特征。此外,我们还研究了希尔伯特曲线作为排序策略的使用情况,发现其排序速度明显快于近邻排序,而且不会影响所提取特征的显著性。
{"title":"Hilbert curves for efficient exploratory landscape analysis neighbourhood sampling","authors":"Johannes J. Pienaar, Anna S. Bosman, Katherine M. Malan","doi":"arxiv-2408.00526","DOIUrl":"https://doi.org/arxiv-2408.00526","url":null,"abstract":"Landscape analysis aims to characterise optimisation problems based on their\u0000objective (or fitness) function landscape properties. The problem search space\u0000is typically sampled, and various landscape features are estimated based on the\u0000samples. One particularly salient set of features is information content, which\u0000requires the samples to be sequences of neighbouring solutions, such that the\u0000local relationships between consecutive sample points are preserved. Generating\u0000such spatially correlated samples that also provide good search space coverage\u0000is challenging. It is therefore common to first obtain an unordered sample with\u0000good search space coverage, and then apply an ordering algorithm such as the\u0000nearest neighbour to minimise the distance between consecutive points in the\u0000sample. However, the nearest neighbour algorithm becomes computationally\u0000prohibitive in higher dimensions, thus there is a need for more efficient\u0000alternatives. In this study, Hilbert space-filling curves are proposed as a\u0000method to efficiently obtain high-quality ordered samples. Hilbert curves are a\u0000special case of fractal curves, and guarantee uniform coverage of a bounded\u0000search space while providing a spatially correlated sample. We study the\u0000effectiveness of Hilbert curves as samplers, and discover that they are capable\u0000of extracting salient features at a fraction of the computational cost compared\u0000to Latin hypercube sampling with post-factum ordering. Further, we investigate\u0000the use of Hilbert curves as an ordering strategy, and find that they order the\u0000sample significantly faster than the nearest neighbour ordering, without\u0000sacrificing the saliency of the extracted features.","PeriodicalId":501347,"journal":{"name":"arXiv - CS - Neural and Evolutionary Computing","volume":"17 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141886848","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SHA-CNN: Scalable Hierarchical Aware Convolutional Neural Network for Edge AI SHA-CNN:用于边缘人工智能的可扩展分层感知卷积神经网络
Pub Date : 2024-07-31 DOI: arxiv-2407.21370
Narendra Singh Dhakad, Yuvnish Malhotra, Santosh Kumar Vishvakarma, Kaushik Roy
This paper introduces a Scalable Hierarchical Aware Convolutional NeuralNetwork (SHA-CNN) model architecture for Edge AI applications. The proposedhierarchical CNN model is meticulously crafted to strike a balance betweencomputational efficiency and accuracy, addressing the challenges posed byresource-constrained edge devices. SHA-CNN demonstrates its efficacy byachieving accuracy comparable to state-of-the-art hierarchical models whileoutperforming baseline models in accuracy metrics. The key innovation lies inthe model's hierarchical awareness, enabling it to discern and prioritizerelevant features at multiple levels of abstraction. The proposed architectureclassifies data in a hierarchical manner, facilitating a nuanced understandingof complex features within the datasets. Moreover, SHA-CNN exhibits aremarkable capacity for scalability, allowing for the seamless incorporation ofnew classes. This flexibility is particularly advantageous in dynamicenvironments where the model needs to adapt to evolving datasets andaccommodate additional classes without the need for extensive retraining.Testing has been conducted on the PYNQ Z2 FPGA board to validate the proposedmodel. The results achieved an accuracy of 99.34%, 83.35%, and 63.66% forMNIST, CIFAR-10, and CIFAR-100 datasets, respectively. For CIFAR-100, ourproposed architecture performs hierarchical classification with 10% reducedcomputation while compromising only 0.7% accuracy with the state-of-the-art.The adaptability of SHA-CNN to FPGA architecture underscores its potential fordeployment in edge devices, where computational resources are limited. TheSHA-CNN framework thus emerges as a promising advancement in the intersectionof hierarchical CNNs, scalability, and FPGA-based Edge AI.
本文介绍了面向边缘人工智能应用的可扩展分层感知卷积神经网络(SHA-CNN)模型架构。所提出的分层 CNN 模型经过精心设计,在计算效率和准确性之间取得了平衡,从而应对了资源受限的边缘设备所带来的挑战。SHA-CNN 的准确度与最先进的分层模型不相上下,同时在准确度指标上优于基线模型,从而证明了它的功效。其关键创新在于模型的分层意识,使其能够在多个抽象层次上识别并优先处理相关特征。所提出的架构以分层的方式对数据进行分类,有助于深入理解数据集中的复杂特征。此外,SHA-CNN 还具有显著的可扩展性,可以无缝地纳入新的类别。这种灵活性在动态环境中尤为有利,因为在这种环境中,模型需要适应不断变化的数据集,并容纳更多的类别,而无需进行大量的重新训练。测试结果表明,MNIST、CIFAR-10 和 CIFAR-100 数据集的准确率分别为 99.34%、83.35% 和 63.66%。对于 CIFAR-100,我们提出的架构在执行分层分类时减少了 10% 的计算量,而准确率与最先进架构相比仅降低了 0.7%。SHA-CNN 对 FPGA 架构的适应性凸显了它在计算资源有限的边缘设备中部署的潜力。因此,SHA-CNN 框架是分层 CNN、可扩展性和基于 FPGA 的边缘人工智能交叉领域的一个有前途的进步。
{"title":"SHA-CNN: Scalable Hierarchical Aware Convolutional Neural Network for Edge AI","authors":"Narendra Singh Dhakad, Yuvnish Malhotra, Santosh Kumar Vishvakarma, Kaushik Roy","doi":"arxiv-2407.21370","DOIUrl":"https://doi.org/arxiv-2407.21370","url":null,"abstract":"This paper introduces a Scalable Hierarchical Aware Convolutional Neural\u0000Network (SHA-CNN) model architecture for Edge AI applications. The proposed\u0000hierarchical CNN model is meticulously crafted to strike a balance between\u0000computational efficiency and accuracy, addressing the challenges posed by\u0000resource-constrained edge devices. SHA-CNN demonstrates its efficacy by\u0000achieving accuracy comparable to state-of-the-art hierarchical models while\u0000outperforming baseline models in accuracy metrics. The key innovation lies in\u0000the model's hierarchical awareness, enabling it to discern and prioritize\u0000relevant features at multiple levels of abstraction. The proposed architecture\u0000classifies data in a hierarchical manner, facilitating a nuanced understanding\u0000of complex features within the datasets. Moreover, SHA-CNN exhibits a\u0000remarkable capacity for scalability, allowing for the seamless incorporation of\u0000new classes. This flexibility is particularly advantageous in dynamic\u0000environments where the model needs to adapt to evolving datasets and\u0000accommodate additional classes without the need for extensive retraining.\u0000Testing has been conducted on the PYNQ Z2 FPGA board to validate the proposed\u0000model. The results achieved an accuracy of 99.34%, 83.35%, and 63.66% for\u0000MNIST, CIFAR-10, and CIFAR-100 datasets, respectively. For CIFAR-100, our\u0000proposed architecture performs hierarchical classification with 10% reduced\u0000computation while compromising only 0.7% accuracy with the state-of-the-art.\u0000The adaptability of SHA-CNN to FPGA architecture underscores its potential for\u0000deployment in edge devices, where computational resources are limited. The\u0000SHA-CNN framework thus emerges as a promising advancement in the intersection\u0000of hierarchical CNNs, scalability, and FPGA-based Edge AI.","PeriodicalId":501347,"journal":{"name":"arXiv - CS - Neural and Evolutionary Computing","volume":"241 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-07-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141863431","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Interactive embodied evolution for socially adept Artificial General Creatures 善于社交的人工普通生物的交互式体现进化
Pub Date : 2024-07-31 DOI: arxiv-2407.21357
Kevin Godin-Dubois, Olivier Weissl, Karine Miras, Anna V. Kononova
We introduce here the concept of Artificial General Creatures (AGC) whichencompasses "robotic or virtual agents with a wide enough range of capabilitiesto ensure their continued survival". With this in mind, we propose a researchline aimed at incrementally building both the technology and thetrustworthiness of AGC. The core element in this approach is that trust canonly be built over time, through demonstrably mutually beneficial interactions. To this end, we advocate starting from unobtrusive, nonthreatening artificialagents that would explicitly collaborate with humans, similarly to whatdomestic animals do. By combining multiple research fields, from EvolutionaryRobotics to Neuroscience, from Ethics to Human-Machine Interaction, we aim atcreating embodied, self-sustaining Artificial General Creatures that would formsocial and emotional connections with humans. Although they would not be ableto play competitive online games or generate poems, we argue that creaturesakin to artificial pets would be invaluable stepping stones toward symbioticArtificial General Intelligence.
在此,我们提出了 "人工通用生物"(AGC)的概念,它包括 "具有足够广泛能力以确保其持续生存的机器人或虚拟代理"。有鉴于此,我们提出了一条研究路线,旨在逐步建立 AGC 的技术和可信度。这种方法的核心要素是,信任可以通过明显互利的互动逐步建立。为此,我们主张从不引人注目、不具威胁性的人工智能开始,让它们明确地与人类合作,就像家养动物所做的那样。通过结合多个研究领域,从进化机器人学到神经科学,从伦理学到人机交互学,我们的目标是创造出能够与人类建立社会和情感联系的、可体现的、可自我维持的人造生物。虽然它们不能玩网络竞技游戏,也不能创作诗歌,但我们认为,与人工宠物类似的生物将是迈向共生人工通用智能的宝贵垫脚石。
{"title":"Interactive embodied evolution for socially adept Artificial General Creatures","authors":"Kevin Godin-Dubois, Olivier Weissl, Karine Miras, Anna V. Kononova","doi":"arxiv-2407.21357","DOIUrl":"https://doi.org/arxiv-2407.21357","url":null,"abstract":"We introduce here the concept of Artificial General Creatures (AGC) which\u0000encompasses \"robotic or virtual agents with a wide enough range of capabilities\u0000to ensure their continued survival\". With this in mind, we propose a research\u0000line aimed at incrementally building both the technology and the\u0000trustworthiness of AGC. The core element in this approach is that trust can\u0000only be built over time, through demonstrably mutually beneficial interactions. To this end, we advocate starting from unobtrusive, nonthreatening artificial\u0000agents that would explicitly collaborate with humans, similarly to what\u0000domestic animals do. By combining multiple research fields, from Evolutionary\u0000Robotics to Neuroscience, from Ethics to Human-Machine Interaction, we aim at\u0000creating embodied, self-sustaining Artificial General Creatures that would form\u0000social and emotional connections with humans. Although they would not be able\u0000to play competitive online games or generate poems, we argue that creatures\u0000akin to artificial pets would be invaluable stepping stones toward symbiotic\u0000Artificial General Intelligence.","PeriodicalId":501347,"journal":{"name":"arXiv - CS - Neural and Evolutionary Computing","volume":"22 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-07-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141863432","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Lexicase-based Selection Methods with Down-sampling for Symbolic Regression Problems: Overview and Benchmark 基于词典的选择方法与符号回归问题的向下采样:概述与基准
Pub Date : 2024-07-31 DOI: arxiv-2407.21632
Alina Geiger, Dominik Sobania, Franz Rothlauf
In recent years, several new lexicase-based selection variants have emergeddue to the success of standard lexicase selection in various applicationdomains. For symbolic regression problems, variants that use anepsilon-threshold or batches of training cases, among others, have led toperformance improvements. Lately, especially variants that combine lexicaseselection and down-sampling strategies have received a lot of attention. Thispaper evaluates random as well as informed down-sampling in combination withthe relevant lexicase-based selection methods on a wide range of symbolicregression problems. In contrast to most work, we not only compare the methodsover a given evaluation budget, but also over a given time as time is usuallylimited in practice. We find that for a given evaluation budget,epsilon-lexicase selection in combination with random or informed down-samplingoutperforms all other methods. Only for a rather long running time of 24h, thebest performing method is tournament selection in combination with informeddown-sampling. If the given running time is very short, lexicase variants usingbatches of training cases perform best.
近年来,由于标准lexicase选择在不同应用领域的成功,出现了几种新的基于lexicase选择的变体。对于符号回归问题,使用epsilon阈值或成批训练案例等的变体提高了性能。最近,结合词典选择和向下抽样策略的变体受到了广泛关注。本文结合相关的基于词法的选择方法,对各种符号回归问题进行了随机和知情向下采样的评估。与大多数工作不同的是,我们不仅在给定的评估预算内比较了这些方法,而且还在给定的时间内比较了这些方法,因为在实践中时间通常是有限的。我们发现,在给定的评估预算下,ε-lexicase 选择与随机或知情向下采样相结合的方法优于所有其他方法。只有在运行时间相当长(24 小时)的情况下,表现最好的方法才是锦标赛选择与知情向下抽样相结合的方法。如果给定的运行时间很短,则使用训练案例批次的词法变体表现最佳。
{"title":"Lexicase-based Selection Methods with Down-sampling for Symbolic Regression Problems: Overview and Benchmark","authors":"Alina Geiger, Dominik Sobania, Franz Rothlauf","doi":"arxiv-2407.21632","DOIUrl":"https://doi.org/arxiv-2407.21632","url":null,"abstract":"In recent years, several new lexicase-based selection variants have emerged\u0000due to the success of standard lexicase selection in various application\u0000domains. For symbolic regression problems, variants that use an\u0000epsilon-threshold or batches of training cases, among others, have led to\u0000performance improvements. Lately, especially variants that combine lexicase\u0000selection and down-sampling strategies have received a lot of attention. This\u0000paper evaluates random as well as informed down-sampling in combination with\u0000the relevant lexicase-based selection methods on a wide range of symbolic\u0000regression problems. In contrast to most work, we not only compare the methods\u0000over a given evaluation budget, but also over a given time as time is usually\u0000limited in practice. We find that for a given evaluation budget,\u0000epsilon-lexicase selection in combination with random or informed down-sampling\u0000outperforms all other methods. Only for a rather long running time of 24h, the\u0000best performing method is tournament selection in combination with informed\u0000down-sampling. If the given running time is very short, lexicase variants using\u0000batches of training cases perform best.","PeriodicalId":501347,"journal":{"name":"arXiv - CS - Neural and Evolutionary Computing","volume":"48 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-07-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141863430","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Neuromorphic on-chip reservoir computing with spiking neural network architectures 采用尖峰神经网络架构的神经形态片上水库计算
Pub Date : 2024-07-30 DOI: arxiv-2407.20547
Samip Karki, Diego Chavez Arana, Andrew Sornborger, Francesco Caravelli
Reservoir computing is a promising approach for harnessing the computationalpower of recurrent neural networks while dramatically simplifying training.This paper investigates the application of integrate-and-fire neurons withinreservoir computing frameworks for two distinct tasks: capturing chaoticdynamics of the H'enon map and forecasting the Mackey-Glass time series.Integrate-and-fire neurons can be implemented in low-power neuromorphicarchitectures such as Intel Loihi. We explore the impact of network topologiescreated through random interactions on the reservoir's performance. Our studyreveals task-specific variations in network effectiveness, highlighting theimportance of tailored architectures for distinct computational tasks. Toidentify optimal network configurations, we employ a meta-learning approachcombined with simulated annealing. This method efficiently explores the spaceof possible network structures, identifying architectures that excel indifferent scenarios. The resulting networks demonstrate a range of behaviors,showcasing how inherent architectural features influence task-specificcapabilities. We study the reservoir computing performance using a customintegrate-and-fire code, Intel's Lava neuromorphic computing softwareframework, and via an on-chip implementation in Loihi. We conclude with ananalysis of the energy performance of the Loihi architecture.
储层计算是一种很有前途的方法,它可以利用递归神经网络的计算能力,同时大大简化训练。本文研究了在储层计算框架中应用集成-发射神经元的两个不同任务:捕捉 H'enon map 的混沌动力学和预测 Mackey-Glass 时间序列。我们探索了通过随机交互创建的网络拓扑结构对水库性能的影响。我们的研究揭示了特定任务在网络有效性方面的差异,突出了针对不同计算任务定制架构的重要性。为了确定最佳网络配置,我们采用了元学习方法与模拟退火相结合。这种方法能有效地探索可能的网络结构空间,识别出在各种情况下都表现出色的架构。由此产生的网络表现出一系列行为,展示了固有架构特性如何影响特定任务的能力。我们使用英特尔的 Lava 神经形态计算软件框架和 Loihi 中的片上实现,研究了定制的 "集成-发射 "代码的水库计算性能。最后,我们对 Loihi 架构的能耗性能进行了分析。
{"title":"Neuromorphic on-chip reservoir computing with spiking neural network architectures","authors":"Samip Karki, Diego Chavez Arana, Andrew Sornborger, Francesco Caravelli","doi":"arxiv-2407.20547","DOIUrl":"https://doi.org/arxiv-2407.20547","url":null,"abstract":"Reservoir computing is a promising approach for harnessing the computational\u0000power of recurrent neural networks while dramatically simplifying training.\u0000This paper investigates the application of integrate-and-fire neurons within\u0000reservoir computing frameworks for two distinct tasks: capturing chaotic\u0000dynamics of the H'enon map and forecasting the Mackey-Glass time series.\u0000Integrate-and-fire neurons can be implemented in low-power neuromorphic\u0000architectures such as Intel Loihi. We explore the impact of network topologies\u0000created through random interactions on the reservoir's performance. Our study\u0000reveals task-specific variations in network effectiveness, highlighting the\u0000importance of tailored architectures for distinct computational tasks. To\u0000identify optimal network configurations, we employ a meta-learning approach\u0000combined with simulated annealing. This method efficiently explores the space\u0000of possible network structures, identifying architectures that excel in\u0000different scenarios. The resulting networks demonstrate a range of behaviors,\u0000showcasing how inherent architectural features influence task-specific\u0000capabilities. We study the reservoir computing performance using a custom\u0000integrate-and-fire code, Intel's Lava neuromorphic computing software\u0000framework, and via an on-chip implementation in Loihi. We conclude with an\u0000analysis of the energy performance of the Loihi architecture.","PeriodicalId":501347,"journal":{"name":"arXiv - CS - Neural and Evolutionary Computing","volume":"75 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-07-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141863435","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
arXiv - CS - Neural and Evolutionary Computing
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1