首页 > 最新文献

arXiv - CS - Neural and Evolutionary Computing最新文献

英文 中文
ORS: A novel Olive Ridley Survival inspired Meta-heuristic Optimization Algorithm ORS:新颖的 Olive Ridley 生存启发元启发式优化算法
Pub Date : 2024-09-13 DOI: arxiv-2409.09210
Niranjan Panigrahi, Sourav Kumar Bhoi, Debasis Mohapatra, Rashmi Ranjan Sahoo, Kshira Sagar Sahoo, Anil Mohapatra
Meta-heuristic algorithmic development has been a thrust area of researchsince its inception. In this paper, a novel meta-heuristic optimizationalgorithm, Olive Ridley Survival (ORS), is proposed which is inspired fromsurvival challenges faced by hatchlings of Olive Ridley sea turtle. A majorfact about survival of Olive Ridley reveals that out of one thousand OliveRidley hatchlings which emerge from nest, only one survive at sea due tovarious environmental and other factors. This fact acts as the backbone fordeveloping the proposed algorithm. The algorithm has two major phases:hatchlings survival through environmental factors and impact of movementtrajectory on its survival. The phases are mathematically modelled andimplemented along with suitable input representation and fitness function. Thealgorithm is analysed theoretically. To validate the algorithm, fourteenmathematical benchmark functions from standard CEC test suites are evaluatedand statistically tested. Also, to study the efficacy of ORS on recent complexbenchmark functions, ten benchmark functions of CEC-06-2019 are evaluated.Further, three well-known engineering problems are solved by ORS and comparedwith other state-of-the-art meta-heuristics. Simulation results show that inmany cases, the proposed ORS algorithm outperforms some state-of-the-artmeta-heuristic optimization algorithms. The sub-optimal behavior of ORS in somerecent benchmark functions is also observed.
元启发式算法自诞生以来一直是研究的重点领域。本文提出了一种新颖的元启发式优化算法--Olive Ridley Survival(ORS),其灵感来源于 Olive Ridley 海龟幼体面临的生存挑战。有关 Olive Ridley 海龟生存的一个重要事实表明,由于各种环境和其他因素,在一千只出巢的 Olive Ridley 海龟幼体中,只有一只能在海上存活。这一事实是开发拟议算法的基础。该算法分为两个主要阶段:幼鸟在环境因素中的存活率和运动轨迹对其存活率的影响。这两个阶段通过数学模型和适当的输入表示和适应度函数得以实现。对算法进行了理论分析。为了验证该算法,对标准 CEC 测试套件中的 14 个数学基准函数进行了评估和统计测试。此外,为了研究 ORS 对最新复杂基准函数的功效,还评估了 CEC-06-2019 中的十个基准函数。仿真结果表明,在许多情况下,所提出的 ORS 算法优于一些最先进的元启发式优化算法。此外,还观察到 ORS 在某些最新基准函数中的次优行为。
{"title":"ORS: A novel Olive Ridley Survival inspired Meta-heuristic Optimization Algorithm","authors":"Niranjan Panigrahi, Sourav Kumar Bhoi, Debasis Mohapatra, Rashmi Ranjan Sahoo, Kshira Sagar Sahoo, Anil Mohapatra","doi":"arxiv-2409.09210","DOIUrl":"https://doi.org/arxiv-2409.09210","url":null,"abstract":"Meta-heuristic algorithmic development has been a thrust area of research\u0000since its inception. In this paper, a novel meta-heuristic optimization\u0000algorithm, Olive Ridley Survival (ORS), is proposed which is inspired from\u0000survival challenges faced by hatchlings of Olive Ridley sea turtle. A major\u0000fact about survival of Olive Ridley reveals that out of one thousand Olive\u0000Ridley hatchlings which emerge from nest, only one survive at sea due to\u0000various environmental and other factors. This fact acts as the backbone for\u0000developing the proposed algorithm. The algorithm has two major phases:\u0000hatchlings survival through environmental factors and impact of movement\u0000trajectory on its survival. The phases are mathematically modelled and\u0000implemented along with suitable input representation and fitness function. The\u0000algorithm is analysed theoretically. To validate the algorithm, fourteen\u0000mathematical benchmark functions from standard CEC test suites are evaluated\u0000and statistically tested. Also, to study the efficacy of ORS on recent complex\u0000benchmark functions, ten benchmark functions of CEC-06-2019 are evaluated.\u0000Further, three well-known engineering problems are solved by ORS and compared\u0000with other state-of-the-art meta-heuristics. Simulation results show that in\u0000many cases, the proposed ORS algorithm outperforms some state-of-the-art\u0000meta-heuristic optimization algorithms. The sub-optimal behavior of ORS in some\u0000recent benchmark functions is also observed.","PeriodicalId":501347,"journal":{"name":"arXiv - CS - Neural and Evolutionary Computing","volume":"13 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142248990","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Training Spiking Neural Networks via Augmented Direct Feedback Alignment 通过增强直接反馈排列训练尖峰神经网络
Pub Date : 2024-09-12 DOI: arxiv-2409.07776
Yongbo Zhang, Katsuma Inoue, Mitsumasa Nakajima, Toshikazu Hashimoto, Yasuo Kuniyoshi, Kohei Nakajima
Spiking neural networks (SNNs), the models inspired by the mechanisms of realneurons in the brain, transmit and represent information by employing discreteaction potentials or spikes. The sparse, asynchronous properties of informationprocessing make SNNs highly energy efficient, leading to SNNs being promisingsolutions for implementing neural networks in neuromorphic devices. However,the nondifferentiable nature of SNN neurons makes it a challenge to train them.The current training methods of SNNs that are based on error backpropagation(BP) and precisely designing surrogate gradient are difficult to implement andbiologically implausible, hindering the implementation of SNNs on neuromorphicdevices. Thus, it is important to train SNNs with a method that is bothphysically implementatable and biologically plausible. In this paper, wepropose using augmented direct feedback alignment (aDFA), a gradient-freeapproach based on random projection, to train SNNs. This method requires onlypartial information of the forward process during training, so it is easy toimplement and biologically plausible. We systematically demonstrate thefeasibility of the proposed aDFA-SNNs scheme, propose its effective workingrange, and analyze its well-performing settings by employing genetic algorithm.We also analyze the impact of crucial features of SNNs on the scheme, thusdemonstrating its superiority and stability over BP and conventional directfeedback alignment. Our scheme can achieve competitive performance withoutaccurate prior knowledge about the utilized system, thus providing a valuablereference for physically training SNNs.
尖峰神经网络(SNN)是受大脑中真实神经元机制启发而建立的模型,通过使用离散动作电位或尖峰来传输和表示信息。信息处理的稀疏性和异步性使 SNN 具有很高的能效,因此 SNN 有望成为在神经形态设备中实现神经网络的解决方案。目前基于误差反向传播(BP)和精确设计替代梯度的 SNNs 训练方法难以实现,而且在生物学上难以置信,阻碍了 SNNs 在神经形态设备上的实现。因此,使用一种既能在物理学上实现,又能在生物学上合理的方法来训练 SNN 是非常重要的。在本文中,我们提出使用增强直接反馈对齐(aDFA)来训练 SNN,这是一种基于随机投影的无梯度方法。这种方法在训练过程中只需要前向过程的部分信息,因此易于实现,在生物学上也是可行的。我们系统地证明了所提出的 aDFA-SNNs 方案的可行性,提出了其有效的工作范围,并通过遗传算法分析了其性能良好的设置,还分析了 SNNs 的关键特征对该方案的影响,从而证明了其优于 BP 和传统直接反馈配准的稳定性。我们的方案可以在没有关于所用系统的准确先验知识的情况下实现具有竞争力的性能,从而为物理训练 SNNs 提供了有价值的参考。
{"title":"Training Spiking Neural Networks via Augmented Direct Feedback Alignment","authors":"Yongbo Zhang, Katsuma Inoue, Mitsumasa Nakajima, Toshikazu Hashimoto, Yasuo Kuniyoshi, Kohei Nakajima","doi":"arxiv-2409.07776","DOIUrl":"https://doi.org/arxiv-2409.07776","url":null,"abstract":"Spiking neural networks (SNNs), the models inspired by the mechanisms of real\u0000neurons in the brain, transmit and represent information by employing discrete\u0000action potentials or spikes. The sparse, asynchronous properties of information\u0000processing make SNNs highly energy efficient, leading to SNNs being promising\u0000solutions for implementing neural networks in neuromorphic devices. However,\u0000the nondifferentiable nature of SNN neurons makes it a challenge to train them.\u0000The current training methods of SNNs that are based on error backpropagation\u0000(BP) and precisely designing surrogate gradient are difficult to implement and\u0000biologically implausible, hindering the implementation of SNNs on neuromorphic\u0000devices. Thus, it is important to train SNNs with a method that is both\u0000physically implementatable and biologically plausible. In this paper, we\u0000propose using augmented direct feedback alignment (aDFA), a gradient-free\u0000approach based on random projection, to train SNNs. This method requires only\u0000partial information of the forward process during training, so it is easy to\u0000implement and biologically plausible. We systematically demonstrate the\u0000feasibility of the proposed aDFA-SNNs scheme, propose its effective working\u0000range, and analyze its well-performing settings by employing genetic algorithm.\u0000We also analyze the impact of crucial features of SNNs on the scheme, thus\u0000demonstrating its superiority and stability over BP and conventional direct\u0000feedback alignment. Our scheme can achieve competitive performance without\u0000accurate prior knowledge about the utilized system, thus providing a valuable\u0000reference for physically training SNNs.","PeriodicalId":501347,"journal":{"name":"arXiv - CS - Neural and Evolutionary Computing","volume":"37 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142188211","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Classifying Images with CoLaNET Spiking Neural Network -- the MNIST Example 使用 CoLaNET 尖峰神经网络进行图像分类 - MNIST 示例
Pub Date : 2024-09-12 DOI: arxiv-2409.07833
Mikhail Kiselev
In the present paper, it is shown how the columnar/layered CoLaNET spikingneural network (SNN) architecture can be used in supervised learning imageclassification tasks. Image pixel brightness is coded by the spike count duringimage presentation period. Image class label is indicated by activity ofspecial SNN input nodes (one node per class). The CoLaNET classificationaccuracy is evaluated on the MNIST benchmark. It is demonstrated that CoLaNETis almost as accurate as the most advanced machine learning algorithms (notusing convolutional approach).
本文展示了柱状/层状 CoLaNET 尖峰神经网络(SNN)架构如何用于监督学习图像分类任务。图像像素亮度由图像呈现期间的尖峰计数编码。图像类别标签由特殊 SNN 输入节点(每个类别一个节点)的活动指示。CoLaNET 的分类准确率在 MNIST 基准上进行了评估。结果表明,CoLaNET 的准确度几乎与最先进的机器学习算法(不使用卷积方法)相当。
{"title":"Classifying Images with CoLaNET Spiking Neural Network -- the MNIST Example","authors":"Mikhail Kiselev","doi":"arxiv-2409.07833","DOIUrl":"https://doi.org/arxiv-2409.07833","url":null,"abstract":"In the present paper, it is shown how the columnar/layered CoLaNET spiking\u0000neural network (SNN) architecture can be used in supervised learning image\u0000classification tasks. Image pixel brightness is coded by the spike count during\u0000image presentation period. Image class label is indicated by activity of\u0000special SNN input nodes (one node per class). The CoLaNET classification\u0000accuracy is evaluated on the MNIST benchmark. It is demonstrated that CoLaNET\u0000is almost as accurate as the most advanced machine learning algorithms (not\u0000using convolutional approach).","PeriodicalId":501347,"journal":{"name":"arXiv - CS - Neural and Evolutionary Computing","volume":"13 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142188210","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Optimizing Neural Network Performance and Interpretability with Diophantine Equation Encoding 用二叉方程编码优化神经网络性能和可解释性
Pub Date : 2024-09-11 DOI: arxiv-2409.07310
Ronald Katende
This paper explores the integration of Diophantine equations into neuralnetwork (NN) architectures to improve model interpretability, stability, andefficiency. By encoding and decoding neural network parameters as integersolutions to Diophantine equations, we introduce a novel approach that enhancesboth the precision and robustness of deep learning models. Our methodintegrates a custom loss function that enforces Diophantine constraints duringtraining, leading to better generalization, reduced error bounds, and enhancedresilience against adversarial attacks. We demonstrate the efficacy of thisapproach through several tasks, including image classification and naturallanguage processing, where improvements in accuracy, convergence, androbustness are observed. This study offers a new perspective on combiningmathematical theory and machine learning to create more interpretable andefficient models.
本文探讨了如何将 Diophantine 方程整合到神经网络(NN)架构中,以提高模型的可解释性、稳定性和效率。通过将神经网络参数编码和解码为 Diophantine 方程的整数解,我们引入了一种新颖的方法来提高深度学习模型的精度和鲁棒性。我们的方法集成了一个自定义损失函数,在训练过程中强制执行 Diophantine 约束,从而实现更好的泛化、降低误差边界,并增强对对抗性攻击的复原力。我们通过包括图像分类和自然语言处理在内的几项任务证明了这种方法的有效性,在准确性、收敛性和稳健性方面都有所改进。这项研究为数学理论与机器学习的结合提供了一个新的视角,以创建更可解释、更高效的模型。
{"title":"Optimizing Neural Network Performance and Interpretability with Diophantine Equation Encoding","authors":"Ronald Katende","doi":"arxiv-2409.07310","DOIUrl":"https://doi.org/arxiv-2409.07310","url":null,"abstract":"This paper explores the integration of Diophantine equations into neural\u0000network (NN) architectures to improve model interpretability, stability, and\u0000efficiency. By encoding and decoding neural network parameters as integer\u0000solutions to Diophantine equations, we introduce a novel approach that enhances\u0000both the precision and robustness of deep learning models. Our method\u0000integrates a custom loss function that enforces Diophantine constraints during\u0000training, leading to better generalization, reduced error bounds, and enhanced\u0000resilience against adversarial attacks. We demonstrate the efficacy of this\u0000approach through several tasks, including image classification and natural\u0000language processing, where improvements in accuracy, convergence, and\u0000robustness are observed. This study offers a new perspective on combining\u0000mathematical theory and machine learning to create more interpretable and\u0000efficient models.","PeriodicalId":501347,"journal":{"name":"arXiv - CS - Neural and Evolutionary Computing","volume":"7 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142188212","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Y-Drop: A Conductance based Dropout for fully connected layers Y-Drop:基于电导的全连接层滤除器
Pub Date : 2024-09-11 DOI: arxiv-2409.09088
Efthymios Georgiou, Georgios Paraskevopoulos, Alexandros Potamianos
In this work, we introduce Y-Drop, a regularization method that biases thedropout algorithm towards dropping more important neurons with higherprobability. The backbone of our approach is neuron conductance, aninterpretable measure of neuron importance that calculates the contribution ofeach neuron towards the end-to-end mapping of the network. We investigate theimpact of the uniform dropout selection criterion on performance by assigninghigher dropout probability to the more important units. We show that forcingthe network to solve the task at hand in the absence of its important unitsyields a strong regularization effect. Further analysis indicates that Y-Dropyields solutions where more neurons are important, i.e have high conductance,and yields robust networks. In our experiments we show that the regularizationeffect of Y-Drop scales better than vanilla dropout w.r.t. the architecturesize and consistently yields superior performance over multiple datasets andarchitecture combinations, with little tuning.
在这项工作中,我们引入了 Y-Drop,这是一种正则化方法,它能使丢弃算法偏向于以更高的概率丢弃更重要的神经元。我们方法的支柱是神经元电导,这是一种可解释的神经元重要性度量,它计算每个神经元对网络端到端映射的贡献。我们通过为更重要的单元分配更高的辍学概率,研究了均匀辍学选择标准对性能的影响。我们发现,迫使网络在没有重要单元的情况下解决手头的任务会产生很强的正则化效应。进一步的分析表明,Y-正则化能产生更多重要神经元(即具有高传导性)的解决方案,并产生稳健的网络。在实验中,我们发现 Y-Drop 的正则化效果比 vanilla dropout 更好地扩展了架构规模,而且在多个数据集和架构组合中,Y-Drop 只需进行少量调整,就能始终如一地获得卓越性能。
{"title":"Y-Drop: A Conductance based Dropout for fully connected layers","authors":"Efthymios Georgiou, Georgios Paraskevopoulos, Alexandros Potamianos","doi":"arxiv-2409.09088","DOIUrl":"https://doi.org/arxiv-2409.09088","url":null,"abstract":"In this work, we introduce Y-Drop, a regularization method that biases the\u0000dropout algorithm towards dropping more important neurons with higher\u0000probability. The backbone of our approach is neuron conductance, an\u0000interpretable measure of neuron importance that calculates the contribution of\u0000each neuron towards the end-to-end mapping of the network. We investigate the\u0000impact of the uniform dropout selection criterion on performance by assigning\u0000higher dropout probability to the more important units. We show that forcing\u0000the network to solve the task at hand in the absence of its important units\u0000yields a strong regularization effect. Further analysis indicates that Y-Drop\u0000yields solutions where more neurons are important, i.e have high conductance,\u0000and yields robust networks. In our experiments we show that the regularization\u0000effect of Y-Drop scales better than vanilla dropout w.r.t. the architecture\u0000size and consistently yields superior performance over multiple datasets and\u0000architecture combinations, with little tuning.","PeriodicalId":501347,"journal":{"name":"arXiv - CS - Neural and Evolutionary Computing","volume":"190 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142248997","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Comprehensive Comparison Between ANNs and KANs For Classifying EEG Alzheimer's Data ANN 与 KAN 在脑电图阿尔茨海默病数据分类方面的综合比较
Pub Date : 2024-09-09 DOI: arxiv-2409.05989
Akshay Sunkara, Sriram Sattiraju, Aakarshan Kumar, Zaryab Kanjiani, Himesh Anumala
Alzheimer's Disease is an incurable cognitive condition that affectsthousands of people globally. While some diagnostic methods exist forAlzheimer's Disease, many of these methods cannot detect Alzheimer's in itsearlier stages. Recently, researchers have explored the use ofElectroencephalogram (EEG) technology for diagnosing Alzheimer's. EEG is anoninvasive method of recording the brain's electrical signals, and EEG datahas shown distinct differences between patients with and without Alzheimer's.In the past, Artificial Neural Networks (ANNs) have been used to predictAlzheimer's from EEG data, but these models sometimes produce false positivediagnoses. This study aims to compare losses between ANNs and Kolmogorov-ArnoldNetworks (KANs) across multiple types of epochs, learning rates, and nodes. Theresults show that across these different parameters, ANNs are more accurate inpredicting Alzheimer's Disease from EEG signals.
阿尔茨海默病是一种无法治愈的认知疾病,影响着全球成千上万的人。虽然目前已有一些诊断阿尔茨海默病的方法,但其中许多方法无法检测到早期阶段的阿尔茨海默病。最近,研究人员探索使用脑电图(EEG)技术诊断阿尔茨海默病。脑电图是一种记录大脑电信号的非侵入性方法,脑电图数据显示阿尔茨海默病患者和非阿尔茨海默病患者之间存在明显差异。过去,人工神经网络(ANN)曾被用于从脑电图数据中预测阿尔茨海默病,但这些模型有时会产生误诊。本研究旨在比较人工神经网络和柯尔莫哥洛夫-阿诺德网络(KAN)在不同类型的历时、学习率和节点上的损失。结果表明,在这些不同的参数中,ANN 在从脑电图信号预测阿尔茨海默病方面更为准确。
{"title":"A Comprehensive Comparison Between ANNs and KANs For Classifying EEG Alzheimer's Data","authors":"Akshay Sunkara, Sriram Sattiraju, Aakarshan Kumar, Zaryab Kanjiani, Himesh Anumala","doi":"arxiv-2409.05989","DOIUrl":"https://doi.org/arxiv-2409.05989","url":null,"abstract":"Alzheimer's Disease is an incurable cognitive condition that affects\u0000thousands of people globally. While some diagnostic methods exist for\u0000Alzheimer's Disease, many of these methods cannot detect Alzheimer's in its\u0000earlier stages. Recently, researchers have explored the use of\u0000Electroencephalogram (EEG) technology for diagnosing Alzheimer's. EEG is a\u0000noninvasive method of recording the brain's electrical signals, and EEG data\u0000has shown distinct differences between patients with and without Alzheimer's.\u0000In the past, Artificial Neural Networks (ANNs) have been used to predict\u0000Alzheimer's from EEG data, but these models sometimes produce false positive\u0000diagnoses. This study aims to compare losses between ANNs and Kolmogorov-Arnold\u0000Networks (KANs) across multiple types of epochs, learning rates, and nodes. The\u0000results show that across these different parameters, ANNs are more accurate in\u0000predicting Alzheimer's Disease from EEG signals.","PeriodicalId":501347,"journal":{"name":"arXiv - CS - Neural and Evolutionary Computing","volume":"160 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142188213","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Advancing Automated Knowledge Transfer in Evolutionary Multitasking via Large Language Models 通过大型语言模型推进进化多任务中的自动知识转移
Pub Date : 2024-09-06 DOI: arxiv-2409.04270
Yuxiao Huang, Xuebin Lv, Shenghao Wu, Jibin Wu, Liang Feng, Kay Chen Tan
Evolutionary Multi-task Optimization (EMTO) is a paradigm that leveragesknowledge transfer across simultaneously optimized tasks for enhanced searchperformance. To facilitate EMTO's performance, various knowledge transfermodels have been developed for specific optimization tasks. However, designingthese models often requires substantial expert knowledge. Recently, largelanguage models (LLMs) have achieved remarkable success in autonomousprogramming, aiming to produce effective solvers for specific problems. In thiswork, a LLM-based optimization paradigm is introduced to establish anautonomous model factory for generating knowledge transfer models, ensuringeffective and efficient knowledge transfer across various optimization tasks.To evaluate the performance of the proposed method, we conducted comprehensiveempirical studies comparing the knowledge transfer model generated by the LLMwith existing state-of-the-art knowledge transfer methods. The resultsdemonstrate that the generated model is able to achieve superior or competitiveperformance against hand-crafted knowledge transfer models in terms of bothefficiency and effectiveness.
进化多任务优化(EMTO)是一种利用跨同时优化任务的知识转移来提高搜索性能的范式。为了提高 EMTO 的性能,针对特定优化任务开发了各种知识转移模型。然而,设计这些模型往往需要大量的专家知识。最近,大型语言模型(LLM)在自主编程方面取得了显著的成功,其目的是为特定问题生成有效的求解器。为了评估所提出方法的性能,我们进行了全面的实证研究,将 LLM 生成的知识转移模型与现有最先进的知识转移方法进行了比较。研究结果表明,与手工创建的知识转移模型相比,LLM 生成的知识转移模型在效率和效果方面都具有优势或竞争力。
{"title":"Advancing Automated Knowledge Transfer in Evolutionary Multitasking via Large Language Models","authors":"Yuxiao Huang, Xuebin Lv, Shenghao Wu, Jibin Wu, Liang Feng, Kay Chen Tan","doi":"arxiv-2409.04270","DOIUrl":"https://doi.org/arxiv-2409.04270","url":null,"abstract":"Evolutionary Multi-task Optimization (EMTO) is a paradigm that leverages\u0000knowledge transfer across simultaneously optimized tasks for enhanced search\u0000performance. To facilitate EMTO's performance, various knowledge transfer\u0000models have been developed for specific optimization tasks. However, designing\u0000these models often requires substantial expert knowledge. Recently, large\u0000language models (LLMs) have achieved remarkable success in autonomous\u0000programming, aiming to produce effective solvers for specific problems. In this\u0000work, a LLM-based optimization paradigm is introduced to establish an\u0000autonomous model factory for generating knowledge transfer models, ensuring\u0000effective and efficient knowledge transfer across various optimization tasks.\u0000To evaluate the performance of the proposed method, we conducted comprehensive\u0000empirical studies comparing the knowledge transfer model generated by the LLM\u0000with existing state-of-the-art knowledge transfer methods. The results\u0000demonstrate that the generated model is able to achieve superior or competitive\u0000performance against hand-crafted knowledge transfer models in terms of both\u0000efficiency and effectiveness.","PeriodicalId":501347,"journal":{"name":"arXiv - CS - Neural and Evolutionary Computing","volume":"23 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142188234","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Pareto Set Prediction Assisted Bilevel Multi-objective Optimization 帕累托集合预测辅助双层多目标优化
Pub Date : 2024-09-05 DOI: arxiv-2409.03328
Bing Wang, Hemant K. Singh, Tapabrata Ray
Bilevel optimization problems comprise an upper level optimization task thatcontains a lower level optimization task as a constraint. While there is asignificant and growing literature devoted to solving bilevel problems withsingle objective at both levels using evolutionary computation, there isrelatively scarce work done to address problems with multiple objectives(BLMOP) at both levels. For black-box BLMOPs, the existing evolutionarytechniques typically utilize nested search, which in its native form consumeslarge number of function evaluations. In this work, we propose to reduce thisexpense by predicting the lower level Pareto set for a candidate upper levelsolution directly, instead of conducting an optimization from scratch. Such aprediction is significantly challenging for BLMOPs as it involves one-to-manymapping scenario. We resolve this bottleneck by supplementing the dataset usinga helper variable and construct a neural network, which can then be trained tomap the variables in a meaningful manner. Then, we embed this initializationwithin a bilevel optimization framework, termed Pareto set prediction assistedevolutionary bilevel multi-objective optimization (PSP-BLEMO). Systematicexperiments with existing state-of-the-art methods are presented to demonstrateits benefit. The experiments show that the proposed approach is competitiveacross a range of problems, including both deceptive and non-deceptive problems
双层优化问题包括一个上层优化任务,该任务包含一个下层优化任务作为约束条件。虽然利用进化计算解决双层单一目标问题的文献数量可观且在不断增加,但解决双层多目标(BLMOP)问题的文献却相对较少。对于黑盒子 BLMOP,现有的进化技术通常使用嵌套搜索,其原始形式会消耗大量的函数评估。在这项工作中,我们建议直接预测候选上层解决方案的下层帕累托集合,而不是从头开始优化,从而减少这种消耗。这种预测对于 BLMOPs 来说具有很大的挑战性,因为它涉及一对多的映射场景。为了解决这一瓶颈,我们使用辅助变量对数据集进行补充,并构建一个神经网络,然后对其进行训练,使其能够以有意义的方式映射变量。然后,我们将这一初始化嵌入到双层优化框架中,即帕累托集预测辅助进化双层多目标优化(PSP-BLEMO)。为了证明这种方法的优势,我们对现有的最先进方法进行了系统实验。实验表明,所提出的方法在包括欺骗性和非欺骗性问题在内的一系列问题上都具有竞争力。
{"title":"Pareto Set Prediction Assisted Bilevel Multi-objective Optimization","authors":"Bing Wang, Hemant K. Singh, Tapabrata Ray","doi":"arxiv-2409.03328","DOIUrl":"https://doi.org/arxiv-2409.03328","url":null,"abstract":"Bilevel optimization problems comprise an upper level optimization task that\u0000contains a lower level optimization task as a constraint. While there is a\u0000significant and growing literature devoted to solving bilevel problems with\u0000single objective at both levels using evolutionary computation, there is\u0000relatively scarce work done to address problems with multiple objectives\u0000(BLMOP) at both levels. For black-box BLMOPs, the existing evolutionary\u0000techniques typically utilize nested search, which in its native form consumes\u0000large number of function evaluations. In this work, we propose to reduce this\u0000expense by predicting the lower level Pareto set for a candidate upper level\u0000solution directly, instead of conducting an optimization from scratch. Such a\u0000prediction is significantly challenging for BLMOPs as it involves one-to-many\u0000mapping scenario. We resolve this bottleneck by supplementing the dataset using\u0000a helper variable and construct a neural network, which can then be trained to\u0000map the variables in a meaningful manner. Then, we embed this initialization\u0000within a bilevel optimization framework, termed Pareto set prediction assisted\u0000evolutionary bilevel multi-objective optimization (PSP-BLEMO). Systematic\u0000experiments with existing state-of-the-art methods are presented to demonstrate\u0000its benefit. The experiments show that the proposed approach is competitive\u0000across a range of problems, including both deceptive and non-deceptive problems","PeriodicalId":501347,"journal":{"name":"arXiv - CS - Neural and Evolutionary Computing","volume":"33 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142188233","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Training-free Conversion of Pretrained ANNs to SNNs for Low-Power and High-Performance Applications 免训练将预训练 ANN 转换为 SNN,以实现低功耗和高性能应用
Pub Date : 2024-09-05 DOI: arxiv-2409.03368
Tong Bu, Maohua Li, Zhaofei Yu
Spiking Neural Networks (SNNs) have emerged as a promising substitute forArtificial Neural Networks (ANNs) due to their advantages of fast inference andlow power consumption. However, the lack of efficient training algorithms hashindered their widespread adoption. Existing supervised learning algorithms forSNNs require significantly more memory and time than their ANN counterparts.Even commonly used ANN-SNN conversion methods necessitate re-training of ANNsto enhance conversion efficiency, incurring additional computational costs. Toaddress these challenges, we propose a novel training-free ANN-SNN conversionpipeline. Our approach directly converts pre-trained ANN models intohigh-performance SNNs without additional training. The conversion pipelineincludes a local-learning-based threshold balancing algorithm, which enablesefficient calculation of the optimal thresholds and fine-grained adjustment ofthreshold value by channel-wise scaling. We demonstrate the scalability of ourframework across three typical computer vision tasks: image classification,semantic segmentation, and object detection. This showcases its applicabilityto both classification and regression tasks. Moreover, we have evaluated theenergy consumption of the converted SNNs, demonstrating their superiorlow-power advantage compared to conventional ANNs. Our training-free algorithmoutperforms existing methods, highlighting its practical applicability andefficiency. This approach simplifies the deployment of SNNs by leveragingopen-source pre-trained ANN models and neuromorphic hardware, enabling fast,low-power inference with negligible performance reduction.
尖峰神经网络(SNN)具有推理速度快、功耗低等优点,因此有望取代人工神经网络(ANN)。然而,高效训练算法的缺乏阻碍了其广泛应用。即使是常用的 ANN-SNN 转换方法,也需要重新训练 ANNN 以提高转换效率,从而产生额外的计算成本。为了应对这些挑战,我们提出了一种新型免训练 ANN-SNN 转换管道。我们的方法可直接将预先训练好的 ANN 模型转换为高性能 SNN,无需额外训练。转换管道包括基于本地学习的阈值平衡算法,该算法可以高效计算最佳阈值,并通过信道缩放对阈值进行细粒度调整。我们在三个典型的计算机视觉任务中展示了我们框架的可扩展性:图像分类、语义分割和物体检测。这展示了它对分类和回归任务的适用性。此外,我们还对转换后的 SNN 的能耗进行了评估,证明与传统 ANN 相比,SNN 具有更低功耗的优势。我们的免训练算法优于现有方法,凸显了其实用性和高效性。这种方法通过利用开源预训练 ANN 模型和神经形态硬件,简化了 SNN 的部署,实现了快速、低功耗推理,性能降低可忽略不计。
{"title":"Training-free Conversion of Pretrained ANNs to SNNs for Low-Power and High-Performance Applications","authors":"Tong Bu, Maohua Li, Zhaofei Yu","doi":"arxiv-2409.03368","DOIUrl":"https://doi.org/arxiv-2409.03368","url":null,"abstract":"Spiking Neural Networks (SNNs) have emerged as a promising substitute for\u0000Artificial Neural Networks (ANNs) due to their advantages of fast inference and\u0000low power consumption. However, the lack of efficient training algorithms has\u0000hindered their widespread adoption. Existing supervised learning algorithms for\u0000SNNs require significantly more memory and time than their ANN counterparts.\u0000Even commonly used ANN-SNN conversion methods necessitate re-training of ANNs\u0000to enhance conversion efficiency, incurring additional computational costs. To\u0000address these challenges, we propose a novel training-free ANN-SNN conversion\u0000pipeline. Our approach directly converts pre-trained ANN models into\u0000high-performance SNNs without additional training. The conversion pipeline\u0000includes a local-learning-based threshold balancing algorithm, which enables\u0000efficient calculation of the optimal thresholds and fine-grained adjustment of\u0000threshold value by channel-wise scaling. We demonstrate the scalability of our\u0000framework across three typical computer vision tasks: image classification,\u0000semantic segmentation, and object detection. This showcases its applicability\u0000to both classification and regression tasks. Moreover, we have evaluated the\u0000energy consumption of the converted SNNs, demonstrating their superior\u0000low-power advantage compared to conventional ANNs. Our training-free algorithm\u0000outperforms existing methods, highlighting its practical applicability and\u0000efficiency. This approach simplifies the deployment of SNNs by leveraging\u0000open-source pre-trained ANN models and neuromorphic hardware, enabling fast,\u0000low-power inference with negligible performance reduction.","PeriodicalId":501347,"journal":{"name":"arXiv - CS - Neural and Evolutionary Computing","volume":"12 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142188232","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SNNAX -- Spiking Neural Networks in JAX SNNAX -- JAX 中的尖峰神经网络
Pub Date : 2024-09-04 DOI: arxiv-2409.02842
Jamie Lohoff, Jan Finkbeiner, Emre Neftci
Spiking Neural Networks (SNNs) simulators are essential tools to prototypebiologically inspired models and neuromorphic hardware architectures andpredict their performance. For such a tool, ease of use and flexibility arecritical, but so is simulation speed especially given the complexity inherentto simulating SNN. Here, we present SNNAX, a JAX-based framework for simulatingand training such models with PyTorch-like intuitiveness and JAX-like executionspeed. SNNAX models are easily extended and customized to fit the desired modelspecifications and target neuromorphic hardware. Additionally, SNNAX offers keyfeatures for optimizing the training and deployment of SNNs such as flexibleautomatic differentiation and just-in-time compilation. We evaluate and compareSNNAX to other commonly used machine learning (ML) frameworks used forprogramming SNNs. We provide key performance metrics, best practices,documented examples for simulating SNNs in SNNAX, and implement severalbenchmarks used in the literature.
尖峰神经网络(SNN)模拟器是对受生物学启发的模型和神经形态硬件架构进行原型设计并预测其性能的重要工具。对于这样一种工具来说,易用性和灵活性至关重要,但仿真速度也同样重要,尤其是考虑到尖峰神经网络仿真固有的复杂性。在此,我们介绍 SNNAX,这是一个基于 JAX 的框架,用于模拟和训练此类模型,具有 PyTorch 的直观性和 JAX 的执行速度。SNNAX 模型可轻松扩展和定制,以适应所需的模型规格和目标神经形态硬件。此外,SNNAX 还提供了优化 SNN 训练和部署的关键功能,如灵活的自动区分和即时编译。我们评估了 SNNAX,并将其与用于编程 SNN 的其他常用机器学习(ML)框架进行了比较。我们提供了在 SNNAX 中模拟 SNN 的关键性能指标、最佳实践和文档示例,并实现了文献中使用的多个基准。
{"title":"SNNAX -- Spiking Neural Networks in JAX","authors":"Jamie Lohoff, Jan Finkbeiner, Emre Neftci","doi":"arxiv-2409.02842","DOIUrl":"https://doi.org/arxiv-2409.02842","url":null,"abstract":"Spiking Neural Networks (SNNs) simulators are essential tools to prototype\u0000biologically inspired models and neuromorphic hardware architectures and\u0000predict their performance. For such a tool, ease of use and flexibility are\u0000critical, but so is simulation speed especially given the complexity inherent\u0000to simulating SNN. Here, we present SNNAX, a JAX-based framework for simulating\u0000and training such models with PyTorch-like intuitiveness and JAX-like execution\u0000speed. SNNAX models are easily extended and customized to fit the desired model\u0000specifications and target neuromorphic hardware. Additionally, SNNAX offers key\u0000features for optimizing the training and deployment of SNNs such as flexible\u0000automatic differentiation and just-in-time compilation. We evaluate and compare\u0000SNNAX to other commonly used machine learning (ML) frameworks used for\u0000programming SNNs. We provide key performance metrics, best practices,\u0000documented examples for simulating SNNs in SNNAX, and implement several\u0000benchmarks used in the literature.","PeriodicalId":501347,"journal":{"name":"arXiv - CS - Neural and Evolutionary Computing","volume":"28 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142188235","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
arXiv - CS - Neural and Evolutionary Computing
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1