首页 > 最新文献

arXiv - CS - Neural and Evolutionary Computing最新文献

英文 中文
When In-memory Computing Meets Spiking Neural Networks -- A Perspective on Device-Circuit-System-and-Algorithm Co-design 当内存计算遇到尖峰神经网络--设备-电路-系统-算法协同设计透视
Pub Date : 2024-08-22 DOI: arxiv-2408.12767
Abhishek Moitra, Abhiroop Bhattacharjee, Yuhang Li, Youngeun Kim, Priyadarshini Panda
This review explores the intersection of bio-plausible artificialintelligence in the form of Spiking Neural Networks (SNNs) with the analogIn-Memory Computing (IMC) domain, highlighting their collective potential forlow-power edge computing environments. Through detailed investigation at thedevice, circuit, and system levels, we highlight the pivotal synergies betweenSNNs and IMC architectures. Additionally, we emphasize the critical need forcomprehensive system-level analyses, considering the inter-dependencies betweenalgorithms, devices, circuit & system parameters, crucial for optimalperformance. An in-depth analysis leads to identification of key system-levelbottlenecks arising from device limitations which can be addressed usingSNN-specific algorithm-hardware co-design techniques. This review underscoresthe imperative for holistic device to system design space co-exploration,highlighting the critical aspects of hardware and algorithm research endeavorsfor low-power neuromorphic solutions.
这篇综述探讨了以尖峰神经网络(SNN)为形式的仿生人工智能与模拟内存计算(IMC)领域的交叉点,强调了它们在低功耗边缘计算环境中的共同潜力。通过对设备、电路和系统层面的详细研究,我们强调了 SNN 与 IMC 架构之间的关键协同作用。此外,我们还强调了全面系统级分析的关键需求,考虑了算法、设备、电路和系统参数之间的相互依存关系,这对实现最佳性能至关重要。通过深入分析,可以识别出由于器件限制而产生的关键系统级瓶颈,这些瓶颈可以通过特定于 SNN 的算法-硬件协同设计技术来解决。这篇综述强调了从器件到系统设计空间的整体共同探索的必要性,突出了低功耗神经形态解决方案的硬件和算法研究工作的关键方面。
{"title":"When In-memory Computing Meets Spiking Neural Networks -- A Perspective on Device-Circuit-System-and-Algorithm Co-design","authors":"Abhishek Moitra, Abhiroop Bhattacharjee, Yuhang Li, Youngeun Kim, Priyadarshini Panda","doi":"arxiv-2408.12767","DOIUrl":"https://doi.org/arxiv-2408.12767","url":null,"abstract":"This review explores the intersection of bio-plausible artificial\u0000intelligence in the form of Spiking Neural Networks (SNNs) with the analog\u0000In-Memory Computing (IMC) domain, highlighting their collective potential for\u0000low-power edge computing environments. Through detailed investigation at the\u0000device, circuit, and system levels, we highlight the pivotal synergies between\u0000SNNs and IMC architectures. Additionally, we emphasize the critical need for\u0000comprehensive system-level analyses, considering the inter-dependencies between\u0000algorithms, devices, circuit & system parameters, crucial for optimal\u0000performance. An in-depth analysis leads to identification of key system-level\u0000bottlenecks arising from device limitations which can be addressed using\u0000SNN-specific algorithm-hardware co-design techniques. This review underscores\u0000the imperative for holistic device to system design space co-exploration,\u0000highlighting the critical aspects of hardware and algorithm research endeavors\u0000for low-power neuromorphic solutions.","PeriodicalId":501347,"journal":{"name":"arXiv - CS - Neural and Evolutionary Computing","volume":"38 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-08-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142188271","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Towards Efficient Formal Verification of Spiking Neural Network 实现尖峰神经网络的高效形式验证
Pub Date : 2024-08-20 DOI: arxiv-2408.10900
Baekryun Seong, Jieung Kim, Sang-Ki Ko
Recently, AI research has primarily focused on large language models (LLMs),and increasing accuracy often involves scaling up and consuming more power. Thepower consumption of AI has become a significant societal issue; in thiscontext, spiking neural networks (SNNs) offer a promising solution. SNNsoperate event-driven, like the human brain, and compress informationtemporally. These characteristics allow SNNs to significantly reduce powerconsumption compared to perceptron-based artificial neural networks (ANNs),highlighting them as a next-generation neural network technology. However,societal concerns regarding AI go beyond power consumption, with thereliability of AI models being a global issue. For instance, adversarialattacks on AI models are a well-studied problem in the context of traditionalneural networks. Despite their importance, the stability and propertyverification of SNNs remains in the early stages of research. Most SNNverification methods are time-consuming and barely scalable, making practicalapplications challenging. In this paper, we introduce temporal encoding toachieve practical performance in verifying the adversarial robustness of SNNs.We conduct a theoretical analysis of this approach and demonstrate its successin verifying SNNs at previously unmanageable scales. Our contribution advancesSNN verification to a practical level, facilitating the safer application ofSNNs.
最近,人工智能研究主要集中在大型语言模型(LLM)上,而要提高准确性,往往需要扩大规模,消耗更多电力。在这种情况下,尖峰神经网络(SNN)提供了一个很有前景的解决方案。尖峰神经网络像人脑一样由事件驱动运行,并按时间压缩信息。与基于感知器的人工神经网络(ANN)相比,尖峰神经网络的这些特点使其能够显著降低功耗,从而成为下一代神经网络技术。然而,社会对人工智能的关注不仅限于功耗,人工智能模型的可靠性也是一个全球性问题。例如,在传统神经网络中,对人工智能模型的对抗性攻击是一个经过深入研究的问题。尽管 SNNs 十分重要,但其稳定性和属性验证仍处于早期研究阶段。大多数 SNN 验证方法既耗时又难以扩展,使实际应用面临挑战。我们对这种方法进行了理论分析,并证明它能在以前无法管理的规模上成功验证 SNN。我们的贡献将 SNN 验证提升到了实用水平,从而促进了 SNN 的更安全应用。
{"title":"Towards Efficient Formal Verification of Spiking Neural Network","authors":"Baekryun Seong, Jieung Kim, Sang-Ki Ko","doi":"arxiv-2408.10900","DOIUrl":"https://doi.org/arxiv-2408.10900","url":null,"abstract":"Recently, AI research has primarily focused on large language models (LLMs),\u0000and increasing accuracy often involves scaling up and consuming more power. The\u0000power consumption of AI has become a significant societal issue; in this\u0000context, spiking neural networks (SNNs) offer a promising solution. SNNs\u0000operate event-driven, like the human brain, and compress information\u0000temporally. These characteristics allow SNNs to significantly reduce power\u0000consumption compared to perceptron-based artificial neural networks (ANNs),\u0000highlighting them as a next-generation neural network technology. However,\u0000societal concerns regarding AI go beyond power consumption, with the\u0000reliability of AI models being a global issue. For instance, adversarial\u0000attacks on AI models are a well-studied problem in the context of traditional\u0000neural networks. Despite their importance, the stability and property\u0000verification of SNNs remains in the early stages of research. Most SNN\u0000verification methods are time-consuming and barely scalable, making practical\u0000applications challenging. In this paper, we introduce temporal encoding to\u0000achieve practical performance in verifying the adversarial robustness of SNNs.\u0000We conduct a theoretical analysis of this approach and demonstrate its success\u0000in verifying SNNs at previously unmanageable scales. Our contribution advances\u0000SNN verification to a practical level, facilitating the safer application of\u0000SNNs.","PeriodicalId":501347,"journal":{"name":"arXiv - CS - Neural and Evolutionary Computing","volume":"30 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-08-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142188308","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Physics-Driven AI Correction in Laser Absorption Sensing Quantification 激光吸收传感定量中的物理驱动人工智能校正
Pub Date : 2024-08-20 DOI: arxiv-2408.10714
Ruiyuan Kang, Panos Liatsis, Meixia Geng, Qingjie Yang
Laser absorption spectroscopy (LAS) quantification is a popular tool used inmeasuring temperature and concentration of gases. It has low error tolerance,whereas current ML-based solutions cannot guarantee their measure reliability.In this work, we propose a new framework, SPEC, to address this issue. Inaddition to the conventional ML estimator-based estimation mode, SPEC alsoincludes a Physics-driven Anomaly Detection module (PAD) to assess the error ofthe estimation. And a Correction mode is designed to correct the unreliableestimation. The correction mode is a network-based optimization algorithm,which uses the guidance of error to iteratively correct the estimation. Ahybrid surrogate error model is proposed to estimate the error distribution,which contains an ensemble of networks to simulate reconstruction error, andtrue feasible error computation. A greedy ensemble search is proposed to findthe optimal correction robustly and efficiently from the gradient guidance ofsurrogate model. The proposed SPEC is validated on the test scenarios which areoutside the training distribution. The results show that SPEC can significantlyimprove the estimation quality, and the correction mode outperforms currentnetwork-based optimization algorithms. In addition, SPEC has thereconfigurability, which can be easily adapted to different quantificationtasks via changing PAD without retraining the ML estimator.
激光吸收光谱(LAS)定量是测量温度和气体浓度的常用工具。在这项工作中,我们提出了一个新的框架 SPEC 来解决这个问题。除了传统的基于 ML 估算器的估算模式外,SPEC 还包括一个物理驱动的异常检测模块(PAD),用于评估估算误差。此外,还设计了一种修正模式来纠正不可靠的估计。修正模式是一种基于网络的优化算法,它利用误差的指导来迭代修正估算。提出了一种混合代用误差模型来估计误差分布,该模型包含模拟重建误差的网络集合和真实可行误差计算。提出了一种贪婪集合搜索方法,以便从代理模型的梯度引导中稳健高效地找到最优修正。提出的 SPEC 在训练分布之外的测试场景中进行了验证。结果表明,SPEC 可以显著提高估计质量,其修正模式优于当前基于网络的优化算法。此外,SPEC 还具有可配置性,可以通过改变 PAD 轻松适应不同的量化任务,而无需重新训练 ML 估计器。
{"title":"Physics-Driven AI Correction in Laser Absorption Sensing Quantification","authors":"Ruiyuan Kang, Panos Liatsis, Meixia Geng, Qingjie Yang","doi":"arxiv-2408.10714","DOIUrl":"https://doi.org/arxiv-2408.10714","url":null,"abstract":"Laser absorption spectroscopy (LAS) quantification is a popular tool used in\u0000measuring temperature and concentration of gases. It has low error tolerance,\u0000whereas current ML-based solutions cannot guarantee their measure reliability.\u0000In this work, we propose a new framework, SPEC, to address this issue. In\u0000addition to the conventional ML estimator-based estimation mode, SPEC also\u0000includes a Physics-driven Anomaly Detection module (PAD) to assess the error of\u0000the estimation. And a Correction mode is designed to correct the unreliable\u0000estimation. The correction mode is a network-based optimization algorithm,\u0000which uses the guidance of error to iteratively correct the estimation. A\u0000hybrid surrogate error model is proposed to estimate the error distribution,\u0000which contains an ensemble of networks to simulate reconstruction error, and\u0000true feasible error computation. A greedy ensemble search is proposed to find\u0000the optimal correction robustly and efficiently from the gradient guidance of\u0000surrogate model. The proposed SPEC is validated on the test scenarios which are\u0000outside the training distribution. The results show that SPEC can significantly\u0000improve the estimation quality, and the correction mode outperforms current\u0000network-based optimization algorithms. In addition, SPEC has the\u0000reconfigurability, which can be easily adapted to different quantification\u0000tasks via changing PAD without retraining the ML estimator.","PeriodicalId":501347,"journal":{"name":"arXiv - CS - Neural and Evolutionary Computing","volume":"76 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-08-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142188277","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Improved Differential Evolution based Feature Selection through Quantum, Chaos, and Lasso 通过量子、混沌和拉索改进基于差分进化的特征选择
Pub Date : 2024-08-20 DOI: arxiv-2408.10693
Yelleti Vivek, Sri Krishna Vadlamani, Vadlamani Ravi, P. Radha Krishna
Modern deep learning continues to achieve outstanding performance on anastounding variety of high-dimensional tasks. In practice, this is obtained byfitting deep neural models to all the input data with minimal featureengineering, thus sacrificing interpretability in many cases. However, inapplications such as medicine, where interpretability is crucial, featuresubset selection becomes an important problem. Metaheuristics such as BinaryDifferential Evolution are a popular approach to feature selection, and theresearch literature continues to introduce novel ideas, drawn from quantumcomputing and chaos theory, for instance, to improve them. In this paper, wedemonstrate that introducing chaos-generated variables, generated fromconsiderations of the Lyapunov time, in place of random variables inquantum-inspired metaheuristics significantly improves their performance onhigh-dimensional medical classification tasks and outperforms other approaches.We show that this chaos-induced improvement is a general phenomenon bydemonstrating it for multiple varieties of underlying quantum-inspiredmetaheuristics. Performance is further enhanced through Lasso-assisted featurepruning. At the implementation level, we vastly speed up our algorithms througha scalable island-based computing cluster parallelization technique.
现代深度学习不断在各种高维任务中取得出色的性能。在实践中,这是通过将深度神经模型与所有输入数据相匹配,并尽量减少特征工程来实现的,因此在很多情况下牺牲了可解释性。然而,在医学等应用中,可解释性至关重要,特征子集的选择就成了一个重要问题。二元差分进化论等元搜索算法是一种流行的特征选择方法,研究文献不断引入量子计算和混沌理论等新思想对其进行改进。在本文中,我们证明了在量子启发元heuristics中引入混沌生成的变量(由Lyapunov时间的考虑而生成)来代替随机变量,可以显著提高它们在高维医学分类任务中的性能,并且优于其他方法。通过 Lasso 辅助特征剪枝,性能得到了进一步提升。在实现层面,我们通过可扩展的岛式计算集群并行化技术大大加快了算法的速度。
{"title":"Improved Differential Evolution based Feature Selection through Quantum, Chaos, and Lasso","authors":"Yelleti Vivek, Sri Krishna Vadlamani, Vadlamani Ravi, P. Radha Krishna","doi":"arxiv-2408.10693","DOIUrl":"https://doi.org/arxiv-2408.10693","url":null,"abstract":"Modern deep learning continues to achieve outstanding performance on an\u0000astounding variety of high-dimensional tasks. In practice, this is obtained by\u0000fitting deep neural models to all the input data with minimal feature\u0000engineering, thus sacrificing interpretability in many cases. However, in\u0000applications such as medicine, where interpretability is crucial, feature\u0000subset selection becomes an important problem. Metaheuristics such as Binary\u0000Differential Evolution are a popular approach to feature selection, and the\u0000research literature continues to introduce novel ideas, drawn from quantum\u0000computing and chaos theory, for instance, to improve them. In this paper, we\u0000demonstrate that introducing chaos-generated variables, generated from\u0000considerations of the Lyapunov time, in place of random variables in\u0000quantum-inspired metaheuristics significantly improves their performance on\u0000high-dimensional medical classification tasks and outperforms other approaches.\u0000We show that this chaos-induced improvement is a general phenomenon by\u0000demonstrating it for multiple varieties of underlying quantum-inspired\u0000metaheuristics. Performance is further enhanced through Lasso-assisted feature\u0000pruning. At the implementation level, we vastly speed up our algorithms through\u0000a scalable island-based computing cluster parallelization technique.","PeriodicalId":501347,"journal":{"name":"arXiv - CS - Neural and Evolutionary Computing","volume":"60 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-08-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142188278","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Recurrent Neural Networks Learn to Store and Generate Sequences using Non-Linear Representations 递归神经网络利用非线性表征学习存储和生成序列
Pub Date : 2024-08-20 DOI: arxiv-2408.10920
Róbert Csordás, Christopher Potts, Christopher D. Manning, Atticus Geiger
The Linear Representation Hypothesis (LRH) states that neural networks learnto encode concepts as directions in activation space, and a strong version ofthe LRH states that models learn only such encodings. In this paper, we presenta counterexample to this strong LRH: when trained to repeat an input tokensequence, gated recurrent neural networks (RNNs) learn to represent the tokenat each position with a particular order of magnitude, rather than a direction.These representations have layered features that are impossible to locate indistinct linear subspaces. To show this, we train interventions to predict andmanipulate tokens by learning the scaling factor corresponding to each sequenceposition. These interventions indicate that the smallest RNNs find only thismagnitude-based solution, while larger RNNs have linear representations. Thesefindings strongly indicate that interpretability research should not beconfined by the LRH.
线性表征假说(Larine Representation Hypothesis,LRH)指出,神经网络学习将概念编码为激活空间中的方向,而 LRH 的强版本指出,模型只学习这样的编码。在本文中,我们提出了这个强 LRH 的反例:当训练重复输入的标记序列时,门控递归神经网络(RNN)会学习用特定的数量级而不是方向来表示每个位置上的标记。为了说明这一点,我们通过学习与每个序列位置相对应的缩放因子来训练预测和操纵标记的干预。这些干预表明,最小的 RNN 只能找到这种基于幅度的解决方案,而较大的 RNN 则具有线性表征。这些发现有力地表明,可解释性研究不应受限于 LRH。
{"title":"Recurrent Neural Networks Learn to Store and Generate Sequences using Non-Linear Representations","authors":"Róbert Csordás, Christopher Potts, Christopher D. Manning, Atticus Geiger","doi":"arxiv-2408.10920","DOIUrl":"https://doi.org/arxiv-2408.10920","url":null,"abstract":"The Linear Representation Hypothesis (LRH) states that neural networks learn\u0000to encode concepts as directions in activation space, and a strong version of\u0000the LRH states that models learn only such encodings. In this paper, we present\u0000a counterexample to this strong LRH: when trained to repeat an input token\u0000sequence, gated recurrent neural networks (RNNs) learn to represent the token\u0000at each position with a particular order of magnitude, rather than a direction.\u0000These representations have layered features that are impossible to locate in\u0000distinct linear subspaces. To show this, we train interventions to predict and\u0000manipulate tokens by learning the scaling factor corresponding to each sequence\u0000position. These interventions indicate that the smallest RNNs find only this\u0000magnitude-based solution, while larger RNNs have linear representations. These\u0000findings strongly indicate that interpretability research should not be\u0000confined by the LRH.","PeriodicalId":501347,"journal":{"name":"arXiv - CS - Neural and Evolutionary Computing","volume":"30 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-08-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142188297","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Neural Exploratory Landscape Analysis 神经探索性景观分析
Pub Date : 2024-08-20 DOI: arxiv-2408.10672
Zeyuan Ma, Jiacheng Chen, Hongshu Guo, Yue-Jiao Gong
Recent research in Meta-Black-Box Optimization (MetaBBO) have shown thatmeta-trained neural networks can effectively guide the design of black-boxoptimizers, significantly reducing the need for expert tuning and deliveringrobust performance across complex problem distributions. Despite their success,a paradox remains: MetaBBO still rely on human-crafted Exploratory LandscapeAnalysis features to inform the meta-level agent about the low-leveloptimization progress. To address the gap, this paper proposes NeuralExploratory Landscape Analysis (NeurELA), a novel framework that dynamicallyprofiles landscape features through a two-stage, attention-based neuralnetwork, executed in an entirely end-to-end fashion. NeurELA is pre-trainedover a variety of MetaBBO algorithms using a multi-task neuroevolutionstrategy. Extensive experiments show that NeurELA achieves consistentlysuperior performance when integrated into different and even unseen MetaBBOtasks and can be efficiently fine-tuned for further performance boost. Thisadvancement marks a pivotal step in making MetaBBO algorithms more autonomousand broadly applicable.
元黑盒优化(MetaBBO)领域的最新研究表明,经过元训练的神经网络可以有效地指导黑盒优化器的设计,大大减少对专家调整的需求,并在复杂的问题分布中提供可靠的性能。尽管取得了成功,但矛盾依然存在:元BBO仍然依赖于人类创建的 "探索性景观分析"(Exploratory LandscapeAnalysis)功能来告知元级代理低开发优化的进展情况。为了弥补这一差距,本文提出了神经探索性景观分析(NeurELA),这是一个新颖的框架,它通过一个基于注意力的两阶段神经网络,以完全端到端方式执行,动态地描述景观特征。NeurELA 使用多任务神经进化策略对各种 MetaBBO 算法进行预训练。广泛的实验表明,NeurELA 在集成到不同甚至未见过的元博狗网好不好任务中时,性能始终保持在较高水平,并且可以有效地进行微调,以进一步提高性能。这一进步标志着在使元BBO算法更具自主性和广泛适用性方面迈出了关键的一步。
{"title":"Neural Exploratory Landscape Analysis","authors":"Zeyuan Ma, Jiacheng Chen, Hongshu Guo, Yue-Jiao Gong","doi":"arxiv-2408.10672","DOIUrl":"https://doi.org/arxiv-2408.10672","url":null,"abstract":"Recent research in Meta-Black-Box Optimization (MetaBBO) have shown that\u0000meta-trained neural networks can effectively guide the design of black-box\u0000optimizers, significantly reducing the need for expert tuning and delivering\u0000robust performance across complex problem distributions. Despite their success,\u0000a paradox remains: MetaBBO still rely on human-crafted Exploratory Landscape\u0000Analysis features to inform the meta-level agent about the low-level\u0000optimization progress. To address the gap, this paper proposes Neural\u0000Exploratory Landscape Analysis (NeurELA), a novel framework that dynamically\u0000profiles landscape features through a two-stage, attention-based neural\u0000network, executed in an entirely end-to-end fashion. NeurELA is pre-trained\u0000over a variety of MetaBBO algorithms using a multi-task neuroevolution\u0000strategy. Extensive experiments show that NeurELA achieves consistently\u0000superior performance when integrated into different and even unseen MetaBBO\u0000tasks and can be efficiently fine-tuned for further performance boost. This\u0000advancement marks a pivotal step in making MetaBBO algorithms more autonomous\u0000and broadly applicable.","PeriodicalId":501347,"journal":{"name":"arXiv - CS - Neural and Evolutionary Computing","volume":"8 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-08-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142188298","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Event Stream based Sign Language Translation: A High-Definition Benchmark Dataset and A New Algorithm 基于事件流的手语翻译:高清基准数据集与新算法
Pub Date : 2024-08-20 DOI: arxiv-2408.10488
Xiao Wang, Yao Rong, Fuling Wang, Jianing Li, Lin Zhu, Bo Jiang, Yaowei Wang
Sign Language Translation (SLT) is a core task in the field of AI-assisteddisability. Unlike traditional SLT based on visible light videos, which iseasily affected by factors such as lighting, rapid hand movements, and privacybreaches, this paper proposes the use of high-definition Event streams for SLT,effectively mitigating the aforementioned issues. This is primarily becauseEvent streams have a high dynamic range and dense temporal signals, which canwithstand low illumination and motion blur well. Additionally, due to theirsparsity in space, they effectively protect the privacy of the target person.More specifically, we propose a new high-resolution Event stream sign languagedataset, termed Event-CSL, which effectively fills the data gap in this area ofresearch. It contains 14,827 videos, 14,821 glosses, and 2,544 Chinese words inthe text vocabulary. These samples are collected in a variety of indoor andoutdoor scenes, encompassing multiple angles, light intensities, and cameramovements. We have benchmarked existing mainstream SLT works to enable faircomparison for future efforts. Based on this dataset and several otherlarge-scale datasets, we propose a novel baseline method that fully leveragesthe Mamba model's ability to integrate temporal information of CNN features,resulting in improved sign language translation outcomes. Both the benchmarkdataset and source code will be released onhttps://github.com/Event-AHU/OpenESL
手语翻译(SLT)是人工智能辅助残疾领域的一项核心任务。传统的手语翻译基于可见光视频,容易受到光线、快速手部动作和隐私泄露等因素的影响,而本文提出使用高清事件流进行手语翻译,有效缓解了上述问题。这主要是因为事件流具有高动态范围和密集的时间信号,能够很好地抵御低照度和运动模糊。更具体地说,我们提出了一个新的高分辨率事件流手势语言数据集,称为 Event-CSL,它有效地填补了这一研究领域的数据空白。它包含 14,827 个视频、14,821 个词汇和 2,544 个中文文本词汇。这些样本是在各种室内和室外场景中收集的,包括多角度、光照强度和摄像机运动。我们对现有的主流 SLT 作品进行了基准测试,以便为今后的工作提供公平的比较。基于该数据集和其他几个大规模数据集,我们提出了一种新颖的基准方法,该方法充分利用了 Mamba 模型整合 CNN 特征的时间信息的能力,从而提高了手语翻译效果。基准数据集和源代码都将在 https://github.com/Event-AHU/OpenESL 上发布。
{"title":"Event Stream based Sign Language Translation: A High-Definition Benchmark Dataset and A New Algorithm","authors":"Xiao Wang, Yao Rong, Fuling Wang, Jianing Li, Lin Zhu, Bo Jiang, Yaowei Wang","doi":"arxiv-2408.10488","DOIUrl":"https://doi.org/arxiv-2408.10488","url":null,"abstract":"Sign Language Translation (SLT) is a core task in the field of AI-assisted\u0000disability. Unlike traditional SLT based on visible light videos, which is\u0000easily affected by factors such as lighting, rapid hand movements, and privacy\u0000breaches, this paper proposes the use of high-definition Event streams for SLT,\u0000effectively mitigating the aforementioned issues. This is primarily because\u0000Event streams have a high dynamic range and dense temporal signals, which can\u0000withstand low illumination and motion blur well. Additionally, due to their\u0000sparsity in space, they effectively protect the privacy of the target person.\u0000More specifically, we propose a new high-resolution Event stream sign language\u0000dataset, termed Event-CSL, which effectively fills the data gap in this area of\u0000research. It contains 14,827 videos, 14,821 glosses, and 2,544 Chinese words in\u0000the text vocabulary. These samples are collected in a variety of indoor and\u0000outdoor scenes, encompassing multiple angles, light intensities, and camera\u0000movements. We have benchmarked existing mainstream SLT works to enable fair\u0000comparison for future efforts. Based on this dataset and several other\u0000large-scale datasets, we propose a novel baseline method that fully leverages\u0000the Mamba model's ability to integrate temporal information of CNN features,\u0000resulting in improved sign language translation outcomes. Both the benchmark\u0000dataset and source code will be released on\u0000https://github.com/Event-AHU/OpenESL","PeriodicalId":501347,"journal":{"name":"arXiv - CS - Neural and Evolutionary Computing","volume":"52 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-08-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142188299","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Evaluation Framework for AI-driven Molecular Design of Multi-target Drugs: Brain Diseases as a Case Study 人工智能驱动的多靶点药物分子设计评估框架:以脑部疾病为例
Pub Date : 2024-08-20 DOI: arxiv-2408.10482
Arthur Cerveira, Frederico Kremer, Darling de Andrade Lourenço, Ulisses B Corrêa
The widespread application of Artificial Intelligence (AI) techniques hassignificantly influenced the development of new therapeutic agents. Thesecomputational methods can be used to design and predict the properties ofgenerated molecules. Multi-target Drug Discovery (MTDD) is an emerging paradigmfor discovering drugs against complex disorders that do not respond well tomore traditional target-specific treatments, such as central nervous system,immune system, and cardiovascular diseases. Still, there is yet to be anestablished benchmark suite for assessing the effectiveness of AI tools fordesigning multi-target compounds. Standardized benchmarks allow for comparingexisting techniques and promote rapid research progress. Hence, this workproposes an evaluation framework for molecule generation techniques in MTDDscenarios, considering brain diseases as a case study. Our methodology involvesusing large language models to select the appropriate molecular targets,gathering and preprocessing the bioassay datasets, training quantitativestructure-activity relationship models to predict target modulation, andassessing other essential drug-likeness properties for implementing thebenchmarks. Additionally, this work will assess the performance of four deepgenerative models and evolutionary algorithms over our benchmark suite. In ourfindings, both evolutionary algorithms and generative models can achievecompetitive results across the proposed benchmarks.
人工智能(AI)技术的广泛应用对新型治疗药物的开发产生了重大影响。这些计算方法可用于设计和预测生成分子的特性。多靶点药物发现(MTDD)是一种新兴的范式,用于发现治疗复杂疾病的药物,这些疾病对传统的特异性靶点治疗效果不佳,如中枢神经系统、免疫系统和心血管疾病。不过,目前还没有一个成熟的基准套件来评估人工智能工具在设计多靶点化合物方面的有效性。标准化的基准可以对现有技术进行比较,促进研究的快速发展。因此,本研究以脑部疾病为案例,为 MTDD 场景中的分子生成技术提出了一个评估框架。我们的方法包括使用大型语言模型来选择合适的分子靶点,收集和预处理生物测定数据集,训练定量结构-活性关系模型来预测靶点调节,以及评估实施基准的其他基本药物相似性。此外,这项工作还将评估四种深度生成模型和进化算法在我们的基准套件中的性能。我们发现,进化算法和生成模型都能在所提出的基准中取得具有竞争力的结果。
{"title":"Evaluation Framework for AI-driven Molecular Design of Multi-target Drugs: Brain Diseases as a Case Study","authors":"Arthur Cerveira, Frederico Kremer, Darling de Andrade Lourenço, Ulisses B Corrêa","doi":"arxiv-2408.10482","DOIUrl":"https://doi.org/arxiv-2408.10482","url":null,"abstract":"The widespread application of Artificial Intelligence (AI) techniques has\u0000significantly influenced the development of new therapeutic agents. These\u0000computational methods can be used to design and predict the properties of\u0000generated molecules. Multi-target Drug Discovery (MTDD) is an emerging paradigm\u0000for discovering drugs against complex disorders that do not respond well to\u0000more traditional target-specific treatments, such as central nervous system,\u0000immune system, and cardiovascular diseases. Still, there is yet to be an\u0000established benchmark suite for assessing the effectiveness of AI tools for\u0000designing multi-target compounds. Standardized benchmarks allow for comparing\u0000existing techniques and promote rapid research progress. Hence, this work\u0000proposes an evaluation framework for molecule generation techniques in MTDD\u0000scenarios, considering brain diseases as a case study. Our methodology involves\u0000using large language models to select the appropriate molecular targets,\u0000gathering and preprocessing the bioassay datasets, training quantitative\u0000structure-activity relationship models to predict target modulation, and\u0000assessing other essential drug-likeness properties for implementing the\u0000benchmarks. Additionally, this work will assess the performance of four deep\u0000generative models and evolutionary algorithms over our benchmark suite. In our\u0000findings, both evolutionary algorithms and generative models can achieve\u0000competitive results across the proposed benchmarks.","PeriodicalId":501347,"journal":{"name":"arXiv - CS - Neural and Evolutionary Computing","volume":"23 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-08-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142188279","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Mutation Strength Adaptation of the $(μ/μ_I, λ)$-ES for Large Population Sizes on the Sphere Function 球函数上大种群规模的$(μ/μ_I, λ)$-ES突变强度适应性研究
Pub Date : 2024-08-19 DOI: arxiv-2408.09761
Amir Omeradzic, Hans-Georg Beyer
The mutation strength adaptation properties of a multi-recombinative$(mu/mu_I, lambda)$-ES are studied for isotropic mutations. To this end,standard implementations of cumulative step-size adaptation (CSA) and mutativeself-adaptation ($sigma$SA) are investigated experimentally and theoreticallyby assuming large population sizes ($mu$) in relation to the search spacedimensionality ($N$). The adaptation is characterized in terms of thescale-invariant mutation strength on the sphere in relation to its maximumachievable value for positive progress. %The results show how the different$sigma$-adaptation variants behave as $mu$ and $N$ are varied. StandardCSA-variants show notably different adaptation properties and progress rates onthe sphere, becoming slower or faster as $mu$ or $N$ are varied. This is shownby investigating common choices for the cumulation and damping parameters.Standard $sigma$SA-variants (with default learning parameter settings) canachieve faster adaptation and larger progress rates compared to the CSA.However, it is shown how self-adaptation affects the progress rate levelsnegatively. Furthermore, differences regarding the adaptation and stability of$sigma$SA with log-normal and normal mutation sampling are elaborated.
针对各向同性突变,研究了多重组$(mu/mu_I, lambda)$-ES的突变强度适应特性。为此,实验和理论研究了累积步长适应(CSA)和突变自适应($sigma$SA)的标准实现,假设种群规模($mu$)与搜索间隔维度($N$)相关较大。适应性的特征是球体上的规模不变突变强度与正进展的最大可实现值的关系。结果显示了不同的$sigma$适应变体在$mu$和$N$变化时的表现。标准 CSA 变体在球面上显示出明显不同的适应特性和进展速度,随着 $mu$ 或 $N$ 的变化而变慢或变快。与 CSA 相比,标准的 $sigma$SA 变体(使用默认学习参数设置)可以获得更快的适应性和更大的进展率。此外,还阐述了采用对数正态和正态突变采样的 CSA 在适应性和稳定性方面的差异。
{"title":"Mutation Strength Adaptation of the $(μ/μ_I, λ)$-ES for Large Population Sizes on the Sphere Function","authors":"Amir Omeradzic, Hans-Georg Beyer","doi":"arxiv-2408.09761","DOIUrl":"https://doi.org/arxiv-2408.09761","url":null,"abstract":"The mutation strength adaptation properties of a multi-recombinative\u0000$(mu/mu_I, lambda)$-ES are studied for isotropic mutations. To this end,\u0000standard implementations of cumulative step-size adaptation (CSA) and mutative\u0000self-adaptation ($sigma$SA) are investigated experimentally and theoretically\u0000by assuming large population sizes ($mu$) in relation to the search space\u0000dimensionality ($N$). The adaptation is characterized in terms of the\u0000scale-invariant mutation strength on the sphere in relation to its maximum\u0000achievable value for positive progress. %The results show how the different\u0000$sigma$-adaptation variants behave as $mu$ and $N$ are varied. Standard\u0000CSA-variants show notably different adaptation properties and progress rates on\u0000the sphere, becoming slower or faster as $mu$ or $N$ are varied. This is shown\u0000by investigating common choices for the cumulation and damping parameters.\u0000Standard $sigma$SA-variants (with default learning parameter settings) can\u0000achieve faster adaptation and larger progress rates compared to the CSA.\u0000However, it is shown how self-adaptation affects the progress rate levels\u0000negatively. Furthermore, differences regarding the adaptation and stability of\u0000$sigma$SA with log-normal and normal mutation sampling are elaborated.","PeriodicalId":501347,"journal":{"name":"arXiv - CS - Neural and Evolutionary Computing","volume":"3 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-08-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142188300","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Liquid Fourier Latent Dynamics Networks for fast GPU-based numerical simulations in computational cardiology 液体傅立叶潜动力网络用于计算心脏病学中基于 GPU 的快速数值模拟
Pub Date : 2024-08-19 DOI: arxiv-2408.09818
Matteo Salvador, Alison L. Marsden
Scientific Machine Learning (ML) is gaining momentum as a cost-effectivealternative to physics-based numerical solvers in many engineeringapplications. In fact, scientific ML is currently being used to build accurateand efficient surrogate models starting from high-fidelity numericalsimulations, effectively encoding the parameterized temporal dynamicsunderlying Ordinary Differential Equations (ODEs), or even the spatio-temporalbehavior underlying Partial Differential Equations (PDEs), in appropriatelydesigned neural networks. We propose an extension of Latent Dynamics Networks(LDNets), namely Liquid Fourier LDNets (LFLDNets), to create parameterizedspace-time surrogate models for multiscale and multiphysics sets of highlynonlinear differential equations on complex geometries. LFLDNets employ aneurologically-inspired, sparse, liquid neural network for temporal dynamics,relaxing the requirement of a numerical solver for time advancement and leadingto superior performance in terms of tunable parameters, accuracy, efficiencyand learned trajectories with respect to neural ODEs based on feedforwardfully-connected neural networks. Furthermore, in our implementation ofLFLDNets, we use a Fourier embedding with a tunable kernel in thereconstruction network to learn high-frequency functions better and faster thanusing space coordinates directly as input. We challenge LFLDNets in theframework of computational cardiology and evaluate their capabilities on two3-dimensional test cases arising from multiscale cardiac electrophysiology andcardiovascular hemodynamics. This paper illustrates the capability to runArtificial Intelligence-based numerical simulations on single or multiple GPUsin a matter of minutes and represents a significant step forward in thedevelopment of physics-informed digital twins.
在许多工程应用中,科学机器学习(ML)作为基于物理的数值求解器的一种经济高效的替代方法,正获得越来越大的发展势头。事实上,科学机器学习目前正被用于从高保真数值模拟出发建立精确高效的代理模型,从而有效地将常微分方程(ODE)或偏微分方程(PDE)的时空动态参数化编码到适当设计的神经网络中。我们提出了潜在动力学网络(LDNets)的扩展,即液体傅立叶 LDNets(LFLDNets),用于创建复杂几何体上多尺度和多物理场高非线性微分方程组的参数化时空代理模型。LFLDNets 采用受神经学启发的稀疏液体神经网络来处理时间动力学,放宽了对时间推进数值求解器的要求,与基于全连接神经网络的神经 ODE 相比,在可调参数、准确性、效率和学习轨迹方面具有更优越的性能。此外,在我们的 LFLDNets 实现中,我们在其构建网络中使用了带有可调内核的傅立叶嵌入,从而比直接将空间坐标作为输入更好、更快地学习高频函数。我们在计算心脏病学的框架内对 LFLDNets 提出了挑战,并在多尺度心脏电生理学和心血管血流动力学的二三维测试案例中对其能力进行了评估。本文展示了在单个或多个 GPU 上运行基于人工智能的数值模拟只需几分钟的能力,标志着在开发物理信息数字双胞胎方面迈出了重要一步。
{"title":"Liquid Fourier Latent Dynamics Networks for fast GPU-based numerical simulations in computational cardiology","authors":"Matteo Salvador, Alison L. Marsden","doi":"arxiv-2408.09818","DOIUrl":"https://doi.org/arxiv-2408.09818","url":null,"abstract":"Scientific Machine Learning (ML) is gaining momentum as a cost-effective\u0000alternative to physics-based numerical solvers in many engineering\u0000applications. In fact, scientific ML is currently being used to build accurate\u0000and efficient surrogate models starting from high-fidelity numerical\u0000simulations, effectively encoding the parameterized temporal dynamics\u0000underlying Ordinary Differential Equations (ODEs), or even the spatio-temporal\u0000behavior underlying Partial Differential Equations (PDEs), in appropriately\u0000designed neural networks. We propose an extension of Latent Dynamics Networks\u0000(LDNets), namely Liquid Fourier LDNets (LFLDNets), to create parameterized\u0000space-time surrogate models for multiscale and multiphysics sets of highly\u0000nonlinear differential equations on complex geometries. LFLDNets employ a\u0000neurologically-inspired, sparse, liquid neural network for temporal dynamics,\u0000relaxing the requirement of a numerical solver for time advancement and leading\u0000to superior performance in terms of tunable parameters, accuracy, efficiency\u0000and learned trajectories with respect to neural ODEs based on feedforward\u0000fully-connected neural networks. Furthermore, in our implementation of\u0000LFLDNets, we use a Fourier embedding with a tunable kernel in the\u0000reconstruction network to learn high-frequency functions better and faster than\u0000using space coordinates directly as input. We challenge LFLDNets in the\u0000framework of computational cardiology and evaluate their capabilities on two\u00003-dimensional test cases arising from multiscale cardiac electrophysiology and\u0000cardiovascular hemodynamics. This paper illustrates the capability to run\u0000Artificial Intelligence-based numerical simulations on single or multiple GPUs\u0000in a matter of minutes and represents a significant step forward in the\u0000development of physics-informed digital twins.","PeriodicalId":501347,"journal":{"name":"arXiv - CS - Neural and Evolutionary Computing","volume":"23 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-08-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142188304","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
arXiv - CS - Neural and Evolutionary Computing
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1