首页 > 最新文献

International Conference on Machine Learning, Optimization, and Data Science最新文献

英文 中文
A Comparison of SVM against Pre-trained Language Models (PLMs) for Text Classification Tasks 支持向量机与预训练语言模型在文本分类任务中的比较
Pub Date : 2022-11-04 DOI: 10.48550/arXiv.2211.02563
Yasmen Wahba, N. Madhavji, John Steinbacher
The emergence of pre-trained language models (PLMs) has shown great success in many Natural Language Processing (NLP) tasks including text classification. Due to the minimal to no feature engineering required when using these models, PLMs are becoming the de facto choice for any NLP task. However, for domain-specific corpora (e.g., financial, legal, and industrial), fine-tuning a pre-trained model for a specific task has shown to provide a performance improvement. In this paper, we compare the performance of four different PLMs on three public domain-free datasets and a real-world dataset containing domain-specific words, against a simple SVM linear classifier with TFIDF vectorized text. The experimental results on the four datasets show that using PLMs, even fine-tuned, do not provide significant gain over the linear SVM classifier. Hence, we recommend that for text classification tasks, traditional SVM along with careful feature engineering can pro-vide a cheaper and superior performance than PLMs.
预训练语言模型(PLMs)的出现在包括文本分类在内的许多自然语言处理(NLP)任务中取得了巨大的成功。由于在使用这些模型时几乎不需要特征工程,因此plm正在成为任何NLP任务的事实上的选择。然而,对于特定领域的语料库(例如,金融、法律和工业),为特定任务微调预训练的模型可以提供性能改进。在本文中,我们比较了四种不同的PLMs在三个公共无领域数据集和一个包含领域特定词的真实数据集上的性能,并与具有TFIDF矢量化文本的简单SVM线性分类器进行了比较。在四个数据集上的实验结果表明,使用PLMs,即使经过微调,也不会提供比线性SVM分类器显著的增益。因此,我们建议,对于文本分类任务,传统的SVM加上仔细的特征工程可以提供比PLMs更便宜和更好的性能。
{"title":"A Comparison of SVM against Pre-trained Language Models (PLMs) for Text Classification Tasks","authors":"Yasmen Wahba, N. Madhavji, John Steinbacher","doi":"10.48550/arXiv.2211.02563","DOIUrl":"https://doi.org/10.48550/arXiv.2211.02563","url":null,"abstract":"The emergence of pre-trained language models (PLMs) has shown great success in many Natural Language Processing (NLP) tasks including text classification. Due to the minimal to no feature engineering required when using these models, PLMs are becoming the de facto choice for any NLP task. However, for domain-specific corpora (e.g., financial, legal, and industrial), fine-tuning a pre-trained model for a specific task has shown to provide a performance improvement. In this paper, we compare the performance of four different PLMs on three public domain-free datasets and a real-world dataset containing domain-specific words, against a simple SVM linear classifier with TFIDF vectorized text. The experimental results on the four datasets show that using PLMs, even fine-tuned, do not provide significant gain over the linear SVM classifier. Hence, we recommend that for text classification tasks, traditional SVM along with careful feature engineering can pro-vide a cheaper and superior performance than PLMs.","PeriodicalId":432112,"journal":{"name":"International Conference on Machine Learning, Optimization, and Data Science","volume":"29 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-11-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130027693","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Hierarchical Decentralized Deep Reinforcement Learning Architecture for a Simulated Four-Legged Agent 模拟四足智能体的分层分散深度强化学习体系结构
Pub Date : 2022-09-21 DOI: 10.48550/arXiv.2210.08003
W. Z. E. Amri, L. Hermes, M. Schilling
Legged locomotion is widespread in nature and has inspired the design of current robots. The controller of these legged robots is often realized as one centralized instance. However, in nature, control of movement happens in a hierarchical and decentralized fashion. Introducing these biological design principles into robotic control systems has motivated this work. We tackle the question whether decentralized and hierarchical control is beneficial for legged robots and present a novel decentral, hierarchical architecture to control a simulated legged agent. Three different tasks varying in complexity are designed to benchmark five architectures (centralized, decentralized, hierarchical and two different combinations of hierarchical decentralized architectures). The results demonstrate that decentralizing the different levels of the hierarchical architectures facilitates learning of the agent, ensures more energy efficient movements as well as robustness towards new unseen environments. Furthermore, this comparison sheds light on the importance of modularity in hierarchical architectures to solve complex goal-directed tasks. We provide an open-source code implementation of our architecture (https://github.com/wzaielamri/hddrl).
腿式运动在自然界中广泛存在,并启发了当前机器人的设计。这些足式机器人的控制器通常作为一个集中实例来实现。然而,在自然界中,对运动的控制是以分层和分散的方式发生的。将这些生物设计原理引入机器人控制系统激发了这项工作。我们解决了去中心化和分层控制是否对有腿机器人有益的问题,并提出了一种新的去中心化、分层结构来控制模拟的有腿机器人。设计了复杂度不同的三种不同任务,对五种架构(集中式、分散式、分层式和分层式分散式架构的两种不同组合)进行基准测试。结果表明,分散层次结构的不同层次有助于智能体的学习,确保更节能的运动以及对新的未知环境的鲁棒性。此外,这种比较揭示了模块化在分层体系结构中解决复杂目标导向任务的重要性。我们提供了架构的开源代码实现(https://github.com/wzaielamri/hddrl)。
{"title":"Hierarchical Decentralized Deep Reinforcement Learning Architecture for a Simulated Four-Legged Agent","authors":"W. Z. E. Amri, L. Hermes, M. Schilling","doi":"10.48550/arXiv.2210.08003","DOIUrl":"https://doi.org/10.48550/arXiv.2210.08003","url":null,"abstract":"Legged locomotion is widespread in nature and has inspired the design of current robots. The controller of these legged robots is often realized as one centralized instance. However, in nature, control of movement happens in a hierarchical and decentralized fashion. Introducing these biological design principles into robotic control systems has motivated this work. We tackle the question whether decentralized and hierarchical control is beneficial for legged robots and present a novel decentral, hierarchical architecture to control a simulated legged agent. Three different tasks varying in complexity are designed to benchmark five architectures (centralized, decentralized, hierarchical and two different combinations of hierarchical decentralized architectures). The results demonstrate that decentralizing the different levels of the hierarchical architectures facilitates learning of the agent, ensures more energy efficient movements as well as robustness towards new unseen environments. Furthermore, this comparison sheds light on the importance of modularity in hierarchical architectures to solve complex goal-directed tasks. We provide an open-source code implementation of our architecture (https://github.com/wzaielamri/hddrl).","PeriodicalId":432112,"journal":{"name":"International Conference on Machine Learning, Optimization, and Data Science","volume":"481 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-09-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114782452","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
On the utility and protection of optimization with differential privacy and classic regularization techniques 差分隐私和经典正则化技术优化的效用和保护
Pub Date : 2022-09-07 DOI: 10.48550/arXiv.2209.03175
Eugenio Lomurno, Matteo Matteucci
Nowadays, owners and developers of deep learning models must consider stringent privacy-preservation rules of their training data, usually crowd-sourced and retaining sensitive information. The most widely adopted method to enforce privacy guarantees of a deep learning model nowadays relies on optimization techniques enforcing differential privacy. According to the literature, this approach has proven to be a successful defence against several models' privacy attacks, but its downside is a substantial degradation of the models' performance. In this work, we compare the effectiveness of the differentially-private stochastic gradient descent (DP-SGD) algorithm against standard optimization practices with regularization techniques. We analyze the resulting models' utility, training performance, and the effectiveness of membership inference and model inversion attacks against the learned models. Finally, we discuss differential privacy's flaws and limits and empirically demonstrate the often superior privacy-preserving properties of dropout and l2-regularization.
如今,深度学习模型的所有者和开发人员必须考虑严格的训练数据隐私保护规则,这些数据通常是众包的,并保留敏感信息。目前采用最广泛的方法来执行深度学习模型的隐私保证依赖于执行差分隐私的优化技术。根据文献,这种方法已被证明是一种成功的防御几个模型的隐私攻击,但它的缺点是模型的性能大幅下降。在这项工作中,我们比较了微分私有随机梯度下降(DP-SGD)算法与正则化技术的标准优化实践的有效性。我们分析了所得模型的效用、训练性能、隶属推理和模型反演攻击对学习模型的有效性。最后,我们讨论了差分隐私的缺陷和局限性,并实证证明了dropout和12 -正则化通常具有优越的隐私保护特性。
{"title":"On the utility and protection of optimization with differential privacy and classic regularization techniques","authors":"Eugenio Lomurno, Matteo Matteucci","doi":"10.48550/arXiv.2209.03175","DOIUrl":"https://doi.org/10.48550/arXiv.2209.03175","url":null,"abstract":"Nowadays, owners and developers of deep learning models must consider stringent privacy-preservation rules of their training data, usually crowd-sourced and retaining sensitive information. The most widely adopted method to enforce privacy guarantees of a deep learning model nowadays relies on optimization techniques enforcing differential privacy. According to the literature, this approach has proven to be a successful defence against several models' privacy attacks, but its downside is a substantial degradation of the models' performance. In this work, we compare the effectiveness of the differentially-private stochastic gradient descent (DP-SGD) algorithm against standard optimization practices with regularization techniques. We analyze the resulting models' utility, training performance, and the effectiveness of membership inference and model inversion attacks against the learned models. Finally, we discuss differential privacy's flaws and limits and empirically demonstrate the often superior privacy-preserving properties of dropout and l2-regularization.","PeriodicalId":432112,"journal":{"name":"International Conference on Machine Learning, Optimization, and Data Science","volume":"40 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-09-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130289948","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Adaptive Zeroth-Order Optimisation of Nonconvex Composite Objectives 非凸复合目标的自适应零阶优化
Pub Date : 2022-08-09 DOI: 10.48550/arXiv.2208.04579
Weijia Shao, S. Albayrak
In this paper, we propose and analyze algorithms for zeroth-order optimization of non-convex composite objectives, focusing on reducing the complexity dependence on dimensionality. This is achieved by exploiting the low dimensional structure of the decision set using the stochastic mirror descent method with an entropy alike function, which performs gradient descent in the space equipped with the maximum norm. To improve the gradient estimation, we replace the classic Gaussian smoothing method with a sampling method based on the Rademacher distribution and show that the mini-batch method copes with the non-Euclidean geometry. To avoid tuning hyperparameters, we analyze the adaptive stepsizes for the general stochastic mirror descent and show that the adaptive version of the proposed algorithm converges without requiring prior knowledge about the problem.
在本文中,我们提出并分析了非凸复合目标的零阶优化算法,重点是降低复杂度对维数的依赖。这是通过使用具有熵相似函数的随机镜像下降方法利用决策集的低维结构来实现的,该方法在具有最大范数的空间中执行梯度下降。为了改进梯度估计,我们用基于Rademacher分布的抽样方法取代了经典的高斯平滑方法,并证明了小批量方法可以处理非欧几里得几何。为了避免调优超参数,我们分析了一般随机镜像下降的自适应步长,并证明了所提出算法的自适应版本在不需要先验知识的情况下收敛。
{"title":"Adaptive Zeroth-Order Optimisation of Nonconvex Composite Objectives","authors":"Weijia Shao, S. Albayrak","doi":"10.48550/arXiv.2208.04579","DOIUrl":"https://doi.org/10.48550/arXiv.2208.04579","url":null,"abstract":"In this paper, we propose and analyze algorithms for zeroth-order optimization of non-convex composite objectives, focusing on reducing the complexity dependence on dimensionality. This is achieved by exploiting the low dimensional structure of the decision set using the stochastic mirror descent method with an entropy alike function, which performs gradient descent in the space equipped with the maximum norm. To improve the gradient estimation, we replace the classic Gaussian smoothing method with a sampling method based on the Rademacher distribution and show that the mini-batch method copes with the non-Euclidean geometry. To avoid tuning hyperparameters, we analyze the adaptive stepsizes for the general stochastic mirror descent and show that the adaptive version of the proposed algorithm converges without requiring prior knowledge about the problem.","PeriodicalId":432112,"journal":{"name":"International Conference on Machine Learning, Optimization, and Data Science","volume":"13 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-08-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132546629","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Robust PCA for Anomaly Detection and Data Imputation in Seasonal Time Series 基于鲁棒主成分分析的季节时间序列异常检测与数据代入
Pub Date : 2022-08-03 DOI: 10.48550/arXiv.2208.01998
Hông-Lan Botterman, Julien Roussel, Thomas Morzadec, A. Jabbari, N. Brunel
We propose a robust principal component analysis (RPCA) framework to recover low-rank and sparse matrices from temporal observations. We develop an online version of the batch temporal algorithm in order to process larger datasets or streaming data. We empirically compare the proposed approaches with different RPCA frameworks and show their effectiveness in practical situations.
我们提出了一个鲁棒主成分分析(RPCA)框架来从时间观测中恢复低秩和稀疏矩阵。为了处理更大的数据集或流数据,我们开发了一个在线版本的批处理时序算法。我们将所提出的方法与不同的RPCA框架进行了实证比较,并在实际情况中展示了它们的有效性。
{"title":"Robust PCA for Anomaly Detection and Data Imputation in Seasonal Time Series","authors":"Hông-Lan Botterman, Julien Roussel, Thomas Morzadec, A. Jabbari, N. Brunel","doi":"10.48550/arXiv.2208.01998","DOIUrl":"https://doi.org/10.48550/arXiv.2208.01998","url":null,"abstract":"We propose a robust principal component analysis (RPCA) framework to recover low-rank and sparse matrices from temporal observations. We develop an online version of the batch temporal algorithm in order to process larger datasets or streaming data. We empirically compare the proposed approaches with different RPCA frameworks and show their effectiveness in practical situations.","PeriodicalId":432112,"journal":{"name":"International Conference on Machine Learning, Optimization, and Data Science","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-08-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134499057","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Parallel Bayesian Optimization of Agent-based Transportation Simulation 基于agent的交通仿真并行贝叶斯优化
Pub Date : 2022-07-11 DOI: 10.48550/arXiv.2207.05041
K. Chhatre, Sidney A. Feygin, C. Sheppard, R. Waraich
MATSim (Multi-Agent Transport Simulation Toolkit) is an open source large-scale agent-based transportation planning project applied to various areas like road transport, public transport, freight transport, regional evacuation, etc. BEAM (Behavior, Energy, Autonomy, and Mobility) framework extends MATSim to enable powerful and scalable analysis of urban transportation systems. The agents from the BEAM simulation exhibit 'mode choice' behavior based on multinomial logit model. In our study, we consider eight mode choices viz. bike, car, walk, ride hail, driving to transit, walking to transit, ride hail to transit, and ride hail pooling. The 'alternative specific constants' for each mode choice are critical hyperparameters in a configuration file related to a particular scenario under experimentation. We use the 'Urbansim-10k' BEAM scenario (with 10,000 population size) for all our experiments. Since these hyperparameters affect the simulation in complex ways, manual calibration methods are time consuming. We present a parallel Bayesian optimization method with early stopping rule to achieve fast convergence for the given multi-in-multi-out problem to its optimal configurations. Our model is based on an open source HpBandSter package. This approach combines hierarchy of several 1D Kernel Density Estimators (KDE) with a cheap evaluator (Hyperband, a single multidimensional KDE). Our model has also incorporated extrapolation based early stopping rule. With our model, we could achieve a 25% L1 norm for a large-scale BEAM simulation in fully autonomous manner. To the best of our knowledge, our work is the first of its kind applied to large-scale multi-agent transportation simulations. This work can be useful for surrogate modeling of scenarios with very large populations.
MATSim (Multi-Agent Transport Simulation Toolkit)是一个开源的大规模基于agent的交通规划项目,应用于道路运输、公共交通、货运、区域疏散等各个领域。BEAM(行为、能源、自主和移动)框架扩展了MATSim,使其能够对城市交通系统进行强大且可扩展的分析。BEAM仿真中的智能体表现出基于多项logit模型的“模式选择”行为。在我们的研究中,我们考虑了八种模式的选择,即自行车、汽车、步行、打车、开车到公交、步行到公交、打车到公交和拼车。每种模式选择的“可选特定常数”是与实验中的特定场景相关的配置文件中的关键超参数。我们使用“Urbansim-10k”BEAM场景(人口规模为10,000)进行所有实验。由于这些超参数以复杂的方式影响仿真,因此手动校准方法非常耗时。针对给定的多入多出问题,提出了一种具有提前停止规则的并行贝叶斯优化方法,使其快速收敛到最优配置。我们的模型是基于一个开源的HpBandSter包。这种方法结合了几个1D核密度估计器(KDE)的层次结构和一个便宜的求值器(Hyperband,一个单一的多维KDE)。我们的模型还纳入了基于外推的早期停止规则。利用我们的模型,我们可以在完全自主的方式下实现大规模BEAM模拟的25% L1规范。据我们所知,我们的工作是第一个应用于大规模多智能体运输模拟的研究。这项工作对于具有非常大的人口的情景的代理建模是有用的。
{"title":"Parallel Bayesian Optimization of Agent-based Transportation Simulation","authors":"K. Chhatre, Sidney A. Feygin, C. Sheppard, R. Waraich","doi":"10.48550/arXiv.2207.05041","DOIUrl":"https://doi.org/10.48550/arXiv.2207.05041","url":null,"abstract":"MATSim (Multi-Agent Transport Simulation Toolkit) is an open source large-scale agent-based transportation planning project applied to various areas like road transport, public transport, freight transport, regional evacuation, etc. BEAM (Behavior, Energy, Autonomy, and Mobility) framework extends MATSim to enable powerful and scalable analysis of urban transportation systems. The agents from the BEAM simulation exhibit 'mode choice' behavior based on multinomial logit model. In our study, we consider eight mode choices viz. bike, car, walk, ride hail, driving to transit, walking to transit, ride hail to transit, and ride hail pooling. The 'alternative specific constants' for each mode choice are critical hyperparameters in a configuration file related to a particular scenario under experimentation. We use the 'Urbansim-10k' BEAM scenario (with 10,000 population size) for all our experiments. Since these hyperparameters affect the simulation in complex ways, manual calibration methods are time consuming. We present a parallel Bayesian optimization method with early stopping rule to achieve fast convergence for the given multi-in-multi-out problem to its optimal configurations. Our model is based on an open source HpBandSter package. This approach combines hierarchy of several 1D Kernel Density Estimators (KDE) with a cheap evaluator (Hyperband, a single multidimensional KDE). Our model has also incorporated extrapolation based early stopping rule. With our model, we could achieve a 25% L1 norm for a large-scale BEAM simulation in fully autonomous manner. To the best of our knowledge, our work is the first of its kind applied to large-scale multi-agent transportation simulations. This work can be useful for surrogate modeling of scenarios with very large populations.","PeriodicalId":432112,"journal":{"name":"International Conference on Machine Learning, Optimization, and Data Science","volume":"61 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-07-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121080367","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Brain-like combination of feedforward and recurrent network components achieves prototype extraction and robust pattern recognition 将前馈和递归网络成分结合,实现了原型提取和鲁棒模式识别
Pub Date : 2022-06-30 DOI: 10.48550/arXiv.2206.15036
Naresh B. Ravichandran, A. Lansner, P. Herman
Associative memory has been a prominent candidate for the computation performed by the massively recurrent neocortical networks. Attractor networks implementing associative memory have offered mechanistic explanation for many cognitive phenomena. However, attractor memory models are typically trained using orthogonal or random patterns to avoid interference between memories, which makes them unfeasible for naturally occurring complex correlated stimuli like images. We approach this problem by combining a recurrent attractor network with a feedforward network that learns distributed representations using an unsupervised Hebbian-Bayesian learning rule. The resulting network model incorporates many known biological properties: unsupervised learning, Hebbian plasticity, sparse distributed activations, sparse connectivity, columnar and laminar cortical architecture, etc. We evaluate the synergistic effects of the feedforward and recurrent network components in complex pattern recognition tasks on the MNIST handwritten digits dataset. We demonstrate that the recurrent attractor component implements associative memory when trained on the feedforward-driven internal (hidden) representations. The associative memory is also shown to perform prototype extraction from the training data and make the representations robust to severely distorted input. We argue that several aspects of the proposed integration of feedforward and recurrent computations are particularly attractive from a machine learning perspective.
联想记忆是由大量循环的新皮层网络进行计算的一个突出候选。实现联想记忆的吸引子网络为许多认知现象提供了机制解释。然而,吸引子记忆模型通常使用正交或随机模式进行训练,以避免记忆之间的干扰,这使得它们对于自然发生的复杂相关刺激(如图像)不可行。我们通过将循环吸引子网络与使用无监督Hebbian-Bayesian学习规则学习分布式表示的前馈网络相结合来解决这个问题。由此产生的网络模型融合了许多已知的生物学特性:无监督学习、Hebbian可塑性、稀疏分布激活、稀疏连接、柱状和层状皮质结构等。我们在MNIST手写数字数据集上评估了复杂模式识别任务中前馈和循环网络组件的协同效应。我们证明了循环吸引子组件在前馈驱动的内部(隐藏)表征上进行训练时实现了联想记忆。联想记忆还可以从训练数据中进行原型提取,并使表征对严重扭曲的输入具有鲁棒性。我们认为,从机器学习的角度来看,前馈和循环计算集成的几个方面特别有吸引力。
{"title":"Brain-like combination of feedforward and recurrent network components achieves prototype extraction and robust pattern recognition","authors":"Naresh B. Ravichandran, A. Lansner, P. Herman","doi":"10.48550/arXiv.2206.15036","DOIUrl":"https://doi.org/10.48550/arXiv.2206.15036","url":null,"abstract":"Associative memory has been a prominent candidate for the computation performed by the massively recurrent neocortical networks. Attractor networks implementing associative memory have offered mechanistic explanation for many cognitive phenomena. However, attractor memory models are typically trained using orthogonal or random patterns to avoid interference between memories, which makes them unfeasible for naturally occurring complex correlated stimuli like images. We approach this problem by combining a recurrent attractor network with a feedforward network that learns distributed representations using an unsupervised Hebbian-Bayesian learning rule. The resulting network model incorporates many known biological properties: unsupervised learning, Hebbian plasticity, sparse distributed activations, sparse connectivity, columnar and laminar cortical architecture, etc. We evaluate the synergistic effects of the feedforward and recurrent network components in complex pattern recognition tasks on the MNIST handwritten digits dataset. We demonstrate that the recurrent attractor component implements associative memory when trained on the feedforward-driven internal (hidden) representations. The associative memory is also shown to perform prototype extraction from the training data and make the representations robust to severely distorted input. We argue that several aspects of the proposed integration of feedforward and recurrent computations are particularly attractive from a machine learning perspective.","PeriodicalId":432112,"journal":{"name":"International Conference on Machine Learning, Optimization, and Data Science","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-06-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130978720","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MicroRacer: a didactic environment for Deep Reinforcement Learning MicroRacer:深度强化学习的教学环境
Pub Date : 2022-03-20 DOI: 10.48550/arXiv.2203.10494
A. Asperti, Marco Del Brutto
MicroRacer is a simple, open source environment inspired by car racing especially meant for the didactics of Deep Reinforcement Learning. The complexity of the environment has been explicitly calibrated to allow users to experiment with many different methods, networks and hyperparameters settings without requiring sophisticated software or the need of exceedingly long training times. Baseline agents for major learning algorithms such as DDPG, PPO, SAC, TD2 and DSAC are provided too, along with a preliminary comparison in terms of training time and performance.
MicroRacer是一个简单的开源环境,受赛车的启发,特别适用于深度强化学习的教学。环境的复杂性已经被明确校准,允许用户尝试许多不同的方法、网络和超参数设置,而不需要复杂的软件或超长的培训时间。提供了DDPG、PPO、SAC、TD2、DSAC等主要学习算法的基线代理,并对训练时间和性能进行了初步比较。
{"title":"MicroRacer: a didactic environment for Deep Reinforcement Learning","authors":"A. Asperti, Marco Del Brutto","doi":"10.48550/arXiv.2203.10494","DOIUrl":"https://doi.org/10.48550/arXiv.2203.10494","url":null,"abstract":"MicroRacer is a simple, open source environment inspired by car racing especially meant for the didactics of Deep Reinforcement Learning. The complexity of the environment has been explicitly calibrated to allow users to experiment with many different methods, networks and hyperparameters settings without requiring sophisticated software or the need of exceedingly long training times. Baseline agents for major learning algorithms such as DDPG, PPO, SAC, TD2 and DSAC are provided too, along with a preliminary comparison in terms of training time and performance.","PeriodicalId":432112,"journal":{"name":"International Conference on Machine Learning, Optimization, and Data Science","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-03-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121095916","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Algorithms that Get Old: The Case of Generative Deep Neural Networks 过时的算法:生成式深度神经网络的案例
Pub Date : 2022-02-07 DOI: 10.1007/978-3-031-25891-6_14
G. Turinici
{"title":"Algorithms that Get Old: The Case of Generative Deep Neural Networks","authors":"G. Turinici","doi":"10.1007/978-3-031-25891-6_14","DOIUrl":"https://doi.org/10.1007/978-3-031-25891-6_14","url":null,"abstract":"","PeriodicalId":432112,"journal":{"name":"International Conference on Machine Learning, Optimization, and Data Science","volume":"18 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-02-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126706235","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Brain Structural Saliency Over The Ages 多年来大脑结构的显著性
Pub Date : 2022-01-12 DOI: 10.1007/978-3-031-25891-6_40
Daniel Taylor, Jonathan Shock, Deshendran Moodley, J. Ipser, M. Treder
{"title":"Brain Structural Saliency Over The Ages","authors":"Daniel Taylor, Jonathan Shock, Deshendran Moodley, J. Ipser, M. Treder","doi":"10.1007/978-3-031-25891-6_40","DOIUrl":"https://doi.org/10.1007/978-3-031-25891-6_40","url":null,"abstract":"","PeriodicalId":432112,"journal":{"name":"International Conference on Machine Learning, Optimization, and Data Science","volume":"44 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-01-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133554114","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
International Conference on Machine Learning, Optimization, and Data Science
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1