A Novel Deep Learning ANN Supported on Langevin-Neelakanta Machine

D. De Groff, P. Neelakanta
{"title":"A Novel Deep Learning ANN Supported on Langevin-Neelakanta Machine","authors":"D. De Groff, P. Neelakanta","doi":"10.14738/tmlai.104.13100","DOIUrl":null,"url":null,"abstract":"In the contexts of deep learning (DL) considered in artificial intelligence (AI) efforts, relevant machine learning (ML) algorithms adopted refer to using a class of deep artificial neural network (ANN) that supports a learning process exercised with an enormous set of input data (labeled and/or unlabeled) so to predict at the output details on accurate features of labeled data present in the input data set. In the present study, a deep ANN is proposed thereof conceived with certain novel considerations: The proposed deep architecture consists of a large number of consequently placed structures of paired-layers. Each layer hosts identical number of neuronal units for computation and the neuronal units are massively interconnected across the entire network. Further, each paired-layer is independently subjected to unsupervised learning (USL). Hence, commencing from the input layer-pair, the excitatory (input) data supplied flows across the interconnected neurons of paired layers, terminating eventually at the final pair of layers, where the output is recovered. That is, the converged neuronal states at any given pair is iteratively passed on to the next pair and so on. The USL suite involves collectively gathering the details of neural information across a pair of the layers constituting the network. This summed data is then limited with a specific choice of a squashing (sigmoidal) function; and, the resulting scaled value is used to adjust the coefficients of interconnection weights seeking a convergence criterion. The associated learning rate on weight adjustment is uniquely designed to facilitate fast learning towards convergence. The unique aspects of deep learning proposed here refer to: (i) Deducing the learning coefficient with a compatible algorithm so as to realize a fast convergence; and, (ii) the adopted sigmoidal function in the USL loop conforms to the heuristics of the so-called Langevin-Neelakanta machine. The paper describes the proposed deep ANN architecture with necessary details on structural considerations, sigmoidal selection, prescribing required learning rate and operational (training and predictive phase) routines. Results are furnished to demonstrate the performance efficacy of the test ANN.","PeriodicalId":119801,"journal":{"name":"Transactions on Machine Learning and Artificial Intelligence","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2022-09-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Transactions on Machine Learning and Artificial Intelligence","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.14738/tmlai.104.13100","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

In the contexts of deep learning (DL) considered in artificial intelligence (AI) efforts, relevant machine learning (ML) algorithms adopted refer to using a class of deep artificial neural network (ANN) that supports a learning process exercised with an enormous set of input data (labeled and/or unlabeled) so to predict at the output details on accurate features of labeled data present in the input data set. In the present study, a deep ANN is proposed thereof conceived with certain novel considerations: The proposed deep architecture consists of a large number of consequently placed structures of paired-layers. Each layer hosts identical number of neuronal units for computation and the neuronal units are massively interconnected across the entire network. Further, each paired-layer is independently subjected to unsupervised learning (USL). Hence, commencing from the input layer-pair, the excitatory (input) data supplied flows across the interconnected neurons of paired layers, terminating eventually at the final pair of layers, where the output is recovered. That is, the converged neuronal states at any given pair is iteratively passed on to the next pair and so on. The USL suite involves collectively gathering the details of neural information across a pair of the layers constituting the network. This summed data is then limited with a specific choice of a squashing (sigmoidal) function; and, the resulting scaled value is used to adjust the coefficients of interconnection weights seeking a convergence criterion. The associated learning rate on weight adjustment is uniquely designed to facilitate fast learning towards convergence. The unique aspects of deep learning proposed here refer to: (i) Deducing the learning coefficient with a compatible algorithm so as to realize a fast convergence; and, (ii) the adopted sigmoidal function in the USL loop conforms to the heuristics of the so-called Langevin-Neelakanta machine. The paper describes the proposed deep ANN architecture with necessary details on structural considerations, sigmoidal selection, prescribing required learning rate and operational (training and predictive phase) routines. Results are furnished to demonstrate the performance efficacy of the test ANN.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
基于Langevin-Neelakanta机器的新型深度学习人工神经网络
在人工智能(AI)工作中考虑的深度学习(DL)背景下,采用的相关机器学习(ML)算法是指使用一类深度人工神经网络(ANN),该网络支持使用大量输入数据(标记和/或未标记)进行学习过程,以便预测输入数据集中存在的标记数据的准确特征的输出细节。在本研究中,提出了一个深度人工神经网络,其中考虑了一些新颖的因素:所提出的深度架构由大量成对层的相应放置结构组成。每一层承载相同数量的用于计算的神经元单元,并且神经元单元在整个网络中大规模互连。此外,每个配对层都独立地进行无监督学习(USL)。因此,从输入层对开始,提供的兴奋性(输入)数据流经成对层的相互连接的神经元,最终终止于最后一层对,在那里输出被恢复。也就是说,任意给定对的收敛神经元状态迭代地传递给下一对,以此类推。USL套件包括在构成网络的一对层上集体收集神经信息的细节。然后,这个汇总的数据被限制为一个特定的压缩(s型)函数的选择;并利用得到的缩放值来调整互连权值系数,寻求收敛准则。权重调整的相关学习率设计独特,以促进快速学习趋于收敛。这里提出的深度学习的独特之处在于:(i)用兼容的算法推导学习系数,从而实现快速收敛;(ii) USL回路中采用的s型函数符合所谓Langevin-Neelakanta机的启发式。本文描述了所提出的深度人工神经网络架构,并提供了有关结构考虑、s型选择、规定所需学习率和操作(训练和预测阶段)例程的必要细节。实验结果验证了测试人工神经网络的性能有效性。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Addressing Challenges Encountered by English Language Teachers in Imparting Communication Skills among Higher Secondary Students: A Critical Overview Singing Voice Melody Detection Inquiring About The Memetic Relationships People Have with Societal Collapse Natural Ventilation in a Semi-Confined Enclosure Heated by a Linear Heat Source NMC: A Fast and Secure ARX Cipher
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1