{"title":"Hyperbolic Tangent-Type Variant-Parameter and Robust ZNN Solutions for Resolving Time-Variant Sylvester Equation in Preassigned-Time","authors":"Jiawei Luo, Lei Yu, Bangshu Xiong","doi":"10.1007/s11063-024-11482-7","DOIUrl":null,"url":null,"abstract":"<p>To solve a general time-variant Sylvester equation, two novel zeroing neural networks (ZNNs) solutions are designed and analyzed. In the foregoing ZNN solutions, the design convergent parameters (CPs) before the nonlinear stimulated functions are very pivotal because CPs basically decide the convergent speeds. Nonetheless, the CPs are generally set to be constants, which is not feasible because CPs are generally time-variant in practical hardware conditions particularly when the external noises invade. So, a lot of variant-parameter ZNNs (VP-ZNNs) with time-variant CPs have been come up with. Comparing with fixed-parameter ZNNs, the foregoing VP-ZNNs have been illustrated to own better convergence, the downside is that the CPs generally increases over time, and will be probably infinite at last. Obviously, infinite large CPs would lead to be non-robustness of the ZNN schemes, which are not permitted in reality when the exterior noises inject. Moreover, even though VP-ZNNs are convergent over time, the growth of CPs will waste tremendous computing resources. Based on these factors, 2 hyperbolic tangent-type variant-parameter robust ZNNs (HTVPR-ZNNs) have been proposed in this paper. Both the convergent preassigned-time of the HTVPR-ZNN and top-time boundary of CPs are theoretically investigated. Many numerical simulations substantiated the admirable validity of the HTVPR-ZNN solutions.</p>","PeriodicalId":51144,"journal":{"name":"Neural Processing Letters","volume":"26 1","pages":""},"PeriodicalIF":2.6000,"publicationDate":"2024-05-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Neural Processing Letters","FirstCategoryId":"94","ListUrlMain":"https://doi.org/10.1007/s11063-024-11482-7","RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
Abstract
To solve a general time-variant Sylvester equation, two novel zeroing neural networks (ZNNs) solutions are designed and analyzed. In the foregoing ZNN solutions, the design convergent parameters (CPs) before the nonlinear stimulated functions are very pivotal because CPs basically decide the convergent speeds. Nonetheless, the CPs are generally set to be constants, which is not feasible because CPs are generally time-variant in practical hardware conditions particularly when the external noises invade. So, a lot of variant-parameter ZNNs (VP-ZNNs) with time-variant CPs have been come up with. Comparing with fixed-parameter ZNNs, the foregoing VP-ZNNs have been illustrated to own better convergence, the downside is that the CPs generally increases over time, and will be probably infinite at last. Obviously, infinite large CPs would lead to be non-robustness of the ZNN schemes, which are not permitted in reality when the exterior noises inject. Moreover, even though VP-ZNNs are convergent over time, the growth of CPs will waste tremendous computing resources. Based on these factors, 2 hyperbolic tangent-type variant-parameter robust ZNNs (HTVPR-ZNNs) have been proposed in this paper. Both the convergent preassigned-time of the HTVPR-ZNN and top-time boundary of CPs are theoretically investigated. Many numerical simulations substantiated the admirable validity of the HTVPR-ZNN solutions.
期刊介绍:
Neural Processing Letters is an international journal publishing research results and innovative ideas on all aspects of artificial neural networks. Coverage includes theoretical developments, biological models, new formal modes, learning, applications, software and hardware developments, and prospective researches.
The journal promotes fast exchange of information in the community of neural network researchers and users. The resurgence of interest in the field of artificial neural networks since the beginning of the 1980s is coupled to tremendous research activity in specialized or multidisciplinary groups. Research, however, is not possible without good communication between people and the exchange of information, especially in a field covering such different areas; fast communication is also a key aspect, and this is the reason for Neural Processing Letters