{"title":"Real-time fast learning hardware implementation","authors":"M. Zhang, Samuel Garcia, M. Terré","doi":"10.1051/smdo/2023001","DOIUrl":null,"url":null,"abstract":"Machine learning algorithms are widely used in many intelligent applications and cloud services. Currently, the hottest topic in this field is Deep Learning represented often by neural network structures. Deep learning is fully known as deep neural network, and artificial neural network is a typical machine learning method and an important way of deep learning. With the massive growth of data, deep learning research has made significant achievements and is widely used in natural language processing (NLP), image recognition, and autonomous driving. However, there are still many breakthroughs needed in the training time and energy consumption of deep learning. Based on our previous research on fast learning architecture for neural network, in this paper, a solution to minimize the learning time of a fully connected neural network is analysed theoretically. Therefore, we propose a new parallel algorithm structure and a training method with over-tuned parameters. This strategy finally leads to an adaptation delay and the impact of this delay on the learning performance is analyzed using a simple benchmark case study. It is shown that a reduction of the adaptation step size could be proposed to compensate errors due to the delayed adaptation, then the gain in processing time for the learning phase is analysed as a function of the network parameters chosen in this study. Finally, to realize the real-time learning, this solution is implemented with a FPGA due to the parallelism architecture and flexibility, this integration shows a good performance and low power consumption.","PeriodicalId":37601,"journal":{"name":"International Journal for Simulation and Multidisciplinary Design Optimization","volume":"1 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2023-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"International Journal for Simulation and Multidisciplinary Design Optimization","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1051/smdo/2023001","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"Mathematics","Score":null,"Total":0}
引用次数: 0
Abstract
Machine learning algorithms are widely used in many intelligent applications and cloud services. Currently, the hottest topic in this field is Deep Learning represented often by neural network structures. Deep learning is fully known as deep neural network, and artificial neural network is a typical machine learning method and an important way of deep learning. With the massive growth of data, deep learning research has made significant achievements and is widely used in natural language processing (NLP), image recognition, and autonomous driving. However, there are still many breakthroughs needed in the training time and energy consumption of deep learning. Based on our previous research on fast learning architecture for neural network, in this paper, a solution to minimize the learning time of a fully connected neural network is analysed theoretically. Therefore, we propose a new parallel algorithm structure and a training method with over-tuned parameters. This strategy finally leads to an adaptation delay and the impact of this delay on the learning performance is analyzed using a simple benchmark case study. It is shown that a reduction of the adaptation step size could be proposed to compensate errors due to the delayed adaptation, then the gain in processing time for the learning phase is analysed as a function of the network parameters chosen in this study. Finally, to realize the real-time learning, this solution is implemented with a FPGA due to the parallelism architecture and flexibility, this integration shows a good performance and low power consumption.
期刊介绍:
The International Journal for Simulation and Multidisciplinary Design Optimization is a peer-reviewed journal covering all aspects related to the simulation and multidisciplinary design optimization. It is devoted to publish original work related to advanced design methodologies, theoretical approaches, contemporary computers and their applications to different fields such as engineering software/hardware developments, science, computing techniques, aerospace, automobile, aeronautic, business, management, manufacturing,... etc. Front-edge research topics related to topology optimization, composite material design, numerical simulation of manufacturing process, advanced optimization algorithms, industrial applications of optimization methods are highly suggested. The scope includes, but is not limited to original research contributions, reviews in the following topics: Parameter identification & Surface Response (all aspects of characterization and modeling of materials and structural behaviors, Artificial Neural Network, Parametric Programming, approximation methods,…etc.) Optimization Strategies (optimization methods that involve heuristic or Mathematics approaches, Control Theory, Linear & Nonlinear Programming, Stochastic Programming, Discrete & Dynamic Programming, Operational Research, Algorithms in Optimization based on nature behaviors,….etc.) Structural Optimization (sizing, shape and topology optimizations with or without external constraints for materials and structures) Dynamic and Vibration (cover modelling and simulation for dynamic and vibration analysis, shape and topology optimizations with or without external constraints for materials and structures) Industrial Applications (Applications Related to Optimization, Modelling for Engineering applications are very welcome. Authors should underline the technological, numerical or integration of the mentioned scopes.).