Youngseok Kim, T. Gokmen, H. Miyazoe, P. Solomon, Seyoung Kim, Asit Ray, J. Doevenspeck, R. S. Khan, V. Narayanan, T. Ando
{"title":"Neural network learning using non-ideal resistive memory devices","authors":"Youngseok Kim, T. Gokmen, H. Miyazoe, P. Solomon, Seyoung Kim, Asit Ray, J. Doevenspeck, R. S. Khan, V. Narayanan, T. Ando","doi":"10.3389/fnano.2022.1008266","DOIUrl":null,"url":null,"abstract":"We demonstrate a modified stochastic gradient (Tiki-Taka v2 or TTv2) algorithm for deep learning network training in a cross-bar array architecture based on ReRAM cells. There have been limited discussions on cross-bar arrays for training applications due to the challenges in the switching behavior of nonvolatile memory materials. TTv2 algorithm is known to overcome the device non-idealities for deep learning training. We demonstrate the feasibility of the algorithm for a linear regression task using 1R and 1T1R ReRAM devices. Using the measured device properties, we project the performance of a long short-term memory (LSTM) network with 78 K parameters. We show that TTv2 algorithm relaxes the criteria for symmetric device update response. In addition, further optimization of the algorithm increases noise robustness and significantly reduces the required number of states, thereby drastically improving the model accuracy even with non-ideal devices and achieving the test error close to that of the conventional learning algorithm with an ideal device.","PeriodicalId":34432,"journal":{"name":"Frontiers in Nanotechnology","volume":" ","pages":""},"PeriodicalIF":4.1000,"publicationDate":"2022-10-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Frontiers in Nanotechnology","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.3389/fnano.2022.1008266","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"MATERIALS SCIENCE, MULTIDISCIPLINARY","Score":null,"Total":0}
引用次数: 0
Abstract
We demonstrate a modified stochastic gradient (Tiki-Taka v2 or TTv2) algorithm for deep learning network training in a cross-bar array architecture based on ReRAM cells. There have been limited discussions on cross-bar arrays for training applications due to the challenges in the switching behavior of nonvolatile memory materials. TTv2 algorithm is known to overcome the device non-idealities for deep learning training. We demonstrate the feasibility of the algorithm for a linear regression task using 1R and 1T1R ReRAM devices. Using the measured device properties, we project the performance of a long short-term memory (LSTM) network with 78 K parameters. We show that TTv2 algorithm relaxes the criteria for symmetric device update response. In addition, further optimization of the algorithm increases noise robustness and significantly reduces the required number of states, thereby drastically improving the model accuracy even with non-ideal devices and achieving the test error close to that of the conventional learning algorithm with an ideal device.
我们展示了一种改进的随机梯度(Tiki Taka v2或TTv2)算法,用于基于ReRAM单元的横杆阵列结构中的深度学习网络训练。由于非易失性存储器材料的开关行为方面的挑战,关于用于训练应用的横杆阵列的讨论有限。已知TTv2算法可以克服深度学习训练的设备非理想性。我们使用1R和1T1R ReRAM设备证明了该算法用于线性回归任务的可行性。使用测量的设备特性,我们预测了具有78K参数的长短期存储器(LSTM)网络的性能。我们证明了TTv2算法放宽了对称设备更新响应的标准。此外,算法的进一步优化提高了噪声鲁棒性,并显著减少了所需的状态数量,从而即使在非理想设备的情况下也能大幅提高模型精度,并实现了与理想设备的传统学习算法接近的测试误差。