{"title":"推动机器学习的未来发展","authors":"O. Temam","doi":"10.1109/VLSIC.2016.7573457","DOIUrl":null,"url":null,"abstract":"Amazing progress in machine-learning, largely based on deep neural networks, has started to make applications once considered impossible, such as real-time translation or self-driving cars, a reality. However, even if, on some restricted problems, machine-learning is getting close to human-level performance, we are still far from the capabilities of the human brain. Machine-learning researchers themselves acknowledge that the progress observed in the past 10 years has been largely due to rapid increase in computing performance, allowing to tackle larger neural networks and larger training sets. So the computer systems and circuits communities can play a very significant role in enabling future progress. While GPUs have been a major driver of this recent progress, both the slowing rate of improvement of standard CMOS technology and the need for even faster progress suggest to at least explore alternative approaches. In this talk, we will discuss lessons learned from research on architectures for machine-learning, and that some of the hurdles ahead largely lie at the circuit level, but can possibly be overcome in the near future.","PeriodicalId":6512,"journal":{"name":"2016 IEEE Symposium on VLSI Circuits (VLSI-Circuits)","volume":"28 1","pages":"1-3"},"PeriodicalIF":0.0000,"publicationDate":"2016-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"5","resultStr":"{\"title\":\"Enabling future progress in machine-learning\",\"authors\":\"O. Temam\",\"doi\":\"10.1109/VLSIC.2016.7573457\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Amazing progress in machine-learning, largely based on deep neural networks, has started to make applications once considered impossible, such as real-time translation or self-driving cars, a reality. However, even if, on some restricted problems, machine-learning is getting close to human-level performance, we are still far from the capabilities of the human brain. Machine-learning researchers themselves acknowledge that the progress observed in the past 10 years has been largely due to rapid increase in computing performance, allowing to tackle larger neural networks and larger training sets. So the computer systems and circuits communities can play a very significant role in enabling future progress. While GPUs have been a major driver of this recent progress, both the slowing rate of improvement of standard CMOS technology and the need for even faster progress suggest to at least explore alternative approaches. In this talk, we will discuss lessons learned from research on architectures for machine-learning, and that some of the hurdles ahead largely lie at the circuit level, but can possibly be overcome in the near future.\",\"PeriodicalId\":6512,\"journal\":{\"name\":\"2016 IEEE Symposium on VLSI Circuits (VLSI-Circuits)\",\"volume\":\"28 1\",\"pages\":\"1-3\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2016-06-15\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"5\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2016 IEEE Symposium on VLSI Circuits (VLSI-Circuits)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/VLSIC.2016.7573457\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2016 IEEE Symposium on VLSI Circuits (VLSI-Circuits)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/VLSIC.2016.7573457","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Amazing progress in machine-learning, largely based on deep neural networks, has started to make applications once considered impossible, such as real-time translation or self-driving cars, a reality. However, even if, on some restricted problems, machine-learning is getting close to human-level performance, we are still far from the capabilities of the human brain. Machine-learning researchers themselves acknowledge that the progress observed in the past 10 years has been largely due to rapid increase in computing performance, allowing to tackle larger neural networks and larger training sets. So the computer systems and circuits communities can play a very significant role in enabling future progress. While GPUs have been a major driver of this recent progress, both the slowing rate of improvement of standard CMOS technology and the need for even faster progress suggest to at least explore alternative approaches. In this talk, we will discuss lessons learned from research on architectures for machine-learning, and that some of the hurdles ahead largely lie at the circuit level, but can possibly be overcome in the near future.