首页 > 最新文献

Proceedings of the 2007 international symposium on Low power electronics and design (ISLPED '07)最新文献

英文 中文
A Fully Onchip Binarized Convolutional Neural Network FPGA Impelmentation with Accurate Inference 具有精确推理的全片上二值化卷积神经网络FPGA实现
Li Yang, Zhezhi He, Deliang Fan
Deep convolutional neural network has taken an important role in machine learning algorithm which has been widely used in computer vision tasks. However, its enormous model size and massive computation cost have became the main obstacle for deployment of such powerful algorithm in low power and resource limited embedded system, such as FPGA. Recent works have shown the binarized neural networks (BNN), utilizing binarized (i.e. +1 and -1) convolution kernel and binary activation function, can significantly reduce the model size and computation complexity, which paves a new road for energy-efficient FPGA implementation. In this work, we first propose a new BNN algorithm, called Parallel-Convolution BNN (i.e. PC-BNN), which replaces the original binary convolution layer in conventional BNN with two parallel binary convolution layers. PC-BNN achieves ~86% on CIFAR-10 dataset with only 2.3Mb parameter size. We then deploy our proposed PC-BNN into the Xilinx PYNQ Z1 FPGA board with only 4.9Mb on-chip RAM. Since the ultra-small network parameter, it is feasible to store the whole network parameter into on-chip RAM, which could greatly reduce the energy and delay overhead to load network parameter from off-chip memory. Meanwhile, a new data streaming pipeline architecture is proposed in PC-BNN FPGA implementation to further improve throughput. The experiment results show that our PC-BNN based FPGA implementation achieves 930 frames per second, 387.5 FPS/Watt and 396x10-4 FPS/LUT, which are among the best throughput and energy efficiency compared to most recent works.
深度卷积神经网络在机器学习算法中起着重要的作用,在计算机视觉任务中得到了广泛的应用。然而,其庞大的模型尺寸和庞大的计算成本已经成为这种强大算法在低功耗和资源有限的嵌入式系统(如FPGA)中部署的主要障碍。最近的研究表明,利用二值化(即+1和-1)卷积核和二值激活函数的二值化神经网络(BNN)可以显著减小模型尺寸和计算复杂度,为高效节能的FPGA实现铺平了新的道路。在这项工作中,我们首先提出了一种新的BNN算法,称为parallel - convolution BNN(即PC-BNN),它将传统BNN中的原始二进制卷积层替换为两个并行二进制卷积层。PC-BNN在只有2.3Mb参数大小的CIFAR-10数据集上达到~86%。然后,我们将我们提出的PC-BNN部署到只有4.9Mb片上RAM的Xilinx PYNQ Z1 FPGA板上。由于网络参数超小,将整个网络参数存储在片上RAM中是可行的,这样可以大大减少从片外存储器加载网络参数的能量和延迟开销。同时,在PC-BNN FPGA实现中提出了一种新的数据流管道架构,进一步提高了吞吐量。实验结果表明,基于PC-BNN的FPGA实现实现了930帧/秒,387.5 FPS/Watt和396x10-4 FPS/LUT,与最近的工作相比,具有最佳的吞吐量和能效。
{"title":"A Fully Onchip Binarized Convolutional Neural Network FPGA Impelmentation with Accurate Inference","authors":"Li Yang, Zhezhi He, Deliang Fan","doi":"10.1145/3218603.3218615","DOIUrl":"https://doi.org/10.1145/3218603.3218615","url":null,"abstract":"Deep convolutional neural network has taken an important role in machine learning algorithm which has been widely used in computer vision tasks. However, its enormous model size and massive computation cost have became the main obstacle for deployment of such powerful algorithm in low power and resource limited embedded system, such as FPGA. Recent works have shown the binarized neural networks (BNN), utilizing binarized (i.e. +1 and -1) convolution kernel and binary activation function, can significantly reduce the model size and computation complexity, which paves a new road for energy-efficient FPGA implementation. In this work, we first propose a new BNN algorithm, called Parallel-Convolution BNN (i.e. PC-BNN), which replaces the original binary convolution layer in conventional BNN with two parallel binary convolution layers. PC-BNN achieves ~86% on CIFAR-10 dataset with only 2.3Mb parameter size. We then deploy our proposed PC-BNN into the Xilinx PYNQ Z1 FPGA board with only 4.9Mb on-chip RAM. Since the ultra-small network parameter, it is feasible to store the whole network parameter into on-chip RAM, which could greatly reduce the energy and delay overhead to load network parameter from off-chip memory. Meanwhile, a new data streaming pipeline architecture is proposed in PC-BNN FPGA implementation to further improve throughput. The experiment results show that our PC-BNN based FPGA implementation achieves 930 frames per second, 387.5 FPS/Watt and 396x10-4 FPS/LUT, which are among the best throughput and energy efficiency compared to most recent works.","PeriodicalId":20456,"journal":{"name":"Proceedings of the 2007 international symposium on Low power electronics and design (ISLPED '07)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2018-07-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"90679359","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 36
In-situ Stochastic Training of MTJ Crossbar based Neural Networks MTJ交叉棒神经网络的原位随机训练
Ankit Mondal, Ankur Srivastava
Owing to high device density, scalability and non-volatility, Magnetic Tunnel Junction-based crossbars have garnered significant interest for implementing the weights of an artificial neural network. The existence of only two stable states in MTJs implies a high overhead of obtaining optimal binary weights in software. We illustrate that the inherent parallelism in the crossbar structure makes it highly appropriate for in-situ training, wherein the network is taught directly on the hardware. It leads to significantly smaller training overhead as the training time is independent of the size of the network, while also circumventing the effects of alternate current paths in the crossbar and accounting for manufacturing variations in the device. We show how the stochastic switching characteristics of MTJs can be leveraged to perform probabilistic weight updates using the gradient descent algorithm. We describe how the update operations can be performed on crossbars both with and without access transistors and perform simulations on them to demonstrate the effectiveness of our techniques. The results reveal that stochastically trained MTJ-crossbar NNs achieve a classification accuracy nearly same as that of real-valued-weight networks trained in software and exhibit immunity to device variations.
由于高器件密度、可扩展性和非易失性,基于磁隧道结的交叉棒在实现人工神经网络的权重方面获得了极大的兴趣。mtj中只有两个稳定状态的存在意味着在软件中获得最优二进制权值的开销很大。我们说明了交叉杆结构固有的并行性使其非常适合于原位训练,其中网络直接在硬件上进行教学。由于训练时间与网络的大小无关,它可以显著减少训练开销,同时还可以规避交叉棒中交变电流路径的影响,并考虑到设备中的制造变化。我们展示了如何利用mtj的随机切换特性来使用梯度下降算法执行概率权重更新。我们描述了如何在有和没有接入晶体管的交叉栅上执行更新操作,并对它们进行了模拟以证明我们技术的有效性。结果表明,随机训练的mtj交叉棒神经网络的分类精度与软件训练的实值权重网络几乎相同,并且对设备变化具有免疫力。
{"title":"In-situ Stochastic Training of MTJ Crossbar based Neural Networks","authors":"Ankit Mondal, Ankur Srivastava","doi":"10.1145/3218603.3218616","DOIUrl":"https://doi.org/10.1145/3218603.3218616","url":null,"abstract":"Owing to high device density, scalability and non-volatility, Magnetic Tunnel Junction-based crossbars have garnered significant interest for implementing the weights of an artificial neural network. The existence of only two stable states in MTJs implies a high overhead of obtaining optimal binary weights in software. We illustrate that the inherent parallelism in the crossbar structure makes it highly appropriate for in-situ training, wherein the network is taught directly on the hardware. It leads to significantly smaller training overhead as the training time is independent of the size of the network, while also circumventing the effects of alternate current paths in the crossbar and accounting for manufacturing variations in the device. We show how the stochastic switching characteristics of MTJs can be leveraged to perform probabilistic weight updates using the gradient descent algorithm. We describe how the update operations can be performed on crossbars both with and without access transistors and perform simulations on them to demonstrate the effectiveness of our techniques. The results reveal that stochastically trained MTJ-crossbar NNs achieve a classification accuracy nearly same as that of real-valued-weight networks trained in software and exhibit immunity to device variations.","PeriodicalId":20456,"journal":{"name":"Proceedings of the 2007 international symposium on Low power electronics and design (ISLPED '07)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2018-06-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"72772373","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Deploying Customized Data Representation and Approximate Computing in Machine Learning Applications 在机器学习应用中部署自定义数据表示和近似计算
M. Nazemi, Massoud Pedram
Major advancements in building general-purpose and customized hardware have been one of the key enablers of versatility and pervasiveness of machine learning models such as deep neural networks. To sustain this ubiquitous deployment of machine learning models and cope with their computational and storage complexity, several solutions such as low-precision representation of model parameters using fixed-point representation and deploying approximate arithmetic operations have been employed. Studying the potency of such solutions in different applications requires integrating them into existing machine learning frameworks for high-level simulations as well as implementing them in hardware to analyze their effects on power/energy dissipation, throughput, and chip area. Lop is a library for design space exploration that bridges the gap between machine learning and efficient hardware realization. It comprises a Python module, which can be integrated with some of the existing machine learning frameworks and implements various customizable data representations including fixed-point and floating-point as well as approximate arithmetic operations. Furthermore, it includes a highly-parameterized Scala module, which allows synthesizing hardware based on the said data representations and arithmetic operations. Lop allows researchers and designers to quickly compare quality of their models using various data representations and arithmetic operations in Python and contrast the hardware cost of viable representations by synthesizing them on their target platforms (e.g., FPGA or ASIC). To the best of our knowledge, Lop is the first library that allows both software simulation and hardware realization using customized data representations and approximate computing techniques.
构建通用和定制硬件方面的重大进展是机器学习模型(如深度神经网络)的多功能性和普遍性的关键推动因素之一。为了维持这种无处不在的机器学习模型的部署,并处理它们的计算和存储复杂性,已经采用了几种解决方案,例如使用定点表示和部署近似算术运算来低精度表示模型参数。研究这些解决方案在不同应用中的效力需要将它们集成到现有的机器学习框架中进行高级模拟,并在硬件中实现它们以分析它们对功率/能量消耗,吞吐量和芯片面积的影响。Lop是一个用于设计空间探索的库,它在机器学习和高效硬件实现之间架起了桥梁。它包含一个Python模块,该模块可以与一些现有的机器学习框架集成,并实现各种可定制的数据表示,包括定点和浮点以及近似算术运算。此外,它还包括一个高度参数化的Scala模块,该模块允许基于上述数据表示和算术运算来合成硬件。Lop允许研究人员和设计人员使用Python中的各种数据表示和算术运算快速比较模型的质量,并通过在目标平台(例如FPGA或ASIC)上合成可行表示来对比硬件成本。据我们所知,Lop是第一个允许使用自定义数据表示和近似计算技术进行软件模拟和硬件实现的库。
{"title":"Deploying Customized Data Representation and Approximate Computing in Machine Learning Applications","authors":"M. Nazemi, Massoud Pedram","doi":"10.1145/3218603.3218612","DOIUrl":"https://doi.org/10.1145/3218603.3218612","url":null,"abstract":"Major advancements in building general-purpose and customized hardware have been one of the key enablers of versatility and pervasiveness of machine learning models such as deep neural networks. To sustain this ubiquitous deployment of machine learning models and cope with their computational and storage complexity, several solutions such as low-precision representation of model parameters using fixed-point representation and deploying approximate arithmetic operations have been employed. Studying the potency of such solutions in different applications requires integrating them into existing machine learning frameworks for high-level simulations as well as implementing them in hardware to analyze their effects on power/energy dissipation, throughput, and chip area. Lop is a library for design space exploration that bridges the gap between machine learning and efficient hardware realization. It comprises a Python module, which can be integrated with some of the existing machine learning frameworks and implements various customizable data representations including fixed-point and floating-point as well as approximate arithmetic operations. Furthermore, it includes a highly-parameterized Scala module, which allows synthesizing hardware based on the said data representations and arithmetic operations. Lop allows researchers and designers to quickly compare quality of their models using various data representations and arithmetic operations in Python and contrast the hardware cost of viable representations by synthesizing them on their target platforms (e.g., FPGA or ASIC). To the best of our knowledge, Lop is the first library that allows both software simulation and hardware realization using customized data representations and approximate computing techniques.","PeriodicalId":20456,"journal":{"name":"Proceedings of the 2007 international symposium on Low power electronics and design (ISLPED '07)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2018-06-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"89960323","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
AxTrain: Hardware-Oriented Neural Network Training for Approximate Inference 面向硬件的神经网络近似推理训练
Xin He, Liu Ke, Wenyan Lu, Guihai Yan, Xuan Zhang
The intrinsic error tolerance of neural network (NN) makes approximate computing a promising technique to improve the energy efficiency of NN inference. Conventional approximate computing focuses on balancing the efficiency-accuracy trade-off for existing pre-trained networks, which can lead to suboptimal solutions. In this paper, we propose AxTrain, a hardware-oriented training framework to facilitate approximate computing for NN inference. Specifically, AxTrain leverages the synergy between two orthogonal methods---one actively searches for a network parameters distribution with high error tolerance, and the other passively learns resilient weights by numerically incorporating the noise distributions of the approximate hardware in the forward pass during the training phase. Experimental results from various datasets with near-threshold computing and approximation multiplication strategies demonstrate AxTrain's ability to obtain resilient neural network parameters and system energy efficiency improvement.
神经网络固有的容错性使近似计算成为提高神经网络推理能量效率的一种很有前途的技术。传统的近似计算侧重于平衡现有预训练网络的效率和精度之间的权衡,这可能导致次优解。在本文中,我们提出了AxTrain,一个面向硬件的训练框架,以促进神经网络推理的近似计算。具体来说,AxTrain利用了两种正交方法之间的协同作用——一种是主动搜索具有高容错性的网络参数分布,另一种是通过在训练阶段将近似硬件的噪声分布数值结合起来,被动地学习弹性权重。使用近阈值计算和近似乘法策略的各种数据集的实验结果表明,AxTrain能够获得弹性神经网络参数并提高系统能效。
{"title":"AxTrain: Hardware-Oriented Neural Network Training for Approximate Inference","authors":"Xin He, Liu Ke, Wenyan Lu, Guihai Yan, Xuan Zhang","doi":"10.1145/3218603.3218643","DOIUrl":"https://doi.org/10.1145/3218603.3218643","url":null,"abstract":"The intrinsic error tolerance of neural network (NN) makes approximate computing a promising technique to improve the energy efficiency of NN inference. Conventional approximate computing focuses on balancing the efficiency-accuracy trade-off for existing pre-trained networks, which can lead to suboptimal solutions. In this paper, we propose AxTrain, a hardware-oriented training framework to facilitate approximate computing for NN inference. Specifically, AxTrain leverages the synergy between two orthogonal methods---one actively searches for a network parameters distribution with high error tolerance, and the other passively learns resilient weights by numerically incorporating the noise distributions of the approximate hardware in the forward pass during the training phase. Experimental results from various datasets with near-threshold computing and approximation multiplication strategies demonstrate AxTrain's ability to obtain resilient neural network parameters and system energy efficiency improvement.","PeriodicalId":20456,"journal":{"name":"Proceedings of the 2007 international symposium on Low power electronics and design (ISLPED '07)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2018-05-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"80241455","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 30
Keynote: Peering into the post Moore's Law world 主题演讲:展望后摩尔定律时代
T. Austin
For decades, Moore's Law dimensional scaling has been the fuel that propelled the computing industry forward, by delivering performance, power and cost advantages with each new generation of silicon. Today, these scaling benefits are slowing to a crawl. If the computing industry wants to continue to make scalability the primary source of value in tomorrow's computing systems, we will have to quickly find new and productive ways to scale future systems. In this talk, I will highlight my work and the work of others that is rejuvenating scaling through the application of heterogeneous parallel designs. Leveraging these technologies to solve the scaling problem will be a significant challenge, as future scalability success will ultimately be less about “how” to do it and more about “how much” will it cost.
几十年来,摩尔定律的尺寸缩放一直是推动计算行业向前发展的燃料,通过每一代新硅提供性能、功率和成本优势。如今,这些规模效益正在放缓。如果计算行业希望继续使可伸缩性成为未来计算系统的主要价值来源,我们必须迅速找到新的、有效的方法来扩展未来的系统。在这次演讲中,我将重点介绍我的工作和其他人的工作,这些工作是通过应用异构并行设计来恢复可伸缩性的。利用这些技术来解决可伸缩性问题将是一个重大的挑战,因为未来可伸缩性的成功最终将不再是“如何”去做,而是更多地是“花费多少”。
{"title":"Keynote: Peering into the post Moore's Law world","authors":"T. Austin","doi":"10.1109/ISLPED.2017.8009150","DOIUrl":"https://doi.org/10.1109/ISLPED.2017.8009150","url":null,"abstract":"For decades, Moore's Law dimensional scaling has been the fuel that propelled the computing industry forward, by delivering performance, power and cost advantages with each new generation of silicon. Today, these scaling benefits are slowing to a crawl. If the computing industry wants to continue to make scalability the primary source of value in tomorrow's computing systems, we will have to quickly find new and productive ways to scale future systems. In this talk, I will highlight my work and the work of others that is rejuvenating scaling through the application of heterogeneous parallel designs. Leveraging these technologies to solve the scaling problem will be a significant challenge, as future scalability success will ultimately be less about “how” to do it and more about “how much” will it cost.","PeriodicalId":20456,"journal":{"name":"Proceedings of the 2007 international symposium on Low power electronics and design (ISLPED '07)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2017-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"91332962","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Keynote: Architecture and software for emerging low-power systems 主题演讲:新兴低功耗系统的架构和软件
Wen-mei W. Hwu
We have been experiencing two very important developments in computing. On the one hand, a tremendous amount of resources have been invested into innovative applications such as first-principle based models, deep learning and cognitive computing. On the other hand, the industry has been taking a technological path where application performance and power efficiency vary by more than two orders of magnitude depending on their parallelism, heterogeneity, and locality. We envision a “perfect storm” is coming for future computing resulting from the fact that data movement has become the dominating factor for both power and performance of high-valued applications. It will be critical to match the compute throughput to the data access bandwidth and to locate the compute at where the data is. Much has been and continuously needs to be learned about of algorithms, languages, compilers and hardware architecture in this movement. What are the killer applications that may become the new diver for future technology development? How hard is it to program existing systems to address the date movement issues today? How will we program these systems in the future? How will innovations in memory devices present further opportunities and challenges in designing new systems? What is the impact on long-term software engineering cost on applications (and legacy applications in particular)? In this talk, I will present some lessons learned as we design the IBM-Illinois C3SR Erudite system inside this perfect storm.
我们在计算机领域经历了两个非常重要的发展。一方面,大量的资源被投入到创新应用中,如基于第一性原理的模型、深度学习和认知计算。另一方面,该行业一直在采用一种技术途径,根据其并行性、异构性和局部性,应用程序性能和功率效率的变化超过两个数量级。我们预计,由于数据移动已经成为高价值应用程序的功率和性能的主导因素,未来的计算将迎来一场“完美风暴”。将计算吞吐量与数据访问带宽相匹配并将计算定位在数据所在位置将是至关重要的。在这场运动中,关于算法、语言、编译器和硬件架构,我们已经学习了很多,而且还需要继续学习。哪些杀手级应用可能成为未来技术发展的新潜水员?现在对现有系统进行编程以解决日期移动问题有多难?未来我们将如何对这些系统进行编程?存储设备的创新将如何为设计新系统带来更多的机遇和挑战?对应用程序(特别是遗留应用程序)的长期软件工程成本有什么影响?在这次演讲中,我将介绍我们在这场完美风暴中设计IBM-Illinois C3SR博学系统时学到的一些经验教训。
{"title":"Keynote: Architecture and software for emerging low-power systems","authors":"Wen-mei W. Hwu","doi":"10.1109/ISLPED.2017.8009151","DOIUrl":"https://doi.org/10.1109/ISLPED.2017.8009151","url":null,"abstract":"We have been experiencing two very important developments in computing. On the one hand, a tremendous amount of resources have been invested into innovative applications such as first-principle based models, deep learning and cognitive computing. On the other hand, the industry has been taking a technological path where application performance and power efficiency vary by more than two orders of magnitude depending on their parallelism, heterogeneity, and locality. We envision a “perfect storm” is coming for future computing resulting from the fact that data movement has become the dominating factor for both power and performance of high-valued applications. It will be critical to match the compute throughput to the data access bandwidth and to locate the compute at where the data is. Much has been and continuously needs to be learned about of algorithms, languages, compilers and hardware architecture in this movement. What are the killer applications that may become the new diver for future technology development? How hard is it to program existing systems to address the date movement issues today? How will we program these systems in the future? How will innovations in memory devices present further opportunities and challenges in designing new systems? What is the impact on long-term software engineering cost on applications (and legacy applications in particular)? In this talk, I will present some lessons learned as we design the IBM-Illinois C3SR Erudite system inside this perfect storm.","PeriodicalId":20456,"journal":{"name":"Proceedings of the 2007 international symposium on Low power electronics and design (ISLPED '07)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2017-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"77640367","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Keynote: A new Silicon Age 4.0: Generating semiconductor-intelligence paradigm with a Virtual Moore's Law Economics and Heterogeneous technologies 主题演讲:新硅时代4.0:用虚拟摩尔定律经济学和异构技术生成半导体智能范式
Nicky Liu
The future of the silicon-based economy will not be as pessimistic as some commentators have argued, given their predictions of the end of Moore's Law Economy (ME) by the early 2020s. On the contrary, a Virtual Moore's Law Economy (VME) will develop and thrive, advancing innovation by a new Silicon Way of producing various application-driven Heterogeneous Integrated (HI) Nano-systems by optimization of physics, materials, devices, circuits/chips, software and systems to enable exciting applications for business growth. The semiconductor industry will enjoy sufficient financial returns from new application and system-product sales, even considering more expensive silicon investment. Such a technological approach based on a (Function × Value)-Scaling Down-Plus-Up Methodology, in addition to Linear-Scaling, Area-Scaling and Volumetric-Scaling Methodologies, can fundamentally change the way of thinking and execution toward optimizing coherently both technology definition and final system design with an holistic HIDAS (HI Design/Architecture/System) method. This will drive IC scaling to an effective 1-Nanometer Realm, stimulating a thriving silicon industry which can have at least 30 more years of growth toward a 1 trillion-dollar size.
硅基经济的未来不会像一些评论家所说的那样悲观,因为他们预测摩尔定律经济(ME)将在本世纪20年代初终结。相反,虚拟摩尔定律经济(Virtual Moore’s Law Economy, VME)将会蓬勃发展,通过优化物理、材料、器件、电路/芯片、软件和系统,通过一种新的硅方式来生产各种应用驱动的异构集成(HI)纳米系统,从而推动创新,为业务增长提供令人兴奋的应用。即使考虑到更昂贵的硅投资,半导体行业也将从新的应用和系统产品销售中获得足够的财务回报。这种基于(函数×值)向下加向上缩放方法的技术方法,以及线性缩放、面积缩放和体积缩放方法,可以从根本上改变思维和执行方式,以整体HIDAS (HI设计/架构/系统)方法连贯地优化技术定义和最终系统设计。这将推动集成电路扩展到有效的1纳米领域,刺激蓬勃发展的硅产业,至少有30年的时间增长到1万亿美元的规模。
{"title":"Keynote: A new Silicon Age 4.0: Generating semiconductor-intelligence paradigm with a Virtual Moore's Law Economics and Heterogeneous technologies","authors":"Nicky Liu","doi":"10.1109/ISLPED.2017.8009149","DOIUrl":"https://doi.org/10.1109/ISLPED.2017.8009149","url":null,"abstract":"The future of the silicon-based economy will not be as pessimistic as some commentators have argued, given their predictions of the end of Moore's Law Economy (ME) by the early 2020s. On the contrary, a Virtual Moore's Law Economy (VME) will develop and thrive, advancing innovation by a new Silicon Way of producing various application-driven Heterogeneous Integrated (HI) Nano-systems by optimization of physics, materials, devices, circuits/chips, software and systems to enable exciting applications for business growth. The semiconductor industry will enjoy sufficient financial returns from new application and system-product sales, even considering more expensive silicon investment. Such a technological approach based on a (Function × Value)-Scaling Down-Plus-Up Methodology, in addition to Linear-Scaling, Area-Scaling and Volumetric-Scaling Methodologies, can fundamentally change the way of thinking and execution toward optimizing coherently both technology definition and final system design with an holistic HIDAS (HI Design/Architecture/System) method. This will drive IC scaling to an effective 1-Nanometer Realm, stimulating a thriving silicon industry which can have at least 30 more years of growth toward a 1 trillion-dollar size.","PeriodicalId":20456,"journal":{"name":"Proceedings of the 2007 international symposium on Low power electronics and design (ISLPED '07)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2017-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82559147","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Message from the program co-chairs 来自项目联合主席的信息
J. Kulkarni, T. Wenisch
Over the past few years, however, issues related to the broader context for Dublin Core have come to the fore, such as XML and, since 1997, the Resource Description Framework, now part of a Semantic Web Activity; harvesting approaches such as Open Archives Initiative; domainspecific metadata standards and their relation to a “core”; and the constructs and policies necessary for managing namespaces and “application profiles” for automatic processing by machines.
{"title":"Message from the program co-chairs","authors":"J. Kulkarni, T. Wenisch","doi":"10.1109/ISLPED.2017.8009141","DOIUrl":"https://doi.org/10.1109/ISLPED.2017.8009141","url":null,"abstract":"Over the past few years, however, issues related to the broader context for Dublin Core have come to the fore, such as XML and, since 1997, the Resource Description Framework, now part of a Semantic Web Activity; harvesting approaches such as Open Archives Initiative; domainspecific metadata standards and their relation to a “core”; and the constructs and policies necessary for managing namespaces and “application profiles” for automatic processing by machines.","PeriodicalId":20456,"journal":{"name":"Proceedings of the 2007 international symposium on Low power electronics and design (ISLPED '07)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2017-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"79622275","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Let's get physical: Adding physical dimensions to cyber systems 让我们把物理维度添加到网络系统中
A. S. Vincentelli
Technology advances are creating major shifts in the industrial landscape. Traditional sectors such as transportation, medical and avionics, are witnessing fundamental changes in the supply chain and in the content where the interactions between the physical world and the computing world are becoming increasingly tight. Cyber Physical Systems, Systems of Systems, Internet of Things, Industrie 4.0, Swarm Systems and The Fog are all sectors that attract massive attention from the research communities and massive investment from industry. These concepts are tightly intertwined and describe a movement towards a fully interconnected planet where billions of devices interact via a complex mesh of wireless and wired communication infrastructures. The most compelling vision for the future of technology and industry is one where a swarm of devices is connected with the cloud to provide platforms for myriad of new applications. In this new world, new companies will arise and established ones will have to change radically their business model. The increasing sophistication and heterogeneity of these systems requires radical changes in the way sense-and-control platforms are designed to regulate them. In this presentation, I highlight some of the design challenges due to the complexity, heterogeneity and power consumption of CPS. Indeed, low power consumption is an essential requirement for the swarm of devices especially in the domain of wearable devices for healthcare. Coupled with low cost and reliability, power consumption has to be taken into consideration for any CPS deployment.
技术进步正在使工业格局发生重大变化。运输、医疗和航空电子等传统部门正在见证供应链和内容的根本变化,其中物理世界和计算世界之间的相互作用变得越来越紧密。网络物理系统、系统的系统、物联网、工业4.0、群系统和雾都是备受研究界关注和工业界大量投资的领域。这些概念紧密交织在一起,描述了一个走向完全互联的星球的运动,在这个星球上,数十亿设备通过无线和有线通信基础设施的复杂网络进行交互。技术和工业的未来最引人注目的愿景是大量设备与云连接,为无数新应用程序提供平台。在这个新世界里,新公司将会出现,老牌公司将不得不彻底改变它们的商业模式。这些系统的复杂性和异质性日益增加,需要对设计用于管理它们的感知和控制平台的方式进行根本性的改变。在这次演讲中,我将重点介绍由于CPS的复杂性、异质性和功耗而带来的一些设计挑战。事实上,低功耗是设备群的基本要求,特别是在医疗保健可穿戴设备领域。再加上低成本和可靠性,任何CPS部署都必须考虑功耗。
{"title":"Let's get physical: Adding physical dimensions to cyber systems","authors":"A. S. Vincentelli","doi":"10.1109/ISLPED.2015.7273478","DOIUrl":"https://doi.org/10.1109/ISLPED.2015.7273478","url":null,"abstract":"Technology advances are creating major shifts in the industrial landscape. Traditional sectors such as transportation, medical and avionics, are witnessing fundamental changes in the supply chain and in the content where the interactions between the physical world and the computing world are becoming increasingly tight. Cyber Physical Systems, Systems of Systems, Internet of Things, Industrie 4.0, Swarm Systems and The Fog are all sectors that attract massive attention from the research communities and massive investment from industry. These concepts are tightly intertwined and describe a movement towards a fully interconnected planet where billions of devices interact via a complex mesh of wireless and wired communication infrastructures. The most compelling vision for the future of technology and industry is one where a swarm of devices is connected with the cloud to provide platforms for myriad of new applications. In this new world, new companies will arise and established ones will have to change radically their business model. The increasing sophistication and heterogeneity of these systems requires radical changes in the way sense-and-control platforms are designed to regulate them. In this presentation, I highlight some of the design challenges due to the complexity, heterogeneity and power consumption of CPS. Indeed, low power consumption is an essential requirement for the swarm of devices especially in the domain of wearable devices for healthcare. Coupled with low cost and reliability, power consumption has to be taken into consideration for any CPS deployment.","PeriodicalId":20456,"journal":{"name":"Proceedings of the 2007 international symposium on Low power electronics and design (ISLPED '07)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2015-07-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82244860","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 19
Opportunities in system power management for high performance mixed signal platforms 高性能混合信号平台的系统电源管理机会
Jose Pineda de Jyvez
In an era in which foundry technologies are a commodity, product differentiation comes by design. As fabless becomes mainstream, a paradigm shift demands innovation and collaboration among industrial and research thinking to further lower the cost of ICs, as well as to address upcoming power-performance challenges. High performance mixed signal (HPMS) platforms require stringent overall system and subsystem performance. The ability to design ultra-low power systems is used in a wide range of platforms including consumer, mobile, identification, healthcare products and microcontrollers. This presentation explores low power design techniques, challenges and opportunities faced in an industrial research environment. The overview addresses design tradeoffs, design implications, and measurement results during their deployment in HPMS platforms.
在一个铸造技术成为商品的时代,产品的差异化来自于设计。随着无晶圆厂成为主流,一种范式的转变需要工业和研究思维之间的创新和合作,以进一步降低集成电路的成本,并应对即将到来的功耗性能挑战。高性能混合信号(HPMS)平台要求严格的系统和子系统的整体性能。设计超低功耗系统的能力被广泛应用于各种平台,包括消费、移动、识别、医疗保健产品和微控制器。本演讲探讨了低功耗设计技术、工业研究环境中面临的挑战和机遇。概述了在HPMS平台部署期间的设计权衡、设计含义和测量结果。
{"title":"Opportunities in system power management for high performance mixed signal platforms","authors":"Jose Pineda de Jyvez","doi":"10.1109/ISLPED.2015.7273479","DOIUrl":"https://doi.org/10.1109/ISLPED.2015.7273479","url":null,"abstract":"In an era in which foundry technologies are a commodity, product differentiation comes by design. As fabless becomes mainstream, a paradigm shift demands innovation and collaboration among industrial and research thinking to further lower the cost of ICs, as well as to address upcoming power-performance challenges. High performance mixed signal (HPMS) platforms require stringent overall system and subsystem performance. The ability to design ultra-low power systems is used in a wide range of platforms including consumer, mobile, identification, healthcare products and microcontrollers. This presentation explores low power design techniques, challenges and opportunities faced in an industrial research environment. The overview addresses design tradeoffs, design implications, and measurement results during their deployment in HPMS platforms.","PeriodicalId":20456,"journal":{"name":"Proceedings of the 2007 international symposium on Low power electronics and design (ISLPED '07)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2015-07-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"81466671","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Proceedings of the 2007 international symposium on Low power electronics and design (ISLPED '07)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1