首页 > 最新文献

2020 Design, Automation & Test in Europe Conference & Exhibition (DATE)最新文献

英文 中文
On the Task Mapping and Scheduling for DAG-based Embedded Vision Applications on Heterogeneous Multi/Many-core Architectures 异构多核/多核架构下基于dag的嵌入式视觉应用任务映射与调度研究
Pub Date : 2020-03-01 DOI: 10.23919/DATE48585.2020.9116462
Stefano Aldegheri, N. Bombieri, Hiren D. Patel
In this work, we show that applying the heterogeneous earliest finish time (HEFT) heuristic for the task scheduling of embedded vision applications can improve the system performance up to 70% w.r.t. the scheduling solutions at the state of the art. We propose an algorithm called exclusive earliest finish time (XEFT) that introduces the notion of exclusive overlap between application primitives to improve the load balancing. We show that XEFT can improve the system performance up to 33% over HEFT, and 82% over the state of the art approaches. We present the results on different benchmarks, including a real-world localization and mapping application (ORB-SLAM) combined with the NVIDIA object detection application based on deep-learning.
在这项工作中,我们证明了将异构最早完成时间(HEFT)启发式方法应用于嵌入式视觉应用的任务调度可以将系统性能提高到目前最先进的调度解决方案的70%。我们提出了一种称为排他性最早完成时间(XEFT)的算法,它引入了应用程序原语之间的排他性重叠的概念,以改善负载平衡。我们表明,XEFT可以比HEFT提高33%的系统性能,比目前最先进的方法提高82%。我们展示了不同基准测试的结果,包括一个现实世界的定位和地图应用程序(ORB-SLAM)与基于深度学习的NVIDIA目标检测应用程序相结合。
{"title":"On the Task Mapping and Scheduling for DAG-based Embedded Vision Applications on Heterogeneous Multi/Many-core Architectures","authors":"Stefano Aldegheri, N. Bombieri, Hiren D. Patel","doi":"10.23919/DATE48585.2020.9116462","DOIUrl":"https://doi.org/10.23919/DATE48585.2020.9116462","url":null,"abstract":"In this work, we show that applying the heterogeneous earliest finish time (HEFT) heuristic for the task scheduling of embedded vision applications can improve the system performance up to 70% w.r.t. the scheduling solutions at the state of the art. We propose an algorithm called exclusive earliest finish time (XEFT) that introduces the notion of exclusive overlap between application primitives to improve the load balancing. We show that XEFT can improve the system performance up to 33% over HEFT, and 82% over the state of the art approaches. We present the results on different benchmarks, including a real-world localization and mapping application (ORB-SLAM) combined with the NVIDIA object detection application based on deep-learning.","PeriodicalId":289525,"journal":{"name":"2020 Design, Automation & Test in Europe Conference & Exhibition (DATE)","volume":"5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131072278","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Towards Specification and Testing of RISC-V ISA Compliance⋆ RISC-V ISA合规性规范与测试
Pub Date : 2020-03-01 DOI: 10.23919/DATE48585.2020.9116193
V. Herdt, Daniel Große, R. Drechsler
Compliance testing for RISC-V is very important. Therefore, an official hand-written compliance test-suite is being actively developed. However, this requires significant manual effort in particular to achieve a high test coverage.In this paper we propose a test-suite specification mechanism in combination with a first set of instruction constraints and coverage requirements for the base RISC-V ISA. In addition, we present an automated method to generate a test-suite that satisfies the specification. Our evaluation demonstrates the effectiveness and potential of our method.
RISC-V的合规性测试非常重要。因此,官方手写的法规遵循测试套件正在积极开发中。然而,这需要大量的手工工作来实现高测试覆盖率。在本文中,我们提出了一种测试套件规范机制,结合第一组指令约束和基本RISC-V ISA的覆盖要求。另外,我们提出了一种自动化的方法来生成满足规范的测试套件。我们的评估证明了我们的方法的有效性和潜力。
{"title":"Towards Specification and Testing of RISC-V ISA Compliance⋆","authors":"V. Herdt, Daniel Große, R. Drechsler","doi":"10.23919/DATE48585.2020.9116193","DOIUrl":"https://doi.org/10.23919/DATE48585.2020.9116193","url":null,"abstract":"Compliance testing for RISC-V is very important. Therefore, an official hand-written compliance test-suite is being actively developed. However, this requires significant manual effort in particular to achieve a high test coverage.In this paper we propose a test-suite specification mechanism in combination with a first set of instruction constraints and coverage requirements for the base RISC-V ISA. In addition, we present an automated method to generate a test-suite that satisfies the specification. Our evaluation demonstrates the effectiveness and potential of our method.","PeriodicalId":289525,"journal":{"name":"2020 Design, Automation & Test in Europe Conference & Exhibition (DATE)","volume":"356 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132895100","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 16
Lazy Event Prediction using Defining Trees and Schedule Bypass for Out-of-Order PDES 对乱序PDES使用定义树和调度旁路的延迟事件预测
Pub Date : 2020-03-01 DOI: 10.23919/DATE48585.2020.9116512
Daniel Mendoza, Zhongqi Cheng, E. Arasteh, R. Dömer
Out-of-order parallel discrete event simulation (PDES) has been shown to be very effective in speeding up system design by utilizing parallel processors on multi- and many-core hosts. As the number of threads in the design model grows larger, however, the original scheduling approach does not scale. In this work, we analyze the out-of-order scheduler and identify a bottleneck with quadratic complexity in event prediction. We propose a more efficient lazy strategy based on defining trees and a schedule bypass with O(m log2 m) complexity which shows sustained and improved performance gains in simulation of SystemC models with many processes. For models containing over 1000 processes, experimental results show simulation run time speedups of up to 90x using lazy event prediction against the original out-of-order PDES approach.
乱序并行离散事件仿真(PDES)通过在多核和多核主机上使用并行处理器,在加速系统设计方面非常有效。然而,当设计模型中的线程数量增加时,原来的调度方法无法扩展。在本文中,我们分析了乱序调度程序,并确定了事件预测中具有二次复杂度的瓶颈。我们提出了一种更有效的懒惰策略,基于定义树和复杂度为0 (m log2 m)的调度绕过,该策略在具有许多进程的SystemC模型仿真中显示出持续和改进的性能增益。对于包含超过1000个进程的模型,实验结果表明,与原始的无序PDES方法相比,使用延迟事件预测,模拟运行时速度可提高90倍。
{"title":"Lazy Event Prediction using Defining Trees and Schedule Bypass for Out-of-Order PDES","authors":"Daniel Mendoza, Zhongqi Cheng, E. Arasteh, R. Dömer","doi":"10.23919/DATE48585.2020.9116512","DOIUrl":"https://doi.org/10.23919/DATE48585.2020.9116512","url":null,"abstract":"Out-of-order parallel discrete event simulation (PDES) has been shown to be very effective in speeding up system design by utilizing parallel processors on multi- and many-core hosts. As the number of threads in the design model grows larger, however, the original scheduling approach does not scale. In this work, we analyze the out-of-order scheduler and identify a bottleneck with quadratic complexity in event prediction. We propose a more efficient lazy strategy based on defining trees and a schedule bypass with O(m log2 m) complexity which shows sustained and improved performance gains in simulation of SystemC models with many processes. For models containing over 1000 processes, experimental results show simulation run time speedups of up to 90x using lazy event prediction against the original out-of-order PDES approach.","PeriodicalId":289525,"journal":{"name":"2020 Design, Automation & Test in Europe Conference & Exhibition (DATE)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134582430","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
L2L: A Highly Accurate Log_2_Lead Quantization of Pre-trained Neural Networks 一种高精度的预训练神经网络log__lead量化
Pub Date : 2020-03-01 DOI: 10.23919/DATE48585.2020.9116373
Salim Ullah, Siddharth Gupta, K. Ahuja, Aruna Tiwari, Akash Kumar
Deep Neural Networks are one of the machine learning techniques which are increasingly used in a variety of applications. However, the significantly high memory and computation demands of deep neural networks often limit their deployment on embedded systems. Many recent works have considered this problem by proposing different types of data quantization schemes. However, most of these techniques either require post-quantization retraining of deep neural networks or bear a significant loss in output accuracy. In this paper, we propose a novel quantization technique for parameters of pre-trained deep neural networks. Our technique significantly maintains the accuracy of the parameters and does not require retraining of the networks. Compared to the single-precision floating-point numbers-based implementation, our proposed 8-bit quantization technique generates only ~1% and the ~0.4%, loss in top-1 and top-5 accuracies respectively for VGG16 network using ImageNet dataset.
深度神经网络是机器学习技术的一种,在各种应用中得到越来越多的应用。然而,深度神经网络的高内存和计算需求往往限制了其在嵌入式系统上的部署。最近的许多工作都通过提出不同类型的数据量化方案来考虑这个问题。然而,这些技术要么需要对深度神经网络进行量化后的再训练,要么在输出精度上有很大的损失。本文提出了一种新的深度神经网络参数量化方法。我们的技术显著地保持了参数的准确性,并且不需要对网络进行再训练。与基于单精度浮点数的实现相比,我们提出的8位量化技术在使用ImageNet数据集的VGG16网络中,top-1和top-5精度的损失分别为~1%和~0.4%。
{"title":"L2L: A Highly Accurate Log_2_Lead Quantization of Pre-trained Neural Networks","authors":"Salim Ullah, Siddharth Gupta, K. Ahuja, Aruna Tiwari, Akash Kumar","doi":"10.23919/DATE48585.2020.9116373","DOIUrl":"https://doi.org/10.23919/DATE48585.2020.9116373","url":null,"abstract":"Deep Neural Networks are one of the machine learning techniques which are increasingly used in a variety of applications. However, the significantly high memory and computation demands of deep neural networks often limit their deployment on embedded systems. Many recent works have considered this problem by proposing different types of data quantization schemes. However, most of these techniques either require post-quantization retraining of deep neural networks or bear a significant loss in output accuracy. In this paper, we propose a novel quantization technique for parameters of pre-trained deep neural networks. Our technique significantly maintains the accuracy of the parameters and does not require retraining of the networks. Compared to the single-precision floating-point numbers-based implementation, our proposed 8-bit quantization technique generates only ~1% and the ~0.4%, loss in top-1 and top-5 accuracies respectively for VGG16 network using ImageNet dataset.","PeriodicalId":289525,"journal":{"name":"2020 Design, Automation & Test in Europe Conference & Exhibition (DATE)","volume":"2011 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125633733","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Opportunities for Cross-Layer Design in High-Performance Computing Systems with Integrated Silicon Photonic Networks 集成硅光子网络高性能计算系统跨层设计的机遇
Pub Date : 2020-03-01 DOI: 10.23919/DATE48585.2020.9116234
Asif Mirza, Shadi Manafi Avari, Ebadollah Taheri, S. Pasricha, M. Nikdast
With the ever growing complexity of high-performance computing (HPC) systems to satisfy emerging application requirements (e.g., high memory bandwidth requirement for machine learning applications), the performance bottleneck in such systems has moved from being computation-centric to be more communication-centric. Silicon photonic interconnection networks have been proposed to address the aggressive communication requirements in HPC systems, to realize higher bandwidth, lower latency, and better energy efficiency. There have been many successful efforts on developing silicon photonic devices, integrated circuits, and architectures for HPC systems. Moreover, many efforts have been made to address and mitigate the impact of different challenges (e.g., fabrication process and thermal variations) in silicon photonic interconnects. However, most of these efforts have focused only on a single design layer in the system design space (e.g., device, circuit or architecture level). Therefore, there is often a gap between what a design technique can improve in one layer, and what it might impair in another one. In this paper, we discuss the promise of cross-layer design methodologies for HPC systems integrating silicon photonic interconnects. In particular, we discuss how such cross-layer design solutions based on cooperatively designing and exchanging design objectives among different system design layers can help achieve the best possible performance when integrating silicon photonics into HPC systems.
随着高性能计算(HPC)系统不断增长的复杂性,以满足新兴的应用需求(例如,机器学习应用程序的高内存带宽需求),这些系统的性能瓶颈已经从以计算为中心转向以通信为中心。硅光子互连网络的提出是为了解决高性能计算系统中积极的通信需求,以实现更高的带宽,更低的延迟和更好的能源效率。在开发用于高性能计算系统的硅光子器件、集成电路和体系结构方面已经有了许多成功的努力。此外,已经做出了许多努力来解决和减轻硅光子互连中不同挑战(例如,制造工艺和热变化)的影响。然而,大多数这些努力只集中在系统设计空间中的单个设计层(例如,器件,电路或体系结构级)。因此,设计技术在某一层中所能改善的内容与在另一层中可能损害的内容之间往往存在差距。在本文中,我们讨论了集成硅光子互连的高性能计算系统的跨层设计方法的前景。我们特别讨论了这种基于协作设计和不同系统设计层之间交换设计目标的跨层设计解决方案如何在将硅光子学集成到高性能计算系统中时帮助实现最佳性能。
{"title":"Opportunities for Cross-Layer Design in High-Performance Computing Systems with Integrated Silicon Photonic Networks","authors":"Asif Mirza, Shadi Manafi Avari, Ebadollah Taheri, S. Pasricha, M. Nikdast","doi":"10.23919/DATE48585.2020.9116234","DOIUrl":"https://doi.org/10.23919/DATE48585.2020.9116234","url":null,"abstract":"With the ever growing complexity of high-performance computing (HPC) systems to satisfy emerging application requirements (e.g., high memory bandwidth requirement for machine learning applications), the performance bottleneck in such systems has moved from being computation-centric to be more communication-centric. Silicon photonic interconnection networks have been proposed to address the aggressive communication requirements in HPC systems, to realize higher bandwidth, lower latency, and better energy efficiency. There have been many successful efforts on developing silicon photonic devices, integrated circuits, and architectures for HPC systems. Moreover, many efforts have been made to address and mitigate the impact of different challenges (e.g., fabrication process and thermal variations) in silicon photonic interconnects. However, most of these efforts have focused only on a single design layer in the system design space (e.g., device, circuit or architecture level). Therefore, there is often a gap between what a design technique can improve in one layer, and what it might impair in another one. In this paper, we discuss the promise of cross-layer design methodologies for HPC systems integrating silicon photonic interconnects. In particular, we discuss how such cross-layer design solutions based on cooperatively designing and exchanging design objectives among different system design layers can help achieve the best possible performance when integrating silicon photonics into HPC systems.","PeriodicalId":289525,"journal":{"name":"2020 Design, Automation & Test in Europe Conference & Exhibition (DATE)","volume":"70 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130297379","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Emergent Control of MPSoC Operation by a Hierarchical Supervisor / Reinforcement Learning Approach 基于分层监督/强化学习方法的MPSoC运行紧急控制
Pub Date : 2020-03-01 DOI: 10.23919/DATE48585.2020.9116574
F. Maurer, Bryan Donyanavard, A. Rahmani, N. Dutt, A. Herkersdorf
MPSoCs increasingly depend on adaptive resource management strategies at runtime for efficient utilization of resources when executing complex application workloads. In particular, conflicting demands for adequate computation performance and power-/energy-efficiency constraints make desired application goals hard to achieve. We present a hierarchical, cross-layer hardware/software resource manager capable of adapting to changing workloads and system dynamics with zero initial knowledge. The manager uses rule-based reinforcement learning classifier tables (LCTs) with an archive-based backup policy as leaf controllers. The LCTs directly manipulate and enforce MPSoC building block operation parameters in order to explore and optimize potentially conflicting system requirements (e.g., meeting a performance target while staying within the power constraint). A supervisor translates system requirements and application goals into per-LCT objective functions (e.g., core instructions-per-second (IPS). Thus, the supervisor manages the possibly emergent behavior of the low-level LCT controllers in response to 1) switching between operation strategies (e.g., maximize performance vs. minimize power; and 2) changing application requirements. This hierarchical manager leverages the dual benefits of a software supervisor (enabling flexibility), together with hardware learners (allowing quick and efficient optimization). Experiments on an FPGA prototype confirmed the ability of our approach to identify optimized MPSoC operation parameters at runtime while strictly obeying given power constraints.
mpsoc在运行时越来越依赖于自适应资源管理策略,以便在执行复杂的应用程序工作负载时有效地利用资源。特别是,对足够的计算性能和功率/能源效率约束的冲突需求使得期望的应用程序目标难以实现。我们提出了一种分层的、跨层的硬件/软件资源管理器,能够适应不断变化的工作负载和系统动态,而无需初始知识。管理器使用基于规则的强化学习分类器表(lct)和基于存档的备份策略作为叶子控制器。lct直接操纵和执行MPSoC构建块操作参数,以探索和优化潜在的冲突系统需求(例如,在保持功率约束的同时满足性能目标)。主管将系统需求和应用程序目标转换为每个lct的目标函数(例如,每秒核心指令数(IPS))。因此,管理者管理低层LCT控制器可能出现的紧急行为,以响应1)操作策略之间的切换(例如,性能最大化vs功率最小化;2)不断变化的应用需求。这个分层管理器利用了软件管理器(实现灵活性)和硬件学习器(允许快速有效的优化)的双重好处。在FPGA原型上的实验证实了我们的方法能够在严格遵守给定功率限制的情况下,在运行时识别优化的MPSoC操作参数。
{"title":"Emergent Control of MPSoC Operation by a Hierarchical Supervisor / Reinforcement Learning Approach","authors":"F. Maurer, Bryan Donyanavard, A. Rahmani, N. Dutt, A. Herkersdorf","doi":"10.23919/DATE48585.2020.9116574","DOIUrl":"https://doi.org/10.23919/DATE48585.2020.9116574","url":null,"abstract":"MPSoCs increasingly depend on adaptive resource management strategies at runtime for efficient utilization of resources when executing complex application workloads. In particular, conflicting demands for adequate computation performance and power-/energy-efficiency constraints make desired application goals hard to achieve. We present a hierarchical, cross-layer hardware/software resource manager capable of adapting to changing workloads and system dynamics with zero initial knowledge. The manager uses rule-based reinforcement learning classifier tables (LCTs) with an archive-based backup policy as leaf controllers. The LCTs directly manipulate and enforce MPSoC building block operation parameters in order to explore and optimize potentially conflicting system requirements (e.g., meeting a performance target while staying within the power constraint). A supervisor translates system requirements and application goals into per-LCT objective functions (e.g., core instructions-per-second (IPS). Thus, the supervisor manages the possibly emergent behavior of the low-level LCT controllers in response to 1) switching between operation strategies (e.g., maximize performance vs. minimize power; and 2) changing application requirements. This hierarchical manager leverages the dual benefits of a software supervisor (enabling flexibility), together with hardware learners (allowing quick and efficient optimization). Experiments on an FPGA prototype confirmed the ability of our approach to identify optimized MPSoC operation parameters at runtime while strictly obeying given power constraints.","PeriodicalId":289525,"journal":{"name":"2020 Design, Automation & Test in Europe Conference & Exhibition (DATE)","volume":"217 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132310413","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
In-Memory Resistive RAM Implementation of Binarized Neural Networks for Medical Applications 用于医疗应用的二值化神经网络的内存电阻性RAM实现
Pub Date : 2020-03-01 DOI: 10.23919/DATE48585.2020.9116439
Bogdan Penkovsky, M. Bocquet, T. Hirtzlin, Jacques-Olivier Klein, E. Nowak, E. Vianello, J. Portal, D. Querlioz
The advent of deep learning has considerably accelerated machine learning development. The deployment of deep neural networks at the edge is however limited by their high memory and energy consumption requirements. With new memory technology available, emerging Binarized Neural Networks (BNNs) are promising to reduce the energy impact of the forthcoming machine learning hardware generation, enabling machine learning on the edge devices and avoiding data transfer over the network. In this work, after presenting our implementation employing a hybrid CMOS - hafnium oxide resistive memory technology, we suggest strategies to apply BNNs to biomedical signals such as electrocardiography and electroencephalography, keeping accuracy level and reducing memory requirements. We investigate the memory-accuracy trade-off when binarizing whole network and binarizing solely the classifier part. We also discuss how these results translate to the edge-oriented Mobilenet V1 neural network on the Imagenet task. The final goal of this research is to enable smart autonomous healthcare devices.
深度学习的出现大大加速了机器学习的发展。然而,边缘深度神经网络的部署受到其高内存和能耗要求的限制。随着新的存储技术的出现,新兴的二值化神经网络(bnn)有望减少即将到来的机器学习硬件产生的能量影响,从而在边缘设备上实现机器学习,并避免通过网络传输数据。在这项工作中,在介绍了我们采用混合CMOS -氧化铪电阻式记忆技术的实现之后,我们提出了将bnn应用于心电图和脑电图等生物医学信号的策略,以保持准确性水平并降低记忆要求。我们研究了二值化整个网络和二值化分类器部分时的记忆-精度权衡。我们还讨论了这些结果如何在Imagenet任务上转化为面向边缘的Mobilenet V1神经网络。本研究的最终目标是实现智能自主医疗保健设备。
{"title":"In-Memory Resistive RAM Implementation of Binarized Neural Networks for Medical Applications","authors":"Bogdan Penkovsky, M. Bocquet, T. Hirtzlin, Jacques-Olivier Klein, E. Nowak, E. Vianello, J. Portal, D. Querlioz","doi":"10.23919/DATE48585.2020.9116439","DOIUrl":"https://doi.org/10.23919/DATE48585.2020.9116439","url":null,"abstract":"The advent of deep learning has considerably accelerated machine learning development. The deployment of deep neural networks at the edge is however limited by their high memory and energy consumption requirements. With new memory technology available, emerging Binarized Neural Networks (BNNs) are promising to reduce the energy impact of the forthcoming machine learning hardware generation, enabling machine learning on the edge devices and avoiding data transfer over the network. In this work, after presenting our implementation employing a hybrid CMOS - hafnium oxide resistive memory technology, we suggest strategies to apply BNNs to biomedical signals such as electrocardiography and electroencephalography, keeping accuracy level and reducing memory requirements. We investigate the memory-accuracy trade-off when binarizing whole network and binarizing solely the classifier part. We also discuss how these results translate to the edge-oriented Mobilenet V1 neural network on the Imagenet task. The final goal of this research is to enable smart autonomous healthcare devices.","PeriodicalId":289525,"journal":{"name":"2020 Design, Automation & Test in Europe Conference & Exhibition (DATE)","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114055454","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Exact DAG-Aware Rewriting 精确的dag感知重写
Pub Date : 2020-03-01 DOI: 10.23919/DATE48585.2020.9116379
Heinz Riener, A. Mishchenko, Mathias Soeken
We present a generic resynthesis framework for optimizing Boolean networks parameterized with a multi-level logic representation, a cut-computation algorithm, and a resynthesis algorithm. The framework allows us to realize powerful optimization algorithms in a plug-and-play fashion. We show the framework’s versatility by composing an exact DAG-aware rewriting engine. Disjoint-support decomposition and SAT-based exact synthesis together with efficient caching strategies enable the algorithm to resynthesize larger parts of the logic. DAGaware rewriting is used to compute the gain of resynthesis while taking the benefit of structural hashing into account.
我们提出了一个通用的重组框架,用于优化布尔网络参数化的多层次逻辑表示,切割计算算法和重组算法。该框架允许我们以即插即用的方式实现强大的优化算法。我们通过编写一个精确的支持dag的重写引擎来展示框架的多功能性。分离支持分解和基于sat的精确合成以及高效的缓存策略使算法能够重新合成更大的逻辑部分。在考虑结构哈希的好处的同时,使用感知重写来计算重合成的增益。
{"title":"Exact DAG-Aware Rewriting","authors":"Heinz Riener, A. Mishchenko, Mathias Soeken","doi":"10.23919/DATE48585.2020.9116379","DOIUrl":"https://doi.org/10.23919/DATE48585.2020.9116379","url":null,"abstract":"We present a generic resynthesis framework for optimizing Boolean networks parameterized with a multi-level logic representation, a cut-computation algorithm, and a resynthesis algorithm. The framework allows us to realize powerful optimization algorithms in a plug-and-play fashion. We show the framework’s versatility by composing an exact DAG-aware rewriting engine. Disjoint-support decomposition and SAT-based exact synthesis together with efficient caching strategies enable the algorithm to resynthesize larger parts of the logic. DAGaware rewriting is used to compute the gain of resynthesis while taking the benefit of structural hashing into account.","PeriodicalId":289525,"journal":{"name":"2020 Design, Automation & Test in Europe Conference & Exhibition (DATE)","volume":"39 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127997810","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Network Synthesis for Industry 4.0 面向工业4.0的网络综合
Pub Date : 2020-03-01 DOI: 10.23919/DATE48585.2020.9116407
Enrico Fraccaroli, Alan Michael Padovani, D. Quaglia, F. Fummi
Today’s factory machines are ever more connected with SCADA, MES, ERP applications as well as external systems for data analysis. Different types of network architectures must be used for this purpose. For instance, control applications at the lowest level are susceptible to delays and errors while data analysis with machine learning procedures requires to move a large amount of data without real-time constraints. Standard data formats, like Automation Markup Language (AML), have been established to document factory environment, machine placement and network deployment, however, no automatic technique is currently available in the context of Industry 4.0 to choose the best mix of network architectures according to spacial constraints, cost, and performance. We propose to fill this gap by formulating an optimization problem. First of all, spatial and communication requirements are extracted from the AML description. Then, the optimal interconnection of wired or wireless channels is obtained according to application objectives. Finally, this result is back-annotated to AML to be used in the life cycle of the production system. The proposed methodology is described through a small, but complete, smart production plant.
今天的工厂机器与SCADA、MES、ERP应用程序以及用于数据分析的外部系统的连接越来越紧密。为此,必须使用不同类型的网络体系结构。例如,最低级别的控制应用程序容易受到延迟和错误的影响,而使用机器学习过程的数据分析需要在没有实时限制的情况下移动大量数据。标准的数据格式,如自动化标记语言(AML),已经被建立来记录工厂环境、机器放置和网络部署,然而,在工业4.0的背景下,目前还没有自动化技术可以根据空间限制、成本和性能选择网络架构的最佳组合。我们建议通过制定一个优化问题来填补这一空白。首先,从AML描述中提取空间和通信需求。然后,根据应用目标得到有线或无线信道的最优互连。最后,将此结果反向注释为AML,以便在生产系统的生命周期中使用。提出的方法是通过一个小而完整的智能生产工厂来描述的。
{"title":"Network Synthesis for Industry 4.0","authors":"Enrico Fraccaroli, Alan Michael Padovani, D. Quaglia, F. Fummi","doi":"10.23919/DATE48585.2020.9116407","DOIUrl":"https://doi.org/10.23919/DATE48585.2020.9116407","url":null,"abstract":"Today’s factory machines are ever more connected with SCADA, MES, ERP applications as well as external systems for data analysis. Different types of network architectures must be used for this purpose. For instance, control applications at the lowest level are susceptible to delays and errors while data analysis with machine learning procedures requires to move a large amount of data without real-time constraints. Standard data formats, like Automation Markup Language (AML), have been established to document factory environment, machine placement and network deployment, however, no automatic technique is currently available in the context of Industry 4.0 to choose the best mix of network architectures according to spacial constraints, cost, and performance. We propose to fill this gap by formulating an optimization problem. First of all, spatial and communication requirements are extracted from the AML description. Then, the optimal interconnection of wired or wireless channels is obtained according to application objectives. Finally, this result is back-annotated to AML to be used in the life cycle of the production system. The proposed methodology is described through a small, but complete, smart production plant.","PeriodicalId":289525,"journal":{"name":"2020 Design, Automation & Test in Europe Conference & Exhibition (DATE)","volume":"49 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121433666","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Design-flow Methodology for Secure Group Anonymous Authentication 安全组匿名认证的设计流程方法
Pub Date : 2020-03-01 DOI: 10.23919/DATE48585.2020.9116290
R. Agrawal, Lake Bu, Eliakin Del Rosario, M. Kinsy
In heterogeneous distributed systems, computing devices and software components often come from different providers and have different security, trust, and privacy levels. In many of these systems, the need frequently arises to (i) control the access to services and resources granted to individual devices or components in a context-aware manner and (ii) establish and enforce data sharing policies that preserve the privacy of the critical information on end users. In essence, the need is to authenticate and anonymize an entity or device simultaneously, two seemingly contradictory goals. The design challenge is further complicated by potential security problems, such as man-in-the-middle attacks, hijacked devices, and counterfeits. In this work, we present a system design flow for a trustworthy group anonymous authentication protocol (GAAP), which not only fulfills the desired functionality for authentication and privacy, but also provides strong security guarantees.
在异构分布式系统中,计算设备和软件组件通常来自不同的提供商,具有不同的安全性、信任和隐私级别。在许多这样的系统中,经常需要(i)以上下文感知的方式控制对授予单个设备或组件的服务和资源的访问,以及(ii)建立和执行数据共享策略,以保护最终用户关键信息的隐私。从本质上讲,需要同时对实体或设备进行身份验证和匿名化,这是两个看似矛盾的目标。潜在的安全问题(如中间人攻击、被劫持的设备和假冒产品)使设计挑战进一步复杂化。在这项工作中,我们提出了一个可信组匿名认证协议(GAAP)的系统设计流程,该协议不仅满足了期望的认证和隐私功能,而且提供了强大的安全保证。
{"title":"Design-flow Methodology for Secure Group Anonymous Authentication","authors":"R. Agrawal, Lake Bu, Eliakin Del Rosario, M. Kinsy","doi":"10.23919/DATE48585.2020.9116290","DOIUrl":"https://doi.org/10.23919/DATE48585.2020.9116290","url":null,"abstract":"In heterogeneous distributed systems, computing devices and software components often come from different providers and have different security, trust, and privacy levels. In many of these systems, the need frequently arises to (i) control the access to services and resources granted to individual devices or components in a context-aware manner and (ii) establish and enforce data sharing policies that preserve the privacy of the critical information on end users. In essence, the need is to authenticate and anonymize an entity or device simultaneously, two seemingly contradictory goals. The design challenge is further complicated by potential security problems, such as man-in-the-middle attacks, hijacked devices, and counterfeits. In this work, we present a system design flow for a trustworthy group anonymous authentication protocol (GAAP), which not only fulfills the desired functionality for authentication and privacy, but also provides strong security guarantees.","PeriodicalId":289525,"journal":{"name":"2020 Design, Automation & Test in Europe Conference & Exhibition (DATE)","volume":"129 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115965098","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
期刊
2020 Design, Automation & Test in Europe Conference & Exhibition (DATE)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1