首页 > 最新文献

Proceedings of the 2nd International Workshop on Embedded and Mobile Deep Learning最新文献

英文 中文
Mobile Based Continuous Authentication Using Deep Features 使用深度特征的基于移动的连续认证
Mario Parreño Centeno, Yu Guan, A. Moorsel
Continuous authentication is a promising approach to validate the user's identity during a work session, e.g., for mobile banking applications. Recently, it has been demonstrated that changes in the motion patterns of the user may help to note the unauthorised use of mobile devices. Several approaches have been proposed in this area but with relatively weak performance results. In this work, we propose an approach which uses a Siamese convolutional neural network to learn the signatures of the motion patterns from users and achieve a competitive verification accuracy up to 97.8%. We also find our algorithm is not very sensitive to sampling frequency and the length of the sequence.
连续身份验证是在工作会话期间验证用户身份的一种很有前途的方法,例如,用于移动银行应用程序。最近,有研究表明,用户运动模式的变化可能有助于注意到未经授权使用移动设备。在这方面已经提出了几种方法,但性能结果相对较弱。在这项工作中,我们提出了一种使用暹罗卷积神经网络从用户那里学习运动模式签名的方法,并实现了高达97.8%的竞争性验证准确率。我们还发现我们的算法对采样频率和序列长度不是很敏感。
{"title":"Mobile Based Continuous Authentication Using Deep Features","authors":"Mario Parreño Centeno, Yu Guan, A. Moorsel","doi":"10.1145/3212725.3212732","DOIUrl":"https://doi.org/10.1145/3212725.3212732","url":null,"abstract":"Continuous authentication is a promising approach to validate the user's identity during a work session, e.g., for mobile banking applications. Recently, it has been demonstrated that changes in the motion patterns of the user may help to note the unauthorised use of mobile devices. Several approaches have been proposed in this area but with relatively weak performance results. In this work, we propose an approach which uses a Siamese convolutional neural network to learn the signatures of the motion patterns from users and achieve a competitive verification accuracy up to 97.8%. We also find our algorithm is not very sensitive to sampling frequency and the length of the sequence.","PeriodicalId":419019,"journal":{"name":"Proceedings of the 2nd International Workshop on Embedded and Mobile Deep Learning","volume":"66 ","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131550514","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 36
Proceedings of the 2nd International Workshop on Embedded and Mobile Deep Learning 第二届嵌入式和移动深度学习国际研讨会论文集
Hongkai Wen, Petko Georgiev, Erran L. Li, Samir Kumar, A. Balasubramanian, Youngki Lee
In recent years, breakthroughs from the field of deep learning have transformed how sensor data (e.g., images, audio, and even accelerometers and GPS) can be interpreted to extract the high-level information needed by bleeding-edge sensor-driven systems like smartphone apps and wearable devices. Today, the state-of-the-art in computational models that, for example, recognize a face, track user emotions, or monitor physical activities are increasingly based on deep learning principles and algorithms. Unfortunately, deep models typically exert severe demands on local device resources and this conventionally limits their adoption within mobile and embedded platforms. As a result, in far too many cases existing systems process sensor data with machine learning methods that have been superseded by deep learning years ago.
近年来,深度学习领域的突破已经改变了传感器数据(例如图像,音频,甚至加速度计和GPS)的解释方式,以提取智能手机应用程序和可穿戴设备等前沿传感器驱动系统所需的高级信息。如今,最先进的计算模型,例如识别人脸、跟踪用户情绪或监控身体活动,越来越多地基于深度学习原理和算法。不幸的是,深度模型通常会对本地设备资源产生严重的需求,这通常限制了它们在移动和嵌入式平台中的应用。因此,在很多情况下,现有系统使用机器学习方法处理传感器数据,而这些方法在几年前就被深度学习所取代。
{"title":"Proceedings of the 2nd International Workshop on Embedded and Mobile Deep Learning","authors":"Hongkai Wen, Petko Georgiev, Erran L. Li, Samir Kumar, A. Balasubramanian, Youngki Lee","doi":"10.1145/3212725","DOIUrl":"https://doi.org/10.1145/3212725","url":null,"abstract":"In recent years, breakthroughs from the field of deep learning have transformed how sensor data (e.g., images, audio, and even accelerometers and GPS) can be interpreted to extract the high-level information needed by bleeding-edge sensor-driven systems like smartphone apps and wearable devices. Today, the state-of-the-art in computational models that, for example, recognize a face, track user emotions, or monitor physical activities are increasingly based on deep learning principles and algorithms. Unfortunately, deep models typically exert severe demands on local device resources and this conventionally limits their adoption within mobile and embedded platforms. As a result, in far too many cases existing systems process sensor data with machine learning methods that have been superseded by deep learning years ago.","PeriodicalId":419019,"journal":{"name":"Proceedings of the 2nd International Workshop on Embedded and Mobile Deep Learning","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121057996","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
On-the-fly deterministic binary filters for memory efficient keyword spotting applications on embedded devices 嵌入式设备上的高效内存关键字定位应用的动态确定性二进制滤波器
J. Fernández-Marqués, V. W. Tseng, S. Bhattacharya, N. Lane
Lightweight keyword spotting (KWS) applications are often used to trigger the execution of more complex speech recognition algorithms that are computationally demanding and therefore cannot be constantly running on the device. Often KWS applications are executed in small microcontrollers with very constrained memory (e.g. 128kB) and compute capabilities (e.g. CPU at 80MHz) limiting the complexity of deployable KWS systems. We present a compact binary architecture with 60% fewer parameters and 50% fewer operations (OP) during inference compared to the current state of the art for KWS applications at the cost of 3.4% accuracy drop. It makes use of binary orthogonal codes to analyse speech features from a voice command resulting in a model with minimal memory footprint and computationally cheap, making possible its deployment in very resource-constrained microcontrollers with less than 30kB of on-chip memory. Our technique offers a different perspective to how filters in neural networks could be constructed at inference time instead of directly loading them from disk.
轻量级关键字识别(KWS)应用程序通常用于触发更复杂的语音识别算法的执行,这些算法的计算要求很高,因此不能在设备上持续运行。通常,KWS应用程序在具有非常有限的内存(例如128kB)和计算能力(例如80MHz的CPU)的小型微控制器中执行,限制了可部署KWS系统的复杂性。我们提出了一种紧凑的二进制架构,与目前的KWS应用程序相比,在推理过程中参数减少了60%,操作(OP)减少了50%,但精度下降了3.4%。它利用二进制正交代码来分析语音命令的语音特征,从而产生一个内存占用最小且计算成本低廉的模型,使其能够部署在资源非常有限的微控制器中,其片上内存少于30kB。我们的技术为神经网络中的过滤器如何在推理时构建提供了不同的视角,而不是直接从磁盘加载它们。
{"title":"On-the-fly deterministic binary filters for memory efficient keyword spotting applications on embedded devices","authors":"J. Fernández-Marqués, V. W. Tseng, S. Bhattacharya, N. Lane","doi":"10.1145/3212725.3212731","DOIUrl":"https://doi.org/10.1145/3212725.3212731","url":null,"abstract":"Lightweight keyword spotting (KWS) applications are often used to trigger the execution of more complex speech recognition algorithms that are computationally demanding and therefore cannot be constantly running on the device. Often KWS applications are executed in small microcontrollers with very constrained memory (e.g. 128kB) and compute capabilities (e.g. CPU at 80MHz) limiting the complexity of deployable KWS systems. We present a compact binary architecture with 60% fewer parameters and 50% fewer operations (OP) during inference compared to the current state of the art for KWS applications at the cost of 3.4% accuracy drop. It makes use of binary orthogonal codes to analyse speech features from a voice command resulting in a model with minimal memory footprint and computationally cheap, making possible its deployment in very resource-constrained microcontrollers with less than 30kB of on-chip memory. Our technique offers a different perspective to how filters in neural networks could be constructed at inference time instead of directly loading them from disk.","PeriodicalId":419019,"journal":{"name":"Proceedings of the 2nd International Workshop on Embedded and Mobile Deep Learning","volume":"69 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134348061","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Neural Network Syntax Analyzer for Embedded Standardized Deep Learning 用于嵌入式标准化深度学习的神经网络语法分析器
Myungjae Shin, Joongheon Kim, Aziz Mohaisen, Jaebok Park, KyungHee Lee
Deep learning frameworks based on the neural network model have attracted a lot of attention recently for their potential in various applications. Accordingly, recent developments in the fields of deep learning configuration platforms have led to renewed interests in neural network unified format (NNUF) for standardized deep learning computation. The attempt of making NNUF becomes quite challenging because primarily used platforms change over time and the structures of deep learning computation models are continuously evolving. This paper presents the design and implementation of a parser of NNUF for standardized deep learning computation. We call the platform implemented with the neural network exchange framework (NNEF) standard as the NNUF. This framework provides platform-independent processes for configuring and training deep learning neural networks, where the independence is offered by the NNUF model. This model allows us to configure all components of neural network graphs. Our framework also allows the resulting graph to be easily shared with other platform-dependent descriptions which configure various neural network architectures in their own ways. This paper presents the details of the parser design, JavaCC-based implementation, and initial results.
基于神经网络模型的深度学习框架因其在各种应用中的潜力而引起了人们的广泛关注。因此,深度学习配置平台领域的最新发展导致了对标准化深度学习计算的神经网络统一格式(NNUF)的新兴趣。制作NNUF的尝试变得相当具有挑战性,因为主要使用的平台会随着时间的推移而变化,深度学习计算模型的结构也在不断发展。本文提出了一种用于标准化深度学习计算的NNUF解析器的设计与实现。我们将采用神经网络交换框架(NNEF)标准实现的平台称为NNUF。该框架为配置和训练深度学习神经网络提供了与平台无关的过程,其中的独立性由NNUF模型提供。该模型允许我们配置神经网络图的所有组件。我们的框架还允许结果图很容易地与其他平台相关的描述共享,这些描述以自己的方式配置各种神经网络架构。本文详细介绍了解析器的设计、基于javac的实现和初步结果。
{"title":"Neural Network Syntax Analyzer for Embedded Standardized Deep Learning","authors":"Myungjae Shin, Joongheon Kim, Aziz Mohaisen, Jaebok Park, KyungHee Lee","doi":"10.1145/3212725.3212727","DOIUrl":"https://doi.org/10.1145/3212725.3212727","url":null,"abstract":"Deep learning frameworks based on the neural network model have attracted a lot of attention recently for their potential in various applications. Accordingly, recent developments in the fields of deep learning configuration platforms have led to renewed interests in neural network unified format (NNUF) for standardized deep learning computation. The attempt of making NNUF becomes quite challenging because primarily used platforms change over time and the structures of deep learning computation models are continuously evolving. This paper presents the design and implementation of a parser of NNUF for standardized deep learning computation. We call the platform implemented with the neural network exchange framework (NNEF) standard as the NNUF. This framework provides platform-independent processes for configuring and training deep learning neural networks, where the independence is offered by the NNUF model. This model allows us to configure all components of neural network graphs. Our framework also allows the resulting graph to be easily shared with other platform-dependent descriptions which configure various neural network architectures in their own ways. This paper presents the details of the parser design, JavaCC-based implementation, and initial results.","PeriodicalId":419019,"journal":{"name":"Proceedings of the 2nd International Workshop on Embedded and Mobile Deep Learning","volume":"79 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133781701","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Efficiently Combining SVD, Pruning, Clustering and Retraining for Enhanced Neural Network Compression 有效结合SVD、剪枝、聚类和再训练的增强神经网络压缩
Koen Goetschalckx, Bert Moons, P. Wambacq, M. Verhelst
{"title":"Efficiently Combining SVD, Pruning, Clustering and Retraining for Enhanced Neural Network Compression","authors":"Koen Goetschalckx, Bert Moons, P. Wambacq, M. Verhelst","doi":"10.1145/3212725.3212733","DOIUrl":"https://doi.org/10.1145/3212725.3212733","url":null,"abstract":"","PeriodicalId":419019,"journal":{"name":"Proceedings of the 2nd International Workshop on Embedded and Mobile Deep Learning","volume":"116 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115507203","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 17
D-Pruner: Filter-Based Pruning Method for Deep Convolutional Neural Network D-Pruner:基于滤波器的深度卷积神经网络剪枝方法
Huynh Nguyen Loc, Youngki Lee, R. Balan
The emergence of augmented reality devices such as Google Glass and Microsoft Hololens has opened up a new class of vision sensing applications. Those applications often require the ability to continuously capture and analyze contextual information from video streams. They often adopt various deep learning algorithms such as convolutional neural networks (CNN) to achieve high recognition accuracy while facing severe challenges to run computationally intensive deep learning algorithms on resource-constrained mobile devices. In this paper, we propose and explore a new class of compression technique called D-Pruner to efficiently prune redundant parameters within a CNN model to run the model efficiently on mobile devices. D-Pruner removes redundancy by embedding a small additional network. This network evaluates the importance of filters and removes them during the fine-tuning phase to efficiently reduce the size of the model while maintaining the accuracy of the original model. We evaluated D-Pruner on various datasets such as CIFAR-10 and CIFAR-100 and showed that D-Pruner could reduce a significant amount of parameters up to 4.4 times on many existing models while maintaining accuracy drop less than 1%.
谷歌Glass和微软Hololens等增强现实设备的出现开辟了一类新的视觉传感应用。这些应用程序通常需要从视频流中连续捕获和分析上下文信息的能力。他们通常采用卷积神经网络(CNN)等各种深度学习算法来实现较高的识别精度,同时面临着在资源受限的移动设备上运行计算密集型深度学习算法的严峻挑战。在本文中,我们提出并探索了一种称为D-Pruner的新型压缩技术,以有效地修剪CNN模型中的冗余参数,以便在移动设备上有效地运行模型。D-Pruner通过嵌入一个小的附加网络来消除冗余。该网络评估过滤器的重要性,并在微调阶段删除它们,以有效地减少模型的大小,同时保持原始模型的准确性。我们在各种数据集(如CIFAR-10和CIFAR-100)上对D-Pruner进行了评估,结果表明D-Pruner可以在许多现有模型上减少大量参数,最多减少4.4倍,同时保持精度下降不到1%。
{"title":"D-Pruner: Filter-Based Pruning Method for Deep Convolutional Neural Network","authors":"Huynh Nguyen Loc, Youngki Lee, R. Balan","doi":"10.1145/3212725.3212730","DOIUrl":"https://doi.org/10.1145/3212725.3212730","url":null,"abstract":"The emergence of augmented reality devices such as Google Glass and Microsoft Hololens has opened up a new class of vision sensing applications. Those applications often require the ability to continuously capture and analyze contextual information from video streams. They often adopt various deep learning algorithms such as convolutional neural networks (CNN) to achieve high recognition accuracy while facing severe challenges to run computationally intensive deep learning algorithms on resource-constrained mobile devices. In this paper, we propose and explore a new class of compression technique called D-Pruner to efficiently prune redundant parameters within a CNN model to run the model efficiently on mobile devices. D-Pruner removes redundancy by embedding a small additional network. This network evaluates the importance of filters and removes them during the fine-tuning phase to efficiently reduce the size of the model while maintaining the accuracy of the original model. We evaluated D-Pruner on various datasets such as CIFAR-10 and CIFAR-100 and showed that D-Pruner could reduce a significant amount of parameters up to 4.4 times on many existing models while maintaining accuracy drop less than 1%.","PeriodicalId":419019,"journal":{"name":"Proceedings of the 2nd International Workshop on Embedded and Mobile Deep Learning","volume":"26 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116555145","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
QualityDeepSense
Shuochao Yao, Yiran Zhao, Shaohan Hu, T. Abdelzaher
Deep neural networks are becoming increasingly popular in mobile sensing and computing applications. Their capability of fusing multiple sensor inputs and extracting temporal relationships can enhance intelligence in a wide range of applications. One key problem however is the noisy on-device sensors, whose characters are heterogeneous and varying over time. The existing mobile deep learning frameworks usually treat every sensor input equally over time, lacking the ability of identifying and exploiting the heterogeneity of sensor noise. In this work, we propose QualityDeepSense, a deep learning framework that can automatically balance the contribution of sensor inputs over time by their sensing qualities. We propose a sensor-temporal attention mechanism to learn the dependencies among sensor inputs over time. These correlations are used to infer the qualities and reassign the contribution of sensor inputs. QualityDeepSense can thus focus on more informative sensor inputs for prediction. We demonstrate the effectiveness of QualityDeepSense using the noise-augmented heterogeneous human activity recognition task. QualityDeepSense outperforms the state-of-the-art methods by a clear margin. In addition, we show QualityDeepSense only impose limited resource-consumption burden on embedded devices.
{"title":"QualityDeepSense","authors":"Shuochao Yao, Yiran Zhao, Shaohan Hu, T. Abdelzaher","doi":"10.1145/3212725.3212729","DOIUrl":"https://doi.org/10.1145/3212725.3212729","url":null,"abstract":"Deep neural networks are becoming increasingly popular in mobile sensing and computing applications. Their capability of fusing multiple sensor inputs and extracting temporal relationships can enhance intelligence in a wide range of applications. One key problem however is the noisy on-device sensors, whose characters are heterogeneous and varying over time. The existing mobile deep learning frameworks usually treat every sensor input equally over time, lacking the ability of identifying and exploiting the heterogeneity of sensor noise. In this work, we propose QualityDeepSense, a deep learning framework that can automatically balance the contribution of sensor inputs over time by their sensing qualities. We propose a sensor-temporal attention mechanism to learn the dependencies among sensor inputs over time. These correlations are used to infer the qualities and reassign the contribution of sensor inputs. QualityDeepSense can thus focus on more informative sensor inputs for prediction. We demonstrate the effectiveness of QualityDeepSense using the noise-augmented heterogeneous human activity recognition task. QualityDeepSense outperforms the state-of-the-art methods by a clear margin. In addition, we show QualityDeepSense only impose limited resource-consumption burden on embedded devices.","PeriodicalId":419019,"journal":{"name":"Proceedings of the 2nd International Workshop on Embedded and Mobile Deep Learning","volume":"27 11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125683134","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 23
Mayo: A Framework for Auto-generating Hardware Friendly Deep Neural Networks 梅奥:一个自动生成硬件友好深度神经网络的框架
Yiren Zhao, Xitong Gao, R. Mullins, Chengzhong Xu
Deep Neural Networks (DNNs) have proved to be a convenient and powerful tool for a wide range of problems. However, the extensive computational and memory resource requirements hinder the adoption of DNNs in resource-constrained scenarios. Existing compression methods have been shown to significantly reduce the computation and memory requirements of many popular DNNs. These methods, however, remain elusive to non-experts, as they demand extensive manual tuning of hyperparameters. The effects of combining various compression techniques lack exploration because of the large design space. To alleviate these challenges, this paper proposes an automated framework, Mayo, which is built on top of TensorFlow and can compress DNNs with minimal human intervention. First, we present overriders which are recursively-compositional and can be configured to effectively compress individual components (e.g. weights, biases, layer computations and gradients) in a DNN. Second, we introduce novel heuristics and a global search algorithm to efficiently optimize hyperparameters. We demonstrate that without any manual tuning, Mayo generates a sparse ResNet-18 that is 5.13x smaller than the baseline with no loss in test accuracy. By composing multiple overriders, our tool produces a sparse 6-bit CIFAR-10 classifier with only 0.16% top-1 accuracy loss and a 34x compression rate. Mayo and all compressed models are publicly available. To our knowledge, Mayo is the first framework that supports overlapping multiple compression techniques and automatically optimizes hyperparameters in them.
深度神经网络(dnn)已被证明是一种方便而强大的工具,可用于解决广泛的问题。然而,大量的计算和内存资源需求阻碍了dnn在资源受限情况下的采用。现有的压缩方法已经被证明可以显著减少许多流行的深度神经网络的计算和内存需求。然而,这些方法对于非专家来说仍然难以捉摸,因为它们需要大量的超参数手动调优。由于设计空间大,多种压缩技术组合的效果缺乏探索。为了缓解这些挑战,本文提出了一个自动化框架,Mayo,它建立在TensorFlow之上,可以在最少的人为干预下压缩dnn。首先,我们提出了递归组合的覆盖器,可以配置为有效地压缩DNN中的单个组件(例如权重,偏差,层计算和梯度)。其次,我们引入了新的启发式算法和全局搜索算法来有效地优化超参数。我们证明,在没有任何手动调优的情况下,Mayo生成的稀疏ResNet-18比基线小5.13倍,测试精度没有损失。通过组合多个重写器,我们的工具生成了一个稀疏的6位CIFAR-10分类器,其top-1精度损失仅为0.16%,压缩率为34倍。Mayo和所有压缩模型都是公开的。据我们所知,Mayo是第一个支持重叠多个压缩技术并自动优化其中的超参数的框架。
{"title":"Mayo: A Framework for Auto-generating Hardware Friendly Deep Neural Networks","authors":"Yiren Zhao, Xitong Gao, R. Mullins, Chengzhong Xu","doi":"10.1145/3212725.3212726","DOIUrl":"https://doi.org/10.1145/3212725.3212726","url":null,"abstract":"Deep Neural Networks (DNNs) have proved to be a convenient and powerful tool for a wide range of problems. However, the extensive computational and memory resource requirements hinder the adoption of DNNs in resource-constrained scenarios. Existing compression methods have been shown to significantly reduce the computation and memory requirements of many popular DNNs. These methods, however, remain elusive to non-experts, as they demand extensive manual tuning of hyperparameters. The effects of combining various compression techniques lack exploration because of the large design space. To alleviate these challenges, this paper proposes an automated framework, Mayo, which is built on top of TensorFlow and can compress DNNs with minimal human intervention. First, we present overriders which are recursively-compositional and can be configured to effectively compress individual components (e.g. weights, biases, layer computations and gradients) in a DNN. Second, we introduce novel heuristics and a global search algorithm to efficiently optimize hyperparameters. We demonstrate that without any manual tuning, Mayo generates a sparse ResNet-18 that is 5.13x smaller than the baseline with no loss in test accuracy. By composing multiple overriders, our tool produces a sparse 6-bit CIFAR-10 classifier with only 0.16% top-1 accuracy loss and a 34x compression rate. Mayo and all compressed models are publicly available. To our knowledge, Mayo is the first framework that supports overlapping multiple compression techniques and automatically optimizes hyperparameters in them.","PeriodicalId":419019,"journal":{"name":"Proceedings of the 2nd International Workshop on Embedded and Mobile Deep Learning","volume":"2649 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127483494","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
HARNet
Prahalathan Sundaramoorthy, Gautham Krishna Gudur, Manav Rajiv Moorthy, R. Bhandari, Vineeth Vijayaraghavan
Recent advancements in the domain of pervasive computing have seen the incorporation of sensor-based Deep Learning algorithms in Human Activity Recognition (HAR). Contemporary Deep Learning models are engineered to alleviate the difficulties posed by conventional Machine Learning algorithms which require extensive domain knowledge to obtain heuristic hand-crafted features. Upon training and deployment of these Deep Learning models on ubiquitous mobile/embedded devices, it must be ensured that the model adheres to their computation and memory limitations, in addition to addressing the various mobile- and user-based heterogeneities prevalent in actuality. To handle this, we propose HARNet - a resource-efficient and computationally viable network to enable on-line Incremental Learning and User Adaptability as a mitigation technique for anomalous user behavior in HAR. Heterogeneity Activity Recognition Dataset was used to evaluate HARNet and other proposed variants by utilizing acceleration data acquired from diverse mobile platforms across three different modes from a practical application perspective. We perform Decimation as a Down-sampling technique for generalizing sampling frequencies across mobile devices, and Discrete Wavelet Transform for preserving information across frequency and time. Systematic evaluation of HARNet on User Adaptability yields an increase in accuracy by ~35% by leveraging the model's capability to extract discriminative features across activities in heterogeneous environments.
{"title":"HARNet","authors":"Prahalathan Sundaramoorthy, Gautham Krishna Gudur, Manav Rajiv Moorthy, R. Bhandari, Vineeth Vijayaraghavan","doi":"10.1145/3212725.3212728","DOIUrl":"https://doi.org/10.1145/3212725.3212728","url":null,"abstract":"Recent advancements in the domain of pervasive computing have seen the incorporation of sensor-based Deep Learning algorithms in Human Activity Recognition (HAR). Contemporary Deep Learning models are engineered to alleviate the difficulties posed by conventional Machine Learning algorithms which require extensive domain knowledge to obtain heuristic hand-crafted features. Upon training and deployment of these Deep Learning models on ubiquitous mobile/embedded devices, it must be ensured that the model adheres to their computation and memory limitations, in addition to addressing the various mobile- and user-based heterogeneities prevalent in actuality. To handle this, we propose HARNet - a resource-efficient and computationally viable network to enable on-line Incremental Learning and User Adaptability as a mitigation technique for anomalous user behavior in HAR. Heterogeneity Activity Recognition Dataset was used to evaluate HARNet and other proposed variants by utilizing acceleration data acquired from diverse mobile platforms across three different modes from a practical application perspective. We perform Decimation as a Down-sampling technique for generalizing sampling frequencies across mobile devices, and Discrete Wavelet Transform for preserving information across frequency and time. Systematic evaluation of HARNet on User Adaptability yields an increase in accuracy by ~35% by leveraging the model's capability to extract discriminative features across activities in heterogeneous environments.","PeriodicalId":419019,"journal":{"name":"Proceedings of the 2nd International Workshop on Embedded and Mobile Deep Learning","volume":"19 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131949776","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 14
期刊
Proceedings of the 2nd International Workshop on Embedded and Mobile Deep Learning
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1