首页 > 最新文献

2019 IEEE International Conference on Artificial Intelligence Circuits and Systems (AICAS)最新文献

英文 中文
A Deep Learning Based Wearable Medicines Recognition System for Visually Impaired People 基于深度学习的视障人士可穿戴药物识别系统
W. Chang, Yue-Xun Yu, Jheng-Hao Chen, Zhi-Yao Zhang, S. Ko, Tsung-Han Yang, Chia-Hao Hsu, Liang-Bi Chen, Ming-Che Chen
This paper proposes a deep learning based wearable medicines recognition system for visually impaired people. The proposed system is composed of a pair of wearable smart glasses, a wearable waist-mounted drug pills recognition device, a mobile device application, and a cloud-based management platform. The proposed system uses deep learning technology to identify drug pills to avoid taking wrong drugs. The experimental results show that the accuracy of the proposed system has reached up to 90% that can really be achieved the purpose of correct medication for visually impaired people.
提出了一种基于深度学习的视障人士可穿戴药品识别系统。该系统由一副可穿戴智能眼镜、一款可穿戴式腰间药丸识别设备、一款移动设备应用程序和一个基于云的管理平台组成。该系统使用深度学习技术来识别药物,以避免服用错误的药物。实验结果表明,本文提出的系统准确率达到90%以上,能够真正达到视障人士正确用药的目的。
{"title":"A Deep Learning Based Wearable Medicines Recognition System for Visually Impaired People","authors":"W. Chang, Yue-Xun Yu, Jheng-Hao Chen, Zhi-Yao Zhang, S. Ko, Tsung-Han Yang, Chia-Hao Hsu, Liang-Bi Chen, Ming-Che Chen","doi":"10.1109/AICAS.2019.8771559","DOIUrl":"https://doi.org/10.1109/AICAS.2019.8771559","url":null,"abstract":"This paper proposes a deep learning based wearable medicines recognition system for visually impaired people. The proposed system is composed of a pair of wearable smart glasses, a wearable waist-mounted drug pills recognition device, a mobile device application, and a cloud-based management platform. The proposed system uses deep learning technology to identify drug pills to avoid taking wrong drugs. The experimental results show that the accuracy of the proposed system has reached up to 90% that can really be achieved the purpose of correct medication for visually impaired people.","PeriodicalId":273095,"journal":{"name":"2019 IEEE International Conference on Artificial Intelligence Circuits and Systems (AICAS)","volume":"30 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128758755","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 12
Spatial Data Dependence Graph Simulator for Convolutional Neural Network Accelerators 卷积神经网络加速器的空间数据依赖图模拟器
Jooho Wang, Ji-Won Kim, Sungmin Moon, Sunwoo Kim, Sungkyung Park, C. Park
A spatial data dependence graph (S-DDG) is newly proposed to model an accelerator dataflow. The pre-RTL simulator based on the S-DDG helps to explore the design space in the early design phase. The simulation results show the impact of memory latency and bandwidth on a convolutional neural network (CNN) accelerator.
提出了一种空间数据依赖图(S-DDG)来对加速器数据流进行建模。基于S-DDG的预rtl模拟器有助于在早期设计阶段探索设计空间。仿真结果显示了存储延迟和带宽对卷积神经网络加速器的影响。
{"title":"Spatial Data Dependence Graph Simulator for Convolutional Neural Network Accelerators","authors":"Jooho Wang, Ji-Won Kim, Sungmin Moon, Sunwoo Kim, Sungkyung Park, C. Park","doi":"10.1109/AICAS.2019.8771561","DOIUrl":"https://doi.org/10.1109/AICAS.2019.8771561","url":null,"abstract":"A spatial data dependence graph (S-DDG) is newly proposed to model an accelerator dataflow. The pre-RTL simulator based on the S-DDG helps to explore the design space in the early design phase. The simulation results show the impact of memory latency and bandwidth on a convolutional neural network (CNN) accelerator.","PeriodicalId":273095,"journal":{"name":"2019 IEEE International Conference on Artificial Intelligence Circuits and Systems (AICAS)","volume":"102 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121333508","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Heart Rate Estimation from Ballistocardiogram Using Hilbert Transform and Viterbi Decoding 基于希尔伯特变换和维特比解码的ballo心电图心率估计
Qingsong Xie, Yongfu Li, Guoxing Wang, Y. Lian
This paper presents a robust algorithm to estimate heart rate (HR) from ballistocardiogram (BCG). The BCG signal can be easily acquired from the vibration or force sensor embedded in a chair or a mattress without any electrode attached to body. The algorithm employs the Hilbert Transform to reveal the frequency content of J-peak in BCG signal. The Viterbi decoding (VD) is used to estimate HR by finding the most likely path through time-frequency state-space plane. The performance of the proposed algorithm is evaluated by BCG recordings from 10 subjects. Mean absolute error (MAE) of 1.35 beats per minute (BPM) and standard deviation of absolute error (STD) of 1.99 BPM are obtained. Pearson correlation coefficient between estimated HR and true HR of 0.94 is also achieved.
提出了一种基于弹道心动图(BCG)估计心率(HR)的鲁棒算法。BCG信号可以很容易地从嵌入椅子或床垫的振动或力传感器中获取,而无需任何电极连接到身体上。该算法利用希尔伯特变换来揭示BCG信号中j峰的频率含量。利用Viterbi译码(VD)方法通过时频状态空间平面寻找最可能的路径来估计HR。通过10个受试者的BCG记录对算法的性能进行了评价。平均绝对误差(MAE)为1.35 BPM,标准绝对误差(STD)为1.99 BPM。估计HR与真实HR之间的Pearson相关系数为0.94。
{"title":"Heart Rate Estimation from Ballistocardiogram Using Hilbert Transform and Viterbi Decoding","authors":"Qingsong Xie, Yongfu Li, Guoxing Wang, Y. Lian","doi":"10.1109/AICAS.2019.8771627","DOIUrl":"https://doi.org/10.1109/AICAS.2019.8771627","url":null,"abstract":"This paper presents a robust algorithm to estimate heart rate (HR) from ballistocardiogram (BCG). The BCG signal can be easily acquired from the vibration or force sensor embedded in a chair or a mattress without any electrode attached to body. The algorithm employs the Hilbert Transform to reveal the frequency content of J-peak in BCG signal. The Viterbi decoding (VD) is used to estimate HR by finding the most likely path through time-frequency state-space plane. The performance of the proposed algorithm is evaluated by BCG recordings from 10 subjects. Mean absolute error (MAE) of 1.35 beats per minute (BPM) and standard deviation of absolute error (STD) of 1.99 BPM are obtained. Pearson correlation coefficient between estimated HR and true HR of 0.94 is also achieved.","PeriodicalId":273095,"journal":{"name":"2019 IEEE International Conference on Artificial Intelligence Circuits and Systems (AICAS)","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114165740","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Edge and Fog Computing Enabled AI for IoT-An Overview 支持边缘和雾计算的物联网AI概述
Z. Zou, Yi Jin, P. Nevalainen, Y. Huan, J. Heikkonen, Tomi Westerlund
In recent years, Artificial Intelligence (AI) has been widely deployed in a variety of business sectors and industries, yielding numbers of revolutionary applications and services that are primarily driven by high-performance computation and storage facilities in the cloud. On the other hand, embedding intelligence into edge devices is highly demanded by emerging applications such as autonomous systems, human-machine interactions, and the Internet of Things (IoT). In these applications, it is advantageous to process data near or at the source of data to improve energy & spectrum efficiency and security, and decrease latency. Although the computation capability of edge devices has increased tremendously during the past decade, it is still challenging to perform sophisticated AI algorithms in these resource-constrained edge devices, which calls for not only low-power chips for energy efficient processing at the edge but also a system-level framework to distribute resources and tasks along the edge-cloud continuum. In this overview, we summarize dedicated edge hardware for machine learning from embedded applications to sub-mW “always-on” IoT nodes. Recent advances of circuits and systems incorporating joint design of architectures and algorithms will be reviewed. Fog computing paradigm that enables processing at the edge while still offering the possibility to interact with the cloud will be covered, with focus on opportunities and challenges of exploiting fog computing in AI as a bridge between the edge device and the cloud.
近年来,人工智能(AI)已广泛应用于各种商业部门和行业,产生了大量革命性的应用和服务,这些应用和服务主要由云中的高性能计算和存储设施驱动。另一方面,自主系统、人机交互和物联网(IoT)等新兴应用对将智能嵌入边缘设备提出了很高的要求。在这些应用中,在数据源附近或数据源处处理数据有利于提高能量和频谱效率和安全性,并减少延迟。尽管边缘设备的计算能力在过去十年中有了极大的提高,但在这些资源受限的边缘设备中执行复杂的人工智能算法仍然具有挑战性,这不仅需要在边缘进行节能处理的低功耗芯片,还需要一个系统级框架来沿着边缘云连续体分配资源和任务。在这篇综述中,我们总结了用于机器学习的专用边缘硬件,从嵌入式应用到低于mw的“永远在线”物联网节点。结合架构和算法联合设计的电路和系统的最新进展将被回顾。雾计算范式将涵盖在边缘进行处理的同时仍提供与云交互的可能性,重点是利用人工智能中的雾计算作为边缘设备和云之间的桥梁的机遇和挑战。
{"title":"Edge and Fog Computing Enabled AI for IoT-An Overview","authors":"Z. Zou, Yi Jin, P. Nevalainen, Y. Huan, J. Heikkonen, Tomi Westerlund","doi":"10.1109/AICAS.2019.8771621","DOIUrl":"https://doi.org/10.1109/AICAS.2019.8771621","url":null,"abstract":"In recent years, Artificial Intelligence (AI) has been widely deployed in a variety of business sectors and industries, yielding numbers of revolutionary applications and services that are primarily driven by high-performance computation and storage facilities in the cloud. On the other hand, embedding intelligence into edge devices is highly demanded by emerging applications such as autonomous systems, human-machine interactions, and the Internet of Things (IoT). In these applications, it is advantageous to process data near or at the source of data to improve energy & spectrum efficiency and security, and decrease latency. Although the computation capability of edge devices has increased tremendously during the past decade, it is still challenging to perform sophisticated AI algorithms in these resource-constrained edge devices, which calls for not only low-power chips for energy efficient processing at the edge but also a system-level framework to distribute resources and tasks along the edge-cloud continuum. In this overview, we summarize dedicated edge hardware for machine learning from embedded applications to sub-mW “always-on” IoT nodes. Recent advances of circuits and systems incorporating joint design of architectures and algorithms will be reviewed. Fog computing paradigm that enables processing at the edge while still offering the possibility to interact with the cloud will be covered, with focus on opportunities and challenges of exploiting fog computing in AI as a bridge between the edge device and the cloud.","PeriodicalId":273095,"journal":{"name":"2019 IEEE International Conference on Artificial Intelligence Circuits and Systems (AICAS)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134532338","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 54
Function-Safe Vehicular AI Processor with Nano Core-In-Memory Architecture 功能安全的车载AI处理器与纳米核心内存架构
Youngsu Kwon, Jeongmin Yang, Yong Cheol Peter Cho, Kyoung-Seon Shin, Jaehoon Chung, Jinho Han, C. Lyuh, Hyun-Mi Kim, Chan Kim, Minseok Choi
State-of-the-art neural network accelerators consist of arithmetic engines organized in a mesh structure datapath surrounded by memory blocks that provide neural data to the datapath. While server-based accelerators coupled with server-class processors are accommodated with large silicon area and consume large amounts of power, electronic control units in autonomous driving vehicles require power-optimized, ‘AI processors’ with a small footprint. An AI processor for mobile applications that integrates general-purpose processor cores with mesh-structured neural network accelerators and high speed memory while achieving high-performance with low-power and compact area constraints necessitates designing a novel AI processor architecture. We present the design of an AI processor for electronic systems in autonomous driving vehicles targeting not only CNN-based object recognition but also MLP-based in-vehicle voice recognition. The AI processor integrates Super-Thread-Cores (STC) for neural network acceleration with function-safe general purpose cores that satisfy vehicular electronics safety requirements. The STC is composed of 16384 programmable nano-cores organized in a mesh-grid structured datapath network. Designed based on thorough analysis of neural network computations, the nano-core-in-memory architecture enhances computation intensity of STC with efficient feeding of multi-dimensional activation and kernel data into the nano-cores. The quad function-safe general purpose cores ensure functional safety of Super-Thread-Core to comply with road vehicle safety standard ISO 26262. The AI processor exhibits 32 Tera FLOPS, enabling hyper real-time execution of CNN, RNN, and FCN.
最先进的神经网络加速器由在网格结构数据路径中组织的算术引擎组成,数据路径被内存块包围,内存块向数据路径提供神经数据。与服务器级处理器相结合的基于服务器的加速器占用了大量的硅面积,消耗了大量的功率,而自动驾驶车辆中的电子控制单元则需要功耗优化、占地面积小的“人工智能处理器”。移动应用的AI处理器将通用处理器核心与网格结构神经网络加速器和高速存储器集成在一起,同时在低功耗和紧凑的面积限制下实现高性能,因此需要设计一种新颖的AI处理器架构。我们设计了一种用于自动驾驶汽车电子系统的人工智能处理器,不仅针对基于cnn的物体识别,还针对基于mlp的车载语音识别。AI处理器集成了用于神经网络加速的超级线程内核(STC)和满足车辆电子安全要求的功能安全通用内核。STC由16384个可编程纳米核组成,组织在网格结构数据通路网络中。基于对神经网络计算的深入分析,纳米核内存架构通过将多维激活和核数据高效地馈送到纳米核中,提高了STC的计算强度。四功能安全通用芯保证了Super-Thread-Core的功能安全,符合道路车辆安全标准ISO 26262。AI处理器具有32 Tera FLOPS,可实现CNN、RNN和FCN的超实时执行。
{"title":"Function-Safe Vehicular AI Processor with Nano Core-In-Memory Architecture","authors":"Youngsu Kwon, Jeongmin Yang, Yong Cheol Peter Cho, Kyoung-Seon Shin, Jaehoon Chung, Jinho Han, C. Lyuh, Hyun-Mi Kim, Chan Kim, Minseok Choi","doi":"10.1109/AICAS.2019.8771603","DOIUrl":"https://doi.org/10.1109/AICAS.2019.8771603","url":null,"abstract":"State-of-the-art neural network accelerators consist of arithmetic engines organized in a mesh structure datapath surrounded by memory blocks that provide neural data to the datapath. While server-based accelerators coupled with server-class processors are accommodated with large silicon area and consume large amounts of power, electronic control units in autonomous driving vehicles require power-optimized, ‘AI processors’ with a small footprint. An AI processor for mobile applications that integrates general-purpose processor cores with mesh-structured neural network accelerators and high speed memory while achieving high-performance with low-power and compact area constraints necessitates designing a novel AI processor architecture. We present the design of an AI processor for electronic systems in autonomous driving vehicles targeting not only CNN-based object recognition but also MLP-based in-vehicle voice recognition. The AI processor integrates Super-Thread-Cores (STC) for neural network acceleration with function-safe general purpose cores that satisfy vehicular electronics safety requirements. The STC is composed of 16384 programmable nano-cores organized in a mesh-grid structured datapath network. Designed based on thorough analysis of neural network computations, the nano-core-in-memory architecture enhances computation intensity of STC with efficient feeding of multi-dimensional activation and kernel data into the nano-cores. The quad function-safe general purpose cores ensure functional safety of Super-Thread-Core to comply with road vehicle safety standard ISO 26262. The AI processor exhibits 32 Tera FLOPS, enabling hyper real-time execution of CNN, RNN, and FCN.","PeriodicalId":273095,"journal":{"name":"2019 IEEE International Conference on Artificial Intelligence Circuits and Systems (AICAS)","volume":"11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133077838","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Performance Trade-offs in Weight Quantization for Memory-Efficient Inference 基于内存效率推断的权重量化的性能权衡
Pablo M. Tostado, B. Pedroni, G. Cauwenberghs
Over the past decade, Deep Neural Networks (DNNs) trained using Deep Learning (DL) frameworks have become the workhorse to solve a wide variety of computational tasks in big data environments. To date, DL DNNs have relied on large amounts of computational power to reach peak performance, typically relying on the high computational bandwidth of GPUs, while straining available memory bandwidth and capacity. With ever increasing data complexity and more stringent energy constraints in Internet-of-Things (IoT) application environments, there has been a growing interest in the development of more efficient DNN inference methods that economize on random-access memory usage in weight access. Herein, we present a systematic analysis of the performance trade-offs of quantized weight representations at variable bit length for memory-efficient inference in pre-trained DNN models. In this work, we vary the mantissa and exponent bit lengths in the representation of the network parameters and examine the effect of DropOut regularization during pre-training and the impact of two different weight truncation mechanisms: stochastic and deterministic rounding. We show drastic reduction in the memory need, down to 4 bits per weight, while maintaining near-optimal test performance of low-complexity DNNs pre-trained on the MNIST and CIFAR-10 datasets. These results offer a simple methodology to achieve high memory and computation efficiency of inference in DNN dedicated low-power hardware for IoT, directly from pre-trained, high-resolution DNNs using standard DL algorithms.
在过去的十年中,使用深度学习(DL)框架训练的深度神经网络(dnn)已经成为解决大数据环境中各种计算任务的主力。迄今为止,深度深度神经网络依赖于大量的计算能力来达到峰值性能,通常依赖于gpu的高计算带宽,同时使可用的内存带宽和容量紧张。随着物联网(IoT)应用环境中不断增加的数据复杂性和更严格的能量限制,人们对开发更有效的DNN推理方法越来越感兴趣,这些方法可以在权重访问中节省随机访问内存的使用。在此,我们系统地分析了在预训练的DNN模型中,可变位长的量化权重表示用于内存高效推理的性能权衡。在这项工作中,我们改变了网络参数表示中的尾数和指数位长度,并检查了预训练期间DropOut正则化的影响以及两种不同权重截断机制(随机和确定性舍入)的影响。在MNIST和CIFAR-10数据集上预训练的低复杂度dnn保持近乎最佳的测试性能的同时,我们显示了内存需求的大幅减少,每个权重降至4位。这些结果提供了一种简单的方法,可以直接从使用标准深度学习算法的预训练的高分辨率DNN中实现用于物联网的DNN专用低功耗硬件的高内存和计算效率。
{"title":"Performance Trade-offs in Weight Quantization for Memory-Efficient Inference","authors":"Pablo M. Tostado, B. Pedroni, G. Cauwenberghs","doi":"10.1109/AICAS.2019.8771473","DOIUrl":"https://doi.org/10.1109/AICAS.2019.8771473","url":null,"abstract":"Over the past decade, Deep Neural Networks (DNNs) trained using Deep Learning (DL) frameworks have become the workhorse to solve a wide variety of computational tasks in big data environments. To date, DL DNNs have relied on large amounts of computational power to reach peak performance, typically relying on the high computational bandwidth of GPUs, while straining available memory bandwidth and capacity. With ever increasing data complexity and more stringent energy constraints in Internet-of-Things (IoT) application environments, there has been a growing interest in the development of more efficient DNN inference methods that economize on random-access memory usage in weight access. Herein, we present a systematic analysis of the performance trade-offs of quantized weight representations at variable bit length for memory-efficient inference in pre-trained DNN models. In this work, we vary the mantissa and exponent bit lengths in the representation of the network parameters and examine the effect of DropOut regularization during pre-training and the impact of two different weight truncation mechanisms: stochastic and deterministic rounding. We show drastic reduction in the memory need, down to 4 bits per weight, while maintaining near-optimal test performance of low-complexity DNNs pre-trained on the MNIST and CIFAR-10 datasets. These results offer a simple methodology to achieve high memory and computation efficiency of inference in DNN dedicated low-power hardware for IoT, directly from pre-trained, high-resolution DNNs using standard DL algorithms.","PeriodicalId":273095,"journal":{"name":"2019 IEEE International Conference on Artificial Intelligence Circuits and Systems (AICAS)","volume":"108 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115073521","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Classification of Cardiac Arrhythmias Based on Artificial Neural Networks and Continuous-in-Time Discrete-in-Amplitude Signal Flow 基于人工神经网络和连续时离散幅值信号流的心律失常分类
Yang Zhao, Simon Lin, Zhongxia Shang, Y. Lian
Conventional Artificial Neural Networks (ANNs) for classification of cardiac arrhythmias are based on Nyquist sampled electrocardiogram (ECG) signals. The uniform sampling scheme introduces large redundancy in the ANN, which results high power and large silicon area. To address these issues, we propose to use continuous-in-time discrete-in-amplitude (CTDA) sampling scheme as the input of the network. The CTDA sampling scheme significantly reduces the sample points on the baseline part while provides more detail on useful features in the ECG signal. It is shown that the CTDA sampling scheme achieves significant savings on arithmetic operations in the ANN while maintains the similar performance as Nyquist sampling in the classification. The proposed method is evaluated by MIT-BIH arrhythmia database following AAMI recommended practice.
传统的心律失常分类人工神经网络(ann)是基于奈奎斯特采样的心电图信号。均匀采样方案给人工神经网络引入了大量冗余,从而实现了高功率和大硅面积。为了解决这些问题,我们建议使用连续时离散幅值(CTDA)采样方案作为网络的输入。CTDA采样方案显著减少了基线部分的采样点,同时提供了心电信号中有用特征的更多细节。研究表明,CTDA采样方案在保持与Nyquist采样相似的分类性能的同时,显著节省了人工神经网络的算术运算。按照AAMI推荐的做法,采用MIT-BIH心律失常数据库对该方法进行评估。
{"title":"Classification of Cardiac Arrhythmias Based on Artificial Neural Networks and Continuous-in-Time Discrete-in-Amplitude Signal Flow","authors":"Yang Zhao, Simon Lin, Zhongxia Shang, Y. Lian","doi":"10.1109/AICAS.2019.8771620","DOIUrl":"https://doi.org/10.1109/AICAS.2019.8771620","url":null,"abstract":"Conventional Artificial Neural Networks (ANNs) for classification of cardiac arrhythmias are based on Nyquist sampled electrocardiogram (ECG) signals. The uniform sampling scheme introduces large redundancy in the ANN, which results high power and large silicon area. To address these issues, we propose to use continuous-in-time discrete-in-amplitude (CTDA) sampling scheme as the input of the network. The CTDA sampling scheme significantly reduces the sample points on the baseline part while provides more detail on useful features in the ECG signal. It is shown that the CTDA sampling scheme achieves significant savings on arithmetic operations in the ANN while maintains the similar performance as Nyquist sampling in the classification. The proposed method is evaluated by MIT-BIH arrhythmia database following AAMI recommended practice.","PeriodicalId":273095,"journal":{"name":"2019 IEEE International Conference on Artificial Intelligence Circuits and Systems (AICAS)","volume":"123 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122891669","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Using A Cropping Technique or Not: Impacts on SVM-based AMD Detection on OCT Images 是否使用裁剪技术:对基于svm的OCT图像AMD检测的影响
C. Ko, Po-Han Chen, Wei-Ming Liao, Cheng-Kai Lu, Cheng-Hung Lin, Jing-Wen Liang
This paper compares the system performance of distinct flows with automatic image cropping to without automatic image cropping for age-related macular degeneration (AMD) detection on optical coherence tomography (OCT) images. Using the image cropping, the computational time of noise removal and feature extraction can be significantly reduced by a small loss of detection accuracy. The simulation results show that using the image cropping at the first stage achieves 93.4% accuracy. Compared to the flow without image cropping, using the image cropping loses only 0.5% accuracy but saves about 12 hours computational time and about a half of memory storages.
本文比较了在光学相干断层扫描(OCT)图像上进行年龄相关性黄斑变性(AMD)检测时,自动图像裁剪与不自动图像裁剪的不同流的系统性能。利用图像裁剪,在检测精度损失很小的情况下,可以显著减少去噪和特征提取的计算时间。仿真结果表明,在第一阶段使用图像裁剪,准确率达到93.4%。与不进行图像裁剪的流程相比,使用图像裁剪仅损失0.5%的精度,但节省了大约12小时的计算时间和大约一半的内存存储。
{"title":"Using A Cropping Technique or Not: Impacts on SVM-based AMD Detection on OCT Images","authors":"C. Ko, Po-Han Chen, Wei-Ming Liao, Cheng-Kai Lu, Cheng-Hung Lin, Jing-Wen Liang","doi":"10.1109/AICAS.2019.8771609","DOIUrl":"https://doi.org/10.1109/AICAS.2019.8771609","url":null,"abstract":"This paper compares the system performance of distinct flows with automatic image cropping to without automatic image cropping for age-related macular degeneration (AMD) detection on optical coherence tomography (OCT) images. Using the image cropping, the computational time of noise removal and feature extraction can be significantly reduced by a small loss of detection accuracy. The simulation results show that using the image cropping at the first stage achieves 93.4% accuracy. Compared to the flow without image cropping, using the image cropping loses only 0.5% accuracy but saves about 12 hours computational time and about a half of memory storages.","PeriodicalId":273095,"journal":{"name":"2019 IEEE International Conference on Artificial Intelligence Circuits and Systems (AICAS)","volume":"59 21","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"120972119","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Exploration of Automatic Mixed-Precision Search for Deep Neural Networks 深度神经网络自动混合精度搜索的探索
Xuyang Guo, Yuanjun Huang, Hsin-Pai Cheng, Bing Li, W. Wen, Siyuan Ma, H. Li, Yiran Chen
Neural networks have shown great performance in cognitive tasks. When deploying network models on mobile devices with limited computation and storage resources, the weight quantization technique has been widely adopted. In practice, 8-bit or 16-bit quantization is mostly likely to be selected in order to maintain the accuracy at the same level as the models in 32-bit floating-point precision. Binary quantization, on the contrary, aims to obtain the highest compression at the cost of much bigger accuracy drop. Applying different precision in different layers/structures can potentially produce the most efficient model. Seeking for the best precision configuration, however, is difficult. In this work, we proposed an automatic search algorithm to address the challenge. By relaxing the search space of quantization bitwidth from discrete to continuous domain, our algorithm can generate a mixed-precision quantization scheme which achieves the compression rate close to the one from the binary-weighted model while maintaining the testing accuracy similar to the original full-precision model.
神经网络在认知任务中表现优异。当在计算和存储资源有限的移动设备上部署网络模型时,权重量化技术被广泛采用。在实践中,为了保持与32位浮点精度模型相同的精度水平,最可能选择8位或16位量化。相反,二值量化的目标是以更大的精度下降为代价获得最高的压缩。在不同的层/结构中应用不同的精度可以产生最有效的模型。然而,寻找最佳精度配置是困难的。在这项工作中,我们提出了一种自动搜索算法来解决这一挑战。通过将量化位宽的搜索空间从离散域放宽到连续域,我们的算法可以生成一种混合精度量化方案,该方案的压缩率接近于二值加权模型的压缩率,同时保持与原全精度模型相似的测试精度。
{"title":"Exploration of Automatic Mixed-Precision Search for Deep Neural Networks","authors":"Xuyang Guo, Yuanjun Huang, Hsin-Pai Cheng, Bing Li, W. Wen, Siyuan Ma, H. Li, Yiran Chen","doi":"10.1109/AICAS.2019.8771498","DOIUrl":"https://doi.org/10.1109/AICAS.2019.8771498","url":null,"abstract":"Neural networks have shown great performance in cognitive tasks. When deploying network models on mobile devices with limited computation and storage resources, the weight quantization technique has been widely adopted. In practice, 8-bit or 16-bit quantization is mostly likely to be selected in order to maintain the accuracy at the same level as the models in 32-bit floating-point precision. Binary quantization, on the contrary, aims to obtain the highest compression at the cost of much bigger accuracy drop. Applying different precision in different layers/structures can potentially produce the most efficient model. Seeking for the best precision configuration, however, is difficult. In this work, we proposed an automatic search algorithm to address the challenge. By relaxing the search space of quantization bitwidth from discrete to continuous domain, our algorithm can generate a mixed-precision quantization scheme which achieves the compression rate close to the one from the binary-weighted model while maintaining the testing accuracy similar to the original full-precision model.","PeriodicalId":273095,"journal":{"name":"2019 IEEE International Conference on Artificial Intelligence Circuits and Systems (AICAS)","volume":"23 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128320736","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Binarized Neural Network with Stochastic Memristors 随机忆阻器二值化神经网络
O. Krestinskaya, Otaniyoz Otaniyozov, A. P. James
This paper proposes the analog hardware implementation of Binarized Neural Network (BNN). Most of the existing hardware implementations of neural networks do not consider the memristor variability issue and its effect on the overall system performance. In this work, we investigate the variability in memristive devices in crossbar dot product computation and leakage currents in the proposed BNN, and show how it effects the overall system performance.
提出了二值化神经网络(BNN)的模拟硬件实现。大多数现有的神经网络硬件实现都没有考虑忆阻变异性问题及其对系统整体性能的影响。在这项工作中,我们研究了记忆器件在交叉棒点积计算和泄漏电流中的可变性,并展示了它如何影响整体系统性能。
{"title":"Binarized Neural Network with Stochastic Memristors","authors":"O. Krestinskaya, Otaniyoz Otaniyozov, A. P. James","doi":"10.1109/AICAS.2019.8771565","DOIUrl":"https://doi.org/10.1109/AICAS.2019.8771565","url":null,"abstract":"This paper proposes the analog hardware implementation of Binarized Neural Network (BNN). Most of the existing hardware implementations of neural networks do not consider the memristor variability issue and its effect on the overall system performance. In this work, we investigate the variability in memristive devices in crossbar dot product computation and leakage currents in the proposed BNN, and show how it effects the overall system performance.","PeriodicalId":273095,"journal":{"name":"2019 IEEE International Conference on Artificial Intelligence Circuits and Systems (AICAS)","volume":"532 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132314997","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
期刊
2019 IEEE International Conference on Artificial Intelligence Circuits and Systems (AICAS)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1