首页 > 最新文献

2018 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA ASC)最新文献

英文 中文
A systematic assessment of a cochlear implant processor's ability to encode interaural time differences 对人工耳蜗处理器对耳际时差进行编码能力的系统评估
A. Kan, Z. Peng, K. Moua, R. Litovsky
Bilateral cochlear implantation is becoming the standard of care for patients with sensorineural hearing loss with demonstrated improvements over unilateral use in everyday tasks, such as sound localization ability. However, even with bilateral implantation, performance in these tasks is still poorer than that of normal hearing listeners. The gap in performance has often been attributed to the poor encoding of fine structure interaural time differences (ITDs) by clinical processor. However, in theory, the signal processing employed in clinical processors should still encode envelope ITDs with some degree of fidelity. In this work, we quantitatively measured the ability of Cochlear CP910 processors to encode envelope ITDs, while running the Advanced Combinational Encoder (ACE) strategy. Results suggest that while the processors are able to support relatively good envelope encoding, the peak-picking approach of the ACE strategy degrades the computation of ITDs by encoding spectral information in different frequency regions in the two ears. Our results may explain the poorer sound localization performance observed in cochlear implant users who use the ACE strategy, but cannot account for the poorer sound localization performance observed in cochlear implant users in general.
双侧人工耳蜗植入术正成为感音神经性听力损失患者的标准治疗方法,与单侧使用相比,双侧人工耳蜗植入术在日常工作中(如声音定位能力)有明显改善。然而,即使双侧植入,在这些任务中的表现仍然不如正常听力的听者。这种性能上的差距通常是由于临床处理器对精细结构间时差(ITDs)编码不佳造成的。然而,理论上,临床处理器所采用的信号处理仍应以一定程度的保真度对包络图进行编码。在这项工作中,我们定量测量了Cochlear CP910处理器在运行高级组合编码器(ACE)策略时编码包络过渡段的能力。结果表明,虽然处理器能够支持相对较好的包络编码,但ACE策略的拾峰方法通过编码两耳不同频率区域的频谱信息来降低过渡段的计算。我们的研究结果可以解释在使用ACE策略的人工耳蜗使用者中观察到的较差的声音定位表现,但不能解释在一般人工耳蜗使用者中观察到的较差的声音定位表现。
{"title":"A systematic assessment of a cochlear implant processor's ability to encode interaural time differences","authors":"A. Kan, Z. Peng, K. Moua, R. Litovsky","doi":"10.23919/APSIPA.2018.8659694","DOIUrl":"https://doi.org/10.23919/APSIPA.2018.8659694","url":null,"abstract":"Bilateral cochlear implantation is becoming the standard of care for patients with sensorineural hearing loss with demonstrated improvements over unilateral use in everyday tasks, such as sound localization ability. However, even with bilateral implantation, performance in these tasks is still poorer than that of normal hearing listeners. The gap in performance has often been attributed to the poor encoding of fine structure interaural time differences (ITDs) by clinical processor. However, in theory, the signal processing employed in clinical processors should still encode envelope ITDs with some degree of fidelity. In this work, we quantitatively measured the ability of Cochlear CP910 processors to encode envelope ITDs, while running the Advanced Combinational Encoder (ACE) strategy. Results suggest that while the processors are able to support relatively good envelope encoding, the peak-picking approach of the ACE strategy degrades the computation of ITDs by encoding spectral information in different frequency regions in the two ears. Our results may explain the poorer sound localization performance observed in cochlear implant users who use the ACE strategy, but cannot account for the poorer sound localization performance observed in cochlear implant users in general.","PeriodicalId":287799,"journal":{"name":"2018 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA ASC)","volume":"69 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128163562","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
Performance Profiling of Cloud Radio Access Networks using OpenAirInterface 使用OpenAirInterface的云无线接入网络性能分析
Po-Chiang Lin, Sheng-Lun Huang
NGFI, the Next Generation Fronthaul Interface, is a promising fronthaul interface for the C-RAN (Cloud Radio Access Network). NGFI is used to connect RCC (Radio Cloud Center) and RRS (Radio Remote System) in order to avoid the drawbacks of traditional CPRI (Common Public Radio Interface). In this paper we investigate the NGFI-based C-RAN. We use the OpenAirInterface (OAI) open source 4G/5G mobile communication software and GPP (general purpose processor) based servers and personal computers to build an OAI C-RAN testbed. We also use the source codes of OAI to run the performance profiling on this OAI C-RAN testbed to understand the behavior of this testbed. The purpose of this paper is to build the comprehensive performance profiling methods and results on the OAI C-RAN system, and to use these results to help designing and optimizing the OAI C-RAN system. Based on the results, we could decide which part of the system software to optimize to improve the system speed and the efficiency of memory usage.
NGFI,下一代前传接口,是一种很有前途的用于C-RAN(云无线接入网)的前传接口。NGFI用于连接RCC (Radio Cloud Center)和RRS (Radio Remote System),避免了传统CPRI (Common Public Radio Interface)的弊端。本文研究了基于ngfi的C-RAN。我们使用OpenAirInterface (OAI)开源4G/5G移动通信软件和基于GPP(通用处理器)的服务器和个人计算机构建了OAI C-RAN测试平台。我们还使用OAI的源代码在这个OAI C-RAN测试平台上运行性能分析,以了解这个测试平台的行为。本文的目的是建立OAI C-RAN系统的综合性能分析方法和结果,并利用这些结果帮助OAI C-RAN系统的设计和优化。根据结果,我们可以决定对系统软件的哪个部分进行优化,以提高系统速度和内存使用效率。
{"title":"Performance Profiling of Cloud Radio Access Networks using OpenAirInterface","authors":"Po-Chiang Lin, Sheng-Lun Huang","doi":"10.23919/APSIPA.2018.8659532","DOIUrl":"https://doi.org/10.23919/APSIPA.2018.8659532","url":null,"abstract":"NGFI, the Next Generation Fronthaul Interface, is a promising fronthaul interface for the C-RAN (Cloud Radio Access Network). NGFI is used to connect RCC (Radio Cloud Center) and RRS (Radio Remote System) in order to avoid the drawbacks of traditional CPRI (Common Public Radio Interface). In this paper we investigate the NGFI-based C-RAN. We use the OpenAirInterface (OAI) open source 4G/5G mobile communication software and GPP (general purpose processor) based servers and personal computers to build an OAI C-RAN testbed. We also use the source codes of OAI to run the performance profiling on this OAI C-RAN testbed to understand the behavior of this testbed. The purpose of this paper is to build the comprehensive performance profiling methods and results on the OAI C-RAN system, and to use these results to help designing and optimizing the OAI C-RAN system. Based on the results, we could decide which part of the system software to optimize to improve the system speed and the efficiency of memory usage.","PeriodicalId":287799,"journal":{"name":"2018 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA ASC)","volume":"48 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125683182","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 18
A Reconfigurable Process Engine for Flexible Convolutional Neural Network Acceleration 柔性卷积神经网络加速的可重构过程引擎
Xiaobai Chen, Shanlin Xiao, Zhiyi Yu
Convolutional neural network (CNN) is the most powerful artificial intelligence algorithm widely used in computer vision due to its state-of-the-art performance. There are many accelerators proposed for CNN to handle its huge computation and communication cost. In this paper we proposed a reconfigurable process engine which can support different data flows, bit-widths, and parallelism strategies for CNNs. The process engine was implemented on Xilinx ZC706 FPGA board, with high flexibility to support all popular CNNs, and better energy efficiency compared to other state-of-the-art designs.
卷积神经网络(CNN)由于其最先进的性能而被广泛应用于计算机视觉领域,是最强大的人工智能算法。为了解决CNN庞大的计算和通信成本,有很多加速器被提出。本文提出了一种可重构的过程引擎,该引擎可以支持不同的数据流、位宽度和并行策略。该处理引擎在赛灵思ZC706 FPGA板上实现,具有高灵活性,可支持所有流行的cnn,并且与其他最先进的设计相比具有更高的能效。
{"title":"A Reconfigurable Process Engine for Flexible Convolutional Neural Network Acceleration","authors":"Xiaobai Chen, Shanlin Xiao, Zhiyi Yu","doi":"10.23919/APSIPA.2018.8659629","DOIUrl":"https://doi.org/10.23919/APSIPA.2018.8659629","url":null,"abstract":"Convolutional neural network (CNN) is the most powerful artificial intelligence algorithm widely used in computer vision due to its state-of-the-art performance. There are many accelerators proposed for CNN to handle its huge computation and communication cost. In this paper we proposed a reconfigurable process engine which can support different data flows, bit-widths, and parallelism strategies for CNNs. The process engine was implemented on Xilinx ZC706 FPGA board, with high flexibility to support all popular CNNs, and better energy efficiency compared to other state-of-the-art designs.","PeriodicalId":287799,"journal":{"name":"2018 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA ASC)","volume":"83 7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127979305","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Integral 3D image coding by using multiview video compression technologies 基于多视点视频压缩技术的整体三维图像编码
Kazuhiro Hara, Miwa Katayama, M. Kawakita, T. Fujii, T. Mishina
Effective compression technology is required to reduce the huge amount of information for integral three-dimensional (3D) television. For compressing an integral 3D image, we propose a compression method of converting elemental images to multiview images and of applying multiview video coding to part of the multiview images and their depth maps. In this method, the relationship between the number of the part of the multiview images and the image quality degradation of a reconstructed 3D image was studied by subjective evaluation experiment, and we confirmed the amount of information required for displaying an acceptable reconstructed 3D image. As a result, the reconstructed 3D images with acceptable image quality were obtained with about 2/9 times the amount of information for coding all the multiview images converted from the elemental images.
整体三维电视需要有效的压缩技术来减少巨大的信息量。为了压缩完整的三维图像,我们提出了一种将元素图像转换为多视图图像并对部分多视图图像及其深度图进行多视图视频编码的压缩方法。该方法通过主观评价实验,研究了多视图图像的部分数量与重建三维图像的图像质量退化之间的关系,确定了显示可接受的重建三维图像所需的信息量。结果,得到了图像质量可接受的重建三维图像,其信息量约为由元素图像转换而来的所有多视图图像编码信息量的2/9倍。
{"title":"Integral 3D image coding by using multiview video compression technologies","authors":"Kazuhiro Hara, Miwa Katayama, M. Kawakita, T. Fujii, T. Mishina","doi":"10.23919/APSIPA.2018.8659709","DOIUrl":"https://doi.org/10.23919/APSIPA.2018.8659709","url":null,"abstract":"Effective compression technology is required to reduce the huge amount of information for integral three-dimensional (3D) television. For compressing an integral 3D image, we propose a compression method of converting elemental images to multiview images and of applying multiview video coding to part of the multiview images and their depth maps. In this method, the relationship between the number of the part of the multiview images and the image quality degradation of a reconstructed 3D image was studied by subjective evaluation experiment, and we confirmed the amount of information required for displaying an acceptable reconstructed 3D image. As a result, the reconstructed 3D images with acceptable image quality were obtained with about 2/9 times the amount of information for coding all the multiview images converted from the elemental images.","PeriodicalId":287799,"journal":{"name":"2018 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA ASC)","volume":"71 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121751092","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Improved Vessel Segmentation Using Curvelet Transform and Line Operators 基于曲线变换和直线算子的改进血管分割
Renoh Johnson Chalakkal, W. Abdulla
Vessel segmentation from the fundus retinal images is highly significant in diagnosing many pathologies related to eye and other systemic diseases. Even though there are many methods in the literature focusing on this task, most of these methods are not focusing on the small peripheral vessels segmentation. In this paper, we propose a new approach based on curvelet transform and line operators which can segment the small peripheral vessels with very high accuracy resulting in a higher sensitivity compared to the other state-of-the-art methods. In the proposed approach, the contrast between the retinal vessels and the background pixels is enhanced by applying a series of image processing steps involving color space transformation, adaptive histogram equalization, and anisotropic diffusion filtering. Then by using the modified curvelet transform coefficients, the retinal vessel edge contrast is further enhanced. Finally, the vessels are segmented out by applying the line operator response, followed by suitable thresholding to obtain the segmented vessels. Post processing is carried out to remove the scattered unwanted background pixels. The performance of the method is compared against the other state-of-the-art methods using DRIVE as a testing database. An average sensitivity, specificity, accuracy and positive predictive value of 0.7653, 0.9735, 0.9542 and 0.7438 are respectively achieved.
眼底视网膜图像的血管分割对诊断许多与眼睛和其他全身性疾病有关的病理具有重要意义。尽管文献中有很多方法致力于这项任务,但大多数方法都没有关注周围小血管的分割。在本文中,我们提出了一种基于曲线变换和线算子的新方法,与其他最先进的方法相比,它可以以非常高的精度分割小的外围血管,从而获得更高的灵敏度。该方法通过色彩空间变换、自适应直方图均衡化和各向异性扩散滤波等一系列图像处理步骤,增强了视网膜血管和背景像素之间的对比度。然后利用改进的曲波变换系数进一步增强视网膜血管边缘对比度。最后,应用线性算子响应分割出血管,然后采用合适的阈值分割得到分割后的血管。进行后处理以去除分散的不需要的背景像素。将该方法的性能与使用DRIVE作为测试数据库的其他最先进的方法进行比较。平均灵敏度、特异度、准确度和阳性预测值分别为0.7653、0.9735、0.9542和0.7438。
{"title":"Improved Vessel Segmentation Using Curvelet Transform and Line Operators","authors":"Renoh Johnson Chalakkal, W. Abdulla","doi":"10.23919/APSIPA.2018.8659682","DOIUrl":"https://doi.org/10.23919/APSIPA.2018.8659682","url":null,"abstract":"Vessel segmentation from the fundus retinal images is highly significant in diagnosing many pathologies related to eye and other systemic diseases. Even though there are many methods in the literature focusing on this task, most of these methods are not focusing on the small peripheral vessels segmentation. In this paper, we propose a new approach based on curvelet transform and line operators which can segment the small peripheral vessels with very high accuracy resulting in a higher sensitivity compared to the other state-of-the-art methods. In the proposed approach, the contrast between the retinal vessels and the background pixels is enhanced by applying a series of image processing steps involving color space transformation, adaptive histogram equalization, and anisotropic diffusion filtering. Then by using the modified curvelet transform coefficients, the retinal vessel edge contrast is further enhanced. Finally, the vessels are segmented out by applying the line operator response, followed by suitable thresholding to obtain the segmented vessels. Post processing is carried out to remove the scattered unwanted background pixels. The performance of the method is compared against the other state-of-the-art methods using DRIVE as a testing database. An average sensitivity, specificity, accuracy and positive predictive value of 0.7653, 0.9735, 0.9542 and 0.7438 are respectively achieved.","PeriodicalId":287799,"journal":{"name":"2018 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA ASC)","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131353059","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
Speech Enhancement with Phase Correction based on Modified DNN Architecture 基于改进DNN结构的相位校正语音增强
Rui Cheng, C. Bao, Yang Xiang
Speech enhancement is an important issue in the field of speech signal processing. With the development of deep learning, speech enhancement technology combined with neural network has provided a more diverse solution for this field. In this paper, we present a new approach to enhance the noisy speech, which is recorded by a single channel. We propose a phase correction method, which is based on the joint optimization of clean speech and noise by deep neural network (DNN). In this method, the ideal ratio masking (IRM) is employed to estimate the clean speech and noise, and the phase correction is combined to get the final clean speech. Experiments are conducted by using TIMIT corpus combined with four types of noises at three different signal to noise ratio (SNR) levels. The results show that the proposed method has a significant improvement over the referenced DNN-based enhancement method for both objective evaluation criterion and subjective evaluation criterion.
语音增强是语音信号处理领域的一个重要问题。随着深度学习的发展,语音增强技术与神经网络的结合为该领域提供了更加多样化的解决方案。本文提出了一种新的方法来增强单通道录制的带噪语音。提出了一种基于深度神经网络(DNN)清洁语音和噪声联合优化的相位校正方法。该方法采用理想比例掩蔽(IRM)估计干净语音和噪声,并结合相位校正得到最终的干净语音。利用TIMIT语料库结合四种不同信噪比(SNR)水平的噪声进行了实验。结果表明,该方法在客观评价准则和主观评价准则方面都比参考的基于dnn的增强方法有显著改进。
{"title":"Speech Enhancement with Phase Correction based on Modified DNN Architecture","authors":"Rui Cheng, C. Bao, Yang Xiang","doi":"10.23919/APSIPA.2018.8659625","DOIUrl":"https://doi.org/10.23919/APSIPA.2018.8659625","url":null,"abstract":"Speech enhancement is an important issue in the field of speech signal processing. With the development of deep learning, speech enhancement technology combined with neural network has provided a more diverse solution for this field. In this paper, we present a new approach to enhance the noisy speech, which is recorded by a single channel. We propose a phase correction method, which is based on the joint optimization of clean speech and noise by deep neural network (DNN). In this method, the ideal ratio masking (IRM) is employed to estimate the clean speech and noise, and the phase correction is combined to get the final clean speech. Experiments are conducted by using TIMIT corpus combined with four types of noises at three different signal to noise ratio (SNR) levels. The results show that the proposed method has a significant improvement over the referenced DNN-based enhancement method for both objective evaluation criterion and subjective evaluation criterion.","PeriodicalId":287799,"journal":{"name":"2018 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA ASC)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129965004","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Block-Matching Convolutional Neural Network (BMCNN): Improving CNN-Based Denoising by Block-Matched Inputs 块匹配卷积神经网络(BMCNN):改进基于cnn的块匹配输入去噪
Byeongyong Ahn, Yoonsik Kim, G. Park, N. Cho
There are two main streams in up-to-date image denoising algorithms: non-local self similarity (NSS) prior based methods and convolutional neural network (CNN) based methods. The NSS based methods are favorable on images with regular and repetitive patterns while the CNN based methods perform better on irregular structures. In this paper, we propose a block-matching convolutional neural network (BMCNN) method that combines NSS prior and CNN. Initially, similar local patches in the input image are integrated into a 3D block. In order to prevent the noise from messing up the block matching, we first apply an existing denoising algorithm on the noisy image. The denoised image is employed as a pilot signal for the block matching, and then denoising function for the block is learned by a CNN structure. Experimental results show that the proposed BMCNN algorithm achieves state-of-the-art performance. In detail, BMCNN can restore both repetitive and irregular structures.
目前图像去噪算法有两大主流:基于非局部自相似先验的方法和基于卷积神经网络的方法。基于NSS的方法在具有规则和重复模式的图像上表现较好,而基于CNN的方法在不规则结构上表现较好。本文提出了一种结合NSS先验和CNN的块匹配卷积神经网络(BMCNN)方法。最初,输入图像中相似的局部补丁被整合到一个3D块中。为了防止噪声干扰块匹配,我们首先对带有噪声的图像应用现有的去噪算法。将去噪后的图像作为导频信号进行分块匹配,然后通过CNN结构学习分块去噪函数。实验结果表明,所提出的BMCNN算法达到了最先进的性能。BMCNN既可以恢复重复结构,也可以恢复不规则结构。
{"title":"Block-Matching Convolutional Neural Network (BMCNN): Improving CNN-Based Denoising by Block-Matched Inputs","authors":"Byeongyong Ahn, Yoonsik Kim, G. Park, N. Cho","doi":"10.23919/APSIPA.2018.8659548","DOIUrl":"https://doi.org/10.23919/APSIPA.2018.8659548","url":null,"abstract":"There are two main streams in up-to-date image denoising algorithms: non-local self similarity (NSS) prior based methods and convolutional neural network (CNN) based methods. The NSS based methods are favorable on images with regular and repetitive patterns while the CNN based methods perform better on irregular structures. In this paper, we propose a block-matching convolutional neural network (BMCNN) method that combines NSS prior and CNN. Initially, similar local patches in the input image are integrated into a 3D block. In order to prevent the noise from messing up the block matching, we first apply an existing denoising algorithm on the noisy image. The denoised image is employed as a pilot signal for the block matching, and then denoising function for the block is learned by a CNN structure. Experimental results show that the proposed BMCNN algorithm achieves state-of-the-art performance. In detail, BMCNN can restore both repetitive and irregular structures.","PeriodicalId":287799,"journal":{"name":"2018 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA ASC)","volume":"306 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133912254","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Restoration of dry electrode EEG using deep convolutional neural network 基于深度卷积神经网络的干电极脑电恢复
Yuki Kojoma, Y. Washizawa
Electroencephalography(EEG) has been used widely in biomedical research and consumer products because of its reasonable size and cost. In order to reduce the electrical impedance between electrodes and skin of the scalp, we use conductive gel. However, it takes time to setup EEG. This problem is solved by dry electrodes, which do not require to use the conductive gel, however, the signal quality of dry electrodes is lower than that of wet electrodes. In this research, we propose a method to improve quality of the dry EEG signal. In order to design a restoration filter, we prepare wet and dry EEG signals recorded simultaneously. Then the filter is trained by both wet and dry EEG signals to restore wet EEG signal from dry EEG signal input. We used the fully connected deep neural network (DNN) and convolutional neural network (CNN). We conducted an experiment using the oddball paradigm to demonstrate the proposed method and compare with the classical Wiener filter.
脑电图(EEG)由于其体积和成本合理,在生物医学研究和消费品中得到了广泛的应用。为了减少电极和头皮皮肤之间的电阻抗,我们使用导电凝胶。但是,EEG的设置需要一定的时间。干电极解决了这个问题,它不需要使用导电凝胶,但是干电极的信号质量比湿电极低。在本研究中,我们提出了一种改善干脑电信号质量的方法。为了设计一个恢复滤波器,我们准备了同时记录的干、湿脑电信号。然后分别对干、湿脑电信号进行滤波训练,将输入的干脑电信号还原为湿脑信号。我们使用了全连接深度神经网络(DNN)和卷积神经网络(CNN)。我们使用奇异范式进行了实验,以验证所提出的方法,并与经典的维纳滤波器进行了比较。
{"title":"Restoration of dry electrode EEG using deep convolutional neural network","authors":"Yuki Kojoma, Y. Washizawa","doi":"10.23919/APSIPA.2018.8659676","DOIUrl":"https://doi.org/10.23919/APSIPA.2018.8659676","url":null,"abstract":"Electroencephalography(EEG) has been used widely in biomedical research and consumer products because of its reasonable size and cost. In order to reduce the electrical impedance between electrodes and skin of the scalp, we use conductive gel. However, it takes time to setup EEG. This problem is solved by dry electrodes, which do not require to use the conductive gel, however, the signal quality of dry electrodes is lower than that of wet electrodes. In this research, we propose a method to improve quality of the dry EEG signal. In order to design a restoration filter, we prepare wet and dry EEG signals recorded simultaneously. Then the filter is trained by both wet and dry EEG signals to restore wet EEG signal from dry EEG signal input. We used the fully connected deep neural network (DNN) and convolutional neural network (CNN). We conducted an experiment using the oddball paradigm to demonstrate the proposed method and compare with the classical Wiener filter.","PeriodicalId":287799,"journal":{"name":"2018 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA ASC)","volume":"21 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134000673","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Use of Claimed Speaker Models for Replay Detection 使用声明扬声器模型进行重放检测
Gajan Suthokumar, Kaavya Sriskandaraja, V. Sethu, C. Wijenayake, E. Ambikairajah, Haizhou Li
Replay attacks are the simplest form of spoofing attacks on automatic speaker verification (ASV) systems and consequently the detection of these attacks is a critical research problem. Currently, most research on replay detection focuses on developing a stand-alone countermeasure that runs independently of a speaker verification system by training a single common spoofed model as well as a single common genuine model. This paper investigates the potential advantages of sharing speaker data between the speaker verification system and the replay detection system. Specifically, it explores the benefits of using the claimed speaker's model in place of the common genuine model. The proposed approach is validated on a modified evaluation set of the ASVspoof 2017 version 2.0 corpus and show that the use of adapted speaker models is far superior to the use of a single common genuine model.
重放攻击是自动说话人验证(ASV)系统中最简单的欺骗攻击形式,因此这些攻击的检测是一个关键的研究问题。目前,大多数重播检测的研究都集中在通过训练单个常见的欺骗模型和单个常见的真实模型来开发独立于说话人验证系统运行的独立对策。本文探讨了在说话人验证系统和重播检测系统之间共享说话人数据的潜在优势。具体来说,它探讨了使用声称的说话人模型代替常见的真实模型的好处。在ASVspoof 2017 2.0版本语料库的改进评估集上验证了所提出的方法,并表明使用适应的说话人模型远远优于使用单一的普通真实模型。
{"title":"Use of Claimed Speaker Models for Replay Detection","authors":"Gajan Suthokumar, Kaavya Sriskandaraja, V. Sethu, C. Wijenayake, E. Ambikairajah, Haizhou Li","doi":"10.23919/APSIPA.2018.8659510","DOIUrl":"https://doi.org/10.23919/APSIPA.2018.8659510","url":null,"abstract":"Replay attacks are the simplest form of spoofing attacks on automatic speaker verification (ASV) systems and consequently the detection of these attacks is a critical research problem. Currently, most research on replay detection focuses on developing a stand-alone countermeasure that runs independently of a speaker verification system by training a single common spoofed model as well as a single common genuine model. This paper investigates the potential advantages of sharing speaker data between the speaker verification system and the replay detection system. Specifically, it explores the benefits of using the claimed speaker's model in place of the common genuine model. The proposed approach is validated on a modified evaluation set of the ASVspoof 2017 version 2.0 corpus and show that the use of adapted speaker models is far superior to the use of a single common genuine model.","PeriodicalId":287799,"journal":{"name":"2018 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA ASC)","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134295562","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
A Deep Learning Approach to the Acoustic Condition Monitoring of a Sintering Plant 烧结厂声学状态监测的深度学习方法
Shahab Pasha, C. Ritz, D. Stirling, P. Zulli, D. Pinson, S. Chew
This paper proposes the use of deep learning classification for acoustic monitoring of an industrial process. Specifically, the application is to process sound recordings to detect when additional air leaks through gaps between grate bars lining the bottom of the sinter strand pallets, caused by thermal cycling, aging and deterioration. Detecting holes is not possible visually as the hole is usually small and covered with a granular bed of sinter/blend material. Acoustic signals from normal operation and periods of air leakage are fed into the basic supervised classification methods (SVM and J48) and the deep learning networks, to learn and distinguish the differences. Results suggest that the applied deep learning approach can effectively detect the acoustic emissions from holes time segments with a minimum 79% of accuracy.
本文提出将深度学习分类用于工业过程的声学监测。具体来说,该应用程序是处理录音,以检测由于热循环、老化和变质引起的额外空气何时通过烧结股托盘底部篦条之间的间隙泄漏。由于孔洞通常很小,并且覆盖着烧结/混合材料的颗粒床,因此肉眼无法检测孔洞。将正常运行和漏气时段的声信号输入到基本监督分类方法(SVM和J48)和深度学习网络中,学习和区分差异。结果表明,应用深度学习方法可以有效地检测孔洞时间段的声发射,准确率至少为79%。
{"title":"A Deep Learning Approach to the Acoustic Condition Monitoring of a Sintering Plant","authors":"Shahab Pasha, C. Ritz, D. Stirling, P. Zulli, D. Pinson, S. Chew","doi":"10.23919/APSIPA.2018.8659486","DOIUrl":"https://doi.org/10.23919/APSIPA.2018.8659486","url":null,"abstract":"This paper proposes the use of deep learning classification for acoustic monitoring of an industrial process. Specifically, the application is to process sound recordings to detect when additional air leaks through gaps between grate bars lining the bottom of the sinter strand pallets, caused by thermal cycling, aging and deterioration. Detecting holes is not possible visually as the hole is usually small and covered with a granular bed of sinter/blend material. Acoustic signals from normal operation and periods of air leakage are fed into the basic supervised classification methods (SVM and J48) and the deep learning networks, to learn and distinguish the differences. Results suggest that the applied deep learning approach can effectively detect the acoustic emissions from holes time segments with a minimum 79% of accuracy.","PeriodicalId":287799,"journal":{"name":"2018 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA ASC)","volume":"39 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130781560","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
期刊
2018 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA ASC)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1