首页 > 最新文献

2016 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA)最新文献

英文 中文
No-reference quality assessment for image sharpness and noise 图像清晰度和噪声的无参考质量评估
Lijuan Tang, Xiongkuo Min, V. Jakhetiya, Ke Gu, Xinfeng Zhang, Shuai Yang
To blindly evaluate the visual quality of image is of great importance in many image processing and computer vision applications. In this paper, we develop a novel training-free no-reference (NR) quality metric (QM) based on a unified brain theory, namely, free energy principle. The free energy principle tells that there always exists a difference between an input true visual signal and its processed one by human brain. The difference encompasses the “surprising” information between the real and processed signals. This difference has been found to be highly related to visual quality and attention. More specifically, given a distorted image signal, we first compute the aforesaid difference to approximate its visual quality and saliency via a semi-parametric method that is constructed by combining bilateral filter and auto-regression model. Afterwards, the computed visual saliency and a new natural scene statistic (NSS) model are used for modification to infer the final visual quality score. Extensive experiments are conducted on popular natural scene image databases and a recently released screen content image database for performance comparison. Results have proved the effectiveness of the proposed blind quality measure compared with classical and state-of-the-art full- and no-reference QMs.
在许多图像处理和计算机视觉应用中,对图像的视觉质量进行盲目评价是非常重要的。本文基于统一脑理论即自由能原理,提出了一种新的无训练无参考质量度量(QM)。自由能原理告诉我们,输入的真实视觉信号与人脑处理后的视觉信号之间总是存在差异。这种差异包含了真实信号和经过处理的信号之间的“惊人”信息。这种差异被发现与视觉质量和注意力高度相关。更具体地说,给定一个扭曲的图像信号,我们首先通过结合双边滤波和自回归模型构建的半参数方法计算上述差异以近似其视觉质量和显著性。然后,利用计算的视觉显著性和一种新的自然场景统计(NSS)模型进行修正,推断出最终的视觉质量分数。在流行的自然场景图像数据库和最近发布的屏幕内容图像数据库上进行了大量的实验,以进行性能比较。结果表明,与经典和最先进的全参考质量管理体系和无参考质量管理体系相比,所提出的盲质量管理体系是有效的。
{"title":"No-reference quality assessment for image sharpness and noise","authors":"Lijuan Tang, Xiongkuo Min, V. Jakhetiya, Ke Gu, Xinfeng Zhang, Shuai Yang","doi":"10.1109/APSIPA.2016.7820693","DOIUrl":"https://doi.org/10.1109/APSIPA.2016.7820693","url":null,"abstract":"To blindly evaluate the visual quality of image is of great importance in many image processing and computer vision applications. In this paper, we develop a novel training-free no-reference (NR) quality metric (QM) based on a unified brain theory, namely, free energy principle. The free energy principle tells that there always exists a difference between an input true visual signal and its processed one by human brain. The difference encompasses the “surprising” information between the real and processed signals. This difference has been found to be highly related to visual quality and attention. More specifically, given a distorted image signal, we first compute the aforesaid difference to approximate its visual quality and saliency via a semi-parametric method that is constructed by combining bilateral filter and auto-regression model. Afterwards, the computed visual saliency and a new natural scene statistic (NSS) model are used for modification to infer the final visual quality score. Extensive experiments are conducted on popular natural scene image databases and a recently released screen content image database for performance comparison. Results have proved the effectiveness of the proposed blind quality measure compared with classical and state-of-the-art full- and no-reference QMs.","PeriodicalId":409448,"journal":{"name":"2016 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA)","volume":"52 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122757284","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Automatic heart and lung sounds classification using convolutional neural networks 使用卷积神经网络的自动心肺音分类
Qiyu Chen, Weibin Zhang, Xiang Tian, Xiaoxue Zhang, Shaoqiong Chen, Wenkang Lei
We study the effectiveness of using convolutional neural networks (CNNs) to automatically detect abnormal heart and lung sounds and classify them into different classes in this paper. Heart and respiratory diseases have been affecting humankind for a long time. An effective and automatic diagnostic method is highly attractive since it can help discover potential threat at the early stage, even at home without a professional doctor. We collected a data set containing normal and abnormal heart and lung sounds. These sounds were then annotated by professional doctors. CNNs based systems were implemented to automatically classify the heart sounds into one of the seven categories: normal, bruit de galop, mitral inadequacy, mitral stenosis, interventricular septal defect (IVSD), aortic incompetence, aorta stenosis, and the lung sounds into one of the three categories: normal, moist rales, wheezing rale.
本文研究了利用卷积神经网络(cnn)自动检测异常心肺音并对其进行分类的有效性。心脏和呼吸系统疾病长期以来一直影响着人类。即使在没有专业医生的情况下,也可以在早期发现潜在的威胁,因此有效的自动诊断方法具有很高的吸引力。我们收集了一组包含正常和异常心肺音的数据集。然后由专业医生对这些声音进行注释。采用基于cnn的系统将心音自动分类为正常、杂音、二尖瓣功能不全、二尖瓣狭窄、室间隔缺损、主动脉功能不全、主动脉狭窄等7类之一,将肺音自动分类为正常、湿润音、喘息音等3类之一。
{"title":"Automatic heart and lung sounds classification using convolutional neural networks","authors":"Qiyu Chen, Weibin Zhang, Xiang Tian, Xiaoxue Zhang, Shaoqiong Chen, Wenkang Lei","doi":"10.1109/APSIPA.2016.7820741","DOIUrl":"https://doi.org/10.1109/APSIPA.2016.7820741","url":null,"abstract":"We study the effectiveness of using convolutional neural networks (CNNs) to automatically detect abnormal heart and lung sounds and classify them into different classes in this paper. Heart and respiratory diseases have been affecting humankind for a long time. An effective and automatic diagnostic method is highly attractive since it can help discover potential threat at the early stage, even at home without a professional doctor. We collected a data set containing normal and abnormal heart and lung sounds. These sounds were then annotated by professional doctors. CNNs based systems were implemented to automatically classify the heart sounds into one of the seven categories: normal, bruit de galop, mitral inadequacy, mitral stenosis, interventricular septal defect (IVSD), aortic incompetence, aorta stenosis, and the lung sounds into one of the three categories: normal, moist rales, wheezing rale.","PeriodicalId":409448,"journal":{"name":"2016 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA)","volume":"401 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122787932","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 26
Gaze estimation using 3-D eyeball model and eyelid shapes 基于三维眼球模型和眼睑形状的凝视估计
S. Han, Insung Hwang, Sang Hwa Lee, N. Cho
This paper proposes a gaze estimation algorithm using 3-D eyeball model and eyelid shape. The gaze estimation suffers from differences of eye shapes and individual behaviors, and requires user-specific gaze calibration. The proposed method exploits the usual 3-D eyeball model and shapes of the eyelid to estimate gaze without user-specific calibration and learning. Since the gaze is closely related to the 3-D rotation of eyeball, this paper first derives the relation between 2-D pupil location extracted in the eye image and 3-D rotation of eyeball. This paper also models the shapes of the eyelid to adjust gaze based on the observation that the shapes of the eyelid are deformed with respect to the gaze. This paper models the curvature of eyelid curve to compensate for the gaze. According to the various experiments, the proposed method shows good results in gaze estimation. The proposed method does not need user-specific calibration or gaze learning since the general 3-D eyeball and eyelid models are exploited in the localized eye region. Therefore, it is expected that the proposed gaze estimation algorithm is suitable for various applications such as VR/AR devices, driver gaze tracking, gaze-based interfaces, and so on.
提出了一种基于三维眼球模型和眼睑形状的注视估计算法。注视估计受到眼形和个体行为差异的影响,需要对用户进行特定的注视校准。该方法利用通常的3d眼球模型和眼睑形状来估计凝视,而无需用户特定的校准和学习。由于注视与眼球的三维旋转密切相关,本文首先推导了眼睛图像中提取的二维瞳孔位置与眼球三维旋转之间的关系。根据观察到的眼睑形状相对于凝视的变形,本文还建立了眼睑形状的模型来调整凝视。本文通过对眼睑曲线曲率的建模来补偿凝视。实验结果表明,该方法具有良好的注视估计效果。该方法不需要用户特定的校准或凝视学习,因为一般的3d眼球和眼睑模型在局部眼睛区域被利用。因此,期望本文提出的注视估计算法适用于VR/AR设备、驾驶员注视跟踪、基于注视的接口等各种应用。
{"title":"Gaze estimation using 3-D eyeball model and eyelid shapes","authors":"S. Han, Insung Hwang, Sang Hwa Lee, N. Cho","doi":"10.1109/APSIPA.2016.7820784","DOIUrl":"https://doi.org/10.1109/APSIPA.2016.7820784","url":null,"abstract":"This paper proposes a gaze estimation algorithm using 3-D eyeball model and eyelid shape. The gaze estimation suffers from differences of eye shapes and individual behaviors, and requires user-specific gaze calibration. The proposed method exploits the usual 3-D eyeball model and shapes of the eyelid to estimate gaze without user-specific calibration and learning. Since the gaze is closely related to the 3-D rotation of eyeball, this paper first derives the relation between 2-D pupil location extracted in the eye image and 3-D rotation of eyeball. This paper also models the shapes of the eyelid to adjust gaze based on the observation that the shapes of the eyelid are deformed with respect to the gaze. This paper models the curvature of eyelid curve to compensate for the gaze. According to the various experiments, the proposed method shows good results in gaze estimation. The proposed method does not need user-specific calibration or gaze learning since the general 3-D eyeball and eyelid models are exploited in the localized eye region. Therefore, it is expected that the proposed gaze estimation algorithm is suitable for various applications such as VR/AR devices, driver gaze tracking, gaze-based interfaces, and so on.","PeriodicalId":409448,"journal":{"name":"2016 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA)","volume":"26 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122831271","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Vehicle verification in two nonoverlapped views using sparse representation 使用稀疏表示的两个非重叠视图中的车辆验证
Shih-Chung Hsu, Chung-Lin Huang
Vehicle verification in two different views can be applied for Intelligent Transportation System. However, object appearance matching in two different views is difficult. The vehicle images captured in two views are represented as a feature pair which can be classified as the same/different pair. Sparse representation (SR) has been applied for reconstruction, recognition, and verification. However, the SR dictionary may not guarantee feature sparsity and effective representation. In the paper, we propose Boost-KSVD method without using initial random atom to generate the SR dictionary which can be applied for object verification with very good accuracy. Then, we develop a discriminative criterion to decide the SR dictionary size. Finally, the experiments show that our method can generate better verification accuracy compared with the other methods.
两种不同视角的车辆验证可以应用于智能交通系统。然而,两种不同视角下的物体外观匹配比较困难。在两个视图中捕获的车辆图像被表示为一个特征对,该特征对可以被分类为相同/不同对。稀疏表示(SR)已被应用于重建、识别和验证。然而,SR字典可能不能保证特征稀疏性和有效表示。本文提出了不使用初始随机原子生成SR字典的Boost-KSVD方法,该方法可用于物体验证,具有很好的精度。然后,我们建立了一个判别标准来确定SR字典的大小。最后,实验表明,与其他方法相比,我们的方法可以产生更好的验证精度。
{"title":"Vehicle verification in two nonoverlapped views using sparse representation","authors":"Shih-Chung Hsu, Chung-Lin Huang","doi":"10.1109/APSIPA.2016.7820685","DOIUrl":"https://doi.org/10.1109/APSIPA.2016.7820685","url":null,"abstract":"Vehicle verification in two different views can be applied for Intelligent Transportation System. However, object appearance matching in two different views is difficult. The vehicle images captured in two views are represented as a feature pair which can be classified as the same/different pair. Sparse representation (SR) has been applied for reconstruction, recognition, and verification. However, the SR dictionary may not guarantee feature sparsity and effective representation. In the paper, we propose Boost-KSVD method without using initial random atom to generate the SR dictionary which can be applied for object verification with very good accuracy. Then, we develop a discriminative criterion to decide the SR dictionary size. Finally, the experiments show that our method can generate better verification accuracy compared with the other methods.","PeriodicalId":409448,"journal":{"name":"2016 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA)","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129412511","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
On the training of DNN-based average voice model for speech synthesis 基于dnn的语音合成平均语音模型的训练
Shan Yang, Zhizheng Wu, Lei Xie
Adaptability and controllability are the major advantages of statistical parametric speech synthesis (SPSS) over unit-selection synthesis. Recently, deep neural networks (DNNs) have significantly improved the performance of SPSS. However, current studies are mainly focusing on the training of speaker-dependent DNNs, which generally requires a significant amount of data from a single speaker. In this work, we perform a systematic analysis of the training of multi-speaker average voice model (AVM), which is the foundation of adaptability and controllability of a DNN-based speech synthesis system. Specifically, we employ the i-vector framework to factorise the speaker specific information, which allows a variety of speakers to share all the hidden layers. And the speaker identity vector is augmented with linguistic features in the DNN input. We systematically analyse the impact of the implementations of i-vectors and speaker normalisation.
自适应性和可控性是统计参数语音合成(SPSS)相对于单元选择合成的主要优点。近年来,深度神经网络(dnn)显著提高了SPSS的性能。然而,目前的研究主要集中在说话人依赖的深度神经网络的训练上,这通常需要来自单个说话人的大量数据。在这项工作中,我们对多说话者平均语音模型(AVM)的训练进行了系统的分析,AVM是基于dnn的语音合成系统的适应性和可控性的基础。具体来说,我们采用i-vector框架来分解扬声器特定的信息,从而允许各种扬声器共享所有隐藏层。在深度神经网络输入中对说话人身份向量进行语言特征增强。我们系统地分析了i向量和说话人归一化实现的影响。
{"title":"On the training of DNN-based average voice model for speech synthesis","authors":"Shan Yang, Zhizheng Wu, Lei Xie","doi":"10.1109/APSIPA.2016.7820818","DOIUrl":"https://doi.org/10.1109/APSIPA.2016.7820818","url":null,"abstract":"Adaptability and controllability are the major advantages of statistical parametric speech synthesis (SPSS) over unit-selection synthesis. Recently, deep neural networks (DNNs) have significantly improved the performance of SPSS. However, current studies are mainly focusing on the training of speaker-dependent DNNs, which generally requires a significant amount of data from a single speaker. In this work, we perform a systematic analysis of the training of multi-speaker average voice model (AVM), which is the foundation of adaptability and controllability of a DNN-based speech synthesis system. Specifically, we employ the i-vector framework to factorise the speaker specific information, which allows a variety of speakers to share all the hidden layers. And the speaker identity vector is augmented with linguistic features in the DNN input. We systematically analyse the impact of the implementations of i-vectors and speaker normalisation.","PeriodicalId":409448,"journal":{"name":"2016 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA)","volume":"70 10","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114105486","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
Tactile brain-computer interface using classification of P300 responses evoked by full body spatial vibrotactile stimuli 基于P300反应分类的触觉脑机接口
Takumi Kodama, S. Makino, Tomasz M. Rutkowski
In this study we propose a novel stimulus-driven brain-computer interface (BCI) paradigm, which generates control commands based on classification of somatosensory modality P300 responses. Six spatial vibrotactile stimulus patterns are applied to entire back and limbs of a user. The aim of the current project is to validate an effectiveness of the vibrotactile stimulus patterns for BCI purposes and to establish a novel concept of tactile modality communication link, which shall help locked-in syndrome (LIS) patients, who lose their sight and hearing due to sensory disabilities. We define this approach as a full-body BCI (fbBCI) and we conduct psychophysical stimulus evaluation and realtime EEG response classification experiments with ten healthy body-able users. The grand mean averaged psychophysical stimulus pattern recognition accuracy have resulted at 98.18%, whereas the realtime EEG accuracy at 53.67%. An information-transfer-rate (ITR) scores of all the tested users have ranged from 0.042 to 4.154 bit/minute.
在这项研究中,我们提出了一种新的刺激驱动的脑机接口(BCI)范式,该范式基于体感模态P300反应的分类生成控制命令。六种空间振动触觉刺激模式应用于用户的整个背部和四肢。本项目旨在验证振动触觉刺激模式在脑机接口(BCI)中的有效性,并建立触觉模态沟通链接的新概念,以帮助因感觉障碍而失去视力和听力的闭锁综合征(LIS)患者。我们将这种方法定义为全身脑机接口(fbBCI),并对10名身体健康的用户进行了心理物理刺激评估和实时脑电反应分类实验。心理物理刺激模式识别正确率为98.18%,实时脑电模式识别正确率为53.67%。所有测试用户的信息传输速率(ITR)分数范围为0.042到4.154比特/分钟。
{"title":"Tactile brain-computer interface using classification of P300 responses evoked by full body spatial vibrotactile stimuli","authors":"Takumi Kodama, S. Makino, Tomasz M. Rutkowski","doi":"10.1109/APSIPA.2016.7820734","DOIUrl":"https://doi.org/10.1109/APSIPA.2016.7820734","url":null,"abstract":"In this study we propose a novel stimulus-driven brain-computer interface (BCI) paradigm, which generates control commands based on classification of somatosensory modality P300 responses. Six spatial vibrotactile stimulus patterns are applied to entire back and limbs of a user. The aim of the current project is to validate an effectiveness of the vibrotactile stimulus patterns for BCI purposes and to establish a novel concept of tactile modality communication link, which shall help locked-in syndrome (LIS) patients, who lose their sight and hearing due to sensory disabilities. We define this approach as a full-body BCI (fbBCI) and we conduct psychophysical stimulus evaluation and realtime EEG response classification experiments with ten healthy body-able users. The grand mean averaged psychophysical stimulus pattern recognition accuracy have resulted at 98.18%, whereas the realtime EEG accuracy at 53.67%. An information-transfer-rate (ITR) scores of all the tested users have ranged from 0.042 to 4.154 bit/minute.","PeriodicalId":409448,"journal":{"name":"2016 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA)","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126445374","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
High accuracy reconstruction algorithm for CS-MRI using SDMM 基于SDMM的CS-MRI高精度重建算法
M. Shibata, Norihito Inamuro, Takashi Ijiri, A. Hirabayashi
We propose a high accuracy algorithm for compressed sensing magnetic resonance imaging (CS-MRI) using a convex optimization technique. Lustig et al. proposed CS-MRI technique based on the minimization of a cost function defined by the sum of the data fidelity term, the 11-norm of sparsifying transform coefficients, and a total variation (TV). This function is not differentiable because of both l1-norm and TV. Hence, they used approximations of the non-differentiable terms and a nonlinear conjugate gradient algorithm was applied to minimize the approximated cost function. The obtained solution was also an approximated one, thus of low-quality. In this paper, we propose an algorithm that obtains the exact solution based on the simultaneous direction method of multipliers (SDMM), which is one of the convex optimization techniques. A simple application of SDMM to CS-MRI cannot be implemented because the transformation matrix size is proportional to the square of the image size. We solve this problem using eigenvalue decompositions. Simulations using real MR images show that the proposed algorithm outperforms the conventional one regardless of compression ratio and random sensing patterns.
我们提出了一种使用凸优化技术的高精度压缩感知磁共振成像(CS-MRI)算法。Lustig等人提出了基于最小化成本函数的CS-MRI技术,该成本函数由数据保真度项、稀疏化变换系数的11范数和总变差(TV)的总和定义。由于11范数和TV的存在,这个函数是不可微的。因此,他们使用不可微项的近似,并应用非线性共轭梯度算法来最小化近似的成本函数。所得溶液也是近似溶液,质量较差。本文提出了一种基于乘法器同时方向法(SDMM)的精确求解算法,该算法是凸优化技术中的一种。由于变换矩阵的大小与图像大小的平方成正比,因此无法实现SDMM在CS-MRI中的简单应用。我们用特征值分解来解决这个问题。仿真结果表明,无论压缩比和随机感知模式如何,该算法都优于传统算法。
{"title":"High accuracy reconstruction algorithm for CS-MRI using SDMM","authors":"M. Shibata, Norihito Inamuro, Takashi Ijiri, A. Hirabayashi","doi":"10.1109/APSIPA.2016.7820788","DOIUrl":"https://doi.org/10.1109/APSIPA.2016.7820788","url":null,"abstract":"We propose a high accuracy algorithm for compressed sensing magnetic resonance imaging (CS-MRI) using a convex optimization technique. Lustig et al. proposed CS-MRI technique based on the minimization of a cost function defined by the sum of the data fidelity term, the 11-norm of sparsifying transform coefficients, and a total variation (TV). This function is not differentiable because of both l1-norm and TV. Hence, they used approximations of the non-differentiable terms and a nonlinear conjugate gradient algorithm was applied to minimize the approximated cost function. The obtained solution was also an approximated one, thus of low-quality. In this paper, we propose an algorithm that obtains the exact solution based on the simultaneous direction method of multipliers (SDMM), which is one of the convex optimization techniques. A simple application of SDMM to CS-MRI cannot be implemented because the transformation matrix size is proportional to the square of the image size. We solve this problem using eigenvalue decompositions. Simulations using real MR images show that the proposed algorithm outperforms the conventional one regardless of compression ratio and random sensing patterns.","PeriodicalId":409448,"journal":{"name":"2016 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA)","volume":"72 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128131593","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
The alpha-trimming mean filter for Video stabilization 用于视频稳定的alpha-修剪平均滤波器
Jinju Lim, Min-Cheol Hong
This paper proposed video stabilization techniques using undesired motion detection and alpha-trimming mean filter. The proposed method consists of detecting undesired motions step and filtering the undesired motions step. The limitation on undesired motions is defined, using the local motion information. The alpha-trimming mean filter's alpha is controlled based on this limitation, so that regenerated video is controlled. The experimental results proved that the superior performance of the proposed algorithm.
本文提出了利用非期望运动检测和α -修剪平均滤波器的视频稳定技术。该方法包括检测不需要的运动和过滤不需要的运动两个步骤。利用局部运动信息定义了对非期望运动的限制。基于这一限制,对均值滤波器的alpha值进行控制,从而对生成的视频进行控制。实验结果证明了该算法的优越性能。
{"title":"The alpha-trimming mean filter for Video stabilization","authors":"Jinju Lim, Min-Cheol Hong","doi":"10.1109/APSIPA.2016.7820770","DOIUrl":"https://doi.org/10.1109/APSIPA.2016.7820770","url":null,"abstract":"This paper proposed video stabilization techniques using undesired motion detection and alpha-trimming mean filter. The proposed method consists of detecting undesired motions step and filtering the undesired motions step. The limitation on undesired motions is defined, using the local motion information. The alpha-trimming mean filter's alpha is controlled based on this limitation, so that regenerated video is controlled. The experimental results proved that the superior performance of the proposed algorithm.","PeriodicalId":409448,"journal":{"name":"2016 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA)","volume":"23 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132491108","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Background priors based saliency object detection 基于背景先验的显著性目标检测
Zexia Liu, Guanghua Gu, Chunxia Chen, D. Cui, Chunyu Lin
Saliency object detection is the key process of identifying the location of the object. It has been widely used in numerous applications, including object recognition, image segmentation, video summarization and so on. In this paper, we proposed a saliency object detection approach based on the background priors. First, we obtain a border set by collecting the image border superpixels, in addition remove the superpixels with strong image edges out of the border set to reduce the foreground noises and obtain the true background superpixels seeds. Then, the initial saliency map can be made by computing a background saliency map based on the background seeds and fusing a centered anisotropic Gaussian distribution. Finally, we refine the initial saliency map via the smoothness constraint which encourages neighbor pixels in the image to have the same label. Experimental results on two large benchmark datasets demonstrate that the proposed algorithm performs favorably against other six state-of-art methods in terms of precision, recall and F-Measure. Our method is demonstrated to be more effective in highlighting the salient objects and reducing the background noise.
显著性目标检测是识别目标位置的关键过程。它被广泛应用于物体识别、图像分割、视频摘要等众多领域。本文提出了一种基于背景先验的显著性目标检测方法。首先,我们通过收集图像的边界超像素得到一个边界集,并将边缘较强的超像素从边界集中去除,以降低前景噪声,得到真正的背景超像素种子。然后,根据背景种子计算背景显著性图,融合中心各向异性高斯分布,得到初始显著性图。最后,我们通过平滑约束来优化初始显著性映射,平滑约束鼓励图像中的相邻像素具有相同的标签。在两个大型基准数据集上的实验结果表明,该算法在精度、召回率和F-Measure方面优于其他六种最新方法。该方法在突出突出目标和降低背景噪声方面具有较好的效果。
{"title":"Background priors based saliency object detection","authors":"Zexia Liu, Guanghua Gu, Chunxia Chen, D. Cui, Chunyu Lin","doi":"10.1109/APSIPA.2016.7820744","DOIUrl":"https://doi.org/10.1109/APSIPA.2016.7820744","url":null,"abstract":"Saliency object detection is the key process of identifying the location of the object. It has been widely used in numerous applications, including object recognition, image segmentation, video summarization and so on. In this paper, we proposed a saliency object detection approach based on the background priors. First, we obtain a border set by collecting the image border superpixels, in addition remove the superpixels with strong image edges out of the border set to reduce the foreground noises and obtain the true background superpixels seeds. Then, the initial saliency map can be made by computing a background saliency map based on the background seeds and fusing a centered anisotropic Gaussian distribution. Finally, we refine the initial saliency map via the smoothness constraint which encourages neighbor pixels in the image to have the same label. Experimental results on two large benchmark datasets demonstrate that the proposed algorithm performs favorably against other six state-of-art methods in terms of precision, recall and F-Measure. Our method is demonstrated to be more effective in highlighting the salient objects and reducing the background noise.","PeriodicalId":409448,"journal":{"name":"2016 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130129278","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Octagonal formation of ultrasonic array sensors for fall detection 用于跌倒检测的八角形超声阵列传感器
Chokemongkol Nadee, K. Chamnongthai
Elderly and patients with abnormalities in the control of body movement need automatic alarm with privacy protection when they fall. this paper proposes octagonal formation of ultrasonic array sensors for fall detection system in a smart room. In the method, octagonal array is designed to install ultrasonic sensors on a wall and roof. Signals of ultrasonic reflected from object in the room are processed to recognize fall. In experiments, the proposed method is proved 94% accuracy in fall detection.
老年人和身体运动控制异常的患者,跌倒时需要自动报警,并有隐私保护。提出了一种用于智能房间跌倒检测系统的八角形超声传感器阵列。该方法将超声波传感器设计成八角形阵列安装在墙壁和屋顶上。通过处理房间内物体反射的超声波信号来识别跌倒。实验证明,该方法对跌倒检测的准确率为94%。
{"title":"Octagonal formation of ultrasonic array sensors for fall detection","authors":"Chokemongkol Nadee, K. Chamnongthai","doi":"10.1109/APSIPA.2016.7820819","DOIUrl":"https://doi.org/10.1109/APSIPA.2016.7820819","url":null,"abstract":"Elderly and patients with abnormalities in the control of body movement need automatic alarm with privacy protection when they fall. this paper proposes octagonal formation of ultrasonic array sensors for fall detection system in a smart room. In the method, octagonal array is designed to install ultrasonic sensors on a wall and roof. Signals of ultrasonic reflected from object in the room are processed to recognize fall. In experiments, the proposed method is proved 94% accuracy in fall detection.","PeriodicalId":409448,"journal":{"name":"2016 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA)","volume":"53 2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133875656","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
期刊
2016 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1