首页 > 最新文献

2016 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA)最新文献

英文 中文
Salient object segmentation using a switch scheme 突出目标分割使用切换方案
Ran Shi, K. Ngan, Songnan Li
In this paper, we propose a novel switch scheme and a saliency map binarization method for salient object segmentation. With the proposed switch scheme, the saliency map can be segmented by different methods according to its quality, which is evaluated by a method proposed in this paper. We also develop a binarization method by integrating three properties of the salient object. This method exclusively derives information from the saliency map (i.e., without referring to the original image). Experimental results demonstrate that the proposed binarization method can generate better segmentation results and the switch scheme can further improve the segmentation results by fully exploiting the merit of both segmentation methods.
本文提出了一种新的切换方案和显著性映射二值化方法用于显著性目标分割。利用所提出的切换方案,可以根据显著性映射的质量采用不同的方法进行分割,并用本文提出的方法对显著性映射的质量进行评价。我们还开发了一种二值化方法,通过积分显著目标的三个属性。这种方法完全从显著性图中提取信息(即,不参考原始图像)。实验结果表明,所提出的二值化方法可以产生更好的分割效果,而切换方案则可以充分利用两种分割方法的优点,进一步提高分割效果。
{"title":"Salient object segmentation using a switch scheme","authors":"Ran Shi, K. Ngan, Songnan Li","doi":"10.1109/APSIPA.2016.7820712","DOIUrl":"https://doi.org/10.1109/APSIPA.2016.7820712","url":null,"abstract":"In this paper, we propose a novel switch scheme and a saliency map binarization method for salient object segmentation. With the proposed switch scheme, the saliency map can be segmented by different methods according to its quality, which is evaluated by a method proposed in this paper. We also develop a binarization method by integrating three properties of the salient object. This method exclusively derives information from the saliency map (i.e., without referring to the original image). Experimental results demonstrate that the proposed binarization method can generate better segmentation results and the switch scheme can further improve the segmentation results by fully exploiting the merit of both segmentation methods.","PeriodicalId":409448,"journal":{"name":"2016 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA)","volume":"430 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116143580","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Style-oriented landmark retrieval and summarization 面向风格的地标检索与总结
Wei-Yi Chang, Yi-Ren Yeh, Y. Wang
While the task of visual summarization aims to select representative images from an image collection, we solve a unique problem of style-oriented landmark retrieval and summarization from photographic images of a city. Instead of performing summarization or clustering on landmark images from a city, we allow the user to provide a query input which is not from the city of interest, while the goal is to retrieve and summarize the landmark images with similar style-dependent landmark images, followed by a style-consistent image summarization across landmark categories. As a result, our summarized outputs from various landmarks would exhibit similar image style as that of the query. Our experiments will confirm that the use of our proposed method is able to perform favorably against existing or baseline approaches with improved query-dependent style consistency.
视觉摘要的任务是从图像集合中选择具有代表性的图像,而我们解决了从城市摄影图像中以风格为导向的地标检索和摘要的独特问题。我们允许用户提供不是来自感兴趣城市的查询输入,而不是对来自城市的地标图像进行汇总或聚类,而目标是检索和汇总具有相似风格依赖的地标图像,然后跨地标类别进行风格一致的图像汇总。因此,我们从各个地标汇总的输出将显示与查询相似的图像样式。我们的实验将证实,使用我们提出的方法能够更好地执行现有方法或基线方法,并改进了依赖于查询的样式一致性。
{"title":"Style-oriented landmark retrieval and summarization","authors":"Wei-Yi Chang, Yi-Ren Yeh, Y. Wang","doi":"10.1109/APSIPA.2016.7820857","DOIUrl":"https://doi.org/10.1109/APSIPA.2016.7820857","url":null,"abstract":"While the task of visual summarization aims to select representative images from an image collection, we solve a unique problem of style-oriented landmark retrieval and summarization from photographic images of a city. Instead of performing summarization or clustering on landmark images from a city, we allow the user to provide a query input which is not from the city of interest, while the goal is to retrieve and summarize the landmark images with similar style-dependent landmark images, followed by a style-consistent image summarization across landmark categories. As a result, our summarized outputs from various landmarks would exhibit similar image style as that of the query. Our experiments will confirm that the use of our proposed method is able to perform favorably against existing or baseline approaches with improved query-dependent style consistency.","PeriodicalId":409448,"journal":{"name":"2016 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA)","volume":"300 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114384609","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Dynamic convolutional neural network for activity recognition 动态卷积神经网络用于活动识别
Chih-Hsiang You, Chen-Kuo Chiang
In this paper, a novel Dynamic Convolutional Neural Network (D-CNN) is proposed using sensor data for activity recognition. Sensor data collected for activity recognition is usually not well-aligned. It may also contains noises and variations from different persons. To overcome these challenges, Gaussian Mixture Models (GMM) is exploited to capture the distribution of each activity. Then, sensor data and the GMMs are screened into different segments. These segments form multiple paths in the Convolutional Neural Network. During testing, Gaussian Mixture Regression (GMR) is applied to dynamically fit segments of test signals into corresponding paths in the CNN. Experimental results demonstrate the superior performance of D-CNN to other learning methods.
本文提出了一种利用传感器数据进行活动识别的动态卷积神经网络(D-CNN)。为活动识别收集的传感器数据通常没有很好地对齐。它也可能包含来自不同人的噪音和变化。为了克服这些挑战,利用高斯混合模型(GMM)来捕获每个活动的分布。然后,传感器数据和GMMs被筛选成不同的片段。这些片段在卷积神经网络中形成多条路径。在测试过程中,使用高斯混合回归(GMR)将测试信号的片段动态拟合到CNN中相应的路径中。实验结果表明,D-CNN的学习性能优于其他学习方法。
{"title":"Dynamic convolutional neural network for activity recognition","authors":"Chih-Hsiang You, Chen-Kuo Chiang","doi":"10.1109/APSIPA.2016.7820749","DOIUrl":"https://doi.org/10.1109/APSIPA.2016.7820749","url":null,"abstract":"In this paper, a novel Dynamic Convolutional Neural Network (D-CNN) is proposed using sensor data for activity recognition. Sensor data collected for activity recognition is usually not well-aligned. It may also contains noises and variations from different persons. To overcome these challenges, Gaussian Mixture Models (GMM) is exploited to capture the distribution of each activity. Then, sensor data and the GMMs are screened into different segments. These segments form multiple paths in the Convolutional Neural Network. During testing, Gaussian Mixture Regression (GMR) is applied to dynamically fit segments of test signals into corresponding paths in the CNN. Experimental results demonstrate the superior performance of D-CNN to other learning methods.","PeriodicalId":409448,"journal":{"name":"2016 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA)","volume":"43 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127539619","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Image copy-move forgery detection using hierarchical feature point matching 基于分层特征点匹配的图像复制-移动伪造检测
Yuanman Li, Jiantao Zhou
Copy-move forgery is one of the most commonly used manipulations for tempering digital images. Keypoint-based detection methods have been reported to be very effective in revealing copy-move evidences, due to their robustness against geometric transforms. However, these methods fail to handle the cases when copy-move forgery only involves small or smooth regions, where the number of keypoints is very limited. To tackle this challenge, we propose a simple yet effective copy-move forgery detection approach. By lowering the contrast threshold and rescaling the input image, we first generate a sufficient number of keypoints that exist even in the small or smooth regions. Then, a novel hierarchical matching strategy is developed for solving the keypoint matching problems. Finally, a novel iterative homography estimation technique is suggested through exploiting the dominant orientation information of each keypoint. Extensive experimental results are provided to demonstrate the superior performance of the proposed scheme.
复制-移动伪造是篡改数字图像最常用的手法之一。据报道,基于关键点的检测方法在揭示复制移动证据方面非常有效,因为它们对几何变换具有鲁棒性。然而,这些方法无法处理复制-移动伪造仅涉及小区域或光滑区域的情况,这些区域的关键点数量非常有限。为了解决这一挑战,我们提出了一种简单而有效的复制-移动伪造检测方法。通过降低对比度阈值和重新缩放输入图像,我们首先生成足够数量的关键点,即使在小区域或光滑区域也存在。然后,提出了一种新的分层匹配策略来解决关键点匹配问题。最后,利用各关键点的优势方向信息,提出了一种新的迭代单应性估计技术。大量的实验结果证明了该方案的优越性能。
{"title":"Image copy-move forgery detection using hierarchical feature point matching","authors":"Yuanman Li, Jiantao Zhou","doi":"10.1109/APSIPA.2016.7820758","DOIUrl":"https://doi.org/10.1109/APSIPA.2016.7820758","url":null,"abstract":"Copy-move forgery is one of the most commonly used manipulations for tempering digital images. Keypoint-based detection methods have been reported to be very effective in revealing copy-move evidences, due to their robustness against geometric transforms. However, these methods fail to handle the cases when copy-move forgery only involves small or smooth regions, where the number of keypoints is very limited. To tackle this challenge, we propose a simple yet effective copy-move forgery detection approach. By lowering the contrast threshold and rescaling the input image, we first generate a sufficient number of keypoints that exist even in the small or smooth regions. Then, a novel hierarchical matching strategy is developed for solving the keypoint matching problems. Finally, a novel iterative homography estimation technique is suggested through exploiting the dominant orientation information of each keypoint. Extensive experimental results are provided to demonstrate the superior performance of the proposed scheme.","PeriodicalId":409448,"journal":{"name":"2016 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA)","volume":"11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125344320","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 12
Computer-assisted pronunciation training: From pronunciation scoring towards spoken language learning 计算机辅助发音训练:从发音评分到口语学习
Nancy F. Chen, Haizhou Li
This paper reviews the research approaches used in computer-assisted pronunciation training (CAPT), addresses the existing challenges, and discusses emerging trends and opportunities. To complement existing work, our analysis places more emphasis on pronunciation teaching and learning (as opposed to pronunciation assessment), prosodic error detection (as opposed to phonetic error detection), and research work from the past five years given the recent rapid development in spoken language technology.
本文回顾了计算机辅助发音训练(CAPT)的研究方法,指出了目前存在的挑战,并讨论了新的趋势和机遇。为了补充现有的工作,我们的分析更侧重于语音教学(而不是语音评估),韵律错误检测(而不是语音错误检测),以及最近五年口语技术快速发展的研究工作。
{"title":"Computer-assisted pronunciation training: From pronunciation scoring towards spoken language learning","authors":"Nancy F. Chen, Haizhou Li","doi":"10.1109/APSIPA.2016.7820782","DOIUrl":"https://doi.org/10.1109/APSIPA.2016.7820782","url":null,"abstract":"This paper reviews the research approaches used in computer-assisted pronunciation training (CAPT), addresses the existing challenges, and discusses emerging trends and opportunities. To complement existing work, our analysis places more emphasis on pronunciation teaching and learning (as opposed to pronunciation assessment), prosodic error detection (as opposed to phonetic error detection), and research work from the past five years given the recent rapid development in spoken language technology.","PeriodicalId":409448,"journal":{"name":"2016 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA)","volume":"21 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127038199","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 35
A novel paragraph embedding method for spoken document summarization 一种新的语音文档摘要段落嵌入方法
Kuan-Yu Chen, Shih-Hung Liu, Berlin Chen, H. Wang
Representation learning has emerged as a newly active research subject in many machine learning applications because of its excellent performance. In the context of natural language processing, paragraph (or sentence and document) embedding learning is more suitable/reasonable for some tasks, such as information retrieval and document summarization. However, as far as we are aware, there is only a dearth of research focusing on launching paragraph embedding methods. Extractive spoken document summarization, which can help us browse and digest multimedia data efficiently, aims at selecting a set of indicative sentences from a source document to express the most important theme of the document. A general consensus is that relevance and redundancy are both critical issues in a realistic summarization scenario. However, most of the existing methods focus on determining only the relevance degree between a pair of sentence and document. Motivated by these observations, three major contributions are proposed in this paper. First, we propose a novel unsupervised paragraph embedding method, named the essence vector model, which aims at not only distilling the most representative information from a paragraph but also getting rid of the general background information to produce a more informative low-dimensional vector representation. Second, we incorporate the deduced essence vectors with a density peaks clustering summarization method, which can take both relevance and redundancy information into account simultaneously, to enhance the spoken document summarization performance. Third, the effectiveness of our proposed methods over several well-practiced and state-of-the-art methods is confirmed by extensive spoken document summarization experiments.
表征学习以其优异的性能在许多机器学习应用中成为一个新兴的活跃研究课题。在自然语言处理的背景下,段落(或句子和文档)嵌入学习更适合于某些任务,如信息检索和文档摘要。然而,据我们所知,目前还缺乏针对启动段落嵌入方法的研究。摘要摘要的目的是从源文档中选择一组指示句来表达该文档最重要的主题,可以帮助我们高效地浏览和消化多媒体数据。一个普遍的共识是,相关性和冗余都是现实总结场景中的关键问题。然而,大多数现有的方法只关注于确定一对句子和文档之间的关联度。在这些观察的激励下,本文提出了三个主要贡献。首先,我们提出了一种新的无监督段落嵌入方法——本质向量模型,该方法既能从段落中提取出最具代表性的信息,又能去除一般的背景信息,产生信息量更大的低维向量表示。其次,我们将推导出的本质向量与同时考虑相关性和冗余信息的密度峰聚类摘要方法相结合,提高了语音文档的摘要性能。第三,我们提出的方法的有效性超过了几个实践良好的和最先进的方法是通过广泛的口头文件摘要实验证实。
{"title":"A novel paragraph embedding method for spoken document summarization","authors":"Kuan-Yu Chen, Shih-Hung Liu, Berlin Chen, H. Wang","doi":"10.1109/APSIPA.2016.7820882","DOIUrl":"https://doi.org/10.1109/APSIPA.2016.7820882","url":null,"abstract":"Representation learning has emerged as a newly active research subject in many machine learning applications because of its excellent performance. In the context of natural language processing, paragraph (or sentence and document) embedding learning is more suitable/reasonable for some tasks, such as information retrieval and document summarization. However, as far as we are aware, there is only a dearth of research focusing on launching paragraph embedding methods. Extractive spoken document summarization, which can help us browse and digest multimedia data efficiently, aims at selecting a set of indicative sentences from a source document to express the most important theme of the document. A general consensus is that relevance and redundancy are both critical issues in a realistic summarization scenario. However, most of the existing methods focus on determining only the relevance degree between a pair of sentence and document. Motivated by these observations, three major contributions are proposed in this paper. First, we propose a novel unsupervised paragraph embedding method, named the essence vector model, which aims at not only distilling the most representative information from a paragraph but also getting rid of the general background information to produce a more informative low-dimensional vector representation. Second, we incorporate the deduced essence vectors with a density peaks clustering summarization method, which can take both relevance and redundancy information into account simultaneously, to enhance the spoken document summarization performance. Third, the effectiveness of our proposed methods over several well-practiced and state-of-the-art methods is confirmed by extensive spoken document summarization experiments.","PeriodicalId":409448,"journal":{"name":"2016 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA)","volume":"11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128122445","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Travel photo album summarization based on aesthetic quality, interestingness, and memorableness 旅游相册总结基于审美质量,趣味性,和记忆性
Jun-Hyuk Kim, Jong-Seok Lee
Photo album summarization refers to the process of choosing a representative subset of photos in a photo album. In this paper, we propose a novel system capable of automatic photo album summarization based on three fundamental criteria, namely, aesthetic quality, interestingness, and memorableness. Based on these criteria, steps for filtering and scoring photos are designed. Through an experiment with photo albums of different sizes, it is demonstrated that the proposed system works well consistently.
相册汇总是指在相册中选择具有代表性的照片子集的过程。本文提出了一种基于美感、趣味性和可记忆性三个基本标准的相册自动摘要系统。基于这些标准,设计了对照片进行过滤和评分的步骤。通过对不同尺寸的相册进行实验,证明了该系统的一致性。
{"title":"Travel photo album summarization based on aesthetic quality, interestingness, and memorableness","authors":"Jun-Hyuk Kim, Jong-Seok Lee","doi":"10.1109/APSIPA.2016.7820889","DOIUrl":"https://doi.org/10.1109/APSIPA.2016.7820889","url":null,"abstract":"Photo album summarization refers to the process of choosing a representative subset of photos in a photo album. In this paper, we propose a novel system capable of automatic photo album summarization based on three fundamental criteria, namely, aesthetic quality, interestingness, and memorableness. Based on these criteria, steps for filtering and scoring photos are designed. Through an experiment with photo albums of different sizes, it is demonstrated that the proposed system works well consistently.","PeriodicalId":409448,"journal":{"name":"2016 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA)","volume":"43 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128160268","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
A discriminative training method incorporating pronunciation variations for dysarthric automatic speech recognition 一种结合发音变化的辨别性训练方法用于困难语音自动识别
Woo Kyeong Seong, Nam Kyun Kim, H. Ha, H. Kim
While dysarthric speech recognition can be a convenient interface for dysarthric speakers, it is hard to collect enough speech data to overcome the underestimation problem of acoustic models. In addition, there are lots of pronunciation variations in the collected database due to the paralysis of the articulator of dysarthric speakers. Thus, a discriminative training method is proposed for improving the performance of such resource-limited dysarthric speech recognition. The proposed method is applied to subspace Gaussian mixture modeling by incorporating pronunciation variations into a conventional minimum phone error discriminative training method.
虽然困难语音识别可以为困难语音说话者提供方便的接口,但很难收集足够的语音数据来克服声学模型的低估问题。此外,由于发音困难的说话者的发音麻痹,在收集到的数据库中存在大量的发音变异。因此,我们提出了一种判别训练方法来提高这种资源有限的困难语音识别的性能。该方法将语音变化与传统的最小电话误差判别训练方法相结合,应用于子空间高斯混合建模。
{"title":"A discriminative training method incorporating pronunciation variations for dysarthric automatic speech recognition","authors":"Woo Kyeong Seong, Nam Kyun Kim, H. Ha, H. Kim","doi":"10.1109/APSIPA.2016.7820840","DOIUrl":"https://doi.org/10.1109/APSIPA.2016.7820840","url":null,"abstract":"While dysarthric speech recognition can be a convenient interface for dysarthric speakers, it is hard to collect enough speech data to overcome the underestimation problem of acoustic models. In addition, there are lots of pronunciation variations in the collected database due to the paralysis of the articulator of dysarthric speakers. Thus, a discriminative training method is proposed for improving the performance of such resource-limited dysarthric speech recognition. The proposed method is applied to subspace Gaussian mixture modeling by incorporating pronunciation variations into a conventional minimum phone error discriminative training method.","PeriodicalId":409448,"journal":{"name":"2016 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA)","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125218903","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Sparse spatial filtering in frequency domain of multi-channel EEG for frequency and phase detection 多通道脑电图频域稀疏空间滤波用于频率和相位检测
Naoki Morikawa, Toshihisa Tanaka
A brain-computer interface (BCI) based on steady state visual evoked potentials (SSVEPs) is one of the most practical BCI, because of high recognition accuracies and short time training. To increase the number of commands of SSVEP-based BCI, recently a frequency and phase mixed-coded SSVEP BCI has been proposed. However, in order to detect frequency and phase of SSVEPs accurately, it is required to treat multi-channel phases to select useful channels for detecting commands. In this paper, we propose a novel method for estimating both frequency and phase of SSVEPs with sparse complex spatial filters. We conducted experiments for evaluating the performance of the proposed method in a mixed-coded SSVEP based BCI. As a result, the proposed method showed higher recognition accuracies and lower calculation cost of command detection than conventional methods. Moreover, the proposed method achieved automatic channel selection.
基于稳态视觉诱发电位(SSVEPs)的脑机接口(BCI)具有识别准确率高、训练时间短等优点,是目前最实用的脑机接口之一。为了增加基于SSVEP的BCI的命令数量,最近提出了一种频率和相位混合编码的SSVEP BCI。然而,为了准确地检测ssvep的频率和相位,需要对多通道相位进行处理,以选择有用的通道来检测命令。在本文中,我们提出了一种利用稀疏复空间滤波器估计ssvep频率和相位的新方法。我们进行了实验,以评估所提出的方法在基于混合编码SSVEP的BCI中的性能。结果表明,该方法具有较高的识别精度和较低的指令检测计算成本。此外,该方法实现了信道的自动选择。
{"title":"Sparse spatial filtering in frequency domain of multi-channel EEG for frequency and phase detection","authors":"Naoki Morikawa, Toshihisa Tanaka","doi":"10.1109/APSIPA.2016.7820779","DOIUrl":"https://doi.org/10.1109/APSIPA.2016.7820779","url":null,"abstract":"A brain-computer interface (BCI) based on steady state visual evoked potentials (SSVEPs) is one of the most practical BCI, because of high recognition accuracies and short time training. To increase the number of commands of SSVEP-based BCI, recently a frequency and phase mixed-coded SSVEP BCI has been proposed. However, in order to detect frequency and phase of SSVEPs accurately, it is required to treat multi-channel phases to select useful channels for detecting commands. In this paper, we propose a novel method for estimating both frequency and phase of SSVEPs with sparse complex spatial filters. We conducted experiments for evaluating the performance of the proposed method in a mixed-coded SSVEP based BCI. As a result, the proposed method showed higher recognition accuracies and lower calculation cost of command detection than conventional methods. Moreover, the proposed method achieved automatic channel selection.","PeriodicalId":409448,"journal":{"name":"2016 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA)","volume":"31 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122092336","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
A fast multi-focus image fusion algorithm by DWT and focused region decision map 一种基于小波变换和聚焦区域决策图的快速多焦点图像融合算法
Shumin Liu, Jiajia Chen
To comprise the advantages of both the spatial domain and transform domain methods, this paper presents a novel hybrid algorithm for multi-focus images fusion, which reduces the error rate of sub-band coefficients selection in the transform domain and reduce the artificial discontinuities created in the spatial domain algorithms. In this method, wavelet transforms are firstly performed on each input image, and a focused region decision map is established based on the high-frequency sub-bands extraction. The fusion rules are then guided by this map, and the fused coefficients are transformed back to form the fused image. Experimental results demonstrate that the proposed method is better than various existing methods, in term of fusion quality benchmarks. In addition, the proposed algorithm has a complexity proportional to the total number of pixels in the image, which is lower than some other algorithm which may produce similar fusion quality with the proposed algorithm. Furthermore, the proposed algorithm only requires one level wavelet decomposition, again reducing the processing time. With the proposed method, high quality and fast multi-focus image fusion is made possible.
结合空间域和变换域方法的优点,提出了一种新的多焦点图像融合混合算法,降低了变换域子带系数选择的错误率,减少了空间域算法中产生的人为不连续。该方法首先对输入图像进行小波变换,并在提取高频子带的基础上建立焦点区域决策图。然后利用该映射来指导融合规则,并将融合系数变换回来形成融合图像。实验结果表明,该方法在融合质量基准方面优于现有的各种方法。此外,本文算法的复杂度与图像中像素的总数成正比,比其他可能产生与本文算法相似的融合质量的算法要低。此外,该算法只需要一级小波分解,再次减少了处理时间。该方法可实现高质量、快速的多焦点图像融合。
{"title":"A fast multi-focus image fusion algorithm by DWT and focused region decision map","authors":"Shumin Liu, Jiajia Chen","doi":"10.1109/APSIPA.2016.7820864","DOIUrl":"https://doi.org/10.1109/APSIPA.2016.7820864","url":null,"abstract":"To comprise the advantages of both the spatial domain and transform domain methods, this paper presents a novel hybrid algorithm for multi-focus images fusion, which reduces the error rate of sub-band coefficients selection in the transform domain and reduce the artificial discontinuities created in the spatial domain algorithms. In this method, wavelet transforms are firstly performed on each input image, and a focused region decision map is established based on the high-frequency sub-bands extraction. The fusion rules are then guided by this map, and the fused coefficients are transformed back to form the fused image. Experimental results demonstrate that the proposed method is better than various existing methods, in term of fusion quality benchmarks. In addition, the proposed algorithm has a complexity proportional to the total number of pixels in the image, which is lower than some other algorithm which may produce similar fusion quality with the proposed algorithm. Furthermore, the proposed algorithm only requires one level wavelet decomposition, again reducing the processing time. With the proposed method, high quality and fast multi-focus image fusion is made possible.","PeriodicalId":409448,"journal":{"name":"2016 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA)","volume":"107 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123233314","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
期刊
2016 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1