首页 > 最新文献

2016 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA)最新文献

英文 中文
Mask design for pinhole-array-based hand-held light field cameras with applications in depth estimation 基于针孔阵列的手持光场相机掩模设计及其在深度估计中的应用
Chen-Wei Chang, Min-Hung Chen, Kuan-Chang Chen, Chi-Ming Yeh, Yi-Chang Lu
Pinhole-array-based hand-held light field cameras can be used to capture 4-dimensional light field data for different applications such as digital refocusing and depth estimation. Our previous experiences suggest the design of the pinhole array mask is very critical to the performance of the camera, and the selection of mask parameters could be very different between applications. In this paper, we derive equations for determining the parameters of pinhole masks. The proposed physically-based model can be applied to cameras of different pixel sizes. The experimental results which match the proposed model are also provided at the end of this paper.
基于针孔阵列的手持式光场相机可用于捕获四维光场数据,用于数字重聚焦和深度估计等不同应用。我们以往的经验表明,针孔阵列掩模的设计对相机的性能至关重要,并且在不同的应用中,掩模参数的选择可能会有很大的不同。本文导出了确定针孔掩模参数的方程。提出的基于物理的模型可以应用于不同像素大小的相机。最后给出了与该模型相匹配的实验结果。
{"title":"Mask design for pinhole-array-based hand-held light field cameras with applications in depth estimation","authors":"Chen-Wei Chang, Min-Hung Chen, Kuan-Chang Chen, Chi-Ming Yeh, Yi-Chang Lu","doi":"10.1109/APSIPA.2016.7820688","DOIUrl":"https://doi.org/10.1109/APSIPA.2016.7820688","url":null,"abstract":"Pinhole-array-based hand-held light field cameras can be used to capture 4-dimensional light field data for different applications such as digital refocusing and depth estimation. Our previous experiences suggest the design of the pinhole array mask is very critical to the performance of the camera, and the selection of mask parameters could be very different between applications. In this paper, we derive equations for determining the parameters of pinhole masks. The proposed physically-based model can be applied to cameras of different pixel sizes. The experimental results which match the proposed model are also provided at the end of this paper.","PeriodicalId":409448,"journal":{"name":"2016 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA)","volume":"51 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121208538","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Stereo matching using census transform of adaptive window sizes with gradient images 基于自适应窗口大小普查变换的梯度图像立体匹配
Jaeryun Ko, Yo-Sung Ho
The census transform in computing the matching cost of stereo matching is simple and robust under luminance variations in stereo image pairs; however, different disparity maps are generated depending on the shape and size of the census transform window. In this paper, we propose a stereo matching method with variable sizes of census transform windows based on the gradients of stereo images. Our experiment shows higher accuracy of disparity values in the area of depth discontinuities.
在立体图像对亮度变化的情况下,普查变换计算立体匹配的匹配代价简单、鲁棒;然而,根据人口普查变换窗口的形状和大小,会生成不同的视差图。本文提出了一种基于立体图像梯度的人口普查变换窗口大小可变的立体匹配方法。实验表明,深度不连续区域的视差值精度较高。
{"title":"Stereo matching using census transform of adaptive window sizes with gradient images","authors":"Jaeryun Ko, Yo-Sung Ho","doi":"10.1109/APSIPA.2016.7820827","DOIUrl":"https://doi.org/10.1109/APSIPA.2016.7820827","url":null,"abstract":"The census transform in computing the matching cost of stereo matching is simple and robust under luminance variations in stereo image pairs; however, different disparity maps are generated depending on the shape and size of the census transform window. In this paper, we propose a stereo matching method with variable sizes of census transform windows based on the gradients of stereo images. Our experiment shows higher accuracy of disparity values in the area of depth discontinuities.","PeriodicalId":409448,"journal":{"name":"2016 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA)","volume":"71 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126815594","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Frame rate up conversion for multiview video 帧率上升转换为多视图视频
Yoonmo Yang, Dohoon Lee, Byung Tae Oh
In this paper, we propose a new frame rate up conversion method for multiview video. The proposed method uses the depth map and neighboring view information for the improvement of motion estimation and compensation accuracy. In details, it decomposes a block into multiple layers with depth map. Then it estimates the occluded regions in the lower layer using their neighboring view information, which consequently leads more accurate motion estimation and compensation. The experimental results show that the proposed method highly improves the quality of the interpolated frames compared to the conventional methods.
本文提出了一种新的多视点视频帧率提升转换方法。该方法利用深度图和相邻视图信息,提高了运动估计和补偿精度。具体来说,它通过深度图将一个块分解成多个层。然后利用被遮挡区域的相邻视图信息对被遮挡区域进行估计,从而得到更精确的运动估计和补偿。实验结果表明,与传统插值方法相比,该方法大大提高了插值帧的质量。
{"title":"Frame rate up conversion for multiview video","authors":"Yoonmo Yang, Dohoon Lee, Byung Tae Oh","doi":"10.1109/APSIPA.2016.7820714","DOIUrl":"https://doi.org/10.1109/APSIPA.2016.7820714","url":null,"abstract":"In this paper, we propose a new frame rate up conversion method for multiview video. The proposed method uses the depth map and neighboring view information for the improvement of motion estimation and compensation accuracy. In details, it decomposes a block into multiple layers with depth map. Then it estimates the occluded regions in the lower layer using their neighboring view information, which consequently leads more accurate motion estimation and compensation. The experimental results show that the proposed method highly improves the quality of the interpolated frames compared to the conventional methods.","PeriodicalId":409448,"journal":{"name":"2016 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA)","volume":"05 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127367435","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Discrete feature transform for low-complexity single-image super-resolution 低复杂度单图像超分辨率的离散特征变换
Jonghee Kim, Changick Kim
Dictionary-based super-resolution is actively studied with successful achievements. However, previous dictionary-based super-resolution methods exploit optimization or nearest neighbor search which has high complexity. In this paper, we propose a low-complexity super-resolution method called the discrete feature transform which performs feature extraction and nearest neighbor search at once. As a result, the proposed method achieves the lowest complexity among dictionary-based super-resolution methods with a comparable performance.
基于字典的超分辨技术得到了积极的研究,并取得了成功的成果。然而,以往基于字典的超分辨方法采用优化或最近邻搜索,其复杂度较高。本文提出了一种低复杂度的超分辨率方法——离散特征变换,它可以同时进行特征提取和最近邻搜索。结果表明,在性能相当的情况下,该方法是基于字典的超分辨率方法中复杂度最低的方法。
{"title":"Discrete feature transform for low-complexity single-image super-resolution","authors":"Jonghee Kim, Changick Kim","doi":"10.1109/APSIPA.2016.7820852","DOIUrl":"https://doi.org/10.1109/APSIPA.2016.7820852","url":null,"abstract":"Dictionary-based super-resolution is actively studied with successful achievements. However, previous dictionary-based super-resolution methods exploit optimization or nearest neighbor search which has high complexity. In this paper, we propose a low-complexity super-resolution method called the discrete feature transform which performs feature extraction and nearest neighbor search at once. As a result, the proposed method achieves the lowest complexity among dictionary-based super-resolution methods with a comparable performance.","PeriodicalId":409448,"journal":{"name":"2016 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA)","volume":"31 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127469234","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Adaptive subspace-constrained diagonal loading 自适应子空间约束对角加载
Yueh-Ting Tsai, B. Su, Yu Tsao, Syu-Siang Wang
Recently, a subspace-constrained diagonal loading (SSC-DL) method has been proposed for robust beamforming against the mismatched direction of arrival (DoA) issue. Although SSC-DL has outstanding output SINR performance, it is not clear how to choose the DL factor and subspace dimension in practice. The aim of the present study is to further investigate conditions on optimal parameters for SSC-DL and algorithms to determine them in realistic test conditions. First, we proposed to use the Capon power spectrum density to determine the desired signal power, which is then used to compute the optimal DL factor for SSC-DL. Next, a novel adaptive SSC-DL approach (adaptive-SSC-DL) is proposed, which can dynamically optimize the sub-space dimension based on the test conditions. Simulation results show that adaptive-SSC-DL provides higher output SINR than several existing methods and achieves comparable performance comparing to SSC-DL with ideal parameter setup.
近年来,针对到达方向不匹配问题,提出了一种子空间约束对角加载(SSC-DL)的鲁棒波束形成方法。虽然SSC-DL具有出色的输出SINR性能,但在实际中如何选择DL因子和子空间维数并不是很明确。本研究的目的是进一步研究SSC-DL最优参数的条件和在实际测试条件下确定它们的算法。首先,我们提出使用Capon功率谱密度来确定所需的信号功率,然后使用该功率谱密度来计算SSC-DL的最优DL因子。其次,提出了一种新的自适应SSC-DL方法(adaptive-SSC-DL),该方法可以根据测试条件动态优化子空间维度。仿真结果表明,自适应SSC-DL方法的输出信噪比高于几种现有方法,并且在理想参数设置下与SSC-DL方法性能相当。
{"title":"Adaptive subspace-constrained diagonal loading","authors":"Yueh-Ting Tsai, B. Su, Yu Tsao, Syu-Siang Wang","doi":"10.1109/APSIPA.2016.7820704","DOIUrl":"https://doi.org/10.1109/APSIPA.2016.7820704","url":null,"abstract":"Recently, a subspace-constrained diagonal loading (SSC-DL) method has been proposed for robust beamforming against the mismatched direction of arrival (DoA) issue. Although SSC-DL has outstanding output SINR performance, it is not clear how to choose the DL factor and subspace dimension in practice. The aim of the present study is to further investigate conditions on optimal parameters for SSC-DL and algorithms to determine them in realistic test conditions. First, we proposed to use the Capon power spectrum density to determine the desired signal power, which is then used to compute the optimal DL factor for SSC-DL. Next, a novel adaptive SSC-DL approach (adaptive-SSC-DL) is proposed, which can dynamically optimize the sub-space dimension based on the test conditions. Simulation results show that adaptive-SSC-DL provides higher output SINR than several existing methods and achieves comparable performance comparing to SSC-DL with ideal parameter setup.","PeriodicalId":409448,"journal":{"name":"2016 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA)","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129176065","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Size-Invariant Fully Convolutional Neural Network for vessel segmentation of digital retinal images 基于全卷积神经网络的数字视网膜图像血管分割
Yuan-sheng Luo, Hong Cheng, Lu Yang
Vessel segmentation of digital retinal images plays an important role in diagnosis of diseases such as diabetics, hypertension and retinopathy of prematurity due to these diseases impact the retina. In this paper, a novel Size-Invariant Fully Convolutional Neural Network (SIFCN) is proposed to address the automatic retinal vessel segmentation problems. The input data of the network is the patches of images and the corresponding pixel-wise labels. A consecutive convolution layers and pooling layers follow the input data, so that the network can learn the abstract features to segment retinal vessel. Our network is designed to hold the height and width of data of each layer with padding and assign pooling stride so that the spatial information maintain and up-sample is not required. Compared with the pixel-wise retinal vessel segmentation approaches, our patch-wise segmentation is much more efficient since in each cycle it can predict all the pixels of the patch. Our overlapped SIFCN approach achieves accuracy of 0.9471, with the AUC of 0.9682. And our non-overlap SIFCN is the most efficient approach among the deep learning approaches, costing only 3.68 seconds per image, and the overlapped SIFCN costs 31.17 seconds per image.
由于糖尿病、高血压、早产儿视网膜病变等疾病对视网膜的影响,数字视网膜图像的血管分割在这些疾病的诊断中具有重要的作用。本文提出了一种新型的尺寸不变全卷积神经网络(SIFCN)来解决视网膜血管自动分割问题。网络的输入数据是图像的patch和相应的逐像素标签。对输入数据进行连续的卷积层和池化层,使网络能够学习到抽象的特征来分割视网膜血管。我们的网络通过填充来保持每层数据的高度和宽度,并分配池化步幅,从而不需要空间信息维护和上采样。与基于像素的视网膜血管分割方法相比,我们的基于补丁的分割方法更有效,因为它可以在每个周期内预测补丁的所有像素。我们的重叠SIFCN方法准确率为0.9471,AUC为0.9682。我们的非重叠SIFCN是深度学习方法中效率最高的方法,每张图像耗时仅为3.68秒,而重叠SIFCN每张图像耗时为31.17秒。
{"title":"Size-Invariant Fully Convolutional Neural Network for vessel segmentation of digital retinal images","authors":"Yuan-sheng Luo, Hong Cheng, Lu Yang","doi":"10.1109/APSIPA.2016.7820677","DOIUrl":"https://doi.org/10.1109/APSIPA.2016.7820677","url":null,"abstract":"Vessel segmentation of digital retinal images plays an important role in diagnosis of diseases such as diabetics, hypertension and retinopathy of prematurity due to these diseases impact the retina. In this paper, a novel Size-Invariant Fully Convolutional Neural Network (SIFCN) is proposed to address the automatic retinal vessel segmentation problems. The input data of the network is the patches of images and the corresponding pixel-wise labels. A consecutive convolution layers and pooling layers follow the input data, so that the network can learn the abstract features to segment retinal vessel. Our network is designed to hold the height and width of data of each layer with padding and assign pooling stride so that the spatial information maintain and up-sample is not required. Compared with the pixel-wise retinal vessel segmentation approaches, our patch-wise segmentation is much more efficient since in each cycle it can predict all the pixels of the patch. Our overlapped SIFCN approach achieves accuracy of 0.9471, with the AUC of 0.9682. And our non-overlap SIFCN is the most efficient approach among the deep learning approaches, costing only 3.68 seconds per image, and the overlapped SIFCN costs 31.17 seconds per image.","PeriodicalId":409448,"journal":{"name":"2016 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA)","volume":"05 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130364872","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 12
Saliency detection using secondary quantization in DCT domain 基于二次量化的DCT域显著性检测
Xinyu Shen, Chunyu Lin, Yao Zhao, Hongyun Lin, Meiqin Liu
Saliency detection as an image preprocessing has been widely used in many applications such as image segmentation. Since most images stored in DCT domain, we propose an effective saliency detection algorithm, which is mainly based on DCT and secondary quantization. Firstly, the DC coefficient and the first five AC coefficients are used to get the color saliency map. Then, through secondary quantization of a JPEG image, we can obtain the difference of the original image and the quantified image, from which we can get the texture saliency map. Next, considering the center bias theory, the center region is easier to catch people's attention. And then the band-pass filter is used to simulate the behavior that the human visual system detects the salient region. Finally, the final saliency map is generated based on these two maps and two priorities. Experimental results on two datasets show that the proposed method can accurately detect the saliency regions and outperformed existing methods.
显著性检测作为一种图像预处理技术,已广泛应用于图像分割等领域。针对大多数图像存储在DCT域的特点,提出了一种有效的基于DCT和二次量化的显著性检测算法。首先,利用直流系数和前5个AC系数得到颜色显著性图;然后,通过对JPEG图像进行二次量化,得到原始图像与量化后图像的差值,从而得到纹理显著性图。其次,考虑到中心偏置理论,中心区域更容易引起人们的注意。然后利用带通滤波器模拟人类视觉系统检测显著区域的行为。最后,根据这两个映射和两个优先级生成最终的显著性映射。在两个数据集上的实验结果表明,该方法能够准确地检测出显著区域,优于现有方法。
{"title":"Saliency detection using secondary quantization in DCT domain","authors":"Xinyu Shen, Chunyu Lin, Yao Zhao, Hongyun Lin, Meiqin Liu","doi":"10.1109/APSIPA.2016.7820877","DOIUrl":"https://doi.org/10.1109/APSIPA.2016.7820877","url":null,"abstract":"Saliency detection as an image preprocessing has been widely used in many applications such as image segmentation. Since most images stored in DCT domain, we propose an effective saliency detection algorithm, which is mainly based on DCT and secondary quantization. Firstly, the DC coefficient and the first five AC coefficients are used to get the color saliency map. Then, through secondary quantization of a JPEG image, we can obtain the difference of the original image and the quantified image, from which we can get the texture saliency map. Next, considering the center bias theory, the center region is easier to catch people's attention. And then the band-pass filter is used to simulate the behavior that the human visual system detects the salient region. Finally, the final saliency map is generated based on these two maps and two priorities. Experimental results on two datasets show that the proposed method can accurately detect the saliency regions and outperformed existing methods.","PeriodicalId":409448,"journal":{"name":"2016 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA)","volume":"26 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130666618","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Recognition of low-resolution face images using sparse coding of local features 基于局部特征稀疏编码的低分辨率人脸图像识别
M. S. Shakeel, K. Lam
In this paper, we propose a new approach for recognition of low-resolution face images by using sparse coding of local features. The proposed algorithm extracts Gabor features from a low-resolution gallery image and a query image at different scales and orientations, then projects the features separately into a new low-dimensional feature space using sparse coding that preserves the sparse structure of the local features. To determine the similarity between the projected features, a coefficient vector is estimated by using linear regression that determines the relationship between the projected gallery and query features. On the basis of this coefficient vector, residual values will be computed to classify the images. To validate our proposed method, experiments were performed using three databases (ORL, Extended-Yale B, and CAS-PEAL-R1), which contain images with different facial expressions and lighting conditions. Experimental results show that our method outperforms various classical and state-of-the-art face recognition methods.
本文提出了一种基于局部特征稀疏编码的低分辨率人脸图像识别新方法。该算法从不同尺度和方向的低分辨率图库图像和查询图像中提取Gabor特征,然后使用稀疏编码将特征分别投影到新的低维特征空间中,保留了局部特征的稀疏结构。为了确定投影特征之间的相似性,使用线性回归来估计系数向量,该系数向量确定投影图库和查询特征之间的关系。在此系数向量的基础上,计算残差值对图像进行分类。为了验证我们提出的方法,使用三个数据库(ORL, Extended-Yale B和cas - pearl - r1)进行了实验,这些数据库包含不同面部表情和光照条件的图像。实验结果表明,该方法优于各种经典和先进的人脸识别方法。
{"title":"Recognition of low-resolution face images using sparse coding of local features","authors":"M. S. Shakeel, K. Lam","doi":"10.1109/APSIPA.2016.7820829","DOIUrl":"https://doi.org/10.1109/APSIPA.2016.7820829","url":null,"abstract":"In this paper, we propose a new approach for recognition of low-resolution face images by using sparse coding of local features. The proposed algorithm extracts Gabor features from a low-resolution gallery image and a query image at different scales and orientations, then projects the features separately into a new low-dimensional feature space using sparse coding that preserves the sparse structure of the local features. To determine the similarity between the projected features, a coefficient vector is estimated by using linear regression that determines the relationship between the projected gallery and query features. On the basis of this coefficient vector, residual values will be computed to classify the images. To validate our proposed method, experiments were performed using three databases (ORL, Extended-Yale B, and CAS-PEAL-R1), which contain images with different facial expressions and lighting conditions. Experimental results show that our method outperforms various classical and state-of-the-art face recognition methods.","PeriodicalId":409448,"journal":{"name":"2016 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130683806","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Consideration on performance improvement of shadow and reflection removal based on GMM 基于GMM的阴影和反射去除性能改进的思考
K. Nishikawa, Yoshihiro Yamashita, Toru Yamaguchi, T. Nishitani
Wearable devices are expected to provide a ubiquitous network connection in the near future. In this paper, we consider systems which uses human finger gestures as an input device. To assure accurate input characteristics, the shape of arm and fingers should be captured clearly, and for that purpose we consider using the Gaussian mixture model (GMM) foreground segmentation. It is known that shadow or reflection in the frame image affects the performance of GMM foreground segmentation. A low computational shadow or reflection removal methods are proposed [1]-[3] which are suitable to be implemented in wearable devices. Although the methods improve the foreground segmentation performance, the results depend on the characteristics of the video. In this paper, we consider improving the performance of the methods by modifying the equation for deciding the shadow region. Through the computer simulations, we show the effectiveness of the proposed method.
可穿戴设备有望在不久的将来提供无处不在的网络连接。在本文中,我们考虑使用人类手指手势作为输入设备的系统。为了保证准确的输入特征,手臂和手指的形状应该被清晰地捕获,为此我们考虑使用高斯混合模型(GMM)前景分割。已知帧图像中的阴影或反射会影响GMM前景分割的性能。提出了一种适合在可穿戴设备中实现的低计算阴影或反射去除方法[1]-[3]。虽然这些方法提高了前景分割的性能,但结果取决于视频的特性。在本文中,我们考虑通过修改确定阴影区域的方程来提高方法的性能。通过计算机仿真,验证了该方法的有效性。
{"title":"Consideration on performance improvement of shadow and reflection removal based on GMM","authors":"K. Nishikawa, Yoshihiro Yamashita, Toru Yamaguchi, T. Nishitani","doi":"10.1109/APSIPA.2016.7820902","DOIUrl":"https://doi.org/10.1109/APSIPA.2016.7820902","url":null,"abstract":"Wearable devices are expected to provide a ubiquitous network connection in the near future. In this paper, we consider systems which uses human finger gestures as an input device. To assure accurate input characteristics, the shape of arm and fingers should be captured clearly, and for that purpose we consider using the Gaussian mixture model (GMM) foreground segmentation. It is known that shadow or reflection in the frame image affects the performance of GMM foreground segmentation. A low computational shadow or reflection removal methods are proposed [1]-[3] which are suitable to be implemented in wearable devices. Although the methods improve the foreground segmentation performance, the results depend on the characteristics of the video. In this paper, we consider improving the performance of the methods by modifying the equation for deciding the shadow region. Through the computer simulations, we show the effectiveness of the proposed method.","PeriodicalId":409448,"journal":{"name":"2016 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA)","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116982543","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
A facial expression model with generative albedo texture 具有生成反照率纹理的面部表情模型
Songnan Li, Fanzi Wu, Tianhao Zhao, Ran Shi, K. Ngan
A facial expression model (FEM) is developed which can synthesize various face shapes and albedo textures. The face shape varies with individuals and expressions. FEM synthesizes these shape variations by using a bilinear face model built from the Face Warehouse Database. On the other hand, the generative albedo texture is directly extracted from a neutral face model — the Basel Face Model. In this paper, we elaborate the model construction process and demonstrate its application in face reconstruction and expression tracking.
建立了一种能够综合各种面部形状和反照率纹理的面部表情模型。脸型因个人和表情而异。有限元法利用人脸仓库数据库建立的双线性人脸模型综合这些形状变化。另一方面,生成反照率纹理是直接从中性人脸模型-巴塞尔人脸模型中提取的。本文详细阐述了模型的构建过程,并演示了其在人脸重建和表情跟踪中的应用。
{"title":"A facial expression model with generative albedo texture","authors":"Songnan Li, Fanzi Wu, Tianhao Zhao, Ran Shi, K. Ngan","doi":"10.1109/APSIPA.2016.7820866","DOIUrl":"https://doi.org/10.1109/APSIPA.2016.7820866","url":null,"abstract":"A facial expression model (FEM) is developed which can synthesize various face shapes and albedo textures. The face shape varies with individuals and expressions. FEM synthesizes these shape variations by using a bilinear face model built from the Face Warehouse Database. On the other hand, the generative albedo texture is directly extracted from a neutral face model — the Basel Face Model. In this paper, we elaborate the model construction process and demonstrate its application in face reconstruction and expression tracking.","PeriodicalId":409448,"journal":{"name":"2016 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131958865","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
2016 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1