首页 > 最新文献

2018 IEEE 3rd International Conference on Image, Vision and Computing (ICIVC)最新文献

英文 中文
Research on the Counting Algorithm of Bundled Steel Bars Based on the Features Matching of Connected Regions 基于连通区域特征匹配的捆扎钢筋计数算法研究
Pub Date : 2018-06-27 DOI: 10.1109/ICIVC.2018.8492784
Xing Yan, X. Chen
Aiming at the problem that tight arrangement of bundled steel bar, irregular shape of bar end surfaces, adhesions and in the automatic counting prone to multi-count and less-count, a fast and accurate counting method based on single-multi-classification of the connected regions' feature matching is proposed. Firstly, the image of the steel bars' end surface is segmented, morphological and other preprocessing to remove most of the adhesion, and then extracts the features of handled binary image including the area, diameter, center of gravity, shape factor of connected region, according to the area characteristics, the target single - Multi - Classification of the bar image target is classified, and the area feature matching of the single steel bar is counted quickly too, and according to the characteristics of the center of gravity to identify the identification of steel, multiple steel bars establish the template and combine with area, form and other factor to matching counting, so as to achieve the purpose of high efficiency and accurate counting.
针对捆扎钢筋排列紧密、钢筋端面形状不规则、粘连以及自动计数中容易出现多计数和少计数的问题,提出了一种基于连通区域特征匹配的单-多分类快速准确计数方法。首先对钢筋端面图像进行分割、形态学等预处理,去除大部分粘连,然后提取处理后的二值图像特征,包括连通区域的面积、直径、重心、形状因子等,根据面积特征对钢筋图像目标进行单-多分类,并对单个钢筋的面积特征匹配进行快速计数;并根据钢筋的重心识别特征,将多根钢筋建立模板,并结合面积、形状等因素进行匹配计数,从而达到高效准确计数的目的。
{"title":"Research on the Counting Algorithm of Bundled Steel Bars Based on the Features Matching of Connected Regions","authors":"Xing Yan, X. Chen","doi":"10.1109/ICIVC.2018.8492784","DOIUrl":"https://doi.org/10.1109/ICIVC.2018.8492784","url":null,"abstract":"Aiming at the problem that tight arrangement of bundled steel bar, irregular shape of bar end surfaces, adhesions and in the automatic counting prone to multi-count and less-count, a fast and accurate counting method based on single-multi-classification of the connected regions' feature matching is proposed. Firstly, the image of the steel bars' end surface is segmented, morphological and other preprocessing to remove most of the adhesion, and then extracts the features of handled binary image including the area, diameter, center of gravity, shape factor of connected region, according to the area characteristics, the target single - Multi - Classification of the bar image target is classified, and the area feature matching of the single steel bar is counted quickly too, and according to the characteristics of the center of gravity to identify the identification of steel, multiple steel bars establish the template and combine with area, form and other factor to matching counting, so as to achieve the purpose of high efficiency and accurate counting.","PeriodicalId":173981,"journal":{"name":"2018 IEEE 3rd International Conference on Image, Vision and Computing (ICIVC)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128799721","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
An Investigation of Skeleton-Based Optical Flow-Guided Features for 3D Action Recognition Using a Multi-Stream CNN Model 基于多流CNN模型的骨骼光流引导特征在三维动作识别中的研究
Pub Date : 2018-06-27 DOI: 10.1109/ICIVC.2018.8492894
J. Ren, N. Reyes, A. Barczak, C. Scogings, M. Liu
Deep learning-based techniques have recently been found significantly effective for handling skeleton-based action recognition tasks. It was observed that modeling the spatiotemporal variations is the key to effective skeleton-based action recognition approaches. This work proposes an easy and yet effective method for encoding different geometric relational features into static color texture images. Collectively, we refer to these features as skeletal optical flow-guided features. The temporal variations of different features are converted into the color variations of their corresponding images. Then, a multi-stream CNN model is employed to pick up the discriminating patterns that exist in the converted images for subsequent classification. Experimental results demonstrate that our proposed geometric relational features and framework can achieve competitive performances on both MSR Action 3D and NTU RGB+D datasets.
基于深度学习的技术最近被发现在处理基于骨架的动作识别任务方面非常有效。研究发现,对骨骼时空变化进行建模是有效的骨骼动作识别方法的关键。本文提出了一种简单有效的方法,将不同的几何关系特征编码到静态彩色纹理图像中。总的来说,我们把这些特征称为骨骼光流引导特征。将不同特征的时间变化转化为相应图像的颜色变化。然后,采用多流CNN模型提取转换后图像中存在的判别模式,用于后续分类。实验结果表明,我们提出的几何关系特征和框架在MSR Action 3D和NTU RGB+D数据集上都能取得具有竞争力的性能。
{"title":"An Investigation of Skeleton-Based Optical Flow-Guided Features for 3D Action Recognition Using a Multi-Stream CNN Model","authors":"J. Ren, N. Reyes, A. Barczak, C. Scogings, M. Liu","doi":"10.1109/ICIVC.2018.8492894","DOIUrl":"https://doi.org/10.1109/ICIVC.2018.8492894","url":null,"abstract":"Deep learning-based techniques have recently been found significantly effective for handling skeleton-based action recognition tasks. It was observed that modeling the spatiotemporal variations is the key to effective skeleton-based action recognition approaches. This work proposes an easy and yet effective method for encoding different geometric relational features into static color texture images. Collectively, we refer to these features as skeletal optical flow-guided features. The temporal variations of different features are converted into the color variations of their corresponding images. Then, a multi-stream CNN model is employed to pick up the discriminating patterns that exist in the converted images for subsequent classification. Experimental results demonstrate that our proposed geometric relational features and framework can achieve competitive performances on both MSR Action 3D and NTU RGB+D datasets.","PeriodicalId":173981,"journal":{"name":"2018 IEEE 3rd International Conference on Image, Vision and Computing (ICIVC)","volume":"174 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114621320","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
High Accuracy Smartphone Video Calibration for Human Foot Surface Mapping 高精度智能手机视频校准人体足部表面映射
Pub Date : 2018-06-01 DOI: 10.1109/ICIVC.2018.8492792
Ali A. Al-kharaz, A. Chong
Although the digital camera is readily available and the price is decreasing, many users still consider it an expensive device that can be dispensed with by using a smart phone camera. However, both the digital camera and the smartphone need to be calibrated to extract three dimensional (3D) space information from (2D) and to obtain accurate results. This study used close range photogrammetry to calibrate two high resolution digital cameras and a Samsung Galaxy smartphone to find whether any one of them give high accuracy 3D coordinates of the retro-reflective targets that were determined using the self-calibration bundle adjustment method in two phases. The first phase is during walking when 3 trials are conducted. The same three cameras are used for each trial. The second phase is during standing when one trial is conducted. Each of the camera types is placed in front of the platform. The results showed that arguably, the Samsung Galaxy S6 camera is most significant than other cameras in term of accuracy. In addition, this study provides information on how to calibrate one board from other board that has already been calibrated.
虽然数码相机很容易买到,而且价格正在下降,但许多用户仍然认为它是一个昂贵的设备,可以用智能手机相机来代替。然而,数码相机和智能手机都需要校准,以从(2D)中提取三维(3D)空间信息,并获得准确的结果。本研究采用近景摄影测量法对两台高分辨率数码相机和一部三星Galaxy智能手机进行了标定,以确定是否有一台相机能给出两阶段自标定束平差法确定的反反射目标的高精度三维坐标。第一阶段是在行走时进行3次试验。每次试验使用相同的三个摄像机。第二阶段是在听证期间,此时进行一次审判。每种类型的摄像机都放置在平台的前面。结果表明,可以说,三星Galaxy S6相机在精度方面比其他相机最重要。此外,本研究还提供了如何从已经校准的另一块板校准一块板的信息。
{"title":"High Accuracy Smartphone Video Calibration for Human Foot Surface Mapping","authors":"Ali A. Al-kharaz, A. Chong","doi":"10.1109/ICIVC.2018.8492792","DOIUrl":"https://doi.org/10.1109/ICIVC.2018.8492792","url":null,"abstract":"Although the digital camera is readily available and the price is decreasing, many users still consider it an expensive device that can be dispensed with by using a smart phone camera. However, both the digital camera and the smartphone need to be calibrated to extract three dimensional (3D) space information from (2D) and to obtain accurate results. This study used close range photogrammetry to calibrate two high resolution digital cameras and a Samsung Galaxy smartphone to find whether any one of them give high accuracy 3D coordinates of the retro-reflective targets that were determined using the self-calibration bundle adjustment method in two phases. The first phase is during walking when 3 trials are conducted. The same three cameras are used for each trial. The second phase is during standing when one trial is conducted. Each of the camera types is placed in front of the platform. The results showed that arguably, the Samsung Galaxy S6 camera is most significant than other cameras in term of accuracy. In addition, this study provides information on how to calibrate one board from other board that has already been calibrated.","PeriodicalId":173981,"journal":{"name":"2018 IEEE 3rd International Conference on Image, Vision and Computing (ICIVC)","volume":"31 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125122108","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Surface Defect Detection Based on Gradient LBP 基于梯度LBP的表面缺陷检测
Pub Date : 2018-06-01 DOI: 10.1109/ICIVC.2018.8492798
Xiaojing Liu, Feng Xue, Lu Teng
The LBP histogram obtained based on the local binary pattern (LBP) method usually has a higher dimension, and not conducive to calculation. The LBP method adopts the gray difference value between single points as the LBP output value, which is not robust to noise and illumination. Therefore, this paper improves the traditional LBP method and proposes a surface defect detection method based on gradient local binary pattern (GLBP), which uses image sub-blocks to reduce the dimensionality of the LBP data matrix. The method adopts weighted binary output values in eight directions within the neighborhood to indicate local gray changes, which suppresses the effects of light and noise on the detection results. Experiments show that the method can determine the defect location well and provide good feature information for subsequent defect classification.
基于局部二值模式(LBP)方法得到的LBP直方图通常维数较高,不利于计算。LBP方法采用单点灰度差值作为LBP输出值,对噪声和光照的鲁棒性较差。因此,本文对传统的LBP方法进行改进,提出了一种基于梯度局部二值模式(GLBP)的表面缺陷检测方法,该方法利用图像子块对LBP数据矩阵进行降维处理。该方法采用邻域内八个方向的加权二值输出值来表示局部灰度变化,抑制了光和噪声对检测结果的影响。实验表明,该方法可以很好地确定缺陷的位置,为后续的缺陷分类提供良好的特征信息。
{"title":"Surface Defect Detection Based on Gradient LBP","authors":"Xiaojing Liu, Feng Xue, Lu Teng","doi":"10.1109/ICIVC.2018.8492798","DOIUrl":"https://doi.org/10.1109/ICIVC.2018.8492798","url":null,"abstract":"The LBP histogram obtained based on the local binary pattern (LBP) method usually has a higher dimension, and not conducive to calculation. The LBP method adopts the gray difference value between single points as the LBP output value, which is not robust to noise and illumination. Therefore, this paper improves the traditional LBP method and proposes a surface defect detection method based on gradient local binary pattern (GLBP), which uses image sub-blocks to reduce the dimensionality of the LBP data matrix. The method adopts weighted binary output values in eight directions within the neighborhood to indicate local gray changes, which suppresses the effects of light and noise on the detection results. Experiments show that the method can determine the defect location well and provide good feature information for subsequent defect classification.","PeriodicalId":173981,"journal":{"name":"2018 IEEE 3rd International Conference on Image, Vision and Computing (ICIVC)","volume":"2016 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115551158","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 19
Fire Smoke Detection Based on Contextual Object Detection 基于上下文对象检测的火灾烟雾检测
Pub Date : 2018-06-01 DOI: 10.1109/ICIVC.2018.8492823
Xuan Zhaa, Hang Ji, Deng-yin Zhang, H. Bao
Smoke detection based on automatic visual system has been applied to fire alarm in open spaces where traditional smoke detection system is not suitable for it. However, detecting the course of smoke posed great challenges for both systems. To address this problem, we propose a new method that combines context-aware framework with automatic visual smoke detection. The strategy is evaluated on dataset and the results demonstrate the effectiveness of the proposed method.
基于自动视觉系统的烟雾探测已被应用于传统烟雾探测系统不适合的开放空间火灾报警中。然而,探测烟雾的过程对这两个系统都提出了巨大的挑战。为了解决这个问题,我们提出了一种将上下文感知框架与自动视觉烟雾检测相结合的新方法。在数据集上对该策略进行了评估,结果证明了该方法的有效性。
{"title":"Fire Smoke Detection Based on Contextual Object Detection","authors":"Xuan Zhaa, Hang Ji, Deng-yin Zhang, H. Bao","doi":"10.1109/ICIVC.2018.8492823","DOIUrl":"https://doi.org/10.1109/ICIVC.2018.8492823","url":null,"abstract":"Smoke detection based on automatic visual system has been applied to fire alarm in open spaces where traditional smoke detection system is not suitable for it. However, detecting the course of smoke posed great challenges for both systems. To address this problem, we propose a new method that combines context-aware framework with automatic visual smoke detection. The strategy is evaluated on dataset and the results demonstrate the effectiveness of the proposed method.","PeriodicalId":173981,"journal":{"name":"2018 IEEE 3rd International Conference on Image, Vision and Computing (ICIVC)","volume":"365 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122315086","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Adaptive Multi-Level Saliency Network in 3D Generation 三维生成中的自适应多层次显著性网络
Pub Date : 2018-06-01 DOI: 10.1109/ICIVC.2018.8492765
Zhongliang Tang
Many CNNs with encoder-decoder structure are widely used in supervised 3D voxel generation. However, their convolutional encoders are usually too simple which causes some local features to degrade during convolution, so it is difficult to extract a good feature representation from an input image by using a simple encoder. Some CNNs apply skip-connection layer for the encoder to reduce the degradation, but general skip-connection layer such as residual layer is not good enough especially when the quantity of convolutional layers in the encoder is relatively small. In this paper, we propose a novel structure called adaptive multi-level saliency network (AMSN) to reduce the degradation of local features. The major innovations of AMSN are multi-level saliency convolution kernels (MSCK) and saliency fusion layer. Different from the kernels used in general skip-connection layer, MSCK are adaptively learned from multi-level salient feature maps rather than initialized randomly. The salient feature maps are sampled from multiple layers in the encoder. MSCK can acquire multi-level features more easily so that we utilize MSCK to acquire local features from low-level layer before the degradation. After that, the acquired local features are fused back into encoder through a saliency fusion layer to reduce the degradation. We evaluated our approach on the ShapeNet and ModelNet40 dataset. The results indicate that our AMSN performs better than related works.
许多具有编码器-解码器结构的cnn被广泛应用于有监督的三维体素生成。然而,它们的卷积编码器通常过于简单,导致卷积过程中一些局部特征的退化,因此使用简单的编码器很难从输入图像中提取出良好的特征表示。有些cnn在编码器上采用跳接层来减少退化,但是一般的跳接层如残差层的效果不够好,特别是在编码器中卷积层数量比较少的情况下。本文提出了一种新的自适应多层次显著性网络(AMSN)结构来减少局部特征的退化。AMSN的主要创新点是多层显著性卷积核(MSCK)和显著性融合层。与一般跳过连接层使用的核不同,MSCK是从多级显著特征映射中自适应学习的,而不是随机初始化。显著特征映射从编码器的多个层中采样。MSCK可以更容易地获取多层特征,因此我们利用MSCK在退化前从低层获取局部特征。然后,通过显著性融合层将获取的局部特征融合回编码器中,以减少退化。我们在ShapeNet和ModelNet40数据集上评估了我们的方法。结果表明,我们的AMSN性能优于相关工作。
{"title":"Adaptive Multi-Level Saliency Network in 3D Generation","authors":"Zhongliang Tang","doi":"10.1109/ICIVC.2018.8492765","DOIUrl":"https://doi.org/10.1109/ICIVC.2018.8492765","url":null,"abstract":"Many CNNs with encoder-decoder structure are widely used in supervised 3D voxel generation. However, their convolutional encoders are usually too simple which causes some local features to degrade during convolution, so it is difficult to extract a good feature representation from an input image by using a simple encoder. Some CNNs apply skip-connection layer for the encoder to reduce the degradation, but general skip-connection layer such as residual layer is not good enough especially when the quantity of convolutional layers in the encoder is relatively small. In this paper, we propose a novel structure called adaptive multi-level saliency network (AMSN) to reduce the degradation of local features. The major innovations of AMSN are multi-level saliency convolution kernels (MSCK) and saliency fusion layer. Different from the kernels used in general skip-connection layer, MSCK are adaptively learned from multi-level salient feature maps rather than initialized randomly. The salient feature maps are sampled from multiple layers in the encoder. MSCK can acquire multi-level features more easily so that we utilize MSCK to acquire local features from low-level layer before the degradation. After that, the acquired local features are fused back into encoder through a saliency fusion layer to reduce the degradation. We evaluated our approach on the ShapeNet and ModelNet40 dataset. The results indicate that our AMSN performs better than related works.","PeriodicalId":173981,"journal":{"name":"2018 IEEE 3rd International Conference on Image, Vision and Computing (ICIVC)","volume":"127 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122902652","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Comparison of Several Hyperspectral Image Fusion Methods for Superresolution 几种超分辨率高光谱图像融合方法的比较
Pub Date : 2018-06-01 DOI: 10.1109/ICIVC.2018.8492889
Hongwen Lin, Jian Chen
Hyperspectral image applications have been explored in various areas, but they are often suffered from coarser spatial resolutions. In recent years, many hyperspectral image fusion approaches which merge hyperspectral image with multi-spectral or panchromatic one have been presented to improve the spatial resolution of hyperspectral image. In this paper, we compared four state-of-the-art hyperspectral fusion methods, namely coupled nonnegative matrix factorization (CNMF) method, sparse matrix factorization (SPMF) method, hyperspectral Image superresolution (HySure) method and sparse representation (SPRE) method. The main idea of each method is depicted briefly, five statistical assessment parameters, namely cross correlation (CC), root-mean-square error (RMSE), spectral angle mapper (SAM), universal image quality index (UIQI), and relative dimensionless global error in synthesis (ERGAS) are adopted to comparatively analyze the fusion results. The experimental results show that the effect of method based on sparse representation is superior to the others one.
高光谱图像的应用已经在各个领域进行了探索,但它们往往受到较粗的空间分辨率的影响。近年来,为了提高高光谱图像的空间分辨率,提出了许多将高光谱图像与多光谱或全色图像合并的高光谱图像融合方法。本文比较了四种最先进的高光谱融合方法,即耦合非负矩阵分解(CNMF)方法、稀疏矩阵分解(SPMF)方法、高光谱图像超分辨率(HySure)方法和稀疏表示(SPRE)方法。简要介绍了每种方法的主要思想,并采用交叉相关(CC)、均方根误差(RMSE)、光谱角映射器(SAM)、通用图像质量指数(UIQI)和相对无量纲全局合成误差(ERGAS) 5个统计评价参数对融合结果进行对比分析。实验结果表明,基于稀疏表示的方法效果优于其他方法。
{"title":"Comparison of Several Hyperspectral Image Fusion Methods for Superresolution","authors":"Hongwen Lin, Jian Chen","doi":"10.1109/ICIVC.2018.8492889","DOIUrl":"https://doi.org/10.1109/ICIVC.2018.8492889","url":null,"abstract":"Hyperspectral image applications have been explored in various areas, but they are often suffered from coarser spatial resolutions. In recent years, many hyperspectral image fusion approaches which merge hyperspectral image with multi-spectral or panchromatic one have been presented to improve the spatial resolution of hyperspectral image. In this paper, we compared four state-of-the-art hyperspectral fusion methods, namely coupled nonnegative matrix factorization (CNMF) method, sparse matrix factorization (SPMF) method, hyperspectral Image superresolution (HySure) method and sparse representation (SPRE) method. The main idea of each method is depicted briefly, five statistical assessment parameters, namely cross correlation (CC), root-mean-square error (RMSE), spectral angle mapper (SAM), universal image quality index (UIQI), and relative dimensionless global error in synthesis (ERGAS) are adopted to comparatively analyze the fusion results. The experimental results show that the effect of method based on sparse representation is superior to the others one.","PeriodicalId":173981,"journal":{"name":"2018 IEEE 3rd International Conference on Image, Vision and Computing (ICIVC)","volume":"34 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128589318","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Iterative Stochastic Resonance Model for Visual Information Extraction from Noisy Environment 噪声环境下视觉信息提取的迭代随机共振模型
Pub Date : 2018-06-01 DOI: 10.1109/ICIVC.2018.8492835
Ling Wang, Peng Miao
Stochastic Resonance phenomenon brings us a new viewpoint of the relation between noise and information which considers the noise as an interacting factor with the information. In this paper, on the Stochastic Resonance phenomenon of neurons in the human visual system, we propose a new model called Iterative Stochastic Resonance, for the visual information extraction from noisy images. The algorithm introduces appropriate noise into the noisy image so that the signal and noise produce a synergistic effect, thereby increasing the energy of the useful signal. The model is then modeled on the characteristics of the human visual system and the results are iteratively computed several times. The model can give perfect denoising output for both simulated and real laser speckle contrast images which are both disturbed by strong noise. It is a new way to solve the problem of effective information extraction in medical noisy images.
随机共振现象使我们对噪声与信息的关系有了新的认识,认为噪声是与信息相互作用的因素。针对人类视觉系统中神经元的随机共振现象,提出了一种新的迭代随机共振模型,用于从噪声图像中提取视觉信息。该算法在有噪声的图像中引入适当的噪声,使信号和噪声产生协同效应,从而增加有用信号的能量。然后根据人类视觉系统的特点对模型进行建模,并对结果进行多次迭代计算。该模型对受强噪声干扰的仿真和真实激光散斑对比图像都能给出较好的去噪输出。它是解决医学噪声图像中有效信息提取问题的一种新方法。
{"title":"Iterative Stochastic Resonance Model for Visual Information Extraction from Noisy Environment","authors":"Ling Wang, Peng Miao","doi":"10.1109/ICIVC.2018.8492835","DOIUrl":"https://doi.org/10.1109/ICIVC.2018.8492835","url":null,"abstract":"Stochastic Resonance phenomenon brings us a new viewpoint of the relation between noise and information which considers the noise as an interacting factor with the information. In this paper, on the Stochastic Resonance phenomenon of neurons in the human visual system, we propose a new model called Iterative Stochastic Resonance, for the visual information extraction from noisy images. The algorithm introduces appropriate noise into the noisy image so that the signal and noise produce a synergistic effect, thereby increasing the energy of the useful signal. The model is then modeled on the characteristics of the human visual system and the results are iteratively computed several times. The model can give perfect denoising output for both simulated and real laser speckle contrast images which are both disturbed by strong noise. It is a new way to solve the problem of effective information extraction in medical noisy images.","PeriodicalId":173981,"journal":{"name":"2018 IEEE 3rd International Conference on Image, Vision and Computing (ICIVC)","volume":"37 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124778880","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Color Feature Unified-Based Approach for Visual Fixation 基于颜色特征统一的视觉注视方法
Pub Date : 2018-06-01 DOI: 10.1109/ICIVC.2018.8492890
Zhaoxia Xie
Human visual attention can be deal with complex scenes in real time effortlessly and efficiently and detect the most interesting regions quickly. Based on the characteristic of human visual attention, one more comprehensive computation framework is proposed which fully takes the advantage of color contrast to obtain visual fixations of natural color images. Firstly, the color space conversion strategy is employed. The RGB color images are converted into the HSV color space and Lab color space respectively. Then, the superpixels generation algorithm is utilized to segment natural images in the HSV color space and in the Lab color space. Next, color feature-contrast in the two color space is respectively implemented and the corresponding single visual fixation is obtained. Finally, the color feature-fused strategy is adopted in order to get the final visual fixation. Experimental results show that our proposed framework can effectively improve the effect of visual fixations compared with a single color space for the natural color images. Moreover, the full resolution visual fixations can be obtained by employing the proposed framework in this paper compared to the context-aware approach. Meanwhile, these experimental results also clearly demonstrate that the proposed model for saliency estimation is effective.
人类的视觉注意可以毫不费力、高效地实时处理复杂的场景,并快速检测出最有趣的区域。根据人类视觉注意的特点,提出了一种更全面的计算框架,充分利用色彩对比的优势,获得自然彩色图像的视觉注视力。首先,采用色彩空间转换策略;将RGB彩色图像分别转换为HSV色彩空间和Lab色彩空间。然后,利用超像素生成算法在HSV色彩空间和Lab色彩空间对自然图像进行分割。然后分别在两个色彩空间中进行色彩特征对比,得到相应的单视觉固定;最后,采用颜色特征融合策略,得到最终的视觉固定。实验结果表明,与单一色彩空间相比,该框架可以有效地提高自然彩色图像的视觉注视效果。此外,与上下文感知方法相比,采用本文提出的框架可以获得全分辨率的视觉注视。同时,这些实验结果也清楚地证明了所提出的显著性估计模型是有效的。
{"title":"Color Feature Unified-Based Approach for Visual Fixation","authors":"Zhaoxia Xie","doi":"10.1109/ICIVC.2018.8492890","DOIUrl":"https://doi.org/10.1109/ICIVC.2018.8492890","url":null,"abstract":"Human visual attention can be deal with complex scenes in real time effortlessly and efficiently and detect the most interesting regions quickly. Based on the characteristic of human visual attention, one more comprehensive computation framework is proposed which fully takes the advantage of color contrast to obtain visual fixations of natural color images. Firstly, the color space conversion strategy is employed. The RGB color images are converted into the HSV color space and Lab color space respectively. Then, the superpixels generation algorithm is utilized to segment natural images in the HSV color space and in the Lab color space. Next, color feature-contrast in the two color space is respectively implemented and the corresponding single visual fixation is obtained. Finally, the color feature-fused strategy is adopted in order to get the final visual fixation. Experimental results show that our proposed framework can effectively improve the effect of visual fixations compared with a single color space for the natural color images. Moreover, the full resolution visual fixations can be obtained by employing the proposed framework in this paper compared to the context-aware approach. Meanwhile, these experimental results also clearly demonstrate that the proposed model for saliency estimation is effective.","PeriodicalId":173981,"journal":{"name":"2018 IEEE 3rd International Conference on Image, Vision and Computing (ICIVC)","volume":"31 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121295264","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Insect Sound Recognition Based on Convolutional Neural Network 基于卷积神经网络的昆虫声音识别
Pub Date : 2018-06-01 DOI: 10.1109/ICIVC.2018.8492871
X. Dong, Ning Yan, Ying Wei
A novel insect sound recognition system using enhanced spectrogram and convolutional neural network is proposed. Contrast-limit adaptive histogram equalization (CLAHE) is adopted to enhance R-space spectrogram. Traditionally, artificial feature extraction is an essential step of classification, introducing extra noise caused by subjectivity of individual researchers. In this paper, we construct a convolutional neural network (CNN) as classifier, extracting deep feature by machine learning. Mel-Frequency Cepstral Coefficient (MFCC) and chromatic spectrogram have been compared with enhanced R-space spectrogram as feature image. Eventually, 97.8723 % accuracy rate is achieved among 47 types of insect sound from USDA library.
提出了一种基于增强谱图和卷积神经网络的昆虫声音识别系统。采用对比度限制自适应直方图均衡化(CLAHE)增强r空间谱图。传统上,人工特征提取是分类的重要步骤,它引入了由于研究人员个人主观性而产生的额外噪声。在本文中,我们构建了卷积神经网络(CNN)作为分类器,通过机器学习提取深度特征。将Mel-Frequency倒谱系数(MFCC)和彩色谱图与增强r空间谱图作为特征图像进行了比较。最终,在美国农业部文库的47种昆虫声音中,准确率达到97.8723%。
{"title":"Insect Sound Recognition Based on Convolutional Neural Network","authors":"X. Dong, Ning Yan, Ying Wei","doi":"10.1109/ICIVC.2018.8492871","DOIUrl":"https://doi.org/10.1109/ICIVC.2018.8492871","url":null,"abstract":"A novel insect sound recognition system using enhanced spectrogram and convolutional neural network is proposed. Contrast-limit adaptive histogram equalization (CLAHE) is adopted to enhance R-space spectrogram. Traditionally, artificial feature extraction is an essential step of classification, introducing extra noise caused by subjectivity of individual researchers. In this paper, we construct a convolutional neural network (CNN) as classifier, extracting deep feature by machine learning. Mel-Frequency Cepstral Coefficient (MFCC) and chromatic spectrogram have been compared with enhanced R-space spectrogram as feature image. Eventually, 97.8723 % accuracy rate is achieved among 47 types of insect sound from USDA library.","PeriodicalId":173981,"journal":{"name":"2018 IEEE 3rd International Conference on Image, Vision and Computing (ICIVC)","volume":"142 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114344497","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
期刊
2018 IEEE 3rd International Conference on Image, Vision and Computing (ICIVC)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1