首页 > 最新文献

2020 IEEE International Conference on Image Processing (ICIP)最新文献

英文 中文
Context-Aware Hierarchical Feature Attention Network For Multi-Scale Object Detection 面向多尺度目标检测的上下文感知分层特征关注网络
Pub Date : 2020-10-01 DOI: 10.1109/ICIP40778.2020.9190896
Xuelong Xu, Xiangfeng Luo, Liyan Ma
Multi-scale object detection involves classification and regression assignments of objects with variable scales from an image. How to extract discriminative features is a key point for multi-scale object detection. Recent detectors simply fuse pyramidal features extracted from ConvNets, which does not take full advantage of useful features and drop out redundant features. To address this problem, we propose Context-Aware Hierarchical Feature Attention Network (CHFANet) to focus on effective multi-scale feature extraction for object detection. Based on single shot multibox detector (SSD) framework, the CHFANet consists of two components: the context-aware feature extraction (CFE) module to capture rich multi-scale context features and the hierarchical feature fusion (HFF) module followed with the channel-wise attention model to generate deeply fused attentive features. On the Pascal VOC benchmark, our CHFANet can achieve 82.6% mAP. Extensive experiments demonstrate that the CHFANet outperforms a lot of state-of-the-art object detectors in accuracy without any bells and whistles.
多尺度目标检测涉及对图像中不同尺度的目标进行分类和回归赋值。如何提取判别特征是多尺度目标检测的关键。最近的检测器只是简单地融合从卷积神经网络中提取的金字塔特征,这没有充分利用有用的特征,并删除了冗余的特征。为了解决这一问题,我们提出了上下文感知分层特征注意网络(CHFANet),专注于有效的多尺度特征提取用于目标检测。CHFANet基于单镜头多盒检测器(SSD)框架,由上下文感知特征提取(CFE)模块和分层特征融合(HFF)模块组成,前者用于捕获丰富的多尺度上下文特征,后者用于生成深度融合的关注特征。在帕斯卡VOC基准上,我们的CHFANet可以达到82.6%的mAP。大量的实验表明,CHFANet在精度上超过了许多最先进的物体探测器,而没有任何花哨的东西。
{"title":"Context-Aware Hierarchical Feature Attention Network For Multi-Scale Object Detection","authors":"Xuelong Xu, Xiangfeng Luo, Liyan Ma","doi":"10.1109/ICIP40778.2020.9190896","DOIUrl":"https://doi.org/10.1109/ICIP40778.2020.9190896","url":null,"abstract":"Multi-scale object detection involves classification and regression assignments of objects with variable scales from an image. How to extract discriminative features is a key point for multi-scale object detection. Recent detectors simply fuse pyramidal features extracted from ConvNets, which does not take full advantage of useful features and drop out redundant features. To address this problem, we propose Context-Aware Hierarchical Feature Attention Network (CHFANet) to focus on effective multi-scale feature extraction for object detection. Based on single shot multibox detector (SSD) framework, the CHFANet consists of two components: the context-aware feature extraction (CFE) module to capture rich multi-scale context features and the hierarchical feature fusion (HFF) module followed with the channel-wise attention model to generate deeply fused attentive features. On the Pascal VOC benchmark, our CHFANet can achieve 82.6% mAP. Extensive experiments demonstrate that the CHFANet outperforms a lot of state-of-the-art object detectors in accuracy without any bells and whistles.","PeriodicalId":405734,"journal":{"name":"2020 IEEE International Conference on Image Processing (ICIP)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2020-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128212949","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
A Deep Learning Framework for 3D Surface Profiling of the Objects Using Digital Holographic Interferometry 基于数字全息干涉测量的物体三维表面轮廓深度学习框架
Pub Date : 2020-10-01 DOI: 10.1109/ICIP40778.2020.9190669
Krishna Sumanth Vengala, Rama Krishna Sai Subrahmanyam Gorthi
Phase reconstruction in Digital Holographic Interferometry (DHI) is widely employed for 3D deformation measurements of the object surfaces. The key challenge in phase reconstruction in DHI is in the estimation of the absolute phase from noisy reconstructed interference fringes. In this paper, we propose a novel efficient deep learning approach for the phase estimation from noisy interference fringes in DHI. The proposed approach takes noisy reconstructed interference fringes as input and estimates the 3D deformation field or the object surface profile as the output. The 3D deformation field measurement of the object is posed as the absolute phase estimation from the noisy wrapped phase, that can be obtained from the reconstructed interference fringes through arctan function. The proposed deep neural network is trained to predict the fringe-order through a fully convolutional semantic segmentation network, from the noisy wrapped phase. These predictions are improved by simultaneously minimizing the regression error between the true phase corresponding to the object deformation field and the estimated absolute phase considering the predicted fringe order. We compare our method with conventional methods as well as with the recent state-of-the-art deep learning phase unwrapping methods. The proposed method outperforms conventional approaches by a large margin, while we can observe significant improvement even with respect to recently proposed deep learning-based phase unwrapping methods, in the presence of noise as high as 0dB to -5dB.
数字全息干涉测量中的相位重建被广泛应用于物体表面的三维变形测量。相位重建的关键挑战是如何从噪声重建的干涉条纹中估计绝对相位。在本文中,我们提出了一种新的高效的深度学习方法,用于从噪声干扰条纹中进行相位估计。该方法以噪声重构干涉条纹作为输入,估计三维变形场或物体表面轮廓作为输出。物体的三维变形场测量是由干涉条纹重构后通过arctan函数得到的噪声包裹相位的绝对相位估计。该深度神经网络通过全卷积语义分割网络从噪声包裹阶段开始预测条纹阶数。考虑到预测的条纹阶数,通过同时最小化物体变形场对应的真相位与估计的绝对相位之间的回归误差来改进这些预测。我们将我们的方法与传统方法以及最近最先进的深度学习阶段展开方法进行了比较。所提出的方法在很大程度上优于传统方法,而我们可以观察到,即使在存在高达0dB至-5dB的噪声的情况下,与最近提出的基于深度学习的相位展开方法相比,也有显著的改进。
{"title":"A Deep Learning Framework for 3D Surface Profiling of the Objects Using Digital Holographic Interferometry","authors":"Krishna Sumanth Vengala, Rama Krishna Sai Subrahmanyam Gorthi","doi":"10.1109/ICIP40778.2020.9190669","DOIUrl":"https://doi.org/10.1109/ICIP40778.2020.9190669","url":null,"abstract":"Phase reconstruction in Digital Holographic Interferometry (DHI) is widely employed for 3D deformation measurements of the object surfaces. The key challenge in phase reconstruction in DHI is in the estimation of the absolute phase from noisy reconstructed interference fringes. In this paper, we propose a novel efficient deep learning approach for the phase estimation from noisy interference fringes in DHI. The proposed approach takes noisy reconstructed interference fringes as input and estimates the 3D deformation field or the object surface profile as the output. The 3D deformation field measurement of the object is posed as the absolute phase estimation from the noisy wrapped phase, that can be obtained from the reconstructed interference fringes through arctan function. The proposed deep neural network is trained to predict the fringe-order through a fully convolutional semantic segmentation network, from the noisy wrapped phase. These predictions are improved by simultaneously minimizing the regression error between the true phase corresponding to the object deformation field and the estimated absolute phase considering the predicted fringe order. We compare our method with conventional methods as well as with the recent state-of-the-art deep learning phase unwrapping methods. The proposed method outperforms conventional approaches by a large margin, while we can observe significant improvement even with respect to recently proposed deep learning-based phase unwrapping methods, in the presence of noise as high as 0dB to -5dB.","PeriodicalId":405734,"journal":{"name":"2020 IEEE International Conference on Image Processing (ICIP)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2020-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127341272","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Representation Reconstruction Head for Object Detection 用于对象检测的表示重建头
Pub Date : 2020-10-01 DOI: 10.1109/ICIP40778.2020.9191049
Shuyu Miao, Rui Feng, Yuejie Zhang
There are two kinds of detection heads in object detection frameworks. Between them, the heads based on full connection contribute to mapping the learned feature representation to the sample label space, while the heads based on full convolution facilitate preserving location sensitivity information. However, to enjoy the benefits from both detection heads is still underexplored. In this paper, we propose a generalized Representation Reconstruction Head (RRHead) to break through the limitation that most detection heads focus on unilateral self-advantage while ignoring another one. RRHead enhances multi scale feature representation for better feature mapping, and employs location sensitivity representation for better location preservation. These optimize fully-convolutional-based heads and fully-connected-based heads separately. RRHead can be embedded in existing detection frameworks to heighten the rationality and reliability of the detection head representation without any additional modification. Extensive experiments show that our proposed RRHead improves the detection performance of the existing frameworks by a large margin on several challenging benchmarks, and achieves new state-of-the-art performance.
在目标检测框架中有两种检测头。其中,基于完全连接的头有助于将学习到的特征表示映射到样本标签空间,而基于完全卷积的头有助于保留位置灵敏度信息。然而,享受两个探测头的好处仍然没有得到充分的探索。本文提出了一种广义表征重构头(RRHead),以突破大多数检测头只关注单方自利而忽略另一方自利的局限。RRHead改进了多尺度特征表示以获得更好的特征映射,并采用位置敏感性表示来获得更好的位置保存。它们分别优化了基于全卷积的头像和基于全连接的头像。RRHead可以嵌入到现有的检测框架中,以提高检测头表示的合理性和可靠性,而无需进行任何额外的修改。大量的实验表明,我们提出的RRHead在几个具有挑战性的基准测试中大大提高了现有框架的检测性能,并实现了新的最先进的性能。
{"title":"Representation Reconstruction Head for Object Detection","authors":"Shuyu Miao, Rui Feng, Yuejie Zhang","doi":"10.1109/ICIP40778.2020.9191049","DOIUrl":"https://doi.org/10.1109/ICIP40778.2020.9191049","url":null,"abstract":"There are two kinds of detection heads in object detection frameworks. Between them, the heads based on full connection contribute to mapping the learned feature representation to the sample label space, while the heads based on full convolution facilitate preserving location sensitivity information. However, to enjoy the benefits from both detection heads is still underexplored. In this paper, we propose a generalized Representation Reconstruction Head (RRHead) to break through the limitation that most detection heads focus on unilateral self-advantage while ignoring another one. RRHead enhances multi scale feature representation for better feature mapping, and employs location sensitivity representation for better location preservation. These optimize fully-convolutional-based heads and fully-connected-based heads separately. RRHead can be embedded in existing detection frameworks to heighten the rationality and reliability of the detection head representation without any additional modification. Extensive experiments show that our proposed RRHead improves the detection performance of the existing frameworks by a large margin on several challenging benchmarks, and achieves new state-of-the-art performance.","PeriodicalId":405734,"journal":{"name":"2020 IEEE International Conference on Image Processing (ICIP)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2020-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129933038","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Accurate Terahertz Imaging Simulation With Ray Tracing Incorporating Beam Shape and Refraction 结合光束形状和折射的射线追踪精确太赫兹成像模拟
Pub Date : 2020-10-01 DOI: 10.1109/ICIP40778.2020.9190937
P. Paramonov, Lars-Paul Lumbeeck, J. D. Beenhouwer, Jan Sijbers
In this paper, we present an approach to realistically simulate terahertz (THz) transmission mode imaging. We model the THz beam shape and account for the refraction of the THz beam at the different media interfaces using ray optics. Our approach does not require prior knowledge on the interfaces, instead it utilizes the refractive index scalar field. We study the beam shape and refraction effects separately by comparing resulting sinograms with the ones simulated by a Gaussian beam model, as well as with a real acquisition of a plastic object. The proposed forward projection can be utilized in iterative reconstruction algorithms to improve the quality of THz CT images.
在本文中,我们提出了一种真实模拟太赫兹(THz)传输模式成像的方法。我们模拟了太赫兹光束的形状,并利用射线光学解释了太赫兹光束在不同介质界面上的折射。我们的方法不需要对界面的先验知识,而是利用折射率标量场。我们分别研究了光束形状和折射效应,并将所得到的波形图与高斯光束模型模拟的波形图以及实际获取的塑料物体的波形图进行了比较。所提出的正演投影可用于迭代重建算法,以提高太赫兹CT图像的质量。
{"title":"Accurate Terahertz Imaging Simulation With Ray Tracing Incorporating Beam Shape and Refraction","authors":"P. Paramonov, Lars-Paul Lumbeeck, J. D. Beenhouwer, Jan Sijbers","doi":"10.1109/ICIP40778.2020.9190937","DOIUrl":"https://doi.org/10.1109/ICIP40778.2020.9190937","url":null,"abstract":"In this paper, we present an approach to realistically simulate terahertz (THz) transmission mode imaging. We model the THz beam shape and account for the refraction of the THz beam at the different media interfaces using ray optics. Our approach does not require prior knowledge on the interfaces, instead it utilizes the refractive index scalar field. We study the beam shape and refraction effects separately by comparing resulting sinograms with the ones simulated by a Gaussian beam model, as well as with a real acquisition of a plastic object. The proposed forward projection can be utilized in iterative reconstruction algorithms to improve the quality of THz CT images.","PeriodicalId":405734,"journal":{"name":"2020 IEEE International Conference on Image Processing (ICIP)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2020-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129006872","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Semantically Supervised Maximal Correlation For Cross-Modal Retrieval 跨模态检索的语义监督最大相关
Pub Date : 2020-10-01 DOI: 10.1109/ICIP40778.2020.9190873
Mingyang Li, Yongni Li, Shao-Lun Huang, Lin Zhang
With the rapid growth of multimedia data, the cross-modal retrieval problem has attracted a lot of interest in both research and industry in recent years. However, the inconsistency of data distribution from different modalities makes such task challenging. In this paper, we propose Semantically Supervised Maximal Correlation (S2MC) method for cross-modal retrieval by incorporating semantic label information into the traditional maximal correlation framework. Combining with maximal correlation based method for extracting unsupervised pairing information, our method effectively exploits supervised semantic information on both common feature space and label space. Extensive experiments show that our method outperforms other current state-of-the-art methods on cross-modal retrieval tasks on three widely used datasets.
随着多媒体数据量的快速增长,跨模式检索问题近年来引起了学术界和工业界的广泛关注。然而,来自不同模式的数据分布的不一致性给这一任务带来了挑战。本文将语义标签信息整合到传统的最大相关框架中,提出了语义监督最大相关(S2MC)跨模态检索方法。该方法结合基于最大相关的无监督配对信息提取方法,有效地利用了公共特征空间和标签空间上的监督语义信息。大量的实验表明,在三个广泛使用的数据集上,我们的方法在跨模态检索任务上优于其他当前最先进的方法。
{"title":"Semantically Supervised Maximal Correlation For Cross-Modal Retrieval","authors":"Mingyang Li, Yongni Li, Shao-Lun Huang, Lin Zhang","doi":"10.1109/ICIP40778.2020.9190873","DOIUrl":"https://doi.org/10.1109/ICIP40778.2020.9190873","url":null,"abstract":"With the rapid growth of multimedia data, the cross-modal retrieval problem has attracted a lot of interest in both research and industry in recent years. However, the inconsistency of data distribution from different modalities makes such task challenging. In this paper, we propose Semantically Supervised Maximal Correlation (S2MC) method for cross-modal retrieval by incorporating semantic label information into the traditional maximal correlation framework. Combining with maximal correlation based method for extracting unsupervised pairing information, our method effectively exploits supervised semantic information on both common feature space and label space. Extensive experiments show that our method outperforms other current state-of-the-art methods on cross-modal retrieval tasks on three widely used datasets.","PeriodicalId":405734,"journal":{"name":"2020 IEEE International Conference on Image Processing (ICIP)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2020-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129160416","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Accelerated 4d Mr Image Reconstruction Using Joint Higher Degree Total Variation And Local Low-Rank Constraints 基于高阶总变分和局部低秩约束的加速4d Mr图像重建
Pub Date : 2020-10-01 DOI: 10.1109/ICIP40778.2020.9191327
Yue Hu, Disi Lin, Kuangshi Zhao
Four-dimensional magnetic resonance imaging (4D-MRI) can provide 3D tissue properties and the temporal profiles at the same time. However, further applications of 4D-MRI is limited by the long acquisition time and motion artifacts. We introduce a regularized image reconstruction method to recover 4D MR images from their undersampled Fourier coefficients, named HDTV-LLR. We adopt the three-dimensional higher degree total variation and the local low-rank penalties to simultaneously exploit the spatial and temporal correlations of the dataset. In order to solve the resulting optimization problem efficiently, we propose a fast alternating minimization algorithm. The performance of the proposed method is demonstrated in the context of 4D cardiac MR images reconstruction with undersampling factors of 12 and 16. The proposed method is compared with iGRASP, and schemes using either low-rank or sparsity constraint alone. Numerical results show that the proposed method enables accelerated 4D-MRI with improved image quality and reduced artifacts.
四维磁共振成像(4D-MRI)可以同时提供三维组织特性和时间剖面。然而,4D-MRI的进一步应用受到采集时间长和运动伪影的限制。我们引入了一种正则化图像重建方法,从欠采样的傅里叶系数中恢复4D MR图像,称为HDTV-LLR。我们采用三维高阶总变分和局部低秩惩罚来同时挖掘数据集的时空相关性。为了有效地解决最终的优化问题,我们提出了一种快速交替最小化算法。在欠采样因子为12和16的4D心脏MR图像重建中,证明了该方法的性能。将该方法与iGRASP和仅使用低秩约束或稀疏约束的方案进行了比较。数值计算结果表明,该方法能够在提高图像质量和减少伪影的情况下加速4D-MRI。
{"title":"Accelerated 4d Mr Image Reconstruction Using Joint Higher Degree Total Variation And Local Low-Rank Constraints","authors":"Yue Hu, Disi Lin, Kuangshi Zhao","doi":"10.1109/ICIP40778.2020.9191327","DOIUrl":"https://doi.org/10.1109/ICIP40778.2020.9191327","url":null,"abstract":"Four-dimensional magnetic resonance imaging (4D-MRI) can provide 3D tissue properties and the temporal profiles at the same time. However, further applications of 4D-MRI is limited by the long acquisition time and motion artifacts. We introduce a regularized image reconstruction method to recover 4D MR images from their undersampled Fourier coefficients, named HDTV-LLR. We adopt the three-dimensional higher degree total variation and the local low-rank penalties to simultaneously exploit the spatial and temporal correlations of the dataset. In order to solve the resulting optimization problem efficiently, we propose a fast alternating minimization algorithm. The performance of the proposed method is demonstrated in the context of 4D cardiac MR images reconstruction with undersampling factors of 12 and 16. The proposed method is compared with iGRASP, and schemes using either low-rank or sparsity constraint alone. Numerical results show that the proposed method enables accelerated 4D-MRI with improved image quality and reduced artifacts.","PeriodicalId":405734,"journal":{"name":"2020 IEEE International Conference on Image Processing (ICIP)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2020-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122374701","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
An Enhanced Local Texture Descriptor for Image Segmentation 一种用于图像分割的增强局部纹理描述符
Pub Date : 2020-10-01 DOI: 10.1109/ICIP40778.2020.9190895
Sheikh Tania, M. Murshed, S. Teng, G. Karmakar
Texture is an indispensable property to develop many vision based autonomous applications. Compared to colour, feature dimension in a local texture descriptor is quite large as dense texture features need to represent the distribution of pixel intensities in the neighbourhood of each pixel. Large dimensional features require additional time for further processing that often restrict real-time applications. In this paper, a robust local texture descriptor is enhanced by reducing feature dimension by three folds without compromising the accuracy in region-based image segmentation applications. Reduction in feature dimension is achieved by exploiting the mean of neighbourhood pixel intensities radially along lines across a certain radius, which eliminates the need for sampling intensity distribution at three scales. Both the results of benchmark metrics and computational time are promising when the enhanced texture feature is used in a region-based hierarchical segmentation algorithm, a recent state-of-the-art technique.
纹理是开发许多基于视觉的自主应用程序必不可少的属性。与颜色相比,局部纹理描述符中的特征维度非常大,因为密集的纹理特征需要表示每个像素附近像素强度的分布。大维度特征需要额外的时间进行进一步处理,这通常会限制实时应用程序。在基于区域的图像分割应用中,在不影响图像分割精度的前提下,将特征维数降低三倍,增强了鲁棒的局部纹理描述子。特征维数的减少是通过沿一定半径的直线径向利用邻域像素强度的平均值来实现的,这就消除了在三个尺度上采样强度分布的需要。在基于区域的分层分割算法中使用增强的纹理特征,基准度量和计算时间的结果都是有希望的。
{"title":"An Enhanced Local Texture Descriptor for Image Segmentation","authors":"Sheikh Tania, M. Murshed, S. Teng, G. Karmakar","doi":"10.1109/ICIP40778.2020.9190895","DOIUrl":"https://doi.org/10.1109/ICIP40778.2020.9190895","url":null,"abstract":"Texture is an indispensable property to develop many vision based autonomous applications. Compared to colour, feature dimension in a local texture descriptor is quite large as dense texture features need to represent the distribution of pixel intensities in the neighbourhood of each pixel. Large dimensional features require additional time for further processing that often restrict real-time applications. In this paper, a robust local texture descriptor is enhanced by reducing feature dimension by three folds without compromising the accuracy in region-based image segmentation applications. Reduction in feature dimension is achieved by exploiting the mean of neighbourhood pixel intensities radially along lines across a certain radius, which eliminates the need for sampling intensity distribution at three scales. Both the results of benchmark metrics and computational time are promising when the enhanced texture feature is used in a region-based hierarchical segmentation algorithm, a recent state-of-the-art technique.","PeriodicalId":405734,"journal":{"name":"2020 IEEE International Conference on Image Processing (ICIP)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2020-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128904178","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Reliable Temporally Consistent Feature Adaptation for Visual Object Tracking 可靠的时间一致特征自适应视觉目标跟踪
Pub Date : 2020-10-01 DOI: 10.1109/ICIP40778.2020.9190957
Goutam Yelluru Gopal, Maria A. Amer
Correlation Filter (CF) based trackers have been the frontiers on various object tracking benchmarks. Use of multiple features and sophisticated learning methods have increased the accuracy of tracking results. However, the contribution of features are often fixed throughout the video sequence. Unreliable features lead to erroneous target localization and result in tracking failures. To alleviate this problem, we propose a method for online adaptation of feature weights based on their reliability. Our method also includes the notion of temporal consistency, to handle noisy reliability estimates. The two objectives are coupled to model a convex optimization problem for robust learning of feature weights. We also propose an algorithm to efficiently solve the resulting optimization problem, without hindering tracking speed. Results on VOT2018, TC128 and NfS30 datasets show that proposed method improves the performance of baseline CF trackers.
基于相关滤波器(CF)的跟踪器已经成为各种目标跟踪基准的前沿。使用多种特征和复杂的学习方法提高了跟踪结果的准确性。然而,在整个视频序列中,特征的贡献通常是固定的。不可靠的特征会导致目标定位错误,导致跟踪失败。为了解决这一问题,我们提出了一种基于可靠性的特征权值在线自适应方法。我们的方法还包括时间一致性的概念,以处理有噪声的可靠性估计。这两个目标是耦合的,以建立一个凸优化问题的鲁棒学习特征权值。我们还提出了一种算法,在不影响跟踪速度的情况下有效地解决所产生的优化问题。在VOT2018, TC128和NfS30数据集上的结果表明,该方法提高了基线CF跟踪器的性能。
{"title":"Reliable Temporally Consistent Feature Adaptation for Visual Object Tracking","authors":"Goutam Yelluru Gopal, Maria A. Amer","doi":"10.1109/ICIP40778.2020.9190957","DOIUrl":"https://doi.org/10.1109/ICIP40778.2020.9190957","url":null,"abstract":"Correlation Filter (CF) based trackers have been the frontiers on various object tracking benchmarks. Use of multiple features and sophisticated learning methods have increased the accuracy of tracking results. However, the contribution of features are often fixed throughout the video sequence. Unreliable features lead to erroneous target localization and result in tracking failures. To alleviate this problem, we propose a method for online adaptation of feature weights based on their reliability. Our method also includes the notion of temporal consistency, to handle noisy reliability estimates. The two objectives are coupled to model a convex optimization problem for robust learning of feature weights. We also propose an algorithm to efficiently solve the resulting optimization problem, without hindering tracking speed. Results on VOT2018, TC128 and NfS30 datasets show that proposed method improves the performance of baseline CF trackers.","PeriodicalId":405734,"journal":{"name":"2020 IEEE International Conference on Image Processing (ICIP)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2020-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130584297","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Coding Of Non-Rectangular Signals With Block-Based Transforms 基于分块变换的非矩形信号编码
Pub Date : 2020-10-01 DOI: 10.1109/ICIP40778.2020.9191301
P. Das, N. Horst, M. Wien
This paper presents a transform coding technique for non-rectangular 2-D signals by extending the signal into a rectangular block in order to enable conventional block-based transform coding. The technique could be suitable for coding residuals of prediction blocks using geometric partitioning which has been adopted into the draft Versatile Video Coding standard. The extension of the non-rectangular signal is found using a sparse solution set generated by applying Orthogonal Matching Pursuits using partitioned transform bases. The method developed in this paper is based on Discrete Cosine Transform. Results achieved in an experimental setup outside of the video coding loop are presented for signals of triangular and trapezoidal shape in comparison to the shape-adaptive DCT. Encouraging gains are observed specifically for larger block sizes and in dependency of the quantization parameter and the partitioning shape.
为了实现传统的基于分块的变换编码,本文提出了一种非矩形二维信号的变换编码技术,将信号扩展到一个矩形块中。该技术适用于预测块残差的几何分割编码,并已被通用视频编码标准草案采用。利用利用分块变换基的正交匹配追踪产生的稀疏解集,找到了非矩形信号的扩展。本文提出的方法是基于离散余弦变换的。在视频编码环路外的实验装置中,给出了三角形和梯形信号与形状自适应DCT的比较结果。特别是对于较大的块大小和依赖于量化参数和分区形状的块,可以观察到令人鼓舞的增益。
{"title":"Coding Of Non-Rectangular Signals With Block-Based Transforms","authors":"P. Das, N. Horst, M. Wien","doi":"10.1109/ICIP40778.2020.9191301","DOIUrl":"https://doi.org/10.1109/ICIP40778.2020.9191301","url":null,"abstract":"This paper presents a transform coding technique for non-rectangular 2-D signals by extending the signal into a rectangular block in order to enable conventional block-based transform coding. The technique could be suitable for coding residuals of prediction blocks using geometric partitioning which has been adopted into the draft Versatile Video Coding standard. The extension of the non-rectangular signal is found using a sparse solution set generated by applying Orthogonal Matching Pursuits using partitioned transform bases. The method developed in this paper is based on Discrete Cosine Transform. Results achieved in an experimental setup outside of the video coding loop are presented for signals of triangular and trapezoidal shape in comparison to the shape-adaptive DCT. Encouraging gains are observed specifically for larger block sizes and in dependency of the quantization parameter and the partitioning shape.","PeriodicalId":405734,"journal":{"name":"2020 IEEE International Conference on Image Processing (ICIP)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2020-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130587151","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Shape-Adaptive Kernel Network for Dense Object Detection 用于密集目标检测的形状自适应核网络
Pub Date : 2020-10-01 DOI: 10.1109/ICIP40778.2020.9190767
H. Kim, Sunghun Joung, Ig-Jae Kim, K. Sohn
Dense object detectors that are applied over a regular, dense grid have advanced and drawn their attention in recent days. Their fully convolutional nature greatly advances the computational efficiency of object detectors compared to the two-stage detectors. However, the lack of the ability to adjust shape variation on a regular grid is still limited. In this paper we introduce a new framework, shape-adaptive kernel network, to handle spatial manipulation of input data in convolutional kernel space. At the heart of out approach is to align the original kernel space recovering shape variation of each input feature on regular grid. To this end, we propose a shape-adaptive kernel sampler to adjust dynamic convolutional kernel conditioned on input. To increase the flexibility of geometric transformation, a cascade refinement module is designed, which first estimates the global transformation grid and then estimates local offset in convolutional kernel space. Our experiments demonstrate the effectiveness of the shape-adaptive kernel network for dense object detection on various benchmarks.
最近几天,应用于规则密集网格的密集物体探测器取得了进展,并引起了人们的注意。与两级检测器相比,它们的全卷积特性大大提高了目标检测器的计算效率。然而,缺乏在规则网格上调整形状变化的能力仍然是有限的。本文引入了一种新的框架——形状自适应核网络来处理卷积核空间中输入数据的空间处理。该方法的核心是在规则网格上对齐原始核空间,恢复每个输入特征的形状变化。为此,我们提出了一种形状自适应核采样器来调整以输入为条件的动态卷积核。为了提高几何变换的灵活性,设计了级联细化模块,该模块首先估计全局变换网格,然后在卷积核空间中估计局部偏移量。我们的实验证明了形状自适应核网络在密集目标检测中的有效性。
{"title":"Shape-Adaptive Kernel Network for Dense Object Detection","authors":"H. Kim, Sunghun Joung, Ig-Jae Kim, K. Sohn","doi":"10.1109/ICIP40778.2020.9190767","DOIUrl":"https://doi.org/10.1109/ICIP40778.2020.9190767","url":null,"abstract":"Dense object detectors that are applied over a regular, dense grid have advanced and drawn their attention in recent days. Their fully convolutional nature greatly advances the computational efficiency of object detectors compared to the two-stage detectors. However, the lack of the ability to adjust shape variation on a regular grid is still limited. In this paper we introduce a new framework, shape-adaptive kernel network, to handle spatial manipulation of input data in convolutional kernel space. At the heart of out approach is to align the original kernel space recovering shape variation of each input feature on regular grid. To this end, we propose a shape-adaptive kernel sampler to adjust dynamic convolutional kernel conditioned on input. To increase the flexibility of geometric transformation, a cascade refinement module is designed, which first estimates the global transformation grid and then estimates local offset in convolutional kernel space. Our experiments demonstrate the effectiveness of the shape-adaptive kernel network for dense object detection on various benchmarks.","PeriodicalId":405734,"journal":{"name":"2020 IEEE International Conference on Image Processing (ICIP)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2020-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130791269","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
期刊
2020 IEEE International Conference on Image Processing (ICIP)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1