首页 > 最新文献

2016 IEEE International Conference on Image Processing (ICIP)最新文献

英文 中文
Model based image reconstruction with physics based priors 基于物理先验的模型图像重建
Pub Date : 2016-09-01 DOI: 10.1109/ICIP.2016.7532945
M. U. Sadiq, J. Simmons, C. Bouman
Computed tomography is increasingly enabling scientists to study physical processes of materials at micron scales. The MBIR framework provides a powerful method for CT reconstruction by incorporating both a measurement model and prior model. Classically, the choice of prior has been limited to models enforcing local similarity in the image data. In some material science problems, however, much more may be known about the underlying physical process being imaged. Moreover, recent work in Plug-And-Play decoupling of the MBIR problem has enabled researchers to look beyond classical prior models, and innovations in methods of data acquisition such as interlaced view sampling have also shown promise for imaging of dynamic physical processes. In this paper, we propose an MBIR framework with a physics based prior model - namely the Cahn-Hilliard equation. The Cahn-Hilliard equation can be used to describe the spatiotemporal evolution of binary alloys. After formulating the MBIR cost with Cahn-Hilliard prior, we use Plug-And-Play algorithm with ICD optimization to minimize this cost. We apply this method to simulated data using the interlaced-view sampling method of data acquisition. Results show superior reconstruction quality compared to the Filtered Back Projection. Though we use Cahn-Hilliard equation as one instance, the method can be easily extended to use any other physics-based prior model for a different set of applications.
计算机断层扫描越来越使科学家能够在微米尺度上研究材料的物理过程。MBIR框架结合了测量模型和先验模型,为CT重建提供了一种强大的方法。传统上,先验的选择仅限于在图像数据中增强局部相似性的模型。然而,在一些材料科学问题中,对于被成像的潜在物理过程,我们可能知道得更多。此外,最近在MBIR问题的即插即用解耦方面的工作使研究人员能够超越经典的先前模型,数据采集方法的创新,如隔行视图采样,也显示了动态物理过程成像的希望。在本文中,我们提出了一个基于物理先验模型的MBIR框架-即Cahn-Hilliard方程。Cahn-Hilliard方程可以用来描述二元合金的时空演化。在使用Cahn-Hilliard先验法确定MBIR成本后,我们使用即插即用算法和ICD优化来最小化该成本。采用数据采集的隔行视图采样方法,将该方法应用于模拟数据。结果表明,与滤波后的投影相比,重建质量更好。虽然我们使用Cahn-Hilliard方程作为一个例子,但该方法可以很容易地扩展到使用任何其他基于物理的先验模型,用于不同的应用程序集。
{"title":"Model based image reconstruction with physics based priors","authors":"M. U. Sadiq, J. Simmons, C. Bouman","doi":"10.1109/ICIP.2016.7532945","DOIUrl":"https://doi.org/10.1109/ICIP.2016.7532945","url":null,"abstract":"Computed tomography is increasingly enabling scientists to study physical processes of materials at micron scales. The MBIR framework provides a powerful method for CT reconstruction by incorporating both a measurement model and prior model. Classically, the choice of prior has been limited to models enforcing local similarity in the image data. In some material science problems, however, much more may be known about the underlying physical process being imaged. Moreover, recent work in Plug-And-Play decoupling of the MBIR problem has enabled researchers to look beyond classical prior models, and innovations in methods of data acquisition such as interlaced view sampling have also shown promise for imaging of dynamic physical processes. In this paper, we propose an MBIR framework with a physics based prior model - namely the Cahn-Hilliard equation. The Cahn-Hilliard equation can be used to describe the spatiotemporal evolution of binary alloys. After formulating the MBIR cost with Cahn-Hilliard prior, we use Plug-And-Play algorithm with ICD optimization to minimize this cost. We apply this method to simulated data using the interlaced-view sampling method of data acquisition. Results show superior reconstruction quality compared to the Filtered Back Projection. Though we use Cahn-Hilliard equation as one instance, the method can be easily extended to use any other physics-based prior model for a different set of applications.","PeriodicalId":6521,"journal":{"name":"2016 IEEE International Conference on Image Processing (ICIP)","volume":"153 1","pages":"3176-3179"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"86155506","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Trilateral filtering-based hybrid up-sampling in dual domains for single video frame super resolution 基于三边滤波的双域混合上采样单帧超分辨率
Pub Date : 2016-09-01 DOI: 10.1109/ICIP.2016.7532540
Zhenyu Wu, H. Hu
Up-sampling has been one of the key techniques for multimedia processing. Higher resolution videos are always the pursuit goal of major customers. Pervasive applications of multimedia processing starve for this technique to solve mismatch problems in display resolutions between senders and receivers. The hybrid DCT-Wiener-based interpolation method designs a powerful framework to interpolate video frames by mining the information in both spatial and DCT domain briefly. It can provide better objective as well as visual qualities with low complexity than many existing well studied interpolation methods. This paper presents an analysis about the bottleneck of hybrid DCT-Wiener-based interpolation method firstly. And then, proposes a trilateral filtering-based hybrid up-sampling algorithm in dual domains, which has dug the information in spatial and frequency domains more deeply. The proposed spatial domain interpolation scheme is an adaptive Wiener filter with a trilateral filtering enhancement, which possess capability to overcome the quarter-pixel shift mismatch of hybrid DCT-Wiener-based interpolation method and achieve much more accurate detail information estimation. Furthermore, flexible block size chosen mechanism in frequency domain enables the whole proposed up-sampling algorithm achieve further advantages in high frequency coefficients retaining. Experiments have been carried out to demonstrate that the proposed algorithm is able to achieve noticeable gains over state of art methods in both objective and visual qualities measurements.
上采样是多媒体处理的关键技术之一。更高分辨率的视频一直是大客户的追求目标。多媒体处理的普遍应用需要这种技术来解决发送方和接收方之间显示分辨率不匹配的问题。基于DCT- wiener的混合插值方法通过对空间域和DCT域的信息进行简单的挖掘,设计了一个强大的视频帧插值框架。与许多已有的插值方法相比,该方法可以提供更好的客观质量和视觉质量,并且复杂度较低。本文首先分析了基于dct - wiener的混合插值方法存在的瓶颈。然后,提出了一种基于三边滤波的对偶域混合上采样算法,更深入地挖掘了空间域和频率域的信息。本文提出的空域插值方案是一种带有三边滤波增强的自适应维纳滤波器,能够克服基于dct -维纳混合插值方法的四分之一像素偏移失配,实现更精确的细节信息估计。此外,在频域灵活的块大小选择机制使整个上采样算法在高频系数保留方面具有进一步的优势。实验已经进行,以证明所提出的算法能够在客观和视觉质量测量中实现比最先进的方法显著的增益。
{"title":"Trilateral filtering-based hybrid up-sampling in dual domains for single video frame super resolution","authors":"Zhenyu Wu, H. Hu","doi":"10.1109/ICIP.2016.7532540","DOIUrl":"https://doi.org/10.1109/ICIP.2016.7532540","url":null,"abstract":"Up-sampling has been one of the key techniques for multimedia processing. Higher resolution videos are always the pursuit goal of major customers. Pervasive applications of multimedia processing starve for this technique to solve mismatch problems in display resolutions between senders and receivers. The hybrid DCT-Wiener-based interpolation method designs a powerful framework to interpolate video frames by mining the information in both spatial and DCT domain briefly. It can provide better objective as well as visual qualities with low complexity than many existing well studied interpolation methods. This paper presents an analysis about the bottleneck of hybrid DCT-Wiener-based interpolation method firstly. And then, proposes a trilateral filtering-based hybrid up-sampling algorithm in dual domains, which has dug the information in spatial and frequency domains more deeply. The proposed spatial domain interpolation scheme is an adaptive Wiener filter with a trilateral filtering enhancement, which possess capability to overcome the quarter-pixel shift mismatch of hybrid DCT-Wiener-based interpolation method and achieve much more accurate detail information estimation. Furthermore, flexible block size chosen mechanism in frequency domain enables the whole proposed up-sampling algorithm achieve further advantages in high frequency coefficients retaining. Experiments have been carried out to demonstrate that the proposed algorithm is able to achieve noticeable gains over state of art methods in both objective and visual qualities measurements.","PeriodicalId":6521,"journal":{"name":"2016 IEEE International Conference on Image Processing (ICIP)","volume":"15 1","pages":"1160-1164"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"84737057","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Pedestrian detection via a leg-driven physiology framework 通过腿驱动的生理学框架行人检测
Pub Date : 2016-09-01 DOI: 10.1109/ICIP.2016.7532895
Gongbo Liang, Qi Li, Xiangui Kang
In this paper, we propose a leg-driven physiology framework for pedestrian detection. The framework is introduced to reduce the search space of candidate regions of pedestrians. Given a set of vertical line segments, we can generate a space of rectangular candidate regions, based on a model of body proportions. The proposed framework can be either integrated with or without learning-based pedestrian detection methods to validate the candidate regions. A symmetry constraint is then applied to validate each candidate region to decrease the false positive rate. The experiment demonstrates the promising results of the proposed method by comparing it with Dalal & Triggs method. For example, rectangular regions detected by the proposed method has much similar area to the ground truth than regions detected by Dalal & Triggs method.
在本文中,我们提出了一个腿部驱动的行人检测生理学框架。引入该框架来减小行人候选区域的搜索空间。给定一组垂直线段,我们可以根据身体比例模型生成一个矩形候选区域的空间。所提出的框架可以与基于学习的行人检测方法集成或不集成以验证候选区域。然后应用对称约束来验证每个候选区域,以降低误报率。通过与Dalal & Triggs方法的比较,验证了该方法的有效性。例如,该方法检测到的矩形区域比Dalal & Triggs方法检测到的区域具有更接近地面真值的面积。
{"title":"Pedestrian detection via a leg-driven physiology framework","authors":"Gongbo Liang, Qi Li, Xiangui Kang","doi":"10.1109/ICIP.2016.7532895","DOIUrl":"https://doi.org/10.1109/ICIP.2016.7532895","url":null,"abstract":"In this paper, we propose a leg-driven physiology framework for pedestrian detection. The framework is introduced to reduce the search space of candidate regions of pedestrians. Given a set of vertical line segments, we can generate a space of rectangular candidate regions, based on a model of body proportions. The proposed framework can be either integrated with or without learning-based pedestrian detection methods to validate the candidate regions. A symmetry constraint is then applied to validate each candidate region to decrease the false positive rate. The experiment demonstrates the promising results of the proposed method by comparing it with Dalal & Triggs method. For example, rectangular regions detected by the proposed method has much similar area to the ground truth than regions detected by Dalal & Triggs method.","PeriodicalId":6521,"journal":{"name":"2016 IEEE International Conference on Image Processing (ICIP)","volume":"13 1","pages":"2926-2930"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"87222277","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
HEVC still image coding and high efficiency image file format HEVC静态图像编码和高效的图像文件格式
Pub Date : 2016-09-01 DOI: 10.1109/ICIP.2016.7532321
J. Lainema, M. Hannuksela, V. Vadakital, Emre B. Aksu
The High Efficiency Video Coding (HEVC) standard includes support for a large range of image representation formats and provides an excellent image compression capability. The High Efficiency Image File Format (HEIF) offers a convenient way to encapsulate HEVC coded images, image sequences and animations together with associated metadata into a single file. This paper discusses various features and functionalities of the HEIF file format and compares the compression efficiency of HEVC still image coding to that of JPEG 2000. According to the experimental results HEVC provides about 25% bitrate reduction compared to JPEG 2000, while keeping the same objective picture quality.
高效视频编码(HEVC)标准包括对大量图像表示格式的支持,并提供了出色的图像压缩能力。高效图像文件格式(HEIF)提供了一种方便的方式来封装HEVC编码的图像、图像序列和动画以及相关的元数据到一个文件中。本文讨论了HEVC文件格式的各种特性和功能,并比较了HEVC静态图像编码与JPEG 2000的压缩效率。实验结果表明,HEVC在保持客观图像质量不变的情况下,比特率比JPEG 2000降低了25%左右。
{"title":"HEVC still image coding and high efficiency image file format","authors":"J. Lainema, M. Hannuksela, V. Vadakital, Emre B. Aksu","doi":"10.1109/ICIP.2016.7532321","DOIUrl":"https://doi.org/10.1109/ICIP.2016.7532321","url":null,"abstract":"The High Efficiency Video Coding (HEVC) standard includes support for a large range of image representation formats and provides an excellent image compression capability. The High Efficiency Image File Format (HEIF) offers a convenient way to encapsulate HEVC coded images, image sequences and animations together with associated metadata into a single file. This paper discusses various features and functionalities of the HEIF file format and compares the compression efficiency of HEVC still image coding to that of JPEG 2000. According to the experimental results HEVC provides about 25% bitrate reduction compared to JPEG 2000, while keeping the same objective picture quality.","PeriodicalId":6521,"journal":{"name":"2016 IEEE International Conference on Image Processing (ICIP)","volume":"2015 1","pages":"71-75"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"87946528","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 23
Localized region context and object feature fusion for people head detection 基于局部区域上下文和目标特征融合的人头部检测
Pub Date : 2016-09-01 DOI: 10.1109/ICIP.2016.7532426
Yule Li, Y. Dou, Xinwang Liu, Teng Li
People head detection in crowded scenes is challenging due to the large variability in clothing and appearance, small scales of people, and strong partial occlusions. Traditional bottom-up proposal methods and existing region proposal network approaches suffer from either poor recall or low precision. In this paper, we propose to improve both the recall and precision of head detection of region proposal models by integrating the local head information. In specific, we first use a region proposal network to predict the bounding boxes and corresponding scores of multiple instances in the region. A local head classifier network is then trained to score the bounding box generated from the region proposal model. After that, we propose an adaptive fusion method by optimally combining both the region and local scores to obtain the final score of each candidate bounding box. Furthermore, our fusion models can automatically learn the optimal hyper-parameters from data. Our algorithm achieves superior people head detection performance on the crowded scenes data set, which significantly outperforms several recent state-of-the-art baselines in the literature.
在拥挤的场景中,由于服装和外表的巨大变化,人的规模小,以及强烈的局部遮挡,人的头部检测是具有挑战性的。传统的自下而上的提议方法和现有的区域提议网络方法存在查全率差和查准率低的问题。在本文中,我们提出通过整合局部头部信息来提高区域建议模型的头部检测召回率和精度。具体而言,我们首先使用区域建议网络来预测区域内多个实例的边界框和相应的分数。然后训练局部头部分类器网络对区域建议模型生成的边界框进行评分。然后,我们提出了一种自适应融合方法,将区域分数和局部分数最优结合,得到每个候选边界框的最终分数。此外,我们的融合模型可以从数据中自动学习到最优的超参数。我们的算法在拥挤的场景数据集上实现了优越的人员头部检测性能,显著优于文献中最近的几个最先进的基线。
{"title":"Localized region context and object feature fusion for people head detection","authors":"Yule Li, Y. Dou, Xinwang Liu, Teng Li","doi":"10.1109/ICIP.2016.7532426","DOIUrl":"https://doi.org/10.1109/ICIP.2016.7532426","url":null,"abstract":"People head detection in crowded scenes is challenging due to the large variability in clothing and appearance, small scales of people, and strong partial occlusions. Traditional bottom-up proposal methods and existing region proposal network approaches suffer from either poor recall or low precision. In this paper, we propose to improve both the recall and precision of head detection of region proposal models by integrating the local head information. In specific, we first use a region proposal network to predict the bounding boxes and corresponding scores of multiple instances in the region. A local head classifier network is then trained to score the bounding box generated from the region proposal model. After that, we propose an adaptive fusion method by optimally combining both the region and local scores to obtain the final score of each candidate bounding box. Furthermore, our fusion models can automatically learn the optimal hyper-parameters from data. Our algorithm achieves superior people head detection performance on the crowded scenes data set, which significantly outperforms several recent state-of-the-art baselines in the literature.","PeriodicalId":6521,"journal":{"name":"2016 IEEE International Conference on Image Processing (ICIP)","volume":"35 1","pages":"594-598"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"88260530","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
Characterizing distortions in first-person videos 描述第一人称视频中的扭曲
Pub Date : 2016-09-01 DOI: 10.1109/ICIP.2016.7532797
Chen Bai, A. Reibman
First-person videos (FPVs) captured by wearable cameras often contain heavy distortions, including motion blur, rolling shutter artifacts and rotation. Existing image and video quality estimators are inefficient for this type of video. We develop a method specifically to measure the distortions present in FPVs, without using a high quality reference video. Our local visual information (LVI) algorithm measures motion blur, and we combine homography estimation with line angle histogram to measure rolling shutter artifacts and rotation. Our experiments demonstrate that captured FPVs have dramatically different distortions compared to traditional source videos. We also show that LVI is responsive to motion blur, but insensitive to rotation and shear.
可穿戴相机拍摄的第一人称视频(fps)通常存在严重失真,包括运动模糊、滚动快门伪影和旋转。现有的图像和视频质量估计器对于这种类型的视频是低效的。我们开发了一种专门测量fpv中存在的失真的方法,而不使用高质量的参考视频。我们的局部视觉信息(LVI)算法测量运动模糊,并将单应性估计与线角直方图相结合来测量滚动快门伪影和旋转。我们的实验表明,与传统源视频相比,捕获的fpv具有显着不同的失真。我们还表明LVI对运动模糊有响应,但对旋转和剪切不敏感。
{"title":"Characterizing distortions in first-person videos","authors":"Chen Bai, A. Reibman","doi":"10.1109/ICIP.2016.7532797","DOIUrl":"https://doi.org/10.1109/ICIP.2016.7532797","url":null,"abstract":"First-person videos (FPVs) captured by wearable cameras often contain heavy distortions, including motion blur, rolling shutter artifacts and rotation. Existing image and video quality estimators are inefficient for this type of video. We develop a method specifically to measure the distortions present in FPVs, without using a high quality reference video. Our local visual information (LVI) algorithm measures motion blur, and we combine homography estimation with line angle histogram to measure rolling shutter artifacts and rotation. Our experiments demonstrate that captured FPVs have dramatically different distortions compared to traditional source videos. We also show that LVI is responsive to motion blur, but insensitive to rotation and shear.","PeriodicalId":6521,"journal":{"name":"2016 IEEE International Conference on Image Processing (ICIP)","volume":"58 1","pages":"2440-2444"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"86965780","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
H.264 intra coding with transforms based on prediction inaccuracy modeling 基于预测误差建模的H.264帧内编码
Pub Date : 2016-09-01 DOI: 10.1109/ICIP.2016.7532785
Xun Cai, J. Lim
In intra video coding, intra frames are predicted with intra prediction and the prediction residual signal is encoded. In many transform-based video coding systems, intra prediction residuals are encoded with transforms. For example, the Discrete Cosine Transform (DCT) and the Asymmetric Discrete Sine Transform (ADST) are used for intra prediction residuals in many coding systems. In the recent work, a set of transforms based on prediction inaccuracy modeling (PIM) has been proposed. These transforms are developed based on the observation that much of the residual non-stationarity is due to the use of an inaccurate prediction parameter. These transforms are shown to be effective for non-stationarity that arises in directional intra prediction residuals. In this paper, we implement the transforms based on prediction inaccuracy modeling on the H.264 intra coding system. The proposed transform is used in hybrid with the ADST. We compare the performance of the hybrid transform with the ADST and show that a significant bit-rate reduction is obtained with the proposed transform.
在视频内编码中,利用内预测对内帧进行预测,并对预测残差信号进行编码。在许多基于变换的视频编码系统中,用变换对帧内预测残差进行编码。例如,在许多编码系统中,离散余弦变换(DCT)和非对称离散正弦变换(ADST)被用于预测帧内残差。在最近的工作中,提出了一组基于预测不准确性建模的变换。这些变换是基于这样的观察,即大部分剩余非平稳性是由于使用了不准确的预测参数。这些变换被证明是有效的非平稳性产生的方向内预测残差。本文在H.264帧内编码系统上实现了基于预测误差建模的变换。所提出的变换与ADST混合使用。我们将混合变换与ADST的性能进行了比较,结果表明混合变换能显著降低比特率。
{"title":"H.264 intra coding with transforms based on prediction inaccuracy modeling","authors":"Xun Cai, J. Lim","doi":"10.1109/ICIP.2016.7532785","DOIUrl":"https://doi.org/10.1109/ICIP.2016.7532785","url":null,"abstract":"In intra video coding, intra frames are predicted with intra prediction and the prediction residual signal is encoded. In many transform-based video coding systems, intra prediction residuals are encoded with transforms. For example, the Discrete Cosine Transform (DCT) and the Asymmetric Discrete Sine Transform (ADST) are used for intra prediction residuals in many coding systems. In the recent work, a set of transforms based on prediction inaccuracy modeling (PIM) has been proposed. These transforms are developed based on the observation that much of the residual non-stationarity is due to the use of an inaccurate prediction parameter. These transforms are shown to be effective for non-stationarity that arises in directional intra prediction residuals. In this paper, we implement the transforms based on prediction inaccuracy modeling on the H.264 intra coding system. The proposed transform is used in hybrid with the ADST. We compare the performance of the hybrid transform with the ADST and show that a significant bit-rate reduction is obtained with the proposed transform.","PeriodicalId":6521,"journal":{"name":"2016 IEEE International Conference on Image Processing (ICIP)","volume":"16 1","pages":"2380-2384"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"87048311","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Fisher-selective search for object detection 目标检测的费雪选择性搜索
Pub Date : 2016-09-01 DOI: 10.1109/ICIP.2016.7533037
Ilker Buzcu, Aydin Alatan
An enhancement to one of the existing visual object detection approaches is proposed for generating candidate windows that improves detection accuracy at no additional computational cost. Hypothesis windows for object detection are obtained based on Fisher Vector representations over initially obtained superpixels. In order to obtain new window hypotheses, hierarchical merging of superpixel regions are applied, depending upon improvements on some objectiveness measures with no additional cost due to additivity of Fisher Vectors. The proposed technique is further improved by concatenating these representations with that of deep networks. Based on the results of the simulations on typical data sets, it can be argued that the approach is quite promising for its use of handcrafted features left to dust due to the rise of deep learning.
提出了一种对现有视觉目标检测方法的改进,在不增加计算成本的情况下生成候选窗口,从而提高检测精度。目标检测的假设窗口是基于初始获得的超像素上的Fisher向量表示获得的。为了获得新的窗口假设,采用超像素区域的分层合并,这取决于对一些客观度量的改进,而不需要由于Fisher向量的可加性而增加成本。通过将这些表示与深度网络的表示连接起来,所提出的技术进一步得到改进。基于典型数据集的模拟结果,可以认为该方法非常有前途,因为它使用了由于深度学习的兴起而遗留下来的手工特征。
{"title":"Fisher-selective search for object detection","authors":"Ilker Buzcu, Aydin Alatan","doi":"10.1109/ICIP.2016.7533037","DOIUrl":"https://doi.org/10.1109/ICIP.2016.7533037","url":null,"abstract":"An enhancement to one of the existing visual object detection approaches is proposed for generating candidate windows that improves detection accuracy at no additional computational cost. Hypothesis windows for object detection are obtained based on Fisher Vector representations over initially obtained superpixels. In order to obtain new window hypotheses, hierarchical merging of superpixel regions are applied, depending upon improvements on some objectiveness measures with no additional cost due to additivity of Fisher Vectors. The proposed technique is further improved by concatenating these representations with that of deep networks. Based on the results of the simulations on typical data sets, it can be argued that the approach is quite promising for its use of handcrafted features left to dust due to the rise of deep learning.","PeriodicalId":6521,"journal":{"name":"2016 IEEE International Conference on Image Processing (ICIP)","volume":"20 1","pages":"3633-3637"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"85903620","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 12
Structured Discriminative Nonnegative Matrix Factorization for hyperspectral unmixing 高光谱分解的结构判别非负矩阵分解
Pub Date : 2016-09-01 DOI: 10.1109/ICIP.2016.7532678
Xue Li, J. Zhou, Lei Tong, Xun Yu, Jianhui Guo, Chunxia Zhao
Hyperspectral unmixing is an important technique for identifying the constituent spectra and estimating their corresponding fractions in an image. Nonnegative Matrix Factorization (NMF) has recently been widely used for hyperspectral unmixing. However, due to the complex distribution of hyperspectral data, most existing NMF algorithms cannot adequately reflect the intrinsic relationship of the data. In this paper, we propose a novel method, Structured Discriminative Nonnegative Matrix Factorization (SDNMF), to preserve the structural information of hyperspectral data. This is achieved by introducing structured discriminative regularization terms to model both local affinity and distant repulsion of observed spectral responses. Moreover, considering that the abundances of most materials are sparse, a sparseness constraint is also introduced into SDNMF. Experimental results on both synthetic and real data have validated the effectiveness of the proposed method which achieves better unmixing performance than several alternative approaches.
高光谱解混是一种重要的光谱识别技术,用于识别图像中的光谱成分并估计其对应的分数。近年来,非负矩阵分解(NMF)在高光谱解混中得到了广泛的应用。然而,由于高光谱数据的复杂分布,现有的大多数NMF算法不能充分反映数据的内在关系。本文提出了一种保留高光谱数据结构信息的新方法——结构化判别非负矩阵分解(SDNMF)。这是通过引入结构化判别正则化项来模拟观察到的光谱响应的局部亲和和远处排斥来实现的。此外,考虑到大多数材料的丰度是稀疏的,在SDNMF中还引入了稀疏性约束。在合成数据和实际数据上的实验结果验证了该方法的有效性,并取得了较好的解混性能。
{"title":"Structured Discriminative Nonnegative Matrix Factorization for hyperspectral unmixing","authors":"Xue Li, J. Zhou, Lei Tong, Xun Yu, Jianhui Guo, Chunxia Zhao","doi":"10.1109/ICIP.2016.7532678","DOIUrl":"https://doi.org/10.1109/ICIP.2016.7532678","url":null,"abstract":"Hyperspectral unmixing is an important technique for identifying the constituent spectra and estimating their corresponding fractions in an image. Nonnegative Matrix Factorization (NMF) has recently been widely used for hyperspectral unmixing. However, due to the complex distribution of hyperspectral data, most existing NMF algorithms cannot adequately reflect the intrinsic relationship of the data. In this paper, we propose a novel method, Structured Discriminative Nonnegative Matrix Factorization (SDNMF), to preserve the structural information of hyperspectral data. This is achieved by introducing structured discriminative regularization terms to model both local affinity and distant repulsion of observed spectral responses. Moreover, considering that the abundances of most materials are sparse, a sparseness constraint is also introduced into SDNMF. Experimental results on both synthetic and real data have validated the effectiveness of the proposed method which achieves better unmixing performance than several alternative approaches.","PeriodicalId":6521,"journal":{"name":"2016 IEEE International Conference on Image Processing (ICIP)","volume":"113 1","pages":"1848-1852"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"86225580","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Membrane segmentation via active learning with deep networks 基于深度网络主动学习的膜分割
Pub Date : 2016-09-01 DOI: 10.1109/ICIP.2016.7532697
Utkarsh Gaur, M. Kourakis, E. Newman-Smith, William C. Smith, B. S. Manjunath
Segmentation is a key component of several bio-medical image processing systems. Recently, segmentation methods based on supervised learning such as deep convolutional networks have enjoyed immense success for natural image datasets and biological datasets alike. These methods require large volumes of data to avoid overfitting which limits their applicability. In this work, we present a transfer learning mechanism based on active learning which allows us to utilize pre-trained deep networks for segmenting new domains with limited labelled data. We introduce a novel optimization criterion to allow feedback on the most uncertain, yet abundant image patterns thus provisioning for an expert in the loop albeit with minimum amount of guidance. Our experiments demonstrate the effectiveness of the proposed method in improving segmentation performance with very limited labelled data.
分割是生物医学图像处理系统的关键组成部分。最近,基于监督学习的分割方法,如深度卷积网络,在自然图像数据集和生物数据集上都取得了巨大的成功。这些方法需要大量的数据,以避免过度拟合,从而限制了它们的适用性。在这项工作中,我们提出了一种基于主动学习的迁移学习机制,该机制允许我们利用预训练的深度网络来分割具有有限标记数据的新域。我们引入了一种新的优化准则,允许对最不确定的、但丰富的图像模式进行反馈,从而为回路中的专家提供最少的指导。我们的实验证明了该方法在非常有限的标记数据下提高分割性能的有效性。
{"title":"Membrane segmentation via active learning with deep networks","authors":"Utkarsh Gaur, M. Kourakis, E. Newman-Smith, William C. Smith, B. S. Manjunath","doi":"10.1109/ICIP.2016.7532697","DOIUrl":"https://doi.org/10.1109/ICIP.2016.7532697","url":null,"abstract":"Segmentation is a key component of several bio-medical image processing systems. Recently, segmentation methods based on supervised learning such as deep convolutional networks have enjoyed immense success for natural image datasets and biological datasets alike. These methods require large volumes of data to avoid overfitting which limits their applicability. In this work, we present a transfer learning mechanism based on active learning which allows us to utilize pre-trained deep networks for segmenting new domains with limited labelled data. We introduce a novel optimization criterion to allow feedback on the most uncertain, yet abundant image patterns thus provisioning for an expert in the loop albeit with minimum amount of guidance. Our experiments demonstrate the effectiveness of the proposed method in improving segmentation performance with very limited labelled data.","PeriodicalId":6521,"journal":{"name":"2016 IEEE International Conference on Image Processing (ICIP)","volume":"81 1","pages":"1943-1947"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"83140227","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
期刊
2016 IEEE International Conference on Image Processing (ICIP)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1