首页 > 最新文献

2014 IEEE International Conference on Image Processing (ICIP)最新文献

英文 中文
A dataset for Hand-Held Object Recognition 手持物体识别的数据集
Pub Date : 2014-10-01 DOI: 10.1109/ICIP.2014.7026188
Jose Rivera-Rubio, Saad Idrees, I. Alexiou, Lucas Hadjilucas, A. Bharath
Visual object recognition is just one of the many applications of camera-equipped smartphones. The ability to recognise objects through photos taken with wearable and handheld cameras is already possible through some of the larger internet search providers; yet, there is little rigorous analysis of the quality of search results, particularly where there is great disparity in image quality. This has motivated us to develop the Small Hand-held Object Recognition Test (SHORT). This includes a dataset that is suitable for recognising hand-held objects from either snapshots or videos acquired using hand-held or wearable cameras. SHORT provides a collection of images and ground truth that help evaluate the different factors that affect recognition performance. At its present state, the dataset is comprised of a set of high quality training images and a large set of nearly 135,000 smartphone-captured test images of 30 grocery products. In this paper, we will discuss some open challenges in the visual object recognition of objects that are being held by users. We evaluate the performance of a number of popular object recognition algorithms, with differing levels of complexity, when tested against SHORT.
视觉对象识别只是配备摄像头的智能手机的众多应用之一。通过一些大型互联网搜索提供商,通过可穿戴和手持相机拍摄的照片识别物体的能力已经成为可能;然而,很少有对搜索结果质量的严格分析,特别是在图像质量存在巨大差异的情况下。这促使我们开发小型手持对象识别测试(SHORT)。这包括一个适合从使用手持或可穿戴相机获取的快照或视频中识别手持物体的数据集。SHORT提供了一组图像和ground truth,帮助评估影响识别性能的不同因素。在目前的状态下,该数据集由一组高质量的训练图像和一组近13.5万张智能手机捕获的30种杂货产品的测试图像组成。在本文中,我们将讨论在用户持有的对象的视觉对象识别方面的一些开放挑战。我们评估了一些流行的对象识别算法的性能,具有不同的复杂程度,当测试SHORT时。
{"title":"A dataset for Hand-Held Object Recognition","authors":"Jose Rivera-Rubio, Saad Idrees, I. Alexiou, Lucas Hadjilucas, A. Bharath","doi":"10.1109/ICIP.2014.7026188","DOIUrl":"https://doi.org/10.1109/ICIP.2014.7026188","url":null,"abstract":"Visual object recognition is just one of the many applications of camera-equipped smartphones. The ability to recognise objects through photos taken with wearable and handheld cameras is already possible through some of the larger internet search providers; yet, there is little rigorous analysis of the quality of search results, particularly where there is great disparity in image quality. This has motivated us to develop the Small Hand-held Object Recognition Test (SHORT). This includes a dataset that is suitable for recognising hand-held objects from either snapshots or videos acquired using hand-held or wearable cameras. SHORT provides a collection of images and ground truth that help evaluate the different factors that affect recognition performance. At its present state, the dataset is comprised of a set of high quality training images and a large set of nearly 135,000 smartphone-captured test images of 30 grocery products. In this paper, we will discuss some open challenges in the visual object recognition of objects that are being held by users. We evaluate the performance of a number of popular object recognition algorithms, with differing levels of complexity, when tested against SHORT.","PeriodicalId":6856,"journal":{"name":"2014 IEEE International Conference on Image Processing (ICIP)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2014-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"76468036","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Collaborative inter-prediction on CPU+GPU systems CPU+GPU系统协同互预测
Pub Date : 2014-10-01 DOI: 10.1109/ICIP.2014.7025245
S. Momcilovic, A. Ilic, N. Roma, L. Sousa
In this paper we propose an efficient method for collaborative H.264/AVC inter-prediction in heterogeneous CPU+GPU systems. In order to minimize the overall encoding time, the proposed method provides stable and balanced load distribution of the most computationally demanding video encoding modules, by relying on accurate and dynamically built functional performance models. In an extensive RD analysis, an efficient temporary dependent prediction of the search area center is proposed, which allows dependency-aware workload partitioning and efficient GPU parallelization, while preserving high compression efficiency. The proposed method also introduces efficient communication-aware techniques, which maximize data reusing, and decrease the overhead of expensive data transfers in collaborative video encoding. The experimental results show that the proposed method is able of achieving real-time video encoding for very demanding video coding parameters, i.e. full HD video format, 64×64 pixels search area and the exhaustive motion estimation.
本文提出了一种高效的异构CPU+GPU系统协同H.264/AVC互预测方法。为了最大限度地减少整体编码时间,该方法通过精确和动态构建的功能性能模型,为计算量最大的视频编码模块提供稳定和均衡的负载分配。在广泛的RD分析中,提出了一种有效的搜索区域中心临时依赖预测,该预测允许依赖感知的工作负载分区和高效的GPU并行化,同时保持较高的压缩效率。该方法还引入了高效的通信感知技术,最大限度地提高了数据重用,降低了协作视频编码中昂贵的数据传输开销。实验结果表明,该方法能够在全高清视频格式、64×64像素搜索面积和穷举运动估计等对视频编码参数要求很高的情况下实现实时视频编码。
{"title":"Collaborative inter-prediction on CPU+GPU systems","authors":"S. Momcilovic, A. Ilic, N. Roma, L. Sousa","doi":"10.1109/ICIP.2014.7025245","DOIUrl":"https://doi.org/10.1109/ICIP.2014.7025245","url":null,"abstract":"In this paper we propose an efficient method for collaborative H.264/AVC inter-prediction in heterogeneous CPU+GPU systems. In order to minimize the overall encoding time, the proposed method provides stable and balanced load distribution of the most computationally demanding video encoding modules, by relying on accurate and dynamically built functional performance models. In an extensive RD analysis, an efficient temporary dependent prediction of the search area center is proposed, which allows dependency-aware workload partitioning and efficient GPU parallelization, while preserving high compression efficiency. The proposed method also introduces efficient communication-aware techniques, which maximize data reusing, and decrease the overhead of expensive data transfers in collaborative video encoding. The experimental results show that the proposed method is able of achieving real-time video encoding for very demanding video coding parameters, i.e. full HD video format, 64×64 pixels search area and the exhaustive motion estimation.","PeriodicalId":6856,"journal":{"name":"2014 IEEE International Conference on Image Processing (ICIP)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2014-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"76508724","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Vanishing point estimation for challenging road images 挑战性道路图像的消失点估计
Pub Date : 2014-10-01 DOI: 10.1109/ICIP.2014.7025200
Qingyun She, Zongqing Lu, Q. Liao
In this paper, we present an efficient vanishing point detection method for challenging road images. This detection process is based on the geometrical features of the roads. The slope distribution of the line segments is analyzed to reduce the spurious lines. A distance-based weighting scheme is also utilized to eliminate the voting noise in the voting stage. The proposed algorithm has been tested on a natural data set from Defense Advanced Research Projects Agency (DARPA). Experimental results with both quantitative and qualitative analyses are provided, which demonstrate the superiority of the proposed method over some state-of-the-art methods.
本文提出了一种有效的道路图像消失点检测方法。这种检测过程是基于道路的几何特征。分析了线段的斜率分布,减少了伪线的产生。为了消除投票阶段的投票噪声,还采用了基于距离的加权方案。该算法已在美国国防高级研究计划局(DARPA)的自然数据集上进行了测试。定量和定性分析的实验结果表明,该方法优于现有的一些方法。
{"title":"Vanishing point estimation for challenging road images","authors":"Qingyun She, Zongqing Lu, Q. Liao","doi":"10.1109/ICIP.2014.7025200","DOIUrl":"https://doi.org/10.1109/ICIP.2014.7025200","url":null,"abstract":"In this paper, we present an efficient vanishing point detection method for challenging road images. This detection process is based on the geometrical features of the roads. The slope distribution of the line segments is analyzed to reduce the spurious lines. A distance-based weighting scheme is also utilized to eliminate the voting noise in the voting stage. The proposed algorithm has been tested on a natural data set from Defense Advanced Research Projects Agency (DARPA). Experimental results with both quantitative and qualitative analyses are provided, which demonstrate the superiority of the proposed method over some state-of-the-art methods.","PeriodicalId":6856,"journal":{"name":"2014 IEEE International Conference on Image Processing (ICIP)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2014-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"77390017","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
Generic segment-wise DC for 3D-HEVC depth intra coding 用于3D-HEVC深度内部编码的通用分段DC
Pub Date : 2014-10-01 DOI: 10.1109/ICIP.2014.7025651
Hongbin Liu, Ying Chen
In 3D extension of HEVC (High Efficiency Video Coding), namely, 3D-HEVC, segment-wise DC coding (SDC) was adopted to more efficiently represent the depth residual for Intra coded depth blocks. Instead of coding pixel-wise residual as in HEVC, SDC codes one DC residual value for each segment of a Prediction Unit (PU) and skips transform and quantization. SDC was originally proposed for only a couple of modes, including the DC mode, Planar mode and depth modeling mode (DMM), which has an arbitrary straight line separation of a PU. This paper proposes a generic SDC method that applies to the conventional angular Intra modes. For each depth prediction unit coded with Intra prediction mode, encoder can adaptively choose to code pixel-wise residual or segment-wise residual to achieve better compression efficiency. Experimental results show that proposed method can reduce the total bit rate by about 1% even though the depth views altogether consumes relatively low percentage of the total bit rate.
在HEVC (High Efficiency Video Coding)的3D扩展中,即3D-HEVC,采用分段DC编码(segwise DC Coding, SDC)更有效地表示Intra编码深度块的深度残差。与HEVC中逐像素的残差编码不同,SDC为预测单元(PU)的每个片段编码一个DC残差值,并跳过变换和量化。SDC最初只提出了几种模式,包括DC模式、Planar模式和深度建模模式(DMM),其中PU具有任意直线分离。本文提出了一种适用于常规角度内模态的通用SDC方法。对于使用Intra预测模式编码的每个深度预测单元,编码器可以自适应选择编码逐像素残差或逐段残差,以获得更好的压缩效率。实验结果表明,尽管深度视图占用总比特率的百分比相对较低,但该方法仍能将总比特率降低约1%。
{"title":"Generic segment-wise DC for 3D-HEVC depth intra coding","authors":"Hongbin Liu, Ying Chen","doi":"10.1109/ICIP.2014.7025651","DOIUrl":"https://doi.org/10.1109/ICIP.2014.7025651","url":null,"abstract":"In 3D extension of HEVC (High Efficiency Video Coding), namely, 3D-HEVC, segment-wise DC coding (SDC) was adopted to more efficiently represent the depth residual for Intra coded depth blocks. Instead of coding pixel-wise residual as in HEVC, SDC codes one DC residual value for each segment of a Prediction Unit (PU) and skips transform and quantization. SDC was originally proposed for only a couple of modes, including the DC mode, Planar mode and depth modeling mode (DMM), which has an arbitrary straight line separation of a PU. This paper proposes a generic SDC method that applies to the conventional angular Intra modes. For each depth prediction unit coded with Intra prediction mode, encoder can adaptively choose to code pixel-wise residual or segment-wise residual to achieve better compression efficiency. Experimental results show that proposed method can reduce the total bit rate by about 1% even though the depth views altogether consumes relatively low percentage of the total bit rate.","PeriodicalId":6856,"journal":{"name":"2014 IEEE International Conference on Image Processing (ICIP)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2014-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"77569444","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 27
Retrieving images using saliency detection and graph matching 使用显著性检测和图匹配检索图像
Pub Date : 2014-10-01 DOI: 10.1109/ICIP.2014.7025624
Shao Huang, Weiqiang Wang, Hui Zhang
The need for fast retrieving images has recently increased tremendously in many application areas (biomedicine, military, commerce, education, etc.). In this work, we exploit the saliency detection to select a group of salient regions and utilize an undirected graph to model the dependency among these salient regions, so that the similarity of images can be measured by calculating the similarity of the corresponding graphs. Identification of salient pixels can decrease interferences from irrelevant information, and make the image representation more effective. The introduction of the graph model can better characterize the spatial constraints among salient regions. The comparison experiments are carried out on the three representative datasets publicly available (Holidays, UKB, and Oxford 5k), and the experimental results show that the integration of the proposed method and the SIFT-like local descriptors can better improve the existing state-of-the-art retrieval accuracy.
最近,在许多应用领域(生物医学、军事、商业、教育等),对快速检索图像的需求急剧增加。在这项工作中,我们利用显著性检测来选择一组显著区域,并利用无向图来建模这些显著区域之间的依赖关系,从而通过计算相应图的相似度来衡量图像的相似度。显著像素的识别可以减少不相关信息的干扰,使图像表示更有效。图模型的引入可以更好地表征显著区域之间的空间约束。对比实验结果表明,将所提方法与类sift局部描述符相结合,能够较好地提高现有的最优检索精度。
{"title":"Retrieving images using saliency detection and graph matching","authors":"Shao Huang, Weiqiang Wang, Hui Zhang","doi":"10.1109/ICIP.2014.7025624","DOIUrl":"https://doi.org/10.1109/ICIP.2014.7025624","url":null,"abstract":"The need for fast retrieving images has recently increased tremendously in many application areas (biomedicine, military, commerce, education, etc.). In this work, we exploit the saliency detection to select a group of salient regions and utilize an undirected graph to model the dependency among these salient regions, so that the similarity of images can be measured by calculating the similarity of the corresponding graphs. Identification of salient pixels can decrease interferences from irrelevant information, and make the image representation more effective. The introduction of the graph model can better characterize the spatial constraints among salient regions. The comparison experiments are carried out on the three representative datasets publicly available (Holidays, UKB, and Oxford 5k), and the experimental results show that the integration of the proposed method and the SIFT-like local descriptors can better improve the existing state-of-the-art retrieval accuracy.","PeriodicalId":6856,"journal":{"name":"2014 IEEE International Conference on Image Processing (ICIP)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2014-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"77621207","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
Primal-dual first order methods for total variation image restoration in presence of poisson noise 泊松噪声下全变分图像恢复的原始-对偶一阶方法
Pub Date : 2014-10-01 DOI: 10.1109/ICIP.2014.7025844
S. Bonettini, A. Benfenati, V. Ruggiero
Image restoration often requires the minimization of a convex, possibly nonsmooth functional, given by the sum of a data fidelity measure plus a regularization term. In order to face the lack of smoothness, alternative formulations of the minimization problem could be exploited via the duality principle. Indeed, the primal-dual and the dual formulation have been well explored in the literature when the data suffer from Gaussian noise and, thus, the data fidelity term is quadratic. Unfortunately, the most part of the approaches proposed for the Gaussian are difficult to apply to general data discrepancy terms, such as the Kullback-Leibler divergence. In this work we propose primal-dual methods which apply to the minimization of the sum of general convex functions and whose iteration is easy to compute, regardless of the form of the objective function, since it essentially consists in a subgradient projection step. We provide the convergence analysis and we suggest some strategies to improve the convergence speed by means of a careful selection of the steplength parameters. A numerical experience on Total Variation based denoising and deblurring problems from Poisson data shows the behavior of the proposed method with respect to other state-of-the-art algorithms.
图像恢复通常需要最小化凸函数,可能是非光滑函数,由数据保真度度量加上正则化项的总和给出。为了解决光滑性不足的问题,可以利用对偶原理开发最小化问题的备选公式。事实上,当数据受到高斯噪声的影响,因此数据保真度项是二次项时,文献中已经很好地探讨了原始对偶和对偶公式。不幸的是,大多数针对高斯分布提出的方法很难应用于一般的数据差异项,比如Kullback-Leibler散度。在这项工作中,我们提出了适用于一般凸函数和的最小化的原始对偶方法,其迭代易于计算,无论目标函数的形式如何,因为它本质上由一个次梯度投影步骤组成。给出了收敛性分析,并通过对步长参数的选择提出了提高收敛速度的策略。基于全变分的泊松数据去噪和去模糊问题的数值经验表明,该方法相对于其他最先进的算法具有良好的性能。
{"title":"Primal-dual first order methods for total variation image restoration in presence of poisson noise","authors":"S. Bonettini, A. Benfenati, V. Ruggiero","doi":"10.1109/ICIP.2014.7025844","DOIUrl":"https://doi.org/10.1109/ICIP.2014.7025844","url":null,"abstract":"Image restoration often requires the minimization of a convex, possibly nonsmooth functional, given by the sum of a data fidelity measure plus a regularization term. In order to face the lack of smoothness, alternative formulations of the minimization problem could be exploited via the duality principle. Indeed, the primal-dual and the dual formulation have been well explored in the literature when the data suffer from Gaussian noise and, thus, the data fidelity term is quadratic. Unfortunately, the most part of the approaches proposed for the Gaussian are difficult to apply to general data discrepancy terms, such as the Kullback-Leibler divergence. In this work we propose primal-dual methods which apply to the minimization of the sum of general convex functions and whose iteration is easy to compute, regardless of the form of the objective function, since it essentially consists in a subgradient projection step. We provide the convergence analysis and we suggest some strategies to improve the convergence speed by means of a careful selection of the steplength parameters. A numerical experience on Total Variation based denoising and deblurring problems from Poisson data shows the behavior of the proposed method with respect to other state-of-the-art algorithms.","PeriodicalId":6856,"journal":{"name":"2014 IEEE International Conference on Image Processing (ICIP)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2014-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"77715713","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
Who is my parent? Reconstructing video sequences from partially matching shots 谁是我的父母?从部分匹配的镜头重建视频序列
Pub Date : 2014-10-01 DOI: 10.1109/ICIP.2014.7026081
S. Lameri, Paolo Bestagini, A. Melloni, S. Milani, A. Rocha, M. Tagliasacchi, S. Tubaro
Nowadays, a significant fraction of the available video content is created by reusing already existing online videos. In these cases, the source video is seldom reused as is. Conversely, it is typically time clipped to extract only a subset of the original frames, and other transformations are commonly applied (e.g., cropping, logo insertion, etc.). In this paper, we analyze a pool of videos related to the same event or topic. We propose a method that aims at automatically reconstructing the content of the original source videos, i.e., the parent sequences, by splicing together sets of near-duplicate shots seemingly extracted from the same parent sequence. The result of the analysis shows how content is reused, thus revealing the intent of content creators, and enables us to reconstruct a parent sequence also when it is no longer available online. In doing so, we make use of a robust-hash algorithm that allows us to detect whether groups of frames are near-duplicates. Based on that, we developed an algorithm to automatically find near-duplicate matchings between multiple parts of multiple sequences. All the near-duplicate parts are finally temporally aligned to reconstruct the parent sequence. The proposed method is validated with both synthetic and real world datasets downloaded from YouTube.
如今,可用视频内容的很大一部分是通过重用已经存在的在线视频创建的。在这些情况下,源视频很少按原样重复使用。相反,它通常是时间剪辑,只提取原始帧的子集,并且通常应用其他转换(例如,裁剪,徽标插入等)。在本文中,我们分析了与同一事件或主题相关的视频池。我们提出了一种方法,旨在自动重建原始源视频的内容,即父序列,通过将看似从同一父序列中提取的近重复镜头拼接在一起。分析的结果显示了内容是如何被重用的,从而揭示了内容创建者的意图,并使我们能够在不再在线时重建父序列。在此过程中,我们使用了一种鲁棒哈希算法,该算法允许我们检测帧组是否接近重复。在此基础上,我们开发了一种算法来自动查找多个序列的多个部分之间的近重复匹配。最后对所有接近重复的部分进行临时对齐,以重建父序列。该方法通过从YouTube下载的合成数据集和真实数据集进行了验证。
{"title":"Who is my parent? Reconstructing video sequences from partially matching shots","authors":"S. Lameri, Paolo Bestagini, A. Melloni, S. Milani, A. Rocha, M. Tagliasacchi, S. Tubaro","doi":"10.1109/ICIP.2014.7026081","DOIUrl":"https://doi.org/10.1109/ICIP.2014.7026081","url":null,"abstract":"Nowadays, a significant fraction of the available video content is created by reusing already existing online videos. In these cases, the source video is seldom reused as is. Conversely, it is typically time clipped to extract only a subset of the original frames, and other transformations are commonly applied (e.g., cropping, logo insertion, etc.). In this paper, we analyze a pool of videos related to the same event or topic. We propose a method that aims at automatically reconstructing the content of the original source videos, i.e., the parent sequences, by splicing together sets of near-duplicate shots seemingly extracted from the same parent sequence. The result of the analysis shows how content is reused, thus revealing the intent of content creators, and enables us to reconstruct a parent sequence also when it is no longer available online. In doing so, we make use of a robust-hash algorithm that allows us to detect whether groups of frames are near-duplicates. Based on that, we developed an algorithm to automatically find near-duplicate matchings between multiple parts of multiple sequences. All the near-duplicate parts are finally temporally aligned to reconstruct the parent sequence. The proposed method is validated with both synthetic and real world datasets downloaded from YouTube.","PeriodicalId":6856,"journal":{"name":"2014 IEEE International Conference on Image Processing (ICIP)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2014-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"77815042","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 22
Handwritten word spotting based on a hybrid optimal distance 基于混合最优距离的手写单词识别
Pub Date : 2014-10-01 DOI: 10.1109/ICIP.2014.7025522
P. Wang, V. Eglin, C. Largeron, Christophe Garcia
In this paper, we develop a comprehensive representation model for handwriting, which contains both morphological and topological information. An adapted Shape Context descriptor built on structural points is employed to describe the contour of the text. Graphs are first constructed by using the structural points as nodes and the skeleton of the strokes as edges. Based on graphs, Topological Node Features (TNFs) of n-neighbourhood are extracted. Bag-of-Words representation model based on the TNFs is employed to depict the topological characteristics of word images. Moreover, a novel approach for word spotting application by using the proposed model is presented. The final distance is a weighted mixture of the SC cost, and the TNF distribution comparison. Linear Discriminant Analysis (LDA) is used to learn the optimal weight for each part of the distance with the consideration of writing styles. The evaluation of the proposed approach shows the significance of combining the properties of the handwriting from different aspects.
在本文中,我们开发了一个包含形态学和拓扑信息的手写体的综合表示模型。采用基于结构点的自适应形状上下文描述符来描述文本的轮廓。图首先使用结构点作为节点,笔画的骨架作为边来构造。基于图提取n邻域的拓扑节点特征(tnf)。采用基于tnf的词袋表示模型来描述词图像的拓扑特征。此外,本文还提出了一种新的单词识别方法。最终距离是SC成本和TNF分布比较的加权混合物。使用线性判别分析(LDA)在考虑写作风格的情况下,学习距离各部分的最优权重。通过对该方法的评价,可以看出综合不同笔迹特征的重要性。
{"title":"Handwritten word spotting based on a hybrid optimal distance","authors":"P. Wang, V. Eglin, C. Largeron, Christophe Garcia","doi":"10.1109/ICIP.2014.7025522","DOIUrl":"https://doi.org/10.1109/ICIP.2014.7025522","url":null,"abstract":"In this paper, we develop a comprehensive representation model for handwriting, which contains both morphological and topological information. An adapted Shape Context descriptor built on structural points is employed to describe the contour of the text. Graphs are first constructed by using the structural points as nodes and the skeleton of the strokes as edges. Based on graphs, Topological Node Features (TNFs) of n-neighbourhood are extracted. Bag-of-Words representation model based on the TNFs is employed to depict the topological characteristics of word images. Moreover, a novel approach for word spotting application by using the proposed model is presented. The final distance is a weighted mixture of the SC cost, and the TNF distribution comparison. Linear Discriminant Analysis (LDA) is used to learn the optimal weight for each part of the distance with the consideration of writing styles. The evaluation of the proposed approach shows the significance of combining the properties of the handwriting from different aspects.","PeriodicalId":6856,"journal":{"name":"2014 IEEE International Conference on Image Processing (ICIP)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2014-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"79951751","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Automatic fault tracking across seismic volumes via tracking vectors 基于跟踪向量的地震体断层自动跟踪
Pub Date : 2014-10-01 DOI: 10.1109/ICIP.2014.7026182
Zhen Wang, Z. Long, G. Al-Regib, Asjad Amin, Mohamed Deriche
The identification of reservoir regions has a close relationship with the detection of faults in seismic volumes. However, only relying on human intervention, most fault detection algorithms are inefficient. In this paper, we present a new technique that automatically tracks faults across a 3D seismic volume. To implement automation, we propose a two-way fault line projection based on estimated tracking vectors. In the tracking process, projected fault lines are integrated into a synthesized line as the tracked fault line, through an optimization process with local geological constraints. The tracking algorithm is evaluated using real-world seismic data sets with promising results. The proposed method provides comparable accuracy to the detection of faults explicitly in every seismic section, and it also reduces computational complexity.
储层的识别与地震体断层的检测有着密切的关系。然而,大多数故障检测算法仅依靠人工干预,效率低下。在本文中,我们提出了一种自动跟踪三维地震体断层的新技术。为了实现自动化,我们提出了一种基于估计跟踪向量的双向故障线投影。在跟踪过程中,通过局部地质约束的优化过程,将投影断层线整合成一条综合断层线作为跟踪断层线。利用实际地震数据集对跟踪算法进行了评估,结果令人满意。该方法不仅具有相当的精度,而且降低了计算复杂度。
{"title":"Automatic fault tracking across seismic volumes via tracking vectors","authors":"Zhen Wang, Z. Long, G. Al-Regib, Asjad Amin, Mohamed Deriche","doi":"10.1109/ICIP.2014.7026182","DOIUrl":"https://doi.org/10.1109/ICIP.2014.7026182","url":null,"abstract":"The identification of reservoir regions has a close relationship with the detection of faults in seismic volumes. However, only relying on human intervention, most fault detection algorithms are inefficient. In this paper, we present a new technique that automatically tracks faults across a 3D seismic volume. To implement automation, we propose a two-way fault line projection based on estimated tracking vectors. In the tracking process, projected fault lines are integrated into a synthesized line as the tracked fault line, through an optimization process with local geological constraints. The tracking algorithm is evaluated using real-world seismic data sets with promising results. The proposed method provides comparable accuracy to the detection of faults explicitly in every seismic section, and it also reduces computational complexity.","PeriodicalId":6856,"journal":{"name":"2014 IEEE International Conference on Image Processing (ICIP)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2014-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"79071717","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 17
Architecture of a socio-conversational agent in virtual worlds 虚拟世界中社会会话代理的体系结构
Pub Date : 2014-10-01 DOI: 10.1109/ICIP.2014.7025809
Brian Ravenet, M. Ochs, C. Pelachaud
Virtual worlds are more and more populated with autonomous conversational agents embodying different roles like tutor, guide, or personal assistant. In order to create more engaging and natural interactions, these agents should be endowed with social capabilities such as expressing different social attitudes through their behaviors. In this paper, we present the architecture of a socio-conversational agent composed of communicative components to detect and respond verbally and non-verbally to the user's speech and to convey different social attitudes. This paper presents the main components of this architecture. These descrpitions are illustrated with scenarios of interaction.
虚拟世界中越来越多的自主对话代理,它们扮演着不同的角色,如导师、向导或个人助理。为了创造更有吸引力和自然的互动,这些代理应该被赋予社会能力,例如通过他们的行为表达不同的社会态度。在本文中,我们提出了一个由交际组件组成的社会会话代理的架构,用于检测和响应用户的口头和非口头语言,并传达不同的社会态度。本文介绍了该体系结构的主要组成部分。这些描述用交互场景来说明。
{"title":"Architecture of a socio-conversational agent in virtual worlds","authors":"Brian Ravenet, M. Ochs, C. Pelachaud","doi":"10.1109/ICIP.2014.7025809","DOIUrl":"https://doi.org/10.1109/ICIP.2014.7025809","url":null,"abstract":"Virtual worlds are more and more populated with autonomous conversational agents embodying different roles like tutor, guide, or personal assistant. In order to create more engaging and natural interactions, these agents should be endowed with social capabilities such as expressing different social attitudes through their behaviors. In this paper, we present the architecture of a socio-conversational agent composed of communicative components to detect and respond verbally and non-verbally to the user's speech and to convey different social attitudes. This paper presents the main components of this architecture. These descrpitions are illustrated with scenarios of interaction.","PeriodicalId":6856,"journal":{"name":"2014 IEEE International Conference on Image Processing (ICIP)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2014-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"79287407","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
2014 IEEE International Conference on Image Processing (ICIP)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1