首页 > 最新文献

2011 IEEE 10th IVMSP Workshop: Perception and Visual Signal Analysis最新文献

英文 中文
Towards a subject-independent adaptive pupil tracker for automatic eye tracking calibration using a mixture model 基于混合模型的独立于主体的自适应瞳孔自动眼动跟踪标定
Pub Date : 2011-06-16 DOI: 10.1109/IVMSPW.2011.5970369
Thomas B. Kinsman, J. Pelz
This paper describes the initial pre-processing steps used to follow the motions of the human eye in an eye tracking application. The central method models each pixel as a combination of either: a dark pupil pixel, bright highlight pixel, or a neutral pixel. Portable eye tracking involves tracking a subject's pupil over the course of a study. This paper describes very preliminary results from using a mixture model as a processing stage. Technical issues of using a mixture model are discussed. The pixel classifications from the mixture model were fed into a naïve Bayes pupil tracker. Only low-level information is used for pupil identification. No motion tracking is performed, no belief propagation is performed, and no convolutions are computed. The algorithm is well positioned for parallel implementations. The solution surmounts several technical challenges, and initial results are unexpectedly accurate. The technique shows good promise for incorporation into a system for automatic eye-to-scene calibration.
本文描述了在眼动追踪应用程序中用于跟踪人眼运动的初始预处理步骤。中心方法将每个像素建模为:暗瞳像素,亮高光像素或中性像素的组合。便携式眼动追踪技术包括在研究过程中跟踪受试者的瞳孔。本文描述了使用混合模型作为处理阶段的非常初步的结果。讨论了使用混合模型的技术问题。混合模型的像素分类输入naïve贝叶斯瞳孔跟踪器。只有低级别的信息被用于瞳孔识别。不执行运动跟踪,不执行信念传播,也不计算卷积。该算法很适合并行实现。该解决方案克服了几个技术挑战,并且最初的结果出乎意料地准确。该技术在眼到场景自动校准系统中显示出良好的应用前景。
{"title":"Towards a subject-independent adaptive pupil tracker for automatic eye tracking calibration using a mixture model","authors":"Thomas B. Kinsman, J. Pelz","doi":"10.1109/IVMSPW.2011.5970369","DOIUrl":"https://doi.org/10.1109/IVMSPW.2011.5970369","url":null,"abstract":"This paper describes the initial pre-processing steps used to follow the motions of the human eye in an eye tracking application. The central method models each pixel as a combination of either: a dark pupil pixel, bright highlight pixel, or a neutral pixel. Portable eye tracking involves tracking a subject's pupil over the course of a study. This paper describes very preliminary results from using a mixture model as a processing stage. Technical issues of using a mixture model are discussed. The pixel classifications from the mixture model were fed into a naïve Bayes pupil tracker. Only low-level information is used for pupil identification. No motion tracking is performed, no belief propagation is performed, and no convolutions are computed. The algorithm is well positioned for parallel implementations. The solution surmounts several technical challenges, and initial results are unexpectedly accurate. The technique shows good promise for incorporation into a system for automatic eye-to-scene calibration.","PeriodicalId":405588,"journal":{"name":"2011 IEEE 10th IVMSP Workshop: Perception and Visual Signal Analysis","volume":"160 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-06-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133021514","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Despeckling trilateral filter 去斑三边过滤器
Pub Date : 2011-06-16 DOI: 10.1109/IVMSPW.2011.5970352
Yongjian Yu, Gang Dong, Jue Wang
The bilateral filter smoothes noisy signals while preserving the semantic signal features. Its main advantage is being non-iterative. It is effective for a variety of applications in computer vision and computer graphics. However, little is known about the usefulness of bilateral filtering for speckle images. We propose a non-iterative, despeckling trilateral filter (DSTF) for smoothing ultrasound or synthetic aperture radar imagery. This filter combines the spatial closeness, intensity similarity and the coefficient of variation component. It generates outputs with speckle regions smoothed and structural features well preserved. The performance of the method is illustrated using synthetic, ultrasound and radar images. We show that the DSTF improves the bilateral filter with better speckle suppression, and is more computational efficient than the heavily iterative speckle reducing anisotropic diffusion.
双边滤波器在保持信号语义特征的同时对噪声信号进行平滑处理。它的主要优点是非迭代的。它对计算机视觉和计算机图形学的各种应用都是有效的。然而,关于双侧滤波对散斑图像的有用性知之甚少。我们提出了一个非迭代,去斑三边滤波器(DSTF)平滑超声或合成孔径雷达图像。该滤波器结合了空间接近度、强度相似度和变异系数分量。它生成的输出具有平滑的散斑区域和良好的结构特征。用合成图像、超声图像和雷达图像说明了该方法的性能。我们表明,DSTF改进了双边滤波器,具有更好的散斑抑制,并且比重迭代散斑减少各向异性扩散的计算效率更高。
{"title":"Despeckling trilateral filter","authors":"Yongjian Yu, Gang Dong, Jue Wang","doi":"10.1109/IVMSPW.2011.5970352","DOIUrl":"https://doi.org/10.1109/IVMSPW.2011.5970352","url":null,"abstract":"The bilateral filter smoothes noisy signals while preserving the semantic signal features. Its main advantage is being non-iterative. It is effective for a variety of applications in computer vision and computer graphics. However, little is known about the usefulness of bilateral filtering for speckle images. We propose a non-iterative, despeckling trilateral filter (DSTF) for smoothing ultrasound or synthetic aperture radar imagery. This filter combines the spatial closeness, intensity similarity and the coefficient of variation component. It generates outputs with speckle regions smoothed and structural features well preserved. The performance of the method is illustrated using synthetic, ultrasound and radar images. We show that the DSTF improves the bilateral filter with better speckle suppression, and is more computational efficient than the heavily iterative speckle reducing anisotropic diffusion.","PeriodicalId":405588,"journal":{"name":"2011 IEEE 10th IVMSP Workshop: Perception and Visual Signal Analysis","volume":"11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-06-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117083546","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Local masking in natural images measured via a new tree-structured forced-choice technique 通过一种新的树结构强制选择技术测量自然图像中的局部掩蔽
Pub Date : 2011-06-16 DOI: 10.1109/IVMSPW.2011.5970348
Kedarnath P. Vilankar, D. Chandler
It is widely known that natural images can hide or mask visual signals, and that this masking ability can vary across different regions of the image. Previous studies have quantified masking by measuring image-wide detection thresholds or local thresholds for select image regions; however, little effort has focused on measuring local thresholds across entire images so as to achieve ground-truth masking maps. Such maps could prove invaluable for testing and refining masking models; however, obtaining these maps requires a prohibitive number of trials using a traditional forced-choice procedure. Here, we present a tree-structured forced-choice procedure (TS-3AFC) designed to efficiently measure local thresholds across images. TS-3AFC requires fewer trials than normal forced-choice by employing recursive patch subdivision in which the child patches are not tested individually until the target is detectable in the parent patch. We show that TS-3AFC can yield masking maps which demonstrate both intrasubject and inter-subject repeatability, and we analyze the performance of a modern masking model and two quality estimators in predicting the obtained ground-truth maps for a small set of images.
众所周知,自然图像可以隐藏或掩盖视觉信号,并且这种掩盖能力可以在图像的不同区域变化。以前的研究通过测量图像范围的检测阈值或选定图像区域的局部阈值来量化掩蔽;然而,很少有人关注于测量整个图像的局部阈值,从而获得地面真值掩蔽图。这样的地图对于测试和改进掩蔽模型是非常宝贵的;然而,获得这些地图需要使用传统的强制选择程序进行大量的试验。在这里,我们提出了一个树结构的强制选择程序(TS-3AFC),旨在有效地测量图像之间的局部阈值。TS-3AFC通过采用递归贴片细分,比正常的强制选择需要更少的试验,其中子贴片不单独测试,直到目标在父贴片中可检测到。我们证明TS-3AFC可以产生具有主体内和主体间可重复性的掩蔽图,并且我们分析了现代掩蔽模型和两个质量估计器在预测一小组图像获得的地面真值图方面的性能。
{"title":"Local masking in natural images measured via a new tree-structured forced-choice technique","authors":"Kedarnath P. Vilankar, D. Chandler","doi":"10.1109/IVMSPW.2011.5970348","DOIUrl":"https://doi.org/10.1109/IVMSPW.2011.5970348","url":null,"abstract":"It is widely known that natural images can hide or mask visual signals, and that this masking ability can vary across different regions of the image. Previous studies have quantified masking by measuring image-wide detection thresholds or local thresholds for select image regions; however, little effort has focused on measuring local thresholds across entire images so as to achieve ground-truth masking maps. Such maps could prove invaluable for testing and refining masking models; however, obtaining these maps requires a prohibitive number of trials using a traditional forced-choice procedure. Here, we present a tree-structured forced-choice procedure (TS-3AFC) designed to efficiently measure local thresholds across images. TS-3AFC requires fewer trials than normal forced-choice by employing recursive patch subdivision in which the child patches are not tested individually until the target is detectable in the parent patch. We show that TS-3AFC can yield masking maps which demonstrate both intrasubject and inter-subject repeatability, and we analyze the performance of a modern masking model and two quality estimators in predicting the obtained ground-truth maps for a small set of images.","PeriodicalId":405588,"journal":{"name":"2011 IEEE 10th IVMSP Workshop: Perception and Visual Signal Analysis","volume":"13 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-06-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123677195","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
View-based modelling of human visual navigation errors 基于视图的人类视觉导航误差建模
Pub Date : 2011-06-16 DOI: 10.1109/IVMSPW.2011.5970368
L. Pickup, A. Fitzgibbon, S. Gilson, A. Glennerster
View-based and Cartesian representations provide rival accounts of visual navigation in humans, and here we explore possible models for the view-based case. A visual “homing” experiment was under-taken by human participants in immersive virtual reality. The distributions of end-point errors on the ground plane differed significantly in shape and extent depending on visual landmark configuration and relative goal location. A model based on simple visual cues captures important characteristics of these distributions. Augmenting visual features to include 3D elements such as stereo and motion parallax result in a set of models that describe the data accurately, demonstrating the effectiveness of a view-based approach.
基于视图的表示和笛卡尔表示提供了人类视觉导航的竞争帐户,在这里,我们探索基于视图的情况下可能的模型。人类参与者在沉浸式虚拟现实中进行了视觉“归巢”实验。终点误差在地平面上的分布在形状和程度上因视觉地标配置和相对目标位置的不同而有显著差异。基于简单视觉线索的模型可以捕捉到这些分布的重要特征。增强视觉特征,包括立体和运动视差等3D元素,产生一组准确描述数据的模型,证明了基于视图的方法的有效性。
{"title":"View-based modelling of human visual navigation errors","authors":"L. Pickup, A. Fitzgibbon, S. Gilson, A. Glennerster","doi":"10.1109/IVMSPW.2011.5970368","DOIUrl":"https://doi.org/10.1109/IVMSPW.2011.5970368","url":null,"abstract":"View-based and Cartesian representations provide rival accounts of visual navigation in humans, and here we explore possible models for the view-based case. A visual “homing” experiment was under-taken by human participants in immersive virtual reality. The distributions of end-point errors on the ground plane differed significantly in shape and extent depending on visual landmark configuration and relative goal location. A model based on simple visual cues captures important characteristics of these distributions. Augmenting visual features to include 3D elements such as stereo and motion parallax result in a set of models that describe the data accurately, demonstrating the effectiveness of a view-based approach.","PeriodicalId":405588,"journal":{"name":"2011 IEEE 10th IVMSP Workshop: Perception and Visual Signal Analysis","volume":"11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-06-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123813981","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Dichromatic color perception in a two stage model: Testing for cone replacement and cone loss models 两阶段模型中的二色感知:锥体替换和锥体丢失模型的测试
Pub Date : 2011-06-16 DOI: 10.1109/IVMSPW.2011.5970347
C. E. Rodríguez-Pardo, Gaurav Sharma
We formulate a two stage model of dichromatic color perception that consists of a first sensor layer with gain control followed by an opponent encoding transformation. We propose a method for estimating the unknown parameters in the model by utilizing pre-existing data from psychophysical experiments on unilateral dichromats. The model is validated using this existing data and by using predictions on known test images for detecting dichromacy. Using the model and analysis we evaluate the feasibility of cone loss and cone replacement hypotheses that have previously been proposed for modeling dichromatic color vision. Results indicate that the two stage model offers good agreement with test data. The cone loss and cone replacement models are shown to have fundamental limitations in matching psychophysical observations.
我们制定了二色色彩感知的两阶段模型,该模型由具有增益控制的第一传感器层组成,然后是对手编码转换。我们提出了一种估算模型中未知参数的方法,该方法利用了单侧二色者心理物理实验中已有的数据。该模型通过使用现有数据和对已知检测二色的测试图像的预测进行验证。利用模型和分析,我们评估了之前提出的用于二色视觉建模的锥体丢失和锥体替换假设的可行性。结果表明,两阶段模型与试验数据吻合较好。锥体损失和锥体替换模型在匹配心理物理观察方面显示出根本性的局限性。
{"title":"Dichromatic color perception in a two stage model: Testing for cone replacement and cone loss models","authors":"C. E. Rodríguez-Pardo, Gaurav Sharma","doi":"10.1109/IVMSPW.2011.5970347","DOIUrl":"https://doi.org/10.1109/IVMSPW.2011.5970347","url":null,"abstract":"We formulate a two stage model of dichromatic color perception that consists of a first sensor layer with gain control followed by an opponent encoding transformation. We propose a method for estimating the unknown parameters in the model by utilizing pre-existing data from psychophysical experiments on unilateral dichromats. The model is validated using this existing data and by using predictions on known test images for detecting dichromacy. Using the model and analysis we evaluate the feasibility of cone loss and cone replacement hypotheses that have previously been proposed for modeling dichromatic color vision. Results indicate that the two stage model offers good agreement with test data. The cone loss and cone replacement models are shown to have fundamental limitations in matching psychophysical observations.","PeriodicalId":405588,"journal":{"name":"2011 IEEE 10th IVMSP Workshop: Perception and Visual Signal Analysis","volume":"86 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-06-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125868103","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
A fully automatic digital camera image refocusing algorithm 一种全自动数码相机图像调焦算法
Pub Date : 2011-06-16 DOI: 10.1109/IVMSPW.2011.5970359
J. E. Adams
One of the greatest dissatisfiers of consumer digital cameras is autofocus failure. One possible solution currently being investigated by many digital camera manufacturers involves capturing a sequence of through-focus images and postprocessing to produce a desired focused image. Most approaches require significant manual input to define a region of interest (ROI) to be optimized for focus. A new through-focus algorithm is proposed that automatically partitions scene content into regions based on range and then automatically determines the ROI for a given range. When this ROI is combined with standard image fusion operations, this algorithm generates an image with an aesthetically pleasing narrow depth-of-field effect, all with little or no user input, and all within the limited compute environment of a digital camera.
消费者对数码相机最大的不满之一是不能自动对焦。目前许多数码相机制造商正在研究的一种可能的解决方案包括捕捉一系列通过聚焦的图像并进行后处理以产生所需的聚焦图像。大多数方法需要大量的人工输入来定义要优化的关注区域(ROI)。提出了一种新的全聚焦算法,该算法基于距离将场景内容自动划分为区域,然后在给定范围内自动确定ROI。当该ROI与标准图像融合操作相结合时,该算法生成的图像具有美观的窄景深效果,所有这些都很少或没有用户输入,并且都在数字相机的有限计算环境中。
{"title":"A fully automatic digital camera image refocusing algorithm","authors":"J. E. Adams","doi":"10.1109/IVMSPW.2011.5970359","DOIUrl":"https://doi.org/10.1109/IVMSPW.2011.5970359","url":null,"abstract":"One of the greatest dissatisfiers of consumer digital cameras is autofocus failure. One possible solution currently being investigated by many digital camera manufacturers involves capturing a sequence of through-focus images and postprocessing to produce a desired focused image. Most approaches require significant manual input to define a region of interest (ROI) to be optimized for focus. A new through-focus algorithm is proposed that automatically partitions scene content into regions based on range and then automatically determines the ROI for a given range. When this ROI is combined with standard image fusion operations, this algorithm generates an image with an aesthetically pleasing narrow depth-of-field effect, all with little or no user input, and all within the limited compute environment of a digital camera.","PeriodicalId":405588,"journal":{"name":"2011 IEEE 10th IVMSP Workshop: Perception and Visual Signal Analysis","volume":"63 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-06-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127885017","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Visual attention model for target search in cluttered scene 杂乱场景下目标搜索的视觉注意模型
Pub Date : 2011-06-16 DOI: 10.1109/IVMSPW.2011.5970370
Nevrez Imamoglu, Weisi Lin
Visual attention models generate saliency maps in which attentive regions are more distinctive with respect to remaining parts of the scene. In this work, a new model of orientation conspicuity map (OCM) is presented for the computation of saliency. The proposed method is based on the difference of the Gabor filter outputs with orthogonal orientations because vehicles are the targets for the search tasks in this study. Moreover, as another contribution, selective resolution for the input image, according to the distance of the target in the scene, is also utilized with the proposed scheme for the benefit to target search. Experimental results demonstrate that both the OCM model and selective resolution for input images yield promising results for the target search in cluttered scenes.
视觉注意模型生成显著性地图,其中注意区域相对于场景的其余部分更具独特性。本文提出了一种新的方向显著性图(OCM)模型,用于显著性的计算。由于车辆是本研究中搜索任务的目标,因此提出的方法是基于正交方向的Gabor滤波器输出的差异。此外,该方案还根据目标在场景中的距离对输入图像进行了选择性分辨率,有利于目标搜索。实验结果表明,OCM模型和输入图像的选择性分辨率对混乱场景下的目标搜索都有很好的效果。
{"title":"Visual attention model for target search in cluttered scene","authors":"Nevrez Imamoglu, Weisi Lin","doi":"10.1109/IVMSPW.2011.5970370","DOIUrl":"https://doi.org/10.1109/IVMSPW.2011.5970370","url":null,"abstract":"Visual attention models generate saliency maps in which attentive regions are more distinctive with respect to remaining parts of the scene. In this work, a new model of orientation conspicuity map (OCM) is presented for the computation of saliency. The proposed method is based on the difference of the Gabor filter outputs with orthogonal orientations because vehicles are the targets for the search tasks in this study. Moreover, as another contribution, selective resolution for the input image, according to the distance of the target in the scene, is also utilized with the proposed scheme for the benefit to target search. Experimental results demonstrate that both the OCM model and selective resolution for input images yield promising results for the target search in cluttered scenes.","PeriodicalId":405588,"journal":{"name":"2011 IEEE 10th IVMSP Workshop: Perception and Visual Signal Analysis","volume":"49 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-06-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127943896","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Identification and discussion of open issues in perceptual video coding based on image analysis and completion 基于图像分析和补全的感知视频编码中开放性问题的识别和讨论
Pub Date : 2011-06-16 DOI: 10.1109/IVMSPW.2011.5970350
D. Doshkov, H. Kaprykowsky, P. Ndjiki-Nya
Perceptual video coding (VC) based on image analysis and completion (IAC) has enjoyed increasing awareness during the past few years. Many related approaches have been proposed that follow diverging strategies: from full compatibility to hybrid block transform coding to alternative codec design. Hence, in this paper, the most significant issues in IAC coding will be identified and their relevance for the IAC VC design highlighted. It will be analyzed where the most promising pathways lie and justified why others may be limited in their potentialities. Discussions will be substantiated using new methods developed by the authors for block-based and region-based IAC coding additionally to the state-of-the-art approaches.
在过去的几年里,基于图像分析和补全(IAC)的感知视频编码(VC)得到了越来越多的关注。已经提出了许多相关的方法,遵循不同的策略:从完全兼容到混合块转换编码到替代编解码器设计。因此,在本文中,将确定IAC编码中最重要的问题,并强调它们与IAC VC设计的相关性。它将分析最有希望的途径在哪里,并证明为什么其他人的潜力可能受到限制。讨论将使用作者为基于块和基于区域的IAC编码开发的新方法,以及最先进的方法来证实。
{"title":"Identification and discussion of open issues in perceptual video coding based on image analysis and completion","authors":"D. Doshkov, H. Kaprykowsky, P. Ndjiki-Nya","doi":"10.1109/IVMSPW.2011.5970350","DOIUrl":"https://doi.org/10.1109/IVMSPW.2011.5970350","url":null,"abstract":"Perceptual video coding (VC) based on image analysis and completion (IAC) has enjoyed increasing awareness during the past few years. Many related approaches have been proposed that follow diverging strategies: from full compatibility to hybrid block transform coding to alternative codec design. Hence, in this paper, the most significant issues in IAC coding will be identified and their relevance for the IAC VC design highlighted. It will be analyzed where the most promising pathways lie and justified why others may be limited in their potentialities. Discussions will be substantiated using new methods developed by the authors for block-based and region-based IAC coding additionally to the state-of-the-art approaches.","PeriodicalId":405588,"journal":{"name":"2011 IEEE 10th IVMSP Workshop: Perception and Visual Signal Analysis","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-06-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115172536","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Effects of texture on color perception 纹理对颜色感知的影响
Pub Date : 2011-06-16 DOI: 10.1109/IVMSPW.2011.5970346
H. Trussell, Juan Lin, R. Shamey
Textures are common distinguishing features used in segmentation and characterization of images. It is common to characterize textures in a statistical manner using various first and second order statistics. Previous work has shown that texture influences the observer's ability to perceive color differences. By considering the frequency content of the texture patterns in relationship to the color frequency response of the human eye, we hope to explain the results of some perceptual experiments in a more quantitative manner and lay a foundation for improved segmentation in computer vision applications.
纹理是用于图像分割和表征的常见区别特征。使用各种一阶和二阶统计量以统计方式表征纹理是常见的。先前的研究表明,纹理会影响观察者感知颜色差异的能力。通过考虑纹理图案的频率含量与人眼颜色频率响应的关系,我们希望能够更定量地解释一些感知实验的结果,为改进计算机视觉应用中的分割奠定基础。
{"title":"Effects of texture on color perception","authors":"H. Trussell, Juan Lin, R. Shamey","doi":"10.1109/IVMSPW.2011.5970346","DOIUrl":"https://doi.org/10.1109/IVMSPW.2011.5970346","url":null,"abstract":"Textures are common distinguishing features used in segmentation and characterization of images. It is common to characterize textures in a statistical manner using various first and second order statistics. Previous work has shown that texture influences the observer's ability to perceive color differences. By considering the frequency content of the texture patterns in relationship to the color frequency response of the human eye, we hope to explain the results of some perceptual experiments in a more quantitative manner and lay a foundation for improved segmentation in computer vision applications.","PeriodicalId":405588,"journal":{"name":"2011 IEEE 10th IVMSP Workshop: Perception and Visual Signal Analysis","volume":"12 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-06-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128022832","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Using human experts' gaze data to evaluate image processing algorithms 利用人类专家的注视数据来评估图像处理算法
Pub Date : 2011-06-16 DOI: 10.1109/IVMSPW.2011.5970367
Preethi Vaidyanathan, J. Pelz, Rui Li, Sai Mulpuru, Dong Wang, P. Shi, C. Calvelli, Anne R. Haake
Understanding the capabilities of the human visual system with respect to image understanding, in order to inform image processing, remains a challenge. Visual attention deployment strategies of experts can serve as an objective measure to help us understand their learned perceptual and conceptual processes. Understanding these processes will inform and direct image the selection and use of image processing algorithms, such as the dermatological images used in our study. The goal of our research is to extract and utilize the tacit knowledge of domain experts towards building a pipeline of image processing algorithms that could closely parallel the underlying cognitive processes. In this paper we use medical experts' eye movement data, primarily fixations, as a metric to evaluate the correlation of perceptually-relevant regions with individual clusters identified through k-means clustering. This test case demonstrates the potential of this approach to determine whether a particular image processing algorithm will be useful in identifying image regions with high visual interest and whether it could be a component of a processing pipeline.
了解人类视觉系统在图像理解方面的能力,以便为图像处理提供信息,仍然是一个挑战。专家的视觉注意部署策略可以作为一种客观的衡量标准,帮助我们理解他们学习的感知和概念过程。了解这些过程将为图像处理算法的选择和使用提供信息和指导,例如我们研究中使用的皮肤病图像。我们的研究目标是提取和利用领域专家的隐性知识来构建一个可以密切平行于底层认知过程的图像处理算法管道。在本文中,我们使用医学专家的眼动数据(主要是注视)作为度量来评估感知相关区域与通过k-means聚类识别的单个聚类之间的相关性。这个测试用例展示了这种方法的潜力,以确定特定的图像处理算法在识别具有高视觉兴趣的图像区域时是否有用,以及它是否可以成为处理管道的组成部分。
{"title":"Using human experts' gaze data to evaluate image processing algorithms","authors":"Preethi Vaidyanathan, J. Pelz, Rui Li, Sai Mulpuru, Dong Wang, P. Shi, C. Calvelli, Anne R. Haake","doi":"10.1109/IVMSPW.2011.5970367","DOIUrl":"https://doi.org/10.1109/IVMSPW.2011.5970367","url":null,"abstract":"Understanding the capabilities of the human visual system with respect to image understanding, in order to inform image processing, remains a challenge. Visual attention deployment strategies of experts can serve as an objective measure to help us understand their learned perceptual and conceptual processes. Understanding these processes will inform and direct image the selection and use of image processing algorithms, such as the dermatological images used in our study. The goal of our research is to extract and utilize the tacit knowledge of domain experts towards building a pipeline of image processing algorithms that could closely parallel the underlying cognitive processes. In this paper we use medical experts' eye movement data, primarily fixations, as a metric to evaluate the correlation of perceptually-relevant regions with individual clusters identified through k-means clustering. This test case demonstrates the potential of this approach to determine whether a particular image processing algorithm will be useful in identifying image regions with high visual interest and whether it could be a component of a processing pipeline.","PeriodicalId":405588,"journal":{"name":"2011 IEEE 10th IVMSP Workshop: Perception and Visual Signal Analysis","volume":"45 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-06-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133310523","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 14
期刊
2011 IEEE 10th IVMSP Workshop: Perception and Visual Signal Analysis
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1