首页 > 最新文献

2009 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops最新文献

英文 中文
Robust real-time 3D modeling of static scenes using solely a Time-of-Flight sensor 鲁棒的实时三维建模静态场景仅使用一个飞行时间传感器
J. Feulner, J. Penne, E. Kollorz, J. Hornegger
An algorithm is proposed for the 3D modeling of static scenes solely based on the range and intensity data acquired by a time-of-flight camera during an arbitrary movement. No additional scene acquisition devices, like inertia sensor, positioning robots or intensity based cameras are incorporated. The current pose is estimated by maximizing the uncentered correlation coefficient between edges detected in the current and a preceding frame at a minimum frame rate of four fps and an average accuracy of 45 mm. The paper also describes several extensions for robust registration like multiresolution hierarchies and projection Iterative Closest Point algorithm. The basic registration algorithm and its extensions were intensively evaluated against ground truth data to validate the accuracy, robustness and real-time-capability.
提出了一种仅基于任意运动时飞行时间相机获取的距离和强度数据的静态场景三维建模算法。没有额外的场景采集设备,如惯性传感器,定位机器人或基于强度的相机。当前姿态通过最大化当前帧和前一帧检测到的边缘之间的无中心相关系数来估计,最小帧速率为4 fps,平均精度为45 mm。本文还介绍了鲁棒配准的几个扩展,如多分辨率层次和投影迭代最近点算法。针对地面真实数据对基本配准算法及其扩展进行了深入评估,以验证其准确性、鲁棒性和实时性。
{"title":"Robust real-time 3D modeling of static scenes using solely a Time-of-Flight sensor","authors":"J. Feulner, J. Penne, E. Kollorz, J. Hornegger","doi":"10.1109/CVPRW.2009.5205204","DOIUrl":"https://doi.org/10.1109/CVPRW.2009.5205204","url":null,"abstract":"An algorithm is proposed for the 3D modeling of static scenes solely based on the range and intensity data acquired by a time-of-flight camera during an arbitrary movement. No additional scene acquisition devices, like inertia sensor, positioning robots or intensity based cameras are incorporated. The current pose is estimated by maximizing the uncentered correlation coefficient between edges detected in the current and a preceding frame at a minimum frame rate of four fps and an average accuracy of 45 mm. The paper also describes several extensions for robust registration like multiresolution hierarchies and projection Iterative Closest Point algorithm. The basic registration algorithm and its extensions were intensively evaluated against ground truth data to validate the accuracy, robustness and real-time-capability.","PeriodicalId":431981,"journal":{"name":"2009 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":"206 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-08-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115035570","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Transparent watermarking using bidirectional imaging 透明水印使用双向成像
Kristin J. Dana, G. Livescu, R. Makonahalli
We present a method for transparent watermarking using a custom bidirectional imaging device. The two innovative concepts of our approach are reflectance coding and multiview imaging. In reflectance coding, information is embedded in the angular space of the bidirectional reflectance distribution function (BRDF) and this information can vary at each surface point. In order to achieve a transparent watermark, reflectance coding is implemented using a spatial variation of the Brewster angle. The novel multiview imaging method measures the reflectance over a range of viewing and illumination angles in order to instantly reveal the unknown Brewster angle. Unlike typical in-lab measurements of the Brewster angle or the refractive index, this method does not require accurate prior knowledge of the surface normal so that imaging in non-lab conditions is feasible. Furthermore, a range of incident angles are examined simultaneously, eliminating the need for scanning incidence angles. The approach is well-suited for transparent watermarking where the observer cannot see the watermark because it is comprised of spatial variations of refractive index. The transparency and angular coding of the watermark has great utility in deterring counterfeit attempts. In this paper, we present the imaging device and demonstrate it's effectiveness in detecting and measuring changes in refractive index. This device acts as the decoder in a transparent watermark system.
我们提出了一种使用自定义双向成像设备的透明水印方法。我们方法的两个创新概念是反射编码和多视图成像。在反射率编码中,信息嵌入在双向反射率分布函数(BRDF)的角空间中,并且该信息在每个表面点上可以变化。为了实现透明水印,利用布鲁斯特角的空间变化实现了反射率编码。这种新颖的多视点成像方法可以测量不同视角和光照角度的反射率,以便即时显示未知的布鲁斯特角。与典型的实验室测量布鲁斯特角或折射率不同,该方法不需要精确的表面法向先验知识,因此在非实验室条件下成像是可行的。此外,同时检查一系列入射角,消除了扫描入射角的需要。由于透明水印是由折射率的空间变化组成的,因此该方法非常适合于观察者无法看到水印的透明水印。水印的透明性和角度编码在防伪方面有很大的作用。在本文中,我们提出了成像装置,并证明了它在探测和测量折射率变化方面的有效性。该装置在透明水印系统中充当解码器。
{"title":"Transparent watermarking using bidirectional imaging","authors":"Kristin J. Dana, G. Livescu, R. Makonahalli","doi":"10.1109/CVPRW.2009.5204321","DOIUrl":"https://doi.org/10.1109/CVPRW.2009.5204321","url":null,"abstract":"We present a method for transparent watermarking using a custom bidirectional imaging device. The two innovative concepts of our approach are reflectance coding and multiview imaging. In reflectance coding, information is embedded in the angular space of the bidirectional reflectance distribution function (BRDF) and this information can vary at each surface point. In order to achieve a transparent watermark, reflectance coding is implemented using a spatial variation of the Brewster angle. The novel multiview imaging method measures the reflectance over a range of viewing and illumination angles in order to instantly reveal the unknown Brewster angle. Unlike typical in-lab measurements of the Brewster angle or the refractive index, this method does not require accurate prior knowledge of the surface normal so that imaging in non-lab conditions is feasible. Furthermore, a range of incident angles are examined simultaneously, eliminating the need for scanning incidence angles. The approach is well-suited for transparent watermarking where the observer cannot see the watermark because it is comprised of spatial variations of refractive index. The transparency and angular coding of the watermark has great utility in deterring counterfeit attempts. In this paper, we present the imaging device and demonstrate it's effectiveness in detecting and measuring changes in refractive index. This device acts as the decoder in a transparent watermark system.","PeriodicalId":431981,"journal":{"name":"2009 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117172727","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Auditory dialog analysis and understanding by generative modelling of interactional dynamics 通过互动动态的生成建模来分析和理解听觉对话
M. Cristani, Anna Pesarin, C. Drioli, A. Tavano, A. Perina, Vittorio Murino
In the last few years, the interest in the analysis of human behavioral schemes has dramatically grown, in particular for the interpretation of the communication modalities called social signals. They represent well defined interaction patterns, possibly unconscious, characterizing different conversational situations and behaviors in general. In this paper, we illustrate an automatic system based on a generative structure able to analyze conversational scenarios. The generative model is composed by integrating a Gaussian mixture model and the (observed) influence model, and it is fed with a novel kind of simple low-level auditory social signals, which are termed steady conversational periods (SCPs). These are built on duration of continuous slots of silence or speech, taking also into account conversational turn-taking. The interactional dynamics built upon the transitions among SCPs provide a behavioral blueprint of conversational settings without relying on segmental or continuous phonetic features. Our contribution here is to show the effectiveness of our model when applied on dialogs classification and clustering tasks, considering dialogs between adults and between children and adults, in both flat and arguing discussions, and showing excellent performances also in comparison with state-of-the-art frameworks.
在过去的几年里,对人类行为模式分析的兴趣急剧增长,特别是对被称为社会信号的交流方式的解释。它们代表了定义良好的交互模式,可能是无意识的,通常表征不同的会话情境和行为。在本文中,我们演示了一个基于生成结构的自动系统,该系统能够分析会话场景。生成模型由高斯混合模型和(观察到的)影响模型集成而成,并以一种新颖的简单低层次听觉社会信号作为输入,这些信号被称为稳定会话期(SCPs)。这些是建立在沉默或讲话的连续时段的持续时间上的,同时也考虑到对话的轮流。建立在scp之间转换的互动动态提供了会话设置的行为蓝图,而不依赖于分段或连续的语音特征。我们在这里的贡献是展示了我们的模型在应用于对话分类和聚类任务时的有效性,考虑到成人之间以及儿童与成人之间的对话,在扁平和争论的讨论中,并且与最先进的框架相比也显示出出色的性能。
{"title":"Auditory dialog analysis and understanding by generative modelling of interactional dynamics","authors":"M. Cristani, Anna Pesarin, C. Drioli, A. Tavano, A. Perina, Vittorio Murino","doi":"10.1109/CVPRW.2009.5204265","DOIUrl":"https://doi.org/10.1109/CVPRW.2009.5204265","url":null,"abstract":"In the last few years, the interest in the analysis of human behavioral schemes has dramatically grown, in particular for the interpretation of the communication modalities called social signals. They represent well defined interaction patterns, possibly unconscious, characterizing different conversational situations and behaviors in general. In this paper, we illustrate an automatic system based on a generative structure able to analyze conversational scenarios. The generative model is composed by integrating a Gaussian mixture model and the (observed) influence model, and it is fed with a novel kind of simple low-level auditory social signals, which are termed steady conversational periods (SCPs). These are built on duration of continuous slots of silence or speech, taking also into account conversational turn-taking. The interactional dynamics built upon the transitions among SCPs provide a behavioral blueprint of conversational settings without relying on segmental or continuous phonetic features. Our contribution here is to show the effectiveness of our model when applied on dialogs classification and clustering tasks, considering dialogs between adults and between children and adults, in both flat and arguing discussions, and showing excellent performances also in comparison with state-of-the-art frameworks.","PeriodicalId":431981,"journal":{"name":"2009 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":"12 3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123268718","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
Posture invariant gender classification for 3D human models 三维人体模型的姿势不变性别分类
S. Wuhrer, Chang Shu, M. Rioux
We study the behaviorally important task of gender classification based on the human body shape. We propose a new technique to classify by gender human bodies represented by possibly incomplete triangular meshes obtained using laser range scanners. The classification algorithm is invariant of the posture of the human body. Geodesic distances on the mesh are used for classification. Our results indicate that the geodesic distances between the chest and the wrists and the geodesic distances between the lower back and the face are the most important ones for gender classification. The classification is shown to perform well for different postures of the human subjects. We model the geodesic distance distributions as Gaussian distributions and compute the quality of the classification for three standard methods in pattern recognition: linear discriminant functions, Bayesian discriminant functions, and support vector machines. All of the experiments yield high classification accuracy. For instance, when support vector machines are used, the classification accuracy is at least 93% for all of our experiments. This shows that geodesic distances are suitable to discriminate humans by gender.
我们研究了基于人体形状的性别分类这一重要的行为任务。我们提出了一种用激光测距仪获得的可能不完整的三角形网格来对人体进行性别分类的新技术。该分类算法不受人体姿态的影响。网格上的测地线距离用于分类。我们的研究结果表明,胸部和手腕之间的测地线距离和下背部和脸部之间的测地线距离是最重要的性别分类指标。结果表明,这种分类方法对不同姿势的人类受试者表现良好。我们将测地线距离分布建模为高斯分布,并计算模式识别中三种标准方法的分类质量:线性判别函数、贝叶斯判别函数和支持向量机。所有的实验都得到了较高的分类准确率。例如,当使用支持向量机时,我们所有的实验的分类准确率至少为93%。这表明测地线距离适用于性别区分。
{"title":"Posture invariant gender classification for 3D human models","authors":"S. Wuhrer, Chang Shu, M. Rioux","doi":"10.1109/CVPRW.2009.5204295","DOIUrl":"https://doi.org/10.1109/CVPRW.2009.5204295","url":null,"abstract":"We study the behaviorally important task of gender classification based on the human body shape. We propose a new technique to classify by gender human bodies represented by possibly incomplete triangular meshes obtained using laser range scanners. The classification algorithm is invariant of the posture of the human body. Geodesic distances on the mesh are used for classification. Our results indicate that the geodesic distances between the chest and the wrists and the geodesic distances between the lower back and the face are the most important ones for gender classification. The classification is shown to perform well for different postures of the human subjects. We model the geodesic distance distributions as Gaussian distributions and compute the quality of the classification for three standard methods in pattern recognition: linear discriminant functions, Bayesian discriminant functions, and support vector machines. All of the experiments yield high classification accuracy. For instance, when support vector machines are used, the classification accuracy is at least 93% for all of our experiments. This shows that geodesic distances are suitable to discriminate humans by gender.","PeriodicalId":431981,"journal":{"name":"2009 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":"13 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117007756","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
Tunable tensor voting improves grouping of membrane-bound macromolecules 可调张量投票改善了膜结合大分子的分组
Leandro A. Loss, G. Bebis, B. Parvin
Membrane-bound macromolecules are responsible for structural support and mediation of cell-cell adhesion in tissues. Quantitative analysis of these macromolecules provides morphological indices for damage or loss of tissue, for example as a result of exogenous stimuli. From an optical point of view, a membrane signal may have nonuniform intensity around the cell boundary, be punctate or diffused, and may even be perceptual at certain locations along the boundary. In this paper, a method for the detection and grouping of punctate, diffuse curvilinear signals is proposed. Our work builds upon the tensor voting and the iterative voting frameworks to propose an efficient method to detect and refine perceptually interesting curvilinear structures in images. The novelty of our method lies on the idea of iteratively tuning the tensor voting fields, which allows the concentration of the votes only over areas of interest. We validate the utility of our system with synthetic and annotated real data. The effectiveness of the tunable tensor voting is demonstrated on complex phenotypic signals that are representative of membrane-bound macromolecular structures.
膜结合大分子负责组织中细胞-细胞粘附的结构支持和调解。这些大分子的定量分析为组织损伤或损失提供形态学指标,例如作为外源性刺激的结果。从光学角度来看,膜信号在细胞边界周围可能具有不均匀的强度,可能是点状的或弥漫性的,甚至在沿边界的某些位置可能是可感知的。本文提出了一种点状扩散曲线信号的检测与分组方法。我们的工作建立在张量投票和迭代投票框架的基础上,提出了一种有效的方法来检测和改进图像中感知上有趣的曲线结构。我们方法的新颖之处在于迭代调整张量投票场的思想,这允许只在感兴趣的区域集中投票。我们用综合和注释的真实数据验证了系统的实用性。可调张量投票的有效性证明了复杂的表型信号,是膜结合大分子结构的代表。
{"title":"Tunable tensor voting improves grouping of membrane-bound macromolecules","authors":"Leandro A. Loss, G. Bebis, B. Parvin","doi":"10.1109/CVPRW.2009.5204047","DOIUrl":"https://doi.org/10.1109/CVPRW.2009.5204047","url":null,"abstract":"Membrane-bound macromolecules are responsible for structural support and mediation of cell-cell adhesion in tissues. Quantitative analysis of these macromolecules provides morphological indices for damage or loss of tissue, for example as a result of exogenous stimuli. From an optical point of view, a membrane signal may have nonuniform intensity around the cell boundary, be punctate or diffused, and may even be perceptual at certain locations along the boundary. In this paper, a method for the detection and grouping of punctate, diffuse curvilinear signals is proposed. Our work builds upon the tensor voting and the iterative voting frameworks to propose an efficient method to detect and refine perceptually interesting curvilinear structures in images. The novelty of our method lies on the idea of iteratively tuning the tensor voting fields, which allows the concentration of the votes only over areas of interest. We validate the utility of our system with synthetic and annotated real data. The effectiveness of the tunable tensor voting is demonstrated on complex phenotypic signals that are representative of membrane-bound macromolecular structures.","PeriodicalId":431981,"journal":{"name":"2009 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":"17 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128465438","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
High level activity recognition using low resolution wearable vision 使用低分辨率可穿戴视觉的高水平活动识别
Sudeep Sundaram, W. Mayol-Cuevas
This paper presents a system aimed to serve as the enabling platform for a wearable assistant. The method observes manipulations from a wearable camera and classifies activities from roughly stabilized low resolution images (160 × 120 pixels) with the help of a 3-level Dynamic Bayesian Network and adapted temporal templates. Our motivation is to explore robust but computationally inexpensive visual methods to perform as much activity inference as possible without resorting to more complex object or hand detectors. The description of the method and results obtained are presented, as well as the motivation for further work in the area of wearable visual sensing.
本文提出了一个系统,旨在作为可穿戴助手的启用平台。该方法通过可穿戴相机观察操作,并借助3级动态贝叶斯网络和自适应时间模板,从大致稳定的低分辨率图像(160 × 120像素)中对活动进行分类。我们的动机是探索健壮但计算成本低廉的视觉方法,以执行尽可能多的活动推理,而无需诉诸更复杂的物体或手检测器。给出了方法的描述和所获得的结果,以及在可穿戴视觉传感领域进一步工作的动机。
{"title":"High level activity recognition using low resolution wearable vision","authors":"Sudeep Sundaram, W. Mayol-Cuevas","doi":"10.1109/CVPRW.2009.5204355","DOIUrl":"https://doi.org/10.1109/CVPRW.2009.5204355","url":null,"abstract":"This paper presents a system aimed to serve as the enabling platform for a wearable assistant. The method observes manipulations from a wearable camera and classifies activities from roughly stabilized low resolution images (160 × 120 pixels) with the help of a 3-level Dynamic Bayesian Network and adapted temporal templates. Our motivation is to explore robust but computationally inexpensive visual methods to perform as much activity inference as possible without resorting to more complex object or hand detectors. The description of the method and results obtained are presented, as well as the motivation for further work in the area of wearable visual sensing.","PeriodicalId":431981,"journal":{"name":"2009 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":"52 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126969137","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 60
A user-friendly method to geometrically calibrate projector-camera systems 一个用户友好的方法几何校准投影仪-摄像机系统
S. Audet, M. Okutomi
Projector-camera systems drive applications in many fields such as measurement and spatial augmented reality. When needed, we can find their internal and external parameters via geometric calibration. For this process, we have to use both a printed pattern and a projector pattern, but they can easily interfere with each other. Current methods compensate by decoupling their calibrations or by leveraging structured light and color channels, but the required manipulations are not user-friendly. Therefore, we cannot expect normal users to execute the procedure, which can also become a burden for researchers. Although not always required, knowledge of the geometric parameters can often facilitate development of new systems. To make the calibration process easier, we propose a method that uses fiducial markers, from which we can easily derive a prewarp that, once applied to the projector calibration pattern, prevents its interference. Using our method, we confirmed that users can easily calibrate a projector-camera system in less than one minute, which we consider to be user-friendly, while still achieving typical subpixel accuracy.
投影相机系统在测量和空间增强现实等许多领域都有应用。在需要的时候,我们可以通过几何标定找到它们的内外参数。在这个过程中,我们必须同时使用印刷图案和投影仪图案,但它们很容易相互干扰。目前的方法通过解耦校准或利用结构光和颜色通道进行补偿,但所需的操作并不用户友好。因此,我们不能期望普通用户执行该程序,这也可能成为研究人员的负担。虽然并不总是必需的,但几何参数的知识通常可以促进新系统的开发。为了使校准过程更容易,我们提出了一种使用基准标记的方法,从中我们可以很容易地推导出预扭曲,一旦应用于投影仪校准模式,就可以防止其干扰。使用我们的方法,我们证实用户可以在不到一分钟的时间内轻松校准投影仪-摄像机系统,我们认为这是用户友好的,同时仍然达到典型的亚像素精度。
{"title":"A user-friendly method to geometrically calibrate projector-camera systems","authors":"S. Audet, M. Okutomi","doi":"10.1109/CVPRW.2009.5204319","DOIUrl":"https://doi.org/10.1109/CVPRW.2009.5204319","url":null,"abstract":"Projector-camera systems drive applications in many fields such as measurement and spatial augmented reality. When needed, we can find their internal and external parameters via geometric calibration. For this process, we have to use both a printed pattern and a projector pattern, but they can easily interfere with each other. Current methods compensate by decoupling their calibrations or by leveraging structured light and color channels, but the required manipulations are not user-friendly. Therefore, we cannot expect normal users to execute the procedure, which can also become a burden for researchers. Although not always required, knowledge of the geometric parameters can often facilitate development of new systems. To make the calibration process easier, we propose a method that uses fiducial markers, from which we can easily derive a prewarp that, once applied to the projector calibration pattern, prevents its interference. Using our method, we confirmed that users can easily calibrate a projector-camera system in less than one minute, which we consider to be user-friendly, while still achieving typical subpixel accuracy.","PeriodicalId":431981,"journal":{"name":"2009 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":"40 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126973018","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 109
Illustrating motion through DLP photography 通过DLP摄影说明运动
S. Koppal, S. Narasimhan
Strobe-light photography creates beautiful high-frequency effects by capturing multiple object copies. Single-chip DLP projectors produce a similar effect, with two important distinctions. Firstly, strobing occurs at different frequencies: at 10000 Hz, due to the DMD chip, and at 120 Hz, due to the colorwheel. Secondly, DLP illumination lacks the perception of `on-off' flashing that characterizes a strobe-light, since these frequencies are beyond human perception. While higher spatial frequencies are preserved in a motion-blurred DLP photograph, deblurring is still difficult, especially for articulated and deformable objects, since the deconvolution kernel can be different at each pixel. Instead we process DLP photographs to create new images that either summarize a dynamic scene or illustrate its motion. We conclude by discussing the frequencies present in DLP photographs, comparing them to images taken under skylight and fluorescent light.
频闪灯摄影通过捕捉多个物体副本来创造美丽的高频效果。单芯片DLP投影仪产生类似的效果,有两个重要的区别。首先,频闪发生在不同的频率:在10000赫兹,由于DMD芯片,和在120赫兹,由于色轮。其次,DLP照明缺乏频闪灯特有的“开-关”闪烁的感觉,因为这些频率超出了人类的感知。虽然在运动模糊的DLP照片中保留了较高的空间频率,但去模糊仍然很困难,特别是对于关节和可变形的物体,因为每个像素的反卷积核可能不同。相反,我们处理DLP照片来创建新的图像,要么总结动态场景,要么说明其运动。我们通过讨论DLP照片中存在的频率来总结,将它们与在天窗和荧光灯下拍摄的图像进行比较。
{"title":"Illustrating motion through DLP photography","authors":"S. Koppal, S. Narasimhan","doi":"10.1109/CVPRW.2009.5204315","DOIUrl":"https://doi.org/10.1109/CVPRW.2009.5204315","url":null,"abstract":"Strobe-light photography creates beautiful high-frequency effects by capturing multiple object copies. Single-chip DLP projectors produce a similar effect, with two important distinctions. Firstly, strobing occurs at different frequencies: at 10000 Hz, due to the DMD chip, and at 120 Hz, due to the colorwheel. Secondly, DLP illumination lacks the perception of `on-off' flashing that characterizes a strobe-light, since these frequencies are beyond human perception. While higher spatial frequencies are preserved in a motion-blurred DLP photograph, deblurring is still difficult, especially for articulated and deformable objects, since the deconvolution kernel can be different at each pixel. Instead we process DLP photographs to create new images that either summarize a dynamic scene or illustrate its motion. We conclude by discussing the frequencies present in DLP photographs, comparing them to images taken under skylight and fluorescent light.","PeriodicalId":431981,"journal":{"name":"2009 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":"27 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121107289","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Image registration by minimization of Mapping Complexity 最小化映射复杂度的图像配准
A. Myronenko, Xubo B. Song
The criterion for the correct spatial alignment is a key component in image registration. We formulate the registration problem as one that finds the spatial and intensity mappings of minimal complexity that make images exactly equal. We do not assume any parametric forms of these functions, and estimate them within variational calculus. We analytically solve for non-stationary intensity mapping, eliminate it from the objective function and arrive with a new similarity measure. We name it the mapping complexity (MC) similarity measure, because it achieves the optimum when intensity and spatial mappings are of minimal complexity. Due to its general formulation, the similarity measure works both for complex intensity relationships (e.g. multimodal registration) and for spatially-varying intensity distortions. Our similarity measure can be interpreted as the one that favors one image to lie mostly within a span of the leading eigenvectors of the kernel matrix, where the kernel matrix is constructed from the second image. We introduce a fast algorithm to compute the similarity measure. In particular, we introduce a fast kernel vector product (FKVP) algorithm, which is of general interest in computer vision. We demonstrate the accuracy of the new similarity measure on several mono- and multi-modal examples with complex intensity non-uniformities.
正确的空间对齐准则是图像配准的关键。我们将配准问题表述为找到最小复杂度的空间和强度映射,使图像完全相等。我们不假设这些函数的任何参数形式,并在变分演算中估计它们。对非平稳强度映射进行解析求解,将其从目标函数中剔除,得到新的相似性测度。我们将其命名为映射复杂性(MC)相似度度量,因为当强度和空间映射的复杂性最小时,它能达到最优。由于其一般公式,相似性度量既适用于复杂的强度关系(如多模态注册),也适用于空间变化的强度扭曲。我们的相似性度量可以被解释为倾向于一个图像大部分位于核矩阵的主要特征向量的跨度内,其中核矩阵是由第二张图像构造的。我们引入了一种快速的相似度计算算法。特别地,我们介绍了一种快速核向量积(FKVP)算法,这是计算机视觉中普遍感兴趣的算法。我们在几个具有复杂强度非均匀性的单模态和多模态例子上证明了新的相似性度量的准确性。
{"title":"Image registration by minimization of Mapping Complexity","authors":"A. Myronenko, Xubo B. Song","doi":"10.1109/CVPRW.2009.5204345","DOIUrl":"https://doi.org/10.1109/CVPRW.2009.5204345","url":null,"abstract":"The criterion for the correct spatial alignment is a key component in image registration. We formulate the registration problem as one that finds the spatial and intensity mappings of minimal complexity that make images exactly equal. We do not assume any parametric forms of these functions, and estimate them within variational calculus. We analytically solve for non-stationary intensity mapping, eliminate it from the objective function and arrive with a new similarity measure. We name it the mapping complexity (MC) similarity measure, because it achieves the optimum when intensity and spatial mappings are of minimal complexity. Due to its general formulation, the similarity measure works both for complex intensity relationships (e.g. multimodal registration) and for spatially-varying intensity distortions. Our similarity measure can be interpreted as the one that favors one image to lie mostly within a span of the leading eigenvectors of the kernel matrix, where the kernel matrix is constructed from the second image. We introduce a fast algorithm to compute the similarity measure. In particular, we introduce a fast kernel vector product (FKVP) algorithm, which is of general interest in computer vision. We demonstrate the accuracy of the new similarity measure on several mono- and multi-modal examples with complex intensity non-uniformities.","PeriodicalId":431981,"journal":{"name":"2009 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":"39 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124416014","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Measuring changes in face appearance through aging 通过衰老测量面部外观的变化
M. Ortega, L. Brodo, M. Bicego, M. Tistarelli
Facial aging has been only partially studied in the past and mostly in a qualitative way. This paper presents a novel approach to the estimation of facial aging aimed to the quantitative evaluation of the changes in facial appearance over time. In particular, the changes both in face shape and texture, due to short-time aging, are considered. The developed framework exploits the concept of “distinctiveness” of facial features and the temporal evolution of such measure. The analysis is performed both at a global and local level to define the features which are more stable over time. Several experiments are performed on publicly available databases with image sequences densely sampled over a time span of several years. The reported results clearly show the potential of the methodology to a number of applications in biometric identification from human faces.
过去对面部衰老的研究只是部分的,而且大多是定性的。本文提出了一种新的面部老化估计方法,旨在定量评估面部外观随时间的变化。特别是考虑了由于短时间老化而引起的脸型和质地的变化。开发的框架利用了面部特征的“独特性”概念和这种测量的时间演变。分析在全局和局部级别执行,以定义随时间推移更稳定的特征。几个实验进行了公开可用的数据库与图像序列密集采样的时间跨度几年。报告的结果清楚地显示了该方法在人脸生物识别方面的许多应用的潜力。
{"title":"Measuring changes in face appearance through aging","authors":"M. Ortega, L. Brodo, M. Bicego, M. Tistarelli","doi":"10.1109/CVPRW.2009.5204302","DOIUrl":"https://doi.org/10.1109/CVPRW.2009.5204302","url":null,"abstract":"Facial aging has been only partially studied in the past and mostly in a qualitative way. This paper presents a novel approach to the estimation of facial aging aimed to the quantitative evaluation of the changes in facial appearance over time. In particular, the changes both in face shape and texture, due to short-time aging, are considered. The developed framework exploits the concept of “distinctiveness” of facial features and the temporal evolution of such measure. The analysis is performed both at a global and local level to define the features which are more stable over time. Several experiments are performed on publicly available databases with image sequences densely sampled over a time span of several years. The reported results clearly show the potential of the methodology to a number of applications in biometric identification from human faces.","PeriodicalId":431981,"journal":{"name":"2009 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":"86 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126346310","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
期刊
2009 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1