首页 > 最新文献

2012 Western New York Image Processing Workshop最新文献

英文 中文
Detecting hand-palm orientation and hand shapes for sign language gesture recognition using 3D images 利用3D图像检测手心方向和手部形状,用于手语手势识别
Pub Date : 2012-11-01 DOI: 10.1109/WNYIPW.2012.6466652
L. K. Phadtare, R. Kushalnagar, N. Cahill
Automatic gesture recognition, specifically for the purpose of understanding sign language, can be an important aid in communicating with the deaf and hard-of-hearing. Recognition of sign languages requires understanding of various linguistic components such as palm orientation, hand shape, hand location and facial expression. We propose a method and system to estimate the palm orientation and the hand shape of a signer. Our system uses Microsoft Kinect to capture color and the depth images of a signer. It analyzes the depth data corresponding to the hand point region and fits plane to this data and defines the normal to this plane as the orientation of the palm. Then it uses 3-D shape context to determine the hand shape by comparing it to example shapes in the database. Palm orientation of the hand was found to be correct in varying poses. The shape context method for hand shape classification was found to identify 20 test hand shapes correctly and 10 shapes were matched to other but very similar shapes.
自动手势识别,特别是为了理解手语的目的,可以成为与聋哑人和听力障碍者交流的重要辅助。识别手语需要理解各种语言成分,如手掌方向、手的形状、手的位置和面部表情。我们提出了一种方法和系统来估计一个人的手掌方向和手的形状。我们的系统使用微软Kinect来捕捉签名者的颜色和深度图像。分析手点区域对应的深度数据,对该数据进行平面拟合,并定义该平面的法线作为手掌的方向。然后,它使用3-D形状上下文通过将其与数据库中的示例形状进行比较来确定手的形状。研究发现,在不同的姿势下,手掌的方向是正确的。研究发现,形状上下文法可以正确识别20个测试手形,其中10个形状与其他非常相似的形状相匹配。
{"title":"Detecting hand-palm orientation and hand shapes for sign language gesture recognition using 3D images","authors":"L. K. Phadtare, R. Kushalnagar, N. Cahill","doi":"10.1109/WNYIPW.2012.6466652","DOIUrl":"https://doi.org/10.1109/WNYIPW.2012.6466652","url":null,"abstract":"Automatic gesture recognition, specifically for the purpose of understanding sign language, can be an important aid in communicating with the deaf and hard-of-hearing. Recognition of sign languages requires understanding of various linguistic components such as palm orientation, hand shape, hand location and facial expression. We propose a method and system to estimate the palm orientation and the hand shape of a signer. Our system uses Microsoft Kinect to capture color and the depth images of a signer. It analyzes the depth data corresponding to the hand point region and fits plane to this data and defines the normal to this plane as the orientation of the palm. Then it uses 3-D shape context to determine the hand shape by comparing it to example shapes in the database. Palm orientation of the hand was found to be correct in varying poses. The shape context method for hand shape classification was found to identify 20 test hand shapes correctly and 10 shapes were matched to other but very similar shapes.","PeriodicalId":218110,"journal":{"name":"2012 Western New York Image Processing Workshop","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117139377","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
A study of the use of SIMD instructions for two image processing algorithms 研究了使用SIMD指令的两种图像处理算法
Pub Date : 2012-11-01 DOI: 10.1109/WNYIPW.2012.6466650
E. Welch, D. Patru, E. Saber, K. Bengtson
Most image processing algorithms are parallelizable, i.e. the calculation of one pixel does not affect another one. SIMD architectures, including Intel's WMMX and SSE and ARM's NEON, can exploit this fact by processing multiple pixels at a time, which can result in significant speedups. This study investigates the use of NEON SIMD instructions for two image processing algorithms. The latter are altered to process four pixels at a time, for which a theoretical speedup factor of four can be achieved. In addition, parts of the original implementation have been replaced with inline functions or modified at assembly code level. Experimental benchmark data shows the actual execution speed to be between two to three times higher than the original reference. These results prove that SIMD instructions can significantly speedup image processing algorithms through proper code manipulations.
大多数图像处理算法都是可并行的,即一个像素的计算不会影响另一个像素。SIMD架构,包括英特尔的WMMX和SSE以及ARM的NEON,可以通过一次处理多个像素来利用这一事实,这可以带来显着的速度提升。本研究探讨了NEON SIMD指令在两种图像处理算法中的使用。后者被改变为一次处理四个像素,因此理论上可以实现四倍的加速因子。此外,原始实现的部分已被内联函数取代或在汇编代码级别进行了修改。实验基准数据表明,实际执行速度比原始参考高出两到三倍。这些结果证明,通过适当的代码操作,SIMD指令可以显著提高图像处理算法的速度。
{"title":"A study of the use of SIMD instructions for two image processing algorithms","authors":"E. Welch, D. Patru, E. Saber, K. Bengtson","doi":"10.1109/WNYIPW.2012.6466650","DOIUrl":"https://doi.org/10.1109/WNYIPW.2012.6466650","url":null,"abstract":"Most image processing algorithms are parallelizable, i.e. the calculation of one pixel does not affect another one. SIMD architectures, including Intel's WMMX and SSE and ARM's NEON, can exploit this fact by processing multiple pixels at a time, which can result in significant speedups. This study investigates the use of NEON SIMD instructions for two image processing algorithms. The latter are altered to process four pixels at a time, for which a theoretical speedup factor of four can be achieved. In addition, parts of the original implementation have been replaced with inline functions or modified at assembly code level. Experimental benchmark data shows the actual execution speed to be between two to three times higher than the original reference. These results prove that SIMD instructions can significantly speedup image processing algorithms through proper code manipulations.","PeriodicalId":218110,"journal":{"name":"2012 Western New York Image Processing Workshop","volume":"10 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131338106","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 12
Video detection anomaly via low-rank and sparse decompositions 基于低秩和稀疏分解的视频异常检测
Pub Date : 2012-11-01 DOI: 10.1109/WNYIPW.2012.6466649
Lam Tran, C. Navasca, Jiebo Luo
In this paper, we purpose a method for anomaly detection in surveillance video in a tensor framework. We treat a video as a tensor and utilize a stable PCA to decompose it into two tensors, the first tensor is a low rank tensor that consists of background pixels and the second tensor is a sparse tensor that consists of the foreground pixels. The sparse tensor is then analyzed to detect anomaly. The proposed method is a one-shot framework to determine frames that are anomalous in a video.
本文提出了一种基于张量框架的监控视频异常检测方法。我们将视频视为一个张量,并利用稳定的PCA将其分解为两个张量,第一个张量是由背景像素组成的低秩张量,第二个张量是由前景像素组成的稀疏张量。然后对稀疏张量进行分析以检测异常。提出的方法是一个单镜头框架来确定视频中的异常帧。
{"title":"Video detection anomaly via low-rank and sparse decompositions","authors":"Lam Tran, C. Navasca, Jiebo Luo","doi":"10.1109/WNYIPW.2012.6466649","DOIUrl":"https://doi.org/10.1109/WNYIPW.2012.6466649","url":null,"abstract":"In this paper, we purpose a method for anomaly detection in surveillance video in a tensor framework. We treat a video as a tensor and utilize a stable PCA to decompose it into two tensors, the first tensor is a low rank tensor that consists of background pixels and the second tensor is a sparse tensor that consists of the foreground pixels. The sparse tensor is then analyzed to detect anomaly. The proposed method is a one-shot framework to determine frames that are anomalous in a video.","PeriodicalId":218110,"journal":{"name":"2012 Western New York Image Processing Workshop","volume":"115 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124248192","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 19
Lightweight, low-cost, side-mounted mobile eye tracking system 轻便,低成本,侧装移动眼动追踪系统
Pub Date : 2012-11-01 DOI: 10.1109/WNYIPW.2012.6466645
A. K. A. Hong, J. Pelz, J. Cockburn
Commercial mobile eye tracking systems are readily available, but are costly and complex. They have an additional disadvantage in that the eye cameras are placed directly in the field of view of the subject in order to obtain a clear frontal view of the eye. We propose a lightweight, low-cost, side-mounted mobile eye tracking system that uses side-view eye images to estimate the gaze of the subject. Cameras are mounted on the side of the head using curved mirrors to split the captured frames into scene and eye images. A hybrid algorithm using both feature-based models and appearance-based models is designed to accommodate this novel system. Image sequences, consisting of 4339 frames from seven subjects are analyzed by the algorithm, resulting in a successful gaze estimation rate of 95.7%.
商用移动眼动追踪系统很容易获得,但价格昂贵且复杂。它们还有一个额外的缺点,即眼睛摄像机直接放置在对象的视野中,以便获得眼睛的清晰正面视图。我们提出了一种轻量级,低成本,侧面安装的移动眼动追踪系统,该系统使用侧视眼图像来估计受试者的凝视。摄像头安装在头部一侧,使用曲面镜将捕捉到的画面分成场景和眼睛图像。为了适应这种新系统,设计了一种基于特征模型和基于外观模型的混合算法。该算法对7个被测对象共4339帧的图像序列进行了分析,成功的注视估计率达到95.7%。
{"title":"Lightweight, low-cost, side-mounted mobile eye tracking system","authors":"A. K. A. Hong, J. Pelz, J. Cockburn","doi":"10.1109/WNYIPW.2012.6466645","DOIUrl":"https://doi.org/10.1109/WNYIPW.2012.6466645","url":null,"abstract":"Commercial mobile eye tracking systems are readily available, but are costly and complex. They have an additional disadvantage in that the eye cameras are placed directly in the field of view of the subject in order to obtain a clear frontal view of the eye. We propose a lightweight, low-cost, side-mounted mobile eye tracking system that uses side-view eye images to estimate the gaze of the subject. Cameras are mounted on the side of the head using curved mirrors to split the captured frames into scene and eye images. A hybrid algorithm using both feature-based models and appearance-based models is designed to accommodate this novel system. Image sequences, consisting of 4339 frames from seven subjects are analyzed by the algorithm, resulting in a successful gaze estimation rate of 95.7%.","PeriodicalId":218110,"journal":{"name":"2012 Western New York Image Processing Workshop","volume":"35 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116804628","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Multi-view action classification using sparse representations on Motion History Images 基于稀疏表示的运动历史图像多视图动作分类
Pub Date : 2012-11-01 DOI: 10.1109/WNYIPW.2012.6466646
S. Azary, A. Savakis
Multi-view action classification is an important component of real world applications such as automatic surveillance and sports analysis. Motion History Images capture the location and direction of motion in a scene and sparse representations provide a compact representation of high dimensional signals. In this paper, we propose a multi-view action classification algorithm based on sparse representation of spatio-temporal action representations using motion history images. We find that this approach is effective at multi-view action classification and experiments with the i3DPost Multi-view Dataset achieve high classification rates.
多视图动作分类是自动监控和体育分析等现实应用的重要组成部分。运动历史图像捕捉场景中运动的位置和方向,稀疏表示提供高维信号的紧凑表示。本文提出了一种基于运动历史图像时空动作表示稀疏表示的多视点动作分类算法。我们发现这种方法在多视图动作分类中是有效的,并且在i3DPost多视图数据集上的实验取得了很高的分类率。
{"title":"Multi-view action classification using sparse representations on Motion History Images","authors":"S. Azary, A. Savakis","doi":"10.1109/WNYIPW.2012.6466646","DOIUrl":"https://doi.org/10.1109/WNYIPW.2012.6466646","url":null,"abstract":"Multi-view action classification is an important component of real world applications such as automatic surveillance and sports analysis. Motion History Images capture the location and direction of motion in a scene and sparse representations provide a compact representation of high dimensional signals. In this paper, we propose a multi-view action classification algorithm based on sparse representation of spatio-temporal action representations using motion history images. We find that this approach is effective at multi-view action classification and experiments with the i3DPost Multi-view Dataset achieve high classification rates.","PeriodicalId":218110,"journal":{"name":"2012 Western New York Image Processing Workshop","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131105302","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Efficient SMQT features for snow-based classification on face detection and character recognition tasks 有效的SMQT特征在基于雪的人脸检测和字符识别任务分类
Pub Date : 2012-11-01 DOI: 10.1109/WNYIPW.2012.6466644
Y. Artan, A. Burry, V. Kozitsky, P. Paul
Face detection using local successive mean quantization transform (SMQT) features and the sparse network of winnows (SNoW) classifier has received interest in the computer vision community due to its success under varying illumination conditions. Recent work has also demonstrated the effectiveness of this classification technique for character recognition tasks. However, heavy storage requirements of the SNoW classifier necessitate the development of efficient techniques to reduce storage and computational requirements. This study shows that the SNoW classifier built with only a limited number of distinguishing SMQT features provides comparable performance to the original dense snow classifier. Initial results using the well-known CMU-MIT facial image database and a private character database are used to demonstrate the effectiveness of the proposed method.
基于局部连续均值量化变换(SMQT)特征和稀疏窗口网络(SNoW)分类器的人脸检测由于其在不同光照条件下的成功而受到了计算机视觉界的关注。最近的工作也证明了这种分类技术在字符识别任务中的有效性。然而,SNoW分类器的大量存储需求需要开发有效的技术来减少存储和计算需求。本研究表明,仅使用有限数量的可区分的SMQT特征构建的SNoW分类器可以提供与原始密集SNoW分类器相当的性能。使用著名的CMU-MIT面部图像数据库和私有字符数据库进行初步结果验证了所提方法的有效性。
{"title":"Efficient SMQT features for snow-based classification on face detection and character recognition tasks","authors":"Y. Artan, A. Burry, V. Kozitsky, P. Paul","doi":"10.1109/WNYIPW.2012.6466644","DOIUrl":"https://doi.org/10.1109/WNYIPW.2012.6466644","url":null,"abstract":"Face detection using local successive mean quantization transform (SMQT) features and the sparse network of winnows (SNoW) classifier has received interest in the computer vision community due to its success under varying illumination conditions. Recent work has also demonstrated the effectiveness of this classification technique for character recognition tasks. However, heavy storage requirements of the SNoW classifier necessitate the development of efficient techniques to reduce storage and computational requirements. This study shows that the SNoW classifier built with only a limited number of distinguishing SMQT features provides comparable performance to the original dense snow classifier. Initial results using the well-known CMU-MIT facial image database and a private character database are used to demonstrate the effectiveness of the proposed method.","PeriodicalId":218110,"journal":{"name":"2012 Western New York Image Processing Workshop","volume":"12 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122101169","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
A multi-channel approach for segmentation of solar corona in images from the solar dynamics observatory 太阳动力学观测站图像中日冕分割的多通道方法
Pub Date : 2012-11-01 DOI: 10.1109/WNYIPW.2012.6466653
S. Suresh, R. Dube
We present a multi-channel segmentation scheme to identify different features of the solar corona, such as coronal holes, active regions and the quiet sun (especially in the ultraviolet and extreme ultraviolet images). In contrast to common techniques, we use an approach that uses image intensity and relative contribution of each of the wavelengths. This approach is illustrated by using the images taken by the AIA telescopes onboard of the SDO mission. This technique incorporates a nearest-neighbor based classifier followed by Moore-neighbor tracing algorithm to find the boundaries and track the regions of interest. This method requires less computation time as compared to the commonly used fuzzy logic methods and is robust in the sense it performs equally well in both the central and limb regions of the solar disc.
我们提出了一种多通道分割方案,以识别日冕的不同特征,如日冕洞、活动区域和安静太阳(特别是在紫外线和极紫外线图像中)。与常用技术相比,我们使用一种方法,利用图像强度和每个波长的相对贡献。使用SDO任务上的AIA望远镜拍摄的图像说明了这种方法。该技术结合了基于最近邻的分类器和Moore-neighbor跟踪算法来找到边界并跟踪感兴趣的区域。与常用的模糊逻辑方法相比,该方法需要更少的计算时间,并且在太阳盘的中心和边缘区域都具有同样的鲁棒性。
{"title":"A multi-channel approach for segmentation of solar corona in images from the solar dynamics observatory","authors":"S. Suresh, R. Dube","doi":"10.1109/WNYIPW.2012.6466653","DOIUrl":"https://doi.org/10.1109/WNYIPW.2012.6466653","url":null,"abstract":"We present a multi-channel segmentation scheme to identify different features of the solar corona, such as coronal holes, active regions and the quiet sun (especially in the ultraviolet and extreme ultraviolet images). In contrast to common techniques, we use an approach that uses image intensity and relative contribution of each of the wavelengths. This approach is illustrated by using the images taken by the AIA telescopes onboard of the SDO mission. This technique incorporates a nearest-neighbor based classifier followed by Moore-neighbor tracing algorithm to find the boundaries and track the regions of interest. This method requires less computation time as compared to the commonly used fuzzy logic methods and is robust in the sense it performs equally well in both the central and limb regions of the solar disc.","PeriodicalId":218110,"journal":{"name":"2012 Western New York Image Processing Workshop","volume":"4 6 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116826289","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Motion tracking for realtime, offline image stabilization with limited hardware 运动跟踪实时,离线图像稳定与有限的硬件
Pub Date : 2012-11-01 DOI: 10.1109/WNYIPW.2012.6466643
J. Sheaffer, M. Moore, M. Bobrov, A. Webster, M. Torres
We present a work-in-progress system for tracking small objects with limited features to produce stabilized, successive images for frame-to-frame analysis in a specialized, hand-held, digital microscopy environment with limited resolution and processing capability and soft real-time requirements. Our system is able to track dirt or imperfections as small as about 10 μm across the end face of a fiber optic cable under a moving camera. It must locate three distinct features and track each of them from frame to frame. The measured positions of the features are used to calculate transformation matrices relative to a selected basis image and move each image in the image stack into the coordinate frame of the basis image so that we can perform focus stacking on the set of images. All of this must be completed in under a second on a low-power, hand-held device.
我们提出了一个正在开发的系统,用于跟踪具有有限特征的小物体,以在具有有限分辨率和处理能力和软实时要求的专用手持式数字显微镜环境中产生稳定的连续图像,用于帧对帧分析。我们的系统能够在移动摄像机下跟踪光纤端面上小至10 μm的污垢或缺陷。它必须定位三个不同的特征,并从一帧到另一帧地跟踪它们。特征的测量位置用于计算相对于选定的基图像的变换矩阵,并将图像堆栈中的每个图像移动到基图像的坐标帧中,以便我们可以对图像集进行焦点叠加。所有这些都必须在一秒钟内在低功耗的手持设备上完成。
{"title":"Motion tracking for realtime, offline image stabilization with limited hardware","authors":"J. Sheaffer, M. Moore, M. Bobrov, A. Webster, M. Torres","doi":"10.1109/WNYIPW.2012.6466643","DOIUrl":"https://doi.org/10.1109/WNYIPW.2012.6466643","url":null,"abstract":"We present a work-in-progress system for tracking small objects with limited features to produce stabilized, successive images for frame-to-frame analysis in a specialized, hand-held, digital microscopy environment with limited resolution and processing capability and soft real-time requirements. Our system is able to track dirt or imperfections as small as about 10 μm across the end face of a fiber optic cable under a moving camera. It must locate three distinct features and track each of them from frame to frame. The measured positions of the features are used to calculate transformation matrices relative to a selected basis image and move each image in the image stack into the coordinate frame of the basis image so that we can perform focus stacking on the set of images. All of this must be completed in under a second on a low-power, hand-held device.","PeriodicalId":218110,"journal":{"name":"2012 Western New York Image Processing Workshop","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127833651","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Color is not a metric space implications for pattern recognition, machine learning, and computer vision 颜色并不是一个度量空间,它暗示着模式识别、机器学习和计算机视觉
Pub Date : 2012-11-01 DOI: 10.1109/WNYIPW.2012.6466642
Thomas B. Kinsman, M. Fairchild, J. Pelz
Using a metric feature space for pattern recognition, data mining, and machine learning greatly simplifies the mathematics because distances are preserved under rotation and translation in feature space. A metric space also provides a “ruler”, or absolute measure of how different two feature vectors are. In the computer vision community color can easily be miss-treated as a metric distance. This paper serves as an introduction to why using a non-metric space is a challenge, and provides details of why color is not a valid Euclidean distance metric.
度量空间还提供了一个“标尺”,或者是两个特征向量差异的绝对度量。在计算机视觉界,颜色很容易被误认为是度量距离。本文介绍了为什么使用非度量空间是一个挑战,并提供了为什么颜色不是一个有效的欧几里得距离度量的细节。
{"title":"Color is not a metric space implications for pattern recognition, machine learning, and computer vision","authors":"Thomas B. Kinsman, M. Fairchild, J. Pelz","doi":"10.1109/WNYIPW.2012.6466642","DOIUrl":"https://doi.org/10.1109/WNYIPW.2012.6466642","url":null,"abstract":"Using a metric feature space for pattern recognition, data mining, and machine learning greatly simplifies the mathematics because distances are preserved under rotation and translation in feature space. A metric space also provides a “ruler”, or absolute measure of how different two feature vectors are. In the computer vision community color can easily be miss-treated as a metric distance. This paper serves as an introduction to why using a non-metric space is a challenge, and provides details of why color is not a valid Euclidean distance metric.","PeriodicalId":218110,"journal":{"name":"2012 Western New York Image Processing Workshop","volume":"207 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132053030","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Hyperspectral linear unmixing: Quantitative evaluation of novel target design and edge unmixing technique 高光谱线性解混:新型目标设计和边缘解混技术的定量评价
Pub Date : 2012-11-01 DOI: 10.1109/WNYIPW.2012.6466651
D. S. Goldberg, J. Kerekes, K. Canham
Remotely sensed hyperspectral images (HSI) have the potential to provide large amounts of information about a scene. HSI, in this context, are images of the Earth collected with a spatial resolution of 1m to 30m in dozens to hundreds of contiguous narrow spectral bands over different wavelengths so that each pixel is a vector of data. Spectral unmixing is one application which can utilize the large amount of information in HSI. Unmixing is a process used to retrieve a material's spectral profile and its fractional abundance in each pixel since a single pixel contains a mixture of material spectra. Unmixing was used with images collected during an airborne hyperspectral collect at the Rochester Institute of Technology in 2010 with 1m resolution and a 390nm to 2450nm spectral range. The goal of our experiment was to quantitatively evaluate unmixing results by introducing a novel unmixing target. In addition, a single-band, edge unmixing technique is introduced with preliminary experimentation which showed results with mean unmixing fraction error of less than 10%. The results of the methods presented above helped in the design of future collection experiments.
遥感高光谱图像(HSI)具有提供大量场景信息的潜力。在这种情况下,HSI是在不同波长的几十到几百个连续的窄光谱带中以1米到30米的空间分辨率收集的地球图像,这样每个像素都是一个数据向量。光谱解混是HSI中可以利用大量信息的一种应用。解混是一个用于检索材料的光谱轮廓及其在每个像素中的分数丰度的过程,因为单个像素包含材料光谱的混合物。2010年,在罗切斯特理工学院(Rochester Institute of Technology)进行了一次机载高光谱采集,图像分辨率为1m,光谱范围为390nm至2450nm。本实验的目的是通过引入一种新的解混靶来定量评价解混效果。此外,还介绍了一种单波段边缘解混技术,并进行了初步实验,结果表明平均解混分数误差小于10%。上述方法的结果有助于今后收集实验的设计。
{"title":"Hyperspectral linear unmixing: Quantitative evaluation of novel target design and edge unmixing technique","authors":"D. S. Goldberg, J. Kerekes, K. Canham","doi":"10.1109/WNYIPW.2012.6466651","DOIUrl":"https://doi.org/10.1109/WNYIPW.2012.6466651","url":null,"abstract":"Remotely sensed hyperspectral images (HSI) have the potential to provide large amounts of information about a scene. HSI, in this context, are images of the Earth collected with a spatial resolution of 1m to 30m in dozens to hundreds of contiguous narrow spectral bands over different wavelengths so that each pixel is a vector of data. Spectral unmixing is one application which can utilize the large amount of information in HSI. Unmixing is a process used to retrieve a material's spectral profile and its fractional abundance in each pixel since a single pixel contains a mixture of material spectra. Unmixing was used with images collected during an airborne hyperspectral collect at the Rochester Institute of Technology in 2010 with 1m resolution and a 390nm to 2450nm spectral range. The goal of our experiment was to quantitatively evaluate unmixing results by introducing a novel unmixing target. In addition, a single-band, edge unmixing technique is introduced with preliminary experimentation which showed results with mean unmixing fraction error of less than 10%. The results of the methods presented above helped in the design of future collection experiments.","PeriodicalId":218110,"journal":{"name":"2012 Western New York Image Processing Workshop","volume":"179 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126766799","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
期刊
2012 Western New York Image Processing Workshop
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1