首页 > 最新文献

2015 21st Korea-Japan Joint Workshop on Frontiers of Computer Vision (FCV)最新文献

英文 中文
A proposal of ambient light estimation methods for skin region detection 一种用于皮肤区域检测的环境光估计方法
Pub Date : 2015-05-11 DOI: 10.1109/FCV.2015.7103706
Tomoya Kondo, K. Kato, Kazuhiko Yamamoto
Many skin region detection methods have been proposed, and color information of the skin is an easy method to detect the skin region from the image. However, the color of the skin is not stable, because the illumination condition, especially the ambient light changes often. Usually, the color space is used to represent the skin color. Nevertheless, it is needed to adjust the parameters of the skin color in the color space for the ambient light change. In order to solve this problem, we propose two methods to detect skin region by estimating the color of the ambient light. In our method, the color of the ambient light is estimated by using one reference light. In the first method, the color of the ambient light is estimated by subtracting two images; an image taken under the unknown ambient light, and an image irradiated a reference light. To subtract these two images, the influence of the ambient light is suppressed. Then the color of the ambient light is estimated. However, irradiation of the reference light is required by changing the ambient light in this method. In the second method, the color of the ambient light is estimated by using the color of the object in the image. However, usually, there is not an object whose color is known. Thus the color of the object in the image is estimated by irradiating the reference light, first. If there is an immovable object in the image, for example the background, this method can estimate the color of the ambient light without irradiate the reference light again. In this paper, these two methods are described, and comparative experiments were conducted. By experimental results, we confirmed that the skin region was detected correctly under the several ambient light environments.
人们提出了许多皮肤区域检测方法,其中皮肤颜色信息是一种从图像中检测皮肤区域的简便方法。然而,皮肤的颜色并不稳定,因为照明条件,特别是环境光经常变化。通常,颜色空间是用来表示皮肤颜色的。然而,需要根据环境光的变化在色彩空间中调整皮肤颜色的参数。为了解决这一问题,我们提出了两种通过估计环境光的颜色来检测皮肤区域的方法。在我们的方法中,通过使用一个参考光来估计环境光的颜色。在第一种方法中,通过相减两幅图像来估计环境光的颜色;在未知环境光下拍摄的图像,以及在参考光照射下的图像。为了减去这两个图像,环境光的影响被抑制。然后估计环境光的颜色。然而,在这种方法中,需要通过改变环境光来照射参考光。在第二种方法中,使用图像中物体的颜色来估计环境光的颜色。然而,通常没有一个物体的颜色是已知的。因此,首先通过照射参考光来估计图像中物体的颜色。如果图像中存在不可移动的物体,例如背景,该方法可以在不再次照射参考光的情况下估计环境光的颜色。本文对这两种方法进行了描述,并进行了对比实验。通过实验结果,我们证实了在几种环境光环境下皮肤区域的检测是正确的。
{"title":"A proposal of ambient light estimation methods for skin region detection","authors":"Tomoya Kondo, K. Kato, Kazuhiko Yamamoto","doi":"10.1109/FCV.2015.7103706","DOIUrl":"https://doi.org/10.1109/FCV.2015.7103706","url":null,"abstract":"Many skin region detection methods have been proposed, and color information of the skin is an easy method to detect the skin region from the image. However, the color of the skin is not stable, because the illumination condition, especially the ambient light changes often. Usually, the color space is used to represent the skin color. Nevertheless, it is needed to adjust the parameters of the skin color in the color space for the ambient light change. In order to solve this problem, we propose two methods to detect skin region by estimating the color of the ambient light. In our method, the color of the ambient light is estimated by using one reference light. In the first method, the color of the ambient light is estimated by subtracting two images; an image taken under the unknown ambient light, and an image irradiated a reference light. To subtract these two images, the influence of the ambient light is suppressed. Then the color of the ambient light is estimated. However, irradiation of the reference light is required by changing the ambient light in this method. In the second method, the color of the ambient light is estimated by using the color of the object in the image. However, usually, there is not an object whose color is known. Thus the color of the object in the image is estimated by irradiating the reference light, first. If there is an immovable object in the image, for example the background, this method can estimate the color of the ambient light without irradiate the reference light again. In this paper, these two methods are described, and comparative experiments were conducted. By experimental results, we confirmed that the skin region was detected correctly under the several ambient light environments.","PeriodicalId":424974,"journal":{"name":"2015 21st Korea-Japan Joint Workshop on Frontiers of Computer Vision (FCV)","volume":"5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-05-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116926092","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Spatially adaptive image defogging using color characteristics 使用颜色特征的空间自适应图像去雾
Pub Date : 2015-05-07 DOI: 10.1109/FCV.2015.7103717
Y. Minami, K. Enomoto, M. Migita, M. Toda
In this paper, we present a vivid image from a foggy underwater image. On fisheries research, it is important to grasp the situation of underwater. They do the undersea observation by using the underwater camera. However, dust floating underwater or mud alter the scene appearance of the camera. They scatter incident light on the camera. The light scattered is one of the major factor of degradation of the scene quality. We tried to solve this problem by adopting defogging algorithm to the low-quality underwater images. But in the underwater images, the environment is very different from the images on ground. The images on ground, airlight is distributed uniformly. On the other hand, the images in underwater, airlight is significantly different between each frame or partial region. We propose the algorithm that can improve visibility from a underwater image which have violent fluctuations of airlight.
在本文中,我们给出了一个生动的图像从一个雾蒙蒙的水下图像。在渔业研究中,掌握水下的情况是非常重要的。他们用水下摄像机进行海底观察。然而,漂浮在水下的灰尘或泥浆会改变相机的场景外观。它们将入射光散射到摄像机上。光散射是影响场景质量的主要因素之一。我们尝试通过对低质量的水下图像采用去雾算法来解决这一问题。但在水下图像中,环境与地面图像非常不同。地面、空中图像分布均匀。另一方面,水下、空中的图像在每帧或部分区域之间存在显著差异。我们提出了一种算法,可以提高空气光波动剧烈的水下图像的能见度。
{"title":"Spatially adaptive image defogging using color characteristics","authors":"Y. Minami, K. Enomoto, M. Migita, M. Toda","doi":"10.1109/FCV.2015.7103717","DOIUrl":"https://doi.org/10.1109/FCV.2015.7103717","url":null,"abstract":"In this paper, we present a vivid image from a foggy underwater image. On fisheries research, it is important to grasp the situation of underwater. They do the undersea observation by using the underwater camera. However, dust floating underwater or mud alter the scene appearance of the camera. They scatter incident light on the camera. The light scattered is one of the major factor of degradation of the scene quality. We tried to solve this problem by adopting defogging algorithm to the low-quality underwater images. But in the underwater images, the environment is very different from the images on ground. The images on ground, airlight is distributed uniformly. On the other hand, the images in underwater, airlight is significantly different between each frame or partial region. We propose the algorithm that can improve visibility from a underwater image which have violent fluctuations of airlight.","PeriodicalId":424974,"journal":{"name":"2015 21st Korea-Japan Joint Workshop on Frontiers of Computer Vision (FCV)","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-05-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128847552","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Phonographic image recognition using fusion of scale invariant descriptor 基于尺度不变描述子融合的留声机图像识别
Pub Date : 2015-05-07 DOI: 10.1109/FCV.2015.7103754
I. Wijaya, I. B. K. Widiartha, K. Uchimura, G. Koutaki
The paper presents a pornographic image recognition using fusion of scale invariant descriptor. The pornographic image means the image contains and shows genital elements of human body having large variability due to poses, lighting, and backgrounds variations. The fusion of scale invariant descriptor that is holistic feature is employed to handle those variability problems. This holistic feature that is pose and scale invariant information of pornographic images is extracted by fusing the scale invariant descriptor of skin region of interests (ROIs) of pornographic images. The skin ROI is used to handle the large variability of pornographic images due to background variations. The main aim of this research finds a good solution for pornographic recognition system, which can be developed to limit the accessing pornographic images by teenagers and children. The experimental results show that the proposed method tends to provide high enough accuracy more than 80%, small enough FNR and FPR bout 2.77% and 28.79%, respectively. It means the proposed method is suitable to develop rejection system of pornographic images. Furthermore, these achievements are much better than the achievements of established methods. This results can be achieved because the fusion of scale invariant descriptor consists rich pornographic information representing holistic feature of pornographic images.
提出了一种融合尺度不变描述子的色情图像识别方法。色情图片是指包含和展示人体的生殖器元素,由于姿势、光线和背景的变化而有很大的变化。采用尺度不变描述子的融合作为整体特征来处理这些变异性问题。通过融合色情图像皮肤兴趣区域(roi)的尺度不变描述子,提取出色情图像的姿态和尺度不变信息的整体特征。皮肤ROI用于处理由于背景变化导致的色情图像的大变异性。本研究的主要目的是为色情识别系统寻找一个良好的解决方案,以限制青少年和儿童对色情图像的访问。实验结果表明,该方法具有较高的准确率(80%以上),较小的FNR(2.77%)和FPR(28.79%)。说明所提出的方法适用于开发色情图像的拒收系统。而且,这些结果比现有方法的结果要好得多。这是因为融合的尺度不变描述子包含了丰富的色情信息,代表了色情图像的整体特征。
{"title":"Phonographic image recognition using fusion of scale invariant descriptor","authors":"I. Wijaya, I. B. K. Widiartha, K. Uchimura, G. Koutaki","doi":"10.1109/FCV.2015.7103754","DOIUrl":"https://doi.org/10.1109/FCV.2015.7103754","url":null,"abstract":"The paper presents a pornographic image recognition using fusion of scale invariant descriptor. The pornographic image means the image contains and shows genital elements of human body having large variability due to poses, lighting, and backgrounds variations. The fusion of scale invariant descriptor that is holistic feature is employed to handle those variability problems. This holistic feature that is pose and scale invariant information of pornographic images is extracted by fusing the scale invariant descriptor of skin region of interests (ROIs) of pornographic images. The skin ROI is used to handle the large variability of pornographic images due to background variations. The main aim of this research finds a good solution for pornographic recognition system, which can be developed to limit the accessing pornographic images by teenagers and children. The experimental results show that the proposed method tends to provide high enough accuracy more than 80%, small enough FNR and FPR bout 2.77% and 28.79%, respectively. It means the proposed method is suitable to develop rejection system of pornographic images. Furthermore, these achievements are much better than the achievements of established methods. This results can be achieved because the fusion of scale invariant descriptor consists rich pornographic information representing holistic feature of pornographic images.","PeriodicalId":424974,"journal":{"name":"2015 21st Korea-Japan Joint Workshop on Frontiers of Computer Vision (FCV)","volume":"10 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-05-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125203887","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Facial features extraction by accelerated implementation of circular hough transform and appearance evaluation 基于加速圆霍夫变换和外观评价的人脸特征提取
Pub Date : 1900-01-01 DOI: 10.1109/FCV.2015.7103709
Hayato Iwasa, W. Ohyama, T. Wakabayashi, F. Kimura
Extraction of facial feature points is essential task for many kinds of applications of face images. The feasibility of facial features extraction is determined by not only its accuracy but processing time. Some applications require real-time detection of facial features. This research aims to propose a method of facial features extraction by an accelerated implementation of circular Hough transform with gradients and appearance evaluation by histogram of gradient features. The acceleration implementation employs General Purpose computing on Graphics Processing Unit. Experiment using FERET database shows that the proposed method successfully extracted eyes and nose for 98.44% and 99.50% of frontal face images in the dataset. And 96.5% of computational time was reduced by accelerated implementation employing GPGPU.
人脸特征点的提取是人脸图像各种应用的基本任务。人脸特征提取的可行性不仅取决于提取的准确性,还取决于提取的时间。一些应用程序需要实时检测面部特征。本研究旨在提出一种利用梯度加速实现圆形霍夫变换并利用梯度特征直方图进行外观评价的人脸特征提取方法。加速实现在图形处理单元上采用通用计算。利用FERET数据库进行的实验表明,该方法成功提取了数据集中98.44%和99.50%的正面人脸图像的眼睛和鼻子。采用GPGPU加速实现后,计算时间减少96.5%。
{"title":"Facial features extraction by accelerated implementation of circular hough transform and appearance evaluation","authors":"Hayato Iwasa, W. Ohyama, T. Wakabayashi, F. Kimura","doi":"10.1109/FCV.2015.7103709","DOIUrl":"https://doi.org/10.1109/FCV.2015.7103709","url":null,"abstract":"Extraction of facial feature points is essential task for many kinds of applications of face images. The feasibility of facial features extraction is determined by not only its accuracy but processing time. Some applications require real-time detection of facial features. This research aims to propose a method of facial features extraction by an accelerated implementation of circular Hough transform with gradients and appearance evaluation by histogram of gradient features. The acceleration implementation employs General Purpose computing on Graphics Processing Unit. Experiment using FERET database shows that the proposed method successfully extracted eyes and nose for 98.44% and 99.50% of frontal face images in the dataset. And 96.5% of computational time was reduced by accelerated implementation employing GPGPU.","PeriodicalId":424974,"journal":{"name":"2015 21st Korea-Japan Joint Workshop on Frontiers of Computer Vision (FCV)","volume":"38 2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126267260","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Fast and robust visual inspection system for tire surface thin defect 快速、鲁棒的轮胎表面薄缺陷视觉检测系统
Pub Date : 1900-01-01 DOI: 10.1109/FCV.2015.7103749
T. Funahashi, K. Taki, H. Koshimizu, Akito Kaneko
In this paper, we propose a new visual inspection method that can analyze the exterior surface of rubber tires with the light stripe projection. Image sensing of the tire surface was implemented by setting the tire on the rotating table so that the normal pattern of the tire surface could be suppressed. The proposed method detects the exterior thin defect after removing three kinds of uneven circularities due to the real shape of tire, the eccentricity of rotating table and the inconsistency between two. Thus, in this research, we realized a visual inspection method by transforming three-dimensional shape of tire surface into two-dimensional image information through the analysis of light stripe images.
本文提出了一种利用光条投影对橡胶轮胎外表面进行视觉检测的新方法。通过将轮胎置于转台上,抑制轮胎表面的法向模式,实现轮胎表面的图像感知。该方法在去除由于轮胎真实形状、转台偏心和二者不一致造成的三种不均匀圆度后,检测出轮胎外薄缺陷。因此,在本研究中,我们通过对光条纹图像的分析,将轮胎表面的三维形状转化为二维图像信息,实现了一种视觉检测方法。
{"title":"Fast and robust visual inspection system for tire surface thin defect","authors":"T. Funahashi, K. Taki, H. Koshimizu, Akito Kaneko","doi":"10.1109/FCV.2015.7103749","DOIUrl":"https://doi.org/10.1109/FCV.2015.7103749","url":null,"abstract":"In this paper, we propose a new visual inspection method that can analyze the exterior surface of rubber tires with the light stripe projection. Image sensing of the tire surface was implemented by setting the tire on the rotating table so that the normal pattern of the tire surface could be suppressed. The proposed method detects the exterior thin defect after removing three kinds of uneven circularities due to the real shape of tire, the eccentricity of rotating table and the inconsistency between two. Thus, in this research, we realized a visual inspection method by transforming three-dimensional shape of tire surface into two-dimensional image information through the analysis of light stripe images.","PeriodicalId":424974,"journal":{"name":"2015 21st Korea-Japan Joint Workshop on Frontiers of Computer Vision (FCV)","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134584388","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Probabilistic contour mapping using oriented gradient features and SVM-bagging 基于定向梯度特征和SVM-bagging的概率等高线映射
Pub Date : 1900-01-01 DOI: 10.1109/FCV.2015.7103732
Shubhra Aich, Yong-Cheol Lee, Chil-Woo Lee
In this paper, we propose a supervised approach to find out the probabilistic mapping of semantic contours in color images. We prepare a new color image modifying the RGB color planes to incorporate reasonable within-object contrasts in all the color planes. Color gradient based features are then extracted from this altered version of color image. Next, multiple support vector machines (SVMs) are trained with disjoint sets of gradient feature sets. Finally, probabilistic decisions on the test images are made using sigmoid estimation based posterior calculations on the ensemble bagging of SVMs. We demonstrate that this SVM-bagging system is capable of boosting the probability of the pixels near the contour regions compared to that of non-contour ones.
在本文中,我们提出了一种有监督的方法来找出彩色图像中语义轮廓的概率映射。我们准备了一个新的彩色图像,修改了RGB颜色平面,在所有的颜色平面中加入了合理的对象内对比度。然后从改变后的彩色图像中提取基于颜色梯度的特征。接下来,使用不相交的梯度特征集集训练多个支持向量机(svm)。最后,使用基于后验计算的s型估计对测试图像进行概率决策。我们证明,与非轮廓区域相比,该SVM-bagging系统能够提高轮廓区域附近像素的概率。
{"title":"Probabilistic contour mapping using oriented gradient features and SVM-bagging","authors":"Shubhra Aich, Yong-Cheol Lee, Chil-Woo Lee","doi":"10.1109/FCV.2015.7103732","DOIUrl":"https://doi.org/10.1109/FCV.2015.7103732","url":null,"abstract":"In this paper, we propose a supervised approach to find out the probabilistic mapping of semantic contours in color images. We prepare a new color image modifying the RGB color planes to incorporate reasonable within-object contrasts in all the color planes. Color gradient based features are then extracted from this altered version of color image. Next, multiple support vector machines (SVMs) are trained with disjoint sets of gradient feature sets. Finally, probabilistic decisions on the test images are made using sigmoid estimation based posterior calculations on the ensemble bagging of SVMs. We demonstrate that this SVM-bagging system is capable of boosting the probability of the pixels near the contour regions compared to that of non-contour ones.","PeriodicalId":424974,"journal":{"name":"2015 21st Korea-Japan Joint Workshop on Frontiers of Computer Vision (FCV)","volume":"34 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129156060","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Estimating spectral reflectance of fabrics from high-resolution multi-band HDR images 利用高分辨率多波段HDR图像估算织物的光谱反射率
Pub Date : 1900-01-01 DOI: 10.1109/FCV.2015.7103731
Shiro Tanaka, Aki Takayanagi, M. Tsuchida, Y. Sakaguchi, Hiromi T. Tanaka
Material perception of woven fabrics is produced by the geometric factors based on microscopic structure of the fabric surface and optical factors based on reflection and transmission properties of fibers and yarns. In order to faithfully reproduce the material perception of the fabric, we observe the high-resolution multi-band high dynamic range (HDR) image of the reflection of fabric surface in one pixel (size: 5 ~ 7μm), and estimate the spectral reflectance for each pixel using the partial least squares (PLS) method. In PLS method, the training samples corresponding to the observed image and the spectral reflectance that was measured by the spectrophotometer are required. However, spectral reflectance of fabric is higher than training samples such as color patches. Therefore, the estimation accuracy decreased in the range of high luminance compared with the training samples. In this paper, we propose a method for estimating the spectral reflectance of having high reflectance, such as fabric, and show the effectiveness of the proposed method. In addition, we have confirmed that the specular reflection containing the object color is stronger on the high light area of the fabric.
机织物的材料感知是由基于织物表面微观结构的几何因素和基于纤维和纱线的反射和透射性能的光学因素共同产生的。为了真实再现织物的材料感觉,我们在单个像素(尺寸:5 ~ 7μm)上观察织物表面反射的高分辨率多波段高动态范围(HDR)图像,并利用偏最小二乘(PLS)方法估计每个像素的光谱反射率。在PLS方法中,需要与观测图像相对应的训练样本和分光光度计测量的光谱反射率。然而,织物的光谱反射率高于色块等训练样本。因此,在高亮度范围内,与训练样本相比,估计精度有所下降。本文提出了一种估算高反射率织物光谱反射率的方法,并验证了该方法的有效性。另外,我们已经确认含有物体颜色的镜面反射在织物的高光区域会更强。
{"title":"Estimating spectral reflectance of fabrics from high-resolution multi-band HDR images","authors":"Shiro Tanaka, Aki Takayanagi, M. Tsuchida, Y. Sakaguchi, Hiromi T. Tanaka","doi":"10.1109/FCV.2015.7103731","DOIUrl":"https://doi.org/10.1109/FCV.2015.7103731","url":null,"abstract":"Material perception of woven fabrics is produced by the geometric factors based on microscopic structure of the fabric surface and optical factors based on reflection and transmission properties of fibers and yarns. In order to faithfully reproduce the material perception of the fabric, we observe the high-resolution multi-band high dynamic range (HDR) image of the reflection of fabric surface in one pixel (size: 5 ~ 7μm), and estimate the spectral reflectance for each pixel using the partial least squares (PLS) method. In PLS method, the training samples corresponding to the observed image and the spectral reflectance that was measured by the spectrophotometer are required. However, spectral reflectance of fabric is higher than training samples such as color patches. Therefore, the estimation accuracy decreased in the range of high luminance compared with the training samples. In this paper, we propose a method for estimating the spectral reflectance of having high reflectance, such as fabric, and show the effectiveness of the proposed method. In addition, we have confirmed that the specular reflection containing the object color is stronger on the high light area of the fabric.","PeriodicalId":424974,"journal":{"name":"2015 21st Korea-Japan Joint Workshop on Frontiers of Computer Vision (FCV)","volume":"27 23 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128677717","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Centralization problem of contacting interaction in multiple object tracking 多目标跟踪中接触交互的集中问题
Pub Date : 1900-01-01 DOI: 10.1109/FCV.2015.7103721
Chi-Min Oh, Abdullah Nazib, Hao Xu, Joopil Moon, Chil-Woo Lee
In this paper we introduce a definition of centralization problem in multiple object tracking. The centralization problem occurs when more than two targets are contacting in close area. As a result, their trackers could be wrongly located in the center position of their targets. It is especially found when the appearances and shape of contacting targets are identical and the likelihood score for each tracker can be maximum value in their central position among their targets. After centralization problem their tracker could track wrong targets. Therefore we propose new interaction model to avoid centralization problem and our performance shows that almost all of centralization problem can be removed.
本文介绍了多目标跟踪中集中问题的定义。当两个以上的目标在近距离接触时,就会出现集中问题。因此,他们的跟踪器可能被错误地定位在目标的中心位置。特别是当接触目标的外观和形状相同时,每个跟踪器的似然得分在其目标的中心位置可以达到最大值。在集中问题之后,他们的跟踪器可能会跟踪错误的目标。因此,我们提出了新的交互模型来避免集中化问题,我们的性能表明,几乎所有的集中化问题都可以被消除。
{"title":"Centralization problem of contacting interaction in multiple object tracking","authors":"Chi-Min Oh, Abdullah Nazib, Hao Xu, Joopil Moon, Chil-Woo Lee","doi":"10.1109/FCV.2015.7103721","DOIUrl":"https://doi.org/10.1109/FCV.2015.7103721","url":null,"abstract":"In this paper we introduce a definition of centralization problem in multiple object tracking. The centralization problem occurs when more than two targets are contacting in close area. As a result, their trackers could be wrongly located in the center position of their targets. It is especially found when the appearances and shape of contacting targets are identical and the likelihood score for each tracker can be maximum value in their central position among their targets. After centralization problem their tracker could track wrong targets. Therefore we propose new interaction model to avoid centralization problem and our performance shows that almost all of centralization problem can be removed.","PeriodicalId":424974,"journal":{"name":"2015 21st Korea-Japan Joint Workshop on Frontiers of Computer Vision (FCV)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130770376","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Object pose estimation using category information from a single image 利用单幅图像的类别信息进行目标姿态估计
Pub Date : 1900-01-01 DOI: 10.1109/FCV.2015.7103728
Shunsuke Shimizu, H. Koyasu, Yoshinori Kobayashi, Y. Kuno
3D object pose estimation is one of the most important challenges in the field of computer vision. A huge amount of image resources such as images on the web or photos taken before can be utilized if the system can estimate the 3D pose from a single image. On the other hand, the object's category and position on the image can be estimated by using state of the techniques in the general object recognition. We propose a method for 3D pose estimation from a single image on the basis of the known object category and position. We employ Regression Forests as the machine learning algorithm and HOG features as the input vectors. The regression function is created based on HOG features which express the differences in shapes depending on the viewing directions and corresponding poses. We evaluate the accuracy of pose estimation by using multiple objects with different categories.
三维物体姿态估计是计算机视觉领域的重要课题之一。如果系统可以从单张图像中估计出3D姿态,则可以利用大量的图像资源,例如网络上的图像或之前拍摄的照片。另一方面,利用一般物体识别技术的状态,可以估计出物体在图像上的类别和位置。提出了一种基于已知目标类别和位置的单幅图像的三维姿态估计方法。我们采用回归森林作为机器学习算法,HOG特征作为输入向量。回归函数是基于HOG特征创建的,HOG特征表达了不同的观察方向和相应的姿态的形状差异。我们通过使用不同类别的多个目标来评估姿态估计的准确性。
{"title":"Object pose estimation using category information from a single image","authors":"Shunsuke Shimizu, H. Koyasu, Yoshinori Kobayashi, Y. Kuno","doi":"10.1109/FCV.2015.7103728","DOIUrl":"https://doi.org/10.1109/FCV.2015.7103728","url":null,"abstract":"3D object pose estimation is one of the most important challenges in the field of computer vision. A huge amount of image resources such as images on the web or photos taken before can be utilized if the system can estimate the 3D pose from a single image. On the other hand, the object's category and position on the image can be estimated by using state of the techniques in the general object recognition. We propose a method for 3D pose estimation from a single image on the basis of the known object category and position. We employ Regression Forests as the machine learning algorithm and HOG features as the input vectors. The regression function is created based on HOG features which express the differences in shapes depending on the viewing directions and corresponding poses. We evaluate the accuracy of pose estimation by using multiple objects with different categories.","PeriodicalId":424974,"journal":{"name":"2015 21st Korea-Japan Joint Workshop on Frontiers of Computer Vision (FCV)","volume":"6 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133180731","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Query expansion with pairwise learning in object retrieval challenge 对象检索挑战中基于成对学习的查询扩展
Pub Date : 1900-01-01 DOI: 10.1109/FCV.2015.7103703
Hao Liu, Atsushi Shimada, Xing Xu, H. Nagahara, Hideaki Uchiyama, R. Taniguchi
Making a reasonable ranking on images in dataset is one of the main objectives for object retrieval challenge, and in this paper we intend to improve the ranking quality. We follow the idea of query expansion in previous researches. Based on the use of bag-of-visual-words model, tf-idf scoring and spatial verification, previous method applied a pointwise style learning in query expansion stage, using but not fully exploring verification results. We intend to extend their learning approach for better discriminative power in retrieval. In re-ranking stage we propose a method using pairwise learning, instead of pointwise learning previously used. We could obtain more reliable ranking on a shortlist of examples. If this verification itself is reliable, a good re-ranking should best preserve this sub-ranking order. Thus in our proposed method, we are motivated to leverage a pairwise learning method to incorporate the ranking sequential information more efficiently. We evaluate and compare our proposed method with previous methods over Oxford 5k dataset, a standard benchmark dataset, where our method achieve better mean average precision and showed better discriminative power.
对数据集中的图像进行合理的排序是对象检索的主要目标之一,本文旨在提高排序质量。我们在之前的研究中遵循了查询扩展的思想。之前的方法在使用视觉词袋模型、tf-idf评分和空间验证的基础上,在查询扩展阶段采用点向风格学习,使用但未充分探索验证结果。我们打算扩展他们的学习方法,以获得更好的检索判别能力。在重新排序阶段,我们提出了一种使用成对学习的方法来代替以前使用的点学习方法。我们可以在一个短名单上获得更可靠的排名。如果这个验证本身是可靠的,那么一个好的重新排序应该最好地保留这个子排序顺序。因此,在我们提出的方法中,我们被激励利用成对学习方法来更有效地整合排名顺序信息。我们在牛津5k数据集(一个标准的基准数据集)上评估并比较了我们提出的方法,在牛津5k数据集上,我们的方法获得了更好的平均精度和更好的判别能力。
{"title":"Query expansion with pairwise learning in object retrieval challenge","authors":"Hao Liu, Atsushi Shimada, Xing Xu, H. Nagahara, Hideaki Uchiyama, R. Taniguchi","doi":"10.1109/FCV.2015.7103703","DOIUrl":"https://doi.org/10.1109/FCV.2015.7103703","url":null,"abstract":"Making a reasonable ranking on images in dataset is one of the main objectives for object retrieval challenge, and in this paper we intend to improve the ranking quality. We follow the idea of query expansion in previous researches. Based on the use of bag-of-visual-words model, tf-idf scoring and spatial verification, previous method applied a pointwise style learning in query expansion stage, using but not fully exploring verification results. We intend to extend their learning approach for better discriminative power in retrieval. In re-ranking stage we propose a method using pairwise learning, instead of pointwise learning previously used. We could obtain more reliable ranking on a shortlist of examples. If this verification itself is reliable, a good re-ranking should best preserve this sub-ranking order. Thus in our proposed method, we are motivated to leverage a pairwise learning method to incorporate the ranking sequential information more efficiently. We evaluate and compare our proposed method with previous methods over Oxford 5k dataset, a standard benchmark dataset, where our method achieve better mean average precision and showed better discriminative power.","PeriodicalId":424974,"journal":{"name":"2015 21st Korea-Japan Joint Workshop on Frontiers of Computer Vision (FCV)","volume":"79 2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131593388","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
期刊
2015 21st Korea-Japan Joint Workshop on Frontiers of Computer Vision (FCV)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1