首页 > 最新文献

2018 IEEE Applied Imagery Pattern Recognition Workshop (AIPR)最新文献

英文 中文
Automated Detection of Colorspace Via Convolutional Neural Network 基于卷积神经网络的色彩空间自动检测
Pub Date : 2018-10-01 DOI: 10.1109/AIPR.2018.8707374
S. Maxwell, M. Kilcher, Alexander Benasutti, Brandon Siebert, Warren Seto, Olivia Shanley, Larry Pearlstein
Prior to the advent of ITU-R Recommendation BT.709 the overwhelming majority of compressed digital video and imagery used the colorspace conversion matrix specified in ITU-R Recommendation BT.601. The introduction of high-definition video formats led to the adoption of Rec. BT.709 for use in colorspace conversion by new systems, and this resulted in confusion in the industry. Specifically, video decoders may not be able to determine the correct matrix to use for converting from the luma/chroma representation used for coding, to the Red-Green-Blue representation needed for display. This confusion has led to a situation where some viewers of decompressed video streams experience subtle, but noticeable, errors in coloration. We have successfully developed and trained a deep convolutional neural network to address this heretofore unsolved problem. We obtained outstanding accuracy on ImageNet data, and on YouTube video frames, and our work can be expected to lead to more accurate color rendering delivered to users of digital imaging and video systems.
在ITU-R建议书BT.709出台之前,绝大多数压缩数字视频和图像使用ITU-R建议书BT.601规定的色彩空间转换矩阵。高清视频格式的引入导致新系统采用Rec. BT.709用于色彩空间转换,这导致了行业的混乱。具体来说,视频解码器可能无法确定用于从用于编码的亮度/色度表示转换到显示所需的红-绿-蓝表示的正确矩阵。这种混淆导致了这样一种情况,即一些观看解压缩视频流的人会经历微妙但明显的色彩错误。我们已经成功地开发并训练了一个深度卷积神经网络来解决这个迄今为止尚未解决的问题。我们在ImageNet数据和YouTube视频帧上获得了出色的准确性,我们的工作有望为数字成像和视频系统的用户提供更准确的色彩渲染。
{"title":"Automated Detection of Colorspace Via Convolutional Neural Network","authors":"S. Maxwell, M. Kilcher, Alexander Benasutti, Brandon Siebert, Warren Seto, Olivia Shanley, Larry Pearlstein","doi":"10.1109/AIPR.2018.8707374","DOIUrl":"https://doi.org/10.1109/AIPR.2018.8707374","url":null,"abstract":"Prior to the advent of ITU-R Recommendation BT.709 the overwhelming majority of compressed digital video and imagery used the colorspace conversion matrix specified in ITU-R Recommendation BT.601. The introduction of high-definition video formats led to the adoption of Rec. BT.709 for use in colorspace conversion by new systems, and this resulted in confusion in the industry. Specifically, video decoders may not be able to determine the correct matrix to use for converting from the luma/chroma representation used for coding, to the Red-Green-Blue representation needed for display. This confusion has led to a situation where some viewers of decompressed video streams experience subtle, but noticeable, errors in coloration. We have successfully developed and trained a deep convolutional neural network to address this heretofore unsolved problem. We obtained outstanding accuracy on ImageNet data, and on YouTube video frames, and our work can be expected to lead to more accurate color rendering delivered to users of digital imaging and video systems.","PeriodicalId":230582,"journal":{"name":"2018 IEEE Applied Imagery Pattern Recognition Workshop (AIPR)","volume":"5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132958460","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An Authentication System based on Hybrid Fusion of Finger-Shapes & Geometry 一种基于手指形状与几何混合融合的身份认证系统
Pub Date : 2018-10-01 DOI: 10.1109/AIPR.2018.8707382
Neha Mittal, M. Hanmandlu, S. Vasikarla
Owing to the ever increasing demand of privacy protection and security concerns, hand geometry based biometric system aimed at addressing these concerns is developed. In this study we have developed an authentication system which is based on finger shapes. The shapes of fingers are considered as patterns for extracting features using Eigen vector method and discrete wavelet transform. There are four feature types that include (i) Frequency and power content using Eigen vector method, (ii) Pisaranko’s method (iii) Wavelet entropy of individual fingers, and (iv) Specific area of finger.
由于对隐私保护和安全问题的需求日益增长,基于手几何的生物识别系统旨在解决这些问题。在这项研究中,我们开发了一个基于手指形状的身份验证系统。采用特征向量法和离散小波变换,将手指的形状作为模式提取特征。有四种特征类型,包括(i)使用特征向量法的频率和功率含量,(ii) Pisaranko的方法,(iii)单个手指的小波熵,(iv)手指的特定面积。
{"title":"An Authentication System based on Hybrid Fusion of Finger-Shapes & Geometry","authors":"Neha Mittal, M. Hanmandlu, S. Vasikarla","doi":"10.1109/AIPR.2018.8707382","DOIUrl":"https://doi.org/10.1109/AIPR.2018.8707382","url":null,"abstract":"Owing to the ever increasing demand of privacy protection and security concerns, hand geometry based biometric system aimed at addressing these concerns is developed. In this study we have developed an authentication system which is based on finger shapes. The shapes of fingers are considered as patterns for extracting features using Eigen vector method and discrete wavelet transform. There are four feature types that include (i) Frequency and power content using Eigen vector method, (ii) Pisaranko’s method (iii) Wavelet entropy of individual fingers, and (iv) Specific area of finger.","PeriodicalId":230582,"journal":{"name":"2018 IEEE Applied Imagery Pattern Recognition Workshop (AIPR)","volume":"21 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134174726","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Robust Multi-object Tracking for Wide Area Motion Imagery 广域运动图像鲁棒多目标跟踪
Pub Date : 2018-10-01 DOI: 10.1109/AIPR.2018.8707377
Noor M. Al-Shakarji, F. Bunyak, G. Seetharaman, K. Palaniappan
Multi-object tracking implemented on airborne wide area motion imagery (WAMI) is still challenging problem in computer vision applications. Extremely camera motion, low frame rate, rapid appearance changes, and occlusion by different objects are the most challenges. Data association, link detected object in the current frame with the existing tracked objects, is the most challenging part for multi-object tracking algorithms. The ambiguity of data association increases in WAMI datasets because objects in the scenes suffer form the lack of rich feature descriptions beside the closeness to each other, and inaccurate object movement displacement. In this paper, detection-based multi-object tracking system that uses a two-step data association scheme to ensure high tracking accuracy and continuity. The first step ensures having reliable short-term tracklets using only spatial information. The second step links tracklets globally and reduces matching hypotheses using discriminative features and tracklets history. Our proposed tracker tested on wide area imagery ABQ dataset [1]. MOTChallage [2] evaluation metrics have been used to evaluate the performance compared to some multi-object-tracking baselines for IWTS42018 [3] and VisDrone2018 [4] challenges. Our tracker shows promising results compared to those trackers.
机载广域运动图像(WAMI)上的多目标跟踪是计算机视觉应用中的一个难题。极端的相机运动,低帧率,快速的外观变化,以及不同物体的遮挡是最大的挑战。数据关联是多目标跟踪算法中最具挑战性的部分,它将当前帧中检测到的目标与已有的跟踪对象联系起来。WAMI数据集中数据关联的模糊性增加,因为场景中的物体除了彼此接近之外,还缺乏丰富的特征描述,物体运动位移不准确。本文提出了基于检测的多目标跟踪系统,该系统采用两步数据关联方案来保证高跟踪精度和连续性。第一步确保仅使用空间信息获得可靠的短期跟踪。第二步,全局链接tracklets,并使用判别特征和tracklets历史减少匹配假设。我们提出的跟踪器在广域图像ABQ数据集[1]上进行了测试。与IWTS42018[3]和VisDrone2018[4]挑战的一些多目标跟踪基线相比,motchallenge[2]评估指标已被用于评估性能。与那些跟踪器相比,我们的跟踪器显示出有希望的结果。
{"title":"Robust Multi-object Tracking for Wide Area Motion Imagery","authors":"Noor M. Al-Shakarji, F. Bunyak, G. Seetharaman, K. Palaniappan","doi":"10.1109/AIPR.2018.8707377","DOIUrl":"https://doi.org/10.1109/AIPR.2018.8707377","url":null,"abstract":"Multi-object tracking implemented on airborne wide area motion imagery (WAMI) is still challenging problem in computer vision applications. Extremely camera motion, low frame rate, rapid appearance changes, and occlusion by different objects are the most challenges. Data association, link detected object in the current frame with the existing tracked objects, is the most challenging part for multi-object tracking algorithms. The ambiguity of data association increases in WAMI datasets because objects in the scenes suffer form the lack of rich feature descriptions beside the closeness to each other, and inaccurate object movement displacement. In this paper, detection-based multi-object tracking system that uses a two-step data association scheme to ensure high tracking accuracy and continuity. The first step ensures having reliable short-term tracklets using only spatial information. The second step links tracklets globally and reduces matching hypotheses using discriminative features and tracklets history. Our proposed tracker tested on wide area imagery ABQ dataset [1]. MOTChallage [2] evaluation metrics have been used to evaluate the performance compared to some multi-object-tracking baselines for IWTS42018 [3] and VisDrone2018 [4] challenges. Our tracker shows promising results compared to those trackers.","PeriodicalId":230582,"journal":{"name":"2018 IEEE Applied Imagery Pattern Recognition Workshop (AIPR)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128075074","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Visualizing Compression of Deep Learning Models for Classification 分类中深度学习模型的可视化压缩
Pub Date : 2018-10-01 DOI: 10.1109/AIPR.2018.8707381
Marissa Dotter, C. Ward
Deep learning models have made great strides in tasks like classification and object detection. However, these models are often computationally intensive, require vast amounts of data in the domain, and typically contain millions or even billions of parameters. They are also relative black-boxes when it comes to being able to interpret and analyze their functionality on data or evaluating the suitability of the network for the data that is available. To address these issues, we investigate compression techniques available off-the-shelf that aid in reducing the dimensionality of the parameter space within a Convolutional Neural Network. In this way, compression will allow us to interpret and evaluate the network more efficiently as only important features will be propagated throughout the network.
深度学习模型在分类和目标检测等任务上取得了巨大的进步。然而,这些模型通常是计算密集型的,需要领域中大量的数据,并且通常包含数百万甚至数十亿个参数。当涉及到能够解释和分析它们对数据的功能或评估网络对可用数据的适用性时,它们也是相对的黑箱。为了解决这些问题,我们研究了现有的压缩技术,这些技术有助于降低卷积神经网络中参数空间的维数。通过这种方式,压缩将允许我们更有效地解释和评估网络,因为只有重要的特征将在整个网络中传播。
{"title":"Visualizing Compression of Deep Learning Models for Classification","authors":"Marissa Dotter, C. Ward","doi":"10.1109/AIPR.2018.8707381","DOIUrl":"https://doi.org/10.1109/AIPR.2018.8707381","url":null,"abstract":"Deep learning models have made great strides in tasks like classification and object detection. However, these models are often computationally intensive, require vast amounts of data in the domain, and typically contain millions or even billions of parameters. They are also relative black-boxes when it comes to being able to interpret and analyze their functionality on data or evaluating the suitability of the network for the data that is available. To address these issues, we investigate compression techniques available off-the-shelf that aid in reducing the dimensionality of the parameter space within a Convolutional Neural Network. In this way, compression will allow us to interpret and evaluate the network more efficiently as only important features will be propagated throughout the network.","PeriodicalId":230582,"journal":{"name":"2018 IEEE Applied Imagery Pattern Recognition Workshop (AIPR)","volume":"5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114909879","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Segmentation of Thermal Breast Images Using Convolutional and Deconvolutional Neural Networks 基于卷积和反卷积神经网络的乳房热图像分割
Pub Date : 2018-10-01 DOI: 10.1109/AIPR.2018.8707379
Shuyue Guan, Nada Kamona, M. Loew
Breast cancer is the second leading cause of death for women in the U.S. Early detection of breast cancer has been shown to be the key to higher survival rates for breast cancer patients. We are investigating infrared thermography as a noninvasive adjunctive to mammography for breast screening. Thermal imaging is safe, radiation-free, pain-free, and non-contact. Segmentation of breast area from the acquired thermal images will help limit the area for tumor search and reduce the time and effort needed for manual hand segmentation. Autoencoder-like convolutional and deconvolutional neural networks (C-DCNN) are promising computational approaches to automatically segment breast areas in thermal images. In this study, we apply the C-DCNN to segment breast areas from our thermal breast images database, which we are collecting in our clinical trials by imaging breast cancer patients with our infrared camera (N2 Imager). For training the C-DCNN, the inputs are 132 gray-value thermal images and the corresponding manually-cropped breast area images (binary masks to designate the breast areas). For testing, we input thermal images to the trained C-DCNN and the output after post-processing are the binary breast-area images. Cross-validation and comparison with the ground-truth images show that the C-DCNN is a promising method to segment breast areas. The results demonstrate the capability of C-DCNN to learn essential features of breast regions and delineate them in thermal images.
在美国,乳腺癌是导致女性死亡的第二大原因。早发现乳腺癌已被证明是提高乳腺癌患者存活率的关键。我们正在研究红外热成像作为乳房x光检查的非侵入性辅助手段。热成像安全、无辐射、无痛、非接触。从获得的热图像中分割乳房区域有助于限制肿瘤搜索的区域,减少人工分割所需的时间和精力。类自编码器卷积和反卷积神经网络(C-DCNN)是热图像中乳房区域自动分割的一种很有前途的计算方法。在这项研究中,我们应用C-DCNN从我们的乳房热图像数据库中分割乳房区域,这些图像是我们在临床试验中通过使用我们的红外相机(N2 Imager)对乳腺癌患者进行成像而收集的。为了训练C-DCNN,输入是132张灰度值热图像和相应的人工裁剪的乳房区域图像(用二值蒙版来指定乳房区域)。为了进行测试,我们将热图像输入训练好的C-DCNN,后处理后输出的是二值乳房区域图像。交叉验证和与真实图像的对比表明,C-DCNN是一种很有前途的乳房区域分割方法。结果表明,C-DCNN能够学习乳房区域的基本特征,并在热图像中描绘它们。
{"title":"Segmentation of Thermal Breast Images Using Convolutional and Deconvolutional Neural Networks","authors":"Shuyue Guan, Nada Kamona, M. Loew","doi":"10.1109/AIPR.2018.8707379","DOIUrl":"https://doi.org/10.1109/AIPR.2018.8707379","url":null,"abstract":"Breast cancer is the second leading cause of death for women in the U.S. Early detection of breast cancer has been shown to be the key to higher survival rates for breast cancer patients. We are investigating infrared thermography as a noninvasive adjunctive to mammography for breast screening. Thermal imaging is safe, radiation-free, pain-free, and non-contact. Segmentation of breast area from the acquired thermal images will help limit the area for tumor search and reduce the time and effort needed for manual hand segmentation. Autoencoder-like convolutional and deconvolutional neural networks (C-DCNN) are promising computational approaches to automatically segment breast areas in thermal images. In this study, we apply the C-DCNN to segment breast areas from our thermal breast images database, which we are collecting in our clinical trials by imaging breast cancer patients with our infrared camera (N2 Imager). For training the C-DCNN, the inputs are 132 gray-value thermal images and the corresponding manually-cropped breast area images (binary masks to designate the breast areas). For testing, we input thermal images to the trained C-DCNN and the output after post-processing are the binary breast-area images. Cross-validation and comparison with the ground-truth images show that the C-DCNN is a promising method to segment breast areas. The results demonstrate the capability of C-DCNN to learn essential features of breast regions and delineate them in thermal images.","PeriodicalId":230582,"journal":{"name":"2018 IEEE Applied Imagery Pattern Recognition Workshop (AIPR)","volume":"26 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133206014","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Acquiring Abstract Visual Knowledge of the Real-World Environment for Autonomous Vehicles 为自动驾驶汽车获取真实世界环境的抽象视觉知识
Pub Date : 2018-07-02 DOI: 10.1109/AIPR.2018.8707386
I. F. Ghalyan, V. Kapila
This paper considers the problem of modeling the surrounding environment of a driven car by using the images captured by a dash cam during the driving process. Inspired from a human driver’s interpretation of the car’s surrounding environment, an abstract representation of the environment is developed that can facilitate in decision-making to prevent the car’s collisions with surrounding objects. The proposed technique for modeling the car’s surrounding environment utilizes the dash cam to capture images as the car is driven facing multiple situations and obstacles. By relying on the human driver’s interpretation of various driving scenarios, the images of the car’s surrounding environment are manually grouped into classes that reflect the driver’s abstract knowledge. Grouping the images allows the formulation of knowledge transfer process from the human driver to an autonomous vehicle as a classification problem, producing a meaningful and efficient representation of models arising from real-world scenarios. The framework of convolutional neural networks (CNN) is employed to model the surrounding environment of the driven car, encapsulating the abstract knowledge of the human driver. The proposed modeling approach is applied to determine its efficacy in two experimental scenarios. In the first experiment, a highway driving scenario is considered with three classes. Alternatively, in the second experiment, a scenario of driving in a residential area is addressed with six classes. Excellent modeling performance is reported for both experiments. Comparisons conducted with alternative image classification techniques reveal the superiority of the CNN for modeling the considered driving scenarios.
本文研究了利用行车记录仪在行驶过程中采集的图像对汽车周围环境进行建模的问题。受人类驾驶员对汽车周围环境的理解启发,开发了一种抽象的环境表示,可以促进决策,防止汽车与周围物体发生碰撞。所提出的汽车周围环境建模技术利用行车记录仪在汽车面临多种情况和障碍物时捕捉图像。通过依靠人类驾驶员对各种驾驶场景的解释,汽车周围环境的图像被手动分组为反映驾驶员抽象知识的类别。将图像分组可以将从人类驾驶员到自动驾驶汽车的知识转移过程作为一个分类问题进行表述,从而产生来自现实世界场景的模型的有意义和有效的表示。采用卷积神经网络(CNN)框架对被驾驶汽车周围环境进行建模,封装人类驾驶员的抽象知识。在两种实验场景下,应用所提出的建模方法来确定其有效性。在第一个实验中,高速公路驾驶场景被考虑为三个类别。另外,在第二个实验中,在居民区驾驶的场景被分为六个类别。这两个实验的建模性能都很好。与其他图像分类技术的比较揭示了CNN在建模所考虑的驾驶场景方面的优势。
{"title":"Acquiring Abstract Visual Knowledge of the Real-World Environment for Autonomous Vehicles","authors":"I. F. Ghalyan, V. Kapila","doi":"10.1109/AIPR.2018.8707386","DOIUrl":"https://doi.org/10.1109/AIPR.2018.8707386","url":null,"abstract":"This paper considers the problem of modeling the surrounding environment of a driven car by using the images captured by a dash cam during the driving process. Inspired from a human driver’s interpretation of the car’s surrounding environment, an abstract representation of the environment is developed that can facilitate in decision-making to prevent the car’s collisions with surrounding objects. The proposed technique for modeling the car’s surrounding environment utilizes the dash cam to capture images as the car is driven facing multiple situations and obstacles. By relying on the human driver’s interpretation of various driving scenarios, the images of the car’s surrounding environment are manually grouped into classes that reflect the driver’s abstract knowledge. Grouping the images allows the formulation of knowledge transfer process from the human driver to an autonomous vehicle as a classification problem, producing a meaningful and efficient representation of models arising from real-world scenarios. The framework of convolutional neural networks (CNN) is employed to model the surrounding environment of the driven car, encapsulating the abstract knowledge of the human driver. The proposed modeling approach is applied to determine its efficacy in two experimental scenarios. In the first experiment, a highway driving scenario is considered with three classes. Alternatively, in the second experiment, a scenario of driving in a residential area is addressed with six classes. Excellent modeling performance is reported for both experiments. Comparisons conducted with alternative image classification techniques reveal the superiority of the CNN for modeling the considered driving scenarios.","PeriodicalId":230582,"journal":{"name":"2018 IEEE Applied Imagery Pattern Recognition Workshop (AIPR)","volume":"68 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-07-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122290969","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Learning Robot-Object Distance Using Bayesian Regression with Application to A Collision Avoidance Scenario 使用贝叶斯回归学习机器人-物体距离及其在避碰场景中的应用
Pub Date : 2018-07-02 DOI: 10.1109/AIPR.2018.8707380
I. F. Ghalyan, Aneesh Jaydeep, V. Kapila
In many practical situations, robots may encounter objects moving in their work space, resulting in undesirable consequences for either the robots or the moving objects. Such situations often call for sensing arrangements that can produce planar images along with depth measurements, e.g., Kinect sensors, to estimate the position of the moving object in 3-D space. In this paper, we aim to estimate the relative distance of a moving object along the axis orthogonal to a camera lens plane, thus relaxing the need to rely on depth measurements that are often noisy when the object is too close to the sensor. Specifically, multiple images of an object, with distinct orthogonal distances, are firstly captured. In this step, the object’s distance from the camera is measured and the normalized area, which is the normalized sum of pixels, of the object is computed. Both computed normalized area and measured distance are filtered using a Gaussian smoothing filter (GSF). Next, a Bayesian statistical model is developed to map the computed normalized area with the measured distance. The developed Bayesian linear model allows to predict the distance between the camera sensor (or robot) and the object given the normalized computed area, obtained from the 2-D images, of the object. To evaluate the performance of the relative distance estimation process, a test stand was built that consists of a robot equipped with a camera. During the learning process of the statistical model, an ultrasonic sensor was used for measuring the distance corresponding to the captured images. After learning the model, the ultrasonic sensor was removed and excellent performance was achieved when using the developed model in estimating the distance of an object, a human hand carrying a measurement tape, moving back and forth along the axis normal to the camera plane.
在许多实际情况下,机器人可能会遇到在其工作空间中移动的物体,从而导致机器人或移动物体的不良后果。这种情况通常需要能够产生平面图像和深度测量的传感装置,例如Kinect传感器,以估计移动物体在3d空间中的位置。在本文中,我们的目标是估计移动物体沿着与相机镜头平面正交的轴的相对距离,从而减少依赖深度测量的需要,当物体太靠近传感器时,深度测量通常会产生噪声。具体来说,首先捕获具有不同正交距离的物体的多幅图像。在这一步中,测量物体到相机的距离,并计算物体的归一化面积,即归一化像素之和。计算归一化面积和测量距离都使用高斯平滑滤波器(GSF)进行滤波。接下来,建立贝叶斯统计模型,将计算的归一化面积与测量的距离进行映射。开发的贝叶斯线性模型允许预测相机传感器(或机器人)和物体之间的距离,给定归一化计算面积,从物体的二维图像中获得。为了评估相对距离估计过程的性能,建立了一个由配备相机的机器人组成的试验台。在统计模型的学习过程中,使用超声波传感器测量捕获图像对应的距离。在学习了模型后,将超声波传感器移除,使用所开发的模型在估计物体的距离时取得了优异的性能,一个人的手拿着卷尺,沿着与相机平面垂直的轴来回移动。
{"title":"Learning Robot-Object Distance Using Bayesian Regression with Application to A Collision Avoidance Scenario","authors":"I. F. Ghalyan, Aneesh Jaydeep, V. Kapila","doi":"10.1109/AIPR.2018.8707380","DOIUrl":"https://doi.org/10.1109/AIPR.2018.8707380","url":null,"abstract":"In many practical situations, robots may encounter objects moving in their work space, resulting in undesirable consequences for either the robots or the moving objects. Such situations often call for sensing arrangements that can produce planar images along with depth measurements, e.g., Kinect sensors, to estimate the position of the moving object in 3-D space. In this paper, we aim to estimate the relative distance of a moving object along the axis orthogonal to a camera lens plane, thus relaxing the need to rely on depth measurements that are often noisy when the object is too close to the sensor. Specifically, multiple images of an object, with distinct orthogonal distances, are firstly captured. In this step, the object’s distance from the camera is measured and the normalized area, which is the normalized sum of pixels, of the object is computed. Both computed normalized area and measured distance are filtered using a Gaussian smoothing filter (GSF). Next, a Bayesian statistical model is developed to map the computed normalized area with the measured distance. The developed Bayesian linear model allows to predict the distance between the camera sensor (or robot) and the object given the normalized computed area, obtained from the 2-D images, of the object. To evaluate the performance of the relative distance estimation process, a test stand was built that consists of a robot equipped with a camera. During the learning process of the statistical model, an ultrasonic sensor was used for measuring the distance corresponding to the captured images. After learning the model, the ultrasonic sensor was removed and excellent performance was achieved when using the developed model in estimating the distance of an object, a human hand carrying a measurement tape, moving back and forth along the axis normal to the camera plane.","PeriodicalId":230582,"journal":{"name":"2018 IEEE Applied Imagery Pattern Recognition Workshop (AIPR)","volume":"21 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-07-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131584509","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
期刊
2018 IEEE Applied Imagery Pattern Recognition Workshop (AIPR)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1