Pub Date : 2012-11-12DOI: 10.1109/MFI.2012.6343033
D. Faria, J. Lobo, J. Dias
In this work we use hand configuration and contact points during in-hand object exploration to identify the manipulated objects. Different contact points associated to an object shape can be represented in a latent space and lie on a lower dimensional non-linear manifold in the contact points space which is suitable for modelling and recognition. Associating and learning hand configurations to specific objects by means of Gaussian mixture models, later by identifying the hand configuration during the in-hand object exploration we can generate hypotheses of candidate objects to be identified. This process selects a set of the most probable objects from a database. The accumulated set of contact points (partial volume of the object shape) during the object in-hand exploration is matched to the set selected from the database (most probable candidate objects). Results are presented for human manipulation of objects, but this can also be applied to artificial hands, although we have not addressed the hand control, only the object identification.
{"title":"Identifying objects from hand configurations during in-hand exploration","authors":"D. Faria, J. Lobo, J. Dias","doi":"10.1109/MFI.2012.6343033","DOIUrl":"https://doi.org/10.1109/MFI.2012.6343033","url":null,"abstract":"In this work we use hand configuration and contact points during in-hand object exploration to identify the manipulated objects. Different contact points associated to an object shape can be represented in a latent space and lie on a lower dimensional non-linear manifold in the contact points space which is suitable for modelling and recognition. Associating and learning hand configurations to specific objects by means of Gaussian mixture models, later by identifying the hand configuration during the in-hand object exploration we can generate hypotheses of candidate objects to be identified. This process selects a set of the most probable objects from a database. The accumulated set of contact points (partial volume of the object shape) during the object in-hand exploration is matched to the set selected from the database (most probable candidate objects). Results are presented for human manipulation of objects, but this can also be applied to artificial hands, although we have not addressed the hand control, only the object identification.","PeriodicalId":103145,"journal":{"name":"2012 IEEE International Conference on Multisensor Fusion and Integration for Intelligent Systems (MFI)","volume":"57 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134078928","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2012-11-12DOI: 10.1109/MFI.2012.6343070
F. García, A. D. L. Escalera, J. M. Armingol, F. Jiménez
Road safety applications require the most reliable and trustable sensors. Context information plays also a key role, adding trustability and allowing the study of the interactions and the danger inherent to them. Vehicle dynamics, dimensions... can be very useful to avoid misdetections when performing vehicle detection and tracking (fusion levels 0 and 1). Traffic safety information is mandatory for fusion levels 2 and 3 by evaluating the interactions and the danger involved in any detection. All this information is context information that was used in this application to enhance the capacity of the sensors, providing a complete and multilevel fusion application. Present application use three sensors: laser scanner, computer vision and inertial system, the information given by these sensors is completed with context information, providing reliable vehicle detection and danger evaluation. Test results are provided to check the usability of the detection algorithm.
{"title":"Context aided fusion procedure for road safety application","authors":"F. García, A. D. L. Escalera, J. M. Armingol, F. Jiménez","doi":"10.1109/MFI.2012.6343070","DOIUrl":"https://doi.org/10.1109/MFI.2012.6343070","url":null,"abstract":"Road safety applications require the most reliable and trustable sensors. Context information plays also a key role, adding trustability and allowing the study of the interactions and the danger inherent to them. Vehicle dynamics, dimensions... can be very useful to avoid misdetections when performing vehicle detection and tracking (fusion levels 0 and 1). Traffic safety information is mandatory for fusion levels 2 and 3 by evaluating the interactions and the danger involved in any detection. All this information is context information that was used in this application to enhance the capacity of the sensors, providing a complete and multilevel fusion application. Present application use three sensors: laser scanner, computer vision and inertial system, the information given by these sensors is completed with context information, providing reliable vehicle detection and danger evaluation. Test results are provided to check the usability of the detection algorithm.","PeriodicalId":103145,"journal":{"name":"2012 IEEE International Conference on Multisensor Fusion and Integration for Intelligent Systems (MFI)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133296079","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2012-11-12DOI: 10.1109/MFI.2012.6343053
Katharine Brigham, B. Kumar, N. Rao, Qiang Liu, Xin Wang
We consider a network of sensors wherein the state estimates are sent from sensors to a fusion center to generate a global state estimate. The underlying fusion algorithm affects the performance measure QCC(τ) (with subscripts CC indicating the effects of the communications and computing quality) of the global state estimate computed within the allocated time τ. We present a probabilistic performance bound on QCC(τ) as a function of the distributions of state estimates, communications parameters as well as the fusion algorithm. We present simulations of simplified scenarios to illustrate the qualitative effects of different fusers, and system-level simulations to complement the analytical results.
{"title":"On state fusers over long-haul sensor networks","authors":"Katharine Brigham, B. Kumar, N. Rao, Qiang Liu, Xin Wang","doi":"10.1109/MFI.2012.6343053","DOIUrl":"https://doi.org/10.1109/MFI.2012.6343053","url":null,"abstract":"We consider a network of sensors wherein the state estimates are sent from sensors to a fusion center to generate a global state estimate. The underlying fusion algorithm affects the performance measure QCC(τ) (with subscripts CC indicating the effects of the communications and computing quality) of the global state estimate computed within the allocated time τ. We present a probabilistic performance bound on QCC(τ) as a function of the distributions of state estimates, communications parameters as well as the fusion algorithm. We present simulations of simplified scenarios to illustrate the qualitative effects of different fusers, and system-level simulations to complement the analytical results.","PeriodicalId":103145,"journal":{"name":"2012 IEEE International Conference on Multisensor Fusion and Integration for Intelligent Systems (MFI)","volume":"54 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127113896","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2012-11-12DOI: 10.1109/MFI.2012.6343064
Aliabbas Vohra, M. Sarkar, Gordon K. Lee
As a nation of an estimated 45 million uninsured and underinsured Americans (almost 15% of the population), out of which over 11 million suffer from chronic diseases who require constant medical supervision, America today is plagued by the national crisis of inadequate and expensive healthcare. This paper introduces an architecture of a multi-tier telemedicine system comprised of strategically placed bio-sensors on a human body capable of collecting vital medical statistics (such as heart rate and blood pressure) and transmitting them (wired or wirelessly)over multiple hops to a remote medical server at a caregiver's location thereby taking telemedicine from the desktop to roaming. However, fundamental wireless networking issues must be addressed and resolved before this dream can be realized. In this regards, this paper proposes a Medium Access Control (MAC) protocol specifically designed for a Wireless Body Area Network (WBAN). Our protocol is designed to cater to the Quality of Service (QoS) requirements that would be essential for an application like WBAN. It fuses data from several biosensors and based on the time criticality of the data, schedules them intelligently such that the data reaches its destination in a timely and energy efficient manner. Simulation results show that the traffic prioritization and scheduling scheme proposed in our MAC architecture surpasses the standard IEEE 802.15.4 MAC protocol in performance.
{"title":"A smart transmission scheme for emergency data from a network of bio-sensors on the human body","authors":"Aliabbas Vohra, M. Sarkar, Gordon K. Lee","doi":"10.1109/MFI.2012.6343064","DOIUrl":"https://doi.org/10.1109/MFI.2012.6343064","url":null,"abstract":"As a nation of an estimated 45 million uninsured and underinsured Americans (almost 15% of the population), out of which over 11 million suffer from chronic diseases who require constant medical supervision, America today is plagued by the national crisis of inadequate and expensive healthcare. This paper introduces an architecture of a multi-tier telemedicine system comprised of strategically placed bio-sensors on a human body capable of collecting vital medical statistics (such as heart rate and blood pressure) and transmitting them (wired or wirelessly)over multiple hops to a remote medical server at a caregiver's location thereby taking telemedicine from the desktop to roaming. However, fundamental wireless networking issues must be addressed and resolved before this dream can be realized. In this regards, this paper proposes a Medium Access Control (MAC) protocol specifically designed for a Wireless Body Area Network (WBAN). Our protocol is designed to cater to the Quality of Service (QoS) requirements that would be essential for an application like WBAN. It fuses data from several biosensors and based on the time criticality of the data, schedules them intelligently such that the data reaches its destination in a timely and energy efficient manner. Simulation results show that the traffic prioritization and scheduling scheme proposed in our MAC architecture surpasses the standard IEEE 802.15.4 MAC protocol in performance.","PeriodicalId":103145,"journal":{"name":"2012 IEEE International Conference on Multisensor Fusion and Integration for Intelligent Systems (MFI)","volume":"40 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126060697","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2012-11-12DOI: 10.1109/MFI.2012.6343001
Li Yan, Liu Jingtai, Li Haifeng, Lu Xiang, Sun Lei
As the competitive networked robot system has the characteristics of strong interaction and high real-time request, the present control methods which are mostly used for the collaborative networked robot system may not be directly applied to the competitive one. Thus the hierarchical control architecture for the competitive networked robot system is proposed in this paper in order to adapt to its two characteristics above. To deal with the system observation uncertainty caused by noise and time delay, and the system action uncertainty caused by the opponent in the meantime, the control architecture based on Partially Observable Markov Decision Processes (POMDP) has been adopted by the executive layer to select the action of the maximum expected reward, thus fulfilling the intention of the strategy layer effectively. In addition, the introduction of the executive layer has successfully freed the strategic layer from the tasks that should have been completed by its bottom layer alone, thus enabling the strategic layer to focus more on its strategic design. In this paper, the networked robot system named Tele-LightSaber with a high degree of confrontation is designed and implemented, and the experiment results show the validity and efficiency of the proposed method on TLS platform.
{"title":"Modeling and control architecture for the competitive networked robot system based on POMDP","authors":"Li Yan, Liu Jingtai, Li Haifeng, Lu Xiang, Sun Lei","doi":"10.1109/MFI.2012.6343001","DOIUrl":"https://doi.org/10.1109/MFI.2012.6343001","url":null,"abstract":"As the competitive networked robot system has the characteristics of strong interaction and high real-time request, the present control methods which are mostly used for the collaborative networked robot system may not be directly applied to the competitive one. Thus the hierarchical control architecture for the competitive networked robot system is proposed in this paper in order to adapt to its two characteristics above. To deal with the system observation uncertainty caused by noise and time delay, and the system action uncertainty caused by the opponent in the meantime, the control architecture based on Partially Observable Markov Decision Processes (POMDP) has been adopted by the executive layer to select the action of the maximum expected reward, thus fulfilling the intention of the strategy layer effectively. In addition, the introduction of the executive layer has successfully freed the strategic layer from the tasks that should have been completed by its bottom layer alone, thus enabling the strategic layer to focus more on its strategic design. In this paper, the networked robot system named Tele-LightSaber with a high degree of confrontation is designed and implemented, and the experiment results show the validity and efficiency of the proposed method on TLS platform.","PeriodicalId":103145,"journal":{"name":"2012 IEEE International Conference on Multisensor Fusion and Integration for Intelligent Systems (MFI)","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114693670","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2012-11-12DOI: 10.1109/MFI.2012.6343065
T. Henderson, E. Cohen, A. Joshi, E. Grant, M. Draelos, N. Deshpande
We propose that robot perception is enabled by means of a common sensorimotor semantics arising from a set of symmetry theories (expressed as symmetry detectors and parsers) embedded a priori in each robot. These theories inform the production of structural representations of sensorimotor processes, and these representations, in turn, permit perceptual fusion to broaden categories of activity. Although the specific knowledge required by a robot will depend on the particular application domain, there is a need for fundamental mechanisms which allow each individual robot to obtain the requisite knowledge. Current methods are too brittle and do not scale very well, and a new approach to perceptual knowledge representation is necessary. Our approach provides firm semantic grounding in the real world, provides for robust dynamic performance in real-time environments with a range of sensors and allows for communication of acquired knowledge in a broad community of other robots and agents, including humans. Our work focuses on symmetry based multisensor knowledge structuring in terms of: (1) symmetry detection in signals, and (2) symmetry parsing for knowledge structure, including structural bootstrapping and knowledge sharing. Operationally, the hypothesis is that group theoretic representations (G-Reps) inform cognitive activity. Our contributions here are to demonstrate symmetry detection and signal analysis and for 1D and 2D signals in a simple office environment; symmetry parsing based on these tokens is left for future work.
{"title":"Symmetry as a basis for perceptual fusion","authors":"T. Henderson, E. Cohen, A. Joshi, E. Grant, M. Draelos, N. Deshpande","doi":"10.1109/MFI.2012.6343065","DOIUrl":"https://doi.org/10.1109/MFI.2012.6343065","url":null,"abstract":"We propose that robot perception is enabled by means of a common sensorimotor semantics arising from a set of symmetry theories (expressed as symmetry detectors and parsers) embedded a priori in each robot. These theories inform the production of structural representations of sensorimotor processes, and these representations, in turn, permit perceptual fusion to broaden categories of activity. Although the specific knowledge required by a robot will depend on the particular application domain, there is a need for fundamental mechanisms which allow each individual robot to obtain the requisite knowledge. Current methods are too brittle and do not scale very well, and a new approach to perceptual knowledge representation is necessary. Our approach provides firm semantic grounding in the real world, provides for robust dynamic performance in real-time environments with a range of sensors and allows for communication of acquired knowledge in a broad community of other robots and agents, including humans. Our work focuses on symmetry based multisensor knowledge structuring in terms of: (1) symmetry detection in signals, and (2) symmetry parsing for knowledge structure, including structural bootstrapping and knowledge sharing. Operationally, the hypothesis is that group theoretic representations (G-Reps) inform cognitive activity. Our contributions here are to demonstrate symmetry detection and signal analysis and for 1D and 2D signals in a simple office environment; symmetry parsing based on these tokens is left for future work.","PeriodicalId":103145,"journal":{"name":"2012 IEEE International Conference on Multisensor Fusion and Integration for Intelligent Systems (MFI)","volume":"13 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129549043","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2012-11-12DOI: 10.1109/MFI.2012.6343056
Shugen Ma, Shuai Guo, Minghui Wang, Bin Li
This paper presents an approach to solve the data association problem for a hybrid metric map representation. The hybrid metric map representation uses Voronoi diagram to partition the global map space into a series of local subregions, and then a local dense map is built in each subregion. Finally the global feature map and the local maps make up of the hybrid metric map, which can represent all the observed environment. In the proposed map representation, there exists an important property that global feature map and local maps have clear one-to-one correspondence. Benefited from this property, an identifying rule of the data association based on compatibility testing is proposed. The identifying rule can efficiently reject the wrong data association hypothesis in the application of dense environment. Two experiments validated the efficiency of data association approach and also demonstrated the feasibility of the hybrid metric map presentation.
{"title":"Data association for a hybrid metric map representation","authors":"Shugen Ma, Shuai Guo, Minghui Wang, Bin Li","doi":"10.1109/MFI.2012.6343056","DOIUrl":"https://doi.org/10.1109/MFI.2012.6343056","url":null,"abstract":"This paper presents an approach to solve the data association problem for a hybrid metric map representation. The hybrid metric map representation uses Voronoi diagram to partition the global map space into a series of local subregions, and then a local dense map is built in each subregion. Finally the global feature map and the local maps make up of the hybrid metric map, which can represent all the observed environment. In the proposed map representation, there exists an important property that global feature map and local maps have clear one-to-one correspondence. Benefited from this property, an identifying rule of the data association based on compatibility testing is proposed. The identifying rule can efficiently reject the wrong data association hypothesis in the application of dense environment. Two experiments validated the efficiency of data association approach and also demonstrated the feasibility of the hybrid metric map presentation.","PeriodicalId":103145,"journal":{"name":"2012 IEEE International Conference on Multisensor Fusion and Integration for Intelligent Systems (MFI)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129726958","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2012-11-12DOI: 10.1109/MFI.2012.6343028
Jiaxi Shi, T. Lüth
This article presents a new method for the data fusion of the intraoperative Electromyography (EMG) signals from a nerve monitor with the position information from a surgical navigation system. In principle, EMG signals and the position information are merged with each other based on the distance-threshold relationship. The merged information can be used to control the power of active surgical instruments intelligently or to improve the surgeon's orientation in operating field during challenging interventions on temporal bones. One part of the described method for stimulation modeling was also verified in an experiment under simplified laboratory conditions. The results show an approximate agreement between the modeled and the measured “distance-threshold” relationship till 2 mm. The curve fitting of the measured and modeled data indicates that a quadratic relationship fits better in the data than a cubic.
{"title":"Data fusion of Electromyography signals from intraoperative nerve monitoring and position information from surgical navigation","authors":"Jiaxi Shi, T. Lüth","doi":"10.1109/MFI.2012.6343028","DOIUrl":"https://doi.org/10.1109/MFI.2012.6343028","url":null,"abstract":"This article presents a new method for the data fusion of the intraoperative Electromyography (EMG) signals from a nerve monitor with the position information from a surgical navigation system. In principle, EMG signals and the position information are merged with each other based on the distance-threshold relationship. The merged information can be used to control the power of active surgical instruments intelligently or to improve the surgeon's orientation in operating field during challenging interventions on temporal bones. One part of the described method for stimulation modeling was also verified in an experiment under simplified laboratory conditions. The results show an approximate agreement between the modeled and the measured “distance-threshold” relationship till 2 mm. The curve fitting of the measured and modeled data indicates that a quadratic relationship fits better in the data than a cubic.","PeriodicalId":103145,"journal":{"name":"2012 IEEE International Conference on Multisensor Fusion and Integration for Intelligent Systems (MFI)","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130247673","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2012-11-12DOI: 10.1109/MFI.2012.6342996
R. Matthias, A. Bihlmaier, H. Wörn
In this paper we address some of the most important aspects in modular self-reconfigurable mobile robotics. Related work indicates that it is not sufficient to just have a flexible, scalable and robust platform but that it is necessary to preserve these capabilities in higher levels of organization to benefit from them. Hence we analyze the way in which similar platforms and the according software are implemented. Then we describe the way our own platform is implemented within the SYMBRION and REPLICATOR projects. Afterwards we show how we manage to preserve the robustness and flexibility for use by other researchers in higher levels of organization. To conclude we provide some measurements that show the general adequacy of our platform architecture to cope with the challenges posed by multi-modular self-reconfigurable robotics.
{"title":"Robustness, scalability and flexibility: Key-features in modular self-reconfigurable mobile robotics","authors":"R. Matthias, A. Bihlmaier, H. Wörn","doi":"10.1109/MFI.2012.6342996","DOIUrl":"https://doi.org/10.1109/MFI.2012.6342996","url":null,"abstract":"In this paper we address some of the most important aspects in modular self-reconfigurable mobile robotics. Related work indicates that it is not sufficient to just have a flexible, scalable and robust platform but that it is necessary to preserve these capabilities in higher levels of organization to benefit from them. Hence we analyze the way in which similar platforms and the according software are implemented. Then we describe the way our own platform is implemented within the SYMBRION and REPLICATOR projects. Afterwards we show how we manage to preserve the robustness and flexibility for use by other researchers in higher levels of organization. To conclude we provide some measurements that show the general adequacy of our platform architecture to cope with the challenges posed by multi-modular self-reconfigurable robotics.","PeriodicalId":103145,"journal":{"name":"2012 IEEE International Conference on Multisensor Fusion and Integration for Intelligent Systems (MFI)","volume":"126 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128204804","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2012-11-12DOI: 10.1109/MFI.2012.6343067
M. Draelos, N. Deshpande, E. Grant
With proper calibration of its color and depth cameras, the Kinect can capture detailed color point clouds at up to 30 frames per second. This capability positions the Kinect for use in robotics as a low-cost navigation sensor. Thus, techniques for efficiently calibrating the Kinect depth camera and altering its optical system to improve suitability for imaging short-range obstacles are presented. To perform depth calibration, a calibration rig and software were developed to automatically map raw depth values to object depths. The calibration rig consisted of a traditional chessboard calibration target with easily locatable features in depth at its exterior corners that facilitated software extraction of corresponding object depths and raw depth values. To modify the Kinect's optics for improved short-range imaging, Nyko's Zoom adapter was used due to its simplicity and low cost. Although effective at reducing the Kinect's minimum range, these optics introduced pronounced distortion in depth. A method based on capturing depth images of planar objects at various depths produced an empirical depth distortion model for correcting such distortion in software. Together, the modified optics and the empirical depth undistortion procedure demonstrated the ability to improve the Kinect's resolution and decrease its minimum range by approximately 30%.
{"title":"The Kinect up close: Adaptations for short-range imaging","authors":"M. Draelos, N. Deshpande, E. Grant","doi":"10.1109/MFI.2012.6343067","DOIUrl":"https://doi.org/10.1109/MFI.2012.6343067","url":null,"abstract":"With proper calibration of its color and depth cameras, the Kinect can capture detailed color point clouds at up to 30 frames per second. This capability positions the Kinect for use in robotics as a low-cost navigation sensor. Thus, techniques for efficiently calibrating the Kinect depth camera and altering its optical system to improve suitability for imaging short-range obstacles are presented. To perform depth calibration, a calibration rig and software were developed to automatically map raw depth values to object depths. The calibration rig consisted of a traditional chessboard calibration target with easily locatable features in depth at its exterior corners that facilitated software extraction of corresponding object depths and raw depth values. To modify the Kinect's optics for improved short-range imaging, Nyko's Zoom adapter was used due to its simplicity and low cost. Although effective at reducing the Kinect's minimum range, these optics introduced pronounced distortion in depth. A method based on capturing depth images of planar objects at various depths produced an empirical depth distortion model for correcting such distortion in software. Together, the modified optics and the empirical depth undistortion procedure demonstrated the ability to improve the Kinect's resolution and decrease its minimum range by approximately 30%.","PeriodicalId":103145,"journal":{"name":"2012 IEEE International Conference on Multisensor Fusion and Integration for Intelligent Systems (MFI)","volume":"38 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128711784","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}