Pub Date : 2016-08-02DOI: 10.1186/s41074-016-0005-0
Saori Uda, Fumihiko Sakaue, J. Sato
{"title":"Variable exposure time imaging for obtaining unblurred HDR images","authors":"Saori Uda, Fumihiko Sakaue, J. Sato","doi":"10.1186/s41074-016-0005-0","DOIUrl":"https://doi.org/10.1186/s41074-016-0005-0","url":null,"abstract":"","PeriodicalId":38957,"journal":{"name":"IPSJ Transactions on Computer Vision and Applications","volume":"8 1","pages":"1-7"},"PeriodicalIF":0.0,"publicationDate":"2016-08-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1186/s41074-016-0005-0","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"65775302","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2016-08-02DOI: 10.1186/s41074-016-0002-3
Yutaro Sako, Y. Sugaya
{"title":"Multibody motion segmentation for an arbitrary number of independent motions","authors":"Yutaro Sako, Y. Sugaya","doi":"10.1186/s41074-016-0002-3","DOIUrl":"https://doi.org/10.1186/s41074-016-0002-3","url":null,"abstract":"","PeriodicalId":38957,"journal":{"name":"IPSJ Transactions on Computer Vision and Applications","volume":"8 1","pages":"1-5"},"PeriodicalIF":0.0,"publicationDate":"2016-08-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1186/s41074-016-0002-3","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"65775221","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2016-08-02DOI: 10.1186/s41074-016-0006-z
Akito Takeki, T. Trinh, Ryota Yoshihashi, Rei Kawakami, M. Iida, T. Naemura
{"title":"Combining deep features for object detection at various scales: finding small birds in landscape images","authors":"Akito Takeki, T. Trinh, Ryota Yoshihashi, Rei Kawakami, M. Iida, T. Naemura","doi":"10.1186/s41074-016-0006-z","DOIUrl":"https://doi.org/10.1186/s41074-016-0006-z","url":null,"abstract":"","PeriodicalId":38957,"journal":{"name":"IPSJ Transactions on Computer Vision and Applications","volume":"8 1","pages":"1-7"},"PeriodicalIF":0.0,"publicationDate":"2016-08-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1186/s41074-016-0006-z","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"65775313","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2016-08-02DOI: 10.1186/s41074-016-0004-1
Kei Terayama, H. Habe, M. Sakagami
{"title":"Multiple fish tracking with an NACA airfoil model for collective behavior analysis","authors":"Kei Terayama, H. Habe, M. Sakagami","doi":"10.1186/s41074-016-0004-1","DOIUrl":"https://doi.org/10.1186/s41074-016-0004-1","url":null,"abstract":"","PeriodicalId":38957,"journal":{"name":"IPSJ Transactions on Computer Vision and Applications","volume":"8 1","pages":"1-7"},"PeriodicalIF":0.0,"publicationDate":"2016-08-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1186/s41074-016-0004-1","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"65775289","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Kosuke Takahashi, Dan Mikami, Mariko Isogawa, Akira Kojima
In this paper, we propose a novel algorithm to extrinsically calibrate a camera to a 3D reference object that is not directly visible from the camera. We use the spherical human cornea as a mirror and calibrate the extrinsic parameters from its reflection of the reference points. The key contribution of this paper is to present a cornea-reflectionbased calibration algorithm with minimal configuration; there are three reference points and one mirror pose. The proposed algorithm introduces two constraints. First constraint is that the cornea is virtually a sphere, which enables us to estimate the center of the cornea sphere from its projection. Second is the equidistance constraint, which enables us to estimate the 3D position of the reference point by assuming that the center of the camera and reference point are located the same distance from the center of the cornea sphere. We demonstrate the advantages of the proposed method with qualitative and quantitative evaluations using synthesized and real data.
{"title":"Extrinsic Camera Calibration with Minimal Configuration Using Cornea Model and Equidistance Constraint","authors":"Kosuke Takahashi, Dan Mikami, Mariko Isogawa, Akira Kojima","doi":"10.2197/ipsjtcva.8.20","DOIUrl":"https://doi.org/10.2197/ipsjtcva.8.20","url":null,"abstract":"In this paper, we propose a novel algorithm to extrinsically calibrate a camera to a 3D reference object that is not directly visible from the camera. We use the spherical human cornea as a mirror and calibrate the extrinsic parameters from its reflection of the reference points. The key contribution of this paper is to present a cornea-reflectionbased calibration algorithm with minimal configuration; there are three reference points and one mirror pose. The proposed algorithm introduces two constraints. First constraint is that the cornea is virtually a sphere, which enables us to estimate the center of the cornea sphere from its projection. Second is the equidistance constraint, which enables us to estimate the 3D position of the reference point by assuming that the center of the camera and reference point are located the same distance from the center of the cornea sphere. We demonstrate the advantages of the proposed method with qualitative and quantitative evaluations using synthesized and real data.","PeriodicalId":38957,"journal":{"name":"IPSJ Transactions on Computer Vision and Applications","volume":"33 1","pages":"20-28"},"PeriodicalIF":0.0,"publicationDate":"2016-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"83718544","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Vision based road recognition and tracking are crucial tasks in a field of autonomous driving. Road recognition methods based on shape analysis of road region have the potential to overcome the limitations of traditional boundary based approaches, but a robust method for road region segmentation is the challenging issue. In our work, we treat the problem of road region segmentation as a classification task, where road pixels are classified by statistical decision rule based on the probability density function (pdf) of road features. This paper presents a new algorithm for the estimation of the pdf, based on sequential Monte-Carlo (SMC) method. The proposed algorithm is evaluated on data sets of three different types of images, and the results of evaluation show the effectiveness of the proposed method.
{"title":"Sequential Monte-Carlo Based Road Region Segmentation Algorithm with Uniform Spatial Sampling","authors":"Z. Procházka","doi":"10.2197/ipsjtcva.8.1","DOIUrl":"https://doi.org/10.2197/ipsjtcva.8.1","url":null,"abstract":"Vision based road recognition and tracking are crucial tasks in a field of autonomous driving. Road recognition methods based on shape analysis of road region have the potential to overcome the limitations of traditional boundary based approaches, but a robust method for road region segmentation is the challenging issue. In our work, we treat the problem of road region segmentation as a classification task, where road pixels are classified by statistical decision rule based on the probability density function (pdf) of road features. This paper presents a new algorithm for the estimation of the pdf, based on sequential Monte-Carlo (SMC) method. The proposed algorithm is evaluated on data sets of three different types of images, and the results of evaluation show the effectiveness of the proposed method.","PeriodicalId":38957,"journal":{"name":"IPSJ Transactions on Computer Vision and Applications","volume":"116 1","pages":"1-10"},"PeriodicalIF":0.0,"publicationDate":"2016-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"79653993","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
This paper presents a novel shape descriptor for topology-based segmentation of 3D video sequences. 3D video is a series of 3D meshes without temporal correspondences which benefit for applications including compression, motion analysis, and kinematic editing. In 3D video, both 3D mesh connectivities and the global surface topology can change frame by frame. This characteristic prevents from making accurate temporal correspondences through the entire 3D mesh series. To overcome this difficulty, we propose a two-step strategy which decomposes the entire sequence into a series of topologically coherent segments using our new shape descriptor, and then estimates temporal correspondences on a per-segment basis. As the result of acquiring temporal correspondences, we could extract rigid parts from the preprocessed 3D video segments to establish partial kinematic structures, and could integrate them into a single unified kinematic model which describes the entire kinematic motion in the 3D video sequence. We demonstrate the robustness and accuracy of the shape descriptor on real data which consist of large non-rigid motion and reconstruction errors.
{"title":"Part-wise Geodesic Histogram Shape Descriptor for Unstructured Mesh Series Segmentation","authors":"T. Mukasa, S. Nobuhara, Tony Tung, T. Matsuyama","doi":"10.2197/ipsjtcva.8.29","DOIUrl":"https://doi.org/10.2197/ipsjtcva.8.29","url":null,"abstract":"This paper presents a novel shape descriptor for topology-based segmentation of 3D video sequences. 3D video is a series of 3D meshes without temporal correspondences which benefit for applications including compression, motion analysis, and kinematic editing. In 3D video, both 3D mesh connectivities and the global surface topology can change frame by frame. This characteristic prevents from making accurate temporal correspondences through the entire 3D mesh series. To overcome this difficulty, we propose a two-step strategy which decomposes the entire sequence into a series of topologically coherent segments using our new shape descriptor, and then estimates temporal correspondences on a per-segment basis. As the result of acquiring temporal correspondences, we could extract rigid parts from the preprocessed 3D video segments to establish partial kinematic structures, and could integrate them into a single unified kinematic model which describes the entire kinematic motion in the 3D video sequence. We demonstrate the robustness and accuracy of the shape descriptor on real data which consist of large non-rigid motion and reconstruction errors.","PeriodicalId":38957,"journal":{"name":"IPSJ Transactions on Computer Vision and Applications","volume":"29 1","pages":"29-39"},"PeriodicalIF":0.0,"publicationDate":"2016-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"77389773","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
This paper is aimed at employing mirrors to estimate relative posture and position of camera, i.e., extrinsic parameters, against a 3D reference object that is not directly visible from the camera. The key contribution of this paper is to propose a novel formulation of extrinsic camera calibration based on orthogonality constraint which should be satisfied by all families of mirror-reflections of a single reference object. This allows us to obtain a larger number of equations which contribute to make the calibration more robust. We demonstrate the advantages of the proposed method in comparison with a state-of-the-art by qualitative and quantitative evaluations using synthesized and real data.
{"title":"Mirror-based Camera Pose Estimation Using an Orthogonality Constraint","authors":"Kosuke Takahashi, S. Nobuhara, T. Matsuyama","doi":"10.2197/ipsjtcva.8.11","DOIUrl":"https://doi.org/10.2197/ipsjtcva.8.11","url":null,"abstract":"This paper is aimed at employing mirrors to estimate relative posture and position of camera, i.e., extrinsic parameters, against a 3D reference object that is not directly visible from the camera. The key contribution of this paper is to propose a novel formulation of extrinsic camera calibration based on orthogonality constraint which should be satisfied by all families of mirror-reflections of a single reference object. This allows us to obtain a larger number of equations which contribute to make the calibration more robust. We demonstrate the advantages of the proposed method in comparison with a state-of-the-art by qualitative and quantitative evaluations using synthesized and real data.","PeriodicalId":38957,"journal":{"name":"IPSJ Transactions on Computer Vision and Applications","volume":"62 1","pages":"11-19"},"PeriodicalIF":0.0,"publicationDate":"2016-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"72804918","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2015-11-01DOI: 10.1186/s41074-016-0011-2
S. Bigdeli, Gregor Budweiser, Matthias Zwicker
{"title":"Temporally coherent disparity maps using CRFs with fast 4D filtering","authors":"S. Bigdeli, Gregor Budweiser, Matthias Zwicker","doi":"10.1186/s41074-016-0011-2","DOIUrl":"https://doi.org/10.1186/s41074-016-0011-2","url":null,"abstract":"","PeriodicalId":38957,"journal":{"name":"IPSJ Transactions on Computer Vision and Applications","volume":"8 1","pages":"1-14"},"PeriodicalIF":0.0,"publicationDate":"2015-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1186/s41074-016-0011-2","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"65774902","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2015-11-01DOI: 10.1186/s41074-017-0017-4
X. Han, Yenwei Chen, Gang Xu
{"title":"Co-occurrence context of the data-driven quantized local ternary patterns for visual recognition","authors":"X. Han, Yenwei Chen, Gang Xu","doi":"10.1186/s41074-017-0017-4","DOIUrl":"https://doi.org/10.1186/s41074-017-0017-4","url":null,"abstract":"","PeriodicalId":38957,"journal":{"name":"IPSJ Transactions on Computer Vision and Applications","volume":"9 1","pages":"1-10"},"PeriodicalIF":0.0,"publicationDate":"2015-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1186/s41074-017-0017-4","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"65774914","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}