Pub Date : 2011-06-16DOI: 10.1109/IVMSPW.2011.5970356
E. Assunção, J. R. Pereira, M. Costa, C. Filho, Rafael Padilla
Subspace methods are frequently used in pattern recognition problems aiming to reduce space dimension by determining its projection vectors. This paper presents subspace methods for feature extraction in an iris image called two-dimensional linear discriminant analysis (2DLDA), diagonal linear discriminant analysis (DiaLDA) and their combination (DiaLDA+2DLDA). The methods were applied in an UBIRIS image database, and the experimental results showed that DiaLDA+2DLDA overcame the 2DLDA method in recognition accuracy. Both methods are powerful in terms of dimension reduction and class discrimination.
{"title":"Representation and classification of iris textures based on diagonal linear discriminant analysis","authors":"E. Assunção, J. R. Pereira, M. Costa, C. Filho, Rafael Padilla","doi":"10.1109/IVMSPW.2011.5970356","DOIUrl":"https://doi.org/10.1109/IVMSPW.2011.5970356","url":null,"abstract":"Subspace methods are frequently used in pattern recognition problems aiming to reduce space dimension by determining its projection vectors. This paper presents subspace methods for feature extraction in an iris image called two-dimensional linear discriminant analysis (2DLDA), diagonal linear discriminant analysis (DiaLDA) and their combination (DiaLDA+2DLDA). The methods were applied in an UBIRIS image database, and the experimental results showed that DiaLDA+2DLDA overcame the 2DLDA method in recognition accuracy. Both methods are powerful in terms of dimension reduction and class discrimination.","PeriodicalId":405588,"journal":{"name":"2011 IEEE 10th IVMSP Workshop: Perception and Visual Signal Analysis","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-06-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124542504","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2011-06-16DOI: 10.1109/IVMSPW.2011.5970364
Anush K. Moorthy, A. Bovik
Although the term ‘visually lossless’ (VL) has been used liberally in the video compression literature, there does not seem to be a systematic evaluation of what it means for a video to be compressed visually lossless-ly. Here, we undertake a psychovisual study to infer the visually lossless threshold for H.264 compression of videos spanning a wide range of contents. Based on results from this study, we then propose a compressibility index which provides a measure of the appropriate bit-rate for VL H.264 compression of a video given texture (i.e., spatial activity) and motion (i.e., temporal activity) information. This compressibility index has been made available online at [1] in order to facilitate practical application of the research presented here and to further research in the area of VL compression.
{"title":"H.264 visually lossless compressibility index: Psychophysics and algorithm design","authors":"Anush K. Moorthy, A. Bovik","doi":"10.1109/IVMSPW.2011.5970364","DOIUrl":"https://doi.org/10.1109/IVMSPW.2011.5970364","url":null,"abstract":"Although the term ‘visually lossless’ (VL) has been used liberally in the video compression literature, there does not seem to be a systematic evaluation of what it means for a video to be compressed visually lossless-ly. Here, we undertake a psychovisual study to infer the visually lossless threshold for H.264 compression of videos spanning a wide range of contents. Based on results from this study, we then propose a compressibility index which provides a measure of the appropriate bit-rate for VL H.264 compression of a video given texture (i.e., spatial activity) and motion (i.e., temporal activity) information. This compressibility index has been made available online at [1] in order to facilitate practical application of the research presented here and to further research in the area of VL compression.","PeriodicalId":405588,"journal":{"name":"2011 IEEE 10th IVMSP Workshop: Perception and Visual Signal Analysis","volume":"110 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-06-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127786994","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2011-06-16DOI: 10.1109/IVMSPW.2011.5970365
Yen-Fu Ou, Yuanyi Xue, Zhan Ma, Yao Wang
In this paper, we investigate the impact of spatial, temporal and amplitude resolution (STAR) on the perceptual quality of a compressed video. Subjective quality tests were carried out on the TI Zoom2 mobile development platform (MDP). Seven source sequences are included in the tests and for each source sequence we have 32 test configurations generated by JSVM encoder (4 QP levels, 5 spatial resolutions, and 3 temporal resolutions) and a total of 224 processed video sequences (PVSs). Videos coded at different spatial resolutions are displayed at the full screen size of the mobile platform. We report the impact of the spatial resolution (SR), temporal resolution (TR) and quantization stepsize (QS) on the perceptual quality, individually as well as jointly. We found that the impact of SR, TR and QS can each be captured by a function with a single content-dependent parameter. The joint impact of SR, TR and QS can be modeled by the product of these three functions. The complete model correlates well with the subjective ratings with a Pearson Correlation Coefficient (PCC) of 0.99.
{"title":"A perceptual video quality model for mobile platform considering impact of spatial, temporal, and amplitude resolutions","authors":"Yen-Fu Ou, Yuanyi Xue, Zhan Ma, Yao Wang","doi":"10.1109/IVMSPW.2011.5970365","DOIUrl":"https://doi.org/10.1109/IVMSPW.2011.5970365","url":null,"abstract":"In this paper, we investigate the impact of spatial, temporal and amplitude resolution (STAR) on the perceptual quality of a compressed video. Subjective quality tests were carried out on the TI Zoom2 mobile development platform (MDP). Seven source sequences are included in the tests and for each source sequence we have 32 test configurations generated by JSVM encoder (4 QP levels, 5 spatial resolutions, and 3 temporal resolutions) and a total of 224 processed video sequences (PVSs). Videos coded at different spatial resolutions are displayed at the full screen size of the mobile platform. We report the impact of the spatial resolution (SR), temporal resolution (TR) and quantization stepsize (QS) on the perceptual quality, individually as well as jointly. We found that the impact of SR, TR and QS can each be captured by a function with a single content-dependent parameter. The joint impact of SR, TR and QS can be modeled by the product of these three functions. The complete model correlates well with the subjective ratings with a Pearson Correlation Coefficient (PCC) of 0.99.","PeriodicalId":405588,"journal":{"name":"2011 IEEE 10th IVMSP Workshop: Perception and Visual Signal Analysis","volume":"5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-06-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124355369","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2011-06-16DOI: 10.1109/IVMSPW.2011.5970349
Ming-Jun Chen, A. Bovik, L. Cormack
We describe a study aimed towards increasing our understanding of the perception of distorted stereoscopic 3D images, by analyzing subjects' performance in locating local distortions in stereoscopically viewed images. Nineteen subjects were recruited for this study. The results indicated that contrast and range variations are correlated with the conspicuity of some distortions, but not others.
{"title":"Study on distortion conspicuity in stereoscopically viewed 3D images","authors":"Ming-Jun Chen, A. Bovik, L. Cormack","doi":"10.1109/IVMSPW.2011.5970349","DOIUrl":"https://doi.org/10.1109/IVMSPW.2011.5970349","url":null,"abstract":"We describe a study aimed towards increasing our understanding of the perception of distorted stereoscopic 3D images, by analyzing subjects' performance in locating local distortions in stereoscopically viewed images. Nineteen subjects were recruited for this study. The results indicated that contrast and range variations are correlated with the conspicuity of some distortions, but not others.","PeriodicalId":405588,"journal":{"name":"2011 IEEE 10th IVMSP Workshop: Perception and Visual Signal Analysis","volume":"36 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-06-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122571868","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2011-06-16DOI: 10.1109/IVMSPW.2011.5970357
Y. Belkhouche, B. Buckles
Object boundary is an important feature for image processing and computer vision applications. In this paper a new method for extracting the non convex boundaries of an object represented by 2D point clouds is established. In order to determine the object boundaries we started by constructing the convex-hull-based Delaunay triangulation using the point clouds. Given the fact that the points are sampled from the object surface using an instrument such as cameras or laser scanners, the distribution of the edges lengths belonging to the objects follows a Gaussian distribution. However this distribution is skewed due to the existence of long edges introduced by the Delaunay triangulation. Removing the skewness will make the convex boundary built by the Delauny algorithm converge to the real boundary of the object. We tested our method using different datasets that includes synthetic data, urban LiDAR (Light Detection and Ranging) data, and binary images. The results show that the proposed method successfully extracts the object boundary.
{"title":"Skewness balancing algorithm for approximation of discrete objects boundaries","authors":"Y. Belkhouche, B. Buckles","doi":"10.1109/IVMSPW.2011.5970357","DOIUrl":"https://doi.org/10.1109/IVMSPW.2011.5970357","url":null,"abstract":"Object boundary is an important feature for image processing and computer vision applications. In this paper a new method for extracting the non convex boundaries of an object represented by 2D point clouds is established. In order to determine the object boundaries we started by constructing the convex-hull-based Delaunay triangulation using the point clouds. Given the fact that the points are sampled from the object surface using an instrument such as cameras or laser scanners, the distribution of the edges lengths belonging to the objects follows a Gaussian distribution. However this distribution is skewed due to the existence of long edges introduced by the Delaunay triangulation. Removing the skewness will make the convex boundary built by the Delauny algorithm converge to the real boundary of the object. We tested our method using different datasets that includes synthetic data, urban LiDAR (Light Detection and Ranging) data, and binary images. The results show that the proposed method successfully extracts the object boundary.","PeriodicalId":405588,"journal":{"name":"2011 IEEE 10th IVMSP Workshop: Perception and Visual Signal Analysis","volume":"358 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-06-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133238872","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2011-06-16DOI: 10.1109/IVMSPW.2011.5970358
Ludovica Lorusso, G. Brelstaff, E. Grosso
We analyse three different visual judgements of human faces: kinship, similarity and dissimilarity. The accepted model that represents the relation between similarity and kinship is one in which notional kinship signals are detected by the human observer in order to judge both kinship and similarity of pairs of faces. We measure observers' response times to face-pair stimuli while they perform judgements of kinship, similarity or dissimilarity. Significant differences in response time are found between the three different tasks and between face-pair categories - suggesting that the strategies employed when extracting information from facial features, may be conditional on both task and stimulus context. We sketch a new model for face processing related to these three judgements.
{"title":"A new model for the visual judgement of kinship","authors":"Ludovica Lorusso, G. Brelstaff, E. Grosso","doi":"10.1109/IVMSPW.2011.5970358","DOIUrl":"https://doi.org/10.1109/IVMSPW.2011.5970358","url":null,"abstract":"We analyse three different visual judgements of human faces: kinship, similarity and dissimilarity. The accepted model that represents the relation between similarity and kinship is one in which notional kinship signals are detected by the human observer in order to judge both kinship and similarity of pairs of faces. We measure observers' response times to face-pair stimuli while they perform judgements of kinship, similarity or dissimilarity. Significant differences in response time are found between the three different tasks and between face-pair categories - suggesting that the strategies employed when extracting information from facial features, may be conditional on both task and stimulus context. We sketch a new model for face processing related to these three judgements.","PeriodicalId":405588,"journal":{"name":"2011 IEEE 10th IVMSP Workshop: Perception and Visual Signal Analysis","volume":"11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-06-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124602329","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 1900-01-01DOI: 10.1109/ivmspw.2011.5970360
R. M. Rodrigues, C. F. F. Costa Filho, M. G. F. Costa
This paper describes a new characteristic vector model for fingerprint representation that uses planar graph and triangulation algorithms. It is shown that this new characteristic vector model presents a better performance in a fingerprint identification system when compared with other vector models already proposed in literature. The minutiae extraction is an essential step of a fingerprint recognition system. In this paper is also presented a new method for minutiae extraction that explores the duality ridge ending/ridge bifurcation that exists when the skeleton image of a fingerprint is inverted. It is shown that this new extraction method simplifies the computational complexity of a fingerprint identification system.
{"title":"Fingerprint verification using characteristic vector based on planar graphics","authors":"R. M. Rodrigues, C. F. F. Costa Filho, M. G. F. Costa","doi":"10.1109/ivmspw.2011.5970360","DOIUrl":"https://doi.org/10.1109/ivmspw.2011.5970360","url":null,"abstract":"This paper describes a new characteristic vector model for fingerprint representation that uses planar graph and triangulation algorithms. It is shown that this new characteristic vector model presents a better performance in a fingerprint identification system when compared with other vector models already proposed in literature. The minutiae extraction is an essential step of a fingerprint recognition system. In this paper is also presented a new method for minutiae extraction that explores the duality ridge ending/ridge bifurcation that exists when the skeleton image of a fingerprint is inverted. It is shown that this new extraction method simplifies the computational complexity of a fingerprint identification system.","PeriodicalId":405588,"journal":{"name":"2011 IEEE 10th IVMSP Workshop: Perception and Visual Signal Analysis","volume":"468 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123382112","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}