The face detection problem is certainly one of the most studied problems in the field of computer vision. It finds indeed application in the human-computer interaction field, automotive, etc. but especially in video surveillance and security systems. In the last years, AdaBoost-based systems showed good performance in both detection rate and computation time allowing its exploitation in realtime face detectors. Although effective, the natural asymmetry, brought by the problem of separating objects from the rest of the world, highlighted the limits of such an algorithm. To overcome this limit the AsymBoost version has been introduced to better distinguish the patterns of the two classes. In this paper, we further optimize the learning strategy by extending the AsymBoost cascade algorithm by introducing a reactive control of the asymmetry at both cascade and classifiers learning stages. The results will point out how the proposed strategy cuts the false negatives by keeping low the false positives.
{"title":"Reactive Learning Strategy for AsymBoost Based Face Detectors","authors":"Ingrid Visentini, C. Micheloni, G. Foresti","doi":"10.1109/ICIAP.2007.106","DOIUrl":"https://doi.org/10.1109/ICIAP.2007.106","url":null,"abstract":"The face detection problem is certainly one of the most studied problems in the field of computer vision. It finds indeed application in the human-computer interaction field, automotive, etc. but especially in video surveillance and security systems. In the last years, AdaBoost-based systems showed good performance in both detection rate and computation time allowing its exploitation in realtime face detectors. Although effective, the natural asymmetry, brought by the problem of separating objects from the rest of the world, highlighted the limits of such an algorithm. To overcome this limit the AsymBoost version has been introduced to better distinguish the patterns of the two classes. In this paper, we further optimize the learning strategy by extending the AsymBoost cascade algorithm by introducing a reactive control of the asymmetry at both cascade and classifiers learning stages. The results will point out how the proposed strategy cuts the false negatives by keeping low the false positives.","PeriodicalId":118466,"journal":{"name":"14th International Conference on Image Analysis and Processing (ICIAP 2007)","volume":"41 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117246012","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Although the way we interact with computers is substantially the same since twenty years based on keyboard, mouse and window metaphor-machine perception could be usefully exploited to enhance the human-computer communication process. In this paper, we present a vision-based user interface where plain, static hand gestures performed nearby the mouse are interpreted as specific input commands. Our tests demonstrate that this new input modality does not interfere with ordinary mouse use and can speed up task execution, while not requiring too much attention from the user.
{"title":"Adding Gestures to Ordinary Mouse Use: a New Input Modality for Improved Human-Computer Interaction","authors":"L. Lombardi, M. Porta","doi":"10.1109/ICIAP.2007.20","DOIUrl":"https://doi.org/10.1109/ICIAP.2007.20","url":null,"abstract":"Although the way we interact with computers is substantially the same since twenty years based on keyboard, mouse and window metaphor-machine perception could be usefully exploited to enhance the human-computer communication process. In this paper, we present a vision-based user interface where plain, static hand gestures performed nearby the mouse are interpreted as specific input commands. Our tests demonstrate that this new input modality does not interfere with ordinary mouse use and can speed up task execution, while not requiring too much attention from the user.","PeriodicalId":118466,"journal":{"name":"14th International Conference on Image Analysis and Processing (ICIAP 2007)","volume":"19 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122423699","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
We address the synchronization of a pair of video sequences captured from moving vehicles and the spatial registration of all the temporally corresponding frames. This is necessary in order to perform the pixel wise comparison of a pair of videos. The novelty of our method is that is free from three common restrictions of most previous works. First, it does not require that the two cameras be rigidly fixed to each other, since they can move independently. Second, the temporal correspondence does not assume a linear mapping. Third, it does not rely on the complete trajectories of image features. We present our results in the context of two applications, outdoor surveillance at night and the comparison of vehicle headlights systems.
{"title":"Alignment of videos recorded from moving vehicles","authors":"J. Serrat, Ferran Diego, F. Lumbreras, J. Álvarez","doi":"10.1109/ICIAP.2007.21","DOIUrl":"https://doi.org/10.1109/ICIAP.2007.21","url":null,"abstract":"We address the synchronization of a pair of video sequences captured from moving vehicles and the spatial registration of all the temporally corresponding frames. This is necessary in order to perform the pixel wise comparison of a pair of videos. The novelty of our method is that is free from three common restrictions of most previous works. First, it does not require that the two cameras be rigidly fixed to each other, since they can move independently. Second, the temporal correspondence does not assume a linear mapping. Third, it does not rely on the complete trajectories of image features. We present our results in the context of two applications, outdoor surveillance at night and the comparison of vehicle headlights systems.","PeriodicalId":118466,"journal":{"name":"14th International Conference on Image Analysis and Processing (ICIAP 2007)","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122792263","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
We describe a new decomposition algorithm for two-dimensional, polygonal shapes. The algorithm first finds a set of overlapping ribbon-like subshapes ("ribbons") by grouping skeleton fragments into long smooth spines. The parts are then obtained by cutting the ribbons along lines of mutual intersection. With this approach we find part cuts that obey heuristic rules of early vision (M. Singh et a., 1999) without using boundary curvature.
我们描述了一种新的二维多边形形状的分解算法。该算法首先通过将骨骼碎片分组成光滑的长刺,找到一组重叠的带状亚形状(“带状”)。然后通过沿相互交点的线切割色带来获得零件。通过这种方法,我们发现部分切割遵循早期视觉的启发式规则(M. Singh et a., 1999),而不使用边界曲率。
{"title":"Parsing Silhouettes without Boundary Curvature","authors":"R. Juengling, Lakshman Prasad","doi":"10.1109/ICIAP.2007.101","DOIUrl":"https://doi.org/10.1109/ICIAP.2007.101","url":null,"abstract":"We describe a new decomposition algorithm for two-dimensional, polygonal shapes. The algorithm first finds a set of overlapping ribbon-like subshapes (\"ribbons\") by grouping skeleton fragments into long smooth spines. The parts are then obtained by cutting the ribbons along lines of mutual intersection. With this approach we find part cuts that obey heuristic rules of early vision (M. Singh et a., 1999) without using boundary curvature.","PeriodicalId":118466,"journal":{"name":"14th International Conference on Image Analysis and Processing (ICIAP 2007)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115411352","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
M. Aguzzi, M. Albanesi, M. Ferretti, M. Garlaschelli
In this paper we present a novel infrastructure for querying heterogeneous databases composed of semantically annotated documents, using XML technology. We have developed a query generalization system in order to make the use of different XML schemas transparent. It uses a preliminary specification of elements considered equivalent in the various schemas. By defining different levels of equivalence, the system is able to translate queries between different annotated XML schemas. We have tested the system on the particular case of image database, more precisely, we consider different tagged image formats. We obtain query translation that preserves the semantics and offers a good match.
{"title":"Query translation based on equivalence classes for heterogeneous multimedia repositories using XML and semantic annotation","authors":"M. Aguzzi, M. Albanesi, M. Ferretti, M. Garlaschelli","doi":"10.1109/ICIAP.2007.105","DOIUrl":"https://doi.org/10.1109/ICIAP.2007.105","url":null,"abstract":"In this paper we present a novel infrastructure for querying heterogeneous databases composed of semantically annotated documents, using XML technology. We have developed a query generalization system in order to make the use of different XML schemas transparent. It uses a preliminary specification of elements considered equivalent in the various schemas. By defining different levels of equivalence, the system is able to translate queries between different annotated XML schemas. We have tested the system on the particular case of image database, more precisely, we consider different tagged image formats. We obtain query translation that preserves the semantics and offers a good match.","PeriodicalId":118466,"journal":{"name":"14th International Conference on Image Analysis and Processing (ICIAP 2007)","volume":"18 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131560268","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Haider Ali, C. Seifert, Nitin Jindal, L. Paletta, G. Paar
This work is about a novel methodology for window detection in urban environments and its multiple use in vision system applications. The presented method for window detection includes appropriate early image processing, provides a multi-scale Haar wavelet representation for the determination of image tiles which is then fed into a cascaded classifier for the task of window detection. The classifier is learned from a Gentle Adaboost driven cascaded decision tree on masked information from training imagery and is tested towards window based ground truth information which is together with the original building image databases publicly available. The experimental results demonstrate that single window detection is to a sufficient degree successful, e.g., for the purpose of building recognition, and, furthermore, that the classifier is in general capable to provide a region of interest operator for the interpretation of urban environments. The extraction of this categorical information is beneficial to index into search spaces for urban object recognition as well as aiming towards providing a semantic focus for accurate post-processing in 3D information processing systems. Targeted applications are (i) mobile services on uncalibrated imagery, e.g. , for tourist guidance, (ii) sparse 3D city modeling, and (iii) deformation analysis from high resolution imagery.
{"title":"Window Detection in Facades","authors":"Haider Ali, C. Seifert, Nitin Jindal, L. Paletta, G. Paar","doi":"10.1109/ICIAP.2007.139","DOIUrl":"https://doi.org/10.1109/ICIAP.2007.139","url":null,"abstract":"This work is about a novel methodology for window detection in urban environments and its multiple use in vision system applications. The presented method for window detection includes appropriate early image processing, provides a multi-scale Haar wavelet representation for the determination of image tiles which is then fed into a cascaded classifier for the task of window detection. The classifier is learned from a Gentle Adaboost driven cascaded decision tree on masked information from training imagery and is tested towards window based ground truth information which is together with the original building image databases publicly available. The experimental results demonstrate that single window detection is to a sufficient degree successful, e.g., for the purpose of building recognition, and, furthermore, that the classifier is in general capable to provide a region of interest operator for the interpretation of urban environments. The extraction of this categorical information is beneficial to index into search spaces for urban object recognition as well as aiming towards providing a semantic focus for accurate post-processing in 3D information processing systems. Targeted applications are (i) mobile services on uncalibrated imagery, e.g. , for tourist guidance, (ii) sparse 3D city modeling, and (iii) deformation analysis from high resolution imagery.","PeriodicalId":118466,"journal":{"name":"14th International Conference on Image Analysis and Processing (ICIAP 2007)","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126912944","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
An algorithm for combining results of different clusterings is presented in this paper, the objective of which is to find groups of patterns which are common to all clusterings. The idea of the proposed combination is to group those samples which are in the same cluster in most cases. We formulate the combination as the resolution of a linear set of equations with binary constraints. The advantage of such a formulation is to provide an objective function for the combination. To optimize the objective function we propose an original unsupervised algorithm. Furthermore, we propose an extension adapted in case of a huge volume of data. The combination of clusterings is performed on the results of different clustering algorithms applied to SPOT5 satellite images and shows the effectiveness of the proposed method.
{"title":"A Method of Clustering Combination Applied to Satellite Image Analysis","authors":"Ivan O. Kyrgyzov, H. Maître, M. Campedel","doi":"10.1109/ICIAP.2007.11","DOIUrl":"https://doi.org/10.1109/ICIAP.2007.11","url":null,"abstract":"An algorithm for combining results of different clusterings is presented in this paper, the objective of which is to find groups of patterns which are common to all clusterings. The idea of the proposed combination is to group those samples which are in the same cluster in most cases. We formulate the combination as the resolution of a linear set of equations with binary constraints. The advantage of such a formulation is to provide an objective function for the combination. To optimize the objective function we propose an original unsupervised algorithm. Furthermore, we propose an extension adapted in case of a huge volume of data. The combination of clusterings is performed on the results of different clustering algorithms applied to SPOT5 satellite images and shows the effectiveness of the proposed method.","PeriodicalId":118466,"journal":{"name":"14th International Conference on Image Analysis and Processing (ICIAP 2007)","volume":"19 20","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"113961908","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
S. Battiato, G. Gallo, G. Puglisi, Salvatore Scellato
This paper presents a video stabilization algorithm based on the extraction and tracking of scale invariant feature transform features through video frames. Implementation of SIFT operator is analyzed and adapted to be used in a feature-based motion estimation algorithm. SIFT features are extracted from video frames and then their trajectory is evaluated to estimate interframe motion. A modified version of iterative least squares method is adopted to avoid estimation errors and features are tracked as they appear in nearby frames to improve video stability. Intentional camera motion is eventually filtered with adaptive motion vector integration. Results confirm the effectiveness of the method.
{"title":"SIFT Features Tracking for Video Stabilization","authors":"S. Battiato, G. Gallo, G. Puglisi, Salvatore Scellato","doi":"10.1109/ICIAP.2007.116","DOIUrl":"https://doi.org/10.1109/ICIAP.2007.116","url":null,"abstract":"This paper presents a video stabilization algorithm based on the extraction and tracking of scale invariant feature transform features through video frames. Implementation of SIFT operator is analyzed and adapted to be used in a feature-based motion estimation algorithm. SIFT features are extracted from video frames and then their trajectory is evaluated to estimate interframe motion. A modified version of iterative least squares method is adopted to avoid estimation errors and features are tracked as they appear in nearby frames to improve video stability. Intentional camera motion is eventually filtered with adaptive motion vector integration. Results confirm the effectiveness of the method.","PeriodicalId":118466,"journal":{"name":"14th International Conference on Image Analysis and Processing (ICIAP 2007)","volume":"41 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125440598","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
We present methods for eliminating or reducing the distortion in a scanned image. Aspects of the present paper allow for the automatic pruning, de-skewing, and unwarping of an image using boundary document layout information. Here, two dominant top/down baselines are selected, in part, by examining the letter spatial locations on boundary baselines rather than examining the entire document layout. It shall be noted that present method is robust enough to handle many types of content, including different languages: Japanese and English, as well as documents with different layouts. The algorithm is applied to images obtained from bound documents and flat documents.
{"title":"Unwarping scanned image of Japanese/English documents","authors":"Ali Zandifar","doi":"10.1109/ICIAP.2007.128","DOIUrl":"https://doi.org/10.1109/ICIAP.2007.128","url":null,"abstract":"We present methods for eliminating or reducing the distortion in a scanned image. Aspects of the present paper allow for the automatic pruning, de-skewing, and unwarping of an image using boundary document layout information. Here, two dominant top/down baselines are selected, in part, by examining the letter spatial locations on boundary baselines rather than examining the entire document layout. It shall be noted that present method is robust enough to handle many types of content, including different languages: Japanese and English, as well as documents with different layouts. The algorithm is applied to images obtained from bound documents and flat documents.","PeriodicalId":118466,"journal":{"name":"14th International Conference on Image Analysis and Processing (ICIAP 2007)","volume":"31 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117082725","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
The paper describes a method for computer-assisted edge tracking and following by local profile matching. It is used primarily to integrate and support manual selection when complex contours have to be found and identified. Such a requirement is quite common in biomedical applications, including a wide spectrum of image modalities, from histological samples to MRI images. The referred example takes part in the analysis of in vitro receptor auto-radiographic data, for the architectonic characterization of transmitter receptors within a defined cortical region. The practical validation we provide of the proposed tracking model reveals that it is a very promising approach, enabling accurate and fast processing of multiple images.
{"title":"Edge Tracking of subjective contours in Biomedical Imaging","authors":"G. Garibotto, V. Garibotto","doi":"10.1109/ICIAP.2007.56","DOIUrl":"https://doi.org/10.1109/ICIAP.2007.56","url":null,"abstract":"The paper describes a method for computer-assisted edge tracking and following by local profile matching. It is used primarily to integrate and support manual selection when complex contours have to be found and identified. Such a requirement is quite common in biomedical applications, including a wide spectrum of image modalities, from histological samples to MRI images. The referred example takes part in the analysis of in vitro receptor auto-radiographic data, for the architectonic characterization of transmitter receptors within a defined cortical region. The practical validation we provide of the proposed tracking model reveals that it is a very promising approach, enabling accurate and fast processing of multiple images.","PeriodicalId":118466,"journal":{"name":"14th International Conference on Image Analysis and Processing (ICIAP 2007)","volume":"21 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121673434","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}