Pub Date : 1997-04-01DOI: 10.1109/ICIP.1996.560736
Dennis F. Dunn, T. Weldon, W. Higgins
Separating halftones from text is an important step in document analysis. We present an algorithm that accurately extracts halftones from other information in printed documents. We treat halftone extraction as a texture-segmentation problem. We show that commonly used halftones, consisting of a pattern of dots, can be viewed as a texture. This texture exhibits a distinct spectral component that can be detected using a properly-tuned Gabor filter. The Gabor (1946) filter essentially transforms halftones into high-contrast regions that can be isolated by thresholding. We propose a filter-design procedure and provide experimental results.
{"title":"Extracting halftones from printed documents using texture analysis","authors":"Dennis F. Dunn, T. Weldon, W. Higgins","doi":"10.1109/ICIP.1996.560736","DOIUrl":"https://doi.org/10.1109/ICIP.1996.560736","url":null,"abstract":"Separating halftones from text is an important step in document analysis. We present an algorithm that accurately extracts halftones from other information in printed documents. We treat halftone extraction as a texture-segmentation problem. We show that commonly used halftones, consisting of a pattern of dots, can be viewed as a texture. This texture exhibits a distinct spectral component that can be detected using a properly-tuned Gabor filter. The Gabor (1946) filter essentially transforms halftones into high-contrast regions that can be isolated by thresholding. We propose a filter-design procedure and provide experimental results.","PeriodicalId":192947,"journal":{"name":"Proceedings of 3rd IEEE International Conference on Image Processing","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1997-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128023273","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 1996-11-01DOI: 10.1109/ICIP.1996.560802
Chiou-Ting Hsu, Ja-Ling Wu
Mosaic techniques have been used to combine two or more signals into a new one with an invisible seam, and with as little distortion of each signal as possible. Multiresolution representation is an effective method for analyzing the information content of signals, and it also fits a wide spectrum of visual signal processing and visual communication applications. The wavelet transform is one kind of multiresolution representations, and has found a wide variety of application in many aspects, including signal analysis, image coding, image processing, computer vision and etc. Due to its characteristic of multiresolution signal decomposition, the wavelet transform is used for the image mosaic by choosing the width of the mosaic transition zone proportional to the frequency represented in the band. Both 1-D and 2-D signal mosaics are described, and some factors which affect the mosaics are discussed.
{"title":"Multiresolution mosaic","authors":"Chiou-Ting Hsu, Ja-Ling Wu","doi":"10.1109/ICIP.1996.560802","DOIUrl":"https://doi.org/10.1109/ICIP.1996.560802","url":null,"abstract":"Mosaic techniques have been used to combine two or more signals into a new one with an invisible seam, and with as little distortion of each signal as possible. Multiresolution representation is an effective method for analyzing the information content of signals, and it also fits a wide spectrum of visual signal processing and visual communication applications. The wavelet transform is one kind of multiresolution representations, and has found a wide variety of application in many aspects, including signal analysis, image coding, image processing, computer vision and etc. Due to its characteristic of multiresolution signal decomposition, the wavelet transform is used for the image mosaic by choosing the width of the mosaic transition zone proportional to the frequency represented in the band. Both 1-D and 2-D signal mosaics are described, and some factors which affect the mosaics are discussed.","PeriodicalId":192947,"journal":{"name":"Proceedings of 3rd IEEE International Conference on Image Processing","volume":"34 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1996-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129811335","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 1996-09-16DOI: 10.1109/ICIP.1996.559555
J. Allebach, Qian Lin
We describe an algorithm to design a frequency modulated screen using the direct binary search algorithm. Compared with the direct binary search algorithm itself, we show that we can maintain halftone image quality while significantly reducing the required computation.
{"title":"FM screen design using DBS algorithm","authors":"J. Allebach, Qian Lin","doi":"10.1109/ICIP.1996.559555","DOIUrl":"https://doi.org/10.1109/ICIP.1996.559555","url":null,"abstract":"We describe an algorithm to design a frequency modulated screen using the direct binary search algorithm. Compared with the direct binary search algorithm itself, we show that we can maintain halftone image quality while significantly reducing the required computation.","PeriodicalId":192947,"journal":{"name":"Proceedings of 3rd IEEE International Conference on Image Processing","volume":"13 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1996-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115433206","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 1996-09-16DOI: 10.1109/ICIP.1996.560986
Vinay P. Kumar, E. Manolakos
Model-based joint segmentation and recognition of objects is proposed in the framework of parameter estimation of hierarchical mixture densities. The maximum a posteriori (MAP) estimate of the parameters is computed by the application of a modified version of the expectation-maximization algorithm (EM with regularizing constraints applied to multiple level hierarchies). The approach is flexible in the sense that it allows for non-stationary pixel statistics, different noise models and is translation and scale invariant. Simulation results suggest that the scheme is well suited for recognition of partially occluded objects and recognition in complex and poorly modeled background.
{"title":"Unsupervised model-based object recognition by parameter estimation of hierarchical mixtures","authors":"Vinay P. Kumar, E. Manolakos","doi":"10.1109/ICIP.1996.560986","DOIUrl":"https://doi.org/10.1109/ICIP.1996.560986","url":null,"abstract":"Model-based joint segmentation and recognition of objects is proposed in the framework of parameter estimation of hierarchical mixture densities. The maximum a posteriori (MAP) estimate of the parameters is computed by the application of a modified version of the expectation-maximization algorithm (EM with regularizing constraints applied to multiple level hierarchies). The approach is flexible in the sense that it allows for non-stationary pixel statistics, different noise models and is translation and scale invariant. Simulation results suggest that the scheme is well suited for recognition of partially occluded objects and recognition in complex and poorly modeled background.","PeriodicalId":192947,"journal":{"name":"Proceedings of 3rd IEEE International Conference on Image Processing","volume":"113 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1996-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123043519","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 1996-09-16DOI: 10.1109/ICIP.1996.559561
Lina Karam
This paper extends the transformation method by developing two new efficient procedures for the design of complex and real, positive- and negative-symmetric, multi-dimensional filters by transforming even-length 1-D prototype filters with complex (or real) coefficients. The first procedure is used to design complex multi-dimensional filters with a rectangular region of support having odd-length sides, while the second procedure is used for filters with a rectangular region of support having even-length sides. The designed filters can be implemented efficiently using special Chebyshev structures. Design examples are presented to illustrate the performance of the design procedures.
{"title":"Design of complex multi-dimensional FIR filters by transformation","authors":"Lina Karam","doi":"10.1109/ICIP.1996.559561","DOIUrl":"https://doi.org/10.1109/ICIP.1996.559561","url":null,"abstract":"This paper extends the transformation method by developing two new efficient procedures for the design of complex and real, positive- and negative-symmetric, multi-dimensional filters by transforming even-length 1-D prototype filters with complex (or real) coefficients. The first procedure is used to design complex multi-dimensional filters with a rectangular region of support having odd-length sides, while the second procedure is used for filters with a rectangular region of support having even-length sides. The designed filters can be implemented efficiently using special Chebyshev structures. Design examples are presented to illustrate the performance of the design procedures.","PeriodicalId":192947,"journal":{"name":"Proceedings of 3rd IEEE International Conference on Image Processing","volume":"43 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1996-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114650341","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 1996-09-16DOI: 10.1109/ICIP.1996.560539
Rafal Foltyniewicz, Maclef Sitnik
This article proposes a new approach for verification of people. The model consists of two parts: face and signature analysis. For face information processing morphological filtering is used to enhance the intrinsic features of a face, reduce the influence of rotation in depth, changes in facial expression, hair style, glasses and lighting conditions. The filtered images are then a subject for learning by a modified high order neural network. In signature analysis the model first traces the signature to extract the dynamical information that is usually lost in an off-line mode. After this step a neural network (neocognitron with switching attention) is used to recognize and finally verify the signature. These two parts can work independently and finally their outputs can be used to form a complex person verifier.
{"title":"Verification of persons via face and signature analysis","authors":"Rafal Foltyniewicz, Maclef Sitnik","doi":"10.1109/ICIP.1996.560539","DOIUrl":"https://doi.org/10.1109/ICIP.1996.560539","url":null,"abstract":"This article proposes a new approach for verification of people. The model consists of two parts: face and signature analysis. For face information processing morphological filtering is used to enhance the intrinsic features of a face, reduce the influence of rotation in depth, changes in facial expression, hair style, glasses and lighting conditions. The filtered images are then a subject for learning by a modified high order neural network. In signature analysis the model first traces the signature to extract the dynamical information that is usually lost in an off-line mode. After this step a neural network (neocognitron with switching attention) is used to recognize and finally verify the signature. These two parts can work independently and finally their outputs can be used to form a complex person verifier.","PeriodicalId":192947,"journal":{"name":"Proceedings of 3rd IEEE International Conference on Image Processing","volume":"90 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1996-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121819719","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 1996-09-16DOI: 10.1109/ICIP.1996.561015
Kyeong-Man Kim, Chae-Soo Lee, Eung-Joo Lee, Yeong-Ho Ha
Color quantization is to design a color palette having almost no noticeably perceived difference between original and quantized images. In this paper, to design such a palette, the color quantization algorithm is considered in two parts. The selection of a proper distortion measure and the design of an optimal palette. The proper selection of distortion measure in the quantization is important to image quality. Since the human eye is the final judge of image quality, it is desirable to use a perception-based distortion measure. Thus we developed an activity-weighted distortion measure considering color visual sensitivity and absorbance of human visual system (HVS) depending on each color component in the local region of image. Then using the distortion measure, a hierarchical quantization algorithm is proposed. The algorithm consists of initial and subdivision steps both to reduce computation time and to minimize the distortion based on spatial masking effect of HVS. The experimental results show that the proposed algorithm shows better visual quality and less computation time comparing to the conventional algorithms.
{"title":"Color image quantization using weighted distortion measure of HVS color activity","authors":"Kyeong-Man Kim, Chae-Soo Lee, Eung-Joo Lee, Yeong-Ho Ha","doi":"10.1109/ICIP.1996.561015","DOIUrl":"https://doi.org/10.1109/ICIP.1996.561015","url":null,"abstract":"Color quantization is to design a color palette having almost no noticeably perceived difference between original and quantized images. In this paper, to design such a palette, the color quantization algorithm is considered in two parts. The selection of a proper distortion measure and the design of an optimal palette. The proper selection of distortion measure in the quantization is important to image quality. Since the human eye is the final judge of image quality, it is desirable to use a perception-based distortion measure. Thus we developed an activity-weighted distortion measure considering color visual sensitivity and absorbance of human visual system (HVS) depending on each color component in the local region of image. Then using the distortion measure, a hierarchical quantization algorithm is proposed. The algorithm consists of initial and subdivision steps both to reduce computation time and to minimize the distortion based on spatial masking effect of HVS. The experimental results show that the proposed algorithm shows better visual quality and less computation time comparing to the conventional algorithms.","PeriodicalId":192947,"journal":{"name":"Proceedings of 3rd IEEE International Conference on Image Processing","volume":"43 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1996-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117135135","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 1996-09-16DOI: 10.1109/ICIP.1996.561030
P. Bocheck, Shih-Fu Chang
We present our work on content based video (CBV) traffic modeling of variable bit rate (VBR) sources. The CBV approach differs from previous works in that it is not based only on matching of various statistics of the original source, but rather on modeling and mapping its visual content into the corresponding bit rate. We show that the CBV model is fully compatible with current and future compression algorithms including those of very low bit rate video coding. We introduce the separation principle between the visual content and encoder dependent bit rate mapping. We construct and verify two experimental CBV models for basic camera operations. The results obtained show that the CBV model can closely match the various statistics of the MPEG-2 VBR stream.
{"title":"A content based video traffic model using camera operations","authors":"P. Bocheck, Shih-Fu Chang","doi":"10.1109/ICIP.1996.561030","DOIUrl":"https://doi.org/10.1109/ICIP.1996.561030","url":null,"abstract":"We present our work on content based video (CBV) traffic modeling of variable bit rate (VBR) sources. The CBV approach differs from previous works in that it is not based only on matching of various statistics of the original source, but rather on modeling and mapping its visual content into the corresponding bit rate. We show that the CBV model is fully compatible with current and future compression algorithms including those of very low bit rate video coding. We introduce the separation principle between the visual content and encoder dependent bit rate mapping. We construct and verify two experimental CBV models for basic camera operations. The results obtained show that the CBV model can closely match the various statistics of the MPEG-2 VBR stream.","PeriodicalId":192947,"journal":{"name":"Proceedings of 3rd IEEE International Conference on Image Processing","volume":"74 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1996-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117204677","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 1996-09-16DOI: 10.1109/ICIP.1996.560769
A. Hess, H. Scheich
Using the Karhunen-Loeve-transformation on sequential images of functional brain activity obtained by optical recording of intrinsic signals we were able to separate different components within the dynamic oximetric response, which are: (1) separate functional representations and vessel artifacts, (2) one component for more consistent local activation could be identified, (3) a second region of more spread activity could be differentiated.
{"title":"Karhunen-Loeve-transformation of functional optical imaging of brain activity-consequences and implications","authors":"A. Hess, H. Scheich","doi":"10.1109/ICIP.1996.560769","DOIUrl":"https://doi.org/10.1109/ICIP.1996.560769","url":null,"abstract":"Using the Karhunen-Loeve-transformation on sequential images of functional brain activity obtained by optical recording of intrinsic signals we were able to separate different components within the dynamic oximetric response, which are: (1) separate functional representations and vessel artifacts, (2) one component for more consistent local activation could be identified, (3) a second region of more spread activity could be differentiated.","PeriodicalId":192947,"journal":{"name":"Proceedings of 3rd IEEE International Conference on Image Processing","volume":"93 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1996-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117286603","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 1996-09-16DOI: 10.1109/ICIP.1996.560741
F. Marqués, P. Salembier, M. Pardàs, Ramon Morros, Isabelle Corset, S. Jeannin, B. Marcotegui, F. Meyer
We present a coding scheme that achieves, for each image in the sequence, the best segmentation in terms of rate-distortion theory. It is obtained from a set of initial regions and a set of available coding techniques. The segmentation combines spatial and motion criteria. It selects at each area of the image the most adequate criterion for defining a partition in order to obtain the best compromise between cost and quality. In addition, the proposed scheme is very suitable for addressing content-based functionalities.
{"title":"A segmentation-based coding system allowing manipulation of objects (SESAME)","authors":"F. Marqués, P. Salembier, M. Pardàs, Ramon Morros, Isabelle Corset, S. Jeannin, B. Marcotegui, F. Meyer","doi":"10.1109/ICIP.1996.560741","DOIUrl":"https://doi.org/10.1109/ICIP.1996.560741","url":null,"abstract":"We present a coding scheme that achieves, for each image in the sequence, the best segmentation in terms of rate-distortion theory. It is obtained from a set of initial regions and a set of available coding techniques. The segmentation combines spatial and motion criteria. It selects at each area of the image the most adequate criterion for defining a partition in order to obtain the best compromise between cost and quality. In addition, the proposed scheme is very suitable for addressing content-based functionalities.","PeriodicalId":192947,"journal":{"name":"Proceedings of 3rd IEEE International Conference on Image Processing","volume":"29 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1996-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121116650","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}