Pub Date : 2007-11-12DOI: 10.1109/ICIP.2007.4379784
Iulian Pruteanu-Malinici, L. Carin
We address the problem of unusual-event detection in a video sequence. Invariant subspace analysis (ISA) is used to extract features from the video, and the time-evolving properties of these features are modeled via an infinite hidden Markov model (iHMM), which is trained using "normal"/"typical" video data. The iHMM automatically determines the proper number of HMM states, and it retains a full posterior density function on all model parameters. Anomalies (unusual events) are detected subsequently if a low likelihood is observed when associated sequential features are submitted to the trained iHMM. A hierarchical Dirichlet process (HDP) framework is employed in the formulation of the iHMM. The evaluation of posterior distributions for the iHMM is achieved in two ways: via MCMC and using a variational Bayes (VB) formulation.
{"title":"Infinite Hidden Markov Models and ISA Features for Unusual-Event Detection in Video","authors":"Iulian Pruteanu-Malinici, L. Carin","doi":"10.1109/ICIP.2007.4379784","DOIUrl":"https://doi.org/10.1109/ICIP.2007.4379784","url":null,"abstract":"We address the problem of unusual-event detection in a video sequence. Invariant subspace analysis (ISA) is used to extract features from the video, and the time-evolving properties of these features are modeled via an infinite hidden Markov model (iHMM), which is trained using \"normal\"/\"typical\" video data. The iHMM automatically determines the proper number of HMM states, and it retains a full posterior density function on all model parameters. Anomalies (unusual events) are detected subsequently if a low likelihood is observed when associated sequential features are submitted to the trained iHMM. A hierarchical Dirichlet process (HDP) framework is employed in the formulation of the iHMM. The evaluation of posterior distributions for the iHMM is achieved in two ways: via MCMC and using a variational Bayes (VB) formulation.","PeriodicalId":131177,"journal":{"name":"2007 IEEE International Conference on Image Processing","volume":"18 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129446379","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2007-11-12DOI: 10.1109/ICIP.2007.4379271
V. Velisavljevic, B. Beferull-Lozano, M. Vetterli
In our previous work we proposed a construction of critically sampled perfect reconstruction transforms with directional vanishing moments (DVMs) imposed in the corresponding basis functions along different directions, called directionlets. Here, we combine the directionlets with the space-frequency quantization (SFQ) image compression method, originally based on the standard two-dimensional (2-D) wavelet transform (WT). We show that our new compression method outperforms the standard SFQ as well as the state-of-the-art compression methods, like SPIHT and JPEG-2000, in terms of the quality of compressed images, especially in a low-rate compression regime. We also show that the order of computational complexity remains the same, as compared to the complexity of the standard SFQ algorithm.
{"title":"Space-Frequency Quantization using Directionlets","authors":"V. Velisavljevic, B. Beferull-Lozano, M. Vetterli","doi":"10.1109/ICIP.2007.4379271","DOIUrl":"https://doi.org/10.1109/ICIP.2007.4379271","url":null,"abstract":"In our previous work we proposed a construction of critically sampled perfect reconstruction transforms with directional vanishing moments (DVMs) imposed in the corresponding basis functions along different directions, called directionlets. Here, we combine the directionlets with the space-frequency quantization (SFQ) image compression method, originally based on the standard two-dimensional (2-D) wavelet transform (WT). We show that our new compression method outperforms the standard SFQ as well as the state-of-the-art compression methods, like SPIHT and JPEG-2000, in terms of the quality of compressed images, especially in a low-rate compression regime. We also show that the order of computational complexity remains the same, as compared to the complexity of the standard SFQ algorithm.","PeriodicalId":131177,"journal":{"name":"2007 IEEE International Conference on Image Processing","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129888642","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2007-11-12DOI: 10.1109/ICIP.2007.4379293
D. Coltuc
This paper continues our researches on high capacity reversible watermarking based on simple transforms. Image pixels obeying some simple constraints are transformed and then, data is embedded by simple additions. The transform induces a congruence equation. At detection, the transformed/not transformed pixels are found by checking the congruence, watermark codes are extracted and original image is recovered. The proposed method is of very low mathematical complexity and does not need additional data compression. The major novelty of the proposed technique is the use of a different transform which allows the embedding of watermark codewords into each transformed pixel. Thus, the improved version provides considerably higher embedding bit-rate. Experimental results for a single embedding level are presented.
{"title":"Improved Capacity Reversible Watermarking","authors":"D. Coltuc","doi":"10.1109/ICIP.2007.4379293","DOIUrl":"https://doi.org/10.1109/ICIP.2007.4379293","url":null,"abstract":"This paper continues our researches on high capacity reversible watermarking based on simple transforms. Image pixels obeying some simple constraints are transformed and then, data is embedded by simple additions. The transform induces a congruence equation. At detection, the transformed/not transformed pixels are found by checking the congruence, watermark codes are extracted and original image is recovered. The proposed method is of very low mathematical complexity and does not need additional data compression. The major novelty of the proposed technique is the use of a different transform which allows the embedding of watermark codewords into each transformed pixel. Thus, the improved version provides considerably higher embedding bit-rate. Experimental results for a single embedding level are presented.","PeriodicalId":131177,"journal":{"name":"2007 IEEE International Conference on Image Processing","volume":"56 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126694187","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2007-11-12DOI: 10.1109/ICIP.2007.4379542
M. Gouiffès
This paper addresses the problem of robust feature points tracking by using specific color invariants -robust to specular reflections, lighting changes and to some extent to color lighting changes-when they are relevant and photometric normalization in the opposite case. Indeed, most color invariants become noisy or irrelevant for low saturation and/or low intensity. They can even make tracking fail. Combining them with luminance information yields to a more performant tracking, whatever the lighting conditions are. A few experiments on real image sequences prove the efficiency of this procedure.
{"title":"Tracking by Combining Photometric Normalization and Color Invariants According to their Relevance","authors":"M. Gouiffès","doi":"10.1109/ICIP.2007.4379542","DOIUrl":"https://doi.org/10.1109/ICIP.2007.4379542","url":null,"abstract":"This paper addresses the problem of robust feature points tracking by using specific color invariants -robust to specular reflections, lighting changes and to some extent to color lighting changes-when they are relevant and photometric normalization in the opposite case. Indeed, most color invariants become noisy or irrelevant for low saturation and/or low intensity. They can even make tracking fail. Combining them with luminance information yields to a more performant tracking, whatever the lighting conditions are. A few experiments on real image sequences prove the efficiency of this procedure.","PeriodicalId":131177,"journal":{"name":"2007 IEEE International Conference on Image Processing","volume":"65 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127028446","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2007-11-12DOI: 10.1109/ICIP.2007.4380068
Xiaoyu Zhang, Jian Cheng, Hanqing Lu, Songde Ma
In relevance feedback, active learning is often used to alleviate the burden of labeling by selecting only the most informative data. Traditional data selection strategies often choose the data closest to the current classification boundary to label, which are in fact not informative enough. In this paper, we propose the moving virtual boundary (MVB) strategy, which is proved to be a more effective way for data selection. The co-SVM algorithm is another powerful method used in relevance feedback. Unfortunately, its basic assumption that each view of the data be sufficient is often untenable in image retrieval. We present our weighted co-SVM as an extension of co-SVM by attaching weight to each view, and thus relax the view sufficiency assumption. The experimental results show that the weighted co-SVM algorithm outperforms co-SVM obviously, especially with the help of MVB data selection strategy.
{"title":"Weighted Co-SVM for Image Retrieval with MVB Strategy","authors":"Xiaoyu Zhang, Jian Cheng, Hanqing Lu, Songde Ma","doi":"10.1109/ICIP.2007.4380068","DOIUrl":"https://doi.org/10.1109/ICIP.2007.4380068","url":null,"abstract":"In relevance feedback, active learning is often used to alleviate the burden of labeling by selecting only the most informative data. Traditional data selection strategies often choose the data closest to the current classification boundary to label, which are in fact not informative enough. In this paper, we propose the moving virtual boundary (MVB) strategy, which is proved to be a more effective way for data selection. The co-SVM algorithm is another powerful method used in relevance feedback. Unfortunately, its basic assumption that each view of the data be sufficient is often untenable in image retrieval. We present our weighted co-SVM as an extension of co-SVM by attaching weight to each view, and thus relax the view sufficiency assumption. The experimental results show that the weighted co-SVM algorithm outperforms co-SVM obviously, especially with the help of MVB data selection strategy.","PeriodicalId":131177,"journal":{"name":"2007 IEEE International Conference on Image Processing","volume":"215 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130619007","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2007-11-12DOI: 10.1109/ICIP.2007.4379959
G. Fernández-Escribano, J. Bialkowski, H. Kalva, P. Cuenca, L. Orozco-Barbosa, André Kaup
In this paper, we propose the use of data mining algorithms to create a macroblock partition mode decision algorithm for inter-frame prediction, to be used as part of a high-efficient H.263 to H.264 transcoder. We use machine learning tools to exploit the correlation and derive decision trees to classify the incoming H.263 MC residual into one of the several coding modes in H.264. The proposed approach reduces the H.264 MB mode computation process into a decision tree lookup with very low complexity. Experimental results show that the proposed approach reduces the inter-prediction complexity by as much as 60% while maintaining the coding efficiency.
{"title":"H.263 to H.264 Transconding using Data Mining","authors":"G. Fernández-Escribano, J. Bialkowski, H. Kalva, P. Cuenca, L. Orozco-Barbosa, André Kaup","doi":"10.1109/ICIP.2007.4379959","DOIUrl":"https://doi.org/10.1109/ICIP.2007.4379959","url":null,"abstract":"In this paper, we propose the use of data mining algorithms to create a macroblock partition mode decision algorithm for inter-frame prediction, to be used as part of a high-efficient H.263 to H.264 transcoder. We use machine learning tools to exploit the correlation and derive decision trees to classify the incoming H.263 MC residual into one of the several coding modes in H.264. The proposed approach reduces the H.264 MB mode computation process into a decision tree lookup with very low complexity. Experimental results show that the proposed approach reduces the inter-prediction complexity by as much as 60% while maintaining the coding efficiency.","PeriodicalId":131177,"journal":{"name":"2007 IEEE International Conference on Image Processing","volume":"6 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123896336","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2007-11-12DOI: 10.1109/ICIP.2007.4379994
Basak Oztan, Gaurav Sharma, R. Loce
Misregistration between the color separations of a printed image, which is often inevitable, can cause objectionable color shifts in average color. We analyze the impact of inter-separation misregistration on clustered-dot halftones using Fourier analysis in a lattice framework. Our analysis provides a complete characterization of the conditions under which the average color is invariant to displacement misregistration. In addition to known conditions on colorant spectra and periodicity of the halftones, the work reveals that invariance can also be obtained when these conditions are violated for suitable dot shapes and displacements. Examples for these conditions are included, as is the consideration of traditional halftone configurations.
{"title":"Conditions for Color Misregistration Sensitivity in Clustered-dot Halftones","authors":"Basak Oztan, Gaurav Sharma, R. Loce","doi":"10.1109/ICIP.2007.4379994","DOIUrl":"https://doi.org/10.1109/ICIP.2007.4379994","url":null,"abstract":"Misregistration between the color separations of a printed image, which is often inevitable, can cause objectionable color shifts in average color. We analyze the impact of inter-separation misregistration on clustered-dot halftones using Fourier analysis in a lattice framework. Our analysis provides a complete characterization of the conditions under which the average color is invariant to displacement misregistration. In addition to known conditions on colorant spectra and periodicity of the halftones, the work reveals that invariance can also be obtained when these conditions are violated for suitable dot shapes and displacements. Examples for these conditions are included, as is the consideration of traditional halftone configurations.","PeriodicalId":131177,"journal":{"name":"2007 IEEE International Conference on Image Processing","volume":"37 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124171129","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2007-11-12DOI: 10.1109/ICIP.2007.4380026
Hang Zhou, D. Suter
We apply Gaussian process classification (GPC) to man-made structure segmentation, treated as a two class problem. GPC is a discriminative approach, and thus focuses on modelling the posterior directly. It relaxes the strong assumption of conditional independence of the observed data (generally used in a generative model). In addition, wavelet transform features, which are effective in describing directional textures, are incorporated in the feature vector. Satisfactory results have been obtained which show the effectiveness of our approach.
{"title":"Man-Made Structure Segmentation using Gaussian Processes and Wavelet Features","authors":"Hang Zhou, D. Suter","doi":"10.1109/ICIP.2007.4380026","DOIUrl":"https://doi.org/10.1109/ICIP.2007.4380026","url":null,"abstract":"We apply Gaussian process classification (GPC) to man-made structure segmentation, treated as a two class problem. GPC is a discriminative approach, and thus focuses on modelling the posterior directly. It relaxes the strong assumption of conditional independence of the observed data (generally used in a generative model). In addition, wavelet transform features, which are effective in describing directional textures, are incorporated in the feature vector. Satisfactory results have been obtained which show the effectiveness of our approach.","PeriodicalId":131177,"journal":{"name":"2007 IEEE International Conference on Image Processing","volume":"48 30 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124218635","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2007-11-12DOI: 10.1109/ICIP.2007.4379995
Naixiang Lian, Yap-Peng Tan
To reduce the cost and size, most digital still cameras (DSCs) capture only one color value at each pixel, and the results - color filter array samples - are then interpolated by a demosaicking method to construct a full-color image. Many advanced demosaicking methods have been proposed recently. However, the high complexity of these methods could prevent them from being used in DSCs. In this paper we propose an efficient and effective demosaicking method, which substitutes high-frequency component of color values in the spatial rather than frequency domain. We also propose a simple ternary, anisotropic interpolation scheme to obtain an initial full-color image required in the spatial-domain high-frequency substitution. Experimental results show that the proposed method can outperform recent state-of-the-art methods in terms of both PSNR performance and perceptual results, at the same time reducing the computational cost substantially.
{"title":"An Efficient and Effective Color Filter Array Demosaicking Method","authors":"Naixiang Lian, Yap-Peng Tan","doi":"10.1109/ICIP.2007.4379995","DOIUrl":"https://doi.org/10.1109/ICIP.2007.4379995","url":null,"abstract":"To reduce the cost and size, most digital still cameras (DSCs) capture only one color value at each pixel, and the results - color filter array samples - are then interpolated by a demosaicking method to construct a full-color image. Many advanced demosaicking methods have been proposed recently. However, the high complexity of these methods could prevent them from being used in DSCs. In this paper we propose an efficient and effective demosaicking method, which substitutes high-frequency component of color values in the spatial rather than frequency domain. We also propose a simple ternary, anisotropic interpolation scheme to obtain an initial full-color image required in the spatial-domain high-frequency substitution. Experimental results show that the proposed method can outperform recent state-of-the-art methods in terms of both PSNR performance and perceptual results, at the same time reducing the computational cost substantially.","PeriodicalId":131177,"journal":{"name":"2007 IEEE International Conference on Image Processing","volume":"96 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116320779","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2007-11-12DOI: 10.1109/ICIP.2007.4380050
Qiming Li, N. Memon
Designing secure digital watermarking schemes resistant to invertibility attacks (or more generally, ambiguity attacks) has been challenging. In a recent work, Li and Chang (IHW'04) give the first stand-alone provably secure non-invertible spread-spectrum watermarking scheme based on cryptographically secure pseudo-random generators. Despite its provable security, there are certain constraints on the security parameters that require further analysis in practice, where it is more important to analyze the exact security instead of theoretical asymptotic bounds. In this paper, we consider a security notion that is slightly weaker theoretically but still reasonable in practice, and show that with this alternative security notion, the exact requirements on the parameters can be analyzed, and such analysis can be used to guide flexible implementations of similar schemes in practice.
{"title":"Practical Security of Non-Invertiblewatermarking Schemes","authors":"Qiming Li, N. Memon","doi":"10.1109/ICIP.2007.4380050","DOIUrl":"https://doi.org/10.1109/ICIP.2007.4380050","url":null,"abstract":"Designing secure digital watermarking schemes resistant to invertibility attacks (or more generally, ambiguity attacks) has been challenging. In a recent work, Li and Chang (IHW'04) give the first stand-alone provably secure non-invertible spread-spectrum watermarking scheme based on cryptographically secure pseudo-random generators. Despite its provable security, there are certain constraints on the security parameters that require further analysis in practice, where it is more important to analyze the exact security instead of theoretical asymptotic bounds. In this paper, we consider a security notion that is slightly weaker theoretically but still reasonable in practice, and show that with this alternative security notion, the exact requirements on the parameters can be analyzed, and such analysis can be used to guide flexible implementations of similar schemes in practice.","PeriodicalId":131177,"journal":{"name":"2007 IEEE International Conference on Image Processing","volume":"95 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116488193","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}