Salim Chebbo, Philippe Durieux, B. Pesquet-Popescu
{"title":"Objective evaluation of compressed video's temporal flickering","authors":"Salim Chebbo, Philippe Durieux, B. Pesquet-Popescu","doi":"10.1109/IPTA.2010.5586728","DOIUrl":null,"url":null,"abstract":"This paper proposes a new method to evaluate the temporal impairments of compressed video signals. The proposed algorithm is block based; each block in a given frame is first classified as stable or in motion with respect to its co-located blocks in the previous and next frames. For each in-motion block, a motion estimation process is used to find its best match. Then the flickering estimation is conducted over a 3D window involving the current block and its temporally co-located or best matched blocks of the previous and next frames. The temporal quality is finally deduced from the flickering estimation. Simulation results show the efficacy of our algorithm.","PeriodicalId":236574,"journal":{"name":"2010 2nd International Conference on Image Processing Theory, Tools and Applications","volume":"25 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2010-07-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"4","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2010 2nd International Conference on Image Processing Theory, Tools and Applications","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/IPTA.2010.5586728","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 4
Abstract
This paper proposes a new method to evaluate the temporal impairments of compressed video signals. The proposed algorithm is block based; each block in a given frame is first classified as stable or in motion with respect to its co-located blocks in the previous and next frames. For each in-motion block, a motion estimation process is used to find its best match. Then the flickering estimation is conducted over a 3D window involving the current block and its temporally co-located or best matched blocks of the previous and next frames. The temporal quality is finally deduced from the flickering estimation. Simulation results show the efficacy of our algorithm.