Pub Date : 2021-04-10DOI: 10.1142/S0219467822500061
M. Salah, Ameni Yengui, M. Neji
In this paper, we present two steps in the process of automatic annotation in archeological images. These steps are feature extraction and feature selection. We focus our research on archeological images which are very much studied in our days. It presents the most important steps in the process of automatic annotation in an image. Feature extraction techniques are applied to get the feature that will be used in classifying and recognizing the images. Also, the selection of characteristics reduces the number of unattractive characteristics. However, we reviewed various images of feature extraction techniques to analyze the archaeological images. Each feature represents one or more feature descriptors in the archeological images. We focus on the descriptor shape of the archaeological objects extraction in the images using contour method-based shape recognition of the monuments. So, the feature selection stage serves to acquire the most interesting characteristics to improve the accuracy of the classification. In the feature selection section, we present a comparative study between feature selection techniques. Then we give our proposal of application of methods of selection of the characteristics of the archaeological images. Finally, we calculate the performance of two steps already mentioned: the extraction of characteristics and the selection of characteristics.
{"title":"Feature Extraction and Selection in Archaeological Images for Automatic Annotation","authors":"M. Salah, Ameni Yengui, M. Neji","doi":"10.1142/S0219467822500061","DOIUrl":"https://doi.org/10.1142/S0219467822500061","url":null,"abstract":"In this paper, we present two steps in the process of automatic annotation in archeological images. These steps are feature extraction and feature selection. We focus our research on archeological images which are very much studied in our days. It presents the most important steps in the process of automatic annotation in an image. Feature extraction techniques are applied to get the feature that will be used in classifying and recognizing the images. Also, the selection of characteristics reduces the number of unattractive characteristics. However, we reviewed various images of feature extraction techniques to analyze the archaeological images. Each feature represents one or more feature descriptors in the archeological images. We focus on the descriptor shape of the archaeological objects extraction in the images using contour method-based shape recognition of the monuments. So, the feature selection stage serves to acquire the most interesting characteristics to improve the accuracy of the classification. In the feature selection section, we present a comparative study between feature selection techniques. Then we give our proposal of application of methods of selection of the characteristics of the archaeological images. Finally, we calculate the performance of two steps already mentioned: the extraction of characteristics and the selection of characteristics.","PeriodicalId":177479,"journal":{"name":"Int. J. Image Graph.","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-04-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122291388","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2021-04-10DOI: 10.1142/S0219467822500036
M. Suresha, D. Raghukumar, Subramanya Kuppa
Among all image enhancement techniques, histogram equalization is the most used technique. However, preserving brightness is the main issue, and it creates a weird look by destroying its originality. This paper proposes a new method that has command on the brightness issue of histogram equalization to enhance the quality of microscopic images. The method splits the histogram of each color channel into two sub-histograms based on their mean as the threshold and supplanting their cumulative distribution with Kumaraswamy distribution. The proposed method is tested with color microscopic images of cancer-affected lymph nodes gathered from Biological Image Repository IICBU, and objective and subjective assessments confirm that the proposed approach performs more efficiently compared to other state-of-the-art methods.
{"title":"Kumaraswamy Distribution Based Bi-histogram Equalization for Enhancement of Microscopic Images","authors":"M. Suresha, D. Raghukumar, Subramanya Kuppa","doi":"10.1142/S0219467822500036","DOIUrl":"https://doi.org/10.1142/S0219467822500036","url":null,"abstract":"Among all image enhancement techniques, histogram equalization is the most used technique. However, preserving brightness is the main issue, and it creates a weird look by destroying its originality. This paper proposes a new method that has command on the brightness issue of histogram equalization to enhance the quality of microscopic images. The method splits the histogram of each color channel into two sub-histograms based on their mean as the threshold and supplanting their cumulative distribution with Kumaraswamy distribution. The proposed method is tested with color microscopic images of cancer-affected lymph nodes gathered from Biological Image Repository IICBU, and objective and subjective assessments confirm that the proposed approach performs more efficiently compared to other state-of-the-art methods.","PeriodicalId":177479,"journal":{"name":"Int. J. Image Graph.","volume":"110 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-04-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127983647","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2021-04-10DOI: 10.1142/S0219467822500115
Anuj Bhardwaj, Vivek Singh Verma, Sandesh Gupta
Image watermarking is one of the most accepted solutions protecting image authenticity. The method presented in this paper not only provides the desired outcome also efficient in terms of memory requirements and preserving image characteristics. This scheme effectively utilizes the concepts of block truncation coding (BTC) and lifting wavelet transform (LWT). The BTC method is applied to observe the binary watermark image corresponding to its gray-scale image. Whereas, the LWT is incorporated to transform the cover image from spatial coordinates to corresponding transform coordinates. In this, a quantization-based approach for watermark bit embedding is applied. And, the extraction of binary watermark data from the attacked watermarked image is based on adaptive thresholding. To show the effectiveness of the proposed scheme, the experiment over different benchmark images is performed. The experimental results and the comparison with state-of-the-art schemes depict not only the good imperceptibility but also high robustness against various attacks.
{"title":"Image Authentication Using Block Truncation Coding in Lifting Wavelet Domain","authors":"Anuj Bhardwaj, Vivek Singh Verma, Sandesh Gupta","doi":"10.1142/S0219467822500115","DOIUrl":"https://doi.org/10.1142/S0219467822500115","url":null,"abstract":"Image watermarking is one of the most accepted solutions protecting image authenticity. The method presented in this paper not only provides the desired outcome also efficient in terms of memory requirements and preserving image characteristics. This scheme effectively utilizes the concepts of block truncation coding (BTC) and lifting wavelet transform (LWT). The BTC method is applied to observe the binary watermark image corresponding to its gray-scale image. Whereas, the LWT is incorporated to transform the cover image from spatial coordinates to corresponding transform coordinates. In this, a quantization-based approach for watermark bit embedding is applied. And, the extraction of binary watermark data from the attacked watermarked image is based on adaptive thresholding. To show the effectiveness of the proposed scheme, the experiment over different benchmark images is performed. The experimental results and the comparison with state-of-the-art schemes depict not only the good imperceptibility but also high robustness against various attacks.","PeriodicalId":177479,"journal":{"name":"Int. J. Image Graph.","volume":"42 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-04-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126951988","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2021-04-10DOI: 10.1142/S021946782250005X
S. Jameel, Jafar Majidpour
Recently, numerous challenging problems have existed for transforming different image types (thermal infrared (TIR), visible spectrum, and near-infrared (NIR)). Other types of cameras may lack the ability and features of certain types of frequently-used cameras that produce different types of images. Based on camera features, different applications might emerge from observing a scenario under specific conditions (darkness, fog, night, day, and artificial light). We need to jump from one field to another to understand the scenario better. This paper proposes a fully automatic model (GVTI-AE) to manipulate the transformation into different types of vibrant, realistic images using the AutoEncoder method, which requires neither pre-nor post-processing or any user input. The experiments carried out using the GVTI-AE model showed that the perceptually realistic results produced in the widely available datasets (Tecnocampus Hand Image Database, Carl dataset, and IRIS Thermal/Visible Face Database).
近年来,不同类型的图像(热红外(TIR)、可见光谱和近红外(NIR))的转换存在许多具有挑战性的问题。其他类型的相机可能缺乏某些类型的常用相机的能力和功能,产生不同类型的图像。根据相机的功能,在特定条件下(黑暗、雾、夜晚、白天和人造光)观察场景可能会出现不同的应用程序。我们需要从一个领域跳到另一个领域,以便更好地理解这个场景。本文提出了一种全自动模型(GVTI-AE),利用AutoEncoder方法将图像转换成不同类型的充满活力的逼真图像,该模型既不需要预处理,也不需要后处理,也不需要用户输入。使用GVTI-AE模型进行的实验表明,在广泛可用的数据集(Tecnocampus Hand Image Database, Carl dataset和IRIS Thermal/Visible Face Database)中产生的感知逼真的结果。
{"title":"Generating Spectrum Images from Different Types - Visible, Thermal, and Infrared Based on Autoencoder Architecture (GVTI-AE)","authors":"S. Jameel, Jafar Majidpour","doi":"10.1142/S021946782250005X","DOIUrl":"https://doi.org/10.1142/S021946782250005X","url":null,"abstract":"Recently, numerous challenging problems have existed for transforming different image types (thermal infrared (TIR), visible spectrum, and near-infrared (NIR)). Other types of cameras may lack the ability and features of certain types of frequently-used cameras that produce different types of images. Based on camera features, different applications might emerge from observing a scenario under specific conditions (darkness, fog, night, day, and artificial light). We need to jump from one field to another to understand the scenario better. This paper proposes a fully automatic model (GVTI-AE) to manipulate the transformation into different types of vibrant, realistic images using the AutoEncoder method, which requires neither pre-nor post-processing or any user input. The experiments carried out using the GVTI-AE model showed that the perceptually realistic results produced in the widely available datasets (Tecnocampus Hand Image Database, Carl dataset, and IRIS Thermal/Visible Face Database).","PeriodicalId":177479,"journal":{"name":"Int. J. Image Graph.","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-04-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134471301","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2021-04-10DOI: 10.1142/S0219467822500012
N. Shrivastava, J. Bharti
Breast cancer is dangerous in women. It is generally found after the symptoms appear. Detecting the breast cancer at an early stage and understanding the treatment are the most important strategies to prevent death from cancer. Generally, for detection of breast cancer, breast Magnetic Resonance Image (MRI) takes place. It is one of the best approaches to detect tumor in women. In this research paper, a combination of selection methods for seed region growing image segmentation is suggested to detect breast tumor. The suggested method has been divided into following parts: First, the pre-processing of breast image is performed. Second, the automatic threshold for binarization process is calculated. Third, the number of seed points and its position in the breast image are determined automatically using density of pixels value. Fourth, a method for calculation of threshold value is proposed for the purpose of region creation in seed region growing. For the evaluation purpose, the proposed method was applied and tested on the RIDER MRI breast dataset from National Biomedical Imaging Archive (NBIA). After the test was performed, it was observed that proposed algorithm gives 90% accuracy, 88% True Negative Fraction, 91% True Positive Fraction, 10% Misclassification Rate, 94% Precision and 86% Relative Overlap which is better than other existing methods. It not only gives better evaluation measure but also provides segmentation method for multiple tumor detection.
{"title":"Breast Tumor Detection in MR Images Based on Density","authors":"N. Shrivastava, J. Bharti","doi":"10.1142/S0219467822500012","DOIUrl":"https://doi.org/10.1142/S0219467822500012","url":null,"abstract":"Breast cancer is dangerous in women. It is generally found after the symptoms appear. Detecting the breast cancer at an early stage and understanding the treatment are the most important strategies to prevent death from cancer. Generally, for detection of breast cancer, breast Magnetic Resonance Image (MRI) takes place. It is one of the best approaches to detect tumor in women. In this research paper, a combination of selection methods for seed region growing image segmentation is suggested to detect breast tumor. The suggested method has been divided into following parts: First, the pre-processing of breast image is performed. Second, the automatic threshold for binarization process is calculated. Third, the number of seed points and its position in the breast image are determined automatically using density of pixels value. Fourth, a method for calculation of threshold value is proposed for the purpose of region creation in seed region growing. For the evaluation purpose, the proposed method was applied and tested on the RIDER MRI breast dataset from National Biomedical Imaging Archive (NBIA). After the test was performed, it was observed that proposed algorithm gives 90% accuracy, 88% True Negative Fraction, 91% True Positive Fraction, 10% Misclassification Rate, 94% Precision and 86% Relative Overlap which is better than other existing methods. It not only gives better evaluation measure but also provides segmentation method for multiple tumor detection.","PeriodicalId":177479,"journal":{"name":"Int. J. Image Graph.","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-04-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128163851","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2021-04-05DOI: 10.1142/S0219467821500510
S. E. Kuzhali, D. Suresh
For handling digital images for various applications, image denoising is considered as a fundamental pre-processing step. Diverse image denoising algorithms have been introduced in the past few decades. The main intent of this proposal is to develop an effective image denoising model on the basis of internal and external patches. This model adopts Non-local means (NLM) for performing the denoising, which uses redundant information of the image in pixel or spatial domain to reduce the noise. While performing the image denoising using NLM, “denoising an image patch using the other noisy patches within the noisy image is done for internal denoising and denoising a patch using the external clean natural patches is done for external denoising”. Here, the selection of optimal block from the entire datasets including internal noisy images and external clean natural images is decided by a new hybrid optimization algorithm. The two renowned optimization algorithms Chicken Swarm Optimization (CSO), and Dragon Fly Algorithm (DA) are merged, and the new hybrid algorithm Rooster-based Levy Updated DA (RLU-DA) is adopted. The experimental results in terms of some relevant performance measures show the promising results of the proposed model with remarkable stability and high accuracy.
{"title":"Automated Image Denoising Model: Contribution Towards Optimized Internal and External Basis","authors":"S. E. Kuzhali, D. Suresh","doi":"10.1142/S0219467821500510","DOIUrl":"https://doi.org/10.1142/S0219467821500510","url":null,"abstract":"For handling digital images for various applications, image denoising is considered as a fundamental pre-processing step. Diverse image denoising algorithms have been introduced in the past few decades. The main intent of this proposal is to develop an effective image denoising model on the basis of internal and external patches. This model adopts Non-local means (NLM) for performing the denoising, which uses redundant information of the image in pixel or spatial domain to reduce the noise. While performing the image denoising using NLM, “denoising an image patch using the other noisy patches within the noisy image is done for internal denoising and denoising a patch using the external clean natural patches is done for external denoising”. Here, the selection of optimal block from the entire datasets including internal noisy images and external clean natural images is decided by a new hybrid optimization algorithm. The two renowned optimization algorithms Chicken Swarm Optimization (CSO), and Dragon Fly Algorithm (DA) are merged, and the new hybrid algorithm Rooster-based Levy Updated DA (RLU-DA) is adopted. The experimental results in terms of some relevant performance measures show the promising results of the proposed model with remarkable stability and high accuracy.","PeriodicalId":177479,"journal":{"name":"Int. J. Image Graph.","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-04-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115005684","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2021-03-18DOI: 10.1142/S0219467821500509
O. Ramwala, Smeet A. Dhakecha, C. Paunwala, M. Paunwala
Documents are an essential source of valuable information and knowledge, and photographs are a great way of reminiscing old memories and past events. However, it becomes difficult to preserve the quality of such ancient documents and old photographs for an extremely long time, as these images usually get damaged or creased due to various extrinsic effects. Utilizing image editing software like Photoshop to manually reconstruct such old photographs and documents is a strenuous and an enduring process. This paper attempts to leverage the generative modeling capabilities of Conditional Generative Adversarial Networks by utilizing specialized architectures for the Generator and the Discriminator. The proposed Reminiscent Net has a U-Net-based Generator with numerous feature maps for complete information transfer with the incorporation of location and contextual details, and the absence of dense layers allows utilization of diverse sized images. Implementation of the PatchGAN-based Discriminator that penalizes the image at the scale of patches has been proposed. NADAM optimizer has been implemented to enable faster and better convergence of the loss function. The proposed method produces visually appealing de-creased images, and experiments indicate that the architecture performs better than various novel approaches, both qualitatively and quantitatively.
{"title":"Reminiscent Net: Conditional GAN-based Old Image De-Creasing","authors":"O. Ramwala, Smeet A. Dhakecha, C. Paunwala, M. Paunwala","doi":"10.1142/S0219467821500509","DOIUrl":"https://doi.org/10.1142/S0219467821500509","url":null,"abstract":"Documents are an essential source of valuable information and knowledge, and photographs are a great way of reminiscing old memories and past events. However, it becomes difficult to preserve the quality of such ancient documents and old photographs for an extremely long time, as these images usually get damaged or creased due to various extrinsic effects. Utilizing image editing software like Photoshop to manually reconstruct such old photographs and documents is a strenuous and an enduring process. This paper attempts to leverage the generative modeling capabilities of Conditional Generative Adversarial Networks by utilizing specialized architectures for the Generator and the Discriminator. The proposed Reminiscent Net has a U-Net-based Generator with numerous feature maps for complete information transfer with the incorporation of location and contextual details, and the absence of dense layers allows utilization of diverse sized images. Implementation of the PatchGAN-based Discriminator that penalizes the image at the scale of patches has been proposed. NADAM optimizer has been implemented to enable faster and better convergence of the loss function. The proposed method produces visually appealing de-creased images, and experiments indicate that the architecture performs better than various novel approaches, both qualitatively and quantitatively.","PeriodicalId":177479,"journal":{"name":"Int. J. Image Graph.","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-03-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126445827","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2021-03-17DOI: 10.1142/S0219467821500522
E. Ehsaeyan, A. Zolghadrasli
Image segmentation is a prime operation to understand the content of images. Multilevel thresholding is applied in image segmentation because of its speed and accuracy. In this paper, a novel multilevel thresholding algorithm based on Cuckoo search (CS) is introduced. One of the major drawbacks of metaheuristic algorithms is the stagnation phenomenon which leads to a fall into local optimums and premature convergence. To overcome this shortcoming, the idea of Darwinian theory is incorporated with CS algorithm to increase the diversity and quality of the individuals without decreasing the convergence speed of CS algorithm. A policy of encouragement and punishment is considered to lead searching agents in the search space and reduce the computational time. The algorithm is implemented based on dividing the population into specified groups and each group tries to find a better location. Ten test images are selected to verify the ability of our algorithm using the famous energy curve method. Two popular entropies criteria, Otsu and Kapur, are employed to evaluate the capability of the introduced algorithm. Eight different search algorithms are also implemented and compared with our method. Experimental results manifest that DCS is a powerful tool for multilevel thresholding and the obtained results outperform the CS algorithm and other heuristic search methods.
{"title":"A Multilevel Image Thresholding Method Using the Darwinian Cuckoo Search Algorithm","authors":"E. Ehsaeyan, A. Zolghadrasli","doi":"10.1142/S0219467821500522","DOIUrl":"https://doi.org/10.1142/S0219467821500522","url":null,"abstract":"Image segmentation is a prime operation to understand the content of images. Multilevel thresholding is applied in image segmentation because of its speed and accuracy. In this paper, a novel multilevel thresholding algorithm based on Cuckoo search (CS) is introduced. One of the major drawbacks of metaheuristic algorithms is the stagnation phenomenon which leads to a fall into local optimums and premature convergence. To overcome this shortcoming, the idea of Darwinian theory is incorporated with CS algorithm to increase the diversity and quality of the individuals without decreasing the convergence speed of CS algorithm. A policy of encouragement and punishment is considered to lead searching agents in the search space and reduce the computational time. The algorithm is implemented based on dividing the population into specified groups and each group tries to find a better location. Ten test images are selected to verify the ability of our algorithm using the famous energy curve method. Two popular entropies criteria, Otsu and Kapur, are employed to evaluate the capability of the introduced algorithm. Eight different search algorithms are also implemented and compared with our method. Experimental results manifest that DCS is a powerful tool for multilevel thresholding and the obtained results outperform the CS algorithm and other heuristic search methods.","PeriodicalId":177479,"journal":{"name":"Int. J. Image Graph.","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-03-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131271595","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2021-03-15DOI: 10.1142/S0219467821500480
A. Swamy, N. Shylashree
HDR images are inherently very large in size compared to normal images. Hence, storage and communication overheads of HDR images are expensive to be used in mobile devices. Hence, invariably image compression is adopted for HDR images. In this paper, HDR image compression is achieved by down sampling the intensity levels while maintaining the dynamic range same as that of the original. This aspect retains the edge information of the images almost intact. Spatial down-sampling process is used to reduce the number of intensity samples. Consequently, this operation lowers the bit depth required to store the corresponding index file which in turn results in image compression.
{"title":"HDR Image Compression by Multi-Scale down Sampling of Intensity Levels","authors":"A. Swamy, N. Shylashree","doi":"10.1142/S0219467821500480","DOIUrl":"https://doi.org/10.1142/S0219467821500480","url":null,"abstract":"HDR images are inherently very large in size compared to normal images. Hence, storage and communication overheads of HDR images are expensive to be used in mobile devices. Hence, invariably image compression is adopted for HDR images. In this paper, HDR image compression is achieved by down sampling the intensity levels while maintaining the dynamic range same as that of the original. This aspect retains the edge information of the images almost intact. Spatial down-sampling process is used to reduce the number of intensity samples. Consequently, this operation lowers the bit depth required to store the corresponding index file which in turn results in image compression.","PeriodicalId":177479,"journal":{"name":"Int. J. Image Graph.","volume":"29 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-03-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121581848","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Computer vision-based recognition of Bangle vehicle license plates (LPs) is an arduous task in dirty and muddy situations. This paper proposes an efficient method for real-time computer vision-based recognition of Bangla vehicle LPs using contour analysis and prediction algorithms. The method initially applies gray scaling the input RGB images, histogram equalization to improve the grayscale image quality, edge detection using Sobel edge detector, and adaptive thresholding to convert it to a binary image. The system localizes the vehicle LP based on the maximum rectangular contour area and converts it into a predefined size. Noise removal technique using morphological dilation and erosion operation is used, followed by Gaussian filtering on binary image to improve the image quality further. The system clusters the two-lined LP into seven clusters. The sub-clustering is applied on specific clusters and makes 68 individual sub-clusters. The system extracts vector contour (VC) from each 68 individual classes. After VC extraction, the system normalizes it into a q predefined length. The system applies inter co-relation function (ICF) to categorize each sub-cluster to its previously defined individual class. For that, it calculates the maximum similarity between test and previously trained VCs. The system applies the dependency prediction algorithm in parallel to predict the whole string (district name) in the cluster-1 based on previously categorized class or classes (starting character or characters of the district part). [Formula: see text] (Metro) or (null) from cluster-2, “-” (hyphen) from cluster-3 and 6 are predicted automatically as their positions are fixed. The system is trained using 68 classes in which each class contains 100 samples generated by the augmentation technique. The system is tested using another set of 68 classes with a total of [Formula: see text] images acquiring the recognition accuracy of 96.62% with the mean computational cost of 8.363[Formula: see text]ms/f. The system is also tested using 500 vehicle whole Bangla LPs achieving the mean whole LP recognition accuracy of 95.41% with a mean computational cost of 35.803[Formula: see text]ms/f.
{"title":"Real-Time Computer Vision-Based Bangla Vehicle License Plate Recognition using Contour Analysis and Prediction Algorithm","authors":"Masud Pervej, Sabuj Das, Md. Parvez Hossain, Md. Atikuzzaman, Md. Mahin, Muhammad Aminur Rahaman","doi":"10.1142/S021946782150042X","DOIUrl":"https://doi.org/10.1142/S021946782150042X","url":null,"abstract":"Computer vision-based recognition of Bangle vehicle license plates (LPs) is an arduous task in dirty and muddy situations. This paper proposes an efficient method for real-time computer vision-based recognition of Bangla vehicle LPs using contour analysis and prediction algorithms. The method initially applies gray scaling the input RGB images, histogram equalization to improve the grayscale image quality, edge detection using Sobel edge detector, and adaptive thresholding to convert it to a binary image. The system localizes the vehicle LP based on the maximum rectangular contour area and converts it into a predefined size. Noise removal technique using morphological dilation and erosion operation is used, followed by Gaussian filtering on binary image to improve the image quality further. The system clusters the two-lined LP into seven clusters. The sub-clustering is applied on specific clusters and makes 68 individual sub-clusters. The system extracts vector contour (VC) from each 68 individual classes. After VC extraction, the system normalizes it into a q predefined length. The system applies inter co-relation function (ICF) to categorize each sub-cluster to its previously defined individual class. For that, it calculates the maximum similarity between test and previously trained VCs. The system applies the dependency prediction algorithm in parallel to predict the whole string (district name) in the cluster-1 based on previously categorized class or classes (starting character or characters of the district part). [Formula: see text] (Metro) or (null) from cluster-2, “-” (hyphen) from cluster-3 and 6 are predicted automatically as their positions are fixed. The system is trained using 68 classes in which each class contains 100 samples generated by the augmentation technique. The system is tested using another set of 68 classes with a total of [Formula: see text] images acquiring the recognition accuracy of 96.62% with the mean computational cost of 8.363[Formula: see text]ms/f. The system is also tested using 500 vehicle whole Bangla LPs achieving the mean whole LP recognition accuracy of 95.41% with a mean computational cost of 35.803[Formula: see text]ms/f.","PeriodicalId":177479,"journal":{"name":"Int. J. Image Graph.","volume":"17 3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-03-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125916277","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}