A disaster is a devastating incident that causes a serious disruption of the functions of a community. It leads to loss of human life and environmental and financial losses. Natural disasters cause damage and privation that could last for months and even years. Immediate steps need to be taken and social media platforms like Twitter help to provide relief to the affected public. However, it is difficult to analyze high-volume data obtained from social media posts. Therefore, the efficiency and accuracy of useful data extracted from the enormous posts related to disaster are low. Satellite imagery is gaining popularity because of its ability to cover large temporal and spatial areas. But, both the social media and satellite imagery require the use of automated methods to avoid the errors caused by humans. Deep learning and machine learning have become extremely popular for text and image classification tasks. In this paper, a review has been done on natural disaster detection through information obtained from social media and satellite images using deep learning and machine learning.
{"title":"A Review on Natural Disaster Detection in Social Media and Satellite Imagery Using Machine Learning and Deep Learning","authors":"Swapandeep Kaur, Sheifali Gupta, Swati Singh, Tanvi Arora","doi":"10.1142/s0219467822500401","DOIUrl":"https://doi.org/10.1142/s0219467822500401","url":null,"abstract":"A disaster is a devastating incident that causes a serious disruption of the functions of a community. It leads to loss of human life and environmental and financial losses. Natural disasters cause damage and privation that could last for months and even years. Immediate steps need to be taken and social media platforms like Twitter help to provide relief to the affected public. However, it is difficult to analyze high-volume data obtained from social media posts. Therefore, the efficiency and accuracy of useful data extracted from the enormous posts related to disaster are low. Satellite imagery is gaining popularity because of its ability to cover large temporal and spatial areas. But, both the social media and satellite imagery require the use of automated methods to avoid the errors caused by humans. Deep learning and machine learning have become extremely popular for text and image classification tasks. In this paper, a review has been done on natural disaster detection through information obtained from social media and satellite images using deep learning and machine learning.","PeriodicalId":177479,"journal":{"name":"Int. J. Image Graph.","volume":"56 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-10-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"120916198","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2021-10-06DOI: 10.1142/s0219467822500528
R. Jatain, Manisha Jailia
Effective face recognition is accomplished using the extraction of features and classification. Though there are multiple techniques for face image recognition, full face recognition in real-time is quite difficult. One of the emerging and promising methods to address this challenge in face recognition is deep learning networks. The inevitable network tool associated with the face recognition method with deep learning systems is convolutional neural networks (CNNs). This research intends to develop a new method for face recognition using adaptive intelligent methods. The main phases of the proposed method are (a) data collection, (b) image pre-processing, (c) normalization, (d) pattern extraction, and (e) recognition. Initially, the images for face recognition are gathered from CPFW, Yale datasets, and the MIT-CBCL dataset. The image pre-processing is performed by the Gaussian filtering method. Further, the normalization of the image will be done, which is a process that alters the range of pixel intensities and can handle the poor contrast due to glare. Then a new descriptor called adaptive local tri Weber pattern (ALTrWP) acts as a pattern extractor. In the recognition phase, the VGG16 architecture with new chick updated-chicken swarm optimization (NSU-CSO) is used. As the modification, VGG16 architecture will be enhanced by this optimization technique. The performance of the developed method is analyzed on two standards face database. Experimental results are compared with different machine learning approaches concerned with noteworthy measures, which demonstrate the efficiency of the considered classifier.
{"title":"Enhanced Face Recognition Using Adaptive Local Tri Weber Pattern with Improved Deep Learning Architecture","authors":"R. Jatain, Manisha Jailia","doi":"10.1142/s0219467822500528","DOIUrl":"https://doi.org/10.1142/s0219467822500528","url":null,"abstract":"Effective face recognition is accomplished using the extraction of features and classification. Though there are multiple techniques for face image recognition, full face recognition in real-time is quite difficult. One of the emerging and promising methods to address this challenge in face recognition is deep learning networks. The inevitable network tool associated with the face recognition method with deep learning systems is convolutional neural networks (CNNs). This research intends to develop a new method for face recognition using adaptive intelligent methods. The main phases of the proposed method are (a) data collection, (b) image pre-processing, (c) normalization, (d) pattern extraction, and (e) recognition. Initially, the images for face recognition are gathered from CPFW, Yale datasets, and the MIT-CBCL dataset. The image pre-processing is performed by the Gaussian filtering method. Further, the normalization of the image will be done, which is a process that alters the range of pixel intensities and can handle the poor contrast due to glare. Then a new descriptor called adaptive local tri Weber pattern (ALTrWP) acts as a pattern extractor. In the recognition phase, the VGG16 architecture with new chick updated-chicken swarm optimization (NSU-CSO) is used. As the modification, VGG16 architecture will be enhanced by this optimization technique. The performance of the developed method is analyzed on two standards face database. Experimental results are compared with different machine learning approaches concerned with noteworthy measures, which demonstrate the efficiency of the considered classifier.","PeriodicalId":177479,"journal":{"name":"Int. J. Image Graph.","volume":"73 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-10-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123628745","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2021-10-06DOI: 10.1142/s0219467822500255
S. Sridhar, K. Anbarasan
Data imbalance occurring among multiclass datasets is very common in real-world applications. Existing studies reveal that various attempts were made in the past to overcome this multiclass imbalance problem, which is a severe issue related to the typical supervised machine learning methods such as classification and regression. But, still there exists a need to handle the imbalance problem efficiently as the datasets include both safe and unsafe minority samples. Most of the widely used oversampling techniques like SMOTE and its variants face challenges in replicating or generating the new data instances for balancing them across multiple classes, particularly when the imbalance is high and the number of rare samples count is too minimal thus leading the classifier to misclassify the data instances. To lessen this problem, we proposed a new data balancing method namely a two-stage iterative ensemble method to tackle the imbalance in multiclass environment. The proposed approach focuses on the rare minority sample’s influence on learning from imbalanced datasets and the main idea of the proposed approach is to balance the data without any change in class distribution before it gets trained by the learner such that it improves the learner’s learning process. Also, the proposed approach is compared against two widely used oversampling techniques and the results reveals that the proposed approach shows a much significant improvement in the learning process among the multiclass imbalanced data.
{"title":"Performance Analysis of Two-Stage Iterative Ensemble Method over Random Oversampling Methods on Multiclass Imbalanced Datasets","authors":"S. Sridhar, K. Anbarasan","doi":"10.1142/s0219467822500255","DOIUrl":"https://doi.org/10.1142/s0219467822500255","url":null,"abstract":"Data imbalance occurring among multiclass datasets is very common in real-world applications. Existing studies reveal that various attempts were made in the past to overcome this multiclass imbalance problem, which is a severe issue related to the typical supervised machine learning methods such as classification and regression. But, still there exists a need to handle the imbalance problem efficiently as the datasets include both safe and unsafe minority samples. Most of the widely used oversampling techniques like SMOTE and its variants face challenges in replicating or generating the new data instances for balancing them across multiple classes, particularly when the imbalance is high and the number of rare samples count is too minimal thus leading the classifier to misclassify the data instances. To lessen this problem, we proposed a new data balancing method namely a two-stage iterative ensemble method to tackle the imbalance in multiclass environment. The proposed approach focuses on the rare minority sample’s influence on learning from imbalanced datasets and the main idea of the proposed approach is to balance the data without any change in class distribution before it gets trained by the learner such that it improves the learner’s learning process. Also, the proposed approach is compared against two widely used oversampling techniques and the results reveals that the proposed approach shows a much significant improvement in the learning process among the multiclass imbalanced data.","PeriodicalId":177479,"journal":{"name":"Int. J. Image Graph.","volume":"13 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-10-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122548635","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2021-10-06DOI: 10.1142/s0219467822400034
Naresh Sammeta, L. Parthiban
Recent healthcare systems are defined as highly complex and expensive. But it can be decreased with enhanced electronic health records (EHR) management, using blockchain technology. The healthcare sector in today’s world needs to address two major issues, namely data ownership and data security. Therefore, blockchain technology is employed to access and distribute the EHRs. With this motivation, this paper presents novel data ownership and secure medical data transmission model using optimal multiple key-based homomorphic encryption (MHE) with Hyperledger blockchain (OMHE-HBC). The presented OMHE-HBC model enables the patients to access their own data, provide permission to hospital authorities, revoke permission from hospital authorities, and permit emergency contacts. The proposed model involves the MHE technique to securely transmit the data to the cloud and prevent unauthorized access to it. Besides, the optimal key generation process in the MHE technique takes place using a hosted cuckoo optimization (HCO) algorithm. In addition, the proposed model enables sharing of EHRs by the use of multi-channel HBC, which makes use of one blockchain to save patient visits and another one for the medical institutions in recoding links that point to EHRs stored in external systems. A complete set of experiments were carried out in order to validate the performance of the suggested model, and the results were analyzed under many aspects. A comprehensive comparison of results analysis reveals that the suggested model outperforms the other techniques.
{"title":"Data Ownership and Secure Medical Data Transmission using Optimal Multiple Key-Based Homomorphic Encryption with Hyperledger Blockchain","authors":"Naresh Sammeta, L. Parthiban","doi":"10.1142/s0219467822400034","DOIUrl":"https://doi.org/10.1142/s0219467822400034","url":null,"abstract":"Recent healthcare systems are defined as highly complex and expensive. But it can be decreased with enhanced electronic health records (EHR) management, using blockchain technology. The healthcare sector in today’s world needs to address two major issues, namely data ownership and data security. Therefore, blockchain technology is employed to access and distribute the EHRs. With this motivation, this paper presents novel data ownership and secure medical data transmission model using optimal multiple key-based homomorphic encryption (MHE) with Hyperledger blockchain (OMHE-HBC). The presented OMHE-HBC model enables the patients to access their own data, provide permission to hospital authorities, revoke permission from hospital authorities, and permit emergency contacts. The proposed model involves the MHE technique to securely transmit the data to the cloud and prevent unauthorized access to it. Besides, the optimal key generation process in the MHE technique takes place using a hosted cuckoo optimization (HCO) algorithm. In addition, the proposed model enables sharing of EHRs by the use of multi-channel HBC, which makes use of one blockchain to save patient visits and another one for the medical institutions in recoding links that point to EHRs stored in external systems. A complete set of experiments were carried out in order to validate the performance of the suggested model, and the results were analyzed under many aspects. A comprehensive comparison of results analysis reveals that the suggested model outperforms the other techniques.","PeriodicalId":177479,"journal":{"name":"Int. J. Image Graph.","volume":"28 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-10-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116551502","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2021-10-06DOI: 10.1142/s0219467822500504
Kalpesh R. Jadav, Arvind R. Yadav
Shadow leads to failure of moving target positioning, segmentation, tracking, and classification in the video surveillance system thus shadow detection and removal is essential for further computer vision process. The existing state-of-the-art methods for dynamic shadow detection have produced a high discrimination rate but a poor detection rate (foreground pixels are classified as shadow pixels). This paper proposes an effective method for dynamic shadow detection and removal based on intensity ratio along with frame difference, gamma correction, and morphology operations. The performance of the proposed method has been tested on two outdoor ATON datasets, namely, highway-I and highway-III for vehicle tracking systems. The proposed method has produced a discrimination rate of 89.07% and a detection rate of 80.79% for highway-I video sequences. Similarly, for a highway-III video sequence, the discrimination rate of 85.60% and detection rate of 84.05% have been obtained. Investigational outcomes show that the proposed method is the simple, steadiest, and robust for dynamic shadow detection on the dataset used in this work.
{"title":"Dynamic Shadow Detection and Removal for Vehicle Tracking System","authors":"Kalpesh R. Jadav, Arvind R. Yadav","doi":"10.1142/s0219467822500504","DOIUrl":"https://doi.org/10.1142/s0219467822500504","url":null,"abstract":"Shadow leads to failure of moving target positioning, segmentation, tracking, and classification in the video surveillance system thus shadow detection and removal is essential for further computer vision process. The existing state-of-the-art methods for dynamic shadow detection have produced a high discrimination rate but a poor detection rate (foreground pixels are classified as shadow pixels). This paper proposes an effective method for dynamic shadow detection and removal based on intensity ratio along with frame difference, gamma correction, and morphology operations. The performance of the proposed method has been tested on two outdoor ATON datasets, namely, highway-I and highway-III for vehicle tracking systems. The proposed method has produced a discrimination rate of 89.07% and a detection rate of 80.79% for highway-I video sequences. Similarly, for a highway-III video sequence, the discrimination rate of 85.60% and detection rate of 84.05% have been obtained. Investigational outcomes show that the proposed method is the simple, steadiest, and robust for dynamic shadow detection on the dataset used in this work.","PeriodicalId":177479,"journal":{"name":"Int. J. Image Graph.","volume":"7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-10-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124494124","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2021-10-06DOI: 10.1142/s0219467822500516
C. Indhumathi, V. Murugan, G. Muthulakshmii
Nowadays, action recognition has gained more attention from the computer vision community. Normally for recognizing human actions, spatial and temporal features are extracted. Two-stream convolutional neural network is used commonly for human action recognition in videos. In this paper, Adaptive motion Attentive Correlated Temporal Feature (ACTF) is used for temporal feature extractor. The temporal average pooling in inter-frame is used for extracting the inter-frame regional correlation feature and mean feature. This proposed method has better accuracy of 96.9% for UCF101 and 74.6% for HMDB51 datasets, respectively, which are higher than the other state-of-the-art methods.
{"title":"Human Action Recognition Using Spatio-Temporal Multiplier Network and Attentive Correlated Temporal Feature","authors":"C. Indhumathi, V. Murugan, G. Muthulakshmii","doi":"10.1142/s0219467822500516","DOIUrl":"https://doi.org/10.1142/s0219467822500516","url":null,"abstract":"Nowadays, action recognition has gained more attention from the computer vision community. Normally for recognizing human actions, spatial and temporal features are extracted. Two-stream convolutional neural network is used commonly for human action recognition in videos. In this paper, Adaptive motion Attentive Correlated Temporal Feature (ACTF) is used for temporal feature extractor. The temporal average pooling in inter-frame is used for extracting the inter-frame regional correlation feature and mean feature. This proposed method has better accuracy of 96.9% for UCF101 and 74.6% for HMDB51 datasets, respectively, which are higher than the other state-of-the-art methods.","PeriodicalId":177479,"journal":{"name":"Int. J. Image Graph.","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-10-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130202695","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2021-09-25DOI: 10.1142/s021946782250036x
Hadi Salehi
Images are widely used in engineering. Unfortunately, medical ultrasound images and synthetic aperture radar (SAR) images are mainly degraded by an intrinsic noise called speckle. Therefore, de-speckling is a main pre-processing stage for degraded images. In this paper, first, an optimized adaptive Wiener filter (OAWF) is proposed. OAWF can be applied to the input image without the need for logarithmic transform. In addition its performance is improved. Next, the coefficient of variation (CV) is computed from the input image. With the help of CV, the guided filter converts to an improved guided filter (IGF). Next, the improved guided filter is applied on the image. Subsequently, the fast bilateral filter is applied on the image. The proposed filter has a better image detail preservation compared to some other standard methods. The experimental outcomes show that the proposed denoising algorithm is able to preserve image details and edges compared with other de-speckling methods.
{"title":"Image De-Speckling Based on the Coefficient of Variation, Improved Guided Filter, and Fast Bilateral Filter","authors":"Hadi Salehi","doi":"10.1142/s021946782250036x","DOIUrl":"https://doi.org/10.1142/s021946782250036x","url":null,"abstract":"Images are widely used in engineering. Unfortunately, medical ultrasound images and synthetic aperture radar (SAR) images are mainly degraded by an intrinsic noise called speckle. Therefore, de-speckling is a main pre-processing stage for degraded images. In this paper, first, an optimized adaptive Wiener filter (OAWF) is proposed. OAWF can be applied to the input image without the need for logarithmic transform. In addition its performance is improved. Next, the coefficient of variation (CV) is computed from the input image. With the help of CV, the guided filter converts to an improved guided filter (IGF). Next, the improved guided filter is applied on the image. Subsequently, the fast bilateral filter is applied on the image. The proposed filter has a better image detail preservation compared to some other standard methods. The experimental outcomes show that the proposed denoising algorithm is able to preserve image details and edges compared with other de-speckling methods.","PeriodicalId":177479,"journal":{"name":"Int. J. Image Graph.","volume":"23 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-09-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124956632","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2021-08-11DOI: 10.1142/s0219467822500322
M. Rajalakshmi, K. Annapurani
Image classification is a complicated process of classifying an image based on its visual representation. This paper portrays the need for adapting and applying a suitable image enhancement and denoising technique in order to arrive at a successful classification of data captured remotely. Biometric properties that are widely explored today are very important for authentication purposes. Noise may be the result of incorrect vein detection in the accepted image, thus explaining the need for a better development technique. This work provides subjective and objective analysis of the performance of various image enhancement filters in the spatial domain. After performing these pre-processing steps, the vein map and the corresponding vein graph can be easily obtained with minimal extraction steps, in which the appropriate Graph Matching method can be used to evaluate hand vein graphs thus performing the person authentication. The analysis result shows that the image enhancement filter performs better as an image enhancement filter compared to all other filters. Image quality measures (IQMs) are also tabulated for the evaluation of image quality.
{"title":"Enhancement of Vascular Patterns in Palm Images Using Various Image Enhancement Techniques for Person Identification","authors":"M. Rajalakshmi, K. Annapurani","doi":"10.1142/s0219467822500322","DOIUrl":"https://doi.org/10.1142/s0219467822500322","url":null,"abstract":"Image classification is a complicated process of classifying an image based on its visual representation. This paper portrays the need for adapting and applying a suitable image enhancement and denoising technique in order to arrive at a successful classification of data captured remotely. Biometric properties that are widely explored today are very important for authentication purposes. Noise may be the result of incorrect vein detection in the accepted image, thus explaining the need for a better development technique. This work provides subjective and objective analysis of the performance of various image enhancement filters in the spatial domain. After performing these pre-processing steps, the vein map and the corresponding vein graph can be easily obtained with minimal extraction steps, in which the appropriate Graph Matching method can be used to evaluate hand vein graphs thus performing the person authentication. The analysis result shows that the image enhancement filter performs better as an image enhancement filter compared to all other filters. Image quality measures (IQMs) are also tabulated for the evaluation of image quality.","PeriodicalId":177479,"journal":{"name":"Int. J. Image Graph.","volume":"105 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-08-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134625339","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2021-07-30DOI: 10.1142/s0219467822500346
M. Wagdy, Khaild Amin, Mina Ibrahim
In this work, we aim to solve the multi-warping document image problems. We can overcome the limitations of the previous dewarping algorithms to recover the shape of the document. The proposed method is based on a well-defined pattern to simulate the distorted and undistorted connected component of document images. Some pairs of control points are selected for each connected component and its ground truth pattern to define the mapping function between them. The dewarping process transforms the warping connected component according to the geometric transformation defined by the calculated mapping function. Results on document dewarping dataset CBDAR demonstrate the effectiveness of our method. OCR error metrics are also used to evaluate the performance of the proposed approach.
{"title":"Detection and Correction of Multi-Warping Document Image","authors":"M. Wagdy, Khaild Amin, Mina Ibrahim","doi":"10.1142/s0219467822500346","DOIUrl":"https://doi.org/10.1142/s0219467822500346","url":null,"abstract":"In this work, we aim to solve the multi-warping document image problems. We can overcome the limitations of the previous dewarping algorithms to recover the shape of the document. The proposed method is based on a well-defined pattern to simulate the distorted and undistorted connected component of document images. Some pairs of control points are selected for each connected component and its ground truth pattern to define the mapping function between them. The dewarping process transforms the warping connected component according to the geometric transformation defined by the calculated mapping function. Results on document dewarping dataset CBDAR demonstrate the effectiveness of our method. OCR error metrics are also used to evaluate the performance of the proposed approach.","PeriodicalId":177479,"journal":{"name":"Int. J. Image Graph.","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129309761","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2021-07-23DOI: 10.1142/S0219467822500371
Gaber Hassan, K. Hosny, R. M. Farouk, A. Alzohairy
One of the most often used techniques to represent color images is quaternion algebra. This study introduces the quaternion Krawtchouk moments, QKrMs, as a new set of moments to represent color images. Krawtchouk moments (KrMs) represent one type of discrete moments. QKrMs use traditional Krawtchouk moments of each color channel to describe color images. This new set of moments is defined by using orthogonal polynomials called the Krawtchouk polynomials. The stability against the translation, rotation, and scaling transformations for QKrMs is discussed. The performance of the proposed QKrMs is evaluated against other discrete quaternion moments for image reconstruction capability, toughness against various types of noise, invariance to similarity transformations, color face image recognition, and CPU elapsed times.
{"title":"New Set of Invariant Quaternion Krawtchouk Moments for Color Image Representation and Recognition","authors":"Gaber Hassan, K. Hosny, R. M. Farouk, A. Alzohairy","doi":"10.1142/S0219467822500371","DOIUrl":"https://doi.org/10.1142/S0219467822500371","url":null,"abstract":"One of the most often used techniques to represent color images is quaternion algebra. This study introduces the quaternion Krawtchouk moments, QKrMs, as a new set of moments to represent color images. Krawtchouk moments (KrMs) represent one type of discrete moments. QKrMs use traditional Krawtchouk moments of each color channel to describe color images. This new set of moments is defined by using orthogonal polynomials called the Krawtchouk polynomials. The stability against the translation, rotation, and scaling transformations for QKrMs is discussed. The performance of the proposed QKrMs is evaluated against other discrete quaternion moments for image reconstruction capability, toughness against various types of noise, invariance to similarity transformations, color face image recognition, and CPU elapsed times.","PeriodicalId":177479,"journal":{"name":"Int. J. Image Graph.","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129884956","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}