首页 > 最新文献

International Journal of Image and Graphics最新文献

英文 中文
Bayesian Selective Median Filtering for Reduction of Impulse Noise in Digital Color Images 基于贝叶斯选择中值滤波的彩色数字图像脉冲噪声抑制
IF 1.6 Q4 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2022-12-15 DOI: 10.1142/s0219467824500268
Demudu Naidu Chukka, J. Meka, S. Setty, P. Choppala
The focus of this paper is impulse noise reduction in digital color images. The most popular noise reduction schemes are the vector median filter and its many variants that operate by minimizing the aggregate distance from one pixel to every other pixel in a chosen window. This minimizing operation determines the most confirmative pixel based on its similarity to the chosen window and replaces the central pixel of the window with the determined one. The peer group filters, unlike the vector median filters, determine a set of pixels that are most confirmative to the window and then perform filtering over the determined set. Using a set of pixels in the filtering process rather than one pixel is more helpful as it takes into account the full information of all the pixels that seemingly contribute to the signal. Hence, the peer group filters are found to be more robust to noise. However, the peer group for each pixel is computed deterministically using thresholding schemes. A wrong choice of the threshold will easily impair the filtering performance. In this paper, we propose a peer group filtering approach using principles of Bayesian probability theory and clustering. Here, we present a method to compute the probability that a pixel value is clean (not corrupted by impulse noise) and then apply clustering on the probability measure to determine the peer group. The key benefit of this proposal is that the need for thresholding in peer group filtering is completely avoided. Simulation results show that the proposed method performs better than the conventional vector median and peer group filtering methods in terms of noise reduction and structural similarity, thus validating the proposed approach.
本文的研究重点是数字彩色图像中的脉冲噪声抑制。最流行的降噪方案是矢量中值滤波器及其许多变体,它们通过最小化选定窗口中从一个像素到每个其他像素的总距离来运行。这种最小化操作根据其与所选窗口的相似性确定最确定的像素,并用确定的像素替换窗口的中心像素。与矢量中值过滤器不同,对等组过滤器确定一组对窗口最确定的像素,然后对确定的像素集执行过滤。在滤波过程中使用一组像素而不是一个像素更有帮助,因为它考虑了所有似乎对信号有贡献的像素的全部信息。因此,发现对等组滤波器对噪声具有更强的鲁棒性。然而,每个像素的对等组是使用阈值方案确定性地计算的。阈值选择不当,容易影响滤波性能。在本文中,我们提出了一种利用贝叶斯概率理论和聚类原理的对等组过滤方法。在这里,我们提出了一种方法来计算像素值是干净的(不被脉冲噪声破坏)的概率,然后在概率度量上应用聚类来确定对等组。该建议的主要优点是完全避免了对等组过滤中阈值的需要。仿真结果表明,该方法在降噪和结构相似度方面优于传统的向量中值滤波和对等组滤波方法,验证了所提方法的有效性。
{"title":"Bayesian Selective Median Filtering for Reduction of Impulse Noise in Digital Color Images","authors":"Demudu Naidu Chukka, J. Meka, S. Setty, P. Choppala","doi":"10.1142/s0219467824500268","DOIUrl":"https://doi.org/10.1142/s0219467824500268","url":null,"abstract":"The focus of this paper is impulse noise reduction in digital color images. The most popular noise reduction schemes are the vector median filter and its many variants that operate by minimizing the aggregate distance from one pixel to every other pixel in a chosen window. This minimizing operation determines the most confirmative pixel based on its similarity to the chosen window and replaces the central pixel of the window with the determined one. The peer group filters, unlike the vector median filters, determine a set of pixels that are most confirmative to the window and then perform filtering over the determined set. Using a set of pixels in the filtering process rather than one pixel is more helpful as it takes into account the full information of all the pixels that seemingly contribute to the signal. Hence, the peer group filters are found to be more robust to noise. However, the peer group for each pixel is computed deterministically using thresholding schemes. A wrong choice of the threshold will easily impair the filtering performance. In this paper, we propose a peer group filtering approach using principles of Bayesian probability theory and clustering. Here, we present a method to compute the probability that a pixel value is clean (not corrupted by impulse noise) and then apply clustering on the probability measure to determine the peer group. The key benefit of this proposal is that the need for thresholding in peer group filtering is completely avoided. Simulation results show that the proposed method performs better than the conventional vector median and peer group filtering methods in terms of noise reduction and structural similarity, thus validating the proposed approach.","PeriodicalId":44688,"journal":{"name":"International Journal of Image and Graphics","volume":" ","pages":""},"PeriodicalIF":1.6,"publicationDate":"2022-12-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"46527184","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Chronic Kidney Disease Prediction Using ML-Based Neuro-Fuzzy Model 基于ml的神经模糊模型预测慢性肾脏疾病
IF 1.6 Q4 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2022-12-15 DOI: 10.1142/s0219467823400132
S. Praveen, V. E. Jyothi, Chokka Anuradha, K. VenuGopal, V. Shariff, S. Sindhura
Nowadays, in most countries, the most dangerous and life threatening infection is Chronic Kidney Disease (CKD). A progressive malfunctioning of the kidneys and less effectiveness of the kidney are considered CKD. CKD can be a life threatening disease if it continues for longer period of time. Prediction of chronic disease in early stage is very crucial so that sustainable care of the patient is taken to prevent menacing situations. Most of the developing countries are being affected by this deadly disease and treatment applied for this disease is also very expensive, here in this paper, a Machine Learning (ML)-positioned approach called Neuro-Fuzzy model is used for prediction belonging to CKD. Based on the image processing technique, fibrosis proportions are detected in the kidney tissues. It also builds a system for identifying and detection of CKD at an early stage. Neuro-Fuzzy model is based on ML which can detect risk of CKD patients. Compared with other conventional methods such as Support Vector Machine (SVM) and K-Nearest Neighbor (KNN), the proposed method of this paper — ML-based Neuro-Fuzzy logic method — obtained 97% accuracy in CKD prediction. This method can be evaluated based on various parameters such as Precision, Accuracy, Recall and F1-Score in CKD prediction. From the results, the patients having high risk of chronic disease can be predicted.
目前,在大多数国家,最危险和威胁生命的感染是慢性肾脏疾病(CKD)。肾脏的进行性功能障碍和肾功能下降被认为是慢性肾病。慢性肾病如果持续时间较长,可能是一种危及生命的疾病。在早期阶段预测慢性疾病是非常重要的,以便采取可持续的护理病人,以防止危险的情况。大多数发展中国家都受到这种致命疾病的影响,治疗这种疾病也非常昂贵,在本文中,一种名为神经模糊模型的机器学习(ML)定位方法被用于预测属于CKD。基于图像处理技术,检测肾脏组织的纤维化比例。建立了CKD早期识别和检测系统。基于ML的神经模糊模型可以检测CKD患者的风险。与支持向量机(SVM)和k近邻(KNN)等传统方法相比,本文提出的基于ml的神经模糊逻辑方法对CKD的预测准确率达到97%。该方法可根据CKD预测的Precision、Accuracy、Recall和F1-Score等参数进行评价。根据结果,可以预测慢性疾病的高危患者。
{"title":"Chronic Kidney Disease Prediction Using ML-Based Neuro-Fuzzy Model","authors":"S. Praveen, V. E. Jyothi, Chokka Anuradha, K. VenuGopal, V. Shariff, S. Sindhura","doi":"10.1142/s0219467823400132","DOIUrl":"https://doi.org/10.1142/s0219467823400132","url":null,"abstract":"Nowadays, in most countries, the most dangerous and life threatening infection is Chronic Kidney Disease (CKD). A progressive malfunctioning of the kidneys and less effectiveness of the kidney are considered CKD. CKD can be a life threatening disease if it continues for longer period of time. Prediction of chronic disease in early stage is very crucial so that sustainable care of the patient is taken to prevent menacing situations. Most of the developing countries are being affected by this deadly disease and treatment applied for this disease is also very expensive, here in this paper, a Machine Learning (ML)-positioned approach called Neuro-Fuzzy model is used for prediction belonging to CKD. Based on the image processing technique, fibrosis proportions are detected in the kidney tissues. It also builds a system for identifying and detection of CKD at an early stage. Neuro-Fuzzy model is based on ML which can detect risk of CKD patients. Compared with other conventional methods such as Support Vector Machine (SVM) and K-Nearest Neighbor (KNN), the proposed method of this paper — ML-based Neuro-Fuzzy logic method — obtained 97% accuracy in CKD prediction. This method can be evaluated based on various parameters such as Precision, Accuracy, Recall and F1-Score in CKD prediction. From the results, the patients having high risk of chronic disease can be predicted.","PeriodicalId":44688,"journal":{"name":"International Journal of Image and Graphics","volume":" ","pages":""},"PeriodicalIF":1.6,"publicationDate":"2022-12-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"44430404","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Hybrid Mayfly Lévy Flight Distribution Optimization Algorithm-Tuned Deep Convolutional Neural Network for Indoor–Outdoor Image Classification 基于混合Mayfly lsamvy飞行分布优化算法的深度卷积神经网络室内外图像分类
IF 1.6 Q4 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2022-12-14 DOI: 10.1142/s0219467824500244
J. D. Pakhare, M. Uplane
Image classification in the image is the persistent task to be computed in robotics, automobiles, and machine vision for sustainability. Scene categorization remains one of the challenging parts of various multi-media technologies implied in human–computer communication, robotic navigation, video surveillance, medical diagnosing, tourist guidance, and drone targeting. In this research, a Hybrid Mayfly Lévy flight distribution (MLFD) optimization algorithm-tuned deep convolutional neural network is proposed to effectively classify the image. The feature extraction process is a significant task to be executed as it enhances the classifier performance by reducing the execution time and the computational complexity. Further, the classifier is optimally trained by the Hybrid MLFD algorithm which in turn reduces optimization issues. The accuracy of the proposed MLFD-based Deep-CNN using the SCID-2 dataset is 95.2683% at 80% of training and 97.6425% for 10 K-fold. This manifests that the proposed MLFD-based Deep-CNN outperforms all the conventional methods in terms of accuracy, sensitivity, and specificity.
图像中的图像分类是机器人、汽车和机器视觉中持续需要计算的任务。场景分类仍然是人机通信、机器人导航、视频监控、医疗诊断、旅游指导和无人机瞄准中隐含的各种多媒体技术中具有挑战性的部分之一。在本研究中,提出了一种混合Mayfly l飞行分布(MLFD)优化算法-深度卷积神经网络对图像进行有效分类。特征提取过程是一项重要的任务,它通过减少执行时间和计算复杂度来提高分类器的性能。此外,分类器通过混合MLFD算法进行最佳训练,从而减少了优化问题。使用SCID-2数据集所提出的基于mlfd的Deep-CNN在80%训练时准确率为95.2683%,在10 K-fold时准确率为97.6425%。这表明所提出的基于mlfd的Deep-CNN在准确性、灵敏度和特异性方面优于所有传统方法。
{"title":"Hybrid Mayfly Lévy Flight Distribution Optimization Algorithm-Tuned Deep Convolutional Neural Network for Indoor–Outdoor Image Classification","authors":"J. D. Pakhare, M. Uplane","doi":"10.1142/s0219467824500244","DOIUrl":"https://doi.org/10.1142/s0219467824500244","url":null,"abstract":"Image classification in the image is the persistent task to be computed in robotics, automobiles, and machine vision for sustainability. Scene categorization remains one of the challenging parts of various multi-media technologies implied in human–computer communication, robotic navigation, video surveillance, medical diagnosing, tourist guidance, and drone targeting. In this research, a Hybrid Mayfly Lévy flight distribution (MLFD) optimization algorithm-tuned deep convolutional neural network is proposed to effectively classify the image. The feature extraction process is a significant task to be executed as it enhances the classifier performance by reducing the execution time and the computational complexity. Further, the classifier is optimally trained by the Hybrid MLFD algorithm which in turn reduces optimization issues. The accuracy of the proposed MLFD-based Deep-CNN using the SCID-2 dataset is 95.2683% at 80% of training and 97.6425% for 10 K-fold. This manifests that the proposed MLFD-based Deep-CNN outperforms all the conventional methods in terms of accuracy, sensitivity, and specificity.","PeriodicalId":44688,"journal":{"name":"International Journal of Image and Graphics","volume":" ","pages":""},"PeriodicalIF":1.6,"publicationDate":"2022-12-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"44900484","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Laplace-Based 3D Human Mesh Sequence Compression 基于拉普拉斯的三维人体网格序列压缩
IF 1.6 Q4 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2022-12-14 DOI: 10.1142/s021946782450027x
Shuhan He, Xueming Li, Qiang Fu
Three-dimensional (3D) human mesh sequences obtained by 3D scanning equipment are often used in film and television, games, the internet, and other industries. However, due to the dense point cloud data obtained by 3D scanning equipment, the data of a single frame of a 3D human model is always large. Considering the different topologies of models between different frames, and even the interaction between the human body and other objects, the content of 3D models between different frames is also complex. Therefore, the traditional 3D model compression method always cannot handle the compression of the 3D human mesh sequence. To address this problem, we propose a sequence compression method of 3D human mesh sequence based on the Laplace operator, and test it on the complex interactive behavior of a soccer player bouncing the ball. This method first detects the mesh separation degree of the interactive object and human body, and then divides the sequence into a series of fragments based on the consistency of separation degrees. In each fragment, we employ a deformation algorithm to map keyframe topology to other frames, to improve the compression ratio of the sequence. Our work can be used for the storage of mesh sequences and mobile applications by providing an approach for data compression.
三维扫描设备获得的三维人体网格序列常用于影视、游戏、互联网等行业。然而,由于三维扫描设备获得的点云数据比较密集,使得三维人体模型的单帧数据量往往比较大。考虑到不同帧间模型拓扑结构的不同,甚至人体与其他物体之间的相互作用,不同帧间的三维模型内容也较为复杂。因此,传统的三维模型压缩方法总是不能处理三维人体网格序列的压缩。针对这一问题,提出了一种基于拉普拉斯算子的三维人体网格序列压缩方法,并在足球运动员弹跳球的复杂交互行为上进行了测试。该方法首先检测交互对象与人体的网格分离度,然后根据分离度的一致性将序列划分为一系列片段。在每个片段中,我们采用变形算法将关键帧拓扑映射到其他帧,以提高序列的压缩比。通过提供一种数据压缩方法,我们的工作可以用于网格序列的存储和移动应用程序。
{"title":"Laplace-Based 3D Human Mesh Sequence Compression","authors":"Shuhan He, Xueming Li, Qiang Fu","doi":"10.1142/s021946782450027x","DOIUrl":"https://doi.org/10.1142/s021946782450027x","url":null,"abstract":"Three-dimensional (3D) human mesh sequences obtained by 3D scanning equipment are often used in film and television, games, the internet, and other industries. However, due to the dense point cloud data obtained by 3D scanning equipment, the data of a single frame of a 3D human model is always large. Considering the different topologies of models between different frames, and even the interaction between the human body and other objects, the content of 3D models between different frames is also complex. Therefore, the traditional 3D model compression method always cannot handle the compression of the 3D human mesh sequence. To address this problem, we propose a sequence compression method of 3D human mesh sequence based on the Laplace operator, and test it on the complex interactive behavior of a soccer player bouncing the ball. This method first detects the mesh separation degree of the interactive object and human body, and then divides the sequence into a series of fragments based on the consistency of separation degrees. In each fragment, we employ a deformation algorithm to map keyframe topology to other frames, to improve the compression ratio of the sequence. Our work can be used for the storage of mesh sequences and mobile applications by providing an approach for data compression.","PeriodicalId":44688,"journal":{"name":"International Journal of Image and Graphics","volume":" ","pages":""},"PeriodicalIF":1.6,"publicationDate":"2022-12-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"42933950","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Weighted Graph Embedding Feature with Bi-Directional Long Short-Term Memory Classifier for Multi-Document Text Summarization 基于加权图嵌入特征的双向长短期记忆分类器用于多文档文本摘要
IF 1.6 Q4 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2022-12-10 DOI: 10.1142/s0219467824500220
Samina Mulla, N. Shaikh
In this digital era, there is a tremendous increase in the volume of data, which adds difficulties to the person who utilizes particular applications, such as websites, email, and news. Text summarization targets to reduce the complexity of obtaining statistics from the websites as it compresses the textual document to a short summary without affecting the relevant information. The crucial step in multi-document summarization is obtaining a relationship between the cross-sentence. However, the conventional methods fail to determine the inter-sentence relationship, especially in long documents. This research develops a graph-based neural network to attain an inter-sentence relationship. The significant step in the proposed multi-document text summarization model is forming the weighted graph embedding features. Furthermore, the weighted graph embedding features are utilized to evaluate the relationship between the document’s words and sentences. Finally, the bidirectional long short-term memory (BiLSTM) classifier is utilized to summarize the multi-document text summarization. The experimental analysis uses the three standard datasets, the Daily Mail dataset, Document Understanding Conference (DUC) 2002, and Document Understanding Conference (DUC) 2004 dataset. The experimental outcome demonstrates that the proposed weighted graph embedding feature + BiLSTM model exceeds all the conventional methods with Precision, Recall, and F1 score of 0.5352, 0.6296, and 0.5429, respectively.
在这个数字时代,数据量急剧增加,这给使用特定应用程序(如网站、电子邮件和新闻)的人增加了困难。文本摘要旨在降低从网站获取统计数据的复杂性,因为它将文本文档压缩为简短摘要,而不会影响相关信息。多文档摘要的关键步骤是获取跨句之间的关系。然而,传统的方法无法确定句间关系,尤其是在长文档中。本研究开发了一个基于图的神经网络来获得句子间的关系。所提出的多文档文本摘要模型的重要步骤是形成加权图嵌入特征。此外,利用加权图嵌入特征来评估文档的单词和句子之间的关系。最后,利用双向长短期记忆(BiLSTM)分类器对多文档文本摘要进行了总结。实验分析使用了三个标准数据集,即《每日邮报》数据集、2002年文献理解会议(DUC)数据集和2004年文献理解大会(DUC)数据集。实验结果表明,所提出的加权图嵌入特征+BiLSTM模型超过了所有传统方法,Precision、Recall和F1得分分别为0.5352、0.6296和0.5429。
{"title":"Weighted Graph Embedding Feature with Bi-Directional Long Short-Term Memory Classifier for Multi-Document Text Summarization","authors":"Samina Mulla, N. Shaikh","doi":"10.1142/s0219467824500220","DOIUrl":"https://doi.org/10.1142/s0219467824500220","url":null,"abstract":"In this digital era, there is a tremendous increase in the volume of data, which adds difficulties to the person who utilizes particular applications, such as websites, email, and news. Text summarization targets to reduce the complexity of obtaining statistics from the websites as it compresses the textual document to a short summary without affecting the relevant information. The crucial step in multi-document summarization is obtaining a relationship between the cross-sentence. However, the conventional methods fail to determine the inter-sentence relationship, especially in long documents. This research develops a graph-based neural network to attain an inter-sentence relationship. The significant step in the proposed multi-document text summarization model is forming the weighted graph embedding features. Furthermore, the weighted graph embedding features are utilized to evaluate the relationship between the document’s words and sentences. Finally, the bidirectional long short-term memory (BiLSTM) classifier is utilized to summarize the multi-document text summarization. The experimental analysis uses the three standard datasets, the Daily Mail dataset, Document Understanding Conference (DUC) 2002, and Document Understanding Conference (DUC) 2004 dataset. The experimental outcome demonstrates that the proposed weighted graph embedding feature + BiLSTM model exceeds all the conventional methods with Precision, Recall, and F1 score of 0.5352, 0.6296, and 0.5429, respectively.","PeriodicalId":44688,"journal":{"name":"International Journal of Image and Graphics","volume":"1 1","pages":""},"PeriodicalIF":1.6,"publicationDate":"2022-12-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"41501572","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Convoluted Neighborhood-Based Ordered-Dither Block Truncation Coding for Ear Image Retrieval 基于卷积邻域的有序抖动块截断编码在耳朵图像检索中的应用
IF 1.6 Q4 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2022-12-07 DOI: 10.1142/s0219467824500177
M. N. Sowmya, K. Prasanna
Image retrieval is a significant and hot research topic among researchers that drives the focus of researchers from keyword toward semantic-based image reconstruction. Nevertheless, existing image retrieval investigations still have a shortage of significant semantic image definition and user behavior consideration. Hence, there is a necessity to offer a high level of assistance towards regulating the semantic gap between low-level visual patterns and high-level ideas for a better understanding between humans and machines. Hence, this research devises an effective medical image retrieval strategy using convoluted neighborhood-based Ordered-dither block truncation coding (ODBTC). The developed approach is devised by modifying the ODBTC concept using a convoluted neighborhood mechanism. Here, the convoluted neighborhood-based color co-occurrence feature (CCF) and convoluted neighborhood-based bit pattern feature (BBF) are extracted. Finally, cross-indexing is performed to convert the feature points into binary codes for effective image retrieval. Meanwhile, the proposed convoluted neighborhood-based ODBTC has achieved maximum precision, recall, and f-measure with values of 0.740, 0.680, and 0.709.
图像检索是一个重要而热门的研究课题,它将研究的重点从关键词转向基于语义的图像重建。然而,现有的图像检索研究仍然缺乏重要的语义图像定义和用户行为考虑。因此,有必要为调节低级视觉模式和高级思想之间的语义差距提供高层次的帮助,以便更好地理解人和机器之间的关系。因此,本研究设计了一种有效的基于卷积邻域的有序抖动块截断编码(ODBTC)的医学图像检索策略。所开发的方法是通过使用复杂的邻域机制修改ODBTC概念来设计的。在这里,提取了基于卷积邻域的颜色共现特征(CCF)和基于卷积邻域的位模式特征(BBF)。最后,进行交叉索引,将特征点转换为二进制代码,实现有效的图像检索。同时,本文提出的基于卷积邻域的ODBTC达到了最高的精度、召回率和f-measure值,分别为0.740、0.680和0.709。
{"title":"Convoluted Neighborhood-Based Ordered-Dither Block Truncation Coding for Ear Image Retrieval","authors":"M. N. Sowmya, K. Prasanna","doi":"10.1142/s0219467824500177","DOIUrl":"https://doi.org/10.1142/s0219467824500177","url":null,"abstract":"Image retrieval is a significant and hot research topic among researchers that drives the focus of researchers from keyword toward semantic-based image reconstruction. Nevertheless, existing image retrieval investigations still have a shortage of significant semantic image definition and user behavior consideration. Hence, there is a necessity to offer a high level of assistance towards regulating the semantic gap between low-level visual patterns and high-level ideas for a better understanding between humans and machines. Hence, this research devises an effective medical image retrieval strategy using convoluted neighborhood-based Ordered-dither block truncation coding (ODBTC). The developed approach is devised by modifying the ODBTC concept using a convoluted neighborhood mechanism. Here, the convoluted neighborhood-based color co-occurrence feature (CCF) and convoluted neighborhood-based bit pattern feature (BBF) are extracted. Finally, cross-indexing is performed to convert the feature points into binary codes for effective image retrieval. Meanwhile, the proposed convoluted neighborhood-based ODBTC has achieved maximum precision, recall, and f-measure with values of 0.740, 0.680, and 0.709.","PeriodicalId":44688,"journal":{"name":"International Journal of Image and Graphics","volume":" ","pages":""},"PeriodicalIF":1.6,"publicationDate":"2022-12-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"43886494","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An Enhanced Deep Neural Network-Based Approach for Speaker Recognition Using Triumvirate Euphemism Strategy 基于增强深度神经网络的三元委婉语说话人识别方法
IF 1.6 Q4 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2022-12-05 DOI: 10.1142/s0219467824500074
P. S. Subhashini Pedalanka, M. Satya Sai Ram, Duggirala Sreenivasa Rao
Automatic Speech Recognition (ASR) has been an intensive research area during the recent years in internet to enable natural human–machine communication. However, the existing Deep Neutral Network (DNN) techniques need more focus on feature extraction process and recognition accuracy. Thus, an enhanced deep neural network (DNN)-based approach for speaker recognition with a novel Triumvirate Euphemism Strategy (TES) is proposed. This overcomes poor feature extraction from Mel-Frequency Cepstral Coefficient (MFCC) map by extracting the features based on petite, hefty and artistry of the features. Then, the features are trained with Silhouette Martyrs Method (SMM) without any inter-class and intra-class separability problems and margins are affixed between classes with three new loss functions, namely A-Loss, AM-Loss and AAM-Loss. Additionally, the parallelization is done by a mini-batch-based BP algorithm in DNN. A novel Frenzied Heap Atrophy (FHA) with a multi-GPU model is introduced in addition with DNN to enhance the parallelized computing that accelerates the training procedures. Thus, the outcome of the proposed technique is highly efficient that provides feasible extraction features and gives incredibly precise results with 97.5% accuracy in the recognition of speakers. Moreover, various parameters were discussed to prove the efficiency of the system and also the proposed method outperformed the existing methods in all aspects.
近年来,自动语音识别(ASR)一直是互联网领域的一个深入研究领域,以实现自然的人机通信。然而,现有的深度神经网络(DNN)技术需要更多地关注特征提取过程和识别精度。因此,提出了一种基于增强深度神经网络(DNN)的说话人识别方法,该方法采用了一种新的三元委婉语策略(TES)。这通过基于特征的小、重和艺术性来提取特征,克服了梅尔频率倒谱系数(MFCC)图中特征提取较差的问题。然后,在没有任何类间和类内可分性问题的情况下,用剪影烈士方法(SMM)训练特征,并用三个新的损失函数(即A-loss、AM loss和AAM loss)在类之间附加裕度。此外,在DNN中使用基于小批量的BP算法进行并行化。除了DNN之外,还引入了一种新的具有多GPU模型的疯狂堆萎缩(FHA),以增强并行计算,从而加速训练过程。因此,所提出的技术的结果是高效的,它提供了可行的提取特征,并在说话人识别中以97.5%的准确率给出了令人难以置信的精确结果。此外,还讨论了各种参数来证明系统的有效性,并且所提出的方法在各个方面都优于现有方法。
{"title":"An Enhanced Deep Neural Network-Based Approach for Speaker Recognition Using Triumvirate Euphemism Strategy","authors":"P. S. Subhashini Pedalanka, M. Satya Sai Ram, Duggirala Sreenivasa Rao","doi":"10.1142/s0219467824500074","DOIUrl":"https://doi.org/10.1142/s0219467824500074","url":null,"abstract":"Automatic Speech Recognition (ASR) has been an intensive research area during the recent years in internet to enable natural human–machine communication. However, the existing Deep Neutral Network (DNN) techniques need more focus on feature extraction process and recognition accuracy. Thus, an enhanced deep neural network (DNN)-based approach for speaker recognition with a novel Triumvirate Euphemism Strategy (TES) is proposed. This overcomes poor feature extraction from Mel-Frequency Cepstral Coefficient (MFCC) map by extracting the features based on petite, hefty and artistry of the features. Then, the features are trained with Silhouette Martyrs Method (SMM) without any inter-class and intra-class separability problems and margins are affixed between classes with three new loss functions, namely A-Loss, AM-Loss and AAM-Loss. Additionally, the parallelization is done by a mini-batch-based BP algorithm in DNN. A novel Frenzied Heap Atrophy (FHA) with a multi-GPU model is introduced in addition with DNN to enhance the parallelized computing that accelerates the training procedures. Thus, the outcome of the proposed technique is highly efficient that provides feasible extraction features and gives incredibly precise results with 97.5% accuracy in the recognition of speakers. Moreover, various parameters were discussed to prove the efficiency of the system and also the proposed method outperformed the existing methods in all aspects.","PeriodicalId":44688,"journal":{"name":"International Journal of Image and Graphics","volume":" ","pages":""},"PeriodicalIF":1.6,"publicationDate":"2022-12-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49309807","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
BCS-AE: Integrated Image Compression-Encryption Model Based on AE and Block-CS BCS-AE:基于AE和块CS的集成图像压缩加密模型
IF 1.6 Q4 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2022-11-24 DOI: 10.1142/s021946782350047x
S. Jameel, Jafar Majidpour
For Compressive Sensing problems, a number of techniques have been introduced, including traditional compressed-sensing (CS) image reconstruction and Deep Neural Network (DNN) models. Unfortunately, due to low sampling rates, the quality of image reconstruction is still poor. This paper proposes a lossy image compression model (i.e. BCS-AE), which combines two different types to produce a model that uses more high-quality low-bitrate CS reconstruction. Initially, block-based compressed sensing (BCS) was utilized, and it was done one block at a time by the same operator. It can correctly extract images with complex geometric configurations. Second, we create an AutoEncoder architecture to replace traditional transforms, and we train it with a rate-distortion loss function. The proposed model is trained and then tested on the CelebA and Kodak databases. According to the results, advanced deep learning-based and iterative optimization-based algorithms perform better in terms of compression ratio and reconstruction quality.
对于压缩传感问题,已经引入了许多技术,包括传统的压缩传感(CS)图像重建和深度神经网络(DNN)模型。不幸的是,由于采样率低,图像重建的质量仍然很差。本文提出了一种有损图像压缩模型(即BCS-AE),它结合了两种不同的类型来产生一种使用更高质量的低比特率CS重建的模型。最初,使用基于块的压缩传感(BCS),由同一操作员一次一个块地完成。它可以正确地提取具有复杂几何配置的图像。其次,我们创建了一个AutoEncoder架构来取代传统的变换,并使用率失真损失函数对其进行训练。所提出的模型经过训练,然后在CelebA和Kodak数据库上进行测试。结果表明,基于深度学习和迭代优化的高级算法在压缩比和重建质量方面表现更好。
{"title":"BCS-AE: Integrated Image Compression-Encryption Model Based on AE and Block-CS","authors":"S. Jameel, Jafar Majidpour","doi":"10.1142/s021946782350047x","DOIUrl":"https://doi.org/10.1142/s021946782350047x","url":null,"abstract":"For Compressive Sensing problems, a number of techniques have been introduced, including traditional compressed-sensing (CS) image reconstruction and Deep Neural Network (DNN) models. Unfortunately, due to low sampling rates, the quality of image reconstruction is still poor. This paper proposes a lossy image compression model (i.e. BCS-AE), which combines two different types to produce a model that uses more high-quality low-bitrate CS reconstruction. Initially, block-based compressed sensing (BCS) was utilized, and it was done one block at a time by the same operator. It can correctly extract images with complex geometric configurations. Second, we create an AutoEncoder architecture to replace traditional transforms, and we train it with a rate-distortion loss function. The proposed model is trained and then tested on the CelebA and Kodak databases. According to the results, advanced deep learning-based and iterative optimization-based algorithms perform better in terms of compression ratio and reconstruction quality.","PeriodicalId":44688,"journal":{"name":"International Journal of Image and Graphics","volume":" ","pages":""},"PeriodicalIF":1.6,"publicationDate":"2022-11-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"41789443","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Semi-Supervised Skin Lesion Segmentation via Iterative Mask Optimization 基于迭代掩模优化的半监督皮肤病灶分割
IF 1.6 Q4 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2022-11-24 DOI: 10.1142/s0219467824500207
Fuhe Du, B. Peng, Zaid Al-Huda, Jing Yao
Deep learning-based skin lesion segmentation methods have achieved promising results in the community. However, they are usually based on fully supervised learning and require many high-quality ground truths. Labeling the ground truths takes a lot of labor, material, and financial resources. We propose a novel semi-supervised skin lesion segmentation method to solve this problem. First, a hierarchical image segmentation algorithm is used to generate optimal segmentation maps. Then, fully supervised training is performed on a small part of the images with ground truths. The resulting pseudo masks are generated to train the rest of the images. The optimal segmentation maps are utilized in this process to refine the pseudo masks. Experiments show that the proposed method can improve the performance of semi-supervised learning for skin lesion segmentation by reducing the gap with fully supervised learning methods. Moreover, it can reduce the workload of labeling the ground truths. Extensive experiments are conducted on the open dataset to validate the efficiency of the proposed method. The results show that our method is competitive in improving the quality of semi-supervised segmentation.
基于深度学习的皮肤病变分割方法在社区中取得了很好的效果。然而,它们通常基于完全监督的学习,需要许多高质量的基本事实。给真相贴标签需要大量的人力、物力和财力。我们提出了一种新的半监督皮肤病变分割方法来解决这个问题。首先,使用分层图像分割算法生成最优分割图。然后,对具有基本事实的图像的一小部分进行完全监督训练。生成所得到的伪掩模以训练图像的其余部分。在该过程中利用最优分割图来细化伪掩模。实验表明,该方法通过缩小与全监督学习方法的差距,可以提高半监督学习在皮肤损伤分割中的性能。此外,它可以减少标记基本事实的工作量。在开放数据集上进行了大量实验,以验证所提出方法的有效性。结果表明,我们的方法在提高半监督分割的质量方面是有竞争力的。
{"title":"Semi-Supervised Skin Lesion Segmentation via Iterative Mask Optimization","authors":"Fuhe Du, B. Peng, Zaid Al-Huda, Jing Yao","doi":"10.1142/s0219467824500207","DOIUrl":"https://doi.org/10.1142/s0219467824500207","url":null,"abstract":"Deep learning-based skin lesion segmentation methods have achieved promising results in the community. However, they are usually based on fully supervised learning and require many high-quality ground truths. Labeling the ground truths takes a lot of labor, material, and financial resources. We propose a novel semi-supervised skin lesion segmentation method to solve this problem. First, a hierarchical image segmentation algorithm is used to generate optimal segmentation maps. Then, fully supervised training is performed on a small part of the images with ground truths. The resulting pseudo masks are generated to train the rest of the images. The optimal segmentation maps are utilized in this process to refine the pseudo masks. Experiments show that the proposed method can improve the performance of semi-supervised learning for skin lesion segmentation by reducing the gap with fully supervised learning methods. Moreover, it can reduce the workload of labeling the ground truths. Extensive experiments are conducted on the open dataset to validate the efficiency of the proposed method. The results show that our method is competitive in improving the quality of semi-supervised segmentation.","PeriodicalId":44688,"journal":{"name":"International Journal of Image and Graphics","volume":" ","pages":""},"PeriodicalIF":1.6,"publicationDate":"2022-11-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"46872995","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Entropy-Based Feature Extraction Model for Fundus Images with Deep Learning Model 基于熵的眼底图像深度学习特征提取模型
IF 1.6 Q4 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2022-11-18 DOI: 10.1142/s0219467823400065
S. Gadde, K. Kiran
Diabetic retinopathy (DR) is stated as a disease in the eyes that affects the retina blood vessels and causes blindness. The early diagnosis and detection of the DR in patients preserve the patient’s vision. In general, for the diagnosis of eye diseases, retinal fundus images are employed. The advancement in the automatic diagnosis of diseases attained higher significance for rapid advancement in computing technology in the medical field. Besides, for the diagnosis of the diseases, fundus image automatic detection is involved in the recognition of blood vessels evaluated based on the length, branching pattern, and width. However, fundus images have low contrast and it is difficult to evaluate the identification of the disease in blood vessels. As a result, it is necessary to adopt a consistent automated method to extract blood vessels in the fundus images for DR. The conventional automated localization of the macula and optic disk in the retinal fundus images needs to be improved for DR disease diagnosis. But existing methods are not sufficient for the early identification and detection of DR. This paper proposed an entropy distributed matching global and local clustering (EDMGL) for fundus images. The developed EDMGL comprises the different uncertainties for the evaluation of the classes based on local and global entropy. The fundus image local entropy is evaluated based on the spatial likelihood fuzzifier membership function estimation for segmentation. The final proposed algorithm membership function is estimated using the addition of weighted parameters through membership estimation based on the global and local entropy. The classification performance of the proposed EDMGL is evaluated based on the dice coefficient, segmentation accuracy, and partition entropy. The performance of the proposed EDMGL is comparatively examined with the conventional technique. The comparative analysis expressed that the performance of the proposed EDMGL exhibits [Formula: see text]5% improved performance in terms of accuracy, precision, recall, and F1-score.
糖尿病视网膜病变(DR)是一种眼部疾病,影响视网膜血管并导致失明。患者的早期诊断和发现可以保护患者的视力。一般情况下,对于眼病的诊断,视网膜眼底图像被使用。随着医疗领域计算技术的飞速发展,疾病自动诊断的进步具有更高的意义。此外,对于疾病的诊断,眼底图像自动检测涉及到基于长度、分支模式和宽度评估血管的识别。然而,眼底图像对比度低,难以评价血管病变的识别。因此,有必要采用一致的自动方法提取眼底图像中的血管进行DR诊断。传统的视网膜眼底图像中黄斑和视盘的自动定位方法有待改进,以用于DR疾病的诊断。本文提出了一种基于眼底图像的熵分布匹配全局局部聚类算法(EDMGL)。所开发的EDMGL包含不同的不确定性,用于基于局部熵和全局熵的分类评估。基于空间似然模糊化隶属函数估计眼底图像局部熵进行分割。最后通过基于全局和局部熵的隶属度估计,利用加权参数的加入来估计算法的隶属度函数。基于骰子系数、分割精度和分割熵对该算法的分类性能进行了评价。并与传统方法进行了性能比较。对比分析表明,提出的EDMGL在准确率、精密度、召回率和f1分数方面的性能提高了5%[公式:见文本]。
{"title":"Entropy-Based Feature Extraction Model for Fundus Images with Deep Learning Model","authors":"S. Gadde, K. Kiran","doi":"10.1142/s0219467823400065","DOIUrl":"https://doi.org/10.1142/s0219467823400065","url":null,"abstract":"Diabetic retinopathy (DR) is stated as a disease in the eyes that affects the retina blood vessels and causes blindness. The early diagnosis and detection of the DR in patients preserve the patient’s vision. In general, for the diagnosis of eye diseases, retinal fundus images are employed. The advancement in the automatic diagnosis of diseases attained higher significance for rapid advancement in computing technology in the medical field. Besides, for the diagnosis of the diseases, fundus image automatic detection is involved in the recognition of blood vessels evaluated based on the length, branching pattern, and width. However, fundus images have low contrast and it is difficult to evaluate the identification of the disease in blood vessels. As a result, it is necessary to adopt a consistent automated method to extract blood vessels in the fundus images for DR. The conventional automated localization of the macula and optic disk in the retinal fundus images needs to be improved for DR disease diagnosis. But existing methods are not sufficient for the early identification and detection of DR. This paper proposed an entropy distributed matching global and local clustering (EDMGL) for fundus images. The developed EDMGL comprises the different uncertainties for the evaluation of the classes based on local and global entropy. The fundus image local entropy is evaluated based on the spatial likelihood fuzzifier membership function estimation for segmentation. The final proposed algorithm membership function is estimated using the addition of weighted parameters through membership estimation based on the global and local entropy. The classification performance of the proposed EDMGL is evaluated based on the dice coefficient, segmentation accuracy, and partition entropy. The performance of the proposed EDMGL is comparatively examined with the conventional technique. The comparative analysis expressed that the performance of the proposed EDMGL exhibits [Formula: see text]5% improved performance in terms of accuracy, precision, recall, and F1-score.","PeriodicalId":44688,"journal":{"name":"International Journal of Image and Graphics","volume":" ","pages":""},"PeriodicalIF":1.6,"publicationDate":"2022-11-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"43949569","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
International Journal of Image and Graphics
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1