首页 > 最新文献

2023 6th International Conference on Pattern Recognition and Image Analysis (IPRIA)最新文献

英文 中文
Diagnosis of COVID-19 cases from viral pneumonia and normal ones based on transfer learning approach: Xception-GRU 基于迁移学习方法的病毒性肺炎与正常肺炎病例诊断:例外- gru
Pub Date : 2023-02-14 DOI: 10.1109/IPRIA59240.2023.10147183
Shahla Najaflou, Fatemeh Sadat Lesani
The World Health Organization (WHO) considered it difficult to describe the information about the spread of critical symptoms of the Coronavirus due to the different behaviors of the COVID −19 virus. Most people only experience symptoms when the symptoms of the Coronavirus reach an acute stage, and others do not experience any symptoms at all. Lung scan images are one of the ways to distinguish COVID-19 from other similar diseases, such as pneumonia. The emerging novel of the coronavirus and the similarity of pulmonary complications cause the doctor to misdiagnose. In this paper, we utilize 13967 samples of lung scan images to diagnose COVID-19 cases from viral pneumonia and normal ones. This paper proposes an Xception based transfer learning approach to extract the deep features of each image based on depthwise separable convolutions. We extend the Xception architecture by adding a Gated Recurrent Unit (GRU) and a fully connected layer and fine-tune the model to adjust a more abstract representation of features to classify them. The obtained results show the effectiveness of our proposed hybrid method in detecting cases of COVID-19 from normal and viral pneumonia with an accuracy and precision of 95.71% and 94.24%, respectively, which improves the state-of-the-art results.
世界卫生组织(WHO)认为,由于新冠病毒的不同行为,很难描述新冠病毒关键症状的传播信息。大多数人只有在冠状病毒症状达到急性阶段时才会出现症状,而其他人根本没有任何症状。肺部扫描图像是区分COVID-19与其他类似疾病(如肺炎)的方法之一。新型冠状病毒的出现和肺部并发症的相似性导致医生误诊。本文利用13967张肺部扫描图像样本,将COVID-19病例从病毒性肺炎和正常肺炎中诊断出来。本文提出了一种基于异常的迁移学习方法,基于深度可分离卷积提取图像的深度特征。我们通过添加门控循环单元(GRU)和全连接层来扩展异常架构,并微调模型以调整更抽象的特征表示来对它们进行分类。结果表明,本文提出的混合方法在正常肺炎和病毒性肺炎中检测COVID-19病例的准确性和精密度分别为95.71%和94.24%,提高了现有结果。
{"title":"Diagnosis of COVID-19 cases from viral pneumonia and normal ones based on transfer learning approach: Xception-GRU","authors":"Shahla Najaflou, Fatemeh Sadat Lesani","doi":"10.1109/IPRIA59240.2023.10147183","DOIUrl":"https://doi.org/10.1109/IPRIA59240.2023.10147183","url":null,"abstract":"The World Health Organization (WHO) considered it difficult to describe the information about the spread of critical symptoms of the Coronavirus due to the different behaviors of the COVID −19 virus. Most people only experience symptoms when the symptoms of the Coronavirus reach an acute stage, and others do not experience any symptoms at all. Lung scan images are one of the ways to distinguish COVID-19 from other similar diseases, such as pneumonia. The emerging novel of the coronavirus and the similarity of pulmonary complications cause the doctor to misdiagnose. In this paper, we utilize 13967 samples of lung scan images to diagnose COVID-19 cases from viral pneumonia and normal ones. This paper proposes an Xception based transfer learning approach to extract the deep features of each image based on depthwise separable convolutions. We extend the Xception architecture by adding a Gated Recurrent Unit (GRU) and a fully connected layer and fine-tune the model to adjust a more abstract representation of features to classify them. The obtained results show the effectiveness of our proposed hybrid method in detecting cases of COVID-19 from normal and viral pneumonia with an accuracy and precision of 95.71% and 94.24%, respectively, which improves the state-of-the-art results.","PeriodicalId":109390,"journal":{"name":"2023 6th International Conference on Pattern Recognition and Image Analysis (IPRIA)","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-02-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126560309","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Application of Explainable Convolutional Neural Networks on the Differential Diagnosis of Covid_19 and Pneumonia using Chest Radiograph 可解释卷积神经网络在新冠肺炎胸片鉴别诊断中的应用
Pub Date : 2023-02-14 DOI: 10.1109/IPRIA59240.2023.10147169
Fereshteh Zandi, H. Ebrahimpour-Komleh, Hassan Homayoun
The Covid_19 disease is one of the deadliest inflammatories and chronic and acute diseases of the human respiratory system, which is the result of the inhibition of a virus called corona in the respiratory organs since the spread of this virus is rapid and has affected many people in the world. a specialist needs to be carefully evaluated to diagnose the disease based on X-ray images because the number of patients with Covid_19 exceeds the capacity of hospitals and taking care of a large number of people is tedious work that can reduce the accuracy of the doctor in diagnosing the disease. In addition, in such cases, the absence of a specialist doctor can lead to misdiagnosis and incorrect prescribing. In this article, we intend to provide an approach to accelerate the diagnosis process and reduce the workload of specialists automatically, which in addition to helping physicians in hospitals that do not have a specialist physician, also allows patients to be diagnosed and treated. we use pre-trained UNet to extract the lung balloons, which eliminates the extra noise and parts in the X-ray image and then we give the generated images to a convolutional neural network model designed to diagnose and classify Covid_19 disease from Pneumonia, and finally, we use Grad-CAM and Vanilla Gradient and Smooth Grad techniques to validate the designed model. according to the results, our proposed approach using evaluation metrics was able to achieve the highest degree of accuracy in distinguishing Covid_19 disease from Pneumonia.
covid - 19疾病是人类呼吸系统最致命的炎症和慢性和急性疾病之一,这是由于呼吸器官中被称为冠状病毒的病毒被抑制的结果,因为这种病毒传播迅速,并影响了世界上许多人。由于患者数量超过医院的能力,而且照顾大量患者是一项繁琐的工作,可能会降低医生诊断疾病的准确性,因此需要对专家进行仔细的评估,才能根据x射线图像诊断疾病。此外,在这种情况下,专科医生的缺席可能导致误诊和错误的处方。在本文中,我们打算提供一种方法来加速诊断过程并自动减少专家的工作量,除了帮助没有专科医生的医院的医生之外,还可以对患者进行诊断和治疗。我们使用预训练的UNet提取肺气球,消除x射线图像中多余的噪声和部分,然后将生成的图像交给卷积神经网络模型,用于肺炎covid - 19疾病的诊断和分类,最后我们使用Grad- cam和Vanilla Gradient和Smooth Grad技术对设计的模型进行验证。根据结果,我们提出的使用评估指标的方法能够在区分covid - 19疾病和肺炎方面达到最高的准确性。
{"title":"Application of Explainable Convolutional Neural Networks on the Differential Diagnosis of Covid_19 and Pneumonia using Chest Radiograph","authors":"Fereshteh Zandi, H. Ebrahimpour-Komleh, Hassan Homayoun","doi":"10.1109/IPRIA59240.2023.10147169","DOIUrl":"https://doi.org/10.1109/IPRIA59240.2023.10147169","url":null,"abstract":"The Covid_19 disease is one of the deadliest inflammatories and chronic and acute diseases of the human respiratory system, which is the result of the inhibition of a virus called corona in the respiratory organs since the spread of this virus is rapid and has affected many people in the world. a specialist needs to be carefully evaluated to diagnose the disease based on X-ray images because the number of patients with Covid_19 exceeds the capacity of hospitals and taking care of a large number of people is tedious work that can reduce the accuracy of the doctor in diagnosing the disease. In addition, in such cases, the absence of a specialist doctor can lead to misdiagnosis and incorrect prescribing. In this article, we intend to provide an approach to accelerate the diagnosis process and reduce the workload of specialists automatically, which in addition to helping physicians in hospitals that do not have a specialist physician, also allows patients to be diagnosed and treated. we use pre-trained UNet to extract the lung balloons, which eliminates the extra noise and parts in the X-ray image and then we give the generated images to a convolutional neural network model designed to diagnose and classify Covid_19 disease from Pneumonia, and finally, we use Grad-CAM and Vanilla Gradient and Smooth Grad techniques to validate the designed model. according to the results, our proposed approach using evaluation metrics was able to achieve the highest degree of accuracy in distinguishing Covid_19 disease from Pneumonia.","PeriodicalId":109390,"journal":{"name":"2023 6th International Conference on Pattern Recognition and Image Analysis (IPRIA)","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-02-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126414468","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Empirical Mode Decomposition Based Morphological Profile For Hyperspectral Image Classification 基于经验模态分解的形态轮廓高光谱图像分类
Pub Date : 2023-02-14 DOI: 10.1109/IPRIA59240.2023.10147181
Kosar Amiri, M. Imani, H. Ghassemian
The empirical mode decomposition (EMD) based morphological profile (MP), called as EMDMP, is proposed for hyperspectral image classification in this work. The EMD algorithm can well decompose the nonlinear spectral feature vector to intrinsic components and the residual term. To extract the main spatial characteristics and shape structures, the closing operators are applied to the intrinsic components. In contrast, to extract details and more abstract contextual features, the opening operators are applied to the residual component. Finally, a multi-resolution morphological profile is provided with concatenation of the intrinsic components-based closing profile and residual component based opening profile. EMDMP achieves 96.54% overall accuracy compared to 95.15% obtained by convolutional neural network (CNN) on Indian dataset with 10% training samples. In University of Pavia with 1% training samples, EMDMP results in 97.66% overall accuracy compared to 95.90% obtained by CNN.
本文提出了基于经验模态分解(EMD)的形态轮廓(MP)方法用于高光谱图像分类。EMD算法能很好地将非线性谱特征向量分解为固有分量和残差项。为了提取主要空间特征和形状结构,对固有分量应用闭合算子。相反,为了提取细节和更抽象的上下文特征,将开放操作符应用于残差分量。最后,提供了基于固有分量的闭合轮廓和基于残差分量的打开轮廓的串联的多分辨率形态轮廓。EMDMP的总体准确率为96.54%,而卷积神经网络(CNN)在印度数据集上使用10%的训练样本获得的总体准确率为95.15%。在帕维亚大学,使用1%的训练样本,EMDMP的总体准确率为97.66%,而CNN的总体准确率为95.90%。
{"title":"Empirical Mode Decomposition Based Morphological Profile For Hyperspectral Image Classification","authors":"Kosar Amiri, M. Imani, H. Ghassemian","doi":"10.1109/IPRIA59240.2023.10147181","DOIUrl":"https://doi.org/10.1109/IPRIA59240.2023.10147181","url":null,"abstract":"The empirical mode decomposition (EMD) based morphological profile (MP), called as EMDMP, is proposed for hyperspectral image classification in this work. The EMD algorithm can well decompose the nonlinear spectral feature vector to intrinsic components and the residual term. To extract the main spatial characteristics and shape structures, the closing operators are applied to the intrinsic components. In contrast, to extract details and more abstract contextual features, the opening operators are applied to the residual component. Finally, a multi-resolution morphological profile is provided with concatenation of the intrinsic components-based closing profile and residual component based opening profile. EMDMP achieves 96.54% overall accuracy compared to 95.15% obtained by convolutional neural network (CNN) on Indian dataset with 10% training samples. In University of Pavia with 1% training samples, EMDMP results in 97.66% overall accuracy compared to 95.90% obtained by CNN.","PeriodicalId":109390,"journal":{"name":"2023 6th International Conference on Pattern Recognition and Image Analysis (IPRIA)","volume":"42 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-02-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121951111","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Quality Assessment of Screen Content Videos 屏幕内容视频的质量评估
Pub Date : 2023-02-14 DOI: 10.1109/IPRIA59240.2023.10147176
Hossein Motamednia, Pooryaa Cheraaqee, Azadeh Mansouri, Ahmad Mahmoudi-Aznaveh
Perceptual quality assessment has always been challenging due to the difficulty in modeling the no-linear human visual system. With the diversity in the contents of multimedia signals, the conventional methods for traditional media seems no longer satisfying. One of these emerging media, is the screen content images/videos (SCINs), Containing texts and computer generated graphics, SCVs cannot be sufficiently expressed with features designed for natural sceneries. Therefore, new researches tried to devise objective quality assessment metrics, specificly for screen contents. Recently, a dataset was proposed for quality assessment of screen content videos. Since screen contents are full of structures that spread in cardinal directions, we were motivated to employ the horizontal and vertical subbands of the wavelet transform to characterize these types of visual contents. The features were incorporated in a full-reference method that showed promising results on the publicly available dataset for SCV quality assessment. The method can bo accessed via: https://github.com/motamedNia/QASCV.
由于人类视觉系统的非线性建模困难,感知质量评估一直是一个具有挑战性的问题。随着多媒体信号内容的多样化,传统媒体的传统方式已经不能满足人们的需求。其中一种新兴媒体是屏幕内容图像/视频(SCINs), SCINs包含文本和计算机生成的图形,无法充分表达为自然风景设计的特征。因此,新的研究试图设计客观的质量评估指标,特别是针对屏幕内容。最近,提出了一个用于屏幕内容视频质量评估的数据集。由于屏幕内容充满了在基本方向上传播的结构,我们被激励使用小波变换的水平和垂直子带来表征这些类型的视觉内容。这些特征被纳入到一个全参考方法中,该方法在公开可用的SCV质量评估数据集上显示出有希望的结果。该方法可以通过:https://github.com/motamedNia/QASCV访问。
{"title":"Quality Assessment of Screen Content Videos","authors":"Hossein Motamednia, Pooryaa Cheraaqee, Azadeh Mansouri, Ahmad Mahmoudi-Aznaveh","doi":"10.1109/IPRIA59240.2023.10147176","DOIUrl":"https://doi.org/10.1109/IPRIA59240.2023.10147176","url":null,"abstract":"Perceptual quality assessment has always been challenging due to the difficulty in modeling the no-linear human visual system. With the diversity in the contents of multimedia signals, the conventional methods for traditional media seems no longer satisfying. One of these emerging media, is the screen content images/videos (SCINs), Containing texts and computer generated graphics, SCVs cannot be sufficiently expressed with features designed for natural sceneries. Therefore, new researches tried to devise objective quality assessment metrics, specificly for screen contents. Recently, a dataset was proposed for quality assessment of screen content videos. Since screen contents are full of structures that spread in cardinal directions, we were motivated to employ the horizontal and vertical subbands of the wavelet transform to characterize these types of visual contents. The features were incorporated in a full-reference method that showed promising results on the publicly available dataset for SCV quality assessment. The method can bo accessed via: https://github.com/motamedNia/QASCV.","PeriodicalId":109390,"journal":{"name":"2023 6th International Conference on Pattern Recognition and Image Analysis (IPRIA)","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-02-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114339680","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Few-shot Learning with Prompting Methods 用提示法进行短时间学习
Pub Date : 2023-02-14 DOI: 10.1109/IPRIA59240.2023.10147172
Morteza Bahrami, Muharram Mansoorizadeh, Hassan Khotanlou
Today, in natural language processing, labeled data is important, however, getting adequate amount of data is a challenging step. There are many tasks for which it is difficult to obtain the required training data. For example, in machine translation, we need to prepare a lot of data in the target language, so that the work performance is acceptable. We may not be able to collect useful data in the target language. Hence, we need to use few-shot learning. Recently, a method called prompting has evolved, in which text inputs are converted into text with a new structure using a certain format, which has a blank space. Given the prompted text, a pre-trained language model replaces the space with the best word. Prompting can help us in the field of few-shot learning; even in cases where there is no data, i.e. zero-shot learning. Recent works use large language models such as GPT-2 and GPT-3, with the prompting method, performed tasks such as machine translation. These efforts do not use any labeled training data. But these types of models with a massive number of parameters require powerful hardware. Pattern-Exploiting Training (PET) and iterative Pattern-Exploiting Training (iPET) were introduced, which perform few-shot learning using prompting and smaller pre-trained language models such as Bert and Roberta. For example, for the Yahoo text classification dataset, using iPET and Roberta and ten labeled datasets, 70% accuracy has been reached. This paper reviews research works in few-shot learning with a new paradigm in natural language processing, which we dub prompt-based learning or in short, prompting.
今天,在自然语言处理中,标记数据很重要,然而,获得足够数量的数据是一个具有挑战性的步骤。有许多任务很难获得所需的训练数据。例如,在机器翻译中,我们需要用目标语言准备大量的数据,以便工作表现可以接受。我们可能无法以目标语言收集有用的数据。因此,我们需要使用几次学习。最近,出现了一种叫做提示的方法,它将文本输入转换为使用特定格式的具有新结构的文本,该格式具有空白。给定提示文本,预训练的语言模型将用最佳单词替换空格。提示可以帮助我们在场上少射学习;即使在没有数据的情况下,也就是零次学习。最近的作品使用了GPT-2和GPT-3等大型语言模型,通过提示的方式,执行机器翻译等任务。这些工作不使用任何标记的训练数据。但这些具有大量参数的模型需要强大的硬件。介绍了模式挖掘训练(PET)和迭代模式挖掘训练(iPET),它们使用提示和较小的预训练语言模型(如Bert和Roberta)进行少量学习。例如,对于Yahoo文本分类数据集,使用iPET和Roberta和10个标记数据集,准确率达到70%。本文以自然语言处理中的一种新范式,即基于提示的学习(prompt-based learning,简称prompt),回顾了在短时学习方面的研究工作。
{"title":"Few-shot Learning with Prompting Methods","authors":"Morteza Bahrami, Muharram Mansoorizadeh, Hassan Khotanlou","doi":"10.1109/IPRIA59240.2023.10147172","DOIUrl":"https://doi.org/10.1109/IPRIA59240.2023.10147172","url":null,"abstract":"Today, in natural language processing, labeled data is important, however, getting adequate amount of data is a challenging step. There are many tasks for which it is difficult to obtain the required training data. For example, in machine translation, we need to prepare a lot of data in the target language, so that the work performance is acceptable. We may not be able to collect useful data in the target language. Hence, we need to use few-shot learning. Recently, a method called prompting has evolved, in which text inputs are converted into text with a new structure using a certain format, which has a blank space. Given the prompted text, a pre-trained language model replaces the space with the best word. Prompting can help us in the field of few-shot learning; even in cases where there is no data, i.e. zero-shot learning. Recent works use large language models such as GPT-2 and GPT-3, with the prompting method, performed tasks such as machine translation. These efforts do not use any labeled training data. But these types of models with a massive number of parameters require powerful hardware. Pattern-Exploiting Training (PET) and iterative Pattern-Exploiting Training (iPET) were introduced, which perform few-shot learning using prompting and smaller pre-trained language models such as Bert and Roberta. For example, for the Yahoo text classification dataset, using iPET and Roberta and ten labeled datasets, 70% accuracy has been reached. This paper reviews research works in few-shot learning with a new paradigm in natural language processing, which we dub prompt-based learning or in short, prompting.","PeriodicalId":109390,"journal":{"name":"2023 6th International Conference on Pattern Recognition and Image Analysis (IPRIA)","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-02-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123735309","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Converge intra-class and Diverge inter-class features for CNN-based Event Detection in football videos 基于cnn的足球视频事件检测的类内收敛和类间发散特征
Pub Date : 2023-02-14 DOI: 10.1109/IPRIA59240.2023.10147187
Amirhosein Zanganeh, E. Sharifi, M. Jampour
Football event detection in videos is very challenging, but challenges on the Penalty and the Free-kick, which have common visual elements, are severe and critical. The existence of common elements between two events causes the extraction of common and ineffective features in recognizing these two events. As a result, the error of recognizing and separating these two events is more than other events. In this paper, we present a new method for filtering the input data to converge the intra-class features and diverge the inter-class features to increase the classification accuracy. For this purpose, using the IAUFD Dataset, we have evaluated images for the Penalty and the Free-kick classes with the criterion of structural similarity. Based on the results, inappropriate images have been ignored according to the average value and standard deviation of each class of data. This filtration leads to ignore of ineffective and common features in the learning process. The results of the proposed method indicate an improvement in the accuracy of distinguishing between two Penalty and Free-kick events using a deep neural network and filtered training images compared to the deep neural network using all training images.
视频中的足球事件检测具有很大的挑战性,其中对具有共同视觉元素的点球和任意球的检测更为严峻和关键。由于两个事件之间存在共同要素,导致在识别这两个事件时对共同特征和无效特征的提取。因此,识别和分离这两个事件的误差大于其他事件。在本文中,我们提出了一种新的过滤输入数据的方法,以收敛类内特征和发散类间特征,以提高分类精度。为此,使用IAUFD数据集,我们用结构相似性标准评估了点球和任意球类的图像。根据结果,根据每一类数据的平均值和标准差,忽略不合适的图像。这种过滤导致了对学习过程中无效和共同特征的忽视。结果表明,与使用所有训练图像的深度神经网络相比,使用深度神经网络和过滤的训练图像区分两个点球和任意球事件的准确性有所提高。
{"title":"Converge intra-class and Diverge inter-class features for CNN-based Event Detection in football videos","authors":"Amirhosein Zanganeh, E. Sharifi, M. Jampour","doi":"10.1109/IPRIA59240.2023.10147187","DOIUrl":"https://doi.org/10.1109/IPRIA59240.2023.10147187","url":null,"abstract":"Football event detection in videos is very challenging, but challenges on the Penalty and the Free-kick, which have common visual elements, are severe and critical. The existence of common elements between two events causes the extraction of common and ineffective features in recognizing these two events. As a result, the error of recognizing and separating these two events is more than other events. In this paper, we present a new method for filtering the input data to converge the intra-class features and diverge the inter-class features to increase the classification accuracy. For this purpose, using the IAUFD Dataset, we have evaluated images for the Penalty and the Free-kick classes with the criterion of structural similarity. Based on the results, inappropriate images have been ignored according to the average value and standard deviation of each class of data. This filtration leads to ignore of ineffective and common features in the learning process. The results of the proposed method indicate an improvement in the accuracy of distinguishing between two Penalty and Free-kick events using a deep neural network and filtered training images compared to the deep neural network using all training images.","PeriodicalId":109390,"journal":{"name":"2023 6th International Conference on Pattern Recognition and Image Analysis (IPRIA)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-02-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130259376","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Deit Model for Iranian Traffic Sign Recognition in Advanced Driver Assistance Systems 先进驾驶辅助系统中伊朗交通标志识别的Deit模型
Pub Date : 2023-02-14 DOI: 10.1109/IPRIA59240.2023.10147174
Marjan. Shahchera, Hossein Ebrahimpour-komleh
Due to the important relationship between the impact of accurate detection of traffic signs in self-driving cars and driver assistance during car movement, it is very challenging and necessary to create a high-accuracy system for interpretation and immediate decision-making. In this research, by applying the new vision transformer Deit approach, a system is designed that can recognize Iranian traffic signs. We trained our model with two collections of traffic sign images (GTSRB and PTSD) that reached higher accuracy levels of 99.5% and 98.8%, respectively, in optimal conditions.
由于自动驾驶汽车准确检测交通标志的影响与汽车运动过程中的驾驶员辅助之间的重要关系,因此创建一个用于解释和即时决策的高精度系统是非常具有挑战性和必要性的。在本研究中,采用新的视觉转换Deit方法,设计了一个能够识别伊朗交通标志的系统。我们使用两个交通标志图像集(GTSRB和PTSD)来训练我们的模型,在最佳条件下,准确率分别达到99.5%和98.8%。
{"title":"Deit Model for Iranian Traffic Sign Recognition in Advanced Driver Assistance Systems","authors":"Marjan. Shahchera, Hossein Ebrahimpour-komleh","doi":"10.1109/IPRIA59240.2023.10147174","DOIUrl":"https://doi.org/10.1109/IPRIA59240.2023.10147174","url":null,"abstract":"Due to the important relationship between the impact of accurate detection of traffic signs in self-driving cars and driver assistance during car movement, it is very challenging and necessary to create a high-accuracy system for interpretation and immediate decision-making. In this research, by applying the new vision transformer Deit approach, a system is designed that can recognize Iranian traffic signs. We trained our model with two collections of traffic sign images (GTSRB and PTSD) that reached higher accuracy levels of 99.5% and 98.8%, respectively, in optimal conditions.","PeriodicalId":109390,"journal":{"name":"2023 6th International Conference on Pattern Recognition and Image Analysis (IPRIA)","volume":"47 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-02-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124983849","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Contextual Information Classification of Remotely Sensed Images 遥感图像上下文信息分类
Pub Date : 2023-02-14 DOI: 10.1109/IPRIA59240.2023.10147193
Alirza Dori, H. Ghassemian, M. Imani
This work proposes a multidisciplinary contextual information extraction and decision fusion approach for increasing the classification accuracy. It improves the image classification with integrating the results of various classifiers. The proposed method is implemented in three-steps: 1) contextual feature extraction using four different feature extractors methods: a) Gray Level Cooccurrence Matrix, b) Gabor filters, c) Laplacian Gaussian filters and d) Gaussian Derivatives Functions; 2) classification of contextual features using four different classification rules (ML, Tree, KNN and SVM) by using only 2% of data for training the classifiers; and 3) finally, decision fusion using six decision fusion rules. The experimental results on real remotely sensed images have been presented.
本文提出了一种多学科的上下文信息提取和决策融合方法,以提高分类精度。它综合了各种分类器的分类结果,提高了图像的分类效率。该方法分三步实现:1)使用四种不同的特征提取方法进行上下文特征提取:a)灰度共生矩阵,b) Gabor滤波器,c)拉普拉斯高斯滤波器和d)高斯导数函数;2)使用四种不同的分类规则(ML、Tree、KNN和SVM)对上下文特征进行分类,仅使用2%的数据进行分类器训练;最后,利用6条决策融合规则进行决策融合。给出了在真实遥感图像上的实验结果。
{"title":"Contextual Information Classification of Remotely Sensed Images","authors":"Alirza Dori, H. Ghassemian, M. Imani","doi":"10.1109/IPRIA59240.2023.10147193","DOIUrl":"https://doi.org/10.1109/IPRIA59240.2023.10147193","url":null,"abstract":"This work proposes a multidisciplinary contextual information extraction and decision fusion approach for increasing the classification accuracy. It improves the image classification with integrating the results of various classifiers. The proposed method is implemented in three-steps: 1) contextual feature extraction using four different feature extractors methods: a) Gray Level Cooccurrence Matrix, b) Gabor filters, c) Laplacian Gaussian filters and d) Gaussian Derivatives Functions; 2) classification of contextual features using four different classification rules (ML, Tree, KNN and SVM) by using only 2% of data for training the classifiers; and 3) finally, decision fusion using six decision fusion rules. The experimental results on real remotely sensed images have been presented.","PeriodicalId":109390,"journal":{"name":"2023 6th International Conference on Pattern Recognition and Image Analysis (IPRIA)","volume":"371 1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-02-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132776062","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
3D Image Annotation using Deep Learning and View-based Image Features 使用深度学习和基于视图的图像特征的3D图像标注
Pub Date : 2023-02-14 DOI: 10.1109/IPRIA59240.2023.10147190
Mohammadiman Hosseinnia, A. Behrad
The act of assigning word labels to images using machine learning algorithms is called automatic image annotation. Automatic annotation of image is used in various applications like media, medical, industrial and archaeological fields. Several methods have been proposed for automatic annotation of images, but most of them are focused on 2D images. In this article, we propose a new approach for 3D image annotation using deep learning and view-based image features. The most challenging issue in the automatic annotation of 3D images is to extract suitable features for image representation. 3D images are generally presented in the form of polygon meshes that are not suitable for deep learning. To counter the problem, we represent 3D images as several view-based images that are captured from different views. This process converts a 3D image into a multi-channel 2D image that can be classified using image-based deep classification networks. We utilized various classification networks for 3D image annotation, and the results showed the F1 score of 0.97 for the best architecture.
使用机器学习算法为图像分配单词标签的行为称为自动图像注释。图像的自动标注用于媒体、医疗、工业和考古等领域的各种应用。目前已经提出了几种图像自动标注的方法,但大多数都是针对二维图像的。在本文中,我们提出了一种使用深度学习和基于视图的图像特征进行3D图像标注的新方法。在三维图像的自动标注中,最具挑战性的问题是如何提取出适合图像表示的特征。3D图像通常以多边形网格的形式呈现,不适合深度学习。为了解决这个问题,我们将3D图像表示为从不同视图捕获的几个基于视图的图像。该过程将3D图像转换为可使用基于图像的深度分类网络进行分类的多通道2D图像。我们利用各种分类网络对三维图像进行标注,结果表明,最佳架构的F1得分为0.97。
{"title":"3D Image Annotation using Deep Learning and View-based Image Features","authors":"Mohammadiman Hosseinnia, A. Behrad","doi":"10.1109/IPRIA59240.2023.10147190","DOIUrl":"https://doi.org/10.1109/IPRIA59240.2023.10147190","url":null,"abstract":"The act of assigning word labels to images using machine learning algorithms is called automatic image annotation. Automatic annotation of image is used in various applications like media, medical, industrial and archaeological fields. Several methods have been proposed for automatic annotation of images, but most of them are focused on 2D images. In this article, we propose a new approach for 3D image annotation using deep learning and view-based image features. The most challenging issue in the automatic annotation of 3D images is to extract suitable features for image representation. 3D images are generally presented in the form of polygon meshes that are not suitable for deep learning. To counter the problem, we represent 3D images as several view-based images that are captured from different views. This process converts a 3D image into a multi-channel 2D image that can be classified using image-based deep classification networks. We utilized various classification networks for 3D image annotation, and the results showed the F1 score of 0.97 for the best architecture.","PeriodicalId":109390,"journal":{"name":"2023 6th International Conference on Pattern Recognition and Image Analysis (IPRIA)","volume":"13 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-02-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115139565","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Bilingual COVID-19 Fake News Detection Based on LDA Topic Modeling and BERT Transformer 基于LDA主题建模和BERT转换器的双语COVID-19假新闻检测
Pub Date : 2023-02-14 DOI: 10.1109/IPRIA59240.2023.10147179
Pouria Omrani, Zahra Ebrahimian, Ramin Toosi, M. Akhaee
The spread of fake news has become more prevalent given the popularity of social media and the various news that circulates on it. As a result, it is crucial to discern between real and fake news. During the COVID-19 pandemic, there have been numerous tweets, posts, and news about this illness in social media and electronic media worldwide. This research presents a bilingual model combining Latent Dirichlet Allocation (LDA) topic modeling and the BERT transformer to detect COVID-19 fake news in both Persian and English. First, the dataset is prepared in Persian and English, and then the proposed method is used to detect COVID-19 fake news on the prepared dataset. Finally, the proposed model is evaluated using various metrics such as accuracy, precision, recall, and the f1-score. As a result of this approach, we achieve 92.18% accuracy, which shows that adding topic information to the pre-trained contextual representations given by the BERT network, significantly improves the solving of instances that are domain-specific. Also, the results show that our proposed approach outperforms previous state-of-the-art methods.
鉴于社交媒体的普及以及在社交媒体上传播的各种新闻,假新闻的传播变得更加普遍。因此,辨别真假新闻至关重要。在2019冠状病毒病大流行期间,全球社交媒体和电子媒体上出现了许多关于这种疾病的推文、帖子和新闻。本研究提出了一种结合潜在狄利克雷分配(Latent Dirichlet Allocation, LDA)主题建模和BERT转换器的双语模型,用于检测波斯语和英语的COVID-19假新闻。首先用波斯语和英语准备数据集,然后使用本文提出的方法在准备好的数据集上检测COVID-19假新闻。最后,使用准确度、精密度、召回率和f1分数等各种指标对所提出的模型进行评估。通过这种方法,我们达到了92.18%的准确率,这表明将主题信息添加到BERT网络给出的预训练上下文表示中,显著提高了特定领域实例的求解效率。此外,结果表明,我们提出的方法优于以前的最先进的方法。
{"title":"Bilingual COVID-19 Fake News Detection Based on LDA Topic Modeling and BERT Transformer","authors":"Pouria Omrani, Zahra Ebrahimian, Ramin Toosi, M. Akhaee","doi":"10.1109/IPRIA59240.2023.10147179","DOIUrl":"https://doi.org/10.1109/IPRIA59240.2023.10147179","url":null,"abstract":"The spread of fake news has become more prevalent given the popularity of social media and the various news that circulates on it. As a result, it is crucial to discern between real and fake news. During the COVID-19 pandemic, there have been numerous tweets, posts, and news about this illness in social media and electronic media worldwide. This research presents a bilingual model combining Latent Dirichlet Allocation (LDA) topic modeling and the BERT transformer to detect COVID-19 fake news in both Persian and English. First, the dataset is prepared in Persian and English, and then the proposed method is used to detect COVID-19 fake news on the prepared dataset. Finally, the proposed model is evaluated using various metrics such as accuracy, precision, recall, and the f1-score. As a result of this approach, we achieve 92.18% accuracy, which shows that adding topic information to the pre-trained contextual representations given by the BERT network, significantly improves the solving of instances that are domain-specific. Also, the results show that our proposed approach outperforms previous state-of-the-art methods.","PeriodicalId":109390,"journal":{"name":"2023 6th International Conference on Pattern Recognition and Image Analysis (IPRIA)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-02-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128522935","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
2023 6th International Conference on Pattern Recognition and Image Analysis (IPRIA)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1