首页 > 最新文献

International Conference on Pattern Recognition Applications and Methods最新文献

英文 中文
Analyzing the Linear and Nonlinear Transformations of AlexNet to Gain Insight into Its Performance 分析AlexNet的线性和非线性变换以深入了解其性能
Pub Date : 2019-02-19 DOI: 10.5220/0007582408600865
Jyoti Nigam, Srishti Barahpuriya, Renu M. Rameshan
AlexNet, one of the earliest and successful deep learning networks, has given great performance in image classification task. There are some fundamental properties for good classification such as: the network preserves the important information of the input data; the network is able to see differently, points from different classes. In this work we experimentally verify that these core properties are followed by the AlexNet architecture. We analyze the effect of linear and nonlinear transformations on input data across the layers. The convolution filters are modeled as linear transformations. The verified results motivate to draw conclusions on the desirable properties of transformation matrix that aid in better classification.
AlexNet是最早和成功的深度学习网络之一,在图像分类任务中表现出色。好的分类有一些基本的性质,如:网络保留了输入数据的重要信息;网络能够看到不同的,来自不同阶层的点。在这项工作中,我们通过实验验证了AlexNet架构遵循这些核心属性。我们分析了线性和非线性变换对各层输入数据的影响。卷积滤波器被建模为线性变换。验证的结果促使我们对变换矩阵的理想性质得出结论,从而有助于更好地分类。
{"title":"Analyzing the Linear and Nonlinear Transformations of AlexNet to Gain Insight into Its Performance","authors":"Jyoti Nigam, Srishti Barahpuriya, Renu M. Rameshan","doi":"10.5220/0007582408600865","DOIUrl":"https://doi.org/10.5220/0007582408600865","url":null,"abstract":"AlexNet, one of the earliest and successful deep learning networks, has given great performance in image classification task. There are some fundamental properties for good classification such as: the network preserves the important information of the input data; the network is able to see differently, points from different classes. In this work we experimentally verify that these core properties are followed by the AlexNet architecture. We analyze the effect of linear and nonlinear transformations on input data across the layers. The convolution filters are modeled as linear transformations. The verified results motivate to draw conclusions on the desirable properties of transformation matrix that aid in better classification.","PeriodicalId":410036,"journal":{"name":"International Conference on Pattern Recognition Applications and Methods","volume":"263 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-02-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133298419","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Semantic Segmentation of Non-linear Multimodal Images for Disease Grading of Inflammatory Bowel Disease: A SegNet-based Application 炎性肠疾病分级非线性多模态图像的语义分割:基于分段网的应用
Pub Date : 2019-02-19 DOI: 10.5220/0007314003960405
Pranita Pradhan, T. Meyer, M. Vieth, A. Stallmach, M. Waldner, M. Schmitt, J. Popp, T. Bocklitz
Non-linear multimodal imaging, the combination of coherent anti-stokes Raman scattering (CARS), two-photon excited fluorescence (TPEF) and second harmonic generation (SHG), has shown its potential to assist the diagnosis of different inflammatory bowel diseases (IBDs). This label-free imaging technique can support the ‘gold-standard’ techniques such as colonoscopy and histopathology to ensure an IBD diagnosis in clinical environment. Moreover, non-linear multimodal imaging can measure biomolecular changes in different tissue regions such as crypt and mucosa region, which serve as a predictive marker for IBD severity. To achieve a real-time assessment of IBD severity, an automatic segmentation of the crypt and mucosa regions is needed. In this paper, we semantically segment the crypt and mucosa region using a deep neural network. We utilized the SegNet architecture (Badrinarayanan et al., 2015) and compared its results with a classical machine learning approach. Our trained SegNet model achieved an overall F1 score of 0.75. This model outperformed the classical machine learning approach for the segmentation of the crypt and mucosa region in our study.
非线性多模态成像,结合相干抗斯托克斯拉曼散射(CARS),双光子激发荧光(TPEF)和二次谐波产生(SHG),已经显示出其协助诊断不同炎症性肠病(IBDs)的潜力。这种无标签成像技术可以支持结肠镜检查和组织病理学等“金标准”技术,以确保在临床环境中对IBD进行诊断。此外,非线性多模态成像可以测量不同组织区域(如隐窝和粘膜区域)的生物分子变化,作为IBD严重程度的预测指标。为了实现对IBD严重程度的实时评估,需要对隐窝和粘膜区域进行自动分割。在本文中,我们使用深度神经网络对隐窝和粘膜区域进行语义分割。我们使用了SegNet架构(Badrinarayanan et al., 2015),并将其结果与经典的机器学习方法进行了比较。我们训练的SegNet模型获得了0.75的F1总分。在我们的研究中,该模型在隐窝和粘膜区域的分割方面优于经典的机器学习方法。
{"title":"Semantic Segmentation of Non-linear Multimodal Images for Disease Grading of Inflammatory Bowel Disease: A SegNet-based Application","authors":"Pranita Pradhan, T. Meyer, M. Vieth, A. Stallmach, M. Waldner, M. Schmitt, J. Popp, T. Bocklitz","doi":"10.5220/0007314003960405","DOIUrl":"https://doi.org/10.5220/0007314003960405","url":null,"abstract":"Non-linear multimodal imaging, the combination of coherent anti-stokes Raman scattering (CARS), two-photon excited fluorescence (TPEF) and second harmonic generation (SHG), has shown its potential to assist the diagnosis of different inflammatory bowel diseases (IBDs). This label-free imaging technique can support the ‘gold-standard’ techniques such as colonoscopy and histopathology to ensure an IBD diagnosis in clinical environment. Moreover, non-linear multimodal imaging can measure biomolecular changes in different tissue regions such as crypt and mucosa region, which serve as a predictive marker for IBD severity. To achieve a real-time assessment of IBD severity, an automatic segmentation of the crypt and mucosa regions is needed. In this paper, we semantically segment the crypt and mucosa region using a deep neural network. We utilized the SegNet architecture (Badrinarayanan et al., 2015) and compared its results with a classical machine learning approach. Our trained SegNet model achieved an overall F1 score of 0.75. This model outperformed the classical machine learning approach for the segmentation of the crypt and mucosa region in our study.","PeriodicalId":410036,"journal":{"name":"International Conference on Pattern Recognition Applications and Methods","volume":"487 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-02-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115880337","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
A Robust Page Frame Detection Method for Complex Historical Document Images 复杂历史文档图像的鲁棒页面帧检测方法
Pub Date : 2019-02-19 DOI: 10.5220/0007382405560564
M. Reza, Md. Ajraf Rakib, S. S. Bukhari, A. Dengel
Document layout analysis is the most important part of converting scanned page images into search-able full text. An intensive amount of research is going on in the field of structured and semi-structured documents (journal articles, books, magazines, invoices) but not much in historical documents. Historical document digitization is a more challenging task than regular structured documents due to poor image quality, damaged characters, big amount of textual and non-textual noise. In the scientific community, the extraneous symbols from the neighboring page are considered as textual noise, while the appearances of black borders, speckles, ruler, different types of image etc. along the border of the documents are considered as non-textual noise. Existing historical document analysis method cannot handle all of this noise which is a very strong reason of getting undesired texts as a result from the output of Optical Character Recognition (OCR) that needs to be removed afterward with a lot of extra afford. This paper presents a new perspective especially for the historical document image cleanup by detecting the page frame of the document. The goal of this method is to find actual contents area of the document and ignore noises along the page border. We use morphological transforms, the line segment detector, and geometric matching algorithm to find an ideal page frame of the document. After the implementation of page frame method, we also evaluate our approach over 16th-19th century printed historical documents. We have noticed in the result that OCR performance for the historical documents increased by 4.49% after applying our page frame detection method. In addition, we are able to increase the OCR accuracy around 6.69% for contemporary documents too.
文档布局分析是将扫描页面图像转换为可搜索全文的最重要部分。在结构化和半结构化文件(期刊文章、书籍、杂志、发票)领域正在进行大量的研究,但在历史文件方面的研究不多。历史文献由于图像质量差、字符损坏、文本和非文本噪声大等问题,比常规结构化文献数字化更具挑战性。在科学界,邻页的无关符号被认为是文本噪声,而文档边缘出现的黑色边框、斑点、标尺、不同类型的图像等被认为是非文本噪声。现有的历史文档分析方法不能处理所有这些噪声,这是光学字符识别(OCR)输出中得到不需要的文本的一个非常重要的原因,这些文本需要在事后花费大量额外的费用来去除。本文提出了一种新的视角,特别是通过检测文档的页框来清理历史文档图像。该方法的目标是找到文档的实际内容区域,并忽略沿页面边界的噪声。我们使用形态变换、线段检测器和几何匹配算法来找到文档的理想页面框架。在页面框架方法实施后,我们还对16 -19世纪的印刷历史文献进行了评估。我们注意到,应用我们的页面框架检测方法后,历史文档的OCR性能提高了4.49%。此外,我们也能够将现代文档的OCR准确率提高到6.69%左右。
{"title":"A Robust Page Frame Detection Method for Complex Historical Document Images","authors":"M. Reza, Md. Ajraf Rakib, S. S. Bukhari, A. Dengel","doi":"10.5220/0007382405560564","DOIUrl":"https://doi.org/10.5220/0007382405560564","url":null,"abstract":"Document layout analysis is the most important part of converting scanned page images into search-able full text. An intensive amount of research is going on in the field of structured and semi-structured documents (journal articles, books, magazines, invoices) but not much in historical documents. Historical document digitization is a more challenging task than regular structured documents due to poor image quality, damaged characters, big amount of textual and non-textual noise. In the scientific community, the extraneous symbols from the neighboring page are considered as textual noise, while the appearances of black borders, speckles, ruler, different types of image etc. along the border of the documents are considered as non-textual noise. Existing historical document analysis method cannot handle all of this noise which is a very strong reason of getting undesired texts as a result from the output of Optical Character Recognition (OCR) that needs to be removed afterward with a lot of extra afford. This paper presents a new perspective especially for the historical document image cleanup by detecting the page frame of the document. The goal of this method is to find actual contents area of the document and ignore noises along the page border. We use morphological transforms, the line segment detector, and geometric matching algorithm to find an ideal page frame of the document. After the implementation of page frame method, we also evaluate our approach over 16th-19th century printed historical documents. We have noticed in the result that OCR performance for the historical documents increased by 4.49% after applying our page frame detection method. In addition, we are able to increase the OCR accuracy around 6.69% for contemporary documents too.","PeriodicalId":410036,"journal":{"name":"International Conference on Pattern Recognition Applications and Methods","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-02-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123414361","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
A Study of Various Text Augmentation Techniques for Relation Classification in Free Text 自由文本中关系分类的各种文本增强技术研究
Pub Date : 2019-02-19 DOI: 10.5220/0007311003600367
Praveen Kumar Badimala Giridhara, Chinmaya Mishra, Reddy Kumar Modam Venkataramana, S. S. Bukhari, A. Dengel
Data augmentation techniques have been widely used in visual recognition tasks as it is easy to generate new data by simple and straight forward image transformations. However, when it comes to text data augmentations, it is difficult to find appropriate transformation techniques which also preserve the contextual and grammatical structure of language texts. In this paper, we explore various text data augmentation techniques in text space and word embedding space. We study the effect of various augmented datasets on the efficiency of different deep learning models for relation classification in text.
数据增强技术在视觉识别任务中得到了广泛的应用,因为它可以通过简单直接的图像变换来生成新的数据。然而,当涉及到文本数据增强时,很难找到适当的转换技术,同时保留语言文本的上下文和语法结构。在本文中,我们探索了文本空间和词嵌入空间中的各种文本数据增强技术。我们研究了各种增强数据集对文本中关系分类的不同深度学习模型效率的影响。
{"title":"A Study of Various Text Augmentation Techniques for Relation Classification in Free Text","authors":"Praveen Kumar Badimala Giridhara, Chinmaya Mishra, Reddy Kumar Modam Venkataramana, S. S. Bukhari, A. Dengel","doi":"10.5220/0007311003600367","DOIUrl":"https://doi.org/10.5220/0007311003600367","url":null,"abstract":"Data augmentation techniques have been widely used in visual recognition tasks as it is easy to generate new data by simple and straight forward image transformations. However, when it comes to text data augmentations, it is difficult to find appropriate transformation techniques which also preserve the contextual and grammatical structure of language texts. In this paper, we explore various text data augmentation techniques in text space and word embedding space. We study the effect of various augmented datasets on the efficiency of different deep learning models for relation classification in text.","PeriodicalId":410036,"journal":{"name":"International Conference on Pattern Recognition Applications and Methods","volume":"66 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-02-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131018825","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 31
Document Image Dewarping using Deep Learning 使用深度学习的文档图像去翘曲
Pub Date : 2019-02-19 DOI: 10.5220/0007368405240531
Vijaya Kumar Bajjer Ramanna, S. S. Bukhari, A. Dengel
The distorted images have been a major problem for Optical Character Recognition (OCR). In order to perform OCR on distorted images, dewarping has become a principal preprocessing step. This paper presents a new document dewarping method that removes curl and geometric distortion of modern and historical documents. Finally, the proposed method is evaluated and compared to the existing Computer Vision based method. Most of the traditional dewarping algorithms are created based on the text line feature extraction and segmentation. However, textual content extraction and segmentation can be sophisticated. Hence, the new technique is proposed, which doesn’t need any complicated methods to process the text lines. The proposed method is based on Deep Learning and it can be applied on all type of text documents and also documents with images and graphics. Moreover, there is no preprocessing required to apply this method on warped images. In the proposed system, the document distortion problem is treated as an image-to-image translation. The new method is implemented using a very powerful pix2pixhd network by utilizing Conditional Generative Adversarial Networks (CGAN). The network is trained on UW3 dataset by supplying distorted document as an input and cleaned image as the target. The generated images from the proposed method are cleanly dewarped and they are of high-resolution. Furthermore, these images can be used to perform OCR.
图像畸变一直是光学字符识别(OCR)中的一个主要问题。为了对失真图像进行OCR处理,去翘曲已经成为一个重要的预处理步骤。提出了一种新的文献去翘曲方法,消除了现代文献和历史文献的卷曲和几何畸变。最后,对该方法进行了评价,并与现有的基于计算机视觉的方法进行了比较。传统的去翘曲算法大多是基于文本行特征的提取和分割。然而,文本内容的提取和分割可能是复杂的。因此,提出了一种不需要任何复杂方法来处理文本行的新技术。该方法基于深度学习,可以应用于所有类型的文本文档以及带有图像和图形的文档。此外,该方法不需要对扭曲图像进行预处理。在该系统中,文档失真问题被视为图像到图像的翻译。该方法利用条件生成对抗网络(CGAN)实现了一个非常强大的pix2pixhd网络。该网络在UW3数据集上进行训练,将扭曲的文档作为输入,将清洗后的图像作为目标。该方法生成的图像经过了清晰的形变处理,具有较高的分辨率。此外,这些图像可以用来执行OCR。
{"title":"Document Image Dewarping using Deep Learning","authors":"Vijaya Kumar Bajjer Ramanna, S. S. Bukhari, A. Dengel","doi":"10.5220/0007368405240531","DOIUrl":"https://doi.org/10.5220/0007368405240531","url":null,"abstract":"The distorted images have been a major problem for Optical Character Recognition (OCR). In order to perform OCR on distorted images, dewarping has become a principal preprocessing step. This paper presents a new document dewarping method that removes curl and geometric distortion of modern and historical documents. Finally, the proposed method is evaluated and compared to the existing Computer Vision based method. Most of the traditional dewarping algorithms are created based on the text line feature extraction and segmentation. However, textual content extraction and segmentation can be sophisticated. Hence, the new technique is proposed, which doesn’t need any complicated methods to process the text lines. The proposed method is based on Deep Learning and it can be applied on all type of text documents and also documents with images and graphics. Moreover, there is no preprocessing required to apply this method on warped images. In the proposed system, the document distortion problem is treated as an image-to-image translation. The new method is implemented using a very powerful pix2pixhd network by utilizing Conditional Generative Adversarial Networks (CGAN). The network is trained on UW3 dataset by supplying distorted document as an input and cleaned image as the target. The generated images from the proposed method are cleanly dewarped and they are of high-resolution. Furthermore, these images can be used to perform OCR.","PeriodicalId":410036,"journal":{"name":"International Conference on Pattern Recognition Applications and Methods","volume":"211 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-02-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116150266","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 15
Online Budgeted Stochastic Coordinate Ascent for Large-Scale Kernelized Dual Support Vector Machine Training 大规模核化双支持向量机训练的在线预算随机坐标上升
Pub Date : 2019-02-19 DOI: 10.1007/978-3-030-40014-9_2
Sahar Qaadan, Abhijeet Pendyala, Merlin Schüler, T. Glasmachers
{"title":"Online Budgeted Stochastic Coordinate Ascent for Large-Scale Kernelized Dual Support Vector Machine Training","authors":"Sahar Qaadan, Abhijeet Pendyala, Merlin Schüler, T. Glasmachers","doi":"10.1007/978-3-030-40014-9_2","DOIUrl":"https://doi.org/10.1007/978-3-030-40014-9_2","url":null,"abstract":"","PeriodicalId":410036,"journal":{"name":"International Conference on Pattern Recognition Applications and Methods","volume":"207 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-02-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133779402","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Understanding of Non-linear Parametric Regression and Classification Models: A Taylor Series based Approach 非线性参数回归和分类模型的理解:基于泰勒级数的方法
Pub Date : 2019-02-19 DOI: 10.5220/0007682008740880
T. Bocklitz
Machine learning methods like classification and regression models are specific solutions for pattern recognition problems. Subsequently, the patterns ’found’ by these methods can be used either in an exploration manner or the model converts the patterns into discriminative values or regression predictions. In both application scenarios it is important to visualize the data-basis of the model, because this unravels the patterns. In case of linear classifiers or linear regression models the task is straight forward, because the model is characterized by a vector which acts as variable weighting and can be visualized. For non-linear models the visualization task is not solved yet and therefore these models act as ’black box’ systems. In this contribution we present a framework, which approximates a given trained parametric model (either classification or regression model) by a series of polynomial models derived from a Taylor expansion of the original non-linear model’s output function. These polynomial models can be visualized until the second order and subsequently interpreted. This visualization opens the ways to understand the data basis of a trained non-linear model and it allows estimating the degree of its non-linearity. By doing so the framework helps to understand non-linear models used for pattern recognition tasks and unravel patterns these methods were using for their predictions.
分类和回归模型等机器学习方法是模式识别问题的具体解决方案。随后,通过这些方法“发现”的模式可以用于探索方式,或者模型将模式转换为判别值或回归预测。在这两种应用程序场景中,可视化模型的数据基础非常重要,因为这将揭示模式。在线性分类器或线性回归模型的情况下,任务是直接的,因为模型的特征是一个向量,它作为可变权重,可以可视化。对于非线性模型,可视化任务尚未解决,因此这些模型充当“黑匣子”系统。在这个贡献中,我们提出了一个框架,它通过一系列多项式模型来近似给定的训练参数模型(分类或回归模型),这些模型来自原始非线性模型的输出函数的泰勒展开。这些多项式模型可以可视化,直到二阶,随后解释。这种可视化打开了理解经过训练的非线性模型的数据基础的方法,并允许估计其非线性程度。通过这样做,该框架有助于理解用于模式识别任务的非线性模型,并揭示这些方法用于预测的模式。
{"title":"Understanding of Non-linear Parametric Regression and Classification Models: A Taylor Series based Approach","authors":"T. Bocklitz","doi":"10.5220/0007682008740880","DOIUrl":"https://doi.org/10.5220/0007682008740880","url":null,"abstract":"Machine learning methods like classification and regression models are specific solutions for pattern recognition problems. Subsequently, the patterns ’found’ by these methods can be used either in an exploration manner or the model converts the patterns into discriminative values or regression predictions. In both application scenarios it is important to visualize the data-basis of the model, because this unravels the patterns. In case of linear classifiers or linear regression models the task is straight forward, because the model is characterized by a vector which acts as variable weighting and can be visualized. For non-linear models the visualization task is not solved yet and therefore these models act as ’black box’ systems. In this contribution we present a framework, which approximates a given trained parametric model (either classification or regression model) by a series of polynomial models derived from a Taylor expansion of the original non-linear model’s output function. These polynomial models can be visualized until the second order and subsequently interpreted. This visualization opens the ways to understand the data basis of a trained non-linear model and it allows estimating the degree of its non-linearity. By doing so the framework helps to understand non-linear models used for pattern recognition tasks and unravel patterns these methods were using for their predictions.","PeriodicalId":410036,"journal":{"name":"International Conference on Pattern Recognition Applications and Methods","volume":"23 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-02-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117120314","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
FoodIE: A Rule-based Named-entity Recognition Method for Food Information Extraction FoodIE:一种基于规则的食品信息抽取命名实体识别方法
Pub Date : 2019-02-19 DOI: 10.5220/0007686309150922
Gorjan Popovski, S. Kochev, B. Korousic-Seljak, T. Eftimov
The application of Natural Language Processing (NLP) methods and resources to biomedical textual data has received growing attention over the past years. Previously organized biomedical NLP-shared tasks (such as, for example, BioNLP Shared Tasks) are related to extracting different biomedical entities (like genes, phenotypes, drugs, diseases, chemical entities) and finding relations between them. However, to the best of our knowledge there are limited NLP methods that can be used for information extraction of entities related to food concepts. For this reason, to extract food entities from unstructured textual data, we propose a rule-based named-entity recognition method for food information extraction, called FoodIE. It is comprised of a small number of rules based on computational linguistics and semantic information that describe the food entities. Experimental results from the evaluation performed using two different datasets showed that very promising results can be achieved. The proposed method achieved 97% precision, 94% recall, and 96% F1 score.
自然语言处理(NLP)方法和资源在生物医学文本数据中的应用近年来受到越来越多的关注。以前组织的生物医学nlp共享任务(如BioNLP共享任务)涉及提取不同的生物医学实体(如基因、表型、药物、疾病、化学实体)并找到它们之间的关系。然而,据我们所知,可以用于与食物概念相关的实体信息提取的NLP方法有限。为此,为了从非结构化文本数据中提取食品实体,我们提出了一种基于规则的食品信息抽取命名实体识别方法FoodIE。它由基于计算语言学和描述食物实体的语义信息的少量规则组成。使用两个不同的数据集进行评估的实验结果表明,可以获得非常有希望的结果。该方法的准确率为97%,召回率为94%,F1得分为96%。
{"title":"FoodIE: A Rule-based Named-entity Recognition Method for Food Information Extraction","authors":"Gorjan Popovski, S. Kochev, B. Korousic-Seljak, T. Eftimov","doi":"10.5220/0007686309150922","DOIUrl":"https://doi.org/10.5220/0007686309150922","url":null,"abstract":"The application of Natural Language Processing (NLP) methods and resources to biomedical textual data has received growing attention over the past years. Previously organized biomedical NLP-shared tasks (such as, for example, BioNLP Shared Tasks) are related to extracting different biomedical entities (like genes, phenotypes, drugs, diseases, chemical entities) and finding relations between them. However, to the best of our knowledge there are limited NLP methods that can be used for information extraction of entities related to food concepts. For this reason, to extract food entities from unstructured textual data, we propose a rule-based named-entity recognition method for food information extraction, called FoodIE. It is comprised of a small number of rules based on computational linguistics and semantic information that describe the food entities. Experimental results from the evaluation performed using two different datasets showed that very promising results can be achieved. The proposed method achieved 97% precision, 94% recall, and 96% F1 score.","PeriodicalId":410036,"journal":{"name":"International Conference on Pattern Recognition Applications and Methods","volume":"21 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-02-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117161124","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 40
Gaussian Model Trees for Traffic Imputation 基于高斯模型树的交通估算
Pub Date : 2019-02-19 DOI: 10.5220/0007690502430254
Sebastian Buschjäger, T. Liebig, K. Morik
Traffic congestion is one of the most pressing issues for smart cities. Information on traffic flow can be used to reduce congestion by predicting vehicle counts at unmonitored locations so that counter-measures can be applied before congestion appears. To do so pricy sensors must be distributed sparsely in the city and at important roads in the city center to collect road and vehicle information throughout the city in real-time. Then, Machine Learning models can be applied to predict vehicle counts at unmonitored locations. To be fault-tolerant and increase coverage of the traffic predictions to the suburbs, rural regions, or even neighboring villages, these Machine Learning models should not operate at a central traffic control room but rather be distributed across the city. Gaussian Processes (GP) work well in the context of traffic count prediction, but cannot capitalize on the vast amount of data available in an entire city. Furthermore, Gaussian Processes are a global and centralized model, which requires all measurements to be available at a central computation node. Product of Expert (PoE) models have been proposed as a scalable alternative to Gaussian Processes. A PoE model trains multiple, independent GPs on different subsets of the data and weight individual predictions based on each experts uncertainty. These methods work well, but they assume that experts are independent even though they may share data points. Furthermore, PoE models require exhaustive communication bandwidth between the individual experts to form the final prediction. In this paper we propose a hierarchical Product of Expert model, which consist of multiple layers of small, independent and local GP experts. We view Gaussian Process induction as regularized optimization procedure and utilize this view to derive an efficient algorithm which selects independent regions of the data. Then, we train local expert models on these regions, so that each expert is responsible for a given region. The resulting algorithm scales well for large amounts of data and outperforms flat PoE models in terms of communication cost, model size and predictive performance. Last, we discuss how to deploy these local expert models onto small devices.
交通拥堵是智慧城市面临的最紧迫问题之一。交通流量信息可以通过预测未监控地点的车辆数量来减少拥堵,以便在拥堵出现之前采取应对措施。要做到这一点,昂贵的传感器必须稀疏地分布在城市和城市中心的重要道路上,以实时收集整个城市的道路和车辆信息。然后,机器学习模型可以应用于预测未监控位置的车辆数量。为了容错并增加对郊区、农村地区甚至邻近村庄的交通预测的覆盖范围,这些机器学习模型不应该在中央交通控制室运行,而是应该分布在整个城市。高斯过程(GP)在交通计数预测方面工作得很好,但无法利用整个城市的大量可用数据。此外,高斯过程是一个全局和集中的模型,它要求所有的测量都在一个中心计算节点上可用。专家产品(PoE)模型已被提出作为高斯过程的可扩展替代方案。PoE模型在不同的数据子集上训练多个独立的gp,并根据每个专家的不确定性对单个预测进行加权。这些方法很有效,但它们假设专家是独立的,即使他们可能共享数据点。此外,PoE模型需要各个专家之间的详尽通信带宽来形成最终预测。本文提出了一种由多层小的、独立的、局部的GP专家组成的分层专家产品模型。我们将高斯过程归纳视为正则化的优化过程,并利用这一观点推导出一种有效的算法来选择数据的独立区域。然后,我们在这些区域上训练局部专家模型,使每个专家负责一个给定的区域。所得到的算法可以很好地适用于大量数据,并且在通信成本、模型大小和预测性能方面优于扁平PoE模型。最后,我们讨论了如何将这些局部专家模型部署到小型设备上。
{"title":"Gaussian Model Trees for Traffic Imputation","authors":"Sebastian Buschjäger, T. Liebig, K. Morik","doi":"10.5220/0007690502430254","DOIUrl":"https://doi.org/10.5220/0007690502430254","url":null,"abstract":"Traffic congestion is one of the most pressing issues for smart cities. Information on traffic flow can be used to reduce congestion by predicting vehicle counts at unmonitored locations so that counter-measures can be applied before congestion appears. To do so pricy sensors must be distributed sparsely in the city and at important roads in the city center to collect road and vehicle information throughout the city in real-time. Then, Machine Learning models can be applied to predict vehicle counts at unmonitored locations. To be fault-tolerant and increase coverage of the traffic predictions to the suburbs, rural regions, or even neighboring villages, these Machine Learning models should not operate at a central traffic control room but rather be distributed across the city. Gaussian Processes (GP) work well in the context of traffic count prediction, but cannot capitalize on the vast amount of data available in an entire city. Furthermore, Gaussian Processes are a global and centralized model, which requires all measurements to be available at a central computation node. Product of Expert (PoE) models have been proposed as a scalable alternative to Gaussian Processes. A PoE model trains multiple, independent GPs on different subsets of the data and weight individual predictions based on each experts uncertainty. These methods work well, but they assume that experts are independent even though they may share data points. Furthermore, PoE models require exhaustive communication bandwidth between the individual experts to form the final prediction. In this paper we propose a hierarchical Product of Expert model, which consist of multiple layers of small, independent and local GP experts. We view Gaussian Process induction as regularized optimization procedure and utilize this view to derive an efficient algorithm which selects independent regions of the data. Then, we train local expert models on these regions, so that each expert is responsible for a given region. The resulting algorithm scales well for large amounts of data and outperforms flat PoE models in terms of communication cost, model size and predictive performance. Last, we discuss how to deploy these local expert models onto small devices.","PeriodicalId":410036,"journal":{"name":"International Conference on Pattern Recognition Applications and Methods","volume":"5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-02-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116281864","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Pedestrian Similarity Extraction to Improve People Counting Accuracy 行人相似度提取提高计数准确率
Pub Date : 2019-02-19 DOI: 10.5220/0007381605480555
Xu Yang, J. Gaspar, W. Ke, C. Lam, Yanwei Zheng, W. Lou, Yapeng Wang
Current state-of-the-art single shot object detection pipelines, composed by an object detector such as Yolo, generate multiple detections for each object, requiring a post-processing Non-Maxima Suppression (NMS) algorithm to remove redundant detections. However, this pipeline struggles to achieve high accuracy, particularly in object counting applications, due to a trade-off between precision and recall rates. A higher NMS threshold results in fewer detections suppressed and, consequently, in a higher recall rate, as well as lower precision and accuracy. In this paper, we have explored a new pedestrian detection pipeline which is more flexible, able to adapt to different scenarios and with improved precision and accuracy. A higher NMS threshold is used to retain all true detections and achieve a high recall rate for different scenarios, and a Pedestrian Similarity Extraction (PSE) algorithm is used to remove redundant detentions, consequently improving counting accuracy. The PSE algorithm significantly reduces the detection accuracy volatility and its dependency on NMS thresholds, improving the mean detection accuracy for different input datasets.
目前最先进的单镜头目标检测管道由Yolo等目标检测器组成,为每个目标生成多个检测,需要后处理非最大抑制(NMS)算法来去除冗余检测。然而,由于精确度和召回率之间的权衡,这种管道难以达到高精度,特别是在对象计数应用中。更高的NMS阈值会导致更少的检测被抑制,从而导致更高的召回率,以及更低的精度和准确性。在本文中,我们探索了一种新的行人检测管道,该管道更加灵活,能够适应不同的场景,并且提高了精度和准确性。使用更高的NMS阈值来保留所有真实检测并实现不同场景的高召回率,并使用行人相似度提取(PSE)算法来去除冗余滞留,从而提高计数精度。PSE算法显著降低了检测精度的波动性及其对NMS阈值的依赖,提高了不同输入数据集的平均检测精度。
{"title":"Pedestrian Similarity Extraction to Improve People Counting Accuracy","authors":"Xu Yang, J. Gaspar, W. Ke, C. Lam, Yanwei Zheng, W. Lou, Yapeng Wang","doi":"10.5220/0007381605480555","DOIUrl":"https://doi.org/10.5220/0007381605480555","url":null,"abstract":"Current state-of-the-art single shot object detection pipelines, composed by an object detector such as Yolo, generate multiple detections for each object, requiring a post-processing Non-Maxima Suppression (NMS) algorithm to remove redundant detections. However, this pipeline struggles to achieve high accuracy, particularly in object counting applications, due to a trade-off between precision and recall rates. A higher NMS threshold results in fewer detections suppressed and, consequently, in a higher recall rate, as well as lower precision and accuracy. In this paper, we have explored a new pedestrian detection pipeline which is more flexible, able to adapt to different scenarios and with improved precision and accuracy. A higher NMS threshold is used to retain all true detections and achieve a high recall rate for different scenarios, and a Pedestrian Similarity Extraction (PSE) algorithm is used to remove redundant detentions, consequently improving counting accuracy. The PSE algorithm significantly reduces the detection accuracy volatility and its dependency on NMS thresholds, improving the mean detection accuracy for different input datasets.","PeriodicalId":410036,"journal":{"name":"International Conference on Pattern Recognition Applications and Methods","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-02-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115447434","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
期刊
International Conference on Pattern Recognition Applications and Methods
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1