首页 > 最新文献

2018 31st SIBGRAPI Conference on Graphics, Patterns and Images (SIBGRAPI)最新文献

英文 中文
Multicenter Imaging Studies: Automated Approach to Evaluating Data Variability and the Role of Outliers 多中心成像研究:评估数据变异性和异常值作用的自动化方法
Pub Date : 2018-10-01 DOI: 10.1109/SIBGRAPI.2018.00030
M. Bento, R. Souza, R. Frayne
Magnetic resonance (MR) as well as other imaging modalities have been used in a large number of clinical and research studies for the analysis and quantification of important structures and the detection of abnormalities. In this context, machine learning is playing an increasingly important role in the development of automated tools for aiding in image quantification, patient diagnosis and follow-up. Normally, these techniques require large, heterogeneous datasets to provide accurate and generalizable results. Large, multi-center studies, for example, can provide such data. Images acquired at different centers, however, can present varying characteristics due to differences in acquisition parameters, site procedures and scanners configuration. While variability in the dataset is required to develop robust, generalizable studies (i.e., independent of the acquisition parameters or center), like all studies there is also a need to ensure overall data quality by prospectively identifying and removing poor-quality data samples that should not be included, e.g., outliers. We wish to keep image samples that are representative of the underlying population (so called inliers), yet removing those samples that are not. We propose a framework to analyze data variability and identify samples that should be removed in order to have more representative, reliable and robust datasets. Our example case study is based on a public dataset containing T1-weighted volumetric head images data acquired at six different centers, using three different scanner vendors and at two commonly used magnetic fields strengths. We propose an algorithm for assessing data robustness and finding the optimal data for study occlusion (i.e., the data size that presents with lowest variability while maintaining generalizability (i.e., using samples from all sites)).
磁共振(MR)以及其他成像方式已在大量的临床和研究中用于重要结构的分析和量化以及异常的检测。在这种情况下,机器学习在辅助图像量化、患者诊断和随访的自动化工具的开发中发挥着越来越重要的作用。通常,这些技术需要大型异构数据集来提供准确和可推广的结果。例如,大型、多中心的研究可以提供这样的数据。然而,由于采集参数、现场程序和扫描仪配置的差异,在不同中心获取的图像可能呈现不同的特征。虽然需要数据集中的可变性来开展稳健的、可推广的研究(即独立于采集参数或中心),但与所有研究一样,还需要通过前瞻性地识别和删除不应包括的低质量数据样本(例如异常值)来确保整体数据质量。我们希望保留代表底层总体的图像样本(即所谓的内线),同时删除那些不代表底层总体的样本。我们提出了一个框架来分析数据变异性,并确定应该删除的样本,以获得更具代表性,可靠和健壮的数据集。我们的示例案例研究基于一个公共数据集,该数据集包含从六个不同的中心获取的t1加权体积头部图像数据,使用三个不同的扫描仪供应商和两种常用的磁场强度。我们提出了一种算法,用于评估数据稳健性和寻找研究遮挡的最佳数据(即,在保持通用性的同时呈现最低变异性的数据大小(即使用来自所有站点的样本))。
{"title":"Multicenter Imaging Studies: Automated Approach to Evaluating Data Variability and the Role of Outliers","authors":"M. Bento, R. Souza, R. Frayne","doi":"10.1109/SIBGRAPI.2018.00030","DOIUrl":"https://doi.org/10.1109/SIBGRAPI.2018.00030","url":null,"abstract":"Magnetic resonance (MR) as well as other imaging modalities have been used in a large number of clinical and research studies for the analysis and quantification of important structures and the detection of abnormalities. In this context, machine learning is playing an increasingly important role in the development of automated tools for aiding in image quantification, patient diagnosis and follow-up. Normally, these techniques require large, heterogeneous datasets to provide accurate and generalizable results. Large, multi-center studies, for example, can provide such data. Images acquired at different centers, however, can present varying characteristics due to differences in acquisition parameters, site procedures and scanners configuration. While variability in the dataset is required to develop robust, generalizable studies (i.e., independent of the acquisition parameters or center), like all studies there is also a need to ensure overall data quality by prospectively identifying and removing poor-quality data samples that should not be included, e.g., outliers. We wish to keep image samples that are representative of the underlying population (so called inliers), yet removing those samples that are not. We propose a framework to analyze data variability and identify samples that should be removed in order to have more representative, reliable and robust datasets. Our example case study is based on a public dataset containing T1-weighted volumetric head images data acquired at six different centers, using three different scanner vendors and at two commonly used magnetic fields strengths. We propose an algorithm for assessing data robustness and finding the optimal data for study occlusion (i.e., the data size that presents with lowest variability while maintaining generalizability (i.e., using samples from all sites)).","PeriodicalId":208985,"journal":{"name":"2018 31st SIBGRAPI Conference on Graphics, Patterns and Images (SIBGRAPI)","volume":"5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114887915","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Hybrid Cloud Rendering System for Massive CAD Models 用于大规模CAD模型的混合云渲染系统
Pub Date : 2018-10-01 DOI: 10.1109/SIBGRAPI.2018.00037
A. Moreira, Paulo Ivson, Waldemar Celes Filho
The recent advances in cloud services enable an increasing number of applications to offload their intensive tasks to remote computers. Cloud rendering comprises a set of services capable of rendering a 3D scene on a remote workstation. Notable progress in this field has been made by cloud gaming services. However, a gap remains between existing cloud rendering systems and other graphics-intensive applications, such as visualization of Computer-Aided Design (CAD) models. Existing cloud gaming services are not suitable to efficiently render these particular 3D scenes. CAD models contain many more objects than a regular game scene, requiring specific assumptions and optimizations to deliver an interactive user experience. In this work, we discuss and propose a novel hybrid cloud rendering system for massive 3D CAD models of industrial plants. The obtained results show that our technique can achieve high frame rates with satisfactory image quality even in a constrained environment, such as a high latency network or obsolete computer hardware.
云服务的最新进展使越来越多的应用程序能够将其繁重的任务卸载到远程计算机上。云渲染包括一组能够在远程工作站上渲染3D场景的服务。云游戏服务在这一领域取得了显著进展。然而,在现有的云渲染系统和其他图形密集型应用程序(如计算机辅助设计(CAD)模型的可视化)之间仍然存在差距。现有的云游戏服务并不适合有效地渲染这些特定的3D场景。CAD模型包含比常规游戏场景更多的对象,需要特定的假设和优化来提供交互式用户体验。在这项工作中,我们讨论并提出了一种新的混合云渲染系统,用于大规模工业厂房的3D CAD模型。结果表明,即使在高延迟网络或过时的计算机硬件等受限环境下,我们的技术也可以实现高帧率和令人满意的图像质量。
{"title":"Hybrid Cloud Rendering System for Massive CAD Models","authors":"A. Moreira, Paulo Ivson, Waldemar Celes Filho","doi":"10.1109/SIBGRAPI.2018.00037","DOIUrl":"https://doi.org/10.1109/SIBGRAPI.2018.00037","url":null,"abstract":"The recent advances in cloud services enable an increasing number of applications to offload their intensive tasks to remote computers. Cloud rendering comprises a set of services capable of rendering a 3D scene on a remote workstation. Notable progress in this field has been made by cloud gaming services. However, a gap remains between existing cloud rendering systems and other graphics-intensive applications, such as visualization of Computer-Aided Design (CAD) models. Existing cloud gaming services are not suitable to efficiently render these particular 3D scenes. CAD models contain many more objects than a regular game scene, requiring specific assumptions and optimizations to deliver an interactive user experience. In this work, we discuss and propose a novel hybrid cloud rendering system for massive 3D CAD models of industrial plants. The obtained results show that our technique can achieve high frame rates with satisfactory image quality even in a constrained environment, such as a high latency network or obsolete computer hardware.","PeriodicalId":208985,"journal":{"name":"2018 31st SIBGRAPI Conference on Graphics, Patterns and Images (SIBGRAPI)","volume":"30 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126415386","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Deep Feature-Based Classifiers for Fruit Fly Identification (Diptera: Tephritidae) 基于深度特征的果蝇识别分类器(双翅目:蝗科)
Pub Date : 2018-10-01 DOI: 10.1109/SIBGRAPI.2018.00012
Matheus Macedo Leonardo, Tiago J. Carvalho, Edmar R. S. Rezende, R. Zucchi, F. Faria
Fruit flies has a big biological and economic importance for the farming of different tropical and subtropical countries in the World. Specifically in Brazil, third largest fruit producer in the world, the direct and indirect losses caused by fruit flies can exceed USD 120 million/year. These losses are related to production, the cost of pest control and export markets. One of the most economically important fruit flies in the America belong to the genus Anastrepha, which has approximately 300 known species, of which 120 are recorded in Brazil. However, less than 10 species are economically important and are considered pests of quarantine significance by regulatory agencies. The extreme similarity among the species of the genus Anastrepha makes its manual taxonomic classification a nontrivial task, causing onerous and very subjective results. In this work, we propose an approach based on deep learning to assist the scarce specialists, reducing the time of analysis, subjectivity of the classifications and consequently, the economic losses related to these agricultural pests. In our experiments, five deep features and nine machine learning techniques have been studied for the target task. Furthermore, the proposed approach have achieved similar effectiveness results to state-of-art approaches.
果蝇对世界上不同热带和亚热带国家的农业有着重要的生物学和经济意义。特别是在世界第三大水果生产国巴西,果蝇造成的直接和间接损失每年可超过1.2亿美元。这些损失与生产、虫害防治费用和出口市场有关。美洲最具经济价值的果蝇之一属于Anastrepha属,已知约有300种,其中120种记录在巴西。然而,只有不到10种具有重要的经济价值,并被监管机构认为具有检疫意义。Anastrepha属的物种之间的极端相似性使得其人工分类分类成为一项艰巨的任务,导致繁重和非常主观的结果。在这项工作中,我们提出了一种基于深度学习的方法来帮助稀缺的专家,减少分析时间,减少分类的主观性,从而减少与这些农业害虫相关的经济损失。在我们的实验中,针对目标任务研究了5种深度特征和9种机器学习技术。此外,所提出的方法取得了与最先进方法相似的有效性结果。
{"title":"Deep Feature-Based Classifiers for Fruit Fly Identification (Diptera: Tephritidae)","authors":"Matheus Macedo Leonardo, Tiago J. Carvalho, Edmar R. S. Rezende, R. Zucchi, F. Faria","doi":"10.1109/SIBGRAPI.2018.00012","DOIUrl":"https://doi.org/10.1109/SIBGRAPI.2018.00012","url":null,"abstract":"Fruit flies has a big biological and economic importance for the farming of different tropical and subtropical countries in the World. Specifically in Brazil, third largest fruit producer in the world, the direct and indirect losses caused by fruit flies can exceed USD 120 million/year. These losses are related to production, the cost of pest control and export markets. One of the most economically important fruit flies in the America belong to the genus Anastrepha, which has approximately 300 known species, of which 120 are recorded in Brazil. However, less than 10 species are economically important and are considered pests of quarantine significance by regulatory agencies. The extreme similarity among the species of the genus Anastrepha makes its manual taxonomic classification a nontrivial task, causing onerous and very subjective results. In this work, we propose an approach based on deep learning to assist the scarce specialists, reducing the time of analysis, subjectivity of the classifications and consequently, the economic losses related to these agricultural pests. In our experiments, five deep features and nine machine learning techniques have been studied for the target task. Furthermore, the proposed approach have achieved similar effectiveness results to state-of-art approaches.","PeriodicalId":208985,"journal":{"name":"2018 31st SIBGRAPI Conference on Graphics, Patterns and Images (SIBGRAPI)","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134416818","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 38
An Architecture for Collision Risk Prediction for Visually Impaired People 视障人群碰撞风险预测体系结构研究
Pub Date : 2018-10-01 DOI: 10.1109/SIBGRAPI.2018.00046
Natal Henrique Cordeiro, E. C. Pedrino
The production of sensory substitution equipment for the visually impaired (VIP) is growing. The aim of this project is to understand the VIP context and predict the risks of collision for the VIP, following an analysis of the position, distance, size and motion of the objects present in their environment. This understanding is refined by data fusion steps applied to the Situation Awareness model to predict possible impacts in the near future. With this goal, a new architecture was designed, composed of systems that detect free passages, static objects, dynamic objects and the paths of these dynamic objects. The detected data was mapped into a 3D plane verifying positions and sizes. For the fusion, a method was developed that compared four more general classifiers in order to verify which presented greater reliability in the given context. These classifiers allowed inferences to be made when analyzing the risks of collision in different directions. The architecture designed for risk prediction is the main contribution of this project.
为视障人士提供感官替代设备的生产正在增加。该项目的目的是了解VIP的环境,并在分析环境中物体的位置、距离、大小和运动后,预测VIP发生碰撞的风险。这种理解通过应用于态势感知模型的数据融合步骤来改进,以预测近期可能的影响。为此,设计了一种新的体系结构,由检测自由通道、静态对象、动态对象以及这些动态对象的路径的系统组成。检测到的数据被映射到一个三维平面,验证位置和尺寸。对于融合,开发了一种方法,比较了四种更一般的分类器,以验证在给定的上下文中哪一种具有更高的可靠性。这些分类器允许在分析不同方向的碰撞风险时做出推断。为风险预测设计的体系结构是该项目的主要贡献。
{"title":"An Architecture for Collision Risk Prediction for Visually Impaired People","authors":"Natal Henrique Cordeiro, E. C. Pedrino","doi":"10.1109/SIBGRAPI.2018.00046","DOIUrl":"https://doi.org/10.1109/SIBGRAPI.2018.00046","url":null,"abstract":"The production of sensory substitution equipment for the visually impaired (VIP) is growing. The aim of this project is to understand the VIP context and predict the risks of collision for the VIP, following an analysis of the position, distance, size and motion of the objects present in their environment. This understanding is refined by data fusion steps applied to the Situation Awareness model to predict possible impacts in the near future. With this goal, a new architecture was designed, composed of systems that detect free passages, static objects, dynamic objects and the paths of these dynamic objects. The detected data was mapped into a 3D plane verifying positions and sizes. For the fusion, a method was developed that compared four more general classifiers in order to verify which presented greater reliability in the given context. These classifiers allowed inferences to be made when analyzing the risks of collision in different directions. The architecture designed for risk prediction is the main contribution of this project.","PeriodicalId":208985,"journal":{"name":"2018 31st SIBGRAPI Conference on Graphics, Patterns and Images (SIBGRAPI)","volume":"41 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133336310","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
xHiPP: eXtended Hierarchical Point Placement Strategy 扩展分层点放置策略
Pub Date : 2018-10-01 DOI: 10.1109/SIBGRAPI.2018.00053
F. Dias, R. Minghim
The complexity and size of data have created challenges to data analysis. Although point placement strategies have gained popularity in the last decade to yield a global view of multidimensional datasets, few attempts have been made to improve visual scalability and offer multilevel exploration in the context of multidimensional projections and point placement strategies. Such approaches can be helpful in improving the analysis capability both by organizing visual spaces and allowing meaningful partitions of larger datasets. In this paper, we extend the Hierarchy Point Placement (HiPP), a strategy for multi-level point placement, in order to enhance its analytical capabilities and flexibility to handle current trends in visual data science. We have provided several combinations of clustering methods and projection approaches to represent and visualize datasets; added a choice to invert the original processing order from cluster-projection to projection-cluster; proposed a better way to initialize the partitions, and added ways to summarize image, audio, text and general data groups. The tool's code is made available online. In this article, we present the new tool (named xHiPP) and provide validation through case studies with simpler and more complex datasets (text and audio) to illustrate that the capabilities afforded by the extensions have managed to provide analysts with the ability to quickly gain insight and adjust the processing pipeline to their needs.
数据的复杂性和规模给数据分析带来了挑战。尽管在过去的十年中,点放置策略已经获得了广泛的应用,以产生多维数据集的全局视图,但在多维投影和点放置策略的背景下,很少有人尝试提高视觉可扩展性并提供多层次的探索。这种方法可以通过组织视觉空间和允许对更大的数据集进行有意义的分区来帮助提高分析能力。在本文中,我们扩展了层次点放置策略(HiPP),以提高其分析能力和灵活性,以应对当前视觉数据科学的发展趋势。我们提供了几种聚类方法和投影方法的组合来表示和可视化数据集;增加了将原始处理顺序从集群-投影转换为投影-集群的选项;提出了一种更好的初始化分区的方法,并增加了对图像、音频、文本和一般数据组进行汇总的方法。该工具的代码可以在网上获得。在本文中,我们介绍了这个新工具(名为xHiPP),并通过使用更简单和更复杂的数据集(文本和音频)的案例研究提供了验证,以说明扩展提供的功能已经成功地为分析人员提供了快速获得洞察力和调整处理管道以满足其需求的能力。
{"title":"xHiPP: eXtended Hierarchical Point Placement Strategy","authors":"F. Dias, R. Minghim","doi":"10.1109/SIBGRAPI.2018.00053","DOIUrl":"https://doi.org/10.1109/SIBGRAPI.2018.00053","url":null,"abstract":"The complexity and size of data have created challenges to data analysis. Although point placement strategies have gained popularity in the last decade to yield a global view of multidimensional datasets, few attempts have been made to improve visual scalability and offer multilevel exploration in the context of multidimensional projections and point placement strategies. Such approaches can be helpful in improving the analysis capability both by organizing visual spaces and allowing meaningful partitions of larger datasets. In this paper, we extend the Hierarchy Point Placement (HiPP), a strategy for multi-level point placement, in order to enhance its analytical capabilities and flexibility to handle current trends in visual data science. We have provided several combinations of clustering methods and projection approaches to represent and visualize datasets; added a choice to invert the original processing order from cluster-projection to projection-cluster; proposed a better way to initialize the partitions, and added ways to summarize image, audio, text and general data groups. The tool's code is made available online. In this article, we present the new tool (named xHiPP) and provide validation through case studies with simpler and more complex datasets (text and audio) to illustrate that the capabilities afforded by the extensions have managed to provide analysts with the ability to quickly gain insight and adjust the processing pipeline to their needs.","PeriodicalId":208985,"journal":{"name":"2018 31st SIBGRAPI Conference on Graphics, Patterns and Images (SIBGRAPI)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131365914","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
A Deep Learning-Based Compatibility Score for Reconstruction of Strip-Shredded Text Documents 基于深度学习的条带粉碎文本文档重建兼容性评分
Pub Date : 2018-10-01 DOI: 10.1109/SIBGRAPI.2018.00018
T. M. Paixão, Rodrigo Berriel, M. C. Boeres, C. Badue, A. D. Souza, Thiago Oliveira-Santos
The use of paper-shredder machines (mechanical shredding) to destroy documents can be illicitly motivated when the purpose is hiding evidence of fraud and other sorts of crimes. Therefore, reconstructing such documents is of great value for forensic investigation, but it is admittedly a stressful and time-consuming task for humans. To address this challenge, several computational techniques have been proposed in literature, particularly for documents with text-based content. In this context, a critical challenge for automated reconstruction is to measure properly the fitting (compatibility) between paper shreds (strips), which has been observed to be the main limitation of literature on this topic. The main contribution of this paper is a deep learning-based compatibility score to be applied in the reconstruction of strip-shredded text documents. Since there is no abundance of real-shredded data, we propose a training scheme based on digital simulated-shredding of documents from a well-known OCR database. The proposed score was coupled to a black-box optimization tool, and the resulting system achieved an average accuracy of 94.58% in the reconstruction of mechanically-shredded documents.
当使用碎纸机(机械碎纸机)销毁文件的目的是隐藏欺诈和其他罪行的证据时,可能是非法动机。因此,重建这些文件对法医调查具有重要的价值,但对于人类来说,这是一项紧张而耗时的任务。为了应对这一挑战,文献中已经提出了几种计算技术,特别是针对具有基于文本内容的文档。在这种情况下,自动重建的一个关键挑战是适当地测量纸碎片(条)之间的拟合(兼容性),这已经被观察到是关于该主题的文献的主要限制。本文的主要贡献是基于深度学习的兼容性评分,用于条带切碎文本文档的重建。由于实际切碎的数据并不丰富,我们提出了一种基于数字模拟切碎文档的训练方案,该训练方案来自一个知名的OCR数据库。将提出的分数与黑盒优化工具耦合,得到的系统在机械撕碎文件的重建中平均准确率达到94.58%。
{"title":"A Deep Learning-Based Compatibility Score for Reconstruction of Strip-Shredded Text Documents","authors":"T. M. Paixão, Rodrigo Berriel, M. C. Boeres, C. Badue, A. D. Souza, Thiago Oliveira-Santos","doi":"10.1109/SIBGRAPI.2018.00018","DOIUrl":"https://doi.org/10.1109/SIBGRAPI.2018.00018","url":null,"abstract":"The use of paper-shredder machines (mechanical shredding) to destroy documents can be illicitly motivated when the purpose is hiding evidence of fraud and other sorts of crimes. Therefore, reconstructing such documents is of great value for forensic investigation, but it is admittedly a stressful and time-consuming task for humans. To address this challenge, several computational techniques have been proposed in literature, particularly for documents with text-based content. In this context, a critical challenge for automated reconstruction is to measure properly the fitting (compatibility) between paper shreds (strips), which has been observed to be the main limitation of literature on this topic. The main contribution of this paper is a deep learning-based compatibility score to be applied in the reconstruction of strip-shredded text documents. Since there is no abundance of real-shredded data, we propose a training scheme based on digital simulated-shredding of documents from a well-known OCR database. The proposed score was coupled to a black-box optimization tool, and the resulting system achieved an average accuracy of 94.58% in the reconstruction of mechanically-shredded documents.","PeriodicalId":208985,"journal":{"name":"2018 31st SIBGRAPI Conference on Graphics, Patterns and Images (SIBGRAPI)","volume":"60 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131488984","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Barrett's Esophagus Identification Using Color Co-Occurrence Matrices 巴雷特食管颜色共现矩阵识别
Pub Date : 2018-10-01 DOI: 10.1109/SIBGRAPI.2018.00028
L. Souza, A. Ebigbo, A. Probst, H. Messmann, J. Papa, R. Mendel, C. Palm
In this work, we propose the use of single channel Color Co-occurrence Matrices for texture description of Barrett's Esophagus (BE) and adenocarcinoma images. Further classification using supervised learning techniques, such as Optimum-Path Forest (OPF), Support Vector Machines with Radial Basis Function (SVM-RBF) and Bayesian classifier supports the context of automatic BE and adenocarcinoma diagnosis. We validated three approaches of classification based on patches, patients and images in two datasets (MICCAI 2015 and Augsburg) using the color-and-texture descriptors and the machine learning techniques. Concerning MICCAI 2015 dataset, the best results were obtained using the blue channel for the descriptors and the supervised OPF for classification purposes in the patch-based approach, with sensitivity nearly to 73% for positive adenocarcinoma identification and specificity close to 77% for BE (non-cancerous) patch classification. Regarding the Augsburg dataset, the most accurate results were also obtained using both OPF classifier and blue channel descriptor for the feature extraction, with sensitivity close to 67% and specificity around to 76%. Our work highlights new advances in the related research area and provides a promising technique that combines color and texture information, allied to three different approaches of dataset pre-processing aiming to configure robust scenarios for the classification step.
在这项工作中,我们提出使用单通道颜色共生矩阵对巴雷特食管(BE)和腺癌图像进行纹理描述。进一步的分类使用监督学习技术,如最优路径森林(OPF)、径向基函数支持向量机(SVM-RBF)和贝叶斯分类器支持自动BE和腺癌诊断。我们在两个数据集(MICCAI 2015和Augsburg)中使用颜色和纹理描述符和机器学习技术验证了基于斑块、患者和图像的三种分类方法。对于MICCAI 2015数据集,在基于补丁的方法中,使用蓝色通道作为描述符和监督OPF用于分类目的获得了最好的结果,对于阳性腺癌识别的敏感性接近73%,对于BE(非癌)补丁分类的特异性接近77%。对于Augsburg数据集,使用OPF分类器和蓝色通道描述符进行特征提取也获得了最准确的结果,灵敏度接近67%,特异性约为76%。我们的工作突出了相关研究领域的新进展,并提供了一种有前途的技术,该技术结合了颜色和纹理信息,结合了三种不同的数据集预处理方法,旨在为分类步骤配置健壮的场景。
{"title":"Barrett's Esophagus Identification Using Color Co-Occurrence Matrices","authors":"L. Souza, A. Ebigbo, A. Probst, H. Messmann, J. Papa, R. Mendel, C. Palm","doi":"10.1109/SIBGRAPI.2018.00028","DOIUrl":"https://doi.org/10.1109/SIBGRAPI.2018.00028","url":null,"abstract":"In this work, we propose the use of single channel Color Co-occurrence Matrices for texture description of Barrett's Esophagus (BE) and adenocarcinoma images. Further classification using supervised learning techniques, such as Optimum-Path Forest (OPF), Support Vector Machines with Radial Basis Function (SVM-RBF) and Bayesian classifier supports the context of automatic BE and adenocarcinoma diagnosis. We validated three approaches of classification based on patches, patients and images in two datasets (MICCAI 2015 and Augsburg) using the color-and-texture descriptors and the machine learning techniques. Concerning MICCAI 2015 dataset, the best results were obtained using the blue channel for the descriptors and the supervised OPF for classification purposes in the patch-based approach, with sensitivity nearly to 73% for positive adenocarcinoma identification and specificity close to 77% for BE (non-cancerous) patch classification. Regarding the Augsburg dataset, the most accurate results were also obtained using both OPF classifier and blue channel descriptor for the feature extraction, with sensitivity close to 67% and specificity around to 76%. Our work highlights new advances in the related research area and provides a promising technique that combines color and texture information, allied to three different approaches of dataset pre-processing aiming to configure robust scenarios for the classification step.","PeriodicalId":208985,"journal":{"name":"2018 31st SIBGRAPI Conference on Graphics, Patterns and Images (SIBGRAPI)","volume":"41 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121739294","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Patch PlaNet: Landmark Recognition with Patch Classification Using Convolutional Neural Networks Patch PlaNet:使用卷积神经网络进行斑块分类的地标识别
Pub Date : 2018-10-01 DOI: 10.1109/SIBGRAPI.2018.00023
K. Cunha, Lucas Maggi, V. Teichrieb, J. P. Lima, J. Quintino, F. Q. Silva, André L. M. Santos, Helder Pinho
In this work we address the problem of landmark recognition. We extend PlaNet, a model based on deep neural networks that approaches the problem of landmark recognition as a classification problem and performs the recognition of places around the world. We propose an extension of the PlaNet technique in which we use a voting scheme to perform the classification, dividing the image into previously defined regions and inferring the landmark based on these regions. The prediction of the model depends not only on the information of the features learned by the deep convolutional neural network architecture during training, but also uses local information from each region in the image for which the classification is made. To validate our proposal, we performed the training of the original PlaNet model and our variation using a database built with images from Flickr, and evaluated the models in the Paris and Oxford Buildings datasets. It was possible to notice that the addition of image division and voting structure improves the accuracy result of the model by 5-11 percentage points on average, reducing the level of ambiguity found during the inference of the model.
在这项工作中,我们解决了地标识别的问题。我们扩展了PlaNet,这是一个基于深度神经网络的模型,它将地标识别问题作为分类问题来处理,并对世界各地的地点进行识别。我们提出了PlaNet技术的扩展,其中我们使用投票方案来执行分类,将图像划分为先前定义的区域,并根据这些区域推断地标。该模型的预测不仅依赖于深度卷积神经网络架构在训练过程中学习到的特征信息,而且还使用了图像中每个分类区域的局部信息。为了验证我们的建议,我们使用一个由Flickr图像构建的数据库对原始行星模型和我们的变体进行了训练,并在巴黎和牛津建筑数据集中评估了模型。可以注意到,图像分割和投票结构的加入使模型的准确率结果平均提高了5-11个百分点,减少了模型推理过程中发现的模糊程度。
{"title":"Patch PlaNet: Landmark Recognition with Patch Classification Using Convolutional Neural Networks","authors":"K. Cunha, Lucas Maggi, V. Teichrieb, J. P. Lima, J. Quintino, F. Q. Silva, André L. M. Santos, Helder Pinho","doi":"10.1109/SIBGRAPI.2018.00023","DOIUrl":"https://doi.org/10.1109/SIBGRAPI.2018.00023","url":null,"abstract":"In this work we address the problem of landmark recognition. We extend PlaNet, a model based on deep neural networks that approaches the problem of landmark recognition as a classification problem and performs the recognition of places around the world. We propose an extension of the PlaNet technique in which we use a voting scheme to perform the classification, dividing the image into previously defined regions and inferring the landmark based on these regions. The prediction of the model depends not only on the information of the features learned by the deep convolutional neural network architecture during training, but also uses local information from each region in the image for which the classification is made. To validate our proposal, we performed the training of the original PlaNet model and our variation using a database built with images from Flickr, and evaluated the models in the Paris and Oxford Buildings datasets. It was possible to notice that the addition of image division and voting structure improves the accuracy result of the model by 5-11 percentage points on average, reducing the level of ambiguity found during the inference of the model.","PeriodicalId":208985,"journal":{"name":"2018 31st SIBGRAPI Conference on Graphics, Patterns and Images (SIBGRAPI)","volume":"30 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122366858","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Hidden Surface Removal for Accurate Painting-Area Calculation on CAD Models 用于CAD模型精确涂漆面积计算的隐藏表面去除
Pub Date : 2018-10-01 DOI: 10.1109/SIBGRAPI.2018.00008
L. Figueiredo, Paulo Ivson, Waldemar Celes Filho
3D CAD models are widely used to improve management of large-scale engineering projects. Examples include Building Information Modeling (BIM) and Oil & Gas industrial plants. Maintaining these facilities is a critical task that often involves anti-corrosive painting of equipment and metallic structures. Existing CAD software estimates the painting area including hidden surfaces that are not actually painted in the field. To improve these computations, we propose an approach based on Adaptively-Sampled Distance Fields (ADFs) exploiting the relationship between object areas and Constructive Solid Geometry (CSG) operations. Tests with synthetic models demonstrate that our technique achieves an accuracy of 99%. In real-world 3D CAD models, we were able to reduce the estimated area by 38% when compared to the naïve calculations. These result in significant cost savings in material provision and workforce required for maintaining these facilities.
三维CAD模型被广泛应用于大型工程项目的管理。例子包括建筑信息模型(BIM)和石油和天然气工业工厂。维护这些设施是一项关键任务,通常涉及设备和金属结构的防腐油漆。现有的CAD软件估计的绘画面积,包括隐藏的表面,实际上没有在现场绘制。为了改进这些计算,我们提出了一种基于自适应采样距离场(adf)的方法,利用目标区域和构造立体几何(CSG)操作之间的关系。综合模型的测试表明,我们的技术达到了99%的准确率。在真实的3D CAD模型中,与naïve计算相比,我们能够将估计面积减少38%。这大大节省了维护这些设施所需的材料供应和劳动力成本。
{"title":"Hidden Surface Removal for Accurate Painting-Area Calculation on CAD Models","authors":"L. Figueiredo, Paulo Ivson, Waldemar Celes Filho","doi":"10.1109/SIBGRAPI.2018.00008","DOIUrl":"https://doi.org/10.1109/SIBGRAPI.2018.00008","url":null,"abstract":"3D CAD models are widely used to improve management of large-scale engineering projects. Examples include Building Information Modeling (BIM) and Oil & Gas industrial plants. Maintaining these facilities is a critical task that often involves anti-corrosive painting of equipment and metallic structures. Existing CAD software estimates the painting area including hidden surfaces that are not actually painted in the field. To improve these computations, we propose an approach based on Adaptively-Sampled Distance Fields (ADFs) exploiting the relationship between object areas and Constructive Solid Geometry (CSG) operations. Tests with synthetic models demonstrate that our technique achieves an accuracy of 99%. In real-world 3D CAD models, we were able to reduce the estimated area by 38% when compared to the naïve calculations. These result in significant cost savings in material provision and workforce required for maintaining these facilities.","PeriodicalId":208985,"journal":{"name":"2018 31st SIBGRAPI Conference on Graphics, Patterns and Images (SIBGRAPI)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131600618","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Real-Time Automatic License Plate Recognition through Deep Multi-Task Networks 基于深度多任务网络的实时自动车牌识别
Pub Date : 2018-10-01 DOI: 10.1109/SIBGRAPI.2018.00021
G. Gonçalves, M. A. Diniz, Rayson Laroca, D. Menotti, W. R. Schwartz
With the increasing number of cameras available in the cities, video traffic analysis can provide useful insights for the transportation segment. One of such analysis is the Automatic License Plate Recognition (ALPR). Previous approaches divided this task into several cascaded subtasks, i.e., vehicle location, license plate detection, character segmentation and optical character recognition. However, since each task has its own accuracy, the error propagation between each subtask is detrimental to the final accuracy. Therefore, focusing on the reduction of error propagation, we propose a technique that is able to perform ALPR using only two deep networks, the first performs license plate detection (LPD) and the second performs license plate recognition (LPR). The latter does not execute explicit character segmentation, which reduces significantly the error propagation. As these deep networks need a large number of samples to converge, we develop new data augmentation techniques that allow them to reach their full potential as well as a new dataset to train and evaluate ALPR approaches. According to experimental results, our approach is able to achieve state-of-the-art results in the SSIG-SegPlate dataset, reaching improvements up to 1.4 percentage point when compared to the best baseline. Furthermore, the approach is also able to perform in real time even in scenarios where many plates are present at the same frame, reaching significantly higher frame rates when compared with previously proposed approaches.
随着城市中摄像头数量的增加,视频交通分析可以为交通部门提供有用的见解。其中一种分析是自动车牌识别(ALPR)。以前的方法将该任务分为几个级联子任务,即车辆定位,车牌检测,字符分割和光学字符识别。但是,由于每个任务都有自己的精度,因此每个子任务之间的错误传播对最终精度是有害的。因此,着眼于减少误差传播,我们提出了一种仅使用两个深度网络就能执行ALPR的技术,第一个深度网络执行车牌检测(LPD),第二个深度网络执行车牌识别(LPR)。后者不执行显式字符分割,这大大减少了错误的传播。由于这些深度网络需要大量的样本来收敛,我们开发了新的数据增强技术,使它们能够充分发挥其潜力,并开发了一个新的数据集来训练和评估ALPR方法。根据实验结果,我们的方法能够在SSIG-SegPlate数据集中获得最先进的结果,与最佳基线相比,提高了1.4个百分点。此外,即使在同一帧中存在许多片的情况下,该方法也能够实时执行,与先前提出的方法相比,可以达到显着更高的帧速率。
{"title":"Real-Time Automatic License Plate Recognition through Deep Multi-Task Networks","authors":"G. Gonçalves, M. A. Diniz, Rayson Laroca, D. Menotti, W. R. Schwartz","doi":"10.1109/SIBGRAPI.2018.00021","DOIUrl":"https://doi.org/10.1109/SIBGRAPI.2018.00021","url":null,"abstract":"With the increasing number of cameras available in the cities, video traffic analysis can provide useful insights for the transportation segment. One of such analysis is the Automatic License Plate Recognition (ALPR). Previous approaches divided this task into several cascaded subtasks, i.e., vehicle location, license plate detection, character segmentation and optical character recognition. However, since each task has its own accuracy, the error propagation between each subtask is detrimental to the final accuracy. Therefore, focusing on the reduction of error propagation, we propose a technique that is able to perform ALPR using only two deep networks, the first performs license plate detection (LPD) and the second performs license plate recognition (LPR). The latter does not execute explicit character segmentation, which reduces significantly the error propagation. As these deep networks need a large number of samples to converge, we develop new data augmentation techniques that allow them to reach their full potential as well as a new dataset to train and evaluate ALPR approaches. According to experimental results, our approach is able to achieve state-of-the-art results in the SSIG-SegPlate dataset, reaching improvements up to 1.4 percentage point when compared to the best baseline. Furthermore, the approach is also able to perform in real time even in scenarios where many plates are present at the same frame, reaching significantly higher frame rates when compared with previously proposed approaches.","PeriodicalId":208985,"journal":{"name":"2018 31st SIBGRAPI Conference on Graphics, Patterns and Images (SIBGRAPI)","volume":"23 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125955270","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 41
期刊
2018 31st SIBGRAPI Conference on Graphics, Patterns and Images (SIBGRAPI)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1