首页 > 最新文献

International Conference on Pattern Recognition Applications and Methods最新文献

英文 中文
State of Health Estimation of Lithium-ion Batteries Using Convolutional Neural Network with Impedance Nyquist Plots 基于阻抗奈奎斯特图卷积神经网络的锂离子电池健康状态估计
Pub Date : 2023-01-01 DOI: 10.5220/0011672300003411
Yichun Li, Mina Maleki, Shadi Banitaan, Ming-Jie Chen
: In order to maintain the Li-ion batteries in a safe operating state and to optimize their performance, a precise estimation of the state of health (SOH), which indicates the degradation level of the Li-ion batteries, has to be taken into consideration urgently. In this paper, we present a regression machine learning framework that combines a convolutional neural network (CNN) with the Nyquist plot of Electrochemical Impedance Spectroscopy (EIS) as features to estimate the SOH of Li-ion batteries with a considerable improvement in the accuracy of SOH estimation. The results indicate that the Nyquist plot based on EIS features provides more detailed information regarding battery aging than simple impedance values due to its ability to reflect impedance change over time. Furthermore, convolutional layers in the CNN model were more effective in extracting different levels of features and characterizing the degradation patterns of Li-ion batteries from EIS measurement data than using simple impedance values with a DNN model, as well as other traditional machine learning methods, such as Gaussian process regression (GPR) and support vector machine (SVM).
{"title":"State of Health Estimation of Lithium-ion Batteries Using Convolutional Neural Network with Impedance Nyquist Plots","authors":"Yichun Li, Mina Maleki, Shadi Banitaan, Ming-Jie Chen","doi":"10.5220/0011672300003411","DOIUrl":"https://doi.org/10.5220/0011672300003411","url":null,"abstract":": In order to maintain the Li-ion batteries in a safe operating state and to optimize their performance, a precise estimation of the state of health (SOH), which indicates the degradation level of the Li-ion batteries, has to be taken into consideration urgently. In this paper, we present a regression machine learning framework that combines a convolutional neural network (CNN) with the Nyquist plot of Electrochemical Impedance Spectroscopy (EIS) as features to estimate the SOH of Li-ion batteries with a considerable improvement in the accuracy of SOH estimation. The results indicate that the Nyquist plot based on EIS features provides more detailed information regarding battery aging than simple impedance values due to its ability to reflect impedance change over time. Furthermore, convolutional layers in the CNN model were more effective in extracting different levels of features and characterizing the degradation patterns of Li-ion batteries from EIS measurement data than using simple impedance values with a DNN model, as well as other traditional machine learning methods, such as Gaussian process regression (GPR) and support vector machine (SVM).","PeriodicalId":410036,"journal":{"name":"International Conference on Pattern Recognition Applications and Methods","volume":"91 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128013382","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Fake it, Mix it, Segment it: Bridging the Domain Gap Between Lidar Sensors 伪造它,混合它,分割它:弥合激光雷达传感器之间的域差距
Pub Date : 2022-12-19 DOI: 10.48550/arXiv.2212.09517
Frederik Hasecke, P. Colling, A. Kummert
Segmentation of lidar data is a task that provides rich, point-wise information about the environment of robots or autonomous vehicles. Currently best performing neural networks for lidar segmentation are fine-tuned to specific datasets. Switching the lidar sensor without retraining on a big set of annotated data from the new sensor creates a domain shift, which causes the network performance to drop drastically. In this work we propose a new method for lidar domain adaption, in which we use annotated panoptic lidar datasets and recreate the recorded scenes in the structure of a different lidar sensor. We narrow the domain gap to the target data by recreating panoptic data from one domain in another and mixing the generated data with parts of (pseudo) labeled target domain data. Our method improves the nuScenes to SemanticKITTI unsupervised domain adaptation performance by 15.2 mean Intersection over Union points (mIoU) and by 48.3 mIoU in our semi-supervised approach. We demonstrate a similar improvement for the SemanticKITTI to nuScenes domain adaptation by 21.8 mIoU and 51.5 mIoU, respectively. We compare our method with two state of the art approaches for semantic lidar segmentation domain adaptation with a significant improvement for unsupervised and semi-supervised domain adaptation. Furthermore we successfully apply our proposed method to two entirely unlabeled datasets of two state of the art lidar sensors Velodyne Alpha Prime and InnovizTwo, and train well performing semantic segmentation networks for both.
激光雷达数据的分割是一项任务,可以提供关于机器人或自动驾驶汽车环境的丰富的、逐点的信息。目前表现最好的激光雷达分割神经网络是针对特定的数据集进行微调的。切换激光雷达传感器而不重新训练来自新传感器的大量注释数据会产生域移位,从而导致网络性能急剧下降。在这项工作中,我们提出了一种新的激光雷达域自适应方法,在该方法中,我们使用带注释的全光激光雷达数据集,并在不同的激光雷达传感器结构中重建记录的场景。我们通过从一个领域在另一个领域重新创建全景数据,并将生成的数据与部分(伪)标记的目标领域数据混合,来缩小与目标数据的领域差距。我们的方法将nuScenes到SemanticKITTI的无监督域自适应性能提高了15.2个平均相交点(Intersection over Union points, mIoU),而我们的半监督方法将nuScenes到SemanticKITTI的无监督域自适应性能提高了48.3个mIoU。我们展示了SemanticKITTI对nuScenes域的适应性的类似改进,分别提高了21.8 mIoU和51.5 mIoU。我们将我们的方法与语义激光雷达分割域自适应的两种最新方法进行了比较,在无监督和半监督域自适应方面有了显着改进。此外,我们成功地将我们提出的方法应用于两种最先进的激光雷达传感器Velodyne Alpha Prime和InnovizTwo的完全未标记数据集,并为两者训练了性能良好的语义分割网络。
{"title":"Fake it, Mix it, Segment it: Bridging the Domain Gap Between Lidar Sensors","authors":"Frederik Hasecke, P. Colling, A. Kummert","doi":"10.48550/arXiv.2212.09517","DOIUrl":"https://doi.org/10.48550/arXiv.2212.09517","url":null,"abstract":"Segmentation of lidar data is a task that provides rich, point-wise information about the environment of robots or autonomous vehicles. Currently best performing neural networks for lidar segmentation are fine-tuned to specific datasets. Switching the lidar sensor without retraining on a big set of annotated data from the new sensor creates a domain shift, which causes the network performance to drop drastically. In this work we propose a new method for lidar domain adaption, in which we use annotated panoptic lidar datasets and recreate the recorded scenes in the structure of a different lidar sensor. We narrow the domain gap to the target data by recreating panoptic data from one domain in another and mixing the generated data with parts of (pseudo) labeled target domain data. Our method improves the nuScenes to SemanticKITTI unsupervised domain adaptation performance by 15.2 mean Intersection over Union points (mIoU) and by 48.3 mIoU in our semi-supervised approach. We demonstrate a similar improvement for the SemanticKITTI to nuScenes domain adaptation by 21.8 mIoU and 51.5 mIoU, respectively. We compare our method with two state of the art approaches for semantic lidar segmentation domain adaptation with a significant improvement for unsupervised and semi-supervised domain adaptation. Furthermore we successfully apply our proposed method to two entirely unlabeled datasets of two state of the art lidar sensors Velodyne Alpha Prime and InnovizTwo, and train well performing semantic segmentation networks for both.","PeriodicalId":410036,"journal":{"name":"International Conference on Pattern Recognition Applications and Methods","volume":"6 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-12-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124423040","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Multi-task Fusion for Efficient Panoptic-Part Segmentation 基于多任务融合的高效全景分割
Pub Date : 2022-12-15 DOI: 10.48550/arXiv.2212.07671
Sravan Kumar Jagadeesh, René Schuster, D. Stricker
In this paper, we introduce a novel network that generates semantic, instance, and part segmentation using a shared encoder and effectively fuses them to achieve panoptic-part segmentation. Unifying these three segmentation problems allows for mutually improved and consistent representation learning. To fuse the predictions of all three heads efficiently, we introduce a parameter-free joint fusion module that dynamically balances the logits and fuses them to create panoptic-part segmentation. Our method is evaluated on the Cityscapes Panoptic Parts (CPP) and Pascal Panoptic Parts (PPP) datasets. For CPP, the PartPQ of our proposed model with joint fusion surpasses the previous state-of-the-art by 1.6 and 4.7 percentage points for all areas and segments with parts, respectively. On PPP, our joint fusion outperforms a model using the previous top-down merging strategy by 3.3 percentage points in PartPQ and 10.5 percentage points in PartPQ for partitionable classes.
在本文中,我们介绍了一种新的网络,该网络使用共享编码器生成语义、实例和部分分割,并有效地融合它们以实现全景部分分割。统一这三个分割问题允许相互改进和一致的表示学习。为了有效地融合所有三个头部的预测,我们引入了一个无参数的关节融合模块,该模块动态平衡逻辑并融合它们以创建全景部分分割。我们的方法在cityscape Panoptic Parts (CPP)和Pascal Panoptic Parts (PPP)数据集上进行了评估。对于CPP,我们提出的具有关节融合的模型的PartPQ在所有有零件的区域和部分上分别超过了以前最先进的1.6和4.7个百分点。在PPP上,我们的联合融合在PartPQ上比使用以前的自顶向下合并策略的模型高出3.3个百分点,在可分区类上比PartPQ高出10.5个百分点。
{"title":"Multi-task Fusion for Efficient Panoptic-Part Segmentation","authors":"Sravan Kumar Jagadeesh, René Schuster, D. Stricker","doi":"10.48550/arXiv.2212.07671","DOIUrl":"https://doi.org/10.48550/arXiv.2212.07671","url":null,"abstract":"In this paper, we introduce a novel network that generates semantic, instance, and part segmentation using a shared encoder and effectively fuses them to achieve panoptic-part segmentation. Unifying these three segmentation problems allows for mutually improved and consistent representation learning. To fuse the predictions of all three heads efficiently, we introduce a parameter-free joint fusion module that dynamically balances the logits and fuses them to create panoptic-part segmentation. Our method is evaluated on the Cityscapes Panoptic Parts (CPP) and Pascal Panoptic Parts (PPP) datasets. For CPP, the PartPQ of our proposed model with joint fusion surpasses the previous state-of-the-art by 1.6 and 4.7 percentage points for all areas and segments with parts, respectively. On PPP, our joint fusion outperforms a model using the previous top-down merging strategy by 3.3 percentage points in PartPQ and 10.5 percentage points in PartPQ for partitionable classes.","PeriodicalId":410036,"journal":{"name":"International Conference on Pattern Recognition Applications and Methods","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-12-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123938766","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Image-Based Fire Detection in Industrial Environments with YOLOv4 基于YOLOv4的工业环境中基于图像的火灾检测
Pub Date : 2022-12-09 DOI: 10.48550/arXiv.2212.04786
O. Zell, Joel Pålsson, Kevin Hernandez-Diaz, F. Alonso-Fernandez, Felix Nilsson
Fires have destructive power when they break out and affect their surroundings on a devastatingly large scale. The best way to minimize their damage is to detect the fire as quickly as possible before it has a chance to grow. Accordingly, this work looks into the potential of AI to detect and recognize fires and reduce detection time using object detection on an image stream. Object detection has made giant leaps in speed and accuracy over the last six years, making real-time detection feasible. To our end, we collected and labeled appropriate data from several public sources, which have been used to train and evaluate several models based on the popular YOLOv4 object detector. Our focus, driven by a collaborating industrial partner, is to implement our system in an industrial warehouse setting, which is characterized by high ceilings. A drawback of traditional smoke detectors in this setup is that the smoke has to rise to a sufficient height. The AI models brought forward in this research managed to outperform these detectors by a significant amount of time, providing precious anticipation that could help to minimize the effects of fires further.
当火灾爆发并对周围环境造成毁灭性的大范围影响时,它们具有破坏性。尽量减少损失的最好方法是在火势有机会扩大之前尽快发现。因此,这项工作研究了人工智能在检测和识别火灾方面的潜力,并利用图像流上的物体检测来缩短检测时间。在过去的六年里,物体检测在速度和准确性上取得了巨大的飞跃,使得实时检测成为可能。最后,我们从几个公共来源收集并标记了适当的数据,这些数据已用于训练和评估基于流行的YOLOv4对象检测器的几个模型。在工业合作伙伴的推动下,我们的重点是在工业仓库环境中实施我们的系统,其特点是天花板很高。在这种情况下,传统烟雾探测器的缺点是烟雾必须上升到足够的高度。本研究中提出的人工智能模型在相当长的时间内成功地超越了这些探测器,提供了宝贵的预测,可以帮助进一步减少火灾的影响。
{"title":"Image-Based Fire Detection in Industrial Environments with YOLOv4","authors":"O. Zell, Joel Pålsson, Kevin Hernandez-Diaz, F. Alonso-Fernandez, Felix Nilsson","doi":"10.48550/arXiv.2212.04786","DOIUrl":"https://doi.org/10.48550/arXiv.2212.04786","url":null,"abstract":"Fires have destructive power when they break out and affect their surroundings on a devastatingly large scale. The best way to minimize their damage is to detect the fire as quickly as possible before it has a chance to grow. Accordingly, this work looks into the potential of AI to detect and recognize fires and reduce detection time using object detection on an image stream. Object detection has made giant leaps in speed and accuracy over the last six years, making real-time detection feasible. To our end, we collected and labeled appropriate data from several public sources, which have been used to train and evaluate several models based on the popular YOLOv4 object detector. Our focus, driven by a collaborating industrial partner, is to implement our system in an industrial warehouse setting, which is characterized by high ceilings. A drawback of traditional smoke detectors in this setup is that the smoke has to rise to a sufficient height. The AI models brought forward in this research managed to outperform these detectors by a significant amount of time, providing precious anticipation that could help to minimize the effects of fires further.","PeriodicalId":410036,"journal":{"name":"International Conference on Pattern Recognition Applications and Methods","volume":"36 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-12-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130697742","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Synthetic Data for Object Classification in Industrial Applications 工业应用中对象分类的合成数据
Pub Date : 2022-12-09 DOI: 10.48550/arXiv.2212.04790
August Baaz, Yonan Yonan, Kevin Hernandez-Diaz, F. Alonso-Fernandez, Felix Nilsson
One of the biggest challenges in machine learning is data collection. Training data is an important part since it determines how the model will behave. In object classification, capturing a large number of images per object and in different conditions is not always possible and can be very time-consuming and tedious. Accordingly, this work explores the creation of artificial images using a game engine to cope with limited data in the training dataset. We combine real and synthetic data to train the object classification engine, a strategy that has shown to be beneficial to increase confidence in the decisions made by the classifier, which is often critical in industrial setups. To combine real and synthetic data, we first train the classifier on a massive amount of synthetic data, and then we fine-tune it on real images. Another important result is that the amount of real images needed for fine-tuning is not very high, reaching top accuracy with just 12 or 24 images per class. This substantially reduces the requirements of capturing a great amount of real data.
机器学习中最大的挑战之一是数据收集。训练数据是一个重要的部分,因为它决定了模型的行为方式。在对象分类中,在不同条件下捕获每个对象的大量图像并不总是可能的,并且可能非常耗时和繁琐。因此,这项工作探索了使用游戏引擎创建人工图像来处理训练数据集中有限的数据。我们将真实数据和合成数据结合起来训练对象分类引擎,这一策略已被证明有利于提高分类器做出决策的信心,这在工业设置中通常是至关重要的。为了结合真实数据和合成数据,我们首先在大量的合成数据上训练分类器,然后在真实图像上对其进行微调。另一个重要的结果是,微调所需的真实图像数量不是很高,每个类只需12或24张图像即可达到最高精度。这大大减少了捕获大量真实数据的需求。
{"title":"Synthetic Data for Object Classification in Industrial Applications","authors":"August Baaz, Yonan Yonan, Kevin Hernandez-Diaz, F. Alonso-Fernandez, Felix Nilsson","doi":"10.48550/arXiv.2212.04790","DOIUrl":"https://doi.org/10.48550/arXiv.2212.04790","url":null,"abstract":"One of the biggest challenges in machine learning is data collection. Training data is an important part since it determines how the model will behave. In object classification, capturing a large number of images per object and in different conditions is not always possible and can be very time-consuming and tedious. Accordingly, this work explores the creation of artificial images using a game engine to cope with limited data in the training dataset. We combine real and synthetic data to train the object classification engine, a strategy that has shown to be beneficial to increase confidence in the decisions made by the classifier, which is often critical in industrial setups. To combine real and synthetic data, we first train the classifier on a massive amount of synthetic data, and then we fine-tune it on real images. Another important result is that the amount of real images needed for fine-tuning is not very high, reaching top accuracy with just 12 or 24 images per class. This substantially reduces the requirements of capturing a great amount of real data.","PeriodicalId":410036,"journal":{"name":"International Conference on Pattern Recognition Applications and Methods","volume":"95 5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-12-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129120383","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Deep Learning for Diagonal Earlobe Crease Detection 对角耳垂折痕检测的深度学习
Pub Date : 2022-10-20 DOI: 10.5220/0011644400003411
Sara L. Almonacid-Uribe, Oliverio J. Santana, D. Hernández-Sosa, David Freire-Obregón
An article published on Medical News Today in June 2022 presented a fundamental question in its title: Can an earlobe crease predict heart attacks? The author explained that end arteries supply the heart and ears. In other words, if they lose blood supply, no other arteries can take over, resulting in tissue damage. Consequently, some earlobes have a diagonal crease, line, or deep fold that resembles a wrinkle. In this paper, we take a step toward detecting this specific marker, commonly known as DELC or Frank's Sign. For this reason, we have made the first DELC dataset available to the public. In addition, we have investigated the performance of numerous cutting-edge backbones on annotated photos. Experimentally, we demonstrate that it is possible to solve this challenge by combining pre-trained encoders with a customized classifier to achieve 97.7% accuracy. Moreover, we have analyzed the backbone trade-off between performance and size, estimating MobileNet as the most promising encoder.
2022年6月发表在《今日医学新闻》上的一篇文章在标题中提出了一个基本问题:耳垂褶皱能预测心脏病发作吗?作者解释说,末梢动脉供应心脏和耳朵。换句话说,如果它们失去血液供应,没有其他动脉可以接管,导致组织损伤。因此,一些耳垂有对角线折痕,线,或深褶皱,类似于皱纹。在本文中,我们朝着检测这种特定标记迈出了一步,通常称为DELC或弗兰克标志。出于这个原因,我们向公众提供了第一个DELC数据集。此外,我们还研究了许多尖端骨干在注释照片上的性能。通过实验,我们证明可以通过将预训练的编码器与定制的分类器相结合来解决这一挑战,达到97.7%的准确率。此外,我们分析了性能和大小之间的主干权衡,估计MobileNet是最有前途的编码器。
{"title":"Deep Learning for Diagonal Earlobe Crease Detection","authors":"Sara L. Almonacid-Uribe, Oliverio J. Santana, D. Hernández-Sosa, David Freire-Obregón","doi":"10.5220/0011644400003411","DOIUrl":"https://doi.org/10.5220/0011644400003411","url":null,"abstract":"An article published on Medical News Today in June 2022 presented a fundamental question in its title: Can an earlobe crease predict heart attacks? The author explained that end arteries supply the heart and ears. In other words, if they lose blood supply, no other arteries can take over, resulting in tissue damage. Consequently, some earlobes have a diagonal crease, line, or deep fold that resembles a wrinkle. In this paper, we take a step toward detecting this specific marker, commonly known as DELC or Frank's Sign. For this reason, we have made the first DELC dataset available to the public. In addition, we have investigated the performance of numerous cutting-edge backbones on annotated photos. Experimentally, we demonstrate that it is possible to solve this challenge by combining pre-trained encoders with a customized classifier to achieve 97.7% accuracy. Moreover, we have analyzed the backbone trade-off between performance and size, estimating MobileNet as the most promising encoder.","PeriodicalId":410036,"journal":{"name":"International Conference on Pattern Recognition Applications and Methods","volume":"68 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-10-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131633194","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
"Why Here and Not There?" - Diverse Contrasting Explanations of Dimensionality Reduction “为什么是这里而不是那里?”-对降维的不同对比解释
Pub Date : 2022-06-15 DOI: 10.48550/arXiv.2206.07391
André Artelt, Alexander Schulz, Barbara Hammer
Dimensionality reduction is a popular preprocessing and a widely used tool in data mining. Transparency, which is usually achieved by means of explanations, is nowadays a widely accepted and crucial requirement of machine learning based systems like classifiers and recommender systems. However, transparency of dimensionality reduction and other data mining tools have not been considered in much depth yet, still it is crucial to understand their behavior -- in particular practitioners might want to understand why a specific sample got mapped to a specific location. In order to (locally) understand the behavior of a given dimensionality reduction method, we introduce the abstract concept of contrasting explanations for dimensionality reduction, and apply a realization of this concept to the specific application of explaining two dimensional data visualization.
降维是一种流行的预处理方法,也是数据挖掘中广泛使用的工具。透明度通常是通过解释来实现的,这是现在基于机器学习的系统(如分类器和推荐系统)广泛接受的关键要求。然而,降维的透明度和其他数据挖掘工具还没有被深入考虑,但理解它们的行为仍然是至关重要的——特别是从业者可能想要理解为什么一个特定的样本被映射到一个特定的位置。为了(局部地)理解给定降维方法的行为,我们引入了降维的对比解释的抽象概念,并将这一概念的实现应用于解释二维数据可视化的具体应用。
{"title":"\"Why Here and Not There?\" - Diverse Contrasting Explanations of Dimensionality Reduction","authors":"André Artelt, Alexander Schulz, Barbara Hammer","doi":"10.48550/arXiv.2206.07391","DOIUrl":"https://doi.org/10.48550/arXiv.2206.07391","url":null,"abstract":"Dimensionality reduction is a popular preprocessing and a widely used tool in data mining. Transparency, which is usually achieved by means of explanations, is nowadays a widely accepted and crucial requirement of machine learning based systems like classifiers and recommender systems. However, transparency of dimensionality reduction and other data mining tools have not been considered in much depth yet, still it is crucial to understand their behavior -- in particular practitioners might want to understand why a specific sample got mapped to a specific location. In order to (locally) understand the behavior of a given dimensionality reduction method, we introduce the abstract concept of contrasting explanations for dimensionality reduction, and apply a realization of this concept to the specific application of explaining two dimensional data visualization.","PeriodicalId":410036,"journal":{"name":"International Conference on Pattern Recognition Applications and Methods","volume":"47 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130868765","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Image-based material analysis of ancient historical documents 基于图像的古代历史文献材料分析
Pub Date : 2022-03-02 DOI: 10.5220/0011743700003411
Thomas Reynolds, Maruf A. Dhali, Lambert Schomaker
Researchers continually perform corroborative tests to classify ancient historical documents based on the physical materials of their writing surfaces. However, these tests, often performed on-site, requires actual access to the manuscript objects. The procedures involve a considerable amount of time and cost, and can damage the manuscripts. Developing a technique to classify such documents using only digital images can be very useful and efficient. In order to tackle this problem, this study uses images of a famous historical collection, the Dead Sea Scrolls, to propose a novel method to classify the materials of the manuscripts. The proposed classifier uses the two-dimensional Fourier Transform to identify patterns within the manuscript surfaces. Combining a binary classification system employing the transform with a majority voting process is shown to be effective for this classification task. This pilot study shows a successful classification percentage of up to 97% for a confined amount of manuscripts produced from either parchment or papyrus material. Feature vectors based on Fourier-space grid representation outperformed a concentric Fourier-space format.
研究人员不断进行确证测试,根据古代历史文献的书写表面的物理材料对其进行分类。然而,这些通常在现场进行的测试需要实际访问手稿对象。这个过程需要相当多的时间和成本,而且可能会损坏手稿。开发一种仅使用数字图像对此类文档进行分类的技术是非常有用和有效的。为了解决这一问题,本研究利用著名的历史收藏品死海古卷的图像,提出了一种新的手稿材料分类方法。提出的分类器使用二维傅里叶变换来识别手稿表面内的模式。将采用该变换的二元分类系统与多数投票过程相结合,对该分类任务是有效的。这项试点研究表明,对于羊皮纸或纸莎草材料制成的有限数量的手稿,成功分类的百分比高达97%。基于傅里叶空间网格表示的特征向量优于同心傅里叶空间格式。
{"title":"Image-based material analysis of ancient historical documents","authors":"Thomas Reynolds, Maruf A. Dhali, Lambert Schomaker","doi":"10.5220/0011743700003411","DOIUrl":"https://doi.org/10.5220/0011743700003411","url":null,"abstract":"Researchers continually perform corroborative tests to classify ancient historical documents based on the physical materials of their writing surfaces. However, these tests, often performed on-site, requires actual access to the manuscript objects. The procedures involve a considerable amount of time and cost, and can damage the manuscripts. Developing a technique to classify such documents using only digital images can be very useful and efficient. In order to tackle this problem, this study uses images of a famous historical collection, the Dead Sea Scrolls, to propose a novel method to classify the materials of the manuscripts. The proposed classifier uses the two-dimensional Fourier Transform to identify patterns within the manuscript surfaces. Combining a binary classification system employing the transform with a majority voting process is shown to be effective for this classification task. This pilot study shows a successful classification percentage of up to 97% for a confined amount of manuscripts produced from either parchment or papyrus material. Feature vectors based on Fourier-space grid representation outperformed a concentric Fourier-space format.","PeriodicalId":410036,"journal":{"name":"International Conference on Pattern Recognition Applications and Methods","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-03-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126182607","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
The influence of labeling techniques in classifying human manipulation movement of different speed 标记技术对不同速度人体操作动作分类的影响
Pub Date : 2022-02-04 DOI: 10.5220/0010906900003122
Sadique Adnan Siddiqui, L. Gutzeit, F. Kirchner
In this work, we investigate the influence of labeling methods on the classification of human movements on data recorded using a marker-based motion capture system. The dataset is labeled using two different approaches, one based on video data of the movements, the other based on the movement trajectories recorded using the motion capture system. The dataset is labeled using two different approaches, one based on video data of the movements, the other based on the movement trajectories recorded using the motion capture system. The data was recorded from one participant performing a stacking scenario comprising simple arm movements at three different speeds (slow, normal, fast). Machine learning algorithms that include k-Nearest Neighbor, Random Forest, Extreme Gradient Boosting classifier, Convolutional Neural networks (CNN), Long Short-Term Memory networks (LSTM), and a combination of CNN-LSTM networks are compared on their performance in recognition of these arm movements. The models were trained on actions performed on slow and normal speed movements segments and generalized on actions consisting of fast-paced human movement. It was observed that all the models trained on normal-paced data labeled using trajectories have almost 20% improvement in accuracy on test data in comparison to the models trained on data labeled using videos of the performed experiments.
在这项工作中,我们研究了标记方法对使用基于标记的运动捕捉系统记录的人类运动分类的影响。数据集使用两种不同的方法进行标记,一种基于运动的视频数据,另一种基于运动捕捉系统记录的运动轨迹。数据集使用两种不同的方法进行标记,一种基于运动的视频数据,另一种基于运动捕捉系统记录的运动轨迹。数据记录于一名参与者以三种不同的速度(慢、正常、快速)进行简单的手臂运动的堆叠场景。机器学习算法包括k-最近邻、随机森林、极端梯度增强分类器、卷积神经网络(CNN)、长短期记忆网络(LSTM)以及CNN-LSTM网络的组合,比较了它们在识别这些手臂运动方面的性能。这些模型在慢速和正常速度的动作片段上进行训练,并在由快节奏的人体动作组成的动作上进行推广。我们观察到,与使用实验视频标记的数据训练的模型相比,使用轨迹标记的正常节奏数据训练的所有模型在测试数据上的准确性提高了近20%。
{"title":"The influence of labeling techniques in classifying human manipulation movement of different speed","authors":"Sadique Adnan Siddiqui, L. Gutzeit, F. Kirchner","doi":"10.5220/0010906900003122","DOIUrl":"https://doi.org/10.5220/0010906900003122","url":null,"abstract":"In this work, we investigate the influence of labeling methods on the classification of human movements on data recorded using a marker-based motion capture system. The dataset is labeled using two different approaches, one based on video data of the movements, the other based on the movement trajectories recorded using the motion capture system. The dataset is labeled using two different approaches, one based on video data of the movements, the other based on the movement trajectories recorded using the motion capture system. The data was recorded from one participant performing a stacking scenario comprising simple arm movements at three different speeds (slow, normal, fast). Machine learning algorithms that include k-Nearest Neighbor, Random Forest, Extreme Gradient Boosting classifier, Convolutional Neural networks (CNN), Long Short-Term Memory networks (LSTM), and a combination of CNN-LSTM networks are compared on their performance in recognition of these arm movements. The models were trained on actions performed on slow and normal speed movements segments and generalized on actions consisting of fast-paced human movement. It was observed that all the models trained on normal-paced data labeled using trajectories have almost 20% improvement in accuracy on test data in comparison to the models trained on data labeled using videos of the performed experiments.","PeriodicalId":410036,"journal":{"name":"International Conference on Pattern Recognition Applications and Methods","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-02-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125450838","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An integrated recurrent neural network and regression model with spatial and climatic couplings for vector-borne disease dynamics 基于空间和气候耦合的媒介传播疾病动力学集成递归神经网络和回归模型
Pub Date : 2022-01-23 DOI: 10.5220/0010762700003122
Zhijian Li, J. Xin, Guofa Zhou
We developed an integrated recurrent neural network and nonlinear regression spatio-temporal model for vector-borne disease evolution. We take into account climate data and seasonality as external factors that correlate with disease transmitting insects (e.g. flies), also spill-over infections from neighboring regions surrounding a region of interest. The climate data is encoded to the model through a quadratic embedding scheme motivated by recommendation systems. The neighboring regions' influence is modeled by a long short-term memory neural network. The integrated model is trained by stochastic gradient descent and tested on leish-maniasis data in Sri Lanka from 2013-2018 where infection outbreaks occurred. Our model outperformed ARIMA models across a number of regions with high infections, and an associated ablation study renders support to our modeling hypothesis and ideas.
我们开发了一个集成的递归神经网络和非线性回归时空模型,用于媒介传播疾病的进化。我们将气候数据和季节性作为与疾病传播昆虫(如苍蝇)相关的外部因素,以及来自感兴趣区域周围邻近地区的溢出感染考虑在内。气候数据通过推荐系统驱动的二次嵌入方案编码到模型中。通过长短期记忆神经网络对相邻区域的影响进行建模。该综合模型采用随机梯度下降法进行训练,并在斯里兰卡2013-2018年暴发感染的利什曼病数据上进行了测试。我们的模型在许多高感染地区的表现优于ARIMA模型,相关的消融研究支持了我们的建模假设和想法。
{"title":"An integrated recurrent neural network and regression model with spatial and climatic couplings for vector-borne disease dynamics","authors":"Zhijian Li, J. Xin, Guofa Zhou","doi":"10.5220/0010762700003122","DOIUrl":"https://doi.org/10.5220/0010762700003122","url":null,"abstract":"We developed an integrated recurrent neural network and nonlinear regression spatio-temporal model for vector-borne disease evolution. We take into account climate data and seasonality as external factors that correlate with disease transmitting insects (e.g. flies), also spill-over infections from neighboring regions surrounding a region of interest. The climate data is encoded to the model through a quadratic embedding scheme motivated by recommendation systems. The neighboring regions' influence is modeled by a long short-term memory neural network. The integrated model is trained by stochastic gradient descent and tested on leish-maniasis data in Sri Lanka from 2013-2018 where infection outbreaks occurred. Our model outperformed ARIMA models across a number of regions with high infections, and an associated ablation study renders support to our modeling hypothesis and ideas.","PeriodicalId":410036,"journal":{"name":"International Conference on Pattern Recognition Applications and Methods","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-01-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134531222","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
International Conference on Pattern Recognition Applications and Methods
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1