首页 > 最新文献

2022 21st IEEE International Conference on Machine Learning and Applications (ICMLA)最新文献

英文 中文
Recurrent Nerual Imaging: An Evolutionary Approach for Mixed Possion-Gaussian Image Denoising 递归神经成像:混合波塞-高斯图像去噪的进化方法
Pub Date : 2022-12-01 DOI: 10.1109/ICMLA55696.2022.00078
A. Ranganath, Omar DeGuchy, Fabian Santiago, Mukesh Singhal, Roummel F. Marcia
Recurrent neural networks (RNNs) are traditionally used for machine learning applications for temporal sequences such as natural language processing. Its application to image processing is relatively new. In this paper, we apply RNNs to denoise images corrupted by mixed Poisson and Gaussian noise. The motivation for using an RNN comes from viewing the denoising of the Poisson-Gaussian realization as a temporal process. The network then attempts to trace back the steps that create the noisy realization in order to arrive at the noiseless reconstruction. Numerical experiments demonstrate that our proposed RNN approach outperforms convolutional autoen-coder methods for denoising and upsampling low-resolution images from the CIFAR-10 dataset.
递归神经网络(rnn)传统上用于时间序列的机器学习应用,如自然语言处理。它在图像处理中的应用相对较新。在本文中,我们应用rnn去噪被混合泊松和高斯噪声破坏的图像。使用RNN的动机来自于将泊松-高斯实现的去噪视为一个时间过程。然后,网络试图追溯产生噪声实现的步骤,以达到无噪声重构。数值实验表明,我们提出的RNN方法在去噪和上采样CIFAR-10数据集中的低分辨率图像方面优于卷积自动编码方法。
{"title":"Recurrent Nerual Imaging: An Evolutionary Approach for Mixed Possion-Gaussian Image Denoising","authors":"A. Ranganath, Omar DeGuchy, Fabian Santiago, Mukesh Singhal, Roummel F. Marcia","doi":"10.1109/ICMLA55696.2022.00078","DOIUrl":"https://doi.org/10.1109/ICMLA55696.2022.00078","url":null,"abstract":"Recurrent neural networks (RNNs) are traditionally used for machine learning applications for temporal sequences such as natural language processing. Its application to image processing is relatively new. In this paper, we apply RNNs to denoise images corrupted by mixed Poisson and Gaussian noise. The motivation for using an RNN comes from viewing the denoising of the Poisson-Gaussian realization as a temporal process. The network then attempts to trace back the steps that create the noisy realization in order to arrive at the noiseless reconstruction. Numerical experiments demonstrate that our proposed RNN approach outperforms convolutional autoen-coder methods for denoising and upsampling low-resolution images from the CIFAR-10 dataset.","PeriodicalId":128160,"journal":{"name":"2022 21st IEEE International Conference on Machine Learning and Applications (ICMLA)","volume":"75 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133036413","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
On the Trade-off Between Benefit and Contribution for Clients in Federated Learning in Healthcare 医疗保健联合学习中客户利益与贡献的权衡
Pub Date : 2022-12-01 DOI: 10.1109/ICMLA55696.2022.00257
Christoph Düsing, P. Cimiano
Federated Learning (FL) is a learning paradigm that allows clients to profit from the data that is available across multiple clients to train a joint model. As FL allows to train such a joint model without explicitly sharing data, but only sharing model updates, it has attained popularity in healthcare settings where patient data is subject to strict privacy policies and needs to be locally stored at each hospital or healthcare provider. A particular challenge for FL settings is data imbalance across clients, as it has been found to be detrimental to model performance and impact the influence of each client on the learning process. Unfortunately, the healthcare domain is particularly prone to such imbalanced data due to regional differences in disease management, prescription behavior etc. In this paper, we introduce the two novel metrics Benefit and Contribution to quantify to which degree individual clients benefit from participation in FL and how they contribute to its success, respectively. Therefore, we measure Benefit and Contribution with respect to four types of imbalances present in data at each client side. Our results show that both client Benefit and Contribution are influenced by data imbalance in such a way that high imbalance in data quantity, label distribution and feature distribution reduces or nullifies clients’ Benefit while increasing their Contribution. Thus, the most valuable clients within a cohort benefit the least from their participation, exposing a critical thread to the success of clinical FL cohorts by withdrawing participation.
联邦学习(FL)是一种学习范例,它允许客户从跨多个客户端可用的数据中获利,以训练联合模型。由于FL允许在不显式共享数据的情况下训练这样的联合模型,而只共享模型更新,因此它在患者数据受严格隐私政策约束并且需要在每个医院或医疗保健提供商本地存储的医疗保健环境中得到了普及。FL设置的一个特殊挑战是客户机之间的数据不平衡,因为它已被发现对模型性能有害,并影响每个客户机对学习过程的影响。不幸的是,由于疾病管理、处方行为等方面的地区差异,医疗保健领域特别容易出现这种数据不平衡。在本文中,我们引入了两个新的指标,分别量化个人客户从参与FL中受益的程度以及他们如何为其成功做出贡献。因此,我们根据每个客户端数据中存在的四种失衡类型来衡量收益和贡献。我们的研究结果表明,客户的利益和贡献都受到数据不平衡的影响,数据数量、标签分布和特征分布的高度不平衡会降低或抵消客户的利益,同时增加客户的贡献。因此,队列中最有价值的客户从他们的参与中获益最少,通过退出参与暴露了临床FL队列成功的关键线索。
{"title":"On the Trade-off Between Benefit and Contribution for Clients in Federated Learning in Healthcare","authors":"Christoph Düsing, P. Cimiano","doi":"10.1109/ICMLA55696.2022.00257","DOIUrl":"https://doi.org/10.1109/ICMLA55696.2022.00257","url":null,"abstract":"Federated Learning (FL) is a learning paradigm that allows clients to profit from the data that is available across multiple clients to train a joint model. As FL allows to train such a joint model without explicitly sharing data, but only sharing model updates, it has attained popularity in healthcare settings where patient data is subject to strict privacy policies and needs to be locally stored at each hospital or healthcare provider. A particular challenge for FL settings is data imbalance across clients, as it has been found to be detrimental to model performance and impact the influence of each client on the learning process. Unfortunately, the healthcare domain is particularly prone to such imbalanced data due to regional differences in disease management, prescription behavior etc. In this paper, we introduce the two novel metrics Benefit and Contribution to quantify to which degree individual clients benefit from participation in FL and how they contribute to its success, respectively. Therefore, we measure Benefit and Contribution with respect to four types of imbalances present in data at each client side. Our results show that both client Benefit and Contribution are influenced by data imbalance in such a way that high imbalance in data quantity, label distribution and feature distribution reduces or nullifies clients’ Benefit while increasing their Contribution. Thus, the most valuable clients within a cohort benefit the least from their participation, exposing a critical thread to the success of clinical FL cohorts by withdrawing participation.","PeriodicalId":128160,"journal":{"name":"2022 21st IEEE International Conference on Machine Learning and Applications (ICMLA)","volume":"263 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116150679","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
A Layer Decomposition Approach to Inference Time Prediction of Deep Learning Architectures 深度学习体系结构推理时间预测的层分解方法
Pub Date : 2022-12-01 DOI: 10.1109/ICMLA55696.2022.00141
Ola Mustafa Alqahtani, Lakshmish Ramaswamy
In recent years, deep learning models have been widely adopted in lots of fields. such as computer vision, pattern recognition, and classification problems like plant disease classification. Due to the large diversity among the computing devices that these models may run on, we need to choose between the appropriate device based on cost and performance. Furthermore, finding the suitable optimal device for a given project is a complex process that needs significant time and resources. Prediction of inference latency DNN models is necessary for many tasks where measuring the latency on real devices is either infeasible or too costly. This is a very challenging problem, and most existing approaches fail to achieve high accuracy of prediction. While some research has been carried out to predict the inference time of DNN models – most existing techniques assume that training time is linearly related to the number of floating-point operations. This paper designs and develops a framework to predict the inference time for deep learning models and is generic to be easily extended for a large set of devices. Our key idea is decomposing a given model inference into layers and conducting layer-level prediction. Our experiments demonstrate that this strategy provides significant benefits in terms of prediction accuracy.
近年来,深度学习模型被广泛应用于许多领域。如计算机视觉、模式识别、植物病害分类等分类问题。由于这些模型可能运行的计算设备之间存在很大的差异,我们需要根据成本和性能在适当的设备之间进行选择。此外,为给定项目找到合适的最佳设备是一个复杂的过程,需要大量的时间和资源。对于在真实设备上测量延迟不可行或成本过高的许多任务,预测推理延迟DNN模型是必要的。这是一个非常具有挑战性的问题,现有的大多数方法都无法达到较高的预测精度。虽然已经进行了一些研究来预测深度神经网络模型的推理时间,但大多数现有技术都假设训练时间与浮点运算次数线性相关。本文设计并开发了一个框架来预测深度学习模型的推理时间,该框架具有通用性,易于扩展到大量设备。我们的关键思想是将给定的模型推理分解成层,并进行层级预测。我们的实验表明,这种策略在预测精度方面提供了显著的好处。
{"title":"A Layer Decomposition Approach to Inference Time Prediction of Deep Learning Architectures","authors":"Ola Mustafa Alqahtani, Lakshmish Ramaswamy","doi":"10.1109/ICMLA55696.2022.00141","DOIUrl":"https://doi.org/10.1109/ICMLA55696.2022.00141","url":null,"abstract":"In recent years, deep learning models have been widely adopted in lots of fields. such as computer vision, pattern recognition, and classification problems like plant disease classification. Due to the large diversity among the computing devices that these models may run on, we need to choose between the appropriate device based on cost and performance. Furthermore, finding the suitable optimal device for a given project is a complex process that needs significant time and resources. Prediction of inference latency DNN models is necessary for many tasks where measuring the latency on real devices is either infeasible or too costly. This is a very challenging problem, and most existing approaches fail to achieve high accuracy of prediction. While some research has been carried out to predict the inference time of DNN models – most existing techniques assume that training time is linearly related to the number of floating-point operations. This paper designs and develops a framework to predict the inference time for deep learning models and is generic to be easily extended for a large set of devices. Our key idea is decomposing a given model inference into layers and conducting layer-level prediction. Our experiments demonstrate that this strategy provides significant benefits in terms of prediction accuracy.","PeriodicalId":128160,"journal":{"name":"2022 21st IEEE International Conference on Machine Learning and Applications (ICMLA)","volume":"165 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116419465","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Application of Machine Learning Techniques in Temperature Forecast 机器学习技术在温度预报中的应用
Pub Date : 2022-12-01 DOI: 10.1109/ICMLA55696.2022.00083
Adrin Issai Arasu, M. Modani, N. R. Vadlamani
Temperature prediction is critical for many industrial and everyday applications. Numerical Weather Prediction (NWP) models using high-performance computing is the most sought technique to forecast weather, including temperature. However, NWP is complex in nature and computationally expensive. In this paper, the temperature is forecast using data-driven Machine Learning techniques, which are not computationally intensive and are further accelerated using GPUs. Two deep learning models: A stacked Long Short-Term Memory (LSTM) and Random Forest Regressor (RFR), are developed and validated using the standard ERA5 data (at 850hPa, above the atmospheric boundary layer). In addition, the models are tested against the ground-level observations (inside the atmospheric boundary layer) for twenty different locations in India. The performance of univariate and multivariate models is also analyzed for the real-time dataset. Root Mean Square Error (RMSE) obtained by the LSTM and RFR are 0.47 and 0.23, respectively, for ERA5 data. When compared to the numerical weather prediction model - operational IFS, the RMSE using LSTM and RFR is smaller by 65% and 83%, respectively. The LSTM and RFR models forecast temperature with an average RMSE of 0.7 for the real-time data at twenty locations. The GPU-enabled LSTM model performed 64 times faster than the CPU-enabled model. The developed RNN models are made publicly available at https://github.com/arasuadrian/RNN-Models.
温度预测对许多工业和日常应用至关重要。使用高性能计算的数值天气预报(NWP)模式是预测天气(包括温度)最受欢迎的技术。然而,NWP本质上是复杂的,计算成本很高。在本文中,使用数据驱动的机器学习技术来预测温度,这些技术不是计算密集型的,并且使用gpu进一步加速。两种深度学习模型:堆叠长短期记忆(LSTM)和随机森林回归(RFR),使用标准ERA5数据(850hPa,大气边界层以上)开发并验证。此外,根据印度20个不同地点的地面观测(大气边界层内)对这些模型进行了测试。针对实时数据集,分析了单变量模型和多变量模型的性能。对于ERA5数据,LSTM和RFR得到的均方根误差(RMSE)分别为0.47和0.23。与数值天气预报模式-实际IFS相比,使用LSTM和RFR的RMSE分别小65%和83%。LSTM和RFR模式预测20个地点的实时数据的平均RMSE为0.7。启用gpu的LSTM模型比启用cpu的模型执行速度快64倍。开发的RNN模型可以在https://github.com/arasuadrian/RNN-Models上公开获得。
{"title":"Application of Machine Learning Techniques in Temperature Forecast","authors":"Adrin Issai Arasu, M. Modani, N. R. Vadlamani","doi":"10.1109/ICMLA55696.2022.00083","DOIUrl":"https://doi.org/10.1109/ICMLA55696.2022.00083","url":null,"abstract":"Temperature prediction is critical for many industrial and everyday applications. Numerical Weather Prediction (NWP) models using high-performance computing is the most sought technique to forecast weather, including temperature. However, NWP is complex in nature and computationally expensive. In this paper, the temperature is forecast using data-driven Machine Learning techniques, which are not computationally intensive and are further accelerated using GPUs. Two deep learning models: A stacked Long Short-Term Memory (LSTM) and Random Forest Regressor (RFR), are developed and validated using the standard ERA5 data (at 850hPa, above the atmospheric boundary layer). In addition, the models are tested against the ground-level observations (inside the atmospheric boundary layer) for twenty different locations in India. The performance of univariate and multivariate models is also analyzed for the real-time dataset. Root Mean Square Error (RMSE) obtained by the LSTM and RFR are 0.47 and 0.23, respectively, for ERA5 data. When compared to the numerical weather prediction model - operational IFS, the RMSE using LSTM and RFR is smaller by 65% and 83%, respectively. The LSTM and RFR models forecast temperature with an average RMSE of 0.7 for the real-time data at twenty locations. The GPU-enabled LSTM model performed 64 times faster than the CPU-enabled model. The developed RNN models are made publicly available at https://github.com/arasuadrian/RNN-Models.","PeriodicalId":128160,"journal":{"name":"2022 21st IEEE International Conference on Machine Learning and Applications (ICMLA)","volume":"29 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121776108","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Deep Learning based Hand Gesture Recognition on a Low-power Microcontroller using IMU Sensors 基于IMU传感器的低功耗微控制器上的深度学习手势识别
Pub Date : 2022-12-01 DOI: 10.1109/ICMLA55696.2022.00122
Daniel Lauss, F. Eibensteiner, P. Petz
In this paper, we demonstrate an inertial measurement unit (IMU) based hand gesture recognition (HGR) on a low-power microcontroller (STM32L476JGY). The focus of this work is to build a reliable hardware prototype by using deep neural networks (DNN) deployed on a resource limited device. To train the DNNs, a dataset was recorded which contains accelerometer and gyroscope readings from three IMUs mounted on the fingertips. With this dataset, various neural networks (NN) were trained and analyzed. The best NN, in terms of accuracy, memory usage and latency, was then selected and ported to the microcontroller. Finally, a runtime analysis of the model has been performed on the controller. The analysis showed that a LSTM is best suited for the detection of hand gestures. The selected model achieves an accuracy of 93% and only takes up around 40KiB of memory. In addition, the model has a throughput time of only 3.52ms, which means that the prototype can be used in real time.
在本文中,我们展示了一种基于惯性测量单元(IMU)的手势识别(HGR)在低功耗微控制器(STM32L476JGY)上。本工作的重点是通过在资源有限的设备上部署深度神经网络(DNN)来构建可靠的硬件原型。为了训练深度神经网络,记录了一个数据集,其中包含安装在指尖上的三个imu的加速度计和陀螺仪读数。利用该数据集,对各种神经网络(NN)进行训练和分析。在准确性、内存使用和延迟方面,选择最好的神经网络并将其移植到微控制器上。最后,在控制器上对模型进行了运行时分析。分析表明LSTM最适合于手势的检测。所选择的模型达到93%的准确率,只占用大约40KiB的内存。此外,该模型的吞吐时间仅为3.52ms,这意味着该原型可以实时使用。
{"title":"A Deep Learning based Hand Gesture Recognition on a Low-power Microcontroller using IMU Sensors","authors":"Daniel Lauss, F. Eibensteiner, P. Petz","doi":"10.1109/ICMLA55696.2022.00122","DOIUrl":"https://doi.org/10.1109/ICMLA55696.2022.00122","url":null,"abstract":"In this paper, we demonstrate an inertial measurement unit (IMU) based hand gesture recognition (HGR) on a low-power microcontroller (STM32L476JGY). The focus of this work is to build a reliable hardware prototype by using deep neural networks (DNN) deployed on a resource limited device. To train the DNNs, a dataset was recorded which contains accelerometer and gyroscope readings from three IMUs mounted on the fingertips. With this dataset, various neural networks (NN) were trained and analyzed. The best NN, in terms of accuracy, memory usage and latency, was then selected and ported to the microcontroller. Finally, a runtime analysis of the model has been performed on the controller. The analysis showed that a LSTM is best suited for the detection of hand gestures. The selected model achieves an accuracy of 93% and only takes up around 40KiB of memory. In addition, the model has a throughput time of only 3.52ms, which means that the prototype can be used in real time.","PeriodicalId":128160,"journal":{"name":"2022 21st IEEE International Conference on Machine Learning and Applications (ICMLA)","volume":"46 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125983775","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Audio Classification of Low Feature Spectrograms Utilizing Convolutional Neural Networks 基于卷积神经网络的低特征谱图音频分类
Pub Date : 2022-12-01 DOI: 10.1109/ICMLA55696.2022.00115
Noel Elias
Modern day audio signal classification techniques lack the ability to classify low feature audio signals in the form of spectrographic temporal frequency data representations. Additionally, currently utilized techniques rely on full diverse data sets that are often not representative of real-world distributions. This paper derives several first-of-its-kind machine learning methodologies to analyze these low feature audio spectrograms given data distributions that may have normalized, skewed, or even limited training sets. In particular, this paper proposes several novel customized convolutional architectures to extract identifying features using binary, one-class, and siamese approaches to identify the spectrographic signature of a given audio signal. Utilizing these novel convolutional architectures as well as the proposed classification methods, these experiments demonstrate state-of-the-art classification accuracy and improved efficiency than traditional audio classification methods.
现代音频信号分类技术缺乏以频谱时间频率数据表示形式对低特征音频信号进行分类的能力。此外,目前使用的技术依赖于完全不同的数据集,这些数据集通常不能代表现实世界的分布。本文提出了几种首创的机器学习方法来分析这些低特征音频频谱图,这些数据分布可能具有规范化、偏斜甚至有限的训练集。特别是,本文提出了几种新的自定义卷积架构,以使用二进制,一类和连体方法提取识别特征,以识别给定音频信号的频谱签名。利用这些新颖的卷积架构和提出的分类方法,这些实验证明了最先进的分类精度和效率比传统的音频分类方法。
{"title":"Audio Classification of Low Feature Spectrograms Utilizing Convolutional Neural Networks","authors":"Noel Elias","doi":"10.1109/ICMLA55696.2022.00115","DOIUrl":"https://doi.org/10.1109/ICMLA55696.2022.00115","url":null,"abstract":"Modern day audio signal classification techniques lack the ability to classify low feature audio signals in the form of spectrographic temporal frequency data representations. Additionally, currently utilized techniques rely on full diverse data sets that are often not representative of real-world distributions. This paper derives several first-of-its-kind machine learning methodologies to analyze these low feature audio spectrograms given data distributions that may have normalized, skewed, or even limited training sets. In particular, this paper proposes several novel customized convolutional architectures to extract identifying features using binary, one-class, and siamese approaches to identify the spectrographic signature of a given audio signal. Utilizing these novel convolutional architectures as well as the proposed classification methods, these experiments demonstrate state-of-the-art classification accuracy and improved efficiency than traditional audio classification methods.","PeriodicalId":128160,"journal":{"name":"2022 21st IEEE International Conference on Machine Learning and Applications (ICMLA)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125490721","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Machine learning protocol from ultrasound data for monitoring, predicting, and supporting the analysis of dam slopes 从超声波数据中提取机器学习协议,用于监测、预测和支持大坝边坡分析
Pub Date : 2022-12-01 DOI: 10.1109/ICMLA55696.2022.00084
W. Rocha, Antônio U Lucena, G. F. Sarmanho, Rodrigo C Félix, S. Miqueleti, T. C. Dourado
Dam monitoring can be used as an important indicator for dam risk management. In this study, a methodology based on machine learning and ultrasound for dam safety monitoring is presented. First, a prototype dam was built to simulate different environmental conditions. Second, ultrasound images were acquired in different areas of a prototype dam. Finally, various machine learning algorithms were applied to distinguish the different regions observed in the prototype dam. The results show that it is possible to distinguish the dam regions, which is of great value for dam safety monitoring and operation.
大坝监测可以作为大坝风险管理的重要指标。本文提出了一种基于机器学习和超声波的大坝安全监测方法。首先,建立了一个原型坝来模拟不同的环境条件。其次,在原型坝的不同区域获取超声图像。最后,应用各种机器学习算法来区分原型坝中观察到的不同区域。结果表明,该方法可实现坝区的划分,对大坝安全监测和运行具有重要价值。
{"title":"Machine learning protocol from ultrasound data for monitoring, predicting, and supporting the analysis of dam slopes","authors":"W. Rocha, Antônio U Lucena, G. F. Sarmanho, Rodrigo C Félix, S. Miqueleti, T. C. Dourado","doi":"10.1109/ICMLA55696.2022.00084","DOIUrl":"https://doi.org/10.1109/ICMLA55696.2022.00084","url":null,"abstract":"Dam monitoring can be used as an important indicator for dam risk management. In this study, a methodology based on machine learning and ultrasound for dam safety monitoring is presented. First, a prototype dam was built to simulate different environmental conditions. Second, ultrasound images were acquired in different areas of a prototype dam. Finally, various machine learning algorithms were applied to distinguish the different regions observed in the prototype dam. The results show that it is possible to distinguish the dam regions, which is of great value for dam safety monitoring and operation.","PeriodicalId":128160,"journal":{"name":"2022 21st IEEE International Conference on Machine Learning and Applications (ICMLA)","volume":"11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125897673","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Dejà vu: Recurrent Neural Networks for health wearables data forecast dejjovu:用于健康可穿戴设备数据预测的递归神经网络
Pub Date : 2022-12-01 DOI: 10.1109/ICMLA55696.2022.00264
Igor Matias, K. Wac
Wearable devices are a useful and widely used source of continuous and temporal dependant data. In contrast to the traditional clinical environment, these devices allow time series data collection in an individual’s daily living environment. However, missing data can occur while using them. Many techniques have been applied to solve these data gaps; nonetheless, missing time series data poses extra challenges, such as maintaining the temporal dependency. In this article, we addressed the forecast of sleep trackers data (sleeping heart rate (HR) and time asleep) for 2 main reasons: (1) to design models capable of accurately forecasting missing data from those devices, and (2) to apply those models to empower sleep interventions that may increase its quality, by forecasting future sleep events. We collected wearables data over 290 days (per individual) from 12 participants using a smartwatch and made this dataset publicly available. We then explored several hyperparameters of 2 Recurrent Neural Networks (RNN), Long Short-Term Memory (LSTM) and Gated Recurrent Unit (GRU). We further elaborated and compared the performance of 3 approaches to training those RNNs. Although similar performance, slightly more accurate results were obtained after training a GRU network on an entire population’s dataset, which was able to forecast the average, minimum, and maximum sleeping HR with a root-mean-squared error (RMSE) of 4.4 (± 1.4), 4.9 (± 2.6), and 12.1 ( 4.0) beats per minute, respectively. However, the total time ±asleep was impossible to forecast with low error.
可穿戴设备是一种有用且广泛使用的连续和时间相关数据来源。与传统的临床环境相比,这些设备允许在个人的日常生活环境中收集时间序列数据。但是,在使用它们时可能会丢失数据。已经应用了许多技术来解决这些数据差距;尽管如此,缺少时间序列数据带来了额外的挑战,例如维护时间依赖性。在本文中,我们讨论了睡眠追踪器数据(睡眠心率(HR)和睡眠时间)的预测,主要有两个原因:(1)设计能够准确预测这些设备缺失数据的模型;(2)通过预测未来的睡眠事件,将这些模型应用于可能提高其质量的睡眠干预。我们从12名使用智能手表的参与者那里收集了超过290天(每个人)的可穿戴设备数据,并将该数据集公开。然后,我们探讨了2种递归神经网络(RNN)、长短期记忆(LSTM)和门控递归单元(GRU)的几个超参数。我们进一步阐述并比较了训练这些rnn的3种方法的性能。虽然表现相似,但在整个人群数据集上训练GRU网络后获得的结果略准确,该网络能够预测平均,最小和最大睡眠HR,均方根误差(RMSE)分别为4.4(±1.4),4.9(±2.6)和12.1(4.0)次/分钟。然而,总睡眠时间±睡眠时间是不可能以低误差预测的。
{"title":"Dejà vu: Recurrent Neural Networks for health wearables data forecast","authors":"Igor Matias, K. Wac","doi":"10.1109/ICMLA55696.2022.00264","DOIUrl":"https://doi.org/10.1109/ICMLA55696.2022.00264","url":null,"abstract":"Wearable devices are a useful and widely used source of continuous and temporal dependant data. In contrast to the traditional clinical environment, these devices allow time series data collection in an individual’s daily living environment. However, missing data can occur while using them. Many techniques have been applied to solve these data gaps; nonetheless, missing time series data poses extra challenges, such as maintaining the temporal dependency. In this article, we addressed the forecast of sleep trackers data (sleeping heart rate (HR) and time asleep) for 2 main reasons: (1) to design models capable of accurately forecasting missing data from those devices, and (2) to apply those models to empower sleep interventions that may increase its quality, by forecasting future sleep events. We collected wearables data over 290 days (per individual) from 12 participants using a smartwatch and made this dataset publicly available. We then explored several hyperparameters of 2 Recurrent Neural Networks (RNN), Long Short-Term Memory (LSTM) and Gated Recurrent Unit (GRU). We further elaborated and compared the performance of 3 approaches to training those RNNs. Although similar performance, slightly more accurate results were obtained after training a GRU network on an entire population’s dataset, which was able to forecast the average, minimum, and maximum sleeping HR with a root-mean-squared error (RMSE) of 4.4 (± 1.4), 4.9 (± 2.6), and 12.1 ( 4.0) beats per minute, respectively. However, the total time ±asleep was impossible to forecast with low error.","PeriodicalId":128160,"journal":{"name":"2022 21st IEEE International Conference on Machine Learning and Applications (ICMLA)","volume":"10 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130282337","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Machine Learning in Personalized Skin Care: A Simulation Scheme for Pattern Recognition in Skin Condition Genome-wide Association Studies 个性化皮肤护理中的机器学习:皮肤状况全基因组关联研究中模式识别的模拟方案
Pub Date : 2022-12-01 DOI: 10.1109/ICMLA55696.2022.00164
Jerry Bonnell, Melanie Xia, Lee Wall, York Eggleston, M. Ogihara, V. Aguiar-Pulido
Personalized medicine is becoming of increasing importance in the study of psoriasis and atopic dermatitis (AD). Because current treatments only target symptoms, early intervention and personalized medicine have a pivotal role in improved health outcomes. To explore this potential, this study investigates the use of direct-to-consumer (DTC) genetic data in devising machine learning models that can pinpoint signatures salient to psoriasis and AD. The study simulates high-dimensional datasets derived from the HapMap 3 and 1000 Genomes Project cohorts (561K and 497K loci, respectively, that act as features). The simulation scheme splits subjects into cases and controls, where randomly selected variants associated with the target phenotypes are introduced into the cases. Unsupervised learning (UMAP) and eight supervised learning techniques are applied to each of the simulated datasets. Our findings suggest that the parametric models tested (SVM, LASSO, and RIDGE) exhibit the best predictive power on the simulated datasets while also yielding high retrieval rates for signatures associated with the target phenotypes.
个体化医疗在银屑病和特应性皮炎(AD)的研究中变得越来越重要。由于目前的治疗只针对症状,早期干预和个性化医疗在改善健康结果方面起着关键作用。为了探索这一潜力,本研究调查了直接面向消费者(DTC)基因数据在设计机器学习模型中的应用,该模型可以确定牛皮癣和AD的显著特征。该研究模拟了来自HapMap 3和1000基因组计划队列的高维数据集(分别为561K和497K位点,作为特征)。模拟方案将受试者分为病例和对照组,其中随机选择与目标表型相关的变体引入病例。将非监督学习(UMAP)和8种监督学习技术应用于每个模拟数据集。我们的研究结果表明,所测试的参数模型(SVM、LASSO和RIDGE)在模拟数据集上表现出最佳的预测能力,同时对与目标表型相关的特征也产生了很高的检索率。
{"title":"Machine Learning in Personalized Skin Care: A Simulation Scheme for Pattern Recognition in Skin Condition Genome-wide Association Studies","authors":"Jerry Bonnell, Melanie Xia, Lee Wall, York Eggleston, M. Ogihara, V. Aguiar-Pulido","doi":"10.1109/ICMLA55696.2022.00164","DOIUrl":"https://doi.org/10.1109/ICMLA55696.2022.00164","url":null,"abstract":"Personalized medicine is becoming of increasing importance in the study of psoriasis and atopic dermatitis (AD). Because current treatments only target symptoms, early intervention and personalized medicine have a pivotal role in improved health outcomes. To explore this potential, this study investigates the use of direct-to-consumer (DTC) genetic data in devising machine learning models that can pinpoint signatures salient to psoriasis and AD. The study simulates high-dimensional datasets derived from the HapMap 3 and 1000 Genomes Project cohorts (561K and 497K loci, respectively, that act as features). The simulation scheme splits subjects into cases and controls, where randomly selected variants associated with the target phenotypes are introduced into the cases. Unsupervised learning (UMAP) and eight supervised learning techniques are applied to each of the simulated datasets. Our findings suggest that the parametric models tested (SVM, LASSO, and RIDGE) exhibit the best predictive power on the simulated datasets while also yielding high retrieval rates for signatures associated with the target phenotypes.","PeriodicalId":128160,"journal":{"name":"2022 21st IEEE International Conference on Machine Learning and Applications (ICMLA)","volume":"35 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126363065","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Spars Kernelized Features for Prediction of Rock’s Carbon Capture using 3D X-Ray Images Spars核特征预测岩石碳捕获使用3D x射线图像
Pub Date : 2022-12-01 DOI: 10.1109/ICMLA55696.2022.00081
S. Sharifzadeh
X-ray Computed Tomography (CT) imaging is used as a non-destructive strategy for characterizing the internal structure of rocks. One important application of such studies is prediction of the relative permeability of CO2 in reservoirs. Estimation of Carbon Capture and Storage (CCS) has a great impact in mitigation strategies for global warming and controlling the effects of climate change. In this paper, 3D Xray Computed Tomography (CT) image volumes of rocks are characterized for prediction of the CO2 relative permeability. A new analysis pipeline is introduced that extracts high dimensional entropy features from the local 3D voxels. That is followed by a sparse kernelized dimensionality reduction step to alleviate the over-fitting issue. Then, regression analysis is performed using Gaussian Process Regression (GPR). Furthermore, the proposed pipeline is compared with two other deep Neural Networks (NN) models including a Convolutional Neural Network (CNN) regression model as well as a transferred pre-trained ResNet50 model using the rock X-ray training data. Experimental results show improvements in CO2 permeability prediction using the proposed analysis pipeline.
x射线计算机断层扫描(CT)成像被用作表征岩石内部结构的非破坏性策略。这类研究的一个重要应用是预测储层中二氧化碳的相对渗透率。碳捕集与封存(CCS)的估算对全球变暖减缓战略和控制气候变化影响具有重要影响。本文对岩石三维x射线计算机断层扫描(CT)成像体积进行了表征,用于预测CO2相对渗透率。提出了一种新的分析管道,从局部三维体素中提取高维熵特征。接下来是一个稀疏核降维步骤,以缓解过度拟合问题。然后,使用高斯过程回归(GPR)进行回归分析。此外,将所提出的管道与另外两种深度神经网络(NN)模型进行了比较,其中包括卷积神经网络(CNN)回归模型以及使用岩石x射线训练数据传输的预训练ResNet50模型。实验结果表明,利用该分析管道对CO2渗透率的预测有一定的提高。
{"title":"Spars Kernelized Features for Prediction of Rock’s Carbon Capture using 3D X-Ray Images","authors":"S. Sharifzadeh","doi":"10.1109/ICMLA55696.2022.00081","DOIUrl":"https://doi.org/10.1109/ICMLA55696.2022.00081","url":null,"abstract":"X-ray Computed Tomography (CT) imaging is used as a non-destructive strategy for characterizing the internal structure of rocks. One important application of such studies is prediction of the relative permeability of CO2 in reservoirs. Estimation of Carbon Capture and Storage (CCS) has a great impact in mitigation strategies for global warming and controlling the effects of climate change. In this paper, 3D Xray Computed Tomography (CT) image volumes of rocks are characterized for prediction of the CO2 relative permeability. A new analysis pipeline is introduced that extracts high dimensional entropy features from the local 3D voxels. That is followed by a sparse kernelized dimensionality reduction step to alleviate the over-fitting issue. Then, regression analysis is performed using Gaussian Process Regression (GPR). Furthermore, the proposed pipeline is compared with two other deep Neural Networks (NN) models including a Convolutional Neural Network (CNN) regression model as well as a transferred pre-trained ResNet50 model using the rock X-ray training data. Experimental results show improvements in CO2 permeability prediction using the proposed analysis pipeline.","PeriodicalId":128160,"journal":{"name":"2022 21st IEEE International Conference on Machine Learning and Applications (ICMLA)","volume":"86 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126354542","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
2022 21st IEEE International Conference on Machine Learning and Applications (ICMLA)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1