首页 > 最新文献

Applied Computing and Geosciences最新文献

英文 中文
Deep learning for real-time P-wave detection: A case study in Indonesia's earthquake early warning system 用于实时 P 波检测的深度学习:印度尼西亚地震预警系统案例研究
IF 2.6 Q2 COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS Pub Date : 2024-09-07 DOI: 10.1016/j.acags.2024.100194
Adi Wibowo , Leni Sophia Heliani , Cecep Pratama , David Prambudi Sahara , Sri Widiyantoro , Dadan Ramdani , Mizan Bustanul Fuady Bisri , Ajat Sudrajat , Sidik Tri Wibowo , Satriawan Rasyid Purnama

Detecting seismic events in real-time for prompt alerts and responses is a challenging task that requires accurately capturing P-wave arrivals. This task becomes even more challenging in regions like Indonesia, where widely spaced seismic stations exist. The wide station spacing makes associating the seismic signals with specific even more difficult. This paper proposes a novel deep learning-based model with three convolutional layers, enriched with dual attention mechanisms—Squeeze, Excitation, and Transformer Encoder (CNN-SE-T) —to refine feature extraction and improve detection sensitivity. We have integrated several post-processing techniques to further bolster the model's robustness against noise. We conducted comprehensive evaluations of our method using three diverse datasets: local earthquake data from East Java, the publicly available Seismic Waveform Data (STEAD), and a continuous waveform dataset spanning 12 h from multiple Indonesian seismic stations. The performance of the CNN-SE-T P-wave detection model yielded exceptionally high F1 scores of 99.10% for East Java, 92.64% for STEAD, and 80% for the 12-h continuous waveforms across Indonesia's network, demonstrating the model's effectiveness and potential for real-world application in earthquake early warning systems.

实时检测地震事件以便及时发出警报和做出反应是一项极具挑战性的任务,需要准确捕捉 P 波到达。在印度尼西亚等地震台站间距较大的地区,这项任务变得更具挑战性。台站间距过大使得将地震信号与具体事件联系起来变得更加困难。本文提出了一种基于深度学习的新型模型,该模型具有三个卷积层,并采用了双重注意机制--挤压、激励和变压器编码器(CNN-SE-T)--以完善特征提取并提高检测灵敏度。我们还集成了几种后处理技术,以进一步增强模型对噪声的鲁棒性。我们使用三个不同的数据集对我们的方法进行了全面评估:东爪哇岛的本地地震数据、公开可用的地震波形数据(STEAD),以及来自多个印尼地震台站、时间跨度达 12 小时的连续波形数据集。CNN-SE-T P 波检测模型在东爪哇的 F1 得分为 99.10%,在 STEAD 的 F1 得分为 92.64%,在印尼网络的 12 小时连续波形的 F1 得分为 80%,表现出该模型在地震预警系统中的有效性和实际应用潜力。
{"title":"Deep learning for real-time P-wave detection: A case study in Indonesia's earthquake early warning system","authors":"Adi Wibowo ,&nbsp;Leni Sophia Heliani ,&nbsp;Cecep Pratama ,&nbsp;David Prambudi Sahara ,&nbsp;Sri Widiyantoro ,&nbsp;Dadan Ramdani ,&nbsp;Mizan Bustanul Fuady Bisri ,&nbsp;Ajat Sudrajat ,&nbsp;Sidik Tri Wibowo ,&nbsp;Satriawan Rasyid Purnama","doi":"10.1016/j.acags.2024.100194","DOIUrl":"10.1016/j.acags.2024.100194","url":null,"abstract":"<div><p>Detecting seismic events in real-time for prompt alerts and responses is a challenging task that requires accurately capturing P-wave arrivals. This task becomes even more challenging in regions like Indonesia, where widely spaced seismic stations exist. The wide station spacing makes associating the seismic signals with specific even more difficult. This paper proposes a novel deep learning-based model with three convolutional layers, enriched with dual attention mechanisms—Squeeze, Excitation, and Transformer Encoder (CNN-SE-T) —to refine feature extraction and improve detection sensitivity. We have integrated several post-processing techniques to further bolster the model's robustness against noise. We conducted comprehensive evaluations of our method using three diverse datasets: local earthquake data from East Java, the publicly available Seismic Waveform Data (STEAD), and a continuous waveform dataset spanning 12 h from multiple Indonesian seismic stations. The performance of the CNN-SE-T P-wave detection model yielded exceptionally high F1 scores of 99.10% for East Java, 92.64% for STEAD, and 80% for the 12-h continuous waveforms across Indonesia's network, demonstrating the model's effectiveness and potential for real-world application in earthquake early warning systems.</p></div>","PeriodicalId":33804,"journal":{"name":"Applied Computing and Geosciences","volume":"24 ","pages":"Article 100194"},"PeriodicalIF":2.6,"publicationDate":"2024-09-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2590197424000417/pdfft?md5=b20faddd2c5b63b0d46b89310f92cfaf&pid=1-s2.0-S2590197424000417-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142161200","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Integration of multi-temporal SAR data and robust machine learning models for improvement of flood susceptibility assessment in the southwest coast of India 整合多时合成孔径雷达数据和稳健的机器学习模型,改进印度西南海岸的洪水易感性评估
IF 2.6 Q2 COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS Pub Date : 2024-09-04 DOI: 10.1016/j.acags.2024.100189
Pankaj Prasad , Sourav Mandal , Sahil Sandeep Naik , Victor Joseph Loveson , Simanku Borah , Priyankar Chandra , Karthik Sudheer

The flood hazards in the southwest coastal region of India in 2018 and 2020 resulted in numerous casualties and the displacement of over a million people from their homes. In order to mitigate the loss of life and resources caused by recurrent major and minor flood events, it is imperative to develop a comprehensive spatial flood zonation map of the entire area. Therefore, the main aim of the present study is to prepare a flood susceptible map of the southwest coastal region of India using synthetic-aperture radar (SAR) data and robust machine learning algorithms. Accurate flood and non-flood locations have been identified from the multi-temporal Sentinel-1 images. These flood locations are correlated with sixteen flood conditioning geo-environmental variables. The Boruta algorithm has been applied to determine the importance of each flood conditioning parameter. Six efficient machine learning models, namely support vector machine (SVM), k-nearest neighbors (KNN), artificial neural network (ANN), random forest (RF), partial least squares (PLS) and penalized discriminant analysis (PDA) have been applied to delineate the flood susceptible areas of the study region. The performance of the models has been evaluated using several statistical criteria, including area under curve (AUC), overall accuracy, specificity, sensitivity and kappa index. The results have revealed that all models have performed more than 90% of AUC due to the high precision of radar data. However, the RF and SVM models have outperformed other models in terms of all statistical parameters. The findings have identified approximately 13% of the study region as highly vulnerable to flood hazards, emphasizing the need for proper planning and management in these areas.

2018 年和 2020 年,印度西南沿海地区的洪水灾害造成大量人员伤亡,超过 100 万人背井离乡。为了减轻经常性大小洪水事件造成的生命和资源损失,当务之急是绘制整个地区的综合空间洪水分区图。因此,本研究的主要目的是利用合成孔径雷达(SAR)数据和强大的机器学习算法,绘制印度西南沿海地区易受洪水影响的地图。从多时相 Sentinel-1 图像中确定了准确的洪水和非洪水位置。这些洪水位置与 16 个洪水条件地质环境变量相关联。Boruta 算法用于确定每个洪水调节参数的重要性。六种高效的机器学习模型,即支持向量机 (SVM)、k-近邻 (KNN)、人工神经网络 (ANN)、随机森林 (RF)、偏最小二乘法 (PLS) 和惩罚性判别分析 (PDA),已被用于划定研究区域的洪水易发区。这些模型的性能采用了多种统计标准进行评估,包括曲线下面积(AUC)、总体准确度、特异性、灵敏度和卡帕指数。结果显示,由于雷达数据精度高,所有模型的 AUC 均超过 90%。不过,RF 和 SVM 模型在所有统计参数方面的表现都优于其他模型。研究结果表明,约有 13% 的研究区域极易受到洪水灾害的影响,强调了在这些区域进行适当规划和管理的必要性。
{"title":"Integration of multi-temporal SAR data and robust machine learning models for improvement of flood susceptibility assessment in the southwest coast of India","authors":"Pankaj Prasad ,&nbsp;Sourav Mandal ,&nbsp;Sahil Sandeep Naik ,&nbsp;Victor Joseph Loveson ,&nbsp;Simanku Borah ,&nbsp;Priyankar Chandra ,&nbsp;Karthik Sudheer","doi":"10.1016/j.acags.2024.100189","DOIUrl":"10.1016/j.acags.2024.100189","url":null,"abstract":"<div><p>The flood hazards in the southwest coastal region of India in 2018 and 2020 resulted in numerous casualties and the displacement of over a million people from their homes. In order to mitigate the loss of life and resources caused by recurrent major and minor flood events, it is imperative to develop a comprehensive spatial flood zonation map of the entire area. Therefore, the main aim of the present study is to prepare a flood susceptible map of the southwest coastal region of India using synthetic-aperture radar (SAR) data and robust machine learning algorithms. Accurate flood and non-flood locations have been identified from the multi-temporal Sentinel-1 images. These flood locations are correlated with sixteen flood conditioning geo-environmental variables. The Boruta algorithm has been applied to determine the importance of each flood conditioning parameter. Six efficient machine learning models, namely support vector machine (SVM), k-nearest neighbors (KNN), artificial neural network (ANN), random forest (RF), partial least squares (PLS) and penalized discriminant analysis (PDA) have been applied to delineate the flood susceptible areas of the study region. The performance of the models has been evaluated using several statistical criteria, including area under curve (AUC), overall accuracy, specificity, sensitivity and kappa index. The results have revealed that all models have performed more than 90% of AUC due to the high precision of radar data. However, the RF and SVM models have outperformed other models in terms of all statistical parameters. The findings have identified approximately 13% of the study region as highly vulnerable to flood hazards, emphasizing the need for proper planning and management in these areas.</p></div>","PeriodicalId":33804,"journal":{"name":"Applied Computing and Geosciences","volume":"24 ","pages":"Article 100189"},"PeriodicalIF":2.6,"publicationDate":"2024-09-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2590197424000363/pdfft?md5=c335020c63eb9eda70216e7662e23b2d&pid=1-s2.0-S2590197424000363-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142161206","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
POSIT: An automated tool for detecting and characterizing diverse morphological features in raster data - Application to pockmarks, mounds, and craters POSIT:用于检测和描述栅格数据中各种形态特征的自动工具--应用于麻坑、土墩和火山口
IF 2.6 Q2 COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS Pub Date : 2024-09-01 DOI: 10.1016/j.acags.2024.100190
José J. Alonso del Rosario , Ariadna Canari , Elízabeth Blázquez Gómez , Sara Martínez-Loriente

Accurate detection and characterization of seafloor morphologies are crucial for marine researchers and industries involved in underwater mapping, environmental monitoring, or resource exploration. Although their detection has relied on visual inspection of detailed bathymetries, few efforts to automate the process can be found in the literature. This study presents a novel MatLab computer code called POSIT (Feature Signature Detection) based on the convolution and correlation with a structural element containing the shape to search for. POSIT is successfully tested on both synthetic and real datasets, encompassing marine and terrestrial digital elevation models of different resolution and on a digital image. The centroids of submarine pockmarks and mounds, terrestrial volcanic craters and lunar craters are calculated with zero dispersion and perfect location, and their geometric parameters and confidence are provided.

对于从事水下测绘、环境监测或资源勘探的海洋研究人员和行业来说,准确检测和描述海底形态至关重要。虽然对海底形态的检测一直依赖于对详细水深测量数据的目测,但文献中鲜有将这一过程自动化的尝试。本研究介绍了一种名为 POSIT(特征签名检测)的新型 MatLab 计算机代码,它基于与包含要搜索的形状的结构元素的卷积和相关性。POSIT 成功地在合成数据集和真实数据集上进行了测试,包括不同分辨率的海洋和陆地数字高程模型以及数字图像。计算出的海底麻坑和土墩、陆地火山口和月球陨石坑的中心点具有零分散和完美定位的特点,并提供了它们的几何参数和置信度。
{"title":"POSIT: An automated tool for detecting and characterizing diverse morphological features in raster data - Application to pockmarks, mounds, and craters","authors":"José J. Alonso del Rosario ,&nbsp;Ariadna Canari ,&nbsp;Elízabeth Blázquez Gómez ,&nbsp;Sara Martínez-Loriente","doi":"10.1016/j.acags.2024.100190","DOIUrl":"10.1016/j.acags.2024.100190","url":null,"abstract":"<div><p>Accurate detection and characterization of seafloor morphologies are crucial for marine researchers and industries involved in underwater mapping, environmental monitoring, or resource exploration. Although their detection has relied on visual inspection of detailed bathymetries, few efforts to automate the process can be found in the literature. This study presents a novel MatLab computer code called POSIT (Feature Signature Detection) based on the convolution and correlation with a structural element containing the shape to search for. POSIT is successfully tested on both synthetic and real datasets, encompassing marine and terrestrial digital elevation models of different resolution and on a digital image. The centroids of submarine pockmarks and mounds, terrestrial volcanic craters and lunar craters are calculated with zero dispersion and perfect location, and their geometric parameters and confidence are provided.</p></div>","PeriodicalId":33804,"journal":{"name":"Applied Computing and Geosciences","volume":"23 ","pages":"Article 100190"},"PeriodicalIF":2.6,"publicationDate":"2024-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2590197424000375/pdfft?md5=dc86b3aae122c80855ab41c6633e87ec&pid=1-s2.0-S2590197424000375-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142096878","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Resolution enhancement and segmentation of polar bubbly ice micro CT scans via 3D convolutional neural network 通过三维卷积神经网络提高极地气泡冰微型 CT 扫描的分辨率并进行分割
IF 2.6 Q2 COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS Pub Date : 2024-09-01 DOI: 10.1016/j.acags.2024.100193
Faramarz Bagherzadeh , Johannes Freitag , Udo Frese , Frank Wilhelms

Accurate segmentation of 3D micro CT scans is a key step in the process of analysis of the microstructure of porous materials. In polar ice core studies, the environmental effects on the firn column could be detected if the microstructure is digitized accurately. The most challenging task is to obtain the microstructure parameters of the bubbly ice section. To identify the minimum, necessary resolution, the bubbly ice micro CT scans with different resolutions (120, 60, 30, 12 μm) were compared object-wise via a region pairing algorithm. When the minimum resolution was found to be 60 μm, for generating the training/validation dataset, 4 ice core samples from a depth range of 96 to 108 meters (bubbly ice) were scanned with 120 μm (input images) and another time with 4 times higher resolution (30 μm) to build ground truth. A specific pipeline was designed with non-rigid image registration to create an accurate ground truth from 4 times higher resolution scans. Then, two SOTA deep learning models (3D-Unet and FCN) were trained and later validated to perform super-resolution segmentation by taking input of 120μm resolution data and giving the output of binary segmented with two times higher resolution (60μm). Finally, the outputs of CNN models were compared with traditional rule-based and unsupervised methods on blind test data. It is observed the 3D-Unet can segment low-resolution scans with an accuracy of 96% and an f1-score of 80.8% while preserving microstructure having less than 2% error in porosity and SSA.

准确分割三维微型 CT 扫描图像是分析多孔材料微观结构的关键步骤。在极地冰芯研究中,如果能对微观结构进行精确的数字化,就能检测出环境对枞树柱的影响。最具挑战性的任务是获取气泡冰断面的微观结构参数。为了确定所需的最小分辨率,通过区域配对算法对不同分辨率(120、60、30、12 μm)的气泡冰微型 CT 扫描进行了对象比较。当发现最小分辨率为 60 μm 时,为生成训练/验证数据集,用 120 μm(输入图像)扫描了 96 至 108 米深度范围内的 4 个冰芯样本(气泡冰),并用高 4 倍的分辨率(30 μm)扫描了另一次,以建立基本真相。设计了一个非刚性图像配准的特定流水线,以便从 4 倍更高分辨率的扫描中创建精确的地面实况。然后,对两个 SOTA 深度学习模型(3D-Unet 和 FCN)进行了训练和验证,以执行超分辨率分割,方法是输入 120 微米分辨率的数据,并输出高两倍分辨率(60 微米)的二进制分割结果。最后,在盲测试数据上将 CNN 模型的输出结果与传统的基于规则的方法和无监督方法进行了比较。结果表明,3D-Unet 能以 96% 的准确率和 80.8% 的 f1 分数分割低分辨率扫描数据,同时保留微观结构,在孔隙度和 SSA 方面的误差小于 2%。
{"title":"Resolution enhancement and segmentation of polar bubbly ice micro CT scans via 3D convolutional neural network","authors":"Faramarz Bagherzadeh ,&nbsp;Johannes Freitag ,&nbsp;Udo Frese ,&nbsp;Frank Wilhelms","doi":"10.1016/j.acags.2024.100193","DOIUrl":"10.1016/j.acags.2024.100193","url":null,"abstract":"<div><p>Accurate segmentation of 3D micro CT scans is a key step in the process of analysis of the microstructure of porous materials. In polar ice core studies, the environmental effects on the firn column could be detected if the microstructure is digitized accurately. The most challenging task is to obtain the microstructure parameters of the bubbly ice section. To identify the minimum, necessary resolution, the bubbly ice micro CT scans with different resolutions (120, 60, 30, 12 <span><math><mrow><mi>μ</mi><mi>m</mi></mrow></math></span>) were compared object-wise via a region pairing algorithm. When the minimum resolution was found to be 60 <span><math><mrow><mi>μ</mi><mi>m</mi></mrow></math></span>, for generating the training/validation dataset, 4 ice core samples from a depth range of 96 to 108 meters (bubbly ice) were scanned with 120 <span><math><mrow><mi>μ</mi><mi>m</mi></mrow></math></span> (input images) and another time with 4 times higher resolution (30 <span><math><mrow><mi>μ</mi><mi>m</mi></mrow></math></span>) to build ground truth. A specific pipeline was designed with non-rigid image registration to create an accurate ground truth from 4 times higher resolution scans. Then, two SOTA deep learning models (3D-Unet and FCN) were trained and later validated to perform super-resolution segmentation by taking input of <span><math><mrow><mn>120</mn><mspace></mspace><mi>μ</mi><mi>m</mi></mrow></math></span> resolution data and giving the output of binary segmented with two times higher resolution (<span><math><mrow><mn>60</mn><mspace></mspace><mi>μ</mi><mi>m</mi></mrow></math></span>). Finally, the outputs of CNN models were compared with traditional rule-based and unsupervised methods on blind test data. It is observed the 3D-Unet can segment low-resolution scans with an accuracy of 96% and an f1-score of 80.8% while preserving microstructure having less than 2% error in porosity and SSA.</p></div>","PeriodicalId":33804,"journal":{"name":"Applied Computing and Geosciences","volume":"23 ","pages":"Article 100193"},"PeriodicalIF":2.6,"publicationDate":"2024-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2590197424000405/pdfft?md5=436bc0a47d2a2e990851e57a7c794d0b&pid=1-s2.0-S2590197424000405-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142158096","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Advancing geological image segmentation: Deep learning approaches for rock type identification and classification 推进地质图像分割:岩石类型识别和分类的深度学习方法
IF 2.6 Q2 COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS Pub Date : 2024-09-01 DOI: 10.1016/j.acags.2024.100192
Amit Kumar Gupta , Priya Mathur , Farhan Sheth , Carlos M. Travieso-Gonzalez , Sandeep Chaurasia

This study aims to tackle the obstacles linked with geological image segmentation by employing sophisticated deep learning techniques. Geological formations, characterized by diverse forms, sizes, textures, and colors, present a complex landscape for traditional image processing techniques. Drawing inspiration from recent advancements in image segmentation, particularly in medical imaging and object recognition, this research proposed a comprehensive methodology tailored to the specific requirements of geological image datasets. To establish the dataset, a minimum of 50 images per rock type was deemed essential, with the majority captured at the University of Las Palmas de Gran Canaria and during a field expedition to La Isla de La Palma, Spain. This dual-source approach ensures diversity in geological formations, enriching the dataset with a comprehensive range of visual characteristics. The study involves the identification of 19 distinct rock types, each documented with 50 samples, resulting in a comprehensive database containing 950 images. The methodology involves two crucial phases: initial preprocessing of the dataset, focusing on formatting and optimization, and subsequent application of deep learning models—ResNets, Inception V3, DenseNets, MobileNets V3, and EfficientNet V2 large. Preparing the dataset is crucial for improving both the quality and relevance, thereby to ensure the optimal performance of deep learning models, the dataset was preprocessed. Following this, transfer learning or more specifically fine-tuning is applied in the subsequent phase with ResNets, Inception V3, DenseNets, MobileNets V3, and EfficientNet V2 large, leveraging pre-trained models to enhance classification task performance. After fine-tuning eight deep learning models with optimal hyperparameters, including ResNet101, ResNet152, Inception-v3, DenseNet169, DenseNet201, MobileNet-v3-small, MobileNet-v3-large, and EfficientNet-v2-large, comprehensive evaluation revealed exceptional performance metrics. DenseNet201 and InceptionV3 attained the highest accuracy of 98.49% when tested on the original dataset, leading in precision, sensitivity, specificity, and F-score. Incorporating preprocessing steps further improved results, with all models exceeding 97.5% accuracy on the preprocessed dataset. In K-Fold cross-validation (k = 5), MobileNet V3 large excelled with the highest accuracy of 99.15%, followed by ResNet101 at 99.08%. Despite varying training times, models on the preprocessed dataset showed faster convergence without overfitting. Minimal misclassifications were observed, mainly among specific classes. Overall, the study's methodologies yielded remarkable results, surpassing 99% accuracy on the preprocessed dataset and in K-Fold cross-validation, affirming the efficacy in advancing rock type understanding.

本研究旨在通过采用复杂的深度学习技术,解决与地质图像分割相关的障碍。地质构造的形态、大小、纹理和颜色多种多样,对于传统的图像处理技术来说是一个复杂的难题。本研究从近年来图像分割领域(尤其是医学成像和物体识别领域)的最新进展中汲取灵感,提出了一种针对地质图像数据集特定要求的综合方法。为了建立数据集,每种岩石类型至少需要 50 张图像,其中大部分图像是在拉斯帕尔马斯德大加那利岛大学和西班牙拉帕尔马岛实地考察期间拍摄的。这种双源方法确保了地质构造的多样性,丰富了数据集的视觉特征。这项研究包括识别 19 种不同的岩石类型,每种类型有 50 个样本,最终形成一个包含 950 幅图像的综合数据库。该方法包括两个关键阶段:对数据集进行初步预处理,重点是格式化和优化;随后应用深度学习模型--ResNets、Inception V3、DenseNets、MobileNets V3 和 EfficientNet V2 large。准备数据集对于提高质量和相关性至关重要,因此,为了确保深度学习模型的最佳性能,我们对数据集进行了预处理。然后,在随后的阶段利用 ResNets、Inception V3、DenseNets、MobileNets V3 和 EfficientNet V2 large 进行迁移学习或更具体的微调,利用预先训练的模型来提高分类任务的性能。在使用最佳超参数对 ResNet101、ResNet152、Inception-v3、DenseNet169、DenseNet201、MobileNet-v3-small、MobileNet-v3-large 和 EfficientNet-v2-large 等八个深度学习模型进行微调后,综合评估显示了卓越的性能指标。在原始数据集上进行测试时,DenseNet201 和 InceptionV3 的准确率最高,达到 98.49%,在精确度、灵敏度、特异性和 F 分数方面均处于领先地位。加入预处理步骤进一步提高了结果,所有模型在预处理数据集上的准确率都超过了 97.5%。在 K-Fold 交叉验证(k = 5)中,MobileNet V3 large 的准确率最高,达到 99.15%,其次是 ResNet101,为 99.08%。尽管训练时间不同,但预处理数据集上的模型收敛速度更快,没有出现过度拟合。误分类现象极少,主要是在特定类别中。总体而言,该研究的方法取得了显著的成果,在预处理数据集和 K-Fold 交叉验证中的准确率超过了 99%,肯定了其在促进岩石类型理解方面的功效。
{"title":"Advancing geological image segmentation: Deep learning approaches for rock type identification and classification","authors":"Amit Kumar Gupta ,&nbsp;Priya Mathur ,&nbsp;Farhan Sheth ,&nbsp;Carlos M. Travieso-Gonzalez ,&nbsp;Sandeep Chaurasia","doi":"10.1016/j.acags.2024.100192","DOIUrl":"10.1016/j.acags.2024.100192","url":null,"abstract":"<div><p>This study aims to tackle the obstacles linked with geological image segmentation by employing sophisticated deep learning techniques. Geological formations, characterized by diverse forms, sizes, textures, and colors, present a complex landscape for traditional image processing techniques. Drawing inspiration from recent advancements in image segmentation, particularly in medical imaging and object recognition, this research proposed a comprehensive methodology tailored to the specific requirements of geological image datasets. To establish the dataset, a minimum of 50 images per rock type was deemed essential, with the majority captured at the University of Las Palmas de Gran Canaria and during a field expedition to La Isla de La Palma, Spain. This dual-source approach ensures diversity in geological formations, enriching the dataset with a comprehensive range of visual characteristics. The study involves the identification of 19 distinct rock types, each documented with 50 samples, resulting in a comprehensive database containing 950 images. The methodology involves two crucial phases: initial preprocessing of the dataset, focusing on formatting and optimization, and subsequent application of deep learning models—ResNets, Inception V3, DenseNets, MobileNets V3, and EfficientNet V2 large. Preparing the dataset is crucial for improving both the quality and relevance, thereby to ensure the optimal performance of deep learning models, the dataset was preprocessed. Following this, transfer learning or more specifically fine-tuning is applied in the subsequent phase with ResNets, Inception V3, DenseNets, MobileNets V3, and EfficientNet V2 large, leveraging pre-trained models to enhance classification task performance. After fine-tuning eight deep learning models with optimal hyperparameters, including ResNet101, ResNet152, Inception-v3, DenseNet169, DenseNet201, MobileNet-v3-small, MobileNet-v3-large, and EfficientNet-v2-large, comprehensive evaluation revealed exceptional performance metrics. DenseNet201 and InceptionV3 attained the highest accuracy of 98.49% when tested on the original dataset, leading in precision, sensitivity, specificity, and F-score. Incorporating preprocessing steps further improved results, with all models exceeding 97.5% accuracy on the preprocessed dataset. In K-Fold cross-validation (k = 5), MobileNet V3 large excelled with the highest accuracy of 99.15%, followed by ResNet101 at 99.08%. Despite varying training times, models on the preprocessed dataset showed faster convergence without overfitting. Minimal misclassifications were observed, mainly among specific classes. Overall, the study's methodologies yielded remarkable results, surpassing 99% accuracy on the preprocessed dataset and in K-Fold cross-validation, affirming the efficacy in advancing rock type understanding.</p></div>","PeriodicalId":33804,"journal":{"name":"Applied Computing and Geosciences","volume":"23 ","pages":"Article 100192"},"PeriodicalIF":2.6,"publicationDate":"2024-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2590197424000399/pdfft?md5=a986940f5d719d111fdfe4229e223af6&pid=1-s2.0-S2590197424000399-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142158095","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Interpretation techniques to explain the output of a spatial land subsidence hazard model in an area with a diverted tributary 解释支流改道地区空间地面沉降危害模型输出结果的解释技术
IF 2.6 Q2 COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS Pub Date : 2024-08-24 DOI: 10.1016/j.acags.2024.100191
Razieh Seihani , Hamid Gholami , Yahya Esmaeilpour , Alireza Kamali , Maryam Zareh

Due to the nature of black-box machine learning (ML) models used in the spatial modelling field of environmental and natural hazards, the interpretation of predictive model outputs is necessary. For this purpose, we applied four interpretation techniques consisting of interaction plot, permutation feature importance (PFI) measure, shapley additive explanation (SHAP) decision plot, and accumulated local effects (ALE) plot to explain and interpret the output of an ML model applied to map land subsidence (LS) in the Nazdasht plain, Hormozgan province, southern Iran. We applied a stepwise regression (SR) algorithm and five ML models (Cforest (as a conditional random forest), generalized linear model (GLM), multivariate linear regression (MLR), partial least squares (PLS) and extreme gradient boosting (XGBoost)) to select important features and to map the LS hazard, respectively. Thereafter, several interpretation techniques were used to explain the spatial ML hazard model output. Our findings revealed that a GLM model was the most accurate approach to map LS in our study area, and that 24.3% of the total study area had a very high susceptibility to the LS hazard. According to the interpretation techniques, land use, elevation, groundwater level and vegetation were the most important variables controlling the LS hazard and also the most important variables contributing to the model’s output. Overall, human activities, especially the diversion of the route of one of the main tributaries feeding the plain and the recharging of groundwater five decades ago, intensified the current LS occurrence. Therefore, management activities such as water spreading projects upstream of the plain can be useful to mitigate LS occurrence in the plain.

由于环境和自然灾害空间建模领域使用的黑盒机器学习(ML)模型的性质,有必要对预测模型的输出结果进行解释。为此,我们应用了四种解释技术,包括交互图、置换特征重要性(PFI)度量、夏普利加法解释(SHAP)决策图和累积局部效应(ALE)图,以解释和解释应用于绘制伊朗南部霍尔木兹甘省纳兹达什特平原土地沉降(LS)地图的 ML 模型的输出结果。我们采用逐步回归 (SR) 算法和五种 ML 模型(Cforest(作为条件随机森林)、广义线性模型 (GLM)、多元线性回归 (MLR)、偏最小二乘 (PLS) 和极梯度提升 (XGBoost)),分别选择重要特征和绘制 LS 危险图。之后,我们使用了几种解释技术来解释空间 ML 危险模型的输出结果。研究结果表明,GLM 模型是绘制研究区 LS 图最准确的方法,研究区总面积的 24.3%极易受到 LS 的危害。根据解释技术,土地利用、海拔高度、地下水位和植被是控制 LS 危险的最重要变量,也是对模型输出贡献最大的变量。总体而言,人类活动,尤其是五十年前对滋养平原的一条主要支流的改道和地下水的补给,加剧了目前的通量损失。因此,平原上游的水利工程等管理活动可以有效缓解平原的 LS 现象。
{"title":"Interpretation techniques to explain the output of a spatial land subsidence hazard model in an area with a diverted tributary","authors":"Razieh Seihani ,&nbsp;Hamid Gholami ,&nbsp;Yahya Esmaeilpour ,&nbsp;Alireza Kamali ,&nbsp;Maryam Zareh","doi":"10.1016/j.acags.2024.100191","DOIUrl":"10.1016/j.acags.2024.100191","url":null,"abstract":"<div><p>Due to the nature of black-box machine learning (ML) models used in the spatial modelling field of environmental and natural hazards, the interpretation of predictive model outputs is necessary. For this purpose, we applied four interpretation techniques consisting of interaction plot, permutation feature importance (PFI) measure, shapley additive explanation (SHAP) decision plot, and accumulated local effects (ALE) plot to explain and interpret the output of an ML model applied to map land subsidence (LS) in the Nazdasht plain, Hormozgan province, southern Iran. We applied a stepwise regression (SR) algorithm and five ML models (Cforest (as a conditional random forest), generalized linear model (GLM), multivariate linear regression (MLR), partial least squares (PLS) and extreme gradient boosting (XGBoost)) to select important features and to map the LS hazard, respectively. Thereafter, several interpretation techniques were used to explain the spatial ML hazard model output. Our findings revealed that a GLM model was the most accurate approach to map LS in our study area, and that 24.3% of the total study area had a very high susceptibility to the LS hazard. According to the interpretation techniques, land use, elevation, groundwater level and vegetation were the most important variables controlling the LS hazard and also the most important variables contributing to the model’s output. Overall, human activities, especially the diversion of the route of one of the main tributaries feeding the plain and the recharging of groundwater five decades ago, intensified the current LS occurrence. Therefore, management activities such as water spreading projects upstream of the plain can be useful to mitigate LS occurrence in the plain.</p></div>","PeriodicalId":33804,"journal":{"name":"Applied Computing and Geosciences","volume":"23 ","pages":"Article 100191"},"PeriodicalIF":2.6,"publicationDate":"2024-08-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2590197424000387/pdfft?md5=aff9aab3e9da8297a983487d668498f5&pid=1-s2.0-S2590197424000387-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142088569","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Improved reservoir characterization of thin beds by advanced deep learning approach 利用先进的深度学习方法改进薄层的储层特征描述
IF 2.6 Q2 COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS Pub Date : 2024-08-22 DOI: 10.1016/j.acags.2024.100188
Umar Manzoor , Muhsan Ehsan , Muyyassar Hussain , Yasir Bashir

Targeting reservoirs below seismic resolution presents a major challenge in reservoir characterization. High-resolution seismic data is critical for imaging the thin gas-bearing Khadro sand facies in several fields within the Lower Indus Basin (LIB). To truly characterize thin beds below tuning thickness, we showcase an optimally developed deep learning technique that can save up to 75% turn-around time while significantly reducing cost. Our workflow generates high-frequency acoustic impedance synthetics by utilizing a deep neural network (DNN) at the reservoir level vis-a-vis validating the results with existing geological facies. Simultaneously, we introduce continuous wavelet transform (CWT); wherein the three components (real, imaginary, and magnitude) are interrelated to obtain a resultant high-frequency seismic volume. A strong agreement is established at available wells to achieve a higher resolution seismic by injecting higher frequencies, which is then populated throughout the 3D cube. An excellent correlation is met with key seismic attributes extracted across the field for original and CWT-based synthetic seismic. The augmented seismic volume with enhanced frequency range substantiates the dominant frequency (Fd) and resolves thin beds, which is also validated with the help of wedge modeling of both acquired and high-frequency datasets. As a geologically valid solution, our approach effectively resolves an initially 54 m bed to ∼25 m. This deep-learning methodology is ideally suited to regions where the acquired seismic has limited resolution and lacks advanced reservoir characterization.

锁定地震分辨率以下的储层是储层特征描述的一大挑战。高分辨率地震数据对于下印度河盆地(LIB)几个油田的薄含气卡德罗砂层成像至关重要。为了真正描述调厚以下薄层的特征,我们展示了一种优化开发的深度学习技术,该技术可节省高达 75% 的周转时间,同时显著降低成本。我们的工作流程通过在储层层面利用深度神经网络(DNN)生成高频声阻抗合成,并根据现有地质面验证结果。同时,我们引入了连续小波变换(CWT),将三个分量(实值、虚值和幅值)相互关联,以获得高频地震量。通过注入更高的频率,在现有的油井中建立了很强的一致性,以获得更高分辨率的地震,然后将其填充到整个三维立方体中。原始地震和基于 CWT 的合成地震与整个油气田提取的关键地震属性具有极好的相关性。频率范围增强的增强地震量证实了主频(Fd)并解析了薄层,这也在采集数据集和高频数据集的楔形建模的帮助下得到了验证。作为一种地质学上有效的解决方案,我们的方法有效地将最初 54 米的薄层解析至 25 米。这种深度学习方法非常适合于地震采集分辨率有限、缺乏先进储层特征描述的地区。
{"title":"Improved reservoir characterization of thin beds by advanced deep learning approach","authors":"Umar Manzoor ,&nbsp;Muhsan Ehsan ,&nbsp;Muyyassar Hussain ,&nbsp;Yasir Bashir","doi":"10.1016/j.acags.2024.100188","DOIUrl":"10.1016/j.acags.2024.100188","url":null,"abstract":"<div><p>Targeting reservoirs below seismic resolution presents a major challenge in reservoir characterization. High-resolution seismic data is critical for imaging the thin gas-bearing Khadro sand facies in several fields within the Lower Indus Basin (LIB). To truly characterize thin beds below tuning thickness, we showcase an optimally developed deep learning technique that can save up to 75% turn-around time while significantly reducing cost. Our workflow generates high-frequency acoustic impedance synthetics by utilizing a deep neural network (DNN) at the reservoir level vis-a-vis validating the results with existing geological facies. Simultaneously, we introduce continuous wavelet transform (CWT); wherein the three components (real, imaginary, and magnitude) are interrelated to obtain a resultant high-frequency seismic volume. A strong agreement is established at available wells to achieve a higher resolution seismic by injecting higher frequencies, which is then populated throughout the 3D cube. An excellent correlation is met with key seismic attributes extracted across the field for original and CWT-based synthetic seismic. The augmented seismic volume with enhanced frequency range substantiates the dominant frequency (F<sub>d</sub>) and resolves thin beds, which is also validated with the help of wedge modeling of both acquired and high-frequency datasets. As a geologically valid solution, our approach effectively resolves an initially 54 m bed to ∼25 m. This deep-learning methodology is ideally suited to regions where the acquired seismic has limited resolution and lacks advanced reservoir characterization.</p></div>","PeriodicalId":33804,"journal":{"name":"Applied Computing and Geosciences","volume":"23 ","pages":"Article 100188"},"PeriodicalIF":2.6,"publicationDate":"2024-08-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2590197424000351/pdfft?md5=80034ccf54e0197dfeb31abc6927a92f&pid=1-s2.0-S2590197424000351-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142088568","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A supervised machine learning procedure for EPMA classification and plotting of mineral groups 用于 EPMA 分类和矿物组绘图的监督机器学习程序
IF 2.6 Q2 COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS Pub Date : 2024-08-19 DOI: 10.1016/j.acags.2024.100186
R. Cossio , S. Ghignone , A. Borghi , A. Corno , G. Vaggelli

An analytical method to automatically characterize rock samples for geological or petrological purposes is here proposed, by applying machine learning approach (ML) as a protocol for saving experimental times and costs.

Proper machine learning algorithms, applied to automatically acquired microanalytical data (i.e., Electron Probe Micro Analysis, EPMA), carried out with a SEM-EDS microprobe on randomly selected areas from a petrographic polished thin section, are trained, used, tested, and reported.

Learning and Validation phases are developed with literature mineral databases of electron microprobe analyses on 15 main rock-forming mineral groups. The Prediction phase is tested using an eclogite rock from the Western Alps, considered as an unknown sample: randomly selected areas are acquired as backscattered images whose intervals of gray levels, appropriately set in the gray level histogram, allow the automated particle mineral separation: automated separating Oxford Instruments Aztec Feature ® packages and a mineral plotting software are applied for mineral particle separation, crystal chemical formula calculation and plotting.

Finally, a microanalytical analysis is performed on each separated mineral particle. The crystal chemical formula is calculated, and the final classification plots are automatically produced for any determined mineral. The final results show good accuracy and analytical ease and assess the proper nature of the unknown eclogite rock sample. Therefore, the proposed analytical protocol is especially recommended in those scenarios where a large flow of microanalytical data is automatically acquired and needs to be processed.

本文提出了一种用于地质或岩石学目的的自动表征岩石样本的分析方法,通过应用机器学习方法(ML)作为节省实验时间和成本的协议、使用电子显微镜-电子显微镜分析(SEM-EDS)微探针对岩石学抛光薄片中随机选取的区域进行电子显微镜分析(EPMA),训练、使用、测试和报告适当的机器学习算法。预测阶段使用来自西阿尔卑斯山的埃克洛辉石岩石进行测试,该岩石被视为未知样本:随机选择区域获取反向散射图像,在灰度直方图中适当设置灰度级间隔,从而实现颗粒矿物的自动分离:应用牛津仪器公司的自动分离 Aztec Feature ® 软件包和矿物绘图软件进行矿物颗粒分离、晶体化学式计算和绘图。最后,对每个分离出来的矿物颗粒进行显微分析,计算晶体化学式,并自动生成任何已确定矿物的最终分类图。最终结果显示了良好的准确性和分析的简易性,并评估了未知斜长岩岩石样本的适当性质。因此,在自动获取大量微量分析数据并需要处理的情况下,特别推荐使用建议的分析方案。
{"title":"A supervised machine learning procedure for EPMA classification and plotting of mineral groups","authors":"R. Cossio ,&nbsp;S. Ghignone ,&nbsp;A. Borghi ,&nbsp;A. Corno ,&nbsp;G. Vaggelli","doi":"10.1016/j.acags.2024.100186","DOIUrl":"10.1016/j.acags.2024.100186","url":null,"abstract":"<div><p>An analytical method to automatically characterize rock samples for geological or petrological purposes is here proposed, by applying machine learning approach (ML) as a protocol for saving experimental times and costs.</p><p>Proper machine learning algorithms, applied to automatically acquired microanalytical data (i.e., Electron Probe Micro Analysis, EPMA), carried out with a SEM-EDS microprobe on randomly selected areas from a petrographic polished thin section, are trained, used, tested, and reported.</p><p>Learning and Validation phases are developed with literature mineral databases of electron microprobe analyses on 15 main rock-forming mineral groups. The Prediction phase is tested using an eclogite rock from the Western Alps, considered as an unknown sample: randomly selected areas are acquired as backscattered images whose intervals of gray levels, appropriately set in the gray level histogram, allow the automated particle mineral separation: automated separating Oxford Instruments Aztec Feature ® packages and a mineral plotting software are applied for mineral particle separation, crystal chemical formula calculation and plotting.</p><p>Finally, a microanalytical analysis is performed on each separated mineral particle. The crystal chemical formula is calculated, and the final classification plots are automatically produced for any determined mineral. The final results show good accuracy and analytical ease and assess the proper nature of the unknown eclogite rock sample. Therefore, the proposed analytical protocol is especially recommended in those scenarios where a large flow of microanalytical data is automatically acquired and needs to be processed.</p></div>","PeriodicalId":33804,"journal":{"name":"Applied Computing and Geosciences","volume":"23 ","pages":"Article 100186"},"PeriodicalIF":2.6,"publicationDate":"2024-08-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2590197424000338/pdfft?md5=5f9a7ff05910f5e248a1bc9ca4b633a6&pid=1-s2.0-S2590197424000338-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142006525","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
LSTM-based DEM generation in riverine environment 基于 LSTM 的河流环境 DEM 生成技术
IF 2.6 Q2 COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS Pub Date : 2024-08-15 DOI: 10.1016/j.acags.2024.100187
Virág Lovász , Ákos Halmai

In the broad field of sensors and 3D information retrieval, bathymetric reconstruction from side-scan sonar imaging is associated with unique technical hurdles. Neural Networks have recently led to promising new solutions in this field, but the available methods tend to be complex and data-intensive in a way typically making their use in a riverine environment impossible. Throughout our work, we have focused on simplifying the problem-handling and treating compatibility with a riverine environment as priority. In our work, Long Short-Term Memory proved to be effective in a surprisingly simple form. Combined with traditional post-processing techniques in the GIS environment, like median filtered focal statistics, our workflow ultimately results in ∼0.259 m median of error on the evaluation dataset of the Dráva River.

在传感器和三维信息检索的广泛领域中,侧扫声纳成像的测深重建存在独特的技术障碍。最近,神经网络为这一领域带来了前景广阔的新解决方案,但现有的方法往往非常复杂且数据密集,通常无法在河流环境中使用。在我们的工作中,我们一直致力于简化问题的处理,并优先考虑与河流环境的兼容性。在我们的工作中,事实证明长短时记忆以一种令人惊讶的简单形式发挥了作用。结合 GIS 环境中的传统后处理技术(如中值过滤焦点统计),我们的工作流程最终使德拉瓦河评估数据集的中值误差降到了 0.259 米以下。
{"title":"LSTM-based DEM generation in riverine environment","authors":"Virág Lovász ,&nbsp;Ákos Halmai","doi":"10.1016/j.acags.2024.100187","DOIUrl":"10.1016/j.acags.2024.100187","url":null,"abstract":"<div><p>In the broad field of sensors and 3D information retrieval, bathymetric reconstruction from side-scan sonar imaging is associated with unique technical hurdles. Neural Networks have recently led to promising new solutions in this field, but the available methods tend to be complex and data-intensive in a way typically making their use in a riverine environment impossible. Throughout our work, we have focused on simplifying the problem-handling and treating compatibility with a riverine environment as priority. In our work, Long Short-Term Memory proved to be effective in a surprisingly simple form. Combined with traditional post-processing techniques in the GIS environment, like median filtered focal statistics, our workflow ultimately results in ∼0.259 m median of error on the evaluation dataset of the Dráva River.</p></div>","PeriodicalId":33804,"journal":{"name":"Applied Computing and Geosciences","volume":"23 ","pages":"Article 100187"},"PeriodicalIF":2.6,"publicationDate":"2024-08-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S259019742400034X/pdfft?md5=17b4129af31ed050fc8151abebd2cdbf&pid=1-s2.0-S259019742400034X-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142011680","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Long-term temperature prediction with hybrid autoencoder algorithms 利用混合自动编码器算法进行长期温度预测
IF 2.6 Q2 COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS Pub Date : 2024-08-08 DOI: 10.1016/j.acags.2024.100185
J. Pérez-Aracil , D. Fister , C.M. Marina , C. Peláez-Rodríguez , L. Cornejo-Bueno , P.A. Gutiérrez , M. Giuliani , A. Castelleti , S. Salcedo-Sanz

This paper proposes two hybrid approaches based on Autoencoders (AEs) for long-term temperature prediction. The first algorithm comprises an AE trained to learn temperature patterns, which is then linked to a second AE, used to detect possible anomalies and provide a final temperature prediction. The second proposed approach involves training an AE and then using the resulting latent space as input of a neural network, which will provide the final prediction output. Both approaches are tested in long-term air temperature prediction in European cities: seven European locations where major heat waves occurred have been considered. The long-term temperature prediction for the entire year of the heatwave events has been analysed. Results show that the proposed approaches can obtain accurate long-term (up to 4 weeks) temperature prediction, improving Persistence and Climatology in the benchmark models compared. In heatwave periods, where the persistence of the temperature is extremely high, our approach beat the persistence operator in three locations and works similarly in the rest of the cases, showing the potential of this AE-based method for long-term temperature prediction.

本文提出了两种基于自动编码器(AE)的混合方法,用于长期温度预测。第一种算法包括一个经过训练的自动编码器,用于学习温度模式,然后将其与第二个自动编码器连接起来,用于检测可能的异常情况并提供最终的温度预测。第二种建议的方法包括训练一个 AE,然后将由此产生的潜在空间作为神经网络的输入,从而提供最终的预测输出。这两种方法都在欧洲城市的长期气温预测中进行了测试:考虑了七个发生过重大热浪的欧洲地点。对热浪事件全年的长期气温预测进行了分析。结果表明,所建议的方法可以获得准确的长期(长达 4 周)气温预测,并改善了基准模型的持久性和气候学。在气温持续性极高的热浪期,我们的方法在三个地点击败了持续性算子,在其他情况下效果类似,显示了这种基于 AE 的方法在长期气温预测方面的潜力。
{"title":"Long-term temperature prediction with hybrid autoencoder algorithms","authors":"J. Pérez-Aracil ,&nbsp;D. Fister ,&nbsp;C.M. Marina ,&nbsp;C. Peláez-Rodríguez ,&nbsp;L. Cornejo-Bueno ,&nbsp;P.A. Gutiérrez ,&nbsp;M. Giuliani ,&nbsp;A. Castelleti ,&nbsp;S. Salcedo-Sanz","doi":"10.1016/j.acags.2024.100185","DOIUrl":"10.1016/j.acags.2024.100185","url":null,"abstract":"<div><p>This paper proposes two hybrid approaches based on Autoencoders (AEs) for long-term temperature prediction. The first algorithm comprises an AE trained to learn temperature patterns, which is then linked to a second AE, used to detect possible anomalies and provide a final temperature prediction. The second proposed approach involves training an AE and then using the resulting latent space as input of a neural network, which will provide the final prediction output. Both approaches are tested in long-term air temperature prediction in European cities: seven European locations where major heat waves occurred have been considered. The long-term temperature prediction for the entire year of the heatwave events has been analysed. Results show that the proposed approaches can obtain accurate long-term (up to 4 weeks) temperature prediction, improving Persistence and Climatology in the benchmark models compared. In heatwave periods, where the persistence of the temperature is extremely high, our approach beat the persistence operator in three locations and works similarly in the rest of the cases, showing the potential of this AE-based method for long-term temperature prediction.</p></div>","PeriodicalId":33804,"journal":{"name":"Applied Computing and Geosciences","volume":"23 ","pages":"Article 100185"},"PeriodicalIF":2.6,"publicationDate":"2024-08-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2590197424000326/pdfft?md5=5456efe65b92894adbcaf61c5ff34ab1&pid=1-s2.0-S2590197424000326-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141993141","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Applied Computing and Geosciences
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1