Pub Date : 2024-09-07DOI: 10.1016/j.acags.2024.100194
Adi Wibowo , Leni Sophia Heliani , Cecep Pratama , David Prambudi Sahara , Sri Widiyantoro , Dadan Ramdani , Mizan Bustanul Fuady Bisri , Ajat Sudrajat , Sidik Tri Wibowo , Satriawan Rasyid Purnama
Detecting seismic events in real-time for prompt alerts and responses is a challenging task that requires accurately capturing P-wave arrivals. This task becomes even more challenging in regions like Indonesia, where widely spaced seismic stations exist. The wide station spacing makes associating the seismic signals with specific even more difficult. This paper proposes a novel deep learning-based model with three convolutional layers, enriched with dual attention mechanisms—Squeeze, Excitation, and Transformer Encoder (CNN-SE-T) —to refine feature extraction and improve detection sensitivity. We have integrated several post-processing techniques to further bolster the model's robustness against noise. We conducted comprehensive evaluations of our method using three diverse datasets: local earthquake data from East Java, the publicly available Seismic Waveform Data (STEAD), and a continuous waveform dataset spanning 12 h from multiple Indonesian seismic stations. The performance of the CNN-SE-T P-wave detection model yielded exceptionally high F1 scores of 99.10% for East Java, 92.64% for STEAD, and 80% for the 12-h continuous waveforms across Indonesia's network, demonstrating the model's effectiveness and potential for real-world application in earthquake early warning systems.
实时检测地震事件以便及时发出警报和做出反应是一项极具挑战性的任务,需要准确捕捉 P 波到达。在印度尼西亚等地震台站间距较大的地区,这项任务变得更具挑战性。台站间距过大使得将地震信号与具体事件联系起来变得更加困难。本文提出了一种基于深度学习的新型模型,该模型具有三个卷积层,并采用了双重注意机制--挤压、激励和变压器编码器(CNN-SE-T)--以完善特征提取并提高检测灵敏度。我们还集成了几种后处理技术,以进一步增强模型对噪声的鲁棒性。我们使用三个不同的数据集对我们的方法进行了全面评估:东爪哇岛的本地地震数据、公开可用的地震波形数据(STEAD),以及来自多个印尼地震台站、时间跨度达 12 小时的连续波形数据集。CNN-SE-T P 波检测模型在东爪哇的 F1 得分为 99.10%,在 STEAD 的 F1 得分为 92.64%,在印尼网络的 12 小时连续波形的 F1 得分为 80%,表现出该模型在地震预警系统中的有效性和实际应用潜力。
{"title":"Deep learning for real-time P-wave detection: A case study in Indonesia's earthquake early warning system","authors":"Adi Wibowo , Leni Sophia Heliani , Cecep Pratama , David Prambudi Sahara , Sri Widiyantoro , Dadan Ramdani , Mizan Bustanul Fuady Bisri , Ajat Sudrajat , Sidik Tri Wibowo , Satriawan Rasyid Purnama","doi":"10.1016/j.acags.2024.100194","DOIUrl":"10.1016/j.acags.2024.100194","url":null,"abstract":"<div><p>Detecting seismic events in real-time for prompt alerts and responses is a challenging task that requires accurately capturing P-wave arrivals. This task becomes even more challenging in regions like Indonesia, where widely spaced seismic stations exist. The wide station spacing makes associating the seismic signals with specific even more difficult. This paper proposes a novel deep learning-based model with three convolutional layers, enriched with dual attention mechanisms—Squeeze, Excitation, and Transformer Encoder (CNN-SE-T) —to refine feature extraction and improve detection sensitivity. We have integrated several post-processing techniques to further bolster the model's robustness against noise. We conducted comprehensive evaluations of our method using three diverse datasets: local earthquake data from East Java, the publicly available Seismic Waveform Data (STEAD), and a continuous waveform dataset spanning 12 h from multiple Indonesian seismic stations. The performance of the CNN-SE-T P-wave detection model yielded exceptionally high F1 scores of 99.10% for East Java, 92.64% for STEAD, and 80% for the 12-h continuous waveforms across Indonesia's network, demonstrating the model's effectiveness and potential for real-world application in earthquake early warning systems.</p></div>","PeriodicalId":33804,"journal":{"name":"Applied Computing and Geosciences","volume":"24 ","pages":"Article 100194"},"PeriodicalIF":2.6,"publicationDate":"2024-09-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2590197424000417/pdfft?md5=b20faddd2c5b63b0d46b89310f92cfaf&pid=1-s2.0-S2590197424000417-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142161200","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
The flood hazards in the southwest coastal region of India in 2018 and 2020 resulted in numerous casualties and the displacement of over a million people from their homes. In order to mitigate the loss of life and resources caused by recurrent major and minor flood events, it is imperative to develop a comprehensive spatial flood zonation map of the entire area. Therefore, the main aim of the present study is to prepare a flood susceptible map of the southwest coastal region of India using synthetic-aperture radar (SAR) data and robust machine learning algorithms. Accurate flood and non-flood locations have been identified from the multi-temporal Sentinel-1 images. These flood locations are correlated with sixteen flood conditioning geo-environmental variables. The Boruta algorithm has been applied to determine the importance of each flood conditioning parameter. Six efficient machine learning models, namely support vector machine (SVM), k-nearest neighbors (KNN), artificial neural network (ANN), random forest (RF), partial least squares (PLS) and penalized discriminant analysis (PDA) have been applied to delineate the flood susceptible areas of the study region. The performance of the models has been evaluated using several statistical criteria, including area under curve (AUC), overall accuracy, specificity, sensitivity and kappa index. The results have revealed that all models have performed more than 90% of AUC due to the high precision of radar data. However, the RF and SVM models have outperformed other models in terms of all statistical parameters. The findings have identified approximately 13% of the study region as highly vulnerable to flood hazards, emphasizing the need for proper planning and management in these areas.
{"title":"Integration of multi-temporal SAR data and robust machine learning models for improvement of flood susceptibility assessment in the southwest coast of India","authors":"Pankaj Prasad , Sourav Mandal , Sahil Sandeep Naik , Victor Joseph Loveson , Simanku Borah , Priyankar Chandra , Karthik Sudheer","doi":"10.1016/j.acags.2024.100189","DOIUrl":"10.1016/j.acags.2024.100189","url":null,"abstract":"<div><p>The flood hazards in the southwest coastal region of India in 2018 and 2020 resulted in numerous casualties and the displacement of over a million people from their homes. In order to mitigate the loss of life and resources caused by recurrent major and minor flood events, it is imperative to develop a comprehensive spatial flood zonation map of the entire area. Therefore, the main aim of the present study is to prepare a flood susceptible map of the southwest coastal region of India using synthetic-aperture radar (SAR) data and robust machine learning algorithms. Accurate flood and non-flood locations have been identified from the multi-temporal Sentinel-1 images. These flood locations are correlated with sixteen flood conditioning geo-environmental variables. The Boruta algorithm has been applied to determine the importance of each flood conditioning parameter. Six efficient machine learning models, namely support vector machine (SVM), k-nearest neighbors (KNN), artificial neural network (ANN), random forest (RF), partial least squares (PLS) and penalized discriminant analysis (PDA) have been applied to delineate the flood susceptible areas of the study region. The performance of the models has been evaluated using several statistical criteria, including area under curve (AUC), overall accuracy, specificity, sensitivity and kappa index. The results have revealed that all models have performed more than 90% of AUC due to the high precision of radar data. However, the RF and SVM models have outperformed other models in terms of all statistical parameters. The findings have identified approximately 13% of the study region as highly vulnerable to flood hazards, emphasizing the need for proper planning and management in these areas.</p></div>","PeriodicalId":33804,"journal":{"name":"Applied Computing and Geosciences","volume":"24 ","pages":"Article 100189"},"PeriodicalIF":2.6,"publicationDate":"2024-09-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2590197424000363/pdfft?md5=c335020c63eb9eda70216e7662e23b2d&pid=1-s2.0-S2590197424000363-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142161206","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-01DOI: 10.1016/j.acags.2024.100190
José J. Alonso del Rosario , Ariadna Canari , Elízabeth Blázquez Gómez , Sara Martínez-Loriente
Accurate detection and characterization of seafloor morphologies are crucial for marine researchers and industries involved in underwater mapping, environmental monitoring, or resource exploration. Although their detection has relied on visual inspection of detailed bathymetries, few efforts to automate the process can be found in the literature. This study presents a novel MatLab computer code called POSIT (Feature Signature Detection) based on the convolution and correlation with a structural element containing the shape to search for. POSIT is successfully tested on both synthetic and real datasets, encompassing marine and terrestrial digital elevation models of different resolution and on a digital image. The centroids of submarine pockmarks and mounds, terrestrial volcanic craters and lunar craters are calculated with zero dispersion and perfect location, and their geometric parameters and confidence are provided.
{"title":"POSIT: An automated tool for detecting and characterizing diverse morphological features in raster data - Application to pockmarks, mounds, and craters","authors":"José J. Alonso del Rosario , Ariadna Canari , Elízabeth Blázquez Gómez , Sara Martínez-Loriente","doi":"10.1016/j.acags.2024.100190","DOIUrl":"10.1016/j.acags.2024.100190","url":null,"abstract":"<div><p>Accurate detection and characterization of seafloor morphologies are crucial for marine researchers and industries involved in underwater mapping, environmental monitoring, or resource exploration. Although their detection has relied on visual inspection of detailed bathymetries, few efforts to automate the process can be found in the literature. This study presents a novel MatLab computer code called POSIT (Feature Signature Detection) based on the convolution and correlation with a structural element containing the shape to search for. POSIT is successfully tested on both synthetic and real datasets, encompassing marine and terrestrial digital elevation models of different resolution and on a digital image. The centroids of submarine pockmarks and mounds, terrestrial volcanic craters and lunar craters are calculated with zero dispersion and perfect location, and their geometric parameters and confidence are provided.</p></div>","PeriodicalId":33804,"journal":{"name":"Applied Computing and Geosciences","volume":"23 ","pages":"Article 100190"},"PeriodicalIF":2.6,"publicationDate":"2024-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2590197424000375/pdfft?md5=dc86b3aae122c80855ab41c6633e87ec&pid=1-s2.0-S2590197424000375-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142096878","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-01DOI: 10.1016/j.acags.2024.100193
Faramarz Bagherzadeh , Johannes Freitag , Udo Frese , Frank Wilhelms
Accurate segmentation of 3D micro CT scans is a key step in the process of analysis of the microstructure of porous materials. In polar ice core studies, the environmental effects on the firn column could be detected if the microstructure is digitized accurately. The most challenging task is to obtain the microstructure parameters of the bubbly ice section. To identify the minimum, necessary resolution, the bubbly ice micro CT scans with different resolutions (120, 60, 30, 12 ) were compared object-wise via a region pairing algorithm. When the minimum resolution was found to be 60 , for generating the training/validation dataset, 4 ice core samples from a depth range of 96 to 108 meters (bubbly ice) were scanned with 120 (input images) and another time with 4 times higher resolution (30 ) to build ground truth. A specific pipeline was designed with non-rigid image registration to create an accurate ground truth from 4 times higher resolution scans. Then, two SOTA deep learning models (3D-Unet and FCN) were trained and later validated to perform super-resolution segmentation by taking input of resolution data and giving the output of binary segmented with two times higher resolution (). Finally, the outputs of CNN models were compared with traditional rule-based and unsupervised methods on blind test data. It is observed the 3D-Unet can segment low-resolution scans with an accuracy of 96% and an f1-score of 80.8% while preserving microstructure having less than 2% error in porosity and SSA.
{"title":"Resolution enhancement and segmentation of polar bubbly ice micro CT scans via 3D convolutional neural network","authors":"Faramarz Bagherzadeh , Johannes Freitag , Udo Frese , Frank Wilhelms","doi":"10.1016/j.acags.2024.100193","DOIUrl":"10.1016/j.acags.2024.100193","url":null,"abstract":"<div><p>Accurate segmentation of 3D micro CT scans is a key step in the process of analysis of the microstructure of porous materials. In polar ice core studies, the environmental effects on the firn column could be detected if the microstructure is digitized accurately. The most challenging task is to obtain the microstructure parameters of the bubbly ice section. To identify the minimum, necessary resolution, the bubbly ice micro CT scans with different resolutions (120, 60, 30, 12 <span><math><mrow><mi>μ</mi><mi>m</mi></mrow></math></span>) were compared object-wise via a region pairing algorithm. When the minimum resolution was found to be 60 <span><math><mrow><mi>μ</mi><mi>m</mi></mrow></math></span>, for generating the training/validation dataset, 4 ice core samples from a depth range of 96 to 108 meters (bubbly ice) were scanned with 120 <span><math><mrow><mi>μ</mi><mi>m</mi></mrow></math></span> (input images) and another time with 4 times higher resolution (30 <span><math><mrow><mi>μ</mi><mi>m</mi></mrow></math></span>) to build ground truth. A specific pipeline was designed with non-rigid image registration to create an accurate ground truth from 4 times higher resolution scans. Then, two SOTA deep learning models (3D-Unet and FCN) were trained and later validated to perform super-resolution segmentation by taking input of <span><math><mrow><mn>120</mn><mspace></mspace><mi>μ</mi><mi>m</mi></mrow></math></span> resolution data and giving the output of binary segmented with two times higher resolution (<span><math><mrow><mn>60</mn><mspace></mspace><mi>μ</mi><mi>m</mi></mrow></math></span>). Finally, the outputs of CNN models were compared with traditional rule-based and unsupervised methods on blind test data. It is observed the 3D-Unet can segment low-resolution scans with an accuracy of 96% and an f1-score of 80.8% while preserving microstructure having less than 2% error in porosity and SSA.</p></div>","PeriodicalId":33804,"journal":{"name":"Applied Computing and Geosciences","volume":"23 ","pages":"Article 100193"},"PeriodicalIF":2.6,"publicationDate":"2024-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2590197424000405/pdfft?md5=436bc0a47d2a2e990851e57a7c794d0b&pid=1-s2.0-S2590197424000405-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142158096","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-01DOI: 10.1016/j.acags.2024.100192
Amit Kumar Gupta , Priya Mathur , Farhan Sheth , Carlos M. Travieso-Gonzalez , Sandeep Chaurasia
This study aims to tackle the obstacles linked with geological image segmentation by employing sophisticated deep learning techniques. Geological formations, characterized by diverse forms, sizes, textures, and colors, present a complex landscape for traditional image processing techniques. Drawing inspiration from recent advancements in image segmentation, particularly in medical imaging and object recognition, this research proposed a comprehensive methodology tailored to the specific requirements of geological image datasets. To establish the dataset, a minimum of 50 images per rock type was deemed essential, with the majority captured at the University of Las Palmas de Gran Canaria and during a field expedition to La Isla de La Palma, Spain. This dual-source approach ensures diversity in geological formations, enriching the dataset with a comprehensive range of visual characteristics. The study involves the identification of 19 distinct rock types, each documented with 50 samples, resulting in a comprehensive database containing 950 images. The methodology involves two crucial phases: initial preprocessing of the dataset, focusing on formatting and optimization, and subsequent application of deep learning models—ResNets, Inception V3, DenseNets, MobileNets V3, and EfficientNet V2 large. Preparing the dataset is crucial for improving both the quality and relevance, thereby to ensure the optimal performance of deep learning models, the dataset was preprocessed. Following this, transfer learning or more specifically fine-tuning is applied in the subsequent phase with ResNets, Inception V3, DenseNets, MobileNets V3, and EfficientNet V2 large, leveraging pre-trained models to enhance classification task performance. After fine-tuning eight deep learning models with optimal hyperparameters, including ResNet101, ResNet152, Inception-v3, DenseNet169, DenseNet201, MobileNet-v3-small, MobileNet-v3-large, and EfficientNet-v2-large, comprehensive evaluation revealed exceptional performance metrics. DenseNet201 and InceptionV3 attained the highest accuracy of 98.49% when tested on the original dataset, leading in precision, sensitivity, specificity, and F-score. Incorporating preprocessing steps further improved results, with all models exceeding 97.5% accuracy on the preprocessed dataset. In K-Fold cross-validation (k = 5), MobileNet V3 large excelled with the highest accuracy of 99.15%, followed by ResNet101 at 99.08%. Despite varying training times, models on the preprocessed dataset showed faster convergence without overfitting. Minimal misclassifications were observed, mainly among specific classes. Overall, the study's methodologies yielded remarkable results, surpassing 99% accuracy on the preprocessed dataset and in K-Fold cross-validation, affirming the efficacy in advancing rock type understanding.
{"title":"Advancing geological image segmentation: Deep learning approaches for rock type identification and classification","authors":"Amit Kumar Gupta , Priya Mathur , Farhan Sheth , Carlos M. Travieso-Gonzalez , Sandeep Chaurasia","doi":"10.1016/j.acags.2024.100192","DOIUrl":"10.1016/j.acags.2024.100192","url":null,"abstract":"<div><p>This study aims to tackle the obstacles linked with geological image segmentation by employing sophisticated deep learning techniques. Geological formations, characterized by diverse forms, sizes, textures, and colors, present a complex landscape for traditional image processing techniques. Drawing inspiration from recent advancements in image segmentation, particularly in medical imaging and object recognition, this research proposed a comprehensive methodology tailored to the specific requirements of geological image datasets. To establish the dataset, a minimum of 50 images per rock type was deemed essential, with the majority captured at the University of Las Palmas de Gran Canaria and during a field expedition to La Isla de La Palma, Spain. This dual-source approach ensures diversity in geological formations, enriching the dataset with a comprehensive range of visual characteristics. The study involves the identification of 19 distinct rock types, each documented with 50 samples, resulting in a comprehensive database containing 950 images. The methodology involves two crucial phases: initial preprocessing of the dataset, focusing on formatting and optimization, and subsequent application of deep learning models—ResNets, Inception V3, DenseNets, MobileNets V3, and EfficientNet V2 large. Preparing the dataset is crucial for improving both the quality and relevance, thereby to ensure the optimal performance of deep learning models, the dataset was preprocessed. Following this, transfer learning or more specifically fine-tuning is applied in the subsequent phase with ResNets, Inception V3, DenseNets, MobileNets V3, and EfficientNet V2 large, leveraging pre-trained models to enhance classification task performance. After fine-tuning eight deep learning models with optimal hyperparameters, including ResNet101, ResNet152, Inception-v3, DenseNet169, DenseNet201, MobileNet-v3-small, MobileNet-v3-large, and EfficientNet-v2-large, comprehensive evaluation revealed exceptional performance metrics. DenseNet201 and InceptionV3 attained the highest accuracy of 98.49% when tested on the original dataset, leading in precision, sensitivity, specificity, and F-score. Incorporating preprocessing steps further improved results, with all models exceeding 97.5% accuracy on the preprocessed dataset. In K-Fold cross-validation (k = 5), MobileNet V3 large excelled with the highest accuracy of 99.15%, followed by ResNet101 at 99.08%. Despite varying training times, models on the preprocessed dataset showed faster convergence without overfitting. Minimal misclassifications were observed, mainly among specific classes. Overall, the study's methodologies yielded remarkable results, surpassing 99% accuracy on the preprocessed dataset and in K-Fold cross-validation, affirming the efficacy in advancing rock type understanding.</p></div>","PeriodicalId":33804,"journal":{"name":"Applied Computing and Geosciences","volume":"23 ","pages":"Article 100192"},"PeriodicalIF":2.6,"publicationDate":"2024-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2590197424000399/pdfft?md5=a986940f5d719d111fdfe4229e223af6&pid=1-s2.0-S2590197424000399-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142158095","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Due to the nature of black-box machine learning (ML) models used in the spatial modelling field of environmental and natural hazards, the interpretation of predictive model outputs is necessary. For this purpose, we applied four interpretation techniques consisting of interaction plot, permutation feature importance (PFI) measure, shapley additive explanation (SHAP) decision plot, and accumulated local effects (ALE) plot to explain and interpret the output of an ML model applied to map land subsidence (LS) in the Nazdasht plain, Hormozgan province, southern Iran. We applied a stepwise regression (SR) algorithm and five ML models (Cforest (as a conditional random forest), generalized linear model (GLM), multivariate linear regression (MLR), partial least squares (PLS) and extreme gradient boosting (XGBoost)) to select important features and to map the LS hazard, respectively. Thereafter, several interpretation techniques were used to explain the spatial ML hazard model output. Our findings revealed that a GLM model was the most accurate approach to map LS in our study area, and that 24.3% of the total study area had a very high susceptibility to the LS hazard. According to the interpretation techniques, land use, elevation, groundwater level and vegetation were the most important variables controlling the LS hazard and also the most important variables contributing to the model’s output. Overall, human activities, especially the diversion of the route of one of the main tributaries feeding the plain and the recharging of groundwater five decades ago, intensified the current LS occurrence. Therefore, management activities such as water spreading projects upstream of the plain can be useful to mitigate LS occurrence in the plain.
由于环境和自然灾害空间建模领域使用的黑盒机器学习(ML)模型的性质,有必要对预测模型的输出结果进行解释。为此,我们应用了四种解释技术,包括交互图、置换特征重要性(PFI)度量、夏普利加法解释(SHAP)决策图和累积局部效应(ALE)图,以解释和解释应用于绘制伊朗南部霍尔木兹甘省纳兹达什特平原土地沉降(LS)地图的 ML 模型的输出结果。我们采用逐步回归 (SR) 算法和五种 ML 模型(Cforest(作为条件随机森林)、广义线性模型 (GLM)、多元线性回归 (MLR)、偏最小二乘 (PLS) 和极梯度提升 (XGBoost)),分别选择重要特征和绘制 LS 危险图。之后,我们使用了几种解释技术来解释空间 ML 危险模型的输出结果。研究结果表明,GLM 模型是绘制研究区 LS 图最准确的方法,研究区总面积的 24.3%极易受到 LS 的危害。根据解释技术,土地利用、海拔高度、地下水位和植被是控制 LS 危险的最重要变量,也是对模型输出贡献最大的变量。总体而言,人类活动,尤其是五十年前对滋养平原的一条主要支流的改道和地下水的补给,加剧了目前的通量损失。因此,平原上游的水利工程等管理活动可以有效缓解平原的 LS 现象。
{"title":"Interpretation techniques to explain the output of a spatial land subsidence hazard model in an area with a diverted tributary","authors":"Razieh Seihani , Hamid Gholami , Yahya Esmaeilpour , Alireza Kamali , Maryam Zareh","doi":"10.1016/j.acags.2024.100191","DOIUrl":"10.1016/j.acags.2024.100191","url":null,"abstract":"<div><p>Due to the nature of black-box machine learning (ML) models used in the spatial modelling field of environmental and natural hazards, the interpretation of predictive model outputs is necessary. For this purpose, we applied four interpretation techniques consisting of interaction plot, permutation feature importance (PFI) measure, shapley additive explanation (SHAP) decision plot, and accumulated local effects (ALE) plot to explain and interpret the output of an ML model applied to map land subsidence (LS) in the Nazdasht plain, Hormozgan province, southern Iran. We applied a stepwise regression (SR) algorithm and five ML models (Cforest (as a conditional random forest), generalized linear model (GLM), multivariate linear regression (MLR), partial least squares (PLS) and extreme gradient boosting (XGBoost)) to select important features and to map the LS hazard, respectively. Thereafter, several interpretation techniques were used to explain the spatial ML hazard model output. Our findings revealed that a GLM model was the most accurate approach to map LS in our study area, and that 24.3% of the total study area had a very high susceptibility to the LS hazard. According to the interpretation techniques, land use, elevation, groundwater level and vegetation were the most important variables controlling the LS hazard and also the most important variables contributing to the model’s output. Overall, human activities, especially the diversion of the route of one of the main tributaries feeding the plain and the recharging of groundwater five decades ago, intensified the current LS occurrence. Therefore, management activities such as water spreading projects upstream of the plain can be useful to mitigate LS occurrence in the plain.</p></div>","PeriodicalId":33804,"journal":{"name":"Applied Computing and Geosciences","volume":"23 ","pages":"Article 100191"},"PeriodicalIF":2.6,"publicationDate":"2024-08-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2590197424000387/pdfft?md5=aff9aab3e9da8297a983487d668498f5&pid=1-s2.0-S2590197424000387-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142088569","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Targeting reservoirs below seismic resolution presents a major challenge in reservoir characterization. High-resolution seismic data is critical for imaging the thin gas-bearing Khadro sand facies in several fields within the Lower Indus Basin (LIB). To truly characterize thin beds below tuning thickness, we showcase an optimally developed deep learning technique that can save up to 75% turn-around time while significantly reducing cost. Our workflow generates high-frequency acoustic impedance synthetics by utilizing a deep neural network (DNN) at the reservoir level vis-a-vis validating the results with existing geological facies. Simultaneously, we introduce continuous wavelet transform (CWT); wherein the three components (real, imaginary, and magnitude) are interrelated to obtain a resultant high-frequency seismic volume. A strong agreement is established at available wells to achieve a higher resolution seismic by injecting higher frequencies, which is then populated throughout the 3D cube. An excellent correlation is met with key seismic attributes extracted across the field for original and CWT-based synthetic seismic. The augmented seismic volume with enhanced frequency range substantiates the dominant frequency (Fd) and resolves thin beds, which is also validated with the help of wedge modeling of both acquired and high-frequency datasets. As a geologically valid solution, our approach effectively resolves an initially 54 m bed to ∼25 m. This deep-learning methodology is ideally suited to regions where the acquired seismic has limited resolution and lacks advanced reservoir characterization.
{"title":"Improved reservoir characterization of thin beds by advanced deep learning approach","authors":"Umar Manzoor , Muhsan Ehsan , Muyyassar Hussain , Yasir Bashir","doi":"10.1016/j.acags.2024.100188","DOIUrl":"10.1016/j.acags.2024.100188","url":null,"abstract":"<div><p>Targeting reservoirs below seismic resolution presents a major challenge in reservoir characterization. High-resolution seismic data is critical for imaging the thin gas-bearing Khadro sand facies in several fields within the Lower Indus Basin (LIB). To truly characterize thin beds below tuning thickness, we showcase an optimally developed deep learning technique that can save up to 75% turn-around time while significantly reducing cost. Our workflow generates high-frequency acoustic impedance synthetics by utilizing a deep neural network (DNN) at the reservoir level vis-a-vis validating the results with existing geological facies. Simultaneously, we introduce continuous wavelet transform (CWT); wherein the three components (real, imaginary, and magnitude) are interrelated to obtain a resultant high-frequency seismic volume. A strong agreement is established at available wells to achieve a higher resolution seismic by injecting higher frequencies, which is then populated throughout the 3D cube. An excellent correlation is met with key seismic attributes extracted across the field for original and CWT-based synthetic seismic. The augmented seismic volume with enhanced frequency range substantiates the dominant frequency (F<sub>d</sub>) and resolves thin beds, which is also validated with the help of wedge modeling of both acquired and high-frequency datasets. As a geologically valid solution, our approach effectively resolves an initially 54 m bed to ∼25 m. This deep-learning methodology is ideally suited to regions where the acquired seismic has limited resolution and lacks advanced reservoir characterization.</p></div>","PeriodicalId":33804,"journal":{"name":"Applied Computing and Geosciences","volume":"23 ","pages":"Article 100188"},"PeriodicalIF":2.6,"publicationDate":"2024-08-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2590197424000351/pdfft?md5=80034ccf54e0197dfeb31abc6927a92f&pid=1-s2.0-S2590197424000351-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142088568","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-19DOI: 10.1016/j.acags.2024.100186
R. Cossio , S. Ghignone , A. Borghi , A. Corno , G. Vaggelli
An analytical method to automatically characterize rock samples for geological or petrological purposes is here proposed, by applying machine learning approach (ML) as a protocol for saving experimental times and costs.
Proper machine learning algorithms, applied to automatically acquired microanalytical data (i.e., Electron Probe Micro Analysis, EPMA), carried out with a SEM-EDS microprobe on randomly selected areas from a petrographic polished thin section, are trained, used, tested, and reported.
Learning and Validation phases are developed with literature mineral databases of electron microprobe analyses on 15 main rock-forming mineral groups. The Prediction phase is tested using an eclogite rock from the Western Alps, considered as an unknown sample: randomly selected areas are acquired as backscattered images whose intervals of gray levels, appropriately set in the gray level histogram, allow the automated particle mineral separation: automated separating Oxford Instruments Aztec Feature ® packages and a mineral plotting software are applied for mineral particle separation, crystal chemical formula calculation and plotting.
Finally, a microanalytical analysis is performed on each separated mineral particle. The crystal chemical formula is calculated, and the final classification plots are automatically produced for any determined mineral. The final results show good accuracy and analytical ease and assess the proper nature of the unknown eclogite rock sample. Therefore, the proposed analytical protocol is especially recommended in those scenarios where a large flow of microanalytical data is automatically acquired and needs to be processed.
{"title":"A supervised machine learning procedure for EPMA classification and plotting of mineral groups","authors":"R. Cossio , S. Ghignone , A. Borghi , A. Corno , G. Vaggelli","doi":"10.1016/j.acags.2024.100186","DOIUrl":"10.1016/j.acags.2024.100186","url":null,"abstract":"<div><p>An analytical method to automatically characterize rock samples for geological or petrological purposes is here proposed, by applying machine learning approach (ML) as a protocol for saving experimental times and costs.</p><p>Proper machine learning algorithms, applied to automatically acquired microanalytical data (i.e., Electron Probe Micro Analysis, EPMA), carried out with a SEM-EDS microprobe on randomly selected areas from a petrographic polished thin section, are trained, used, tested, and reported.</p><p>Learning and Validation phases are developed with literature mineral databases of electron microprobe analyses on 15 main rock-forming mineral groups. The Prediction phase is tested using an eclogite rock from the Western Alps, considered as an unknown sample: randomly selected areas are acquired as backscattered images whose intervals of gray levels, appropriately set in the gray level histogram, allow the automated particle mineral separation: automated separating Oxford Instruments Aztec Feature ® packages and a mineral plotting software are applied for mineral particle separation, crystal chemical formula calculation and plotting.</p><p>Finally, a microanalytical analysis is performed on each separated mineral particle. The crystal chemical formula is calculated, and the final classification plots are automatically produced for any determined mineral. The final results show good accuracy and analytical ease and assess the proper nature of the unknown eclogite rock sample. Therefore, the proposed analytical protocol is especially recommended in those scenarios where a large flow of microanalytical data is automatically acquired and needs to be processed.</p></div>","PeriodicalId":33804,"journal":{"name":"Applied Computing and Geosciences","volume":"23 ","pages":"Article 100186"},"PeriodicalIF":2.6,"publicationDate":"2024-08-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2590197424000338/pdfft?md5=5f9a7ff05910f5e248a1bc9ca4b633a6&pid=1-s2.0-S2590197424000338-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142006525","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-15DOI: 10.1016/j.acags.2024.100187
Virág Lovász , Ákos Halmai
In the broad field of sensors and 3D information retrieval, bathymetric reconstruction from side-scan sonar imaging is associated with unique technical hurdles. Neural Networks have recently led to promising new solutions in this field, but the available methods tend to be complex and data-intensive in a way typically making their use in a riverine environment impossible. Throughout our work, we have focused on simplifying the problem-handling and treating compatibility with a riverine environment as priority. In our work, Long Short-Term Memory proved to be effective in a surprisingly simple form. Combined with traditional post-processing techniques in the GIS environment, like median filtered focal statistics, our workflow ultimately results in ∼0.259 m median of error on the evaluation dataset of the Dráva River.
{"title":"LSTM-based DEM generation in riverine environment","authors":"Virág Lovász , Ákos Halmai","doi":"10.1016/j.acags.2024.100187","DOIUrl":"10.1016/j.acags.2024.100187","url":null,"abstract":"<div><p>In the broad field of sensors and 3D information retrieval, bathymetric reconstruction from side-scan sonar imaging is associated with unique technical hurdles. Neural Networks have recently led to promising new solutions in this field, but the available methods tend to be complex and data-intensive in a way typically making their use in a riverine environment impossible. Throughout our work, we have focused on simplifying the problem-handling and treating compatibility with a riverine environment as priority. In our work, Long Short-Term Memory proved to be effective in a surprisingly simple form. Combined with traditional post-processing techniques in the GIS environment, like median filtered focal statistics, our workflow ultimately results in ∼0.259 m median of error on the evaluation dataset of the Dráva River.</p></div>","PeriodicalId":33804,"journal":{"name":"Applied Computing and Geosciences","volume":"23 ","pages":"Article 100187"},"PeriodicalIF":2.6,"publicationDate":"2024-08-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S259019742400034X/pdfft?md5=17b4129af31ed050fc8151abebd2cdbf&pid=1-s2.0-S259019742400034X-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142011680","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-08DOI: 10.1016/j.acags.2024.100185
J. Pérez-Aracil , D. Fister , C.M. Marina , C. Peláez-Rodríguez , L. Cornejo-Bueno , P.A. Gutiérrez , M. Giuliani , A. Castelleti , S. Salcedo-Sanz
This paper proposes two hybrid approaches based on Autoencoders (AEs) for long-term temperature prediction. The first algorithm comprises an AE trained to learn temperature patterns, which is then linked to a second AE, used to detect possible anomalies and provide a final temperature prediction. The second proposed approach involves training an AE and then using the resulting latent space as input of a neural network, which will provide the final prediction output. Both approaches are tested in long-term air temperature prediction in European cities: seven European locations where major heat waves occurred have been considered. The long-term temperature prediction for the entire year of the heatwave events has been analysed. Results show that the proposed approaches can obtain accurate long-term (up to 4 weeks) temperature prediction, improving Persistence and Climatology in the benchmark models compared. In heatwave periods, where the persistence of the temperature is extremely high, our approach beat the persistence operator in three locations and works similarly in the rest of the cases, showing the potential of this AE-based method for long-term temperature prediction.
{"title":"Long-term temperature prediction with hybrid autoencoder algorithms","authors":"J. Pérez-Aracil , D. Fister , C.M. Marina , C. Peláez-Rodríguez , L. Cornejo-Bueno , P.A. Gutiérrez , M. Giuliani , A. Castelleti , S. Salcedo-Sanz","doi":"10.1016/j.acags.2024.100185","DOIUrl":"10.1016/j.acags.2024.100185","url":null,"abstract":"<div><p>This paper proposes two hybrid approaches based on Autoencoders (AEs) for long-term temperature prediction. The first algorithm comprises an AE trained to learn temperature patterns, which is then linked to a second AE, used to detect possible anomalies and provide a final temperature prediction. The second proposed approach involves training an AE and then using the resulting latent space as input of a neural network, which will provide the final prediction output. Both approaches are tested in long-term air temperature prediction in European cities: seven European locations where major heat waves occurred have been considered. The long-term temperature prediction for the entire year of the heatwave events has been analysed. Results show that the proposed approaches can obtain accurate long-term (up to 4 weeks) temperature prediction, improving Persistence and Climatology in the benchmark models compared. In heatwave periods, where the persistence of the temperature is extremely high, our approach beat the persistence operator in three locations and works similarly in the rest of the cases, showing the potential of this AE-based method for long-term temperature prediction.</p></div>","PeriodicalId":33804,"journal":{"name":"Applied Computing and Geosciences","volume":"23 ","pages":"Article 100185"},"PeriodicalIF":2.6,"publicationDate":"2024-08-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2590197424000326/pdfft?md5=5456efe65b92894adbcaf61c5ff34ab1&pid=1-s2.0-S2590197424000326-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141993141","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}