首页 > 最新文献

Applied Computing and Geosciences最新文献

英文 中文
Automatic seismic fault detection and surface construction 自动地震断层检测和地表施工
IF 3.2 Q2 COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS Pub Date : 2025-09-01 DOI: 10.1016/j.acags.2025.100287
Xin Liu , Xingyu Zhu , Xupeng He , Yuzhu Wang
This paper proposes an effective approach for automatically building the fault model based on the 3D seismic images via two steps of automatic seismic fault detection and fault surface construction. Automatic seismic fault detection is performed to automatically classify the seismic image into two phases of fault and background using a slightly revised deeplabv3_resnet50 architecture with pretrained parameters provided by PyTorch. The output of the automatic seismic fault detection is a binary image contains fault and background, where one fault may be separated into different fault segments, or several faults are connected with each other which need further distinguish. To reassemble these detected fault segments and construct the fault surface model, four steps are implemented including:1) a morphological workflow is used to separate all connected faults into separated fault segments; 2) the moving least square (MLS) method is used to fit each fault segments as a smooth, one-voxel thickness surface; 3) the weighted principle component analysis (WPCA) method is applied to calculate the normal vector of each surface voxel to judge whether two or more adjacent segments should be combined in one fault surface; 4) MLS method is applied again to fit all surface segments from one fault as an unique fault surface. The final output of the proposed method provides a fault model with well-defined, cleanly separated, labeled fault surfaces that is competent for structure modelling.
本文提出了一种基于三维地震图像自动建立断层模型的有效方法,该方法分为地震断层自动检测和断层表面构造两个步骤。使用PyTorch提供的预训练参数,使用稍微修改的deeplabv3_resnet50架构进行自动地震故障检测,将地震图像自动分类为故障和背景两个阶段。地震断层自动检测的输出是包含断层和背景的二值图像,其中一个断层可能被分割成不同的断层段,或者几个断层相互连接,需要进一步区分。为了对检测到的故障段进行重组并构建故障面模型,实现了四个步骤:1)使用形态学工作流将所有连接的故障分离成独立的故障段;2)采用移动最小二乘(MLS)方法拟合各断层段为光滑的单体素厚度面;3)采用加权主成分分析(WPCA)方法计算每个面素的法向量,判断是否需要在一个断层面上合并两个或多个相邻的断层段;4)再次应用MLS方法拟合一个断层的所有面段作为唯一的断层面。该方法的最终输出提供了一个断层模型,该模型具有定义良好,分离清晰,标记的断层面,可用于构造建模。
{"title":"Automatic seismic fault detection and surface construction","authors":"Xin Liu ,&nbsp;Xingyu Zhu ,&nbsp;Xupeng He ,&nbsp;Yuzhu Wang","doi":"10.1016/j.acags.2025.100287","DOIUrl":"10.1016/j.acags.2025.100287","url":null,"abstract":"<div><div>This paper proposes an effective approach for automatically building the fault model based on the 3D seismic images via two steps of automatic seismic fault detection and fault surface construction. Automatic seismic fault detection is performed to automatically classify the seismic image into two phases of fault and background using a slightly revised deeplabv3_resnet50 architecture with pretrained parameters provided by PyTorch. The output of the automatic seismic fault detection is a binary image contains fault and background, where one fault may be separated into different fault segments, or several faults are connected with each other which need further distinguish. To reassemble these detected fault segments and construct the fault surface model, four steps are implemented including:1) a morphological workflow is used to separate all connected faults into separated fault segments; 2) the moving least square (MLS) method is used to fit each fault segments as a smooth, one-voxel thickness surface; 3) the weighted principle component analysis (WPCA) method is applied to calculate the normal vector of each surface voxel to judge whether two or more adjacent segments should be combined in one fault surface; 4) MLS method is applied again to fit all surface segments from one fault as an unique fault surface. The final output of the proposed method provides a fault model with well-defined, cleanly separated, labeled fault surfaces that is competent for structure modelling.</div></div>","PeriodicalId":33804,"journal":{"name":"Applied Computing and Geosciences","volume":"27 ","pages":"Article 100287"},"PeriodicalIF":3.2,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145044135","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Estimating aboveground biomass using environmental covariates and a machine-learning approach in the Lower Brazos River Basin, Texas, USA 利用环境协变量和机器学习方法估算美国德克萨斯州下布拉索斯河流域的地上生物量
IF 3.2 Q2 COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS Pub Date : 2025-09-01 DOI: 10.1016/j.acags.2025.100289
Birhan Getachew Tikuye, Ram Lakhan Ray
Forest ecosystems play a pivotal role in global carbon sequestration, serving as essential carbon sinks for climate change mitigation, while also providing a range of ecosystem services such as seed dispersal, pollination, pest control, and habitat provisioning. This study aimed to estimate aboveground biomass density (AGBD) using environmental covariates and a machine learning approach from the Global Ecosystem Dynamics Investigation Light Detection And Ranging (GEDI-LiDAR) in the Lower Brazos River Watershed, Texas, USA. Specifically, GEDI Level 4A data from the National Aeronautics and Space Administration (NASA) Oak Ridge National Laboratory Distributed Active Archive Center (ORNL DAAC) was integrated with Landsat-9 Operational Land Imagery (OLI) and Shuttle Radar Topographic Mission (SRTM) data to enhance predictive accuracy for AGBD. Spectral indices, including the Normalized Difference Vegetation Index (NDVI) and Enhanced Vegetation Index (EVI), were derived from Landsat 9 to support AGBD prediction. Three machine learning models, such as Multivariate Adaptive Regression Splines (MARS), Random Forest (RF), and eXtreme Gradient Boosting (XGBoost), were deployed, with performance assessed using the coefficient of determination (R2), root mean square error (RMSE), and mean absolute error (MAE). Among the models, XGBoost achieved the highest predictive accuracy (R2 = 0.43, RMSE = 31.03, MAE = 22.49). The modelling indicated that longitude, latitude, moisture stress indices (MSI), and digital elevation model (DEM) are among the critical predictors for AGBD. The mean AGBD across the watershed was estimated at 72.3 Mg ha-1, corresponding to a total biomass of approximately 66.6 million tons. Evergreen forests showed the highest AGBD values at 110 Mg ha-1, while cultivated lands averaged 33 Mg ha-1. These findings highlight the effectiveness of integrating environmental covariates with machine learning to estimate AGBD from GEDI LiDAR across diverse ecosystems. This approach provides a robust tool for advancing carbon management and climate change mitigation efforts, while also supporting data-driven conservation planning in both forested and agricultural landscapes.
森林生态系统在全球固碳中发挥着关键作用,是减缓气候变化的基本碳汇,同时还提供一系列生态系统服务,如种子传播、授粉、虫害防治和栖息地供应。本研究旨在利用全球生态系统动态调查光探测和测距(GEDI-LiDAR)的环境协变量和机器学习方法估计美国德克萨斯州下布拉索斯河流域的地上生物量密度(AGBD)。具体来说,来自美国国家航空航天局(NASA)橡树岭国家实验室分布式主动档案中心(ORNL DAAC)的GEDI 4A级数据与Landsat-9作战陆地图像(OLI)和航天飞机雷达地形任务(SRTM)数据集成,以提高AGBD的预测精度。利用Landsat 9反演的归一化植被指数(NDVI)和增强植被指数(EVI)等光谱指数支持AGBD预测。部署了三种机器学习模型,如多元自适应样条回归(MARS)、随机森林(RF)和极端梯度增强(XGBoost),并使用决定系数(R2)、均方根误差(RMSE)和平均绝对误差(MAE)来评估性能。其中,XGBoost的预测准确率最高(R2 = 0.43, RMSE = 31.03, MAE = 22.49)。模拟结果表明,经度、纬度、水分应力指数(MSI)和数字高程模型(DEM)是AGBD的重要预测因子。整个流域的平均AGBD估计为72.3 Mg ha-1,相当于总生物量约为6660万吨。常绿森林的AGBD值最高,为110 Mg ha-1,而耕地平均为33 Mg ha-1。这些发现强调了将环境协变量与机器学习相结合,以估计GEDI激光雷达在不同生态系统中的AGBD的有效性。这一方法为推进碳管理和减缓气候变化工作提供了强有力的工具,同时也支持以数据为导向的森林和农业景观保护规划。
{"title":"Estimating aboveground biomass using environmental covariates and a machine-learning approach in the Lower Brazos River Basin, Texas, USA","authors":"Birhan Getachew Tikuye,&nbsp;Ram Lakhan Ray","doi":"10.1016/j.acags.2025.100289","DOIUrl":"10.1016/j.acags.2025.100289","url":null,"abstract":"<div><div>Forest ecosystems play a pivotal role in global carbon sequestration, serving as essential carbon sinks for climate change mitigation, while also providing a range of ecosystem services such as seed dispersal, pollination, pest control, and habitat provisioning. This study aimed to estimate aboveground biomass density (AGBD) using environmental covariates and a machine learning approach from the Global Ecosystem Dynamics Investigation Light Detection And Ranging (GEDI-LiDAR) in the Lower Brazos River Watershed, Texas, USA. Specifically, GEDI Level 4A data from the National Aeronautics and Space Administration (NASA) Oak Ridge National Laboratory Distributed Active Archive Center (ORNL DAAC) was integrated with Landsat-9 Operational Land Imagery (OLI) and Shuttle Radar Topographic Mission (SRTM) data to enhance predictive accuracy for AGBD. Spectral indices, including the Normalized Difference Vegetation Index (NDVI) and Enhanced Vegetation Index (EVI), were derived from Landsat 9 to support AGBD prediction. Three machine learning models, such as Multivariate Adaptive Regression Splines (MARS), Random Forest (RF), and eXtreme Gradient Boosting (XGBoost), were deployed, with performance assessed using the coefficient of determination (R<sup>2</sup>), root mean square error (RMSE), and mean absolute error (MAE). Among the models, XGBoost achieved the highest predictive accuracy (R<sup>2</sup> = 0.43, RMSE = 31.03, MAE = 22.49). The modelling indicated that longitude, latitude, moisture stress indices (MSI), and digital elevation model (DEM) are among the critical predictors for AGBD. The mean AGBD across the watershed was estimated at 72.3 Mg ha<sup>-1</sup>, corresponding to a total biomass of approximately 66.6 million tons. Evergreen forests showed the highest AGBD values at 110 Mg ha<sup>-1</sup>, while cultivated lands averaged 33 Mg ha<sup>-1</sup>. These findings highlight the effectiveness of integrating environmental covariates with machine learning to estimate AGBD from GEDI LiDAR across diverse ecosystems. This approach provides a robust tool for advancing carbon management and climate change mitigation efforts, while also supporting data-driven conservation planning in both forested and agricultural landscapes.</div></div>","PeriodicalId":33804,"journal":{"name":"Applied Computing and Geosciences","volume":"27 ","pages":"Article 100289"},"PeriodicalIF":3.2,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145104692","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Machine learning approaches for imputing missing meteorological data in Senegal 塞内加尔丢失气象数据的机器学习方法
IF 3.2 Q2 COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS Pub Date : 2025-08-15 DOI: 10.1016/j.acags.2025.100281
Mory Toure , Nana Ama Browne Klutse , Mamadou Adama Sarr , Md Abul Ehsan Bhuiyan , Annine Duclaire Kenne , Wassila Mamadou Thiaw , Daouda Badiane , Amadou Thierno Gaye , Ousmane Ndiaye , Cheikh Mbow
This study presents the first comprehensive evaluation in West Africa of four imputation methods, Decision Tree (DT), Random Forest (RF), Extreme Gradient Boosting (XGB), and Ordinary Kriging (OK), applied to six core meteorological variables across Senegal over a ten-year period (2015–2024). By simulating realistic missing data scenarios informed by field conditions (e.g., power outages, observer absences, sensor failures), it establishes a robust benchmark for climate data reconstruction using machine learning in resource-constrained settings.
The findings highlight the clear superiority of ensemble learning approaches. XGB consistently outperformed all methods across variables and scenarios, achieving the highest average predictive accuracy with R2 values up to [95 % CI: 0.82–0.88], along with lower Root Mean Square Error (RMSE) and Mean Absolute Error (MAE). RF yielded comparable performance, especially for maximum and minimum temperature (TMAX, TMIN), maintaining strong stability even at 20 % missingness. In contrast, DT performance declined sharply with increased data loss, and OK was constrained by the sparse spatial distribution of meteorological stations, notably impairing its ability to impute precipitation (PRCP) and wind speed (WDSP).
This work contributes a multivariable imputation framework specifically adapted to West African climatic and infrastructural realities. It also integrates block bootstrap methods to quantify uncertainty and derive 95 % confidence intervals for all error metrics. Results confirm that imputation effectiveness is highly variable-dependent: continuous and temporally autocorrelated variables (TMAX, TMIN, dew point temperature — DEWP) are well reconstructed, whereas discontinuous or noisy variables (WDSP and PRCP) remain challenging.
Although ensemble models offer clear advantages, their computational demands and need for hyperparameter tuning may limit real-time implementation in low-resource national meteorological services. Furthermore, the exclusion of satellite or reanalysis inputs may constrain model generalizability.
Ultimately, this study reinforces the role of advanced machine learning methods in improving climate data completeness and reliability in Africa. Although not a substitute for direct observations, imputation emerges as a critical complementary tool to support robust and resilient climate information systems essential for agriculture, public health, and disaster risk management under intensifying climate variability.
本研究首次在西非对决策树(DT)、随机森林(RF)、极端梯度增强(XGB)和普通克里格(OK)四种估算方法进行了综合评估,这些方法应用于塞内加尔10年(2015-2024)期间的六个核心气象变量。通过模拟根据现场条件(例如,停电、观察员缺席、传感器故障)通知的真实丢失数据情景,它为在资源受限的环境下使用机器学习重建气候数据建立了一个强大的基准。研究结果突出了集成学习方法的明显优势。XGB始终优于所有变量和场景的方法,实现最高的平均预测精度,R2值高达[95% CI: 0.82-0.88],同时具有较低的均方根误差(RMSE)和平均绝对误差(MAE)。RF产生了相当的性能,特别是在最高和最低温度(TMAX, TMIN)下,即使在丢失20%时也保持了很强的稳定性。而DT的性能则随着数据丢失的增加而急剧下降,OK受到气象站稀疏空间分布的限制,其估算降水(PRCP)和风速(WDSP)的能力明显受损。这项工作提供了一个特别适应西非气候和基础设施现实的多变量imputation框架。它还集成了块引导方法来量化不确定性,并为所有误差度量导出95%的置信区间。结果证实了插值的有效性是高度变量依赖的:连续和时间自相关的变量(TMAX, TMIN,露点温度- DEWP)可以很好地重建,而不连续或有噪声的变量(WDSP和PRCP)仍然具有挑战性。尽管集成模型具有明显的优势,但其计算需求和对超参数调优的需求可能会限制在资源匮乏的国家气象服务中的实时实施。此外,排除卫星或再分析输入可能会限制模型的泛化性。最终,本研究加强了先进机器学习方法在提高非洲气候数据完整性和可靠性方面的作用。虽然不能替代直接观测,但在气候变异性加剧的情况下,归因作为一种重要的补充工具,可以支持对农业、公共卫生和灾害风险管理至关重要的强大和有弹性的气候信息系统。
{"title":"Machine learning approaches for imputing missing meteorological data in Senegal","authors":"Mory Toure ,&nbsp;Nana Ama Browne Klutse ,&nbsp;Mamadou Adama Sarr ,&nbsp;Md Abul Ehsan Bhuiyan ,&nbsp;Annine Duclaire Kenne ,&nbsp;Wassila Mamadou Thiaw ,&nbsp;Daouda Badiane ,&nbsp;Amadou Thierno Gaye ,&nbsp;Ousmane Ndiaye ,&nbsp;Cheikh Mbow","doi":"10.1016/j.acags.2025.100281","DOIUrl":"10.1016/j.acags.2025.100281","url":null,"abstract":"<div><div>This study presents the first comprehensive evaluation in West Africa of four imputation methods, Decision Tree (DT), Random Forest (RF), Extreme Gradient Boosting (XGB), and Ordinary Kriging (OK), applied to six core meteorological variables across Senegal over a ten-year period (2015–2024). By simulating realistic missing data scenarios informed by field conditions (e.g., power outages, observer absences, sensor failures), it establishes a robust benchmark for climate data reconstruction using machine learning in resource-constrained settings.</div><div>The findings highlight the clear superiority of ensemble learning approaches. XGB consistently outperformed all methods across variables and scenarios, achieving the highest average predictive accuracy with R<sup>2</sup> values up to [95 % CI: 0.82–0.88], along with lower Root Mean Square Error (RMSE) and Mean Absolute Error (MAE). RF yielded comparable performance, especially for maximum and minimum temperature (TMAX, TMIN), maintaining strong stability even at 20 % missingness. In contrast, DT performance declined sharply with increased data loss, and OK was constrained by the sparse spatial distribution of meteorological stations, notably impairing its ability to impute precipitation (PRCP) and wind speed (WDSP).</div><div>This work contributes a multivariable imputation framework specifically adapted to West African climatic and infrastructural realities. It also integrates block bootstrap methods to quantify uncertainty and derive 95 % confidence intervals for all error metrics. Results confirm that imputation effectiveness is highly variable-dependent: continuous and temporally autocorrelated variables (TMAX, TMIN, dew point temperature — DEWP) are well reconstructed, whereas discontinuous or noisy variables (WDSP and PRCP) remain challenging.</div><div>Although ensemble models offer clear advantages, their computational demands and need for hyperparameter tuning may limit real-time implementation in low-resource national meteorological services. Furthermore, the exclusion of satellite or reanalysis inputs may constrain model generalizability.</div><div>Ultimately, this study reinforces the role of advanced machine learning methods in improving climate data completeness and reliability in Africa. Although not a substitute for direct observations, imputation emerges as a critical complementary tool to support robust and resilient climate information systems essential for agriculture, public health, and disaster risk management under intensifying climate variability.</div></div>","PeriodicalId":33804,"journal":{"name":"Applied Computing and Geosciences","volume":"27 ","pages":"Article 100281"},"PeriodicalIF":3.2,"publicationDate":"2025-08-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144885417","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Landslide detection using deep learning on remotely sensed images 基于遥感图像的深度学习滑坡检测
IF 3.2 Q2 COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS Pub Date : 2025-08-13 DOI: 10.1016/j.acags.2025.100278
Yuyang Song , Lina Hao , Weile Li
Natural hazards such as landslides pose significant geological threats that can severely endanger the safety and property of residents in affected areas. Therefore, the prompt detection and accurate localisation of landslides are crucial. With the advancement of remote sensing technology and computational methods, artificial intelligence (AI)-based landslide detection techniques have emerged as effective solutions. Compared to traditional methods, these AI-driven approaches offer enhanced efficiency, accuracy and reliability, improving the speed and precision of landslide detection. They also provide valuable data for disaster prevention, mitigation and the assessment of landslide susceptibility and hazard levels. This study focuses on the western Sichuan region and constructs a historical landslide dataset using Google Earth imagery, which includes 4280 landslide samples (3424 for training and 856 for validation). To augment the dataset, 11 data augmentation techniques were applied, including copy–paste, random horizontal flipping, mosaic, random rotation, random hue, saturation and value transformation, affine transformation, random Gaussian noise, random scaling, random brightness and contrast adjustment, mixup and random cropping. These methods improve the diversity of landslide data, helping deep learning models capture more comprehensive global and local information during optimisation. This research utilises the YOLOv10-n object detection framework, enhanced with RepBlock from EfficientRep, FusedMBConv and MBConv techniques derived from EfficientNetV2, CSCGhostblockv2 from GhostNetv2, CReToNeXt from Damo-YOLO and CSCFocalNeXt. These innovations explore the impact of different backbone architectures on model performance. Additionally, the model incorporates four distinct attention mechanisms—convolutional block attention module (CBAM), global attention mechanism(GAM), sim attention module(SimAM) and selective kernel(SK) attention—to assess their influence on detection accuracy. The detection heads are optimised by substituting with three alternatives—DynamicHead, adaptive spatial feature fusion and real-time detection transformer—to enhance feature integration and investigate their effect on model performance. The results indicate that combining EfficientNetV2 with CBAM and v10Detect yields the highest performance. When applied to the historical landslide dataset from the western Sichuan region, the YOLO-EfficientNetV2 model achieves an average precision of 0.861 and an F1 score of 0.82, with a model size of 5.54 M. This model demonstrates superior capability in accurately identifying landslide locations, addressing the common challenge of balancing detection precision and speed in traditional object detection models, while also reducing parameter size and increasing detection speed.
山体滑坡等自然灾害构成重大地质威胁,严重危及受灾地区居民的安全和财产安全。因此,及时发现和准确定位滑坡是至关重要的。随着遥感技术和计算方法的进步,基于人工智能(AI)的滑坡探测技术已经成为有效的解决方案。与传统方法相比,这些人工智能驱动的方法提高了效率、准确性和可靠性,提高了滑坡探测的速度和精度。它们还为防灾、减灾和评估滑坡易感性和危险程度提供了宝贵的数据。本研究以川西地区为研究对象,利用谷歌地球图像构建了一个滑坡历史数据集,该数据集包含4280个滑坡样本(3424个用于训练,856个用于验证)。为了增强数据集,采用了11种数据增强技术,包括复制粘贴、随机水平翻转、马赛克、随机旋转、随机色调、饱和度和值变换、仿射变换、随机高斯噪声、随机缩放、随机亮度和对比度调整、混合和随机裁剪。这些方法提高了滑坡数据的多样性,帮助深度学习模型在优化过程中捕获更全面的全局和局部信息。本研究利用了YOLOv10-n目标检测框架,增强了来自EfficientRep的RepBlock、来自EfficientNetV2的FusedMBConv和MBConv技术、来自GhostNetv2的CSCGhostblockv2、来自Damo-YOLO的CReToNeXt和CSCFocalNeXt。这些创新探索了不同主干架构对模型性能的影响。此外,该模型还结合了四种不同的注意机制——卷积块注意模块(CBAM)、全局注意机制(GAM)、sim注意模块(SimAM)和选择性核注意(SK)——来评估它们对检测精度的影响。采用动态头、自适应空间特征融合和实时检测变压器三种替代方案对检测头进行优化,以增强特征集成并研究它们对模型性能的影响。结果表明,将EfficientNetV2与CBAM和v10Detect相结合可以产生最高的性能。应用于川西地区历史滑坡数据集,YOLO-EfficientNetV2模型的平均精度为0.861,F1得分为0.82,模型大小为5.54 m,在准确识别滑坡位置方面具有较强的能力,解决了传统目标检测模型在检测精度和速度之间平衡的问题,同时减小了参数大小,提高了检测速度。
{"title":"Landslide detection using deep learning on remotely sensed images","authors":"Yuyang Song ,&nbsp;Lina Hao ,&nbsp;Weile Li","doi":"10.1016/j.acags.2025.100278","DOIUrl":"10.1016/j.acags.2025.100278","url":null,"abstract":"<div><div>Natural hazards such as landslides pose significant geological threats that can severely endanger the safety and property of residents in affected areas. Therefore, the prompt detection and accurate localisation of landslides are crucial. With the advancement of remote sensing technology and computational methods, artificial intelligence (AI)-based landslide detection techniques have emerged as effective solutions. Compared to traditional methods, these AI-driven approaches offer enhanced efficiency, accuracy and reliability, improving the speed and precision of landslide detection. They also provide valuable data for disaster prevention, mitigation and the assessment of landslide susceptibility and hazard levels. This study focuses on the western Sichuan region and constructs a historical landslide dataset using Google Earth imagery, which includes 4280 landslide samples (3424 for training and 856 for validation). To augment the dataset, 11 data augmentation techniques were applied, including copy–paste, random horizontal flipping, mosaic, random rotation, random hue, saturation and value transformation, affine transformation, random Gaussian noise, random scaling, random brightness and contrast adjustment, mixup and random cropping. These methods improve the diversity of landslide data, helping deep learning models capture more comprehensive global and local information during optimisation. This research utilises the YOLOv10-n object detection framework, enhanced with RepBlock from EfficientRep, FusedMBConv and MBConv techniques derived from EfficientNetV2, CSCGhostblockv2 from GhostNetv2, CReToNeXt from Damo-YOLO and CSCFocalNeXt. These innovations explore the impact of different backbone architectures on model performance. Additionally, the model incorporates four distinct attention mechanisms—convolutional block attention module (CBAM), global attention mechanism(GAM), sim attention module(SimAM) and selective kernel(SK) attention—to assess their influence on detection accuracy. The detection heads are optimised by substituting with three alternatives—DynamicHead, adaptive spatial feature fusion and real-time detection transformer—to enhance feature integration and investigate their effect on model performance. The results indicate that combining EfficientNetV2 with CBAM and v10Detect yields the highest performance. When applied to the historical landslide dataset from the western Sichuan region, the YOLO-EfficientNetV2 model achieves an average precision of 0.861 and an F<sub>1</sub> score of 0.82, with a model size of 5.54 M. This model demonstrates superior capability in accurately identifying landslide locations, addressing the common challenge of balancing detection precision and speed in traditional object detection models, while also reducing parameter size and increasing detection speed.</div></div>","PeriodicalId":33804,"journal":{"name":"Applied Computing and Geosciences","volume":"27 ","pages":"Article 100278"},"PeriodicalIF":3.2,"publicationDate":"2025-08-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144917695","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Borehole integrity evaluation utilizing coupled hydraulic thermal and mechanical analyses in robust and pre-optimized finite element simulator 在鲁棒和预优化有限元模拟器中利用耦合水力热力学分析进行井眼完整性评估
IF 3.2 Q2 COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS Pub Date : 2025-08-13 DOI: 10.1016/j.acags.2025.100282
Rached M. Rached , Hussain AlBahrani , Timothy E. Moellendick , J. Carlos Santamarina , Thomas Finkbeiner
A thorough understanding of stress distribution around wellbores is crucial for maintaining wellbore stability, especially in deep wells with complex trajectories and subsurface formations exhibiting coupled mechanical behaviors. This study introduces a new finite-element-based modular simulator designed to address a wide range of challenging drilling and boundary conditions, including the presence or absence of filter cake, high over-pressure, inhomogeneous and anisotropic formations, non-linear constitutive behavior, and deviated wells. The simulator uses finite element modeling to provide accurate stress predictions without the overly conservative assumptions common in existing commercial tools. Each module is pre-tested and validated against published analytical solutions and features a user-friendly interface with minimal input requirements, allowing for quick and robust simulations in both 2D and 3D configurations. The simulator can analyze various phenomena, including time-dependent pore pressure diffusion, temperature-induced stress variations, and the impact of heterogeneous formations and layering on stress concentrations. All pre-tested modules run in <60 s on a mid-range workstation while matching analytical solutions to within 0.2 %. We present several case studies that demonstrate the simulator's advantages over existing commercial tools, with all modules made openly available to facilitate broader application.
深入了解井筒周围的应力分布对于保持井筒稳定性至关重要,特别是在具有复杂轨迹和地表地层耦合力学行为的深井中。该研究介绍了一种新的基于有限元的模块化模拟器,旨在解决各种具有挑战性的钻井和边界条件,包括存在或不存在滤饼、高压、非均质和各向异性地层、非线性本构行为和斜井。该模拟器使用有限元建模来提供准确的应力预测,而不是现有商业工具中常见的过于保守的假设。每个模块都针对发布的分析解决方案进行了预先测试和验证,并具有用户友好的界面,输入要求最小,允许在2D和3D配置中进行快速和强大的模拟。该模拟器可以分析各种现象,包括随时间变化的孔隙压力扩散、温度引起的应力变化,以及非均质地层和分层对应力集中的影响。所有预测试模块在中档工作站上以60秒的速度运行,同时将分析溶液匹配在0.2%以内。我们提供了几个案例研究,展示了模拟器相对于现有商业工具的优势,所有模块都是公开可用的,以促进更广泛的应用。
{"title":"Borehole integrity evaluation utilizing coupled hydraulic thermal and mechanical analyses in robust and pre-optimized finite element simulator","authors":"Rached M. Rached ,&nbsp;Hussain AlBahrani ,&nbsp;Timothy E. Moellendick ,&nbsp;J. Carlos Santamarina ,&nbsp;Thomas Finkbeiner","doi":"10.1016/j.acags.2025.100282","DOIUrl":"10.1016/j.acags.2025.100282","url":null,"abstract":"<div><div>A thorough understanding of stress distribution around wellbores is crucial for maintaining wellbore stability, especially in deep wells with complex trajectories and subsurface formations exhibiting coupled mechanical behaviors. This study introduces a new finite-element-based modular simulator designed to address a wide range of challenging drilling and boundary conditions, including the presence or absence of filter cake, high over-pressure, inhomogeneous and anisotropic formations, non-linear constitutive behavior, and deviated wells. The simulator uses finite element modeling to provide accurate stress predictions without the overly conservative assumptions common in existing commercial tools. Each module is pre-tested and validated against published analytical solutions and features a user-friendly interface with minimal input requirements, allowing for quick and robust simulations in both 2D and 3D configurations. The simulator can analyze various phenomena, including time-dependent pore pressure diffusion, temperature-induced stress variations, and the impact of heterogeneous formations and layering on stress concentrations. All pre-tested modules run in &lt;60 s on a mid-range workstation while matching analytical solutions to within 0.2 %. We present several case studies that demonstrate the simulator's advantages over existing commercial tools, with all modules made openly available to facilitate broader application.</div></div>","PeriodicalId":33804,"journal":{"name":"Applied Computing and Geosciences","volume":"27 ","pages":"Article 100282"},"PeriodicalIF":3.2,"publicationDate":"2025-08-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144865563","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Assessing paleo channel probability for offshore wind farm ground modeling - comparison of multiple-point statistics and sequential indicator simulation 评估海上风电场地面建模的古通道概率——多点统计和顺序指标模拟的比较
IF 3.2 Q2 COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS Pub Date : 2025-08-09 DOI: 10.1016/j.acags.2025.100280
Lennart Siemann, Ramiro Relanez
The presented study investigates the prediction of buried paleo-channels for probabilistic ground modeling of offshore windfarm development areas using geostatistical methods. These channels, common in glaciogenic regions like the North Sea, can pose significant geohazards affecting turbine foundation stability. Conventional 2D seismic data interpretation provides the best estimate of the position but lacks probabilistic assessment, specifically at unexplored locations. Multiple-point statistics (MPS) and sequential indicator simulation (SIS) are applied to quantify the probability of channel features from seismic data, away from seismic lines. MPS utilizes training images to capture complex spatial structures, while SIS relies on variogram models for modeling spatial variability. Results demonstrate that denser seismic line spacing (150 m) yields higher accuracy compared to wider spacings (300 m and 600 m), underscoring the importance of data density in offshore subsurface site characterization. Additionally, the findings indicate that MPS provides lower errors, making it preferable for precise channel location prediction. The selected training image did not have a major impact on the outcome on the tested data. Conversely, SIS offers broader coverage of potential channel locations, which may be advantageous for further de-risking. This research contributes to more informed ground modeling by incorporating probabilistic approaches. Therefore, it supports in offshore wind farm site development by enhancing knowledge of the subsurface at an early stage of wind farm development to aid decisions in windfarm and further site investigation planning.
利用地质统计学方法对海上风电场开发区域概率地面模拟中埋藏古河道的预测进行了研究。这些通道在北海等冰川区很常见,可能会对涡轮机基础的稳定性造成重大的地质危害。传统的二维地震数据解释提供了最佳的位置估计,但缺乏概率评估,特别是在未勘探的位置。采用多点统计(MPS)和顺序指标模拟(SIS)来量化地震数据中远离地震线的通道特征的概率。MPS利用训练图像来捕捉复杂的空间结构,而SIS则依靠变异函数模型来模拟空间变异性。结果表明,较密集的地震线间距(150 m)比较宽的地震线间距(300 m和600 m)具有更高的精度,这强调了数据密度在海上地下场地表征中的重要性。此外,研究结果表明,MPS提供了更低的误差,使其更适合精确的通道位置预测。所选择的训练图像对测试数据的结果没有重大影响。相反,SIS提供了更广泛的潜在渠道位置覆盖范围,这可能有利于进一步降低风险。这项研究通过结合概率方法,有助于更明智的地面建模。因此,它通过在风电场开发的早期阶段增强对地下的了解来支持海上风电场的现场开发,以帮助风电场的决策和进一步的现场调查规划。
{"title":"Assessing paleo channel probability for offshore wind farm ground modeling - comparison of multiple-point statistics and sequential indicator simulation","authors":"Lennart Siemann,&nbsp;Ramiro Relanez","doi":"10.1016/j.acags.2025.100280","DOIUrl":"10.1016/j.acags.2025.100280","url":null,"abstract":"<div><div>The presented study investigates the prediction of buried paleo-channels for probabilistic ground modeling of offshore windfarm development areas using geostatistical methods. These channels, common in glaciogenic regions like the North Sea, can pose significant geohazards affecting turbine foundation stability. Conventional 2D seismic data interpretation provides the best estimate of the position but lacks probabilistic assessment, specifically at unexplored locations. Multiple-point statistics (MPS) and sequential indicator simulation (SIS) are applied to quantify the probability of channel features from seismic data, away from seismic lines. MPS utilizes training images to capture complex spatial structures, while SIS relies on variogram models for modeling spatial variability. Results demonstrate that denser seismic line spacing (150 m) yields higher accuracy compared to wider spacings (300 m and 600 m), underscoring the importance of data density in offshore subsurface site characterization. Additionally, the findings indicate that MPS provides lower errors, making it preferable for precise channel location prediction. The selected training image did not have a major impact on the outcome on the tested data. Conversely, SIS offers broader coverage of potential channel locations, which may be advantageous for further de-risking. This research contributes to more informed ground modeling by incorporating probabilistic approaches. Therefore, it supports in offshore wind farm site development by enhancing knowledge of the subsurface at an early stage of wind farm development to aid decisions in windfarm and further site investigation planning.</div></div>","PeriodicalId":33804,"journal":{"name":"Applied Computing and Geosciences","volume":"27 ","pages":"Article 100280"},"PeriodicalIF":3.2,"publicationDate":"2025-08-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144827361","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Pothole detection and segmentation in the Bushveld Complex using physics-based data augmentation and deep learning 使用基于物理的数据增强和深度学习在Bushveld Complex中进行坑洼检测和分割
IF 3.2 Q2 COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS Pub Date : 2025-08-09 DOI: 10.1016/j.acags.2025.100279
Glen T. Nwaila , Musa S.D. Manzi , Emmanuel John M. Carranza , Raymond J. Durrheim , Hartwig E. Frimmel
Potholes are local depression structures that disrupt stratigraphic continuity, such as in layered igneous intrusions. In the Bushveld Complex (South Africa), potholes range from a few to hundreds of meters in width, and may disrupt orebodies, cause ore loss and pose geotechnical challenges. However, potholes are of scientific value as they are proxies of magma chamber processes that are not directly observable. Unfortunately, it is seldom possible to map the full 3D geometry of potholes directly. Reflection seismics has the potential to map many potholes indirectly. However, the accurate segmentation of potholes in seismic data remains unresolved, particularly using geodata science-based methods. Here, we present a prototype segmentation framework that: (1) uses a physics-based, forward modelling method to synthesize 3D reflection seismic data and augments the training data; and (2) implements a standard deep learning, voxel classification-based pothole detection workflow using the data generated in step (1). Both components of the framework are general enough to permit further development, for example, as deep-learning architectures evolve or as the knowledge of potholes improve. We demonstrate that a self-reinforcing feedback loop of knowledge-driven data engineering and deep learning has the potential to overcome data quality issues in supervised tasks of seismic data analysis. We apply the trained model on augmented data to 3D seismic data acquired from a platinum group element Bushveld Complex orebody and demonstrate that automated pothole prediction is practical. Furthermore, physics-based data augmentation, as opposed to inferential types, provides a realistic path to recursive data augmentation that does not incur problems caused by the use of inferential data synthesis, such as model collapse.
壶穴是破坏地层连续性的局部凹陷结构,如层状火成岩侵入体。在Bushveld Complex(南非),坑洞的宽度从几米到数百米不等,可能会破坏矿体,造成矿石损失,并对岩土工程构成挑战。然而,凹坑是具有科学价值的,因为它们是岩浆房过程的代表,而不是直接观察到的。不幸的是,很少有可能直接绘制坑洞的完整3D几何形状。反射地震有可能间接地绘制出许多坑洞。然而,地震数据中凹坑的准确分割仍然没有解决,特别是使用基于地球数据科学的方法。本文提出了一种原型分割框架:(1)利用基于物理的正演建模方法合成三维反射地震数据,并对训练数据进行增强;(2)使用步骤(1)生成的数据实现了一个标准的深度学习、基于体素分类的坑洞检测工作流。框架的两个组件都足够通用,可以允许进一步开发,例如,随着深度学习架构的发展或对坑洼的了解的提高。我们证明了知识驱动的数据工程和深度学习的自我强化反馈循环有潜力克服地震数据分析监督任务中的数据质量问题。我们将训练好的增强数据模型应用于Bushveld铂族元素复杂矿体的三维地震数据,并证明了自动坑穴预测是可行的。此外,与推理类型相反,基于物理的数据增强为递归数据增强提供了一种现实的途径,不会产生由使用推理数据合成引起的问题,例如模型崩溃。
{"title":"Pothole detection and segmentation in the Bushveld Complex using physics-based data augmentation and deep learning","authors":"Glen T. Nwaila ,&nbsp;Musa S.D. Manzi ,&nbsp;Emmanuel John M. Carranza ,&nbsp;Raymond J. Durrheim ,&nbsp;Hartwig E. Frimmel","doi":"10.1016/j.acags.2025.100279","DOIUrl":"10.1016/j.acags.2025.100279","url":null,"abstract":"<div><div>Potholes are local depression structures that disrupt stratigraphic continuity, such as in layered igneous intrusions. In the Bushveld Complex (South Africa), potholes range from a few to hundreds of meters in width, and may disrupt orebodies, cause ore loss and pose geotechnical challenges. However, potholes are of scientific value as they are proxies of magma chamber processes that are not directly observable. Unfortunately, it is seldom possible to map the full 3D geometry of potholes directly. Reflection seismics has the potential to map many potholes indirectly. However, the accurate segmentation of potholes in seismic data remains unresolved, particularly using geodata science-based methods. Here, we present a prototype segmentation framework that: (1) uses a physics-based, forward modelling method to synthesize 3D reflection seismic data and augments the training data; and (2) implements a standard deep learning, voxel classification-based pothole detection workflow using the data generated in step (1). Both components of the framework are general enough to permit further development, for example, as deep-learning architectures evolve or as the knowledge of potholes improve. We demonstrate that a self-reinforcing feedback loop of knowledge-driven data engineering and deep learning has the potential to overcome data quality issues in supervised tasks of seismic data analysis. We apply the trained model on augmented data to 3D seismic data acquired from a platinum group element Bushveld Complex orebody and demonstrate that automated pothole prediction is practical. Furthermore, physics-based data augmentation, as opposed to inferential types, provides a realistic path to recursive data augmentation that does not incur problems caused by the use of inferential data synthesis, such as model collapse.</div></div>","PeriodicalId":33804,"journal":{"name":"Applied Computing and Geosciences","volume":"27 ","pages":"Article 100279"},"PeriodicalIF":3.2,"publicationDate":"2025-08-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144828175","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Extracting data from maps: Lessons learned from the artificial intelligence for critical mineral assessment competition 从地图中提取数据:从关键矿物评估竞争的人工智能中吸取的经验教训
IF 3.2 Q2 COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS Pub Date : 2025-08-08 DOI: 10.1016/j.acags.2025.100274
Margaret A. Goldman , Graham W. Lederer , Joshua M. Rosera , Garth E. Graham , Asitang Mishra , Alice Yepremyan
The U.S. Geological Survey (USGS), Defense Advanced Projects Research Agency (DARPA), Jet Propulsion Laboratory (JPL), and MITRE ran a 12-week machine learning competition aimed at accelerating development of AI tools for critical mineral assessments. The Artificial Intelligence for Critical Mineral Assessment Competition solicited innovative solutions for two challenges: 1) automated georeferencing of historical geologic and topographic maps, and 2) automated feature extraction from historical maps. Competitors used a new dataset of historical map images to train, validate, and evaluate their models. Automated georeferencing pipelines attained a median root-mean square error of 1.1 km. Prompt-based extraction (i.e., with user input) of polygons, polylines, and points from geologic maps yielded median F1 scores of 0.77, 0.56, 0.35, respectively. Geologic maps pose numerous challenges for AI workflows because they vary significantly. However, despite its short duration, the competition yielded promising results that have since spurred further innovation in this area and led to the development of new AI tools to semi-automate key, time-consuming parts of the assessment workflow.
美国地质调查局(USGS)、国防高级项目研究局(DARPA)、喷气推进实验室(JPL)和MITRE进行了为期12周的机器学习竞赛,旨在加速开发用于关键矿物评估的人工智能工具。关键矿物评估人工智能竞赛针对两个挑战征集创新解决方案:1)历史地质和地形图的自动地理参考,以及2)历史地图的自动特征提取。参赛者使用历史地图图像的新数据集来训练、验证和评估他们的模型。自动地理参考管道的均方根误差中值为1.1公里。从地质图中提取多边形、折线和点(即用户输入)的基于提示的F1得分中值分别为0.77、0.56和0.35。地质图变化很大,给人工智能工作流程带来了许多挑战。然而,尽管比赛持续时间很短,但取得了令人鼓舞的成果,这些成果刺激了该领域的进一步创新,并导致了新的人工智能工具的开发,以实现评估工作流程中关键、耗时部分的半自动化。
{"title":"Extracting data from maps: Lessons learned from the artificial intelligence for critical mineral assessment competition","authors":"Margaret A. Goldman ,&nbsp;Graham W. Lederer ,&nbsp;Joshua M. Rosera ,&nbsp;Garth E. Graham ,&nbsp;Asitang Mishra ,&nbsp;Alice Yepremyan","doi":"10.1016/j.acags.2025.100274","DOIUrl":"10.1016/j.acags.2025.100274","url":null,"abstract":"<div><div>The U.S. Geological Survey (USGS), Defense Advanced Projects Research Agency (DARPA), Jet Propulsion Laboratory (JPL), and MITRE ran a 12-week machine learning competition aimed at accelerating development of AI tools for critical mineral assessments. The Artificial Intelligence for Critical Mineral Assessment Competition solicited innovative solutions for two challenges: 1) automated georeferencing of historical geologic and topographic maps, and 2) automated feature extraction from historical maps. Competitors used a new dataset of historical map images to train, validate, and evaluate their models. Automated georeferencing pipelines attained a median root-mean square error of 1.1 km. Prompt-based extraction (i.e., with user input) of polygons, polylines, and points from geologic maps yielded median F1 scores of 0.77, 0.56, 0.35, respectively. Geologic maps pose numerous challenges for AI workflows because they vary significantly. However, despite its short duration, the competition yielded promising results that have since spurred further innovation in this area and led to the development of new AI tools to semi-automate key, time-consuming parts of the assessment workflow.</div></div>","PeriodicalId":33804,"journal":{"name":"Applied Computing and Geosciences","volume":"27 ","pages":"Article 100274"},"PeriodicalIF":3.2,"publicationDate":"2025-08-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144852248","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Stacking modeling with genetic algorithm-based hyperparameter tuning for uniaxial compressive strength prediction 基于遗传算法的单轴抗压强度预测超参数调整叠加建模
IF 3.2 Q2 COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS Pub Date : 2025-08-07 DOI: 10.1016/j.acags.2025.100276
Tanveer Alam Munshi, Khanum Popi, Labiba Nusrat Jahan, M. Farhad Howladar, Mahamudul Hashan
Measuring rock strength using an uniaxial testing machine is destructive and costly, requiring high-quality rock samples. This work suggests an alternate approach that makes use of machine learning techniques to predict uniaxial compressive strength (UCS). The input parameters for this investigation were derived from 180 datasets containing well log variables such as resistivity (RT), sonic travel time (DT), and gamma-ray (GR), as well as rock properties like density. All these datasets came from a shaly sand reservoir in the Bengal Basin. To forecast UCS, a number of methods were used, such as multilayer perceptron (MLP), random forest (RF), support vector machine (SVM), extreme gradient boosting (XGBoost), and multiple variable regression (MVR). Additionally, a hybrid stacking model that combines these algorithms was developed. Hyperparameter optimization was conducted using grid search and genetic algorithm. A notable contribution of this study lies in the application of both grid search and genetic algorithm (GA) for hyperparameter optimization, implemented across both individual base learners and the stacking ensemble model. Regression metrics including coefficient of determination (R2), mean absolute error (MAE), root mean square error (RMSE), maximum error (MaxE), and minimum error (MinE) were used to assess the effectiveness of the models. The proposed stacking model achieved a high testing R2 of 0.9762, outperforming individual models. The methodology provided in this paper can assist engineers and researchers in quickly and precisely determining the strength of reservoir rock by using a few log features, hence decreasing the reliance on labor-intensive and time-consuming laboratory work.
使用单轴试验机测量岩石强度具有破坏性且成本高昂,需要高质量的岩石样品。这项工作提出了一种替代方法,即利用机器学习技术来预测单轴抗压强度(UCS)。该研究的输入参数来自180个数据集,其中包含电阻率(RT)、声波传播时间(DT)、伽马射线(GR)等测井变量,以及密度等岩石属性。所有这些数据集都来自孟加拉盆地的一个泥质砂储层。为了预测UCS,使用了多种方法,如多层感知器(MLP)、随机森林(RF)、支持向量机(SVM)、极端梯度增强(XGBoost)和多变量回归(MVR)。在此基础上,提出了一种结合上述算法的混合叠加模型。采用网格搜索和遗传算法进行超参数优化。本研究的一个显著贡献在于应用网格搜索和遗传算法(GA)进行超参数优化,在个体基础学习器和堆叠集成模型上实现。采用决定系数(R2)、平均绝对误差(MAE)、均方根误差(RMSE)、最大误差(MaxE)和最小误差(MinE)等回归指标评估模型的有效性。提出的叠加模型的检验R2为0.9762,优于单个模型。本文提供的方法可以帮助工程师和研究人员通过使用几个测井特征快速准确地确定储层岩石的强度,从而减少对劳动密集型和耗时的实验室工作的依赖。
{"title":"Stacking modeling with genetic algorithm-based hyperparameter tuning for uniaxial compressive strength prediction","authors":"Tanveer Alam Munshi,&nbsp;Khanum Popi,&nbsp;Labiba Nusrat Jahan,&nbsp;M. Farhad Howladar,&nbsp;Mahamudul Hashan","doi":"10.1016/j.acags.2025.100276","DOIUrl":"10.1016/j.acags.2025.100276","url":null,"abstract":"<div><div>Measuring rock strength using an uniaxial testing machine is destructive and costly, requiring high-quality rock samples. This work suggests an alternate approach that makes use of machine learning techniques to predict uniaxial compressive strength (UCS). The input parameters for this investigation were derived from 180 datasets containing well log variables such as resistivity (RT), sonic travel time (DT), and gamma-ray (GR), as well as rock properties like density. All these datasets came from a shaly sand reservoir in the Bengal Basin. To forecast UCS, a number of methods were used, such as multilayer perceptron (MLP), random forest (RF), support vector machine (SVM), extreme gradient boosting (XGBoost), and multiple variable regression (MVR). Additionally, a hybrid stacking model that combines these algorithms was developed. Hyperparameter optimization was conducted using grid search and genetic algorithm. A notable contribution of this study lies in the application of both grid search and genetic algorithm (GA) for hyperparameter optimization, implemented across both individual base learners and the stacking ensemble model. Regression metrics including coefficient of determination (R<sup>2</sup>), mean absolute error (MAE), root mean square error (RMSE), maximum error (MaxE), and minimum error (MinE) were used to assess the effectiveness of the models. The proposed stacking model achieved a high testing R<sup>2</sup> of 0.9762, outperforming individual models. The methodology provided in this paper can assist engineers and researchers in quickly and precisely determining the strength of reservoir rock by using a few log features, hence decreasing the reliance on labor-intensive and time-consuming laboratory work.</div></div>","PeriodicalId":33804,"journal":{"name":"Applied Computing and Geosciences","volume":"27 ","pages":"Article 100276"},"PeriodicalIF":3.2,"publicationDate":"2025-08-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144809487","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Development of a technique to identify μm-sized organic matter in asteroidal material: An approach using machine learning 在小行星材料中识别μm大小有机物的技术发展:一种使用机器学习的方法
IF 3.2 Q2 COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS Pub Date : 2025-08-07 DOI: 10.1016/j.acags.2025.100277
Rahul Kumar, Katsura Kobayashi, Christian Potiszil, Tak Kunihiro
Asteroidal materials contain organic matter (OM), which records a number of extraterrestrial environments and thus provides a record of Solar System processes. OM contain essential compounds for the origin of life. To understand the origin and evolution of OM, systematic identification and detailed observation using in-situ techniques is required. While both nm- and μm-sized OM were studied previously, only a small portion of a given sample surface was investigated in each study. Here, a novel workflow was developed and applied to identify and classify μm-sized OM on mm-sized asteroidal materials. The workflow involved image processing and machine learning, enabling a comprehensive and non-biased way of identifying, classifying, and measuring the properties of OM. We found that identifying OM is more accurate by classification with machine learning than by clustering. On the approach of classification with machine learning, five algorithms were tested. The random forest algorithm was selected as it scored the highest in 4 out of 5 accuracy parameters during evaluation. The workflow gave modal OM abundances that were consistent with those identified manually, demonstrating that the workflow can accurately identify 1-15 μm-sized OM. The size distribution of OM was modeled using the power-law distribution, giving slope α values that were consistent with fragmentation processes. The shape of the OM was quantified using circularity and solidity, giving a positive correlation and indicating these properties are closely related. Overall, the workflow enabled identification of many OM quickly and accurately and the obtainment of chemical and petrographic information. Such information can help the selection of OM for further in-situ techniques, and elucidate the origin and evolution of OM preserved in asteroidal materials.
小行星物质中含有有机物(OM),它记录了许多地外环境,从而提供了太阳系过程的记录。OM含有生命起源所必需的化合物。为了了解有机质的起源和演化,需要利用原位技术进行系统的鉴定和详细的观察。虽然以前研究过nm和μm尺寸的OM,但每次研究只研究了给定样品表面的一小部分。本文提出了一种新的工作流程,并将其应用于mm级小行星材料上μm级OM的识别和分类。该工作流程涉及图像处理和机器学习,能够以一种全面、无偏见的方式识别、分类和测量OM的属性。我们发现,通过机器学习分类识别OM比通过聚类更准确。在机器学习的分类方法上,测试了五种算法。随机森林算法在评估的5个精度参数中有4个得分最高,因此选择了随机森林算法。该工作流给出的模态OM丰度与人工识别的丰度一致,表明该工作流可以准确识别1-15 μm大小的OM。OM的大小分布采用幂律分布建模,斜率α值与破碎过程一致。OM的形状用圆度和固体度来量化,给出了正相关关系,表明这些性质密切相关。总的来说,该工作流程能够快速准确地识别许多OM,并获得化学和岩石学信息。这些信息可以帮助进一步的原位技术选择OM,并阐明保存在小行星材料中的OM的起源和演化。
{"title":"Development of a technique to identify μm-sized organic matter in asteroidal material: An approach using machine learning","authors":"Rahul Kumar,&nbsp;Katsura Kobayashi,&nbsp;Christian Potiszil,&nbsp;Tak Kunihiro","doi":"10.1016/j.acags.2025.100277","DOIUrl":"10.1016/j.acags.2025.100277","url":null,"abstract":"<div><div>Asteroidal materials contain organic matter (OM), which records a number of extraterrestrial environments and thus provides a record of Solar System processes. OM contain essential compounds for the origin of life. To understand the origin and evolution of OM, systematic identification and detailed observation using in-situ techniques is required. While both nm- and μm-sized OM were studied previously, only a small portion of a given sample surface was investigated in each study. Here, a novel workflow was developed and applied to identify and classify μm-sized OM on mm-sized asteroidal materials. The workflow involved image processing and machine learning, enabling a comprehensive and non-biased way of identifying, classifying, and measuring the properties of OM. We found that identifying OM is more accurate by classification with machine learning than by clustering. On the approach of classification with machine learning, five algorithms were tested. The random forest algorithm was selected as it scored the highest in 4 out of 5 accuracy parameters during evaluation. The workflow gave modal OM abundances that were consistent with those identified manually, demonstrating that the workflow can accurately identify 1-15 μm-sized OM. The size distribution of OM was modeled using the power-law distribution, giving slope α values that were consistent with fragmentation processes. The shape of the OM was quantified using circularity and solidity, giving a positive correlation and indicating these properties are closely related. Overall, the workflow enabled identification of many OM quickly and accurately and the obtainment of chemical and petrographic information. Such information can help the selection of OM for further in-situ techniques, and elucidate the origin and evolution of OM preserved in asteroidal materials.</div></div>","PeriodicalId":33804,"journal":{"name":"Applied Computing and Geosciences","volume":"27 ","pages":"Article 100277"},"PeriodicalIF":3.2,"publicationDate":"2025-08-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144826846","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Applied Computing and Geosciences
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1