High-dimensional imbalanced data classification is a challenging issue in real-world applications, where massive invalid features and class imbalance severely impede the behavior of classifiers. Due to high-dimensional features, imbalanced approaches suffer hardship in yielding adequate results. To tackle these issues, this paper proposes an enriched multi-view ensemble approach (EMEA), aiming to construct an accurate and resilient classifier ensemble system for high-dimensional class-skewed data. First, an enriched multi-view optimization (EMO) is designed to extract effective and diverse features from high-dimensional imbalanced data, it promotes the classification ability through subview learning on multiple diverse scenarios. Then a prioritized integration of subviews (PIS) is developed to conduct selective integration for subviews, aiming to construct a high-quality view that enhances decision-making for high-dimensional imbalanced data classification. Finally, EMEA employs resampling to construct a balanced subset, mitigating the impact of class imbalance on the base classifier. The experiments on 16 high-dimensional class-skewed datasets demonstrate that EMEA is superior to other mainstream imbalanced ensemble approaches.
{"title":"Enriched multi-view ensemble approach for high-dimensional imbalanced data classification","authors":"Yuhong Xu , Dongyi Ding , Peijie Huang , Zhiwen Yu , C.L. Philip Chen","doi":"10.1016/j.engappai.2026.113940","DOIUrl":"10.1016/j.engappai.2026.113940","url":null,"abstract":"<div><div>High-dimensional imbalanced data classification is a challenging issue in real-world applications, where massive invalid features and class imbalance severely impede the behavior of classifiers. Due to high-dimensional features, imbalanced approaches suffer hardship in yielding adequate results. To tackle these issues, this paper proposes an enriched multi-view ensemble approach (EMEA), aiming to construct an accurate and resilient classifier ensemble system for high-dimensional class-skewed data. First, an enriched multi-view optimization (EMO) is designed to extract effective and diverse features from high-dimensional imbalanced data, it promotes the classification ability through subview learning on multiple diverse scenarios. Then a prioritized integration of subviews (PIS) is developed to conduct selective integration for subviews, aiming to construct a high-quality view that enhances decision-making for high-dimensional imbalanced data classification. Finally, EMEA employs resampling to construct a balanced subset, mitigating the impact of class imbalance on the base classifier. The experiments on 16 high-dimensional class-skewed datasets demonstrate that EMEA is superior to other mainstream imbalanced ensemble approaches.</div></div>","PeriodicalId":50523,"journal":{"name":"Engineering Applications of Artificial Intelligence","volume":"167 ","pages":"Article 113940"},"PeriodicalIF":8.0,"publicationDate":"2026-01-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146038587","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-01-23DOI: 10.1016/j.engappai.2025.113712
Ala Saleh Alluhaidan , Amal M. Aqlan , Mashael Maashi , Ahmed Alsayat , Mashail N. Alkhomsan , Faten Derouez , Rakan Alanazi , Tawfiq Hasanin
The healthcare field has undergone a significant shift in the last few years with the advent of data streaming expertise. Data streaming refers to the constant transfer and study of real-time data from multiple sources. In the healthcare environment, data streaming enables healthcare providers to monitor patients' health, predict health issues, and provide personalized care. Real-time observation of patient well-being and predictive analytics for disease analysis and prevention have become gradually significant in healthcare, as they permit healthcare providers to perceive probable health problems before they arise and occur before they become severe. Consumer electronics health technology has transformed health monitoring by permitting constant tracking of crucial signs, physical activity, and other health restrictions. Incorporating artificial intelligence (AI) and deep learning (DL) into consumer electronic devices promises to improve personalized healthcare by aiding real-time data study and early recognition of health problems. In this manuscript, a Personal Health Monitoring with Predictive Analytics and Consumer Electronics using Dimensionality Reduction and Ensemble Classifiers (PHMPACE-DREC) model is presented. The intention is to propose a consumer electronics method for real-time health monitoring and predictive analytics using advanced models to enable proactive and personalized healthcare solutions. To accomplish that, the PHMPACE-DREC model involves a data pre-processing stage initially by applying min-max normalization to convert the input data into a suitable format. Next, the feature selection step is applied, which is a critical stage as it decreases the data dimensionality and enhances efficiency by using three methods, such as Fast Correlation-based Filter Feature (FCBF), Recursive Feature Elimination (RFE), and Least Absolute Shrinkage and Selection Operator (LASSO). Finally, the classification process is performed by the three ensemble classifiers, such as Elman Neural Network (ENN), Deep Q-Network (DQN), and Conditional Variational Autoencoder (CVAE). The experimental analysis of the PHMPACE-DREC approach portrayed a superior accuracy value of 99.11 % over existing methods under the Wearables dataset.
{"title":"Revolutionizing artificial intelligence enabled predictive analytics with smart consumer electronics for real-time healthcare monitoring","authors":"Ala Saleh Alluhaidan , Amal M. Aqlan , Mashael Maashi , Ahmed Alsayat , Mashail N. Alkhomsan , Faten Derouez , Rakan Alanazi , Tawfiq Hasanin","doi":"10.1016/j.engappai.2025.113712","DOIUrl":"10.1016/j.engappai.2025.113712","url":null,"abstract":"<div><div>The healthcare field has undergone a significant shift in the last few years with the advent of data streaming expertise. Data streaming refers to the constant transfer and study of real-time data from multiple sources. In the healthcare environment, data streaming enables healthcare providers to monitor patients' health, predict health issues, and provide personalized care. Real-time observation of patient well-being and predictive analytics for disease analysis and prevention have become gradually significant in healthcare, as they permit healthcare providers to perceive probable health problems before they arise and occur before they become severe. Consumer electronics health technology has transformed health monitoring by permitting constant tracking of crucial signs, physical activity, and other health restrictions. Incorporating artificial intelligence (AI) and deep learning (DL) into consumer electronic devices promises to improve personalized healthcare by aiding real-time data study and early recognition of health problems. In this manuscript, a Personal Health Monitoring with Predictive Analytics and Consumer Electronics using Dimensionality Reduction and Ensemble Classifiers (PHMPACE-DREC) model is presented. The intention is to propose a consumer electronics method for real-time health monitoring and predictive analytics using advanced models to enable proactive and personalized healthcare solutions. To accomplish that, the PHMPACE-DREC model involves a data pre-processing stage initially by applying min-max normalization to convert the input data into a suitable format. Next, the feature selection step is applied, which is a critical stage as it decreases the data dimensionality and enhances efficiency by using three methods, such as Fast Correlation-based Filter Feature (FCBF), Recursive Feature Elimination (RFE), and Least Absolute Shrinkage and Selection Operator (LASSO). Finally, the classification process is performed by the three ensemble classifiers, such as Elman Neural Network (ENN), Deep Q-Network (DQN), and Conditional Variational Autoencoder (CVAE). The experimental analysis of the PHMPACE-DREC approach portrayed a superior accuracy value of 99.11 % over existing methods under the Wearables dataset.</div></div>","PeriodicalId":50523,"journal":{"name":"Engineering Applications of Artificial Intelligence","volume":"167 ","pages":"Article 113712"},"PeriodicalIF":8.0,"publicationDate":"2026-01-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146038714","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
This paper presents a Fourier-enhanced dynamic sequence-to-sequence latent graph neural network (Seq2SeqLatentGNN), a deep learning architecture for multi-node spatiotemporal forecasting in hydroelectric reservoir systems. The model integrates three key components: (i) a custom Fourier layer that analyzes global temporal patterns through frequency-domain transformations, (ii) a latent correlation graph convolutional network that infers relational structures between monitoring stations without requiring predefined adjacency matrices, and (iii) an attention-based sequence-to-sequence model that processes temporal dependencies while enabling multi-step forecasting. The architecture simultaneously learns graph structure and forecasting tasks, adapting to changing spatial relationships between reservoir nodes. The proposed architecture was evaluated using a comprehensive dataset derived from 19 interconnected hydroelectric reservoirs located in southern Brazil. The dataset encompasses multiple years of high-resolution (hourly) measurements, including reservoir water levels, inflow and outflow rates, precipitation records, and energy production metrics. Experimental results demonstrate that Seq2SeqLatentGNN achieves superior performance compared to conventional statistical models and contemporary machine learning methods, as measured by standard error metrics. Analysis of the learned latent correlations reveals meaningful spatial dependencies that align with hydrological principles. The model exhibits consistent performance across varying temporal patterns, adapts to regime transitions, and captures both periodic and nonstationary dynamics. The proposed architecture contributes to spatiotemporal forecasting by combining spectral processing, dynamic graph learning, and sequence modeling in a unified framework applicable to systems with evolving connectivity patterns.
{"title":"Fourier-enhanced sequence-to-sequence latent graph neural networks for multi-node spatiotemporal forecasting in a hydroelectric reservoir","authors":"Laio Oriel Seman , Stefano Frizzo Stefenon , Kin-Choong Yow , Leandro dos Santos Coelho , Viviana Cocco Mariani","doi":"10.1016/j.engappai.2026.113939","DOIUrl":"10.1016/j.engappai.2026.113939","url":null,"abstract":"<div><div>This paper presents a Fourier-enhanced dynamic sequence-to-sequence latent graph neural network (Seq2SeqLatentGNN), a deep learning architecture for multi-node spatiotemporal forecasting in hydroelectric reservoir systems. The model integrates three key components: (i) a custom Fourier layer that analyzes global temporal patterns through frequency-domain transformations, (ii) a latent correlation graph convolutional network that infers relational structures between monitoring stations without requiring predefined adjacency matrices, and (iii) an attention-based sequence-to-sequence model that processes temporal dependencies while enabling multi-step forecasting. The architecture simultaneously learns graph structure and forecasting tasks, adapting to changing spatial relationships between reservoir nodes. The proposed architecture was evaluated using a comprehensive dataset derived from 19 interconnected hydroelectric reservoirs located in southern Brazil. The dataset encompasses multiple years of high-resolution (hourly) measurements, including reservoir water levels, inflow and outflow rates, precipitation records, and energy production metrics. Experimental results demonstrate that Seq2SeqLatentGNN achieves superior performance compared to conventional statistical models and contemporary machine learning methods, as measured by standard error metrics. Analysis of the learned latent correlations reveals meaningful spatial dependencies that align with hydrological principles. The model exhibits consistent performance across varying temporal patterns, adapts to regime transitions, and captures both periodic and nonstationary dynamics. The proposed architecture contributes to spatiotemporal forecasting by combining spectral processing, dynamic graph learning, and sequence modeling in a unified framework applicable to systems with evolving connectivity patterns.</div></div>","PeriodicalId":50523,"journal":{"name":"Engineering Applications of Artificial Intelligence","volume":"167 ","pages":"Article 113939"},"PeriodicalIF":8.0,"publicationDate":"2026-01-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146038585","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-01-23DOI: 10.1016/j.engappai.2026.113901
Yunyin Li , Shudong Wang , Yuanyuan Zhang , Chuanru Ren , Shanchen Pang , Tiyao Liu , Yingye Liu
The mechanisms of action for numerous drugs involve micro ribonucleic acids (miRNAs), highlighting the significance of studying miRNA-mediated drug sensitivity in drug discovery and disease treatment. Despite advancements in computational approaches, challenges persist in effectively extracting drug and miRNA features and accurately predicting their associations. The existing similarity networks of drugs and miRNAs are in urgent need of supplementing comprehensive similarity information. In addition, most computational methods extract only single-level features without combining information from different levels, limiting the performance of the models. To overcome these challenges, we combine Variational graph auto-encoder and Collaborative matrix factorization to identify MiRNA-Drug Sensitivity (VCMDS). VCMDS figures out the Gaussian Interaction Profile (GIP) kernel similarities between drugs and miRNAs and adds these measurements to each their network. By aggregating multiple sources of information, the GIP kernel similarity provides useful information by considering a wider network of interactions and measuring similarity more accurately. Subsequently, it extracts features of miRNAs and drugs at various levels by applying variational graph auto-encoder and collaborative matrix factorization. Linear and nonlinear features can be combined to produce high-quality features and thus improve the prediction performance. Finally, predicted scores are obtained using a fully connected network. VCMDS achieves an average Area Under Curve (AUC) of 0.9632 in the 5-fold Cross-Validation (CV) experiment, outperforming other competitive methods. Two types of case studies further demonstrate the effectiveness of VCMDS.
{"title":"Micro ribonucleic acids-drug sensitivity prediction by variational graph auto-encoder and collaborative matrix factorization","authors":"Yunyin Li , Shudong Wang , Yuanyuan Zhang , Chuanru Ren , Shanchen Pang , Tiyao Liu , Yingye Liu","doi":"10.1016/j.engappai.2026.113901","DOIUrl":"10.1016/j.engappai.2026.113901","url":null,"abstract":"<div><div>The mechanisms of action for numerous drugs involve micro ribonucleic acids (miRNAs), highlighting the significance of studying miRNA-mediated drug sensitivity in drug discovery and disease treatment. Despite advancements in computational approaches, challenges persist in effectively extracting drug and miRNA features and accurately predicting their associations. The existing similarity networks of drugs and miRNAs are in urgent need of supplementing comprehensive similarity information. In addition, most computational methods extract only single-level features without combining information from different levels, limiting the performance of the models. To overcome these challenges, we combine Variational graph auto-encoder and Collaborative matrix factorization to identify MiRNA-Drug Sensitivity (VCMDS). VCMDS figures out the Gaussian Interaction Profile (GIP) kernel similarities between drugs and miRNAs and adds these measurements to each their network. By aggregating multiple sources of information, the GIP kernel similarity provides useful information by considering a wider network of interactions and measuring similarity more accurately. Subsequently, it extracts features of miRNAs and drugs at various levels by applying variational graph auto-encoder and collaborative matrix factorization. Linear and nonlinear features can be combined to produce high-quality features and thus improve the prediction performance. Finally, predicted scores are obtained using a fully connected network. VCMDS achieves an average Area Under Curve (AUC) of 0.9632 in the 5-fold Cross-Validation (CV) experiment, outperforming other competitive methods. Two types of case studies further demonstrate the effectiveness of VCMDS.</div></div>","PeriodicalId":50523,"journal":{"name":"Engineering Applications of Artificial Intelligence","volume":"167 ","pages":"Article 113901"},"PeriodicalIF":8.0,"publicationDate":"2026-01-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146038590","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-01-23DOI: 10.1016/j.engappai.2026.113739
Yujing Tang , Yang Fu , Qin Cai , Jieping Wu , Qi Wang , Guoqiang Gao
As key equipment for high-speed rail power transmission and the connection of high-voltage systems, the cable terminals are crucial to ensuring the stable operation of the railway system. However, the existing detection methods for cable terminals are easily affected by on-site noise and have low detection accuracy. Therefore, this paper proposes a method for detecting interface defect status of high-speed cable terminals based on the electric field strength feature set and multi-kernel support vector machine (MK-SVM). Firstly, a spatial electric field detection platform was built to extract the electric field intensity of the prefabricated defective cable terminals of different lengths. Secondly, the optimization of the characteristic parameters of electric field strength of defective cable terminals was realized based on the Pearson coefficient method. In order to improve the recognition effect and model generalization ability, a MK-SVM combining linear kernel function and radial basis kernel function was proposed. Finally, a comparative study was conducted on the optimization effects of particle swarm algorithm, firefly algorithm, simulated annealing algorithm and genetic algorithm on MK-SVM. Research has shown that using genetic algorithm for parameter optimization of multi-core SVM has the best performance, with recognition accuracy, average precision, average recall, and average F1 score of 95.6 %, 96 %, 95.6 %, and 0.96, respectively. Compared with the unoptimized SVM, the four feature parameters increased by 8.9 %, 7.9 %, 8.9 %, and 9.6 %, respectively.
{"title":"Research on cable terminal interface defect state detection based on electric field characteristics and multi-core improved support vector machine","authors":"Yujing Tang , Yang Fu , Qin Cai , Jieping Wu , Qi Wang , Guoqiang Gao","doi":"10.1016/j.engappai.2026.113739","DOIUrl":"10.1016/j.engappai.2026.113739","url":null,"abstract":"<div><div>As key equipment for high-speed rail power transmission and the connection of high-voltage systems, the cable terminals are crucial to ensuring the stable operation of the railway system. However, the existing detection methods for cable terminals are easily affected by on-site noise and have low detection accuracy. Therefore, this paper proposes a method for detecting interface defect status of high-speed cable terminals based on the electric field strength feature set and multi-kernel support vector machine (MK-SVM). Firstly, a spatial electric field detection platform was built to extract the electric field intensity of the prefabricated defective cable terminals of different lengths. Secondly, the optimization of the characteristic parameters of electric field strength of defective cable terminals was realized based on the Pearson coefficient method. In order to improve the recognition effect and model generalization ability, a MK-SVM combining linear kernel function and radial basis kernel function was proposed. Finally, a comparative study was conducted on the optimization effects of particle swarm algorithm, firefly algorithm, simulated annealing algorithm and genetic algorithm on MK-SVM. Research has shown that using genetic algorithm for parameter optimization of multi-core SVM has the best performance, with recognition accuracy, average precision, average recall, and average F1 score of 95.6 %, 96 %, 95.6 %, and 0.96, respectively. Compared with the unoptimized SVM, the four feature parameters increased by 8.9 %, 7.9 %, 8.9 %, and 9.6 %, respectively.</div></div>","PeriodicalId":50523,"journal":{"name":"Engineering Applications of Artificial Intelligence","volume":"167 ","pages":"Article 113739"},"PeriodicalIF":8.0,"publicationDate":"2026-01-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146038528","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-01-23DOI: 10.1016/j.engappai.2026.113943
Yafei Qi , Chen Wang , Zhaoning Zhang , Yaping Liu , Yongmin Zhang
Knowledge distillation (KD) represents a fundamental artificial intelligence (AI) technique for model compression and optimization. In computer vision AI applications, most KD methods use Kullback–Leibler (KL) divergence to align teacher–student output probabilities, but often neglect crucial negative aspects of teacher “dark knowledge” by underweighting low-probability signals. This limitation leads to suboptimal logit mimicry and unbalanced knowledge transfer to the student network. In this paper, we investigate the impact of this imbalance and propose a novel method, named Balance Divergence Distillation (BDD). By introducing a compensatory operation using reverse KL divergence, our method can improve the modeling of the extremely small values in the negative from the teacher and preserve the learning capacity for the positive. Furthermore, we test the impact of different temperature coefficients adjustments, which can lead to further balance in knowledge transfer. The evaluation results demonstrate that our method achieves accuracy improvements of for lightweight student networks over standard KD methods on both Canadian Institute for Advanced Research 100 classes(CIFAR-100) and ImageNet datasets. Additionally, when applied to semantic segmentation, our approach enhances the student by 4.55% in mean Intersection over Union (mIoU) compared to the baseline on the Cityscapes dataset. These experiments confirm that our method provides a simple yet highly effective solution that can be seamlessly integrated with various KD frameworks across different vision tasks.
{"title":"Balance divergence for knowledge distillation","authors":"Yafei Qi , Chen Wang , Zhaoning Zhang , Yaping Liu , Yongmin Zhang","doi":"10.1016/j.engappai.2026.113943","DOIUrl":"10.1016/j.engappai.2026.113943","url":null,"abstract":"<div><div>Knowledge distillation (KD) represents a fundamental artificial intelligence (AI) technique for model compression and optimization. In computer vision AI applications, most KD methods use Kullback–Leibler (KL) divergence to align teacher–student output probabilities, but often neglect crucial negative aspects of teacher “dark knowledge” by underweighting low-probability signals. This limitation leads to suboptimal logit mimicry and unbalanced knowledge transfer to the student network. In this paper, we investigate the impact of this imbalance and propose a novel method, named Balance Divergence Distillation (BDD). By introducing a compensatory operation using reverse KL divergence, our method can improve the modeling of the extremely small values in the negative from the teacher and preserve the learning capacity for the positive. Furthermore, we test the impact of different temperature coefficients adjustments, which can lead to further balance in knowledge transfer. The evaluation results demonstrate that our method achieves accuracy improvements of <span><math><mrow><mn>1</mn><mtext>%</mtext><mo>∼</mo><mn>3</mn><mtext>%</mtext></mrow></math></span> for lightweight student networks over standard KD methods on both Canadian Institute for Advanced Research 100 classes(CIFAR-100) and ImageNet datasets. Additionally, when applied to semantic segmentation, our approach enhances the student by 4.55% in mean Intersection over Union (mIoU) compared to the baseline on the Cityscapes dataset. These experiments confirm that our method provides a simple yet highly effective solution that can be seamlessly integrated with various KD frameworks across different vision tasks.</div></div>","PeriodicalId":50523,"journal":{"name":"Engineering Applications of Artificial Intelligence","volume":"167 ","pages":"Article 113943"},"PeriodicalIF":8.0,"publicationDate":"2026-01-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146038535","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-01-23DOI: 10.1016/j.engappai.2026.113935
Yue Zhang , Zheng Gong , Changhai Zhang , Yongquan Zhang , Chao Yin , Xubin Wang , Tiandong Zhang , Xiajie Yi , Huajie Yi , Qi Wang
The design of high-performance polymer dielectrics for capacitor energy storage is crucial but often hindered by time-consuming, resource-intensive development cycles. Polyetherimide is a promising matrix material, yet its performance is limited by a low dielectric constant and breakdown strength. To accelerate the design process, we propose and validate an integrated computational framework combining molecular dynamics simulations with interpretable machine learning. In terms of the Artificial Intelligence contribution, a weighted ensemble model was developed from a dual database of molecular dynamics parameters and molecular descriptors to predict dielectric property in Polyetherimide-based composites. The model was then deconstructed using the SHapley Additive exPlanations framework, which unveiled a multi-scale design hierarchy. This analysis revealed that filler weight fraction and intrinsic dielectric constant are the most dominant predictors, followed by interfacial compatibility and molecular polarity. Regarding the engineering application, to validate our computational approach, model-selected Benzil and Acetophenone were fabricated into composite films. Experimental results confirmed the model's high accuracy, identifying optimal contents of weight percent of 15 wt% for Benzil and 10 wt% for Acetophenone. Notably, the Polyetherimide-based composite with 10 wt% of Acetophenone achieved an excellent discharge energy density of 10.3 J/cm3, representing a 58 % enhancement over pristine Polyetherimide. Ultimately, this study not only developed a promising material but established a reliable data-driven methodology providing clear guidance for designing next-generation polymer dielectrics.
{"title":"Predicting dielectric properties of polyetherimide-based composite via combined molecular dynamics simulation and machine learning","authors":"Yue Zhang , Zheng Gong , Changhai Zhang , Yongquan Zhang , Chao Yin , Xubin Wang , Tiandong Zhang , Xiajie Yi , Huajie Yi , Qi Wang","doi":"10.1016/j.engappai.2026.113935","DOIUrl":"10.1016/j.engappai.2026.113935","url":null,"abstract":"<div><div>The design of high-performance polymer dielectrics for capacitor energy storage is crucial but often hindered by time-consuming, resource-intensive development cycles. Polyetherimide is a promising matrix material, yet its performance is limited by a low dielectric constant and breakdown strength. To accelerate the design process, we propose and validate an integrated computational framework combining molecular dynamics simulations with interpretable machine learning. In terms of the Artificial Intelligence contribution, a weighted ensemble model was developed from a dual database of molecular dynamics parameters and molecular descriptors to predict dielectric property in Polyetherimide-based composites. The model was then deconstructed using the SHapley Additive exPlanations framework, which unveiled a multi-scale design hierarchy. This analysis revealed that filler weight fraction and intrinsic dielectric constant are the most dominant predictors, followed by interfacial compatibility and molecular polarity. Regarding the engineering application, to validate our computational approach, model-selected Benzil and Acetophenone were fabricated into composite films. Experimental results confirmed the model's high accuracy, identifying optimal contents of weight percent of 15 wt% for Benzil and 10 wt% for Acetophenone. Notably, the Polyetherimide-based composite with 10 wt% of Acetophenone achieved an excellent discharge energy density of 10.3 J/cm<sup>3</sup>, representing a 58 % enhancement over pristine Polyetherimide. Ultimately, this study not only developed a promising material but established a reliable data-driven methodology providing clear guidance for designing next-generation polymer dielectrics.</div></div>","PeriodicalId":50523,"journal":{"name":"Engineering Applications of Artificial Intelligence","volume":"167 ","pages":"Article 113935"},"PeriodicalIF":8.0,"publicationDate":"2026-01-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146038589","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
A wide range of mechanical, electrical, and plumbing (MEP) components are mounted along metro tunnel linings, where subtle spatial displacements caused by loosening or deformation may intrude into the train clearance envelope. Detecting such early-stage deviations is challenging due to occlusion, low illumination, and the dense arrangement of facilities in tunnel point clouds acquired using simultaneous localization and mapping (SLAM). To address this problem, this study proposes a training-free few-shot semantic segmentation framework for clearance intrusion detection. The model integrates geometry-based descriptors (linearity, planarity, verticality), a scale factor control mechanism for multi-scale feature enhancement, and a confidence-based filtering strategy to suppress uncertain predictions. Experiments were conducted on metro tunnel point clouds acquired using a backpack-mounted light detection and ranging (LiDAR) SLAM system, with segmentation performed using 1 m under a single-class few-shot setting. The proposed method achieves a mean intersection over union (mIoU) of 78.4 %, while requiring only a small support set of 15 blocks, and the reconstructed axes of MEP facilities enable deviation detection below 1 cm relative to reference inspection epochs. These results demonstrate that the proposed framework provides a practical and robust solution for early-stage clearance intrusion risk assessment in metro tunnel environments.
{"title":"Few-shot semantic segmentation for clearance intrusion risk detection in metro tunnel point clouds","authors":"Wenbo Qin , Yuxiang Wang , Shangbin Gao , Cheng Zhou","doi":"10.1016/j.engappai.2026.113909","DOIUrl":"10.1016/j.engappai.2026.113909","url":null,"abstract":"<div><div>A wide range of mechanical, electrical, and plumbing (MEP) components are mounted along metro tunnel linings, where subtle spatial displacements caused by loosening or deformation may intrude into the train clearance envelope. Detecting such early-stage deviations is challenging due to occlusion, low illumination, and the dense arrangement of facilities in tunnel point clouds acquired using simultaneous localization and mapping (SLAM). To address this problem, this study proposes a training-free few-shot semantic segmentation framework for clearance intrusion detection. The model integrates geometry-based descriptors (linearity, planarity, verticality), a scale factor control mechanism for multi-scale feature enhancement, and a confidence-based filtering strategy to suppress uncertain predictions. Experiments were conducted on metro tunnel point clouds acquired using a backpack-mounted light detection and ranging (LiDAR) SLAM system, with segmentation performed using 1 m under a single-class few-shot setting. The proposed method achieves a mean intersection over union (mIoU) of 78.4 %, while requiring only a small support set of 15 blocks, and the reconstructed axes of MEP facilities enable deviation detection below 1 cm relative to reference inspection epochs. These results demonstrate that the proposed framework provides a practical and robust solution for early-stage clearance intrusion risk assessment in metro tunnel environments.</div></div>","PeriodicalId":50523,"journal":{"name":"Engineering Applications of Artificial Intelligence","volume":"167 ","pages":"Article 113909"},"PeriodicalIF":8.0,"publicationDate":"2026-01-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146038575","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-01-23DOI: 10.1016/j.engappai.2026.113887
Alex J. Garzón-Orduña , Oscar E. Coronado-Hernández , Alfonso Arrieta-Pastrana , Helena M. Ramos , Modesto Pérez-Sánchez
Efficient leakage management in Water Distribution Networks is essential to achieving the objectives of Smart Cities and Sustainable Development Goal 6. Conventional hydraulic models based on Extended Period Simulation cannot reproduce the short transients generated by rapid valve manoeuvres, particularly in pressure-reducing and operational valves. These events create inertial effects and pressure fluctuations that strongly influence leakage behaviour, especially when valve resistance changes with real operational timings, whether manual or actuator-driven. This study develops a hybrid Physics-Informed Machine Learning framework that couples an extended Rigid Water Column Model—derived from mass and momentum conservation and incorporating time-dependent valve resistance—with supervised Machine Learning algorithms. The approach enhances leakage prediction under transient conditions by combining physical interpretability with data-driven adaptability. Simulations from the Rigid Water Column Model and Extended Period Simulation were compared across sixteen synthetic scenarios. Four optimised Machine Learning models—Fine Tree, Bagged Trees, Exponential Gaussian Process Regression, and Wide Neural Network—were trained using physically consistent datasets. They achieved Root Mean Square Error values as low as 0.0019 L per second. Tree-based algorithms proved most robust, while Exponential Gaussian Process Regression and Wide Neural Network models showed reduced extrapolation capacity. Statistical stability verified through 95 percent confidence intervals confirmed the physical coherence of predictions. The proposed framework provides a scalable and transferable tool for real-time leakage prediction and valve operation planning within Digital Twin and Supervisory Control and Data Acquisition environments, supporting predictive control, operational decision-making, and sustainable water-infrastructure management.
{"title":"Physics-informed machine learning for hybrid modelling of water leakages induced by rapid valve manoeuvres in water distribution networks","authors":"Alex J. Garzón-Orduña , Oscar E. Coronado-Hernández , Alfonso Arrieta-Pastrana , Helena M. Ramos , Modesto Pérez-Sánchez","doi":"10.1016/j.engappai.2026.113887","DOIUrl":"10.1016/j.engappai.2026.113887","url":null,"abstract":"<div><div>Efficient leakage management in Water Distribution Networks is essential to achieving the objectives of Smart Cities and Sustainable Development Goal 6. Conventional hydraulic models based on Extended Period Simulation cannot reproduce the short transients generated by rapid valve manoeuvres, particularly in pressure-reducing and operational valves. These events create inertial effects and pressure fluctuations that strongly influence leakage behaviour, especially when valve resistance changes with real operational timings, whether manual or actuator-driven. This study develops a hybrid Physics-Informed Machine Learning framework that couples an extended Rigid Water Column Model—derived from mass and momentum conservation and incorporating time-dependent valve resistance—with supervised Machine Learning algorithms. The approach enhances leakage prediction under transient conditions by combining physical interpretability with data-driven adaptability. Simulations from the Rigid Water Column Model and Extended Period Simulation were compared across sixteen synthetic scenarios. Four optimised Machine Learning models—Fine Tree, Bagged Trees, Exponential Gaussian Process Regression, and Wide Neural Network—were trained using physically consistent datasets. They achieved Root Mean Square Error values as low as 0.0019 L per second. Tree-based algorithms proved most robust, while Exponential Gaussian Process Regression and Wide Neural Network models showed reduced extrapolation capacity. Statistical stability verified through 95 percent confidence intervals confirmed the physical coherence of predictions. The proposed framework provides a scalable and transferable tool for real-time leakage prediction and valve operation planning within Digital Twin and Supervisory Control and Data Acquisition environments, supporting predictive control, operational decision-making, and sustainable water-infrastructure management.</div></div>","PeriodicalId":50523,"journal":{"name":"Engineering Applications of Artificial Intelligence","volume":"167 ","pages":"Article 113887"},"PeriodicalIF":8.0,"publicationDate":"2026-01-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146038530","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-01-23DOI: 10.1016/j.engappai.2026.113937
Fei Teng , Liqiang Jin , Junnian Wang , Feng Xiao , Mengdi Guo , Yanbo Zhou , Jin Zhang
In increasingly complex traffic environments, spatiotemporal attention mechanisms have made remarkable advancements in scene-level interaction modelling. However, the deep and multi-scale spatiotemporal representations required for safe and efficient decision-making in intelligent vehicles remain underexplored. Aiming to address this limitation, this study proposes a multimodal trajectory prediction model based on a cross-layer interleaved spatiotemporal attention (CLISTA) mechanism. Compared with conventional spatiotemporal attention frameworks, CLISTA more effectively captures multi-scale spatiotemporal interactions in complex traffic scenes through the alternating fusion of spatial and temporal features across network layers via a cross-layer interleaving structure. Firstly, spatial, dynamic and heading conflict risks are derived from the relative motion between the target vehicle and its neighbours and aggregated into a social grid weight matrix, through which the neighbours' collective influence on the target vehicle is quantified. Secondly, spatial and temporal multi-head attention modules are designed within each layer. By integrating an interleaved ‘spatial–temporal’ stacking strategy with cross-layer skip connections, the model facilitates progressive alignment and deep fusion, ranging from local interactions to long-range dependencies. Subsequently, an intention recognition module is developed. A second-order gated bilinear fusion mechanism is introduced to adaptively model higher-order couplings between local neighbour dynamics and global interaction semantics, thereby yielding a multimodal probability distribution over the target vehicle's driving intentions. Lastly, multimodal trajectory predictions are generated by decoding the fused spatiotemporal features together with the inferred intention information. Experimental results on three benchmark datasets—NGSIM (Next Generation Simulation), AD4CHE (Aerial Dataset for China Congested Highway and Expressway), and highD—demonstrate that CLISTA consistently outperforms the baseline methods. Relative to the next-best model, it reduces average/final displacement errors by 16.67 %/21.23 %, 12.99 %/21.14 % and 10.53 %/21.59 % on NGSIM, AD4CHE and HighD, respectively. Overall, CLISTA offers reliable multi-hypothesis trajectory priors for safe and efficient decision-making in complex traffic scenarios.
{"title":"High-precision multimodal vehicle trajectory prediction model based on cross-layer interleaved spatiotemporal attention mechanism","authors":"Fei Teng , Liqiang Jin , Junnian Wang , Feng Xiao , Mengdi Guo , Yanbo Zhou , Jin Zhang","doi":"10.1016/j.engappai.2026.113937","DOIUrl":"10.1016/j.engappai.2026.113937","url":null,"abstract":"<div><div>In increasingly complex traffic environments, spatiotemporal attention mechanisms have made remarkable advancements in scene-level interaction modelling. However, the deep and multi-scale spatiotemporal representations required for safe and efficient decision-making in intelligent vehicles remain underexplored. Aiming to address this limitation, this study proposes a multimodal trajectory prediction model based on a cross-layer interleaved spatiotemporal attention (CLISTA) mechanism. Compared with conventional spatiotemporal attention frameworks, CLISTA more effectively captures multi-scale spatiotemporal interactions in complex traffic scenes through the alternating fusion of spatial and temporal features across network layers via a cross-layer interleaving structure. Firstly, spatial, dynamic and heading conflict risks are derived from the relative motion between the target vehicle and its neighbours and aggregated into a social grid weight matrix, through which the neighbours' collective influence on the target vehicle is quantified. Secondly, spatial and temporal multi-head attention modules are designed within each layer. By integrating an interleaved ‘spatial–temporal’ stacking strategy with cross-layer skip connections, the model facilitates progressive alignment and deep fusion, ranging from local interactions to long-range dependencies. Subsequently, an intention recognition module is developed. A second-order gated bilinear fusion mechanism is introduced to adaptively model higher-order couplings between local neighbour dynamics and global interaction semantics, thereby yielding a multimodal probability distribution over the target vehicle's driving intentions. Lastly, multimodal trajectory predictions are generated by decoding the fused spatiotemporal features together with the inferred intention information. Experimental results on three benchmark datasets—NGSIM (Next Generation Simulation), AD4CHE (Aerial Dataset for China Congested Highway and Expressway), and highD—demonstrate that CLISTA consistently outperforms the baseline methods. Relative to the next-best model, it reduces average/final displacement errors by 16.67 %/21.23 %, 12.99 %/21.14 % and 10.53 %/21.59 % on NGSIM, AD4CHE and HighD, respectively. Overall, CLISTA offers reliable multi-hypothesis trajectory priors for safe and efficient decision-making in complex traffic scenarios.</div></div>","PeriodicalId":50523,"journal":{"name":"Engineering Applications of Artificial Intelligence","volume":"167 ","pages":"Article 113937"},"PeriodicalIF":8.0,"publicationDate":"2026-01-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146038537","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}