Pub Date : 2026-03-01Epub Date: 2025-11-26DOI: 10.1016/j.mlwa.2025.100802
Sajjad Saed, Babak Teimourpour
The rapid expansion of the fashion industry and the growing variety of products have made it increasingly challenging for users to identify compatible items on e-commerce platforms. Effective fashion recommendation systems are therefore crucial for filtering irrelevant options and suggesting suitable ones. However, simultaneously addressing outfit compatibility and personalized recommendations remains a significant challenge, as these aspects are typically treated independently in existing studies, thereby overlooking the complex interactions between items and user preferences. This research introduces a new framework named FGAT, which leverages a hierarchical graph representation together with attention mechanisms to address this problem. The framework constructs a three-tier graph of users, outfits, and items, integrating visual and textual features to jointly model outfit compatibility and user preferences. By dynamically weighting node importance during representation propagation, the graph attention mechanism captures key interactions and produces precise embeddings for both user preferences and outfit compatibility. Evaluated on the POG dataset, FGAT outperforms strong baselines such as HFGN, achieving notable improvements in accuracy, precision, hit ratio (HR), recall, and NDCG. These results demonstrate that combining multimodal visual–textual features with a hierarchical graph structure and attention mechanisms significantly enhances the effectiveness and efficiency of personalized fashion recommendation systems.
{"title":"Hybrid-hierarchical fashion graph attention network for compatibility-oriented and personalized outfit recommendation","authors":"Sajjad Saed, Babak Teimourpour","doi":"10.1016/j.mlwa.2025.100802","DOIUrl":"10.1016/j.mlwa.2025.100802","url":null,"abstract":"<div><div>The rapid expansion of the fashion industry and the growing variety of products have made it increasingly challenging for users to identify compatible items on e-commerce platforms. Effective fashion recommendation systems are therefore crucial for filtering irrelevant options and suggesting suitable ones. However, simultaneously addressing outfit compatibility and personalized recommendations remains a significant challenge, as these aspects are typically treated independently in existing studies, thereby overlooking the complex interactions between items and user preferences. This research introduces a new framework named FGAT, which leverages a hierarchical graph representation together with attention mechanisms to address this problem. The framework constructs a three-tier graph of users, outfits, and items, integrating visual and textual features to jointly model outfit compatibility and user preferences. By dynamically weighting node importance during representation propagation, the graph attention mechanism captures key interactions and produces precise embeddings for both user preferences and outfit compatibility. Evaluated on the POG dataset, FGAT outperforms strong baselines such as HFGN, achieving notable improvements in accuracy, precision, hit ratio (HR), recall, and NDCG. These results demonstrate that combining multimodal visual–textual features with a hierarchical graph structure and attention mechanisms significantly enhances the effectiveness and efficiency of personalized fashion recommendation systems.</div></div>","PeriodicalId":74093,"journal":{"name":"Machine learning with applications","volume":"23 ","pages":"Article 100802"},"PeriodicalIF":4.9,"publicationDate":"2026-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145748775","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-03-01Epub Date: 2025-12-02DOI: 10.1016/j.mlwa.2025.100809
Faisal Saeed , Anand Paul
Effectively detecting and assessing real-time structural and ecological parameters in contemporary manufacturing environments poses significant challenges, particularly in identifying minute objects within product images. The swift evolution of the industrial sector underscores the necessity for intelligent manufacturing environments to uphold stringent product quality standards. However, accelerating production processes at high speeds heightens the risk of defective product outcomes. This research addresses the challenges inherent in small object detection within industrial contexts, proposing an innovative detection transformer model tailored to modern manufacturing environments. The proposed model integrates a feature-enhanced multi-head self-attention block (FEMSA), merging cross-channel communication network and multiple multi-head self-attention (MSA) components to refine image features. A query proposal network is also introduced within the detection transformer framework to discern high-ranking proposals using Intersection over Union (IoU) and Non-Maximum Suppression (NMS) algorithms. Through extensive experimentation on custom industrial small objects, our proposed model demonstrates superior performance compared to existing models based on Non-Maximum Suppression and transformers. By tackling the challenges associated with small object detection, our model contributes to the dynamic synchronization between virtual and physical manufacturing realms, enhancing quality control in industrial production.
{"title":"ISO-DeTr: A novel detection transformer for industrial small object detection","authors":"Faisal Saeed , Anand Paul","doi":"10.1016/j.mlwa.2025.100809","DOIUrl":"10.1016/j.mlwa.2025.100809","url":null,"abstract":"<div><div>Effectively detecting and assessing real-time structural and ecological parameters in contemporary manufacturing environments poses significant challenges, particularly in identifying minute objects within product images. The swift evolution of the industrial sector underscores the necessity for intelligent manufacturing environments to uphold stringent product quality standards. However, accelerating production processes at high speeds heightens the risk of defective product outcomes. This research addresses the challenges inherent in small object detection within industrial contexts, proposing an innovative detection transformer model tailored to modern manufacturing environments. The proposed model integrates a feature-enhanced multi-head self-attention block (FEMSA), merging cross-channel communication network and multiple multi-head self-attention (MSA) components to refine image features. A query proposal network is also introduced within the detection transformer framework to discern high-ranking proposals using Intersection over Union (IoU) and Non-Maximum Suppression (NMS) algorithms. Through extensive experimentation on custom industrial small objects, our proposed model demonstrates superior performance compared to existing models based on Non-Maximum Suppression and transformers. By tackling the challenges associated with small object detection, our model contributes to the dynamic synchronization between virtual and physical manufacturing realms, enhancing quality control in industrial production.</div></div>","PeriodicalId":74093,"journal":{"name":"Machine learning with applications","volume":"23 ","pages":"Article 100809"},"PeriodicalIF":4.9,"publicationDate":"2026-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145748778","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
The Internet of Vehicles (IoV) is advancing modern transportation by improving safety, efficiency, and intelligence. However, the reliance on the Controller Area Network (CAN) introduces critical security risks, as CAN-based communication is highly vulnerable to cyberattacks. Addressing this challenge, we propose DAIRE (Detecting Attacks in IoV in REal-time), a lightweight machine learning framework designed for real-time detection and classification of CAN attacks. DAIRE is built on a lightweight artificial neural network (ANN) where each layer contains neurons, with representing the number of neurons in the th layer and corresponding to the total number of attack classes. Other hyperparameters are determined empirically to ensure real-time operation. To support the detection and classification of various IoV attacks, such as Denial-of-Service, Fuzzy, and Spoofing, DAIRE employs the sparse categorical cross-entropy loss function and root mean square propagation for loss minimization. In contrast to more resource-intensive architectures, DAIRE leverages a lightweight ANN to reduce computational demands while still delivering strong performance. Experimental results on the CICIoV2024 and Car-Hacking datasets demonstrate DAIRE’s effectiveness, achieving an average detection rate of 99.88%, a false positive rate of 0.02%, and an overall accuracy of 99.96%. Furthermore, DAIRE significantly outperforms state-of-the-art approaches in inference speed, with a classification time of just 0.03 ms per sample. These results highlight DAIRE’s effectiveness in detecting IoV cyberattacks and its practical suitability for real-time deployment in vehicular systems, underscoring its vital role in strengthening automotive cybersecurity.
车联网(IoV)通过提高安全性、效率和智能,推动着现代交通的发展。然而,对控制器区域网络(CAN)的依赖带来了严重的安全风险,因为基于CAN的通信非常容易受到网络攻击。为了应对这一挑战,我们提出了DAIRE (detection Attacks in IoV in REal-time),这是一个轻量级的机器学习框架,旨在实时检测和分类CAN攻击。DAIRE建立在一个轻量级的人工神经网络(ANN)上,每层包含Ni=i×c神经元,Ni表示第i层神经元的数量,c对应攻击类的总数。其他超参数是经验确定的,以确保实时运行。为了支持各种IoV攻击的检测和分类,如拒绝服务、模糊和欺骗,DAIRE采用稀疏分类交叉熵损失函数和均方根传播来最小化损失。与更多的资源密集型架构相比,DAIRE利用轻量级ANN来减少计算需求,同时仍然提供强大的性能。在CICIoV2024和Car-Hacking数据集上的实验结果证明了DAIRE的有效性,平均检测率为99.88%,假阳性率为0.02%,总体准确率为99.96%。此外,DAIRE在推理速度上明显优于最先进的方法,每个样本的分类时间仅为0.03 ms。这些结果突出了DAIRE在检测车联网网络攻击方面的有效性及其在车载系统中实时部署的实际适用性,强调了其在加强汽车网络安全方面的重要作用。
{"title":"DAIRE: A lightweight AI model for real-time detection of Controller Area Network attacks in the Internet of Vehicles","authors":"Shahid Alam , Amina Jameel , Zahida Parveen , Ehab Alnfrawy , Adeela Ashraf , Raza Uddin , Jamal Aqib","doi":"10.1016/j.mlwa.2026.100859","DOIUrl":"10.1016/j.mlwa.2026.100859","url":null,"abstract":"<div><div>The Internet of Vehicles (IoV) is advancing modern transportation by improving safety, efficiency, and intelligence. However, the reliance on the Controller Area Network (CAN) introduces critical security risks, as CAN-based communication is highly vulnerable to cyberattacks. Addressing this challenge, we propose DAIRE (Detecting Attacks in IoV in REal-time), a lightweight machine learning framework designed for real-time detection and classification of CAN attacks. DAIRE is built on a lightweight artificial neural network (ANN) where each layer contains <span><math><mrow><msub><mrow><mi>N</mi></mrow><mrow><mi>i</mi></mrow></msub><mo>=</mo><mi>i</mi><mo>×</mo><mi>c</mi></mrow></math></span> neurons, with <span><math><msub><mrow><mi>N</mi></mrow><mrow><mi>i</mi></mrow></msub></math></span> representing the number of neurons in the <span><math><mi>i</mi></math></span>th layer and <span><math><mi>c</mi></math></span> corresponding to the total number of attack classes. Other hyperparameters are determined empirically to ensure real-time operation. To support the detection and classification of various IoV attacks, such as <em>Denial-of-Service</em>, <em>Fuzzy</em>, and <em>Spoofing</em>, DAIRE employs the sparse categorical cross-entropy loss function and root mean square propagation for loss minimization. In contrast to more resource-intensive architectures, DAIRE leverages a lightweight ANN to reduce computational demands while still delivering strong performance. Experimental results on the CICIoV2024 and Car-Hacking datasets demonstrate DAIRE’s effectiveness, achieving an average detection rate of 99.88%, a false positive rate of 0.02%, and an overall accuracy of 99.96%. Furthermore, DAIRE significantly outperforms state-of-the-art approaches in inference speed, with a classification time of just 0.03 ms per sample. These results highlight DAIRE’s effectiveness in detecting IoV cyberattacks and its practical suitability for real-time deployment in vehicular systems, underscoring its vital role in strengthening automotive cybersecurity.</div></div>","PeriodicalId":74093,"journal":{"name":"Machine learning with applications","volume":"23 ","pages":"Article 100859"},"PeriodicalIF":4.9,"publicationDate":"2026-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146188221","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-01Epub Date: 2025-09-23DOI: 10.1016/j.mlwa.2025.100732
Britt van Leeuwen , Sandjai Bhulai , Rob van der Mei
Authorship Verification is a key task in Natural Language Processing, essential for applications like plagiarism detection and content authentication. This paper analyzes the use of deep learning models for Authorship Verification, focusing on combining semantic and style features to enhance model performance. We propose three models: the Feature Interaction Network, Pairwise Concatenation Network, and Siamese Network, which aim to determine if two texts are written by the same author. Each model uses RoBERTa embeddings to capture semantic content and incorporates style features such as sentence length, word frequency, and punctuation to differentiate authors based on writing style.
Our results confirm that incorporating style features consistently improves model performance, with the extent of improvement varying by architecture. This demonstrates the value of combining semantic and stylistic information for Authorship Verification. While limitations such as RoBERTa’s fixed input length and the use of predefined style features exist, they do not hinder model effectiveness and point to clear opportunities for future enhancement through extended input handling and dynamic style feature extraction.
In contrast to prior studies such as Bevendorff et al., (2020) and Kestemont, et al., (2022), which relied on balanced and homogeneous datasets with consistent topics and well-formed language, our work evaluates models on a more challenging, imbalanced, and stylistically diverse dataset, better reflecting real-world Authorship Verification conditions. Despite the increased difficulty, our models achieve competitive results, underscoring their robustness and practical applicability.
These findings support the value of combining semantic and style features for real-world Authorship Verification.
{"title":"Combining style and semantics for robust authorship verification","authors":"Britt van Leeuwen , Sandjai Bhulai , Rob van der Mei","doi":"10.1016/j.mlwa.2025.100732","DOIUrl":"10.1016/j.mlwa.2025.100732","url":null,"abstract":"<div><div>Authorship Verification is a key task in Natural Language Processing, essential for applications like plagiarism detection and content authentication. This paper analyzes the use of deep learning models for Authorship Verification, focusing on combining semantic and style features to enhance model performance. We propose three models: the Feature Interaction Network, Pairwise Concatenation Network, and Siamese Network, which aim to determine if two texts are written by the same author. Each model uses RoBERTa embeddings to capture semantic content and incorporates style features such as sentence length, word frequency, and punctuation to differentiate authors based on writing style.</div><div>Our results confirm that incorporating style features consistently improves model performance, with the extent of improvement varying by architecture. This demonstrates the value of combining semantic and stylistic information for Authorship Verification. While limitations such as RoBERTa’s fixed input length and the use of predefined style features exist, they do not hinder model effectiveness and point to clear opportunities for future enhancement through extended input handling and dynamic style feature extraction.</div><div>In contrast to prior studies such as Bevendorff et al., (2020) and Kestemont, et al., (2022), which relied on balanced and homogeneous datasets with consistent topics and well-formed language, our work evaluates models on a more challenging, imbalanced, and stylistically diverse dataset, better reflecting real-world Authorship Verification conditions. Despite the increased difficulty, our models achieve competitive results, underscoring their robustness and practical applicability.</div><div>These findings support the value of combining semantic and style features for real-world Authorship Verification.</div></div>","PeriodicalId":74093,"journal":{"name":"Machine learning with applications","volume":"22 ","pages":"Article 100732"},"PeriodicalIF":4.9,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145222738","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-01Epub Date: 2025-10-10DOI: 10.1016/j.mlwa.2025.100753
Bingbing Guo , Yujie Jiao , Yan Wang , Fengling Zhang , Yuanfei Guo , Qinghao Guan
Reinforced concrete (RC) structures are widely used in civil engineering, and accurate prediction of chloride transport is essential for durability design and service life estimation. Existing machine learning models for predicting the chloride transport in concrete have primarily relied on researchers' expertise for feature construction. However, the factors affecting the chloride transport are numerous and highly complex, making manual feature engineering inefficient and labor-intensive. This study developed text-enhanced multimodal models that integrate natural language processing (NLP) with deep neural network (DNN) to automatically extract features from textual information, including properties of raw materials, experimental methods, chloride attack mechanisms and comments. The results demonstrate that the developed multimodal models have learned prior knowledge, which enables them to achieve significantly higher accuracy than numerical-data-only DNN models. Among these models, the multi-head self-attention model performs the best by capturing features from multiple angles and enabling parallel computation. Crucially, the text-enhanced multimodal models can maintain high accuracy even with limited numerical data.
{"title":"Text-enhanced multimodal deep learning models for predicting chloride transport in concrete","authors":"Bingbing Guo , Yujie Jiao , Yan Wang , Fengling Zhang , Yuanfei Guo , Qinghao Guan","doi":"10.1016/j.mlwa.2025.100753","DOIUrl":"10.1016/j.mlwa.2025.100753","url":null,"abstract":"<div><div>Reinforced concrete (RC) structures are widely used in civil engineering, and accurate prediction of chloride transport is essential for durability design and service life estimation. Existing machine learning models for predicting the chloride transport in concrete have primarily relied on researchers' expertise for feature construction. However, the factors affecting the chloride transport are numerous and highly complex, making manual feature engineering inefficient and labor-intensive. This study developed text-enhanced multimodal models that integrate natural language processing (NLP) with deep neural network (DNN) to automatically extract features from textual information, including properties of raw materials, experimental methods, chloride attack mechanisms and comments. The results demonstrate that the developed multimodal models have learned prior knowledge, which enables them to achieve significantly higher accuracy than numerical-data-only DNN models. Among these models, the multi-head self-attention model performs the best by capturing features from multiple angles and enabling parallel computation. Crucially, the text-enhanced multimodal models can maintain high accuracy even with limited numerical data.</div></div>","PeriodicalId":74093,"journal":{"name":"Machine learning with applications","volume":"22 ","pages":"Article 100753"},"PeriodicalIF":4.9,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145321179","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-01Epub Date: 2025-10-10DOI: 10.1016/j.mlwa.2025.100745
Hy Nguyen , Srikanth Thudumu , Hung Du , Rajesh Vasa , Kon Mouzakis
The exploration of Deep Reinforcement Learning (DRL) in Object Tracking (OT) represents an emerging paradigm and is gaining traction as an alternative to conventional CNN-based methods. DRL’s ability to integrate spatial and temporal context and learn from interactions makes it particularly suited for the sequential decision-making required in OT. The survey reviews a range of DRL-based methods for OT, systematically collating and analyzing existing research to highlight trends and challenges. It also provides an evaluation of different DRL algorithms, categorizing them based on their performance in various dynamic environments. Additionally, we analyze existing evaluation benchmarks and simulators, along with the challenges, potential solutions, and trends in DRL-based OT methods. This paper aims to bridge the fragmented literature on DRL applications in OT, providing a unified view that identifies common approaches, challenges, and potential synergies.
{"title":"A comprehensive survey on deep reinforcement learning in object tracking","authors":"Hy Nguyen , Srikanth Thudumu , Hung Du , Rajesh Vasa , Kon Mouzakis","doi":"10.1016/j.mlwa.2025.100745","DOIUrl":"10.1016/j.mlwa.2025.100745","url":null,"abstract":"<div><div>The exploration of Deep Reinforcement Learning (DRL) in Object Tracking (OT) represents an emerging paradigm and is gaining traction as an alternative to conventional CNN-based methods. DRL’s ability to integrate spatial and temporal context and learn from interactions makes it particularly suited for the sequential decision-making required in OT. The survey reviews a range of DRL-based methods for OT, systematically collating and analyzing existing research to highlight trends and challenges. It also provides an evaluation of different DRL algorithms, categorizing them based on their performance in various dynamic environments. Additionally, we analyze existing evaluation benchmarks and simulators, along with the challenges, potential solutions, and trends in DRL-based OT methods. This paper aims to bridge the fragmented literature on DRL applications in OT, providing a unified view that identifies common approaches, challenges, and potential synergies.</div></div>","PeriodicalId":74093,"journal":{"name":"Machine learning with applications","volume":"22 ","pages":"Article 100745"},"PeriodicalIF":4.9,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145321182","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-01Epub Date: 2025-10-23DOI: 10.1016/j.mlwa.2025.100769
Marco Zanotti
In an era of increasing computational capabilities and growing environmental consciousness, organizations face a critical challenge in balancing the accuracy of forecasting models with computational efficiency and sustainability. Global forecasting models, lowering the computational time, have gained significant attention over the years. However, the common practice of retraining these models with new observations raises important questions about the costs of forecasting. Using ten different machine learning and deep learning models, we analyzed various retraining scenarios, ranging from continuous updates to no retraining at all, across two large retail demand datasets. We showed that less frequent retraining strategies maintain the forecast accuracy while reducing the computational costs, providing a more sustainable approach to large-scale forecasting. We also found that machine learning models are a marginally better choice to reduce the costs of forecasting when coupled with less frequent model retraining strategies as the frequency of the data increases. Our findings challenge the conventional belief that frequent retraining is essential for maintaining forecasting accuracy. Instead, periodic retraining offers a good balance between predictive performance and efficiency, both in the case of point and probabilistic forecasting. These insights provide actionable guidelines for organizations seeking to optimize forecasting pipelines while reducing costs and energy consumption.
{"title":"On the retraining frequency of global models in retail demand forecasting","authors":"Marco Zanotti","doi":"10.1016/j.mlwa.2025.100769","DOIUrl":"10.1016/j.mlwa.2025.100769","url":null,"abstract":"<div><div>In an era of increasing computational capabilities and growing environmental consciousness, organizations face a critical challenge in balancing the accuracy of forecasting models with computational efficiency and sustainability. Global forecasting models, lowering the computational time, have gained significant attention over the years. However, the common practice of retraining these models with new observations raises important questions about the costs of forecasting. Using ten different machine learning and deep learning models, we analyzed various retraining scenarios, ranging from continuous updates to no retraining at all, across two large retail demand datasets. We showed that less frequent retraining strategies maintain the forecast accuracy while reducing the computational costs, providing a more sustainable approach to large-scale forecasting. We also found that machine learning models are a marginally better choice to reduce the costs of forecasting when coupled with less frequent model retraining strategies as the frequency of the data increases. Our findings challenge the conventional belief that frequent retraining is essential for maintaining forecasting accuracy. Instead, periodic retraining offers a good balance between predictive performance and efficiency, both in the case of point and probabilistic forecasting. These insights provide actionable guidelines for organizations seeking to optimize forecasting pipelines while reducing costs and energy consumption.</div></div>","PeriodicalId":74093,"journal":{"name":"Machine learning with applications","volume":"22 ","pages":"Article 100769"},"PeriodicalIF":4.9,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145363478","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-01Epub Date: 2025-10-02DOI: 10.1016/j.mlwa.2025.100747
Pingxian Wu , Junge Wang , Xinyou Chen , Tao Wang , Zongyi Guo , Shuqi Diao , Jinyong Wang
Background
The increasing volume of genome sequencing data poses significant challenges for traditional genome-wide prediction methods in handling large datasets. Machine learning (ML) techniques are well-suited for processing high-dimensional data, and offer promising solutions. This study aimed to identify an optimal genome-wide prediction approach for local pig breeds using 10 datasets with varying single nucleotide polymorphism (SNP) densities, derived from imputed sequencing data of 485 Rongchang pigs and the results of genome-wide association studies (GWAS). Three growth traits, namely, backfat (BF) thickness, loin and thoracic height (LTH), and girth circumference (GC), were predicted using six traditional methods and six ML-based methods, including Kernel Ridge Regression (KRR), Support Vector Regression (SVR), Random Forest, Gradient Boosting Decision Tree, Light Gradient Boosting Machine, and Adaboost.
Results
The efficacy of the different methods was evaluated using a five-fold cross-validation strategy and independent tests. The predictive performance of both the traditional and ML-based methods was initially enhanced through the incorporation of significantly associated SNPs and weighted data, with the KRR method exhibiting exceptional resistance to overfitting at a SNP density of 300,000. The ML-based methods outperformed the traditional methods, with improvements of 6.6–8.1 %. The integration of GWAS data enhanced the prediction accuracy of the ML-based methods. KRR and Gradient Boosting Decision Tree demonstrated significant computational efficiency, indicating their potential as promising strategies for genomic prediction in livestock breeding.
Conclusions
This study provides a comprehensive analysis of genome-wide predictions in Rongchang pigs, and highlights the potential of ML-based techniques in enhancing prediction accuracy and efficiency. The study provides valuable insights into GP and holds key implications for advancing genome breeding practices in local pig breeds.
{"title":"Optimization of genomic breeding value prediction for growth traits in Rongchang pigs through machine learning techniques","authors":"Pingxian Wu , Junge Wang , Xinyou Chen , Tao Wang , Zongyi Guo , Shuqi Diao , Jinyong Wang","doi":"10.1016/j.mlwa.2025.100747","DOIUrl":"10.1016/j.mlwa.2025.100747","url":null,"abstract":"<div><h3>Background</h3><div>The increasing volume of genome sequencing data poses significant challenges for traditional genome-wide prediction methods in handling large datasets. Machine learning (ML) techniques are well-suited for processing high-dimensional data, and offer promising solutions. This study aimed to identify an optimal genome-wide prediction approach for local pig breeds using 10 datasets with varying single nucleotide polymorphism (SNP) densities, derived from imputed sequencing data of 485 Rongchang pigs and the results of genome-wide association studies (GWAS). Three growth traits, namely, backfat (BF) thickness, loin and thoracic height (LTH), and girth circumference (GC), were predicted using six traditional methods and six ML-based methods, including Kernel Ridge Regression (KRR), Support Vector Regression (SVR), Random Forest, Gradient Boosting Decision Tree, Light Gradient Boosting Machine, and Adaboost.</div></div><div><h3>Results</h3><div>The efficacy of the different methods was evaluated using a five-fold cross-validation strategy and independent tests. The predictive performance of both the traditional and ML-based methods was initially enhanced through the incorporation of significantly associated SNPs and weighted data, with the KRR method exhibiting exceptional resistance to overfitting at a SNP density of 300,000. The ML-based methods outperformed the traditional methods, with improvements of 6.6–8.1 %. The integration of GWAS data enhanced the prediction accuracy of the ML-based methods. KRR and Gradient Boosting Decision Tree demonstrated significant computational efficiency, indicating their potential as promising strategies for genomic prediction in livestock breeding.</div></div><div><h3>Conclusions</h3><div>This study provides a comprehensive analysis of genome-wide predictions in Rongchang pigs, and highlights the potential of ML-based techniques in enhancing prediction accuracy and efficiency. The study provides valuable insights into GP and holds key implications for advancing genome breeding practices in local pig breeds.</div></div>","PeriodicalId":74093,"journal":{"name":"Machine learning with applications","volume":"22 ","pages":"Article 100747"},"PeriodicalIF":4.9,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145268843","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-01Epub Date: 2025-11-05DOI: 10.1016/j.mlwa.2025.100776
Prosper Chimunhu , Erkan Topal , Mohammad Waqar Ali Asad , Roohollah Shirani Faradonbeh , Ajak Duany Ajak
For decades, Mixed Integer Programming (MIP) has been successfully utilised to optimise production schedules in underground mining, with increasingly notable results reported. However, recurrent inconsistencies between schedule forecasts and actual production due to imprecise input assumptions, such as mining dilution factors, subtly impair the robustness of optimal solutions, with detrimental hierarchical effects on the business’s cashflow projections and profitability. To address this, this study leverages emerging applications of Machine Learning (ML) and adjacent technologies that are revolutionising intelligent prediction of dilution in underground mining operations. The study proposes a synergistic nexus between MIP and ML models using ML-predicted dilution on a per-stope granularity instead of the traditional single dilution factor to improve the schedule’s forecasting accuracy. A sample of 61 stopes from an underground open-stoping operation was used to create and optimise schedules based on empirically determined and ML-predicted dilution factors. Study findings revealed a 3.1% higher net present value (NPV) for MIP-optimised schedules over manual schedules for the same dilution factor (empirical). Further, it was also noted that the ML-predicted dilution at 74% accuracy on a per-stope granularity enhances the MIP-optimised schedules’ tonnage forecast precision by at least 4 % and the NPV by at least 2 % compared to MIP-optimised schedules using the single dilution factor over a 16-month period. Additionally, results revealed that MIP schedules augmented with ML-predicted dilution demonstrated greater flexibility in navigating schedule constraints, leading to better schedule responsiveness and granularity on forecasts. Thus, the study improves optimal solutions’ robustness, reliability and production scheduling efficacy.
{"title":"Production scheduling optimisation using mixed integer programming with machine learning dilution prediction capabilities for underground open stoping operations","authors":"Prosper Chimunhu , Erkan Topal , Mohammad Waqar Ali Asad , Roohollah Shirani Faradonbeh , Ajak Duany Ajak","doi":"10.1016/j.mlwa.2025.100776","DOIUrl":"10.1016/j.mlwa.2025.100776","url":null,"abstract":"<div><div>For decades, Mixed Integer Programming (MIP) has been successfully utilised to optimise production schedules in underground mining, with increasingly notable results reported. However, recurrent inconsistencies between schedule forecasts and actual production due to imprecise input assumptions, such as mining dilution factors, subtly impair the robustness of optimal solutions, with detrimental hierarchical effects on the business’s cashflow projections and profitability. To address this, this study leverages emerging applications of Machine Learning (ML) and adjacent technologies that are revolutionising intelligent prediction of dilution in underground mining operations. The study proposes a synergistic nexus between MIP and ML models using ML-predicted dilution on a per-stope granularity instead of the traditional single dilution factor to improve the schedule’s forecasting accuracy. A sample of 61 stopes from an underground open-stoping operation was used to create and optimise schedules based on empirically determined and ML-predicted dilution factors. Study findings revealed a 3.1% higher net present value (NPV) for MIP-optimised schedules over manual schedules for the same dilution factor (empirical). Further, it was also noted that the ML-predicted dilution at 74% accuracy on a per-stope granularity enhances the MIP-optimised schedules’ tonnage forecast precision by at least 4 % and the NPV by at least 2 % compared to MIP-optimised schedules using the single dilution factor over a 16-month period. Additionally, results revealed that MIP schedules augmented with ML-predicted dilution demonstrated greater flexibility in navigating schedule constraints, leading to better schedule responsiveness and granularity on forecasts. Thus, the study improves optimal solutions’ robustness, reliability and production scheduling efficacy.</div></div>","PeriodicalId":74093,"journal":{"name":"Machine learning with applications","volume":"22 ","pages":"Article 100776"},"PeriodicalIF":4.9,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145528529","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-01Epub Date: 2025-11-04DOI: 10.1016/j.mlwa.2025.100789
Juan Duran , Yujing Zou , Martin Vallières , Shirin A. Enger
Accurate prediction of progression-free survival (PFS) is critical for precision oncology. However, most existing multimodal survival studies rely on single fusion strategies, one-off cross-validation runs, and focus solely on discrimination metrics, leaving gaps in systematic evaluation and calibration. We evaluated multimodal fusion approaches combining histopathology whole-slide images (via Hierarchical Image Pyramid Transformer) and clinical variables (via Feature Tokenizer-Transformer) across five TCGA cohorts: bladder cancer (BLCA), uterine corpus endometrial carcinoma (UCEC), lung adenocarcinoma (LUAD), breast cancer (BRCA), and head and neck squamous cell carcinoma (HNSC) (N=2,984). Three intermediate (marginal, cross-attention, Variational Autoencoder or VAE) and two late fusion strategies (trainable-weight, meta-learning) were trained end-to-end with DeepSurv. Our 100-repetition 10-fold cross-validation (CV) framework mitigates the variance overlooked in single-run CV evaluations. VAE fusion achieved superior PFS prediction (Concordance-index) in BLCA (0.739±0.019), UCEC (0.770±0.021), LUAD (0.683±0.018), and BRCA (0.760±0.021), while meta-learning was best for HNSC (0.686±0.022). However, Integrated Brier Score values (0.066–0.142) revealed calibration variability. Our findings highlight the importance of multimodal fusion, combined discrimination and calibration metrics, and rigorous validation for clinically meaningful survival modeling.
{"title":"Beyond single-run metrics with CP-fuse: A rigorous multi-cohort evaluation of clinico-pathological fusion for improved survival prediction in TCGA","authors":"Juan Duran , Yujing Zou , Martin Vallières , Shirin A. Enger","doi":"10.1016/j.mlwa.2025.100789","DOIUrl":"10.1016/j.mlwa.2025.100789","url":null,"abstract":"<div><div>Accurate prediction of progression-free survival (PFS) is critical for precision oncology. However, most existing multimodal survival studies rely on single fusion strategies, one-off cross-validation runs, and focus solely on discrimination metrics, leaving gaps in systematic evaluation and calibration. We evaluated multimodal fusion approaches combining histopathology whole-slide images (via Hierarchical Image Pyramid Transformer) and clinical variables (via Feature Tokenizer-Transformer) across five TCGA cohorts: bladder cancer (BLCA), uterine corpus endometrial carcinoma (UCEC), lung adenocarcinoma (LUAD), breast cancer (BRCA), and head and neck squamous cell carcinoma (HNSC) (N=2,984). Three intermediate (marginal, cross-attention, Variational Autoencoder or VAE) and two late fusion strategies (trainable-weight, meta-learning) were trained end-to-end with DeepSurv. Our 100-repetition 10-fold cross-validation (CV) framework mitigates the variance overlooked in single-run CV evaluations. VAE fusion achieved superior PFS prediction (Concordance-index) in BLCA (0.739±0.019), UCEC (0.770±0.021), LUAD (0.683±0.018), and BRCA (0.760±0.021), while meta-learning was best for HNSC (0.686±0.022). However, Integrated Brier Score values (0.066–0.142) revealed calibration variability. Our findings highlight the importance of multimodal fusion, combined discrimination and calibration metrics, and rigorous validation for clinically meaningful survival modeling.</div></div>","PeriodicalId":74093,"journal":{"name":"Machine learning with applications","volume":"22 ","pages":"Article 100789"},"PeriodicalIF":4.9,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145528537","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}