Pub Date : 2026-01-03DOI: 10.1016/j.array.2025.100672
Yingli Yang , Weixing Song , Jingzhe Wang
In order to enhance the utilization and maintenance efficiency of equipment, grounded in the theory of preventive maintenance, the interrelationship between equipment usage and maintenance was analyzed. By proposing optimization objectives aimed at reducing the mean failure rate, preventive maintenance time, standard deviation of monthly motor-hour reserve, and maintenance costs, and comprehensively incorporating both preventive and corrective maintenance strategies, an optimization model for equipment utilization and maintenance was established. A DNA-based population encoding method, along with crossover and mutation operations was designed. Furthermore, an adaptive reinforcement learning algorithm was employed to adjust the reference vectors, and the NSGA-III algorithm is improved for simulation experiments. The model and algorithm are not only practical but also computationally efficient, scientifically sound, and applicable. Upon analyzing the simulation results, optimization strategies for equipment utilization preferences were proposed. These strategies can provide decision-making support for developing equipment deployment and maintenance plans for army.
{"title":"A high-dimensional many-objective co-optimization method on equipment utilization and maintenance","authors":"Yingli Yang , Weixing Song , Jingzhe Wang","doi":"10.1016/j.array.2025.100672","DOIUrl":"10.1016/j.array.2025.100672","url":null,"abstract":"<div><div>In order to enhance the utilization and maintenance efficiency of equipment, grounded in the theory of preventive maintenance, the interrelationship between equipment usage and maintenance was analyzed. By proposing optimization objectives aimed at reducing the mean failure rate, preventive maintenance time, standard deviation of monthly motor-hour reserve, and maintenance costs, and comprehensively incorporating both preventive and corrective maintenance strategies, an optimization model for equipment utilization and maintenance was established. A DNA-based population encoding method, along with crossover and mutation operations was designed. Furthermore, an adaptive reinforcement learning algorithm was employed to adjust the reference vectors, and the NSGA-III algorithm is improved for simulation experiments. The model and algorithm are not only practical but also computationally efficient, scientifically sound, and applicable. Upon analyzing the simulation results, optimization strategies for equipment utilization preferences were proposed. These strategies can provide decision-making support for developing equipment deployment and maintenance plans for army.</div></div>","PeriodicalId":8417,"journal":{"name":"Array","volume":"29 ","pages":"Article 100672"},"PeriodicalIF":4.5,"publicationDate":"2026-01-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145921325","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-01-03DOI: 10.1016/j.array.2025.100674
Hui Han , Silvana Trimi , Sang M. Lee
Tiny Machine Learning (TinyML) enables artificial intelligence on low-power edge devices, yet a quantitative understanding of TinyML research remains limited. This study addresses this gap through a comprehensive bibliometric analysis of 392 peer-reviewed publications (2020–2024) from the Web of Science, using Biblioshiny and VOSviewer. This article contributes by mapping the first bibliometric structure of TinyML, identifying major trends (exponential publication growth, strong international collaboration, core research themes, key contributors, etc.) and proposing future directions (such as sustainable hardware, federated learning, ethical frameworks, etc.). The findings provide a scholarly foundation and strategic roadmap for advancing scalable, energy-efficient, and privacy-preserving TinyML applications.
微型机器学习(TinyML)可以在低功耗边缘设备上实现人工智能,但对TinyML研究的定量理解仍然有限。本研究通过使用Biblioshiny和VOSviewer对来自Web of Science的392篇同行评议出版物(2020-2024)进行全面的文献计量分析,解决了这一差距。本文通过绘制TinyML的第一个文献计量结构,确定主要趋势(指数出版物增长,强大的国际合作,核心研究主题,关键贡献者等)并提出未来方向(如可持续硬件,联邦学习,伦理框架等)。研究结果为推进可扩展、节能和保护隐私的TinyML应用程序提供了学术基础和战略路线图。
{"title":"Tiny Machine Learning (TinyML): Research trends and future application opportunities","authors":"Hui Han , Silvana Trimi , Sang M. Lee","doi":"10.1016/j.array.2025.100674","DOIUrl":"10.1016/j.array.2025.100674","url":null,"abstract":"<div><div>Tiny Machine Learning (TinyML) enables artificial intelligence on low-power edge devices, yet a quantitative understanding of TinyML research remains limited. This study addresses this gap through a comprehensive bibliometric analysis of 392 peer-reviewed publications (2020–2024) from the Web of Science, using Biblioshiny and VOSviewer. This article contributes by mapping the first bibliometric structure of TinyML, identifying major trends (exponential publication growth, strong international collaboration, core research themes, key contributors, etc.) and proposing future directions (such as sustainable hardware, federated learning, ethical frameworks, etc.). The findings provide a scholarly foundation and strategic roadmap for advancing scalable, energy-efficient, and privacy-preserving TinyML applications.</div></div>","PeriodicalId":8417,"journal":{"name":"Array","volume":"29 ","pages":"Article 100674"},"PeriodicalIF":4.5,"publicationDate":"2026-01-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145921319","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-01-03DOI: 10.1016/j.array.2025.100664
Harsh Dadhwal , Mateus de Abreu , Nazanin Parvizi , Sajal Saha
Distributed Denial of Service (DDoS) attacks continue to grow in scale and sophistication, making timely and reliable detection increasingly challenging. Machine learning (ML) models have demonstrated promise in identifying malicious traffic patterns. However, their vulnerability to adversarial manipulation remains a critical security concern. This study benchmarks the adversarial robustness of several standalone ML models trained on clean traffic from the CIC-IDS2017 dataset. Adversarial perturbations are generated using the Fast Gradient Sign Method (FGSM). Baseline robustness is assessed by evaluating each model on FGSM adversarial samples generated at . Adversarial training is then performed by augmenting the clean dataset with FGSM-generated samples, after which models are evaluated across multiple perturbation strengths ( to ). Experimental findings show that boosting-based models, particularly XGBoost and LightGBM, demonstrate the highest resilience under adversarial stress. In contrast, models such as Logistic Regression and MLP experience significant performance degradation, even after adversarial training. Despite this, adversarial training substantially improves robustness across all models, highlighting its effectiveness in mitigating FGSM-induced decision boundary shifts. This benchmarking study underscores the importance of adversarial training and model selection when deploying ML-based intrusion detection systems. Boosting ensembles consistently provide superior robustness, while linear and neural models remain more susceptible to perturbations.
{"title":"Benchmarking the adversarial resilience of machine learning models for DDoS detection","authors":"Harsh Dadhwal , Mateus de Abreu , Nazanin Parvizi , Sajal Saha","doi":"10.1016/j.array.2025.100664","DOIUrl":"10.1016/j.array.2025.100664","url":null,"abstract":"<div><div>Distributed Denial of Service (DDoS) attacks continue to grow in scale and sophistication, making timely and reliable detection increasingly challenging. Machine learning (ML) models have demonstrated promise in identifying malicious traffic patterns. However, their vulnerability to adversarial manipulation remains a critical security concern. This study benchmarks the adversarial robustness of several standalone ML models trained on clean traffic from the CIC-IDS2017 dataset. Adversarial perturbations are generated using the Fast Gradient Sign Method (FGSM). Baseline robustness is assessed by evaluating each model on FGSM adversarial samples generated at <span><math><mrow><mi>ϵ</mi><mo>=</mo><mn>0</mn><mo>.</mo><mn>1</mn></mrow></math></span>. Adversarial training is then performed by augmenting the clean dataset with FGSM-generated samples, after which models are evaluated across multiple perturbation strengths (<span><math><mrow><mi>ϵ</mi><mo>=</mo><mn>0</mn><mo>.</mo><mn>1</mn></mrow></math></span> to <span><math><mrow><mn>3</mn><mo>.</mo><mn>0</mn></mrow></math></span>). Experimental findings show that boosting-based models, particularly XGBoost and LightGBM, demonstrate the highest resilience under adversarial stress. In contrast, models such as Logistic Regression and MLP experience significant performance degradation, even after adversarial training. Despite this, adversarial training substantially improves robustness across all models, highlighting its effectiveness in mitigating FGSM-induced decision boundary shifts. This benchmarking study underscores the importance of adversarial training and model selection when deploying ML-based intrusion detection systems. Boosting ensembles consistently provide superior robustness, while linear and neural models remain more susceptible to perturbations.</div></div>","PeriodicalId":8417,"journal":{"name":"Array","volume":"29 ","pages":"Article 100664"},"PeriodicalIF":4.5,"publicationDate":"2026-01-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146073846","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-01-03DOI: 10.1016/j.array.2025.100673
Ali Asghar , Wareesa Sharif , Amna Shifa
As countries rapidly transition toward smart cities, closed-circuit television (CCTV) surveillance systems are playing an increasingly vital role in ensuring public safety and enabling urban analytics. However, the visual quality of CCTV footage is often degraded by environmental factors (e.g., motion blur, low resolution, and poor illumination), which significantly impact the quality of service, as well as the reliability and effectiveness of these systems. To address these issues, this research proposes a generative adversarial network (GAN)-based framework, named VEGAN (Video Enhancement with Generative Adversarial Network), that combines reconstruction, adversarial, and facial component losses with adaptive balancing to optimise visual sharpness, temporal stability, and identity preservation. VEGAN integrates four key components: (1) a pixel counter, which identifies high and low quality frames; (2) Modified MoCoGAN, which separates foreground and background features to disentangle content from motion; (3) a Recurrent Neural Network, which captures complex temporal and motion patterns; and (4) a Super-Resolution module, which enhances low-quality pixels to recover fine spatial details. The enhanced foreground generated by these combined modules is seamlessly fused with a high-quality background frame, resulting in substantially improved overall video quality. Experimental evaluations demonstrate VEGAN’s effectiveness, achieving an average Learned Perceptual Image Patch Similarity (LPIPS) score of 0.041 and a Video Multimethod Assessment Fusion (VMAF) score of 56.13, indicating significant perceptual and quantitative improvements. These findings highlight VEGAN’s effectiveness in video-based analytics, supporting more accurate performance in tasks such as event detection and activity recognition.
{"title":"VEGAN: CCTV video quality enhancement with GAN-based foreground separation and super-resolution","authors":"Ali Asghar , Wareesa Sharif , Amna Shifa","doi":"10.1016/j.array.2025.100673","DOIUrl":"10.1016/j.array.2025.100673","url":null,"abstract":"<div><div>As countries rapidly transition toward smart cities, closed-circuit television (CCTV) surveillance systems are playing an increasingly vital role in ensuring public safety and enabling urban analytics. However, the visual quality of CCTV footage is often degraded by environmental factors (e.g., motion blur, low resolution, and poor illumination), which significantly impact the quality of service, as well as the reliability and effectiveness of these systems. To address these issues, this research proposes a generative adversarial network (GAN)-based framework, named VEGAN (Video Enhancement with Generative Adversarial Network), that combines reconstruction, adversarial, and facial component losses with adaptive balancing to optimise visual sharpness, temporal stability, and identity preservation. VEGAN integrates four key components: (1) a pixel counter, which identifies high and low quality frames; (2) Modified MoCoGAN, which separates foreground and background features to disentangle content from motion; (3) a Recurrent Neural Network, which captures complex temporal and motion patterns; and (4) a Super-Resolution module, which enhances low-quality pixels to recover fine spatial details. The enhanced foreground generated by these combined modules is seamlessly fused with a high-quality background frame, resulting in substantially improved overall video quality. Experimental evaluations demonstrate VEGAN’s effectiveness, achieving an average Learned Perceptual Image Patch Similarity (LPIPS) score of 0.041 and a Video Multimethod Assessment Fusion (VMAF) score of 56.13, indicating significant perceptual and quantitative improvements. These findings highlight VEGAN’s effectiveness in video-based analytics, supporting more accurate performance in tasks such as event detection and activity recognition.</div></div>","PeriodicalId":8417,"journal":{"name":"Array","volume":"29 ","pages":"Article 100673"},"PeriodicalIF":4.5,"publicationDate":"2026-01-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145921321","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-01-03DOI: 10.1016/j.array.2025.100669
Sunu Wibirama , Muhammad Ainul Fikri , Iman Kahfi Aliza , Kristian Adi Nugraha , Syukron Abu Ishaq Alfarozi , Noor Akhmad Setiawan , Ahmad Riznandi Suhari , Sri Kusrohmaniah
Cognitive load classification during online shopping activities is important to understand the user experience of e-commerce. Traditional classification methods that rely on proprietary software and obtrusive physiological measures often result in inconsistent performance. To address this research gap, we propose a novel approach that leverages deep learning to analyze raw eye movement data during online shopping tasks with low and high cognitive load. The Attention-based Long Short-Term Memory Fully Convolutional Network (ALSTM-FCN) model outperformed other machine learning and deep learning models with an average accuracy and F1 score of 97.70% and 97.69%, respectively. Cognitive load was also measured using the NASA TLX questionnaire, which showed significantly higher scores in high cognitive load tasks for all dimensions: “Mental Demand” (39.37, ), “Performance” (46.12, ), “Effort” (51.92, ), and “Frustration Level” (60.53, ). Based on the analysis of eye movement features used in cognitive load classification, we found that the variability in eye movement during tasks with low and high cognitive loads was predominantly spatial rather than temporal (). Our findings indicate a strong correlation between the deep learning-based classification of raw eye movement data and subjective cognitive load assessments. This study demonstrates the potential of using an affordable eye tracking sensor to classify cognitive load without being constrained by the capability of proprietary software.
{"title":"Cognitive load classification during online shopping using deep learning on time series eye movement indices","authors":"Sunu Wibirama , Muhammad Ainul Fikri , Iman Kahfi Aliza , Kristian Adi Nugraha , Syukron Abu Ishaq Alfarozi , Noor Akhmad Setiawan , Ahmad Riznandi Suhari , Sri Kusrohmaniah","doi":"10.1016/j.array.2025.100669","DOIUrl":"10.1016/j.array.2025.100669","url":null,"abstract":"<div><div>Cognitive load classification during online shopping activities is important to understand the user experience of e-commerce. Traditional classification methods that rely on proprietary software and obtrusive physiological measures often result in inconsistent performance. To address this research gap, we propose a novel approach that leverages deep learning to analyze raw eye movement data during online shopping tasks with low and high cognitive load. The Attention-based Long Short-Term Memory Fully Convolutional Network (ALSTM-FCN) model outperformed other machine learning and deep learning models with an average accuracy and F1 score of 97.70% and 97.69%, respectively. Cognitive load was also measured using the NASA TLX questionnaire, which showed significantly higher scores in high cognitive load tasks for all dimensions: “Mental Demand” (39.37, <span><math><mrow><mi>p</mi><mo>=</mo><mn>0</mn><mo>.</mo><mn>001</mn></mrow></math></span>), “Performance” (46.12, <span><math><mrow><mi>p</mi><mo>=</mo><mn>0</mn><mo>.</mo><mn>004</mn></mrow></math></span>), “Effort” (51.92, <span><math><mrow><mi>p</mi><mo>=</mo><mn>0</mn><mo>.</mo><mn>002</mn></mrow></math></span>), and “Frustration Level” (60.53, <span><math><mrow><mi>p</mi><mo>=</mo><mn>0</mn><mo>.</mo><mn>001</mn></mrow></math></span>). Based on the analysis of eye movement features used in cognitive load classification, we found that the variability in eye movement during tasks with low and high cognitive loads was predominantly spatial rather than temporal (<span><math><mrow><mi>p</mi><mo><</mo><mn>0</mn><mo>.</mo><mn>05</mn></mrow></math></span>). Our findings indicate a strong correlation between the deep learning-based classification of raw eye movement data and subjective cognitive load assessments. This study demonstrates the potential of using an affordable eye tracking sensor to classify cognitive load without being constrained by the capability of proprietary software.</div></div>","PeriodicalId":8417,"journal":{"name":"Array","volume":"29 ","pages":"Article 100669"},"PeriodicalIF":4.5,"publicationDate":"2026-01-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145921320","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-31DOI: 10.1016/j.array.2025.100661
Hossein Taghizad, Michel Lemaire, Daniel Massicotte
This review examines multi-rate real-time simulation (MR-RTS) techniques, models, and frameworks. Although established for years, these approaches continue to evolve to address increasingly complex systems. The paper highlights the advantages of multi-rate simulations over traditional fixed-rate methods, emphasizing their ability to adapt simulation rates to the needs of individual subsystems. It also outlines key evaluation criteria, guidance for selecting suitable frameworks, and major challenges in implementing MR-RTS, along with recommendations to overcome them.
{"title":"Multi-rate real-time simulation: Techniques, models, frameworks, and challenges","authors":"Hossein Taghizad, Michel Lemaire, Daniel Massicotte","doi":"10.1016/j.array.2025.100661","DOIUrl":"10.1016/j.array.2025.100661","url":null,"abstract":"<div><div>This review examines multi-rate real-time simulation (MR-RTS) techniques, models, and frameworks. Although established for years, these approaches continue to evolve to address increasingly complex systems. The paper highlights the advantages of multi-rate simulations over traditional fixed-rate methods, emphasizing their ability to adapt simulation rates to the needs of individual subsystems. It also outlines key evaluation criteria, guidance for selecting suitable frameworks, and major challenges in implementing MR-RTS, along with recommendations to overcome them.</div></div>","PeriodicalId":8417,"journal":{"name":"Array","volume":"29 ","pages":"Article 100661"},"PeriodicalIF":4.5,"publicationDate":"2025-12-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145921323","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Agentic artificial intelligence systems, particularly Large Language Models (LLMs) and Vision-Language Models (VLMs), are a big change in oncology because they can find and diagnose cancer in ways that have never been done before. In accordance with PRISMA 2020 criteria, we conducted a systematic search across nine databases from January 2023 to September 2025, reviewing 3986 records and incorporating 123 papers that assessed agentic AI in cancer detection and diagnosis. Research demonstrated swift expansion (91.9% published in 2024-2025) across various cancer kinds, with breast (22.0%) and lung cancer (13.8%) being the most extensively examined. GPT-4 versions showed performance similar to that of human experts: they found errors better than pathologists (89.5% vs. 88.5%), classified skin lesions as well as dermatologists (84.8% vs. 84.6%), and staged ovarian cancer with 97% accuracy compared to 88% by radiologists. Zero-shot LLMs consistently surpassed conventional supervised models. But there were big problems, like factual errors in 15%–41% of instances, algorithmic bias, and low agreement with tumor boards (50%–70%). Agentic AI has a lot of promise for finding cancer, especially in organized tasks. However, the research so far suggests that it should be used as an aid rather than an independent system. Concerns about reliability and bias in algorithms are two of the most important impediments. Future priorities encompass Retrieval-Augmented Generation(RAG) systems, domain-specific models, and forthcoming trials to ascertain clinical value.
{"title":"Agentic artificial intelligence is the future of cancer detection and diagnosis","authors":"Sayedur Rahman , Md. Tanzib Hosain , Nafiz Fahad , Md. Kishor Morol , Md. Jakir Hossen","doi":"10.1016/j.array.2025.100676","DOIUrl":"10.1016/j.array.2025.100676","url":null,"abstract":"<div><div>Agentic artificial intelligence systems, particularly Large Language Models (LLMs) and Vision-Language Models (VLMs), are a big change in oncology because they can find and diagnose cancer in ways that have never been done before. In accordance with PRISMA 2020 criteria, we conducted a systematic search across nine databases from January 2023 to September 2025, reviewing 3986 records and incorporating 123 papers that assessed agentic AI in cancer detection and diagnosis. Research demonstrated swift expansion (91.9% published in 2024-2025) across various cancer kinds, with breast (22.0%) and lung cancer (13.8%) being the most extensively examined. GPT-4 versions showed performance similar to that of human experts: they found errors better than pathologists (89.5% vs. 88.5%), classified skin lesions as well as dermatologists (84.8% vs. 84.6%), and staged ovarian cancer with 97% accuracy compared to 88% by radiologists. Zero-shot LLMs consistently surpassed conventional supervised models. But there were big problems, like factual errors in 15%–41% of instances, algorithmic bias, and low agreement with tumor boards (50%–70%). Agentic AI has a lot of promise for finding cancer, especially in organized tasks. However, the research so far suggests that it should be used as an aid rather than an independent system. Concerns about reliability and bias in algorithms are two of the most important impediments. Future priorities encompass Retrieval-Augmented Generation(RAG) systems, domain-specific models, and forthcoming trials to ascertain clinical value.</div></div>","PeriodicalId":8417,"journal":{"name":"Array","volume":"29 ","pages":"Article 100676"},"PeriodicalIF":4.5,"publicationDate":"2025-12-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145921324","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-31DOI: 10.1016/j.array.2025.100675
Riadul Islam Rabbi , Poh Ping Em , Md. Jakir Hossen
Drowsy driving is widespread and a significant cause of traffic accidents, and thus poses a serious threat to life and property around the globe. Therefore, real-time driver drowsiness detection has emerged as a primary study area, particularly due to the current advancements that incorporate artificial intelligence (AI) into automobiles. Convolutional Neural Networks (CNNs) have recently been very effective in handling image data and feature extraction for detecting drowsiness based on facial and eye movement patterns. This review paper focuses on the different CNN architectures and models that exist in the field of driver drowsiness detection and their strengths and limitations. Models like VGGNet, ResNet, and Inception V3 that are used in CNN are elaborated using pseudocode for an easy understanding of how they can be implemented practically. This paper also examines new trends in lightweight CNNs for edge computing as a solution to demands for real-time analytics in constrained environments such as vehicles. Moreover, important issues like data bias, model overfitting, and computational constraints are discussed. Additionally, future perspectives are provided to address these challenges, such as the integration of hybrid models and fusion of multimodal data. This review aims to provide a comprehensive understanding of CNN-based drowsiness detection and assist in developing safe and reliable automotive applications.
{"title":"Smart driving with AI: A review of CNN approaches to drowsiness detection","authors":"Riadul Islam Rabbi , Poh Ping Em , Md. Jakir Hossen","doi":"10.1016/j.array.2025.100675","DOIUrl":"10.1016/j.array.2025.100675","url":null,"abstract":"<div><div>Drowsy driving is widespread and a significant cause of traffic accidents, and thus poses a serious threat to life and property around the globe. Therefore, real-time driver drowsiness detection has emerged as a primary study area, particularly due to the current advancements that incorporate artificial intelligence (AI) into automobiles. Convolutional Neural Networks (CNNs) have recently been very effective in handling image data and feature extraction for detecting drowsiness based on facial and eye movement patterns. This review paper focuses on the different CNN architectures and models that exist in the field of driver drowsiness detection and their strengths and limitations. Models like VGGNet, ResNet, and Inception V3 that are used in CNN are elaborated using pseudocode for an easy understanding of how they can be implemented practically. This paper also examines new trends in lightweight CNNs for edge computing as a solution to demands for real-time analytics in constrained environments such as vehicles. Moreover, important issues like data bias, model overfitting, and computational constraints are discussed. Additionally, future perspectives are provided to address these challenges, such as the integration of hybrid models and fusion of multimodal data. This review aims to provide a comprehensive understanding of CNN-based drowsiness detection and assist in developing safe and reliable automotive applications.</div></div>","PeriodicalId":8417,"journal":{"name":"Array","volume":"29 ","pages":"Article 100675"},"PeriodicalIF":4.5,"publicationDate":"2025-12-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145921318","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-30DOI: 10.1016/j.array.2025.100659
Venkaiah Chowdary Bhimineni, Rajiv Senapati
High-dimensional (HD) biomedical data, such as gene expression profiles and ECG signals, pose significant challenges for machine learning (ML) due to limited sample size, feature redundancy, and noisy distributions. Conventional models tend to overfit, whereas boosting and ensemble approaches struggle with irrelevant features. Deep autoencoders (DAE) reduce nonlinear dimensionality but miss complex dependencies, whereas transformers require large datasets to model long-range relationships through self-attention mechanisms. We propose a Transformer-based Attention-guided Deep Ensemble Network (Trans-ADENet) that integrates dimensionality reduction, attention-driven feature learning, and meta-level ensemble fusion in an end-to-end framework. A deep autoencoder compresses HD inputs into compact latent representations, refined by a Transformer Encoder with multi-head self-attention. Refined features are fed to diverse base classifiers (CatBoost, Support Vector Machine (SVM), TabNet, and Generalized Multi-Layer Perceptron (GMLP)), and their outputs are fused by a meta-MLP, which learns adaptive weights to yield robust predictions. Experiments on breast, leukemia, INCART2 and Thyroid-RNA datasets achieved 96.3%, 94.1%, 92.7% and 94.6% accuracy, surpassing state-of-the-art models in terms of accuracy, F1, precision, recall, and AUC. By combining representation learning, attention, and adaptive fusion, Trans-ADENet delivers accurate, interpretable classification for biomedical tasks.
{"title":"Trans-ADENet: Transformer-based Attention-guided Deep Ensemble Network for high-dimensional data classification","authors":"Venkaiah Chowdary Bhimineni, Rajiv Senapati","doi":"10.1016/j.array.2025.100659","DOIUrl":"10.1016/j.array.2025.100659","url":null,"abstract":"<div><div>High-dimensional (HD) biomedical data, such as gene expression profiles and ECG signals, pose significant challenges for machine learning (ML) due to limited sample size, feature redundancy, and noisy distributions. Conventional models tend to overfit, whereas boosting and ensemble approaches struggle with irrelevant features. Deep autoencoders (DAE) reduce nonlinear dimensionality but miss complex dependencies, whereas transformers require large datasets to model long-range relationships through self-attention mechanisms. We propose a Transformer-based Attention-guided Deep Ensemble Network (Trans-ADENet) that integrates dimensionality reduction, attention-driven feature learning, and meta-level ensemble fusion in an end-to-end framework. A deep autoencoder compresses HD inputs into compact latent representations, refined by a Transformer Encoder with multi-head self-attention. Refined features are fed to diverse base classifiers (CatBoost, Support Vector Machine (SVM), TabNet, and Generalized Multi-Layer Perceptron (GMLP)), and their outputs are fused by a meta-MLP, which learns adaptive weights to yield robust predictions. Experiments on breast, leukemia, INCART2 and Thyroid-RNA datasets achieved 96.3%, 94.1%, 92.7% and 94.6% accuracy, surpassing state-of-the-art models in terms of accuracy, F1, precision, recall, and AUC. By combining representation learning, attention, and adaptive fusion, Trans-ADENet delivers accurate, interpretable classification for biomedical tasks.</div></div>","PeriodicalId":8417,"journal":{"name":"Array","volume":"29 ","pages":"Article 100659"},"PeriodicalIF":4.5,"publicationDate":"2025-12-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145921394","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-30DOI: 10.1016/j.array.2025.100667
Khadija Shahzad , Anum Khashir , Hina Tufail , Abdul Ahad , Zahra Ali , Filipe Madeira , Ivan Miguel Pires
Skin lesions include a variety of abnormalities found on the skin. These may be benign (not cancerous) or malignant (cancerous). Every year, the number of cases of skin cancer increases globally, increasing the death rate. Medical data is scarce because people are reluctant to provide their health information due to privacy concerns. In this research, a decentralized machine learning approach, Federated learning, is the primary focus of the discipline to preserve patient data. Using this method, models are trained independently on several dispersed devices without sharing the data. To balance the data and enrich the dataset, the Synthetic Minority Over-sampling technique with Edited Nearest Neighbors (SMOTEENN) is used in this study. The HAM10000 dataset was benchmarked using a Convolutional Neural Network (CNN). Seven classes of HAM10000 include vascular skin lesions, benign keratosis, actinic keratosis, melanoma, dermatofibroma, and melanocytic nevi. A centralized method yields an accuracy of 99.39%, and f1-score, precision, and recall of 99.00%. A simulated Federated learning with three clients, ten rounds, and thirty training epochs produced 93.00% precision, 92.00% recall, 92.00% f1-score, and 91.80% accuracy, respectively. At the same time, an increase to four clients and thirty training epochs produced an accuracy, recall, precision, and f1-score of 97.00% with ten rounds.
{"title":"Federated Convolutional Neural Networks (F-CNNs) for privacy-preserving multi-class skin lesion classification","authors":"Khadija Shahzad , Anum Khashir , Hina Tufail , Abdul Ahad , Zahra Ali , Filipe Madeira , Ivan Miguel Pires","doi":"10.1016/j.array.2025.100667","DOIUrl":"10.1016/j.array.2025.100667","url":null,"abstract":"<div><div>Skin lesions include a variety of abnormalities found on the skin. These may be benign (not cancerous) or malignant (cancerous). Every year, the number of cases of skin cancer increases globally, increasing the death rate. Medical data is scarce because people are reluctant to provide their health information due to privacy concerns. In this research, a decentralized machine learning approach, Federated learning, is the primary focus of the discipline to preserve patient data. Using this method, models are trained independently on several dispersed devices without sharing the data. To balance the data and enrich the dataset, the Synthetic Minority Over-sampling technique with Edited Nearest Neighbors (SMOTEENN) is used in this study. The HAM10000 dataset was benchmarked using a Convolutional Neural Network (CNN). Seven classes of HAM10000 include vascular skin lesions, benign keratosis, actinic keratosis, melanoma, dermatofibroma, and melanocytic nevi. A centralized method yields an accuracy of 99.39%, and f1-score, precision, and recall of 99.00%. A simulated Federated learning with three clients, ten rounds, and thirty training epochs produced 93.00% precision, 92.00% recall, 92.00% f1-score, and 91.80% accuracy, respectively. At the same time, an increase to four clients and thirty training epochs produced an accuracy, recall, precision, and f1-score of 97.00% with ten rounds.</div></div>","PeriodicalId":8417,"journal":{"name":"Array","volume":"29 ","pages":"Article 100667"},"PeriodicalIF":4.5,"publicationDate":"2025-12-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145973363","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}