Autonomous driving in developing regions demands perception systems that can operate reliably in unstructured road environments marked by heterogeneous traffic, weak or missing lane geometry, frequent occlusions, and strong appearance variability. Existing semantic segmentation models, although successful in structured Western datasets, exhibit poor generalization to such chaotic conditions and are often too computationally heavy for real-time deployment on low-power edge hardware. To address these gaps, this paper focuses on the challenge of achieving fast, accurate, and resource-efficient segmentation tailored to complex Indian road scenes. We propose IndiSegNet, a lightweight architecture designed explicitly for this setting. The model introduces two novel components—Multi-Scale Contextual Features (MSCF) for capturing irregular object scales and Encoded Features Refining (EFR) for enhancing thin-structure and boundary detail, resulting in a more stable representation for unstructured environments. IndiSegNet achieves 67.2% mIoU on IDD, 78.9% on Cityscapes, and 74.6% on CamVid, while sustaining 112 FPS on Jetson Nano, outperforming standard baselines by 12%–18% IoU on safety-critical classes such as pedestrians, riders, and vehicles. Real-world evaluation across urban, monsoonal, rural, and mountainous regions shows less than 2.5% variance in mIoU with consistent inference speeds above 108 FPS. These results demonstrate that IndiSegNet offers a practical and hardware-efficient solution for high-speed autonomous navigation in the challenging traffic conditions of developing regions.
{"title":"IndiSegNet: Real-time semantic segmentation for unstructured road scenes in intelligent transportation systems","authors":"Pritam Chakraborty , Anjan Bandyopadhyay , Kushagra Agrawal , Jin Zhang , Man-Fai Leung","doi":"10.1016/j.iswa.2026.200629","DOIUrl":"10.1016/j.iswa.2026.200629","url":null,"abstract":"<div><div>Autonomous driving in developing regions demands perception systems that can operate reliably in unstructured road environments marked by heterogeneous traffic, weak or missing lane geometry, frequent occlusions, and strong appearance variability. Existing semantic segmentation models, although successful in structured Western datasets, exhibit poor generalization to such chaotic conditions and are often too computationally heavy for real-time deployment on low-power edge hardware. To address these gaps, this paper focuses on the challenge of achieving fast, accurate, and resource-efficient segmentation tailored to complex Indian road scenes. We propose IndiSegNet, a lightweight architecture designed explicitly for this setting. The model introduces two novel components—Multi-Scale Contextual Features (MSCF) for capturing irregular object scales and Encoded Features Refining (EFR) for enhancing thin-structure and boundary detail, resulting in a more stable representation for unstructured environments. IndiSegNet achieves 67.2% mIoU on IDD, 78.9% on Cityscapes, and 74.6% on CamVid, while sustaining 112 FPS on Jetson Nano, outperforming standard baselines by 12%–18% IoU on safety-critical classes such as pedestrians, riders, and vehicles. Real-world evaluation across urban, monsoonal, rural, and mountainous regions shows less than 2.5% variance in mIoU with consistent inference speeds above 108 FPS. These results demonstrate that IndiSegNet offers a practical and hardware-efficient solution for high-speed autonomous navigation in the challenging traffic conditions of developing regions.</div></div>","PeriodicalId":100684,"journal":{"name":"Intelligent Systems with Applications","volume":"29 ","pages":"Article 200629"},"PeriodicalIF":4.3,"publicationDate":"2026-01-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146026198","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-01-15DOI: 10.1016/j.iswa.2025.200622
Pei Xue, Yuanchun Ye
We propose a deep reinforcement learning framework for dynamic portfolio optimization that combines a Dirichlet policy with cross-sectional attention mechanisms. The Dirichlet distribution enforces feasibility by construction, accommodates tradability masks, and provides a coherent geometry for exploration. Our architecture integrates per-asset temporal encoders with a global attention layer, allowing the policy to adaptively weight sectoral co-movements, factor spillovers, and other cross-asset dependencies. We evaluate the framework on a comprehensive S&P 500 panel from 2000 to 2025 using purged walk-forward backtesting to prevent look-ahead bias. Empirical results show that attention-enhanced Dirichlet policies deliver higher terminal wealth, Sharpe and Sortino ratios than equal-weight and reinforcement learning baselines, while maintaining realistic turnover and drawdown profiles. Our findings highlight that principled action parameterization and attention-based representation learning materially improve both the stability and interpretability of reinforcement learning methods for portfolio allocation.
{"title":"Attention-enhanced reinforcement learning for dynamic portfolio optimization","authors":"Pei Xue, Yuanchun Ye","doi":"10.1016/j.iswa.2025.200622","DOIUrl":"10.1016/j.iswa.2025.200622","url":null,"abstract":"<div><div>We propose a deep reinforcement learning framework for dynamic portfolio optimization that combines a Dirichlet policy with cross-sectional attention mechanisms. The Dirichlet distribution enforces feasibility by construction, accommodates tradability masks, and provides a coherent geometry for exploration. Our architecture integrates per-asset temporal encoders with a global attention layer, allowing the policy to adaptively weight sectoral co-movements, factor spillovers, and other cross-asset dependencies. We evaluate the framework on a comprehensive S&P 500 panel from 2000 to 2025 using purged walk-forward backtesting to prevent look-ahead bias. Empirical results show that attention-enhanced Dirichlet policies deliver higher terminal wealth, Sharpe and Sortino ratios than equal-weight and reinforcement learning baselines, while maintaining realistic turnover and drawdown profiles. Our findings highlight that principled action parameterization and attention-based representation learning materially improve both the stability and interpretability of reinforcement learning methods for portfolio allocation.</div></div>","PeriodicalId":100684,"journal":{"name":"Intelligent Systems with Applications","volume":"29 ","pages":"Article 200622"},"PeriodicalIF":4.3,"publicationDate":"2026-01-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145976713","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Bound-constrained single-objective optimization and constrained engineering design often feature heterogeneous landscapes and barrier-like structures, motivating search procedures that are scale-aware, robust near constraints, and economical in tuning.
Contributions:
We introduce Quantum Tunneling and Fractional Calculus-Based Metaheuristic (QTFM), a physics-inspired metaheuristic that is parameter-lean and employs bounded, range-aware operators to reduce sensitivity to tuning and to prevent erratic steps close to constraints.
Methodology:
QTFM couples fractional-step dynamics for scale-aware exploitation with a quantum-tunneling jump for barrier crossing, and augments these with a wavefunction-collapse local search that averages a small neighborhood and applies minimal perturbations to accelerate refinement without sacrificing diversity.
Results:
On the IEEE Congress on Evolutionary Computation CEC 2022 single-objective bound-constrained suite, QTFM ranked first on ten of twelve functions; it reached the best optimum on F1 and achieved the best mean values on F2–F8 and F10–F11 with stable standard deviations. In three constrained engineering problems, QTFM produced the lowest mean and the best-found solution for the robotic gripper design, and the lowest mean for the planetary gear train and three-bar truss design.
Findings:
The proposed fractional–quantum approach delivers fast, accurate, and robust search across heterogeneous landscapes and real-world design problems.
{"title":"Novel quantum tunneling and fractional calculus-based metaheuristic for robust global data optimization and its applications in engineering design","authors":"Hussam Fakhouri , Riyad Alrousan , Niveen Halalsheh , Najem Sirhan , Jamal Zraqou , Khalil Omar","doi":"10.1016/j.iswa.2025.200616","DOIUrl":"10.1016/j.iswa.2025.200616","url":null,"abstract":"<div><h3>Background:</h3><div>Bound-constrained single-objective optimization and constrained engineering design often feature heterogeneous landscapes and barrier-like structures, motivating search procedures that are scale-aware, robust near constraints, and economical in tuning.</div></div><div><h3>Contributions:</h3><div>We introduce Quantum Tunneling and Fractional Calculus-Based Metaheuristic (QTFM), a physics-inspired metaheuristic that is parameter-lean and employs bounded, range-aware operators to reduce sensitivity to tuning and to prevent erratic steps close to constraints.</div></div><div><h3>Methodology:</h3><div>QTFM couples fractional-step dynamics for scale-aware exploitation with a quantum-tunneling jump for barrier crossing, and augments these with a wavefunction-collapse local search that averages a small neighborhood and applies minimal perturbations to accelerate refinement without sacrificing diversity.</div></div><div><h3>Results:</h3><div>On the IEEE Congress on Evolutionary Computation CEC 2022 single-objective bound-constrained suite, QTFM ranked first on ten of twelve functions; it reached the best optimum on F1 and achieved the best mean values on F2–F8 and F10–F11 with stable standard deviations. In three constrained engineering problems, QTFM produced the lowest mean and the best-found solution for the robotic gripper design, and the lowest mean for the planetary gear train and three-bar truss design.</div></div><div><h3>Findings:</h3><div>The proposed fractional–quantum approach delivers fast, accurate, and robust search across heterogeneous landscapes and real-world design problems.</div></div>","PeriodicalId":100684,"journal":{"name":"Intelligent Systems with Applications","volume":"29 ","pages":"Article 200616"},"PeriodicalIF":4.3,"publicationDate":"2026-01-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145976712","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-01-08DOI: 10.1016/j.iswa.2026.200628
Thanh Tung Luu , Duy An Huynh
Rolling bearing degradation produces vibration signatures that vary across operating conditions, posing challenges for reliable fault diagnosis. This study proposes an adaptive and lightweight diagnostic framework combining a Depthwise Separable Multi-Scale CNN (DSMSCNN) with Convolutional Block Attention Module (CBAM) and Spatial Pyramid Pooling (SPP) to extract fault-frequency invariant features across different mechanical domains. Wavelet-based time–frequency maps are utilized to suppress noise and preserve multi-resolution spectral characteristics. The multi-scale separable convolutions adaptively capture discriminative frequency patterns, while CBAM highlights informative spectral regions and SPP enhances scale robustness without fixed input sizes. Experiments on the CWRU and HUST bearing datasets demonstrate over 99 % accuracy with significantly fewer parameters than conventional CNNs. The results confirm that the proposed DSMSCNN-CBAM-SPP framework effectively captures invariant fault-frequency features, offering a compact and adaptive solution for intelligent bearing fault diagnosis and real-time predictive maintenance in a noisy environment.
{"title":"An efficient lightweight multi-scale CNN framework with CBAM and SPP for bearing fault diagnosis","authors":"Thanh Tung Luu , Duy An Huynh","doi":"10.1016/j.iswa.2026.200628","DOIUrl":"10.1016/j.iswa.2026.200628","url":null,"abstract":"<div><div>Rolling bearing degradation produces vibration signatures that vary across operating conditions, posing challenges for reliable fault diagnosis. This study proposes an adaptive and lightweight diagnostic framework combining a Depthwise Separable Multi-Scale CNN (DSMSCNN) with Convolutional Block Attention Module (CBAM) and Spatial Pyramid Pooling (SPP) to extract fault-frequency invariant features across different mechanical domains. Wavelet-based time–frequency maps are utilized to suppress noise and preserve multi-resolution spectral characteristics. The multi-scale separable convolutions adaptively capture discriminative frequency patterns, while CBAM highlights informative spectral regions and SPP enhances scale robustness without fixed input sizes. Experiments on the CWRU and HUST bearing datasets demonstrate over 99 % accuracy with significantly fewer parameters than conventional CNNs. The results confirm that the proposed DSMSCNN-CBAM-SPP framework effectively captures invariant fault-frequency features, offering a compact and adaptive solution for intelligent bearing fault diagnosis and real-time predictive maintenance in a noisy environment.</div></div>","PeriodicalId":100684,"journal":{"name":"Intelligent Systems with Applications","volume":"29 ","pages":"Article 200628"},"PeriodicalIF":4.3,"publicationDate":"2026-01-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145976714","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
We evaluate a personalized, two-stage comparison-based FER framework on two datasets of low-to-mid-intensity, near-neutral expressions. The framework consistently outperforms FaceReader and Py-Feat. On the natural-transition younger-adult dataset (Dataset A, n = 9), mean accuracy is 90.22% ± 3.53%, with within-subject median gains of +16.46 percentage points (pp) over FaceReader (95% CI [+11.33, +33.90], p = 0.00195, r = 1.00) and +8.17 pp over Py-Feat (95% CI [+3.39, +21.58], p = 0.00195, r = 1.00). On the older adults dataset (Dataset B, n = 78), mean accuracy is 75.58% ± 9.04%, exceeding FaceReader by +15.47 pp (95% CI [+13.44, +17.21], p = 2.77 × 10–14, r = 0.980) and Py-Feat by +17.67 pp (95% CI [+15.13, +19.34], p = 3.02 × 10–14, r = 0.985). Component analyses are above chance on both datasets (B-stage medians 92.90% and 99.51%), and polarity-specific asymmetries emerge in the C-stage (A: positive > negative, Δ = +4.23 pp, two-sided p = 0.0273; B: negative > positive, Δ = -7.72 pp, p = 0.00442). On a subset of Dataset A emphasizing subtle transitions, the system maintains [78.61%, 85.38%] accuracy where human annotation accuracy ranges [50.00%, 71.47%]. Grad-CAM highlights eyebrows, forehead, and mouth regions consistent with expressive cues. Collectively, these findings demonstrate statistically significant and practically meaningful advantages for low-to-mid-intensity expression recognition and intensity ranking.
我们在两个低到中等强度、接近中性表达的数据集上评估了一个个性化的、基于两阶段比较的FER框架。该框架始终优于FaceReader和Py-Feat。在自然过渡的年轻人-成年人数据集(数据集A, n = 9)上,平均准确率为90.22%±3.53%,比FaceReader (95% CI [+11.33, +33.90], p = 0.00195, r = 1.00)和Py-Feat (95% CI [+3.39, +21.58], p = 0.00195, r = 1.00)的受试者内中位增益+16.46个百分点(pp)。在老年人数据集(数据集B, n = 78)上,平均准确率为75.58%±9.04%,超过FaceReader +15.47 pp (95% CI [+13.44, +17.21], p = 2.77 × 10-14, r = 0.980)和Py-Feat +17.67 pp (95% CI [+15.13, +19.34], p = 3.02 × 10-14, r = 0.985)。成分分析在两个数据集上都高于偶然(B期中位数为92.90%和99.51%),并且极性特异性不对称出现在c期(A:阳性>;阴性,Δ = +4.23 pp,双面p = 0.0273; B:阴性>;阳性,Δ = -7.72 pp, p = 0.00442)。在强调微妙过渡的Dataset a子集上,系统保持了[78.61%,85.38%]的准确率,而人类标注的准确率范围为[50.00%,71.47%]。Grad-CAM突出眉毛、前额和嘴部与表达线索一致。综上所述,这些发现显示了在低到中强度表达识别和强度排序方面具有统计学意义和实际意义的优势。
{"title":"Personalized two-stage comparison-based framework for low-to-mid-intensity facial expression recognition in real-world scenarios","authors":"Junyao Zhang , Kei Shimonishi , Kazuaki Kondo , Yuichi Nakamura","doi":"10.1016/j.iswa.2026.200627","DOIUrl":"10.1016/j.iswa.2026.200627","url":null,"abstract":"<div><div>We evaluate a personalized, two-stage comparison-based FER framework on two datasets of low-to-mid-intensity, near-neutral expressions. The framework consistently outperforms FaceReader and Py-Feat. On the natural-transition younger-adult dataset (Dataset A, <em>n</em> = 9), mean accuracy is 90.22% ± 3.53%, with within-subject median gains of +16.46 percentage points (pp) over FaceReader (95% CI [+11.33, +33.90], <em>p</em> = 0.00195, <em>r</em> = 1.00) and +8.17 pp over Py-Feat (95% CI [+3.39, +21.58], <em>p</em> = 0.00195, <em>r</em> = 1.00). On the older adults dataset (Dataset B, <em>n</em> = 78), mean accuracy is 75.58% ± 9.04%, exceeding FaceReader by +15.47 pp (95% CI [+13.44, +17.21], <em>p</em> = 2.77 × 10<sup>–14</sup>, <em>r</em> = 0.980) and Py-Feat by +17.67 pp (95% CI [+15.13, +19.34], <em>p</em> = 3.02 × 10<sup>–14</sup>, <em>r</em> = 0.985). Component analyses are above chance on both datasets (B-stage medians 92.90% and 99.51%), and polarity-specific asymmetries emerge in the C-stage (A: positive > negative, Δ = +4.23 pp, two-sided <em>p</em> = 0.0273; B: negative > positive, Δ = -7.72 pp, <em>p</em> = 0.00442). On a subset of Dataset A emphasizing subtle transitions, the system maintains [78.61%, 85.38%] accuracy where human annotation accuracy ranges [50.00%, 71.47%]. Grad-CAM highlights eyebrows, forehead, and mouth regions consistent with expressive cues. Collectively, these findings demonstrate statistically significant and practically meaningful advantages for low-to-mid-intensity expression recognition and intensity ranking.</div></div>","PeriodicalId":100684,"journal":{"name":"Intelligent Systems with Applications","volume":"29 ","pages":"Article 200627"},"PeriodicalIF":4.3,"publicationDate":"2026-01-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146026199","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Recent advancements in Large Language Models (LLMs) and Generative Artificial Intelligence (GenAI) have revolutionised software engineering (SE), augmenting practitioners across the SE lifecycle. In this paper, we focus on the application of GenAI within data analytics—considered a subdomain of SE—to address the growing need for reliable, user-friendly tools that bridge the gap between human expertise and automated analytical processes. In our work, we transform a conventional API-based analytics platform into a set of tools that can be used by AI agents and formulate a process to facilitate the communication between the data analyst, the agents and the platform. The result is a chat-based interface that allows analysts to query and execute analytical workflows using natural language, thereby reducing cognitive overhead and technical barriers. To validate our approach, we instantiated the proposed framework with open-source models and achieved a mean overall score increase of 7.2 % compared to other baselines. Complementary user-study data demonstrate that the chat-based analytics interface yielded superior task efficiency and higher user preference scores compared to the traditional form-based baseline.
{"title":"Generative AI for autonomous data analytics","authors":"Mattheos Fikardos , Katerina Lepenioti , Alexandros Bousdekis , Dimitris Apostolou , Gregoris Mentzas","doi":"10.1016/j.iswa.2026.200626","DOIUrl":"10.1016/j.iswa.2026.200626","url":null,"abstract":"<div><div>Recent advancements in Large Language Models (LLMs) and Generative Artificial Intelligence (GenAI) have revolutionised software engineering (SE), augmenting practitioners across the SE lifecycle. In this paper, we focus on the application of GenAI within data analytics—considered a subdomain of SE—to address the growing need for reliable, user-friendly tools that bridge the gap between human expertise and automated analytical processes. In our work, we transform a conventional API-based analytics platform into a set of tools that can be used by AI agents and formulate a process to facilitate the communication between the data analyst, the agents and the platform. The result is a chat-based interface that allows analysts to query and execute analytical workflows using natural language, thereby reducing cognitive overhead and technical barriers. To validate our approach, we instantiated the proposed framework with open-source models and achieved a mean overall score increase of 7.2 % compared to other baselines. Complementary user-study data demonstrate that the chat-based analytics interface yielded superior task efficiency and higher user preference scores compared to the traditional form-based baseline.</div></div>","PeriodicalId":100684,"journal":{"name":"Intelligent Systems with Applications","volume":"29 ","pages":"Article 200626"},"PeriodicalIF":4.3,"publicationDate":"2026-01-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145924622","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Skeleton-based gait recognition has significantly improved due to the advent of graph convolutional networks (GCNs). Nevertheless, the classical ST-GCN has a key drawback: limited receptive fields fail to learn the global correlations of joints, restricting its ability to extract global dependencies effectively. To address this, we present the GSCTN method, a GCN and self-attention contemporary network with temporal convolution. This method combines GCN with a self-attention mechanism using a learnable weighted fusion. By combining local joint details from GCN with the larger context from self-attention, GSCTN creates a strong representation of skeleton movements. Our approach uses decoupled self-attention (DSA) techniques that fragment the tightly coupled (TiC) SA module into two learnable components, unary and pairwise SA, to model joint relationships separately. The unary SA shows an extensive relationship between the single key joint and all additional query joints. The paired SA captures the local gait features from each pair of body joints. We also present a Depthwise Multi-scale Temporal Convolutional Network (DMS-TCN) that smoothly captures the temporal nature of joint movements. DMS-TCN efficiently handles both short-term and long-term motion patterns. To boost the model’s ability to converge spatial and temporal joints dynamically, we applied Global Aware Attention (GAA) to the GSCTN module. We tested our method on the OUMVLP-Pose, CASIA-B, and GREW datasets. The suggested method exhibits remarkable accuracy on widely used CASIA-B datasets, with 97.9% for normal walking, 94.8% for carrying a bag, and 91.91% for clothing conditions. Meanwhile, the OUMVLP-Pose and GREW datasets exhibit a rank-1 accuracy of 93.5% and 75.7%, respectively. Our experimental results demonstrate that the proposed model is a holistic approach for gait recognition by utilizing GCN, DSA, and GAA with DMS-TCN to capture both inter-domain and spatial aspects of human locomotion.
{"title":"A GCN and Graph Self-Attention Contemporary Network with Temporal Depthwise Convolutions for Gait Recognition","authors":"Md. Khaliluzzaman , Kaushik Deb , Pranab Kumar Dhar , Tetsuya Shimamura","doi":"10.1016/j.iswa.2025.200625","DOIUrl":"10.1016/j.iswa.2025.200625","url":null,"abstract":"<div><div>Skeleton-based gait recognition has significantly improved due to the advent of graph convolutional networks (GCNs). Nevertheless, the classical ST-GCN has a key drawback: limited receptive fields fail to learn the global correlations of joints, restricting its ability to extract global dependencies effectively. To address this, we present the GSCTN method, a GCN and self-attention contemporary network with temporal convolution. This method combines GCN with a self-attention mechanism using a learnable weighted fusion. By combining local joint details from GCN with the larger context from self-attention, GSCTN creates a strong representation of skeleton movements. Our approach uses decoupled self-attention (DSA) techniques that fragment the tightly coupled (TiC) SA module into two learnable components, unary and pairwise SA, to model joint relationships separately. The unary SA shows an extensive relationship between the single key joint and all additional query joints. The paired SA captures the local gait features from each pair of body joints. We also present a Depthwise Multi-scale Temporal Convolutional Network (DMS-TCN) that smoothly captures the temporal nature of joint movements. DMS-TCN efficiently handles both short-term and long-term motion patterns. To boost the model’s ability to converge spatial and temporal joints dynamically, we applied Global Aware Attention (GAA) to the GSCTN module. We tested our method on the OUMVLP-Pose, CASIA-B, and GREW datasets. The suggested method exhibits remarkable accuracy on widely used CASIA-B datasets, with 97.9% for normal walking, 94.8% for carrying a bag, and 91.91% for clothing conditions. Meanwhile, the OUMVLP-Pose and GREW datasets exhibit a rank-1 accuracy of 93.5% and 75.7%, respectively. Our experimental results demonstrate that the proposed model is a holistic approach for gait recognition by utilizing GCN, DSA, and GAA with DMS-TCN to capture both inter-domain and spatial aspects of human locomotion.</div></div>","PeriodicalId":100684,"journal":{"name":"Intelligent Systems with Applications","volume":"29 ","pages":"Article 200625"},"PeriodicalIF":4.3,"publicationDate":"2025-12-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145924624","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
The Internet of Things is an enormous network of interrelated devices that makes intelligent interaction and high-level control possible in various environments, such as smart homes, smart cities, and industry, by collecting, processing, and transferring data. The majority of the low-power devices within the network utilize limited sources of energy, such as batteries, and hence energy management is a critical factor in the design and operation of the systems. Current methods, such as reinforcement and evolutionary approaches, have at times been found to provide some enhancements but lacked extensive implementation over broad systems due to computational complexity as well as their inability to adapt to changing environmental settings. The growing number of IoT devices presents challenges in energy management, making it crucial to develop accurate prediction models. This research aims to address this challenge by proposing a novel solution using Long Short-Term Memory (LSTM) networks for energy consumption forecasting. This work suggests an optimal energy usage management model based on Long Short-Term Memory networks. The model collects historical energy usage, activity scheduling, and environmental factors such as temperature and humidity. Following the preprocessing, which includes noise removal and normalisation, it predicts future energy consumption. Scheduling data and the analysis and processing of environmental conditions are done using the short-term memory, while the long-term memory helps the model identify more complex patterns in the energy consumption over time to make more accurate predictions. Based on this prediction, smart policies are made for going to sleep and waking up the devices, so that unnecessary devices are put into sleep mode and only woken up when needed. Adaptive learning algorithms also assist in adjusting to environmental conditions. Results of experiments show that the proposed method can save energy up to 58% and increase device lifetime by 30%, while the prediction of energy consumption has an accuracy of 95%.
{"title":"Optimisation of energy management in IoT devices using LSTM models: Energy consumption prediction with sleep-wake scheduling control","authors":"Nahideh DerakhshanFard, Asra Rajabi Bavil Olyaei, Fahimeh RashidJafari","doi":"10.1016/j.iswa.2025.200624","DOIUrl":"10.1016/j.iswa.2025.200624","url":null,"abstract":"<div><div>The Internet of Things is an enormous network of interrelated devices that makes intelligent interaction and high-level control possible in various environments, such as smart homes, smart cities, and industry, by collecting, processing, and transferring data. The majority of the low-power devices within the network utilize limited sources of energy, such as batteries, and hence energy management is a critical factor in the design and operation of the systems. Current methods, such as reinforcement and evolutionary approaches, have at times been found to provide some enhancements but lacked extensive implementation over broad systems due to computational complexity as well as their inability to adapt to changing environmental settings. The growing number of IoT devices presents challenges in energy management, making it crucial to develop accurate prediction models. This research aims to address this challenge by proposing a novel solution using Long Short-Term Memory (LSTM) networks for energy consumption forecasting. This work suggests an optimal energy usage management model based on Long Short-Term Memory networks. The model collects historical energy usage, activity scheduling, and environmental factors such as temperature and humidity. Following the preprocessing, which includes noise removal and normalisation, it predicts future energy consumption. Scheduling data and the analysis and processing of environmental conditions are done using the short-term memory, while the long-term memory helps the model identify more complex patterns in the energy consumption over time to make more accurate predictions. Based on this prediction, smart policies are made for going to sleep and waking up the devices, so that unnecessary devices are put into sleep mode and only woken up when needed. Adaptive learning algorithms also assist in adjusting to environmental conditions. Results of experiments show that the proposed method can save energy up to 58% and increase device lifetime by 30%, while the prediction of energy consumption has an accuracy of 95%.</div></div>","PeriodicalId":100684,"journal":{"name":"Intelligent Systems with Applications","volume":"29 ","pages":"Article 200624"},"PeriodicalIF":4.3,"publicationDate":"2025-12-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146026200","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-22DOI: 10.1016/j.iswa.2025.200623
Deepa D. Shankar , Adresya Suresh Azhakath
Information technology and digital media have significantly improved in recent years, facilitating the internet as an effective channel for communication and data transmission. Nevertheless, the rapid advancement of technology has rendered data a source of mismanagement and vulnerable to exploitation. Consequently, technologies such as data concealment were devised to mitigate exploitation. Steganalysis is a technique for data concealment. Various processes, including breaches of information security, can be mitigated by steganalysis. This work aims to encapsulate the notion of blind statistical steganalysis within image processing methodologies and ascertain the accuracy percentage of secure transmission. This work discusses the extraction of features that indicate a change during embedding. A specific percentage of text is integrated into a JPEG image of a predetermined size. The text embedding utilizes various steganographic techniques in both the spatial and transform domains. The steganographic techniques include LSB Matching, LSB Replacement, Pixel Value Differencing, and F5. Due to the blind nature of steganalysis, there are no cover images available for comparative analysis. An estimation of the cover image is obtained by a calibration concept. After embedding, the images are partitioned into 8 × 8 blocks, from which certain features are extraction for classification. This paper utilizes interblock dependent features and intrablock dependent features. Both dependencies are regarded as means to mitigate the shortcomings of each individually. The approach of machine learning is employed using a classifier to distinguish between the stego image and the cover image. This research does a comparative investigation of the classifiers SVM and SVM-PSO. Comparative research is frequently performed both with and without use cross-validation methodology. The study incorporates the concept of cross-validation for comparative analysis. There are six unique kernel functions and four sample methods for grouping. The embedding ratio employed in this investigation is 50%.
{"title":"Blind steganalysis-driven secure transmission validation using feature-based classification in JPEG images","authors":"Deepa D. Shankar , Adresya Suresh Azhakath","doi":"10.1016/j.iswa.2025.200623","DOIUrl":"10.1016/j.iswa.2025.200623","url":null,"abstract":"<div><div>Information technology and digital media have significantly improved in recent years, facilitating the internet as an effective channel for communication and data transmission. Nevertheless, the rapid advancement of technology has rendered data a source of mismanagement and vulnerable to exploitation. Consequently, technologies such as data concealment were devised to mitigate exploitation. Steganalysis is a technique for data concealment. Various processes, including breaches of information security, can be mitigated by steganalysis. This work aims to encapsulate the notion of blind statistical steganalysis within image processing methodologies and ascertain the accuracy percentage of secure transmission. This work discusses the extraction of features that indicate a change during embedding. A specific percentage of text is integrated into a JPEG image of a predetermined size. The text embedding utilizes various steganographic techniques in both the spatial and transform domains. The steganographic techniques include LSB Matching, LSB Replacement, Pixel Value Differencing, and F5. Due to the blind nature of steganalysis, there are no cover images available for comparative analysis. An estimation of the cover image is obtained by a calibration concept. After embedding, the images are partitioned into 8 × 8 blocks, from which certain features are extraction for classification. This paper utilizes interblock dependent features and intrablock dependent features. Both dependencies are regarded as means to mitigate the shortcomings of each individually. The approach of machine learning is employed using a classifier to distinguish between the stego image and the cover image. This research does a comparative investigation of the classifiers SVM and SVM-PSO. Comparative research is frequently performed both with and without use cross-validation methodology. The study incorporates the concept of cross-validation for comparative analysis. There are six unique kernel functions and four sample methods for grouping. The embedding ratio employed in this investigation is 50%.</div></div>","PeriodicalId":100684,"journal":{"name":"Intelligent Systems with Applications","volume":"29 ","pages":"Article 200623"},"PeriodicalIF":4.3,"publicationDate":"2025-12-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145976711","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-18DOI: 10.1016/j.iswa.2025.200620
Elaheh Golzardi , Alireza Abdollahpouri
As social networks are constantly changing, decision-making in large groups becomes much more challenging. People form new connections, lose old ones, shift their preferences, and change how much they trust others (Qin, Li, Liang & Pedrycz, 2026). Methods that work well in stable settings often fail to keep pace here, especially when both quick adaptation and the ability to handle scale are essential (Ding et al., 2025). Our approach, called GCD-GNN (Group Consensus Decision using Graph Neural Networks), builds on graph neural networks to track these ongoing changes in structure and preferences. It processes live updates on trust levels, social ties, and preference similarities, then adjusts influence weights in real time to keep the consensus process stable. In experiments using both synthetic and real-world datasets, GCD-GNN delivered higher agreement levels, improved accuracy and precision, and faster execution compared with leading alternatives. These results point to a framework that is not only scalable, but also able to adapt effectiveness in complex, large-scale decision-making environments.
随着社交网络的不断变化,大群体的决策变得更具挑战性。人们建立新的联系,失去旧的联系,改变他们的偏好,并改变他们对他人的信任程度(秦,李,梁,Pedrycz, 2026)。在稳定环境中工作良好的方法往往无法跟上这里的步伐,特别是当快速适应和处理规模的能力都是必不可少的时候(Ding et al., 2025)。我们的方法,称为GCD-GNN(使用图神经网络的群体共识决策),建立在图神经网络的基础上,跟踪这些结构和偏好的持续变化。它处理信任水平、社会关系和偏好相似性的实时更新,然后实时调整影响权重,以保持共识过程的稳定。在使用合成数据集和真实数据集的实验中,与领先的替代方案相比,GCD-GNN提供了更高的一致性水平,提高了准确性和精度,并且执行速度更快。这些结果表明,该框架不仅具有可扩展性,而且能够适应复杂的大规模决策环境的有效性。
{"title":"Scalable and Adaptive Large-Scale Group Decision Making in Dynamic Social Networks via Graph Convolutional Neural Networks#","authors":"Elaheh Golzardi , Alireza Abdollahpouri","doi":"10.1016/j.iswa.2025.200620","DOIUrl":"10.1016/j.iswa.2025.200620","url":null,"abstract":"<div><div>As social networks are constantly changing, decision-making in large groups becomes much more challenging. People form new connections, lose old ones, shift their preferences, and change how much they trust others (<span><span>Qin, Li, Liang & Pedrycz, 2026</span></span>). Methods that work well in stable settings often fail to keep pace here, especially when both quick adaptation and the ability to handle scale are essential (<span><span>Ding et al., 2025</span></span>). Our approach, called GCD-GNN (Group Consensus Decision using Graph Neural Networks), builds on graph neural networks to track these ongoing changes in structure and preferences. It processes live updates on trust levels, social ties, and preference similarities, then adjusts influence weights in real time to keep the consensus process stable. In experiments using both synthetic and real-world datasets, GCD-GNN delivered higher agreement levels, improved accuracy and precision, and faster execution compared with leading alternatives. These results point to a framework that is not only scalable, but also able to adapt effectiveness in complex, large-scale decision-making environments.</div></div>","PeriodicalId":100684,"journal":{"name":"Intelligent Systems with Applications","volume":"29 ","pages":"Article 200620"},"PeriodicalIF":4.3,"publicationDate":"2025-12-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145924623","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}