Pub Date : 2025-04-24DOI: 10.1007/s10489-025-06573-4
Jun Lei, Jiqian Zhao, Jingqi Wang, An-Bao Xu
This paper proposes a novel method for network latency estimation. Network latency estimation is a crucial indicator for evaluating network performance, yet accurate estimation of large-scale network latency requires substantial computation time. Therefore, this paper introduces a method capable of enhancing the speed of network latency estimation. The paper represents the data structure of network nodes as matrices and introduces a time dimension to form a tensor model, thereby transforming the entire network latency estimation problem into a tensor completion problem. The main contributions of this paper include: optimizing leveraged sampling for tensors to improve sampling speed, and on this basis, introducing the Qatar Riyal (QR) decomposition of tensors into the Alternating Direction Method of Multipliers (ADMM) framework to accelerate tensor completion; these two components are combined to form a new model called LNLS-TQR. Numerical experimental results demonstrate that this model significantly improves computation speed while maintaining high accuracy.
{"title":"Tensor completion via leverage sampling and tensor QR decomposition for network latency estimation","authors":"Jun Lei, Jiqian Zhao, Jingqi Wang, An-Bao Xu","doi":"10.1007/s10489-025-06573-4","DOIUrl":"10.1007/s10489-025-06573-4","url":null,"abstract":"<div><p>This paper proposes a novel method for network latency estimation. Network latency estimation is a crucial indicator for evaluating network performance, yet accurate estimation of large-scale network latency requires substantial computation time. Therefore, this paper introduces a method capable of enhancing the speed of network latency estimation. The paper represents the data structure of network nodes as matrices and introduces a time dimension to form a tensor model, thereby transforming the entire network latency estimation problem into a tensor completion problem. The main contributions of this paper include: optimizing leveraged sampling for tensors to improve sampling speed, and on this basis, introducing the Qatar Riyal (QR) decomposition of tensors into the Alternating Direction Method of Multipliers (ADMM) framework to accelerate tensor completion; these two components are combined to form a new model called LNLS-TQR. Numerical experimental results demonstrate that this model significantly improves computation speed while maintaining high accuracy.</p></div>","PeriodicalId":8041,"journal":{"name":"Applied Intelligence","volume":"55 7","pages":""},"PeriodicalIF":3.4,"publicationDate":"2025-04-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143865478","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
The ability of orthogonal frequency division multiplexing (OFDM) to counteract frequency-selective fading channels has made it a popular modem technology in contemporary communication systems. But maintaining dependable signaling is still difficult, especially when the signal-to-noise ratio (SNR) is low. In order to increase the dependability of OFDM systems, this study presents an enhanced LSTM-based autoencoder architecture. The suggested autoencoder efficiently utilizes temporal dependencies and reduces the impacts of channel distortion by encoding and decoding OFDM signals utilizing one-hot encoding employing long short-term memory (LSTM) networks. The outcomes of the simulation show notable gains in performance indicators. The average block error rate (BLER) of the suggested model is 0.0150, as opposed to 0.0296 for traditional autoencoders and 0.0886 for convolutional OFDM systems. Comparably, the average packet error rate (PER) is decreased to 0.0017, surpassing convolutional OFDM systems' 0.2260 and traditional autoencoders' 0.0070. These outcomes highlight the LSTM-based autoencoder's efficacy in enhancing OFDM systems' dependability, especially in demanding settings. This study lays the groundwork for employing cutting-edge deep learning methods to create reliable and effective communication systems.
{"title":"Enhancing Robustness of OFDM Systems Using LSTM-Based Autoencoders","authors":"Rajarajan P, Madona B. Sahaai","doi":"10.1002/dac.70090","DOIUrl":"https://doi.org/10.1002/dac.70090","url":null,"abstract":"<div>\u0000 \u0000 <p>The ability of orthogonal frequency division multiplexing (OFDM) to counteract frequency-selective fading channels has made it a popular modem technology in contemporary communication systems. But maintaining dependable signaling is still difficult, especially when the signal-to-noise ratio (SNR) is low. In order to increase the dependability of OFDM systems, this study presents an enhanced LSTM-based autoencoder architecture. The suggested autoencoder efficiently utilizes temporal dependencies and reduces the impacts of channel distortion by encoding and decoding OFDM signals utilizing one-hot encoding employing long short-term memory (LSTM) networks. The outcomes of the simulation show notable gains in performance indicators. The average block error rate (BLER) of the suggested model is 0.0150, as opposed to 0.0296 for traditional autoencoders and 0.0886 for convolutional OFDM systems. Comparably, the average packet error rate (PER) is decreased to 0.0017, surpassing convolutional OFDM systems' 0.2260 and traditional autoencoders' 0.0070. These outcomes highlight the LSTM-based autoencoder's efficacy in enhancing OFDM systems' dependability, especially in demanding settings. This study lays the groundwork for employing cutting-edge deep learning methods to create reliable and effective communication systems.</p>\u0000 </div>","PeriodicalId":13946,"journal":{"name":"International Journal of Communication Systems","volume":"38 9","pages":""},"PeriodicalIF":1.7,"publicationDate":"2025-04-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143865759","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-04-24DOI: 10.1016/j.swevo.2025.101941
Li Yan , Shunge Guo , Jing Liang , Boyang Qu , Chao Li , Kunjie Yu
Constrained multimodal multiobjective optimization problems (CMMOPs) consist of multiple equivalent constrained Pareto sets (CPSs) that have the identical constrained Pareto front (CPF). The key to solving CMMOPs lies in how to locate and retain CPSs and CPF in search spaces. Thus, this paper proposes a subspace strategy based coevolutionary framework for CMMOPs, named SCCMMO. Firstly, the subspace generation and maintenance strategy is proposed to efficiently locate multiple CPSs within the decision space. Secondly, the subspace-type perception strategy is used to exploit the feasible and infeasible information in subspaces. Finally, a coevolutionary framework is introduced to improve search efficiency. To prove the effectiveness of the algorithm, the proposed method is compared with ten state-of-the-art algorithms on seventeen benchmarks. The results demonstrate the superiority of SCCMMO in solving CMMOPs. Moreover, SCCMMO also achieves better performance on the real-world problem.
{"title":"A subspace strategy based coevolutionary framework for constrained multimodal multiobjective optimization problems","authors":"Li Yan , Shunge Guo , Jing Liang , Boyang Qu , Chao Li , Kunjie Yu","doi":"10.1016/j.swevo.2025.101941","DOIUrl":"10.1016/j.swevo.2025.101941","url":null,"abstract":"<div><div>Constrained multimodal multiobjective optimization problems (CMMOPs) consist of multiple equivalent constrained Pareto sets (CPSs) that have the identical constrained Pareto front (CPF). The key to solving CMMOPs lies in how to locate and retain CPSs and CPF in search spaces. Thus, this paper proposes a subspace strategy based coevolutionary framework for CMMOPs, named SCCMMO. Firstly, the subspace generation and maintenance strategy is proposed to efficiently locate multiple CPSs within the decision space. Secondly, the subspace-type perception strategy is used to exploit the feasible and infeasible information in subspaces. Finally, a coevolutionary framework is introduced to improve search efficiency. To prove the effectiveness of the algorithm, the proposed method is compared with ten state-of-the-art algorithms on seventeen benchmarks. The results demonstrate the superiority of SCCMMO in solving CMMOPs. Moreover, SCCMMO also achieves better performance on the real-world problem.</div></div>","PeriodicalId":48682,"journal":{"name":"Swarm and Evolutionary Computation","volume":"95 ","pages":"Article 101941"},"PeriodicalIF":8.2,"publicationDate":"2025-04-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143863713","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Existing click-based interactive image segmentation methods typically initiate object extraction with the first click and iteratively refine the coarse segmentation through subsequent interactions. Unlike box-based methods, click-based approaches mitigate ambiguity when multiple targets are present within a single bounding box, but suffer from a lack of precise location and outline information. Inspired by instance segmentation, the authors propose a Generated-bbox Guided method that provides location and outline information using an automatically generated bounding box, rather than a manually labelled one, minimising the need for extensive user interaction. Building on the success of vision transformers, the authors adopt them as the network architecture to enhance model's performance. A click-based interactive image segmentation network named the Generated-bbox Guided Coarse-to-Fine Network (GCFN) was proposed. GCFN is a two-stage cascade network comprising two sub-networks: Coarsenet and Finenet. A transformer-based Box Detector was introduced to generate an initial bounding box from a inside click, that can provide location and outline information. Additionally, two feature enhancement modules guided by foreground and background information: the Foreground-Background Feature Enhancement Module (FFEM) and the Pixel Enhancement Module (PEM) were designed. The authors evaluate the GCFN method on five popular benchmark datasets and demonstrate the generalisation capability on three medical image datasets.
{"title":"The Generated-bbox Guided Interactive Image Segmentation With Vision Transformers","authors":"Shiyin Zhang, Yafei Dong, Shuang Qiu","doi":"10.1049/cvi2.70019","DOIUrl":"https://doi.org/10.1049/cvi2.70019","url":null,"abstract":"<p>Existing click-based interactive image segmentation methods typically initiate object extraction with the first click and iteratively refine the coarse segmentation through subsequent interactions. Unlike box-based methods, click-based approaches mitigate ambiguity when multiple targets are present within a single bounding box, but suffer from a lack of precise location and outline information. Inspired by instance segmentation, the authors propose a Generated-bbox Guided method that provides location and outline information using an automatically generated bounding box, rather than a manually labelled one, minimising the need for extensive user interaction. Building on the success of vision transformers, the authors adopt them as the network architecture to enhance model's performance. A click-based interactive image segmentation network named the Generated-bbox Guided Coarse-to-Fine Network (GCFN) was proposed. GCFN is a two-stage cascade network comprising two sub-networks: Coarsenet and Finenet. A transformer-based Box Detector was introduced to generate an initial bounding box from a inside click, that can provide location and outline information. Additionally, two feature enhancement modules guided by foreground and background information: the Foreground-Background Feature Enhancement Module (FFEM) and the Pixel Enhancement Module (PEM) were designed. The authors evaluate the GCFN method on five popular benchmark datasets and demonstrate the generalisation capability on three medical image datasets.</p>","PeriodicalId":56304,"journal":{"name":"IET Computer Vision","volume":"19 1","pages":""},"PeriodicalIF":1.5,"publicationDate":"2025-04-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1049/cvi2.70019","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143865768","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Fog computing extends cloud computing to the edge of the network, bringing processing and storage capabilities closer to end users and Internet of Things (IoT) devices. This paradigm helps to reduce latency, improve response time, and optimize bandwidth usage. In the cloud computing environment, service availability is a criterion for determining user satisfaction, which is strongly influenced by response time and optimal allocation of network resources (communication bandwidth). Service placement in fog computing refers to the process of determining optimal locations for placing services in the network. In this paper, the service placement is done by being aware of the volume of user requests from fog nodes by using neural networks, reinforcement learning, and the improved gray wolf optimization (IGWO) method. Based on the results obtained from simulation, the proposed approach has less response time (between 5% and 21%), more favorable load balance, more utility value (12%) and lower Energy consumption by a minimum of 10% and a maximum of 25%.
{"title":"Service Placement in Fog Computing Using a Combination of Reinforcement Learning and Improved Gray Wolf Optimization Method","authors":"Pouria Ashkani, Seyyed Hamid Ghafouri, Maliheh Hashemipour","doi":"10.1002/cpe.70097","DOIUrl":"https://doi.org/10.1002/cpe.70097","url":null,"abstract":"<div>\u0000 \u0000 <p>Fog computing extends cloud computing to the edge of the network, bringing processing and storage capabilities closer to end users and Internet of Things (IoT) devices. This paradigm helps to reduce latency, improve response time, and optimize bandwidth usage. In the cloud computing environment, service availability is a criterion for determining user satisfaction, which is strongly influenced by response time and optimal allocation of network resources (communication bandwidth). Service placement in fog computing refers to the process of determining optimal locations for placing services in the network. In this paper, the service placement is done by being aware of the volume of user requests from fog nodes by using neural networks, reinforcement learning, and the improved gray wolf optimization (IGWO) method. Based on the results obtained from simulation, the proposed approach has less response time (between 5% and 21%), more favorable load balance, more utility value (12%) and lower Energy consumption by a minimum of 10% and a maximum of 25%.</p>\u0000 </div>","PeriodicalId":55214,"journal":{"name":"Concurrency and Computation-Practice & Experience","volume":"37 9-11","pages":""},"PeriodicalIF":1.5,"publicationDate":"2025-04-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143865629","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-04-24DOI: 10.1016/j.mechatronics.2025.103316
Augusto H.B.M. Tavares , Saulo O.D. Luiz , Tiago P. Nascimento , Antonio M.N. Lima
This paper studies analytically and quantitatively the influence of the power source on the dynamic performance of the altitude control system of a quadrotor powered by an electrochemical battery. This paper also proposes the formulation of the altitude control system design as a constrained optimization problem in which the drone, actuators, control law, and electrochemical battery models are considered, to define a trade-off between the power consumption rate and the closed-loop dynamic performance loss. An analytical representation of the effect of the battery discharge over the altitude dynamics is obtained through a linear approximation, enabling an analysis of the system poles. The problem of designing an altitude controller is then posed as a constrained optimization problem that can include the battery as a factor. A comparison of the error transient response between the cases of the battery-unaware controller design and the battery-aware controller design is performed in simulations and experimental flight tests. The results lead to the following conclusions: i. the analytical demonstration agrees with the worse performance observed in the in-flight dynamics as the battery discharges and ii. through a battery-aware controller design approach this effect can be diminished, at the cost of a trade-off in the battery discharge rate.
{"title":"Trade-off between flight performance and energy consumption of a quadrotor","authors":"Augusto H.B.M. Tavares , Saulo O.D. Luiz , Tiago P. Nascimento , Antonio M.N. Lima","doi":"10.1016/j.mechatronics.2025.103316","DOIUrl":"10.1016/j.mechatronics.2025.103316","url":null,"abstract":"<div><div>This paper studies analytically and quantitatively the influence of the power source on the dynamic performance of the altitude control system of a quadrotor powered by an electrochemical battery. This paper also proposes the formulation of the altitude control system design as a constrained optimization problem in which the drone, actuators, control law, and electrochemical battery models are considered, to define a trade-off between the power consumption rate and the closed-loop dynamic performance loss. An analytical representation of the effect of the battery discharge over the altitude dynamics is obtained through a linear approximation, enabling an analysis of the system poles. The problem of designing an altitude controller is then posed as a constrained optimization problem that can include the battery as a factor. A comparison of the error transient response between the cases of the battery-unaware controller design and the battery-aware controller design is performed in simulations and experimental flight tests. The results lead to the following conclusions: i. the analytical demonstration agrees with the worse performance observed in the in-flight dynamics as the battery discharges and ii. through a battery-aware controller design approach this effect can be diminished, at the cost of a trade-off in the battery discharge rate.</div></div>","PeriodicalId":49842,"journal":{"name":"Mechatronics","volume":"108 ","pages":"Article 103316"},"PeriodicalIF":3.1,"publicationDate":"2025-04-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143865076","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-04-24DOI: 10.1016/j.softx.2025.102140
Rabindra Khadka , Pedro G. Lind , Anis Yazidi , Asma Belhadi
Electroencephalography (EEG) provides a non-invasive way to observe brain activity in real time. Deep learning has enhanced EEG analysis, enabling meaningful pattern detection for clinical and research purposes. However, most existing frameworks for EEG data analysis are either focused on preprocessing techniques or deep learning model development, often overlooking the crucial need for structured documentation and model interpretability. In this paper, we introduce DREAMS (Deep REport for AI ModelS), a Python-based framework designed to generate automated model cards for deep learning models applied to EEG data. Unlike generic model reporting tools, DREAMS is specifically tailored for EEG-based deep learning applications, incorporating domain-specific metadata, preprocessing details, performance metrics, and uncertainty quantification. The framework seamlessly integrates with deep learning pipelines, providing structured YAML-based documentation. We evaluate DREAMS through two case studies: an EEG emotion classification task using the FACED dataset and a abnormal EEG classification task using the Temple University Hospital (TUH) Abnormal dataset. These evaluations demonstrate how the generated model card enhances transparency by documenting model performance, dataset biases, and interpretability limitations. Unlike existing model documentation approaches, DREAMS provides visualized performance metrics, dataset alignment details, and model uncertainty estimations, making it a valuable tool for researchers and clinicians working with EEG-based AI. The source code for DREAMS is open-source, facilitating broad adoption in healthcare AI, research, and ethical AI development.
{"title":"DREAMS: A python framework for training deep learning models on EEG data with model card reporting for medical applications","authors":"Rabindra Khadka , Pedro G. Lind , Anis Yazidi , Asma Belhadi","doi":"10.1016/j.softx.2025.102140","DOIUrl":"10.1016/j.softx.2025.102140","url":null,"abstract":"<div><div>Electroencephalography (EEG) provides a non-invasive way to observe brain activity in real time. Deep learning has enhanced EEG analysis, enabling meaningful pattern detection for clinical and research purposes. However, most existing frameworks for EEG data analysis are either focused on preprocessing techniques or deep learning model development, often overlooking the crucial need for structured documentation and model interpretability. In this paper, we introduce DREAMS (Deep REport for AI ModelS), a Python-based framework designed to generate automated model cards for deep learning models applied to EEG data. Unlike generic model reporting tools, DREAMS is specifically tailored for EEG-based deep learning applications, incorporating domain-specific metadata, preprocessing details, performance metrics, and uncertainty quantification. The framework seamlessly integrates with deep learning pipelines, providing structured YAML-based documentation. We evaluate DREAMS through two case studies: an EEG emotion classification task using the FACED dataset and a abnormal EEG classification task using the Temple University Hospital (TUH) Abnormal dataset. These evaluations demonstrate how the generated model card enhances transparency by documenting model performance, dataset biases, and interpretability limitations. Unlike existing model documentation approaches, DREAMS provides visualized performance metrics, dataset alignment details, and model uncertainty estimations, making it a valuable tool for researchers and clinicians working with EEG-based AI. The source code for DREAMS is open-source, facilitating broad adoption in healthcare AI, research, and ethical AI development.</div></div>","PeriodicalId":21905,"journal":{"name":"SoftwareX","volume":"30 ","pages":"Article 102140"},"PeriodicalIF":2.4,"publicationDate":"2025-04-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143865085","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-04-24DOI: 10.1016/j.engappai.2025.110956
Kun Yue , Liming Wang , Xiaoxi Ding , Wennian Yu , Zaigang Chen , Wenbin Huang
Recent years have seen artificial intelligence algorithms gain considerable popularity in gear fault classification, yet their performance remains hindered by the scarcity of labeled fault data, often leading to suboptimal classification results. Several state-of-the-art studies have demonstrated that incorporating physical information can improve classification accuracy. However, the differences between simulated and measured signals pose a significant challenge in enhancing the performance of physics-informed methods. In order to fill this gap, this paper introduces a novel physics-informed dual guidance (PI-DG) method using physical envelope harmonic distribution (PEHD) and transfer learning (TL) for few-shot gear fault classification. Within the proposed method, we introduce a new concept of PEHD, which is defined as the distribution feature of the shaft frequency and its harmonics in envelope spectrum. A physics-informed parameter optimization model (PI-POM) is developed to minimize the difference between the simulation and measured signals in terms of PEHD, enabling the accurate identification of detailed parameters within the dynamic model. Subsequently, a TL guidance framework is established for the fine-tuning and adaptation of a Long Short-Term Memory-aided Different Kolmogorov-Arnold network (LSTM-DKAN), with the aim of improving classification accuracy. Validation on a constructed back-to-back gear test rig with induced crack and spalling faults demonstrates the PI-DG method's effectiveness in reducing physics-simulation discrepancies, exhibiting superior classification performance especially in few-shot cases.
{"title":"Physics-informed dual guidance method using physical envelope harmonic distribution and transfer learning for few-shot gear fault classification","authors":"Kun Yue , Liming Wang , Xiaoxi Ding , Wennian Yu , Zaigang Chen , Wenbin Huang","doi":"10.1016/j.engappai.2025.110956","DOIUrl":"10.1016/j.engappai.2025.110956","url":null,"abstract":"<div><div>Recent years have seen artificial intelligence algorithms gain considerable popularity in gear fault classification, yet their performance remains hindered by the scarcity of labeled fault data, often leading to suboptimal classification results. Several state-of-the-art studies have demonstrated that incorporating physical information can improve classification accuracy. However, the differences between simulated and measured signals pose a significant challenge in enhancing the performance of physics-informed methods. In order to fill this gap, this paper introduces a novel physics-informed dual guidance (PI-DG) method using physical envelope harmonic distribution (PEHD) and transfer learning (TL) for few-shot gear fault classification. Within the proposed method, we introduce a new concept of PEHD, which is defined as the distribution feature of the shaft frequency and its harmonics in envelope spectrum. A physics-informed parameter optimization model (PI-POM) is developed to minimize the difference between the simulation and measured signals in terms of PEHD, enabling the accurate identification of detailed parameters within the dynamic model. Subsequently, a TL guidance framework is established for the fine-tuning and adaptation of a Long Short-Term Memory-aided Different Kolmogorov-Arnold network (LSTM-DKAN), with the aim of improving classification accuracy. Validation on a constructed back-to-back gear test rig with induced crack and spalling faults demonstrates the PI-DG method's effectiveness in reducing physics-simulation discrepancies, exhibiting superior classification performance especially in few-shot cases.</div></div>","PeriodicalId":50523,"journal":{"name":"Engineering Applications of Artificial Intelligence","volume":"153 ","pages":"Article 110956"},"PeriodicalIF":7.5,"publicationDate":"2025-04-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143864253","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Khalid Zaman, Gan Zengkang, Sun Zhaoyun, Sayyed Mudassar Shah, Waqar Riaz, Jiancheng (Charles) Ji, Tariq Hussain, Razaz Waheeb Attar
Human emotion recognition (HER) has rapidly advanced, with applications in intelligent customer service, adaptive system training, human–robot interaction (HRI), and mental health monitoring. HER’s primary goal is to accurately recognize and classify emotions from digital inputs. Emotion recognition (ER) and feature extraction have long been core elements of HER, with deep neural networks (DNNs), particularly convolutional neural networks (CNNs), playing a critical role due to their superior visual feature extraction capabilities. This study proposes improving HER by integrating EfficientNet with transfer learning (TL) to train CNNs. Initially, an efficient R-CNN accurately recognizes faces in online and offline videos. The ensemble classification model is trained by combining features from four CNN models using feature pooling. The novel VGG-19 block is used to enhance the Faster R-CNN learning block, boosting face recognition efficiency and accuracy. The model benefits from fully connected mean pooling, dense pooling, and global dropout layers, solving the evanescent gradient issue. Tested on CK+, FER-2013, and the custom novel HER dataset (HERD), the approach shows significant accuracy improvements, reaching 89.23% (CK+), 94.36% (FER-2013), and 97.01% (HERD), proving its robustness and effectiveness.
{"title":"A Novel Emotion Recognition System for Human–Robot Interaction (HRI) Using Deep Ensemble Classification","authors":"Khalid Zaman, Gan Zengkang, Sun Zhaoyun, Sayyed Mudassar Shah, Waqar Riaz, Jiancheng (Charles) Ji, Tariq Hussain, Razaz Waheeb Attar","doi":"10.1155/int/6611276","DOIUrl":"https://doi.org/10.1155/int/6611276","url":null,"abstract":"<div>\u0000 <p>Human emotion recognition (HER) has rapidly advanced, with applications in intelligent customer service, adaptive system training, human–robot interaction (HRI), and mental health monitoring. HER’s primary goal is to accurately recognize and classify emotions from digital inputs. Emotion recognition (ER) and feature extraction have long been core elements of HER, with deep neural networks (DNNs), particularly convolutional neural networks (CNNs), playing a critical role due to their superior visual feature extraction capabilities. This study proposes improving HER by integrating EfficientNet with transfer learning (TL) to train CNNs. Initially, an efficient R-CNN accurately recognizes faces in online and offline videos. The ensemble classification model is trained by combining features from four CNN models using feature pooling. The novel VGG-19 block is used to enhance the Faster R-CNN learning block, boosting face recognition efficiency and accuracy. The model benefits from fully connected mean pooling, dense pooling, and global dropout layers, solving the evanescent gradient issue. Tested on CK+, FER-2013, and the custom novel HER dataset (HERD), the approach shows significant accuracy improvements, reaching 89.23% (CK+), 94.36% (FER-2013), and 97.01% (HERD), proving its robustness and effectiveness.</p>\u0000 </div>","PeriodicalId":14089,"journal":{"name":"International Journal of Intelligent Systems","volume":"2025 1","pages":""},"PeriodicalIF":5.0,"publicationDate":"2025-04-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1155/int/6611276","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143865625","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Sunil Prajapat, Mohammad S. Obaidat, Vivek Bharmaik, Garima Thakur, Pankaj Kumar
As quantum technology advances, classical digital signatures exhibit vulnerabilities in preserving security properties during the transmission of information. Working toward a reliable communication protocol, we introduce a proxy blind signature scheme to teleport a single particle qubit state with a message to the receiver, employing a three qubit GHZ entangled state. The blindness property is utilized to secure the message information from the proxy signer. A trusted party, Trent, is introduced to supervise the communication process. Alice blinds the original message and sends the Bell measurements with her entangled particle to proxy signer Charlie. After receiving measurements from Alice and Charlie, Bob verifies the proxy blind signature and performs appropriate unitary operations on his particle. Thereafter, Trent verifies the security of the quantum teleportation setup by matching the output data with the original data sent by Alice. Security analysis results prove that the proposed scheme fulfils the basic security necessities, including undeniability, unforgeability, blindness, verifiability, and traceability.
{"title":"Quantum Safe Proxy Blind Signature Protocol Based on 3D Entangled GHZ-Type States","authors":"Sunil Prajapat, Mohammad S. Obaidat, Vivek Bharmaik, Garima Thakur, Pankaj Kumar","doi":"10.1002/ett.70140","DOIUrl":"https://doi.org/10.1002/ett.70140","url":null,"abstract":"<div>\u0000 \u0000 <p>As quantum technology advances, classical digital signatures exhibit vulnerabilities in preserving security properties during the transmission of information. Working toward a reliable communication protocol, we introduce a proxy blind signature scheme to teleport a single particle qubit state with a message to the receiver, employing a three qubit GHZ entangled state. The blindness property is utilized to secure the message information from the proxy signer. A trusted party, Trent, is introduced to supervise the communication process. Alice blinds the original message and sends the Bell measurements with her entangled particle to proxy signer Charlie. After receiving measurements from Alice and Charlie, Bob verifies the proxy blind signature and performs appropriate unitary operations on his particle. Thereafter, Trent verifies the security of the quantum teleportation setup by matching the output data with the original data sent by Alice. Security analysis results prove that the proposed scheme fulfils the basic security necessities, including undeniability, unforgeability, blindness, verifiability, and traceability.</p>\u0000 </div>","PeriodicalId":23282,"journal":{"name":"Transactions on Emerging Telecommunications Technologies","volume":"36 5","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-04-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143865801","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}