Transfer learning models have proven superior to classical machine learning approaches in various text classification tasks, such as sentiment analysis, question answering, news categorization, and natural language inference. Recently, these models have shown exceptional results in natural language understanding (NLU). Advanced attention-based language models like BERT and XLNet excel at handling complex tasks across diverse contexts. However, they encounter difficulties when applied to specific domains. Platforms like Facebook, characterized by continually evolving casual and sophisticated language, demand meticulous context analysis even from human users. The literature has proposed numerous solutions using statistical and machine learning techniques to predict the sentiment (positive or negative) of online customer reviews, but most of them rely on various business, review, and reviewer features, which leads to generalizability issues. Furthermore, there have been very few studies investigating the effectiveness of state-of-the-art pre-trained language models for sentiment classification in reviews. Therefore, this study aims to assess the effectiveness of BERT, RoBERTa, ALBERT, DistilBERT, and XLNet in sentiment classification using the Yelp reviews dataset. The models were fine-tuned, and the results obtained with the same hyperparameters are as follows: 98.30 for RoBERTa, 98.20 for XLNet, 97.40 for BERT, 97.20 for ALBERT, and 96.00 for DistilBERT.
{"title":"Exploring transformer models for sentiment classification: A comparison of BERT, RoBERTa, ALBERT, DistilBERT, and XLNet","authors":"Ali Areshey, Hassan Mathkour","doi":"10.1111/exsy.13701","DOIUrl":"10.1111/exsy.13701","url":null,"abstract":"<p>Transfer learning models have proven superior to classical machine learning approaches in various text classification tasks, such as sentiment analysis, question answering, news categorization, and natural language inference. Recently, these models have shown exceptional results in natural language understanding (NLU). Advanced attention-based language models like BERT and XLNet excel at handling complex tasks across diverse contexts. However, they encounter difficulties when applied to specific domains. Platforms like Facebook, characterized by continually evolving casual and sophisticated language, demand meticulous context analysis even from human users. The literature has proposed numerous solutions using statistical and machine learning techniques to predict the sentiment (positive or negative) of online customer reviews, but most of them rely on various business, review, and reviewer features, which leads to generalizability issues. Furthermore, there have been very few studies investigating the effectiveness of state-of-the-art pre-trained language models for sentiment classification in reviews. Therefore, this study aims to assess the effectiveness of BERT, RoBERTa, ALBERT, DistilBERT, and XLNet in sentiment classification using the Yelp reviews dataset. The models were fine-tuned, and the results obtained with the same hyperparameters are as follows: 98.30 for RoBERTa, 98.20 for XLNet, 97.40 for BERT, 97.20 for ALBERT, and 96.00 for DistilBERT.</p>","PeriodicalId":51053,"journal":{"name":"Expert Systems","volume":"41 11","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-08-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142227719","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Smart city networks are critical for delivering essential services such as healthcare, education, and business operations. However, these networks are highly susceptible to a range of threats, including natural disasters and intentional cyberattacks, which can severely disrupt their functionality. To address these vulnerabilities, we present the resilient smart city (ResiSC) system, designed to enhance the resilience of smart city communication networks through a topological design approach. Our system employs a graph-theoretic algorithm to determine the optimal network topology for a given set of nodes, aiming to maximize connectivity while minimizing link provisioning costs. We introduce two novel connectivity measurements, All Nodes Reachability (ANR) and Sum of All Nodes Reachability (SANR), to evaluate network resilience. We applied our approach to data from two public universities of different sizes, simulating various attack scenarios to assess the robustness of the resulting network topologies. Evaluation results indicate that our solution improves network resilience against targeted attacks by 38% compared to baseline methods such as k-nearest neighbours (k-NN) graphs, while also reducing the number of additional links and their associated costs. Results also indicate that our proposed solution outperforms baseline methods like k-NN in terms of network resilience against targeted attacks by 41%. This work provides a practical framework for developing robust smart city networks capable of withstanding diverse threats.
{"title":"ResiSC: A system for building resilient smart city communication networks","authors":"Mohammed J. F. Alenazi","doi":"10.1111/exsy.13698","DOIUrl":"10.1111/exsy.13698","url":null,"abstract":"<p>Smart city networks are critical for delivering essential services such as healthcare, education, and business operations. However, these networks are highly susceptible to a range of threats, including natural disasters and intentional cyberattacks, which can severely disrupt their functionality. To address these vulnerabilities, we present the resilient smart city (ResiSC) system, designed to enhance the resilience of smart city communication networks through a topological design approach. Our system employs a graph-theoretic algorithm to determine the optimal network topology for a given set of nodes, aiming to maximize connectivity while minimizing link provisioning costs. We introduce two novel connectivity measurements, All Nodes Reachability (ANR) and Sum of All Nodes Reachability (SANR), to evaluate network resilience. We applied our approach to data from two public universities of different sizes, simulating various attack scenarios to assess the robustness of the resulting network topologies. Evaluation results indicate that our solution improves network resilience against targeted attacks by 38% compared to baseline methods such as k-nearest neighbours (k-NN) graphs, while also reducing the number of additional links and their associated costs. Results also indicate that our proposed solution outperforms baseline methods like k-NN in terms of network resilience against targeted attacks by 41%. This work provides a practical framework for developing robust smart city networks capable of withstanding diverse threats.</p>","PeriodicalId":51053,"journal":{"name":"Expert Systems","volume":"41 11","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-08-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141927028","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Fakhri Alam Khan, Abdullah Khan, Muhammad Imran, Awais Ahmad, Gwanggil Jeon
Alzheimer's disease (AD) is a major global health concern that affects millions of people globally. This study investigates the technical challenges in AD analysis and provides a thorough analysis of AD, emphasizing the disease's worldwide effects as well as the predicted increase. It explores the technological difficulties associated with AD analysis, concentrating on the shift in automated clinical diagnosis using MRI data from conventional machine learning to deep learning techniques. This study advances our knowledge of the effects of AD and provides new developments in deep learning for precise diagnosis, providing insightful information for both clinical and future research. The research introduces an innovative deep learning model, leveraging YOLOv5 and variants of YOLOv8, to classify AD images into four (NC, EMCI, LMCI, AD) categories. This study evaluates the performance of YOLOv5 which achieved high accuracy (97%) in multi‐class classification (classes 0 to 3) with precision, recall, and F1‐score reported for each class. YOLOv8 (Small) and YOLOv8 (Medium) models are also assessed for Alzheimer's disease diagnosis, demonstrating accuracy of 97% and 98%, respectively. Precision, recall, and F1‐score metrics provide detailed insights into the models' effectiveness across different classes. Comparative analysis against a transfer learning model reveals YOLOv5, YOLOv8 (Small), and YOLOv8 (Medium) consistently outperforming across six binary classifications related to cognitive impairment. These models show improved sensitivity and accuracy compared to baseline architectures from [32]. In AD/NC classification, YOLOv8 (Medium) achieves 98.43% accuracy and 97.45% sensitivity, for EMCI/LMCI classification, YOLOv8 (Medium) also excels with 92.12% accuracy and 90.12% sensitivity. The results highlight the effectiveness of YOLOv5 and YOLOv8 variants in neuroimaging tasks, showcasing their potential in clinical applications for cognitive impairment classification. The proposed models showcase superior performance, achieving high accuracy, sensitivity, and F1‐scores, surpassing baseline architectures and previous methods. Comparative analyses highlight the robustness and effectiveness of the proposed models in AD classification tasks, providing valuable insights for future research and clinical applications.
{"title":"Advancements in deep learning for Alzheimer's disease diagnosis: A comprehensive exploration and critical analysis of neuroimaging approaches","authors":"Fakhri Alam Khan, Abdullah Khan, Muhammad Imran, Awais Ahmad, Gwanggil Jeon","doi":"10.1111/exsy.13688","DOIUrl":"https://doi.org/10.1111/exsy.13688","url":null,"abstract":"Alzheimer's disease (AD) is a major global health concern that affects millions of people globally. This study investigates the technical challenges in AD analysis and provides a thorough analysis of AD, emphasizing the disease's worldwide effects as well as the predicted increase. It explores the technological difficulties associated with AD analysis, concentrating on the shift in automated clinical diagnosis using MRI data from conventional machine learning to deep learning techniques. This study advances our knowledge of the effects of AD and provides new developments in deep learning for precise diagnosis, providing insightful information for both clinical and future research. The research introduces an innovative deep learning model, leveraging YOLOv5 and variants of YOLOv8, to classify AD images into four (NC, EMCI, LMCI, AD) categories. This study evaluates the performance of YOLOv5 which achieved high accuracy (97%) in multi‐class classification (classes 0 to 3) with precision, recall, and F1‐score reported for each class. YOLOv8 (Small) and YOLOv8 (Medium) models are also assessed for Alzheimer's disease diagnosis, demonstrating accuracy of 97% and 98%, respectively. Precision, recall, and F1‐score metrics provide detailed insights into the models' effectiveness across different classes. Comparative analysis against a transfer learning model reveals YOLOv5, YOLOv8 (Small), and YOLOv8 (Medium) consistently outperforming across six binary classifications related to cognitive impairment. These models show improved sensitivity and accuracy compared to baseline architectures from [32]. In AD/NC classification, YOLOv8 (Medium) achieves 98.43% accuracy and 97.45% sensitivity, for EMCI/LMCI classification, YOLOv8 (Medium) also excels with 92.12% accuracy and 90.12% sensitivity. The results highlight the effectiveness of YOLOv5 and YOLOv8 variants in neuroimaging tasks, showcasing their potential in clinical applications for cognitive impairment classification. The proposed models showcase superior performance, achieving high accuracy, sensitivity, and F1‐scores, surpassing baseline architectures and previous methods. Comparative analyses highlight the robustness and effectiveness of the proposed models in AD classification tasks, providing valuable insights for future research and clinical applications.","PeriodicalId":51053,"journal":{"name":"Expert Systems","volume":"55 1","pages":""},"PeriodicalIF":3.3,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141886302","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Generative artificial intelligence (AI) has developed as an effective tool for time series predicting, revolutionizing the typical methods of prediction. Different classical approaches that depend on existing approaches and assumptions, generative AI controls advanced deep learning (DL) approaches like generative adversarial networks (GANs) and recurrent neural networks (RNNs), to identify designs and connections in time series data. DL has accomplished major success in optimizing performances connected with AI. In the financial area, it can be extremely utilized for the stock market predictive, trade implementation approaches, and set of optimizers. Stock market predictive is the most important use case in this field. GANs with advanced AI approaches have become more significant in recent times. However, it can be utilized in image‐image‐translation and other computer vision (CV) conditions. GANs could not utilized greatly for stock market prediction because of their effort to establish the proper set of hyperparameters. This study develops an integrated spotted hyena optimization algorithm with generative artificial intelligence for time series forecasting (SHOAGAI‐TSF) technique. The purpose of the SHOAGAI‐TSF technique is to accomplish a forecasting process for the utilization of stock price prediction. The SHOAGAI‐TSF technique uses probabilistic forecasting with a conditional GAN (CGAN) approach for the prediction of stock prices. The CGAN model learns the data generation distribution and determines the probabilistic prediction from it. To boost the prediction results of the CGAN approach, the hyperparameter tuning can be performed by the use of the SHOA. The simulation result analysis of the SHOAGAI‐TSF technique takes place on the stock market dataset. The experimental outcomes determine the significant solution of the SHOAGAI‐TSF algorithm with other compared methods in terms of distinct metrics.
{"title":"Integrating spotted hyena optimization technique with generative artificial intelligence for time series forecasting","authors":"Reda Salama","doi":"10.1111/exsy.13681","DOIUrl":"https://doi.org/10.1111/exsy.13681","url":null,"abstract":"Generative artificial intelligence (AI) has developed as an effective tool for time series predicting, revolutionizing the typical methods of prediction. Different classical approaches that depend on existing approaches and assumptions, generative AI controls advanced deep learning (DL) approaches like generative adversarial networks (GANs) and recurrent neural networks (RNNs), to identify designs and connections in time series data. DL has accomplished major success in optimizing performances connected with AI. In the financial area, it can be extremely utilized for the stock market predictive, trade implementation approaches, and set of optimizers. Stock market predictive is the most important use case in this field. GANs with advanced AI approaches have become more significant in recent times. However, it can be utilized in image‐image‐translation and other computer vision (CV) conditions. GANs could not utilized greatly for stock market prediction because of their effort to establish the proper set of hyperparameters. This study develops an integrated spotted hyena optimization algorithm with generative artificial intelligence for time series forecasting (SHOAGAI‐TSF) technique. The purpose of the SHOAGAI‐TSF technique is to accomplish a forecasting process for the utilization of stock price prediction. The SHOAGAI‐TSF technique uses probabilistic forecasting with a conditional GAN (CGAN) approach for the prediction of stock prices. The CGAN model learns the data generation distribution and determines the probabilistic prediction from it. To boost the prediction results of the CGAN approach, the hyperparameter tuning can be performed by the use of the SHOA. The simulation result analysis of the SHOAGAI‐TSF technique takes place on the stock market dataset. The experimental outcomes determine the significant solution of the SHOAGAI‐TSF algorithm with other compared methods in terms of distinct metrics.","PeriodicalId":51053,"journal":{"name":"Expert Systems","volume":"52 1","pages":""},"PeriodicalIF":3.3,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141886400","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
The infrared and visible fusion technology holds a pivotal position in smart city for cloud and fog computing, particularly in security system. By fusing infrared and visible image information, this technology enhances target identification, tracking and monitoring precision, bolstering overall system security. However, existing deep learning‐based methods rely heavily on convolutional operations, which excel at extracting local features but have limited receptive fields, hampering global information capture. To overcome this difficulty, we introduce GRDATFusion, a novel end‐to‐end network comprising three key modules: transformer, gradient residual dense and attention residual. The gradient residual dense module extracts local complementary features, leveraging a dense‐shaped network to retain potentially lost information. The attention residual module focuses on crucial input image details, while the transformer module captures global information and models long‐range dependencies. Experiments on public datasets show that GRDATFusion outperforms state‐of‐the‐art algorithms in qualitative and quantitative assessments. Ablation studies validate our approach's advantages, and efficiency comparisons demonstrate its computational efficiency. Therefore, our method makes the security systems in smart city with shorter delay and satisfies the real‐time requirement.
{"title":"GRDATFusion: A gradient residual dense and attention transformer infrared and visible image fusion network for smart city security systems in cloud and fog computing","authors":"Jian Zheng, Seunggil Jeon, Xiaomin Yang","doi":"10.1111/exsy.13685","DOIUrl":"https://doi.org/10.1111/exsy.13685","url":null,"abstract":"The infrared and visible fusion technology holds a pivotal position in smart city for cloud and fog computing, particularly in security system. By fusing infrared and visible image information, this technology enhances target identification, tracking and monitoring precision, bolstering overall system security. However, existing deep learning‐based methods rely heavily on convolutional operations, which excel at extracting local features but have limited receptive fields, hampering global information capture. To overcome this difficulty, we introduce GRDATFusion, a novel end‐to‐end network comprising three key modules: transformer, gradient residual dense and attention residual. The gradient residual dense module extracts local complementary features, leveraging a dense‐shaped network to retain potentially lost information. The attention residual module focuses on crucial input image details, while the transformer module captures global information and models long‐range dependencies. Experiments on public datasets show that GRDATFusion outperforms state‐of‐the‐art algorithms in qualitative and quantitative assessments. Ablation studies validate our approach's advantages, and efficiency comparisons demonstrate its computational efficiency. Therefore, our method makes the security systems in smart city with shorter delay and satisfies the real‐time requirement.","PeriodicalId":51053,"journal":{"name":"Expert Systems","volume":"44 1","pages":""},"PeriodicalIF":3.3,"publicationDate":"2024-07-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141871075","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Unexpected faults in rotating machinery can lead to cascading disruptions of the entire work process, emphasizing the importance of early detection of performance degradation and identification of the current state. To accurately assess the health of a machine, this study introduces an FFT-based raw vibration data preprocessing and graph representation technique, which analyses changes in frequency bands to detect early degradation trends in vibration data that may appear normal. The approach proposes a methodology that utilizes a graph convolutional autoencoder trained using only normal data to extract health indicators using the differences in the vectors as degradation progresses. This approach has the advantage of using only normal data to detect subtle performance degradation early and effectively represent health indicators accordingly.
{"title":"Health indicator construction based on normal states through FFT-graph embedding","authors":"GwanPil Kim, Jason J. Jung, David Camacho","doi":"10.1111/exsy.13689","DOIUrl":"10.1111/exsy.13689","url":null,"abstract":"<p>Unexpected faults in rotating machinery can lead to cascading disruptions of the entire work process, emphasizing the importance of early detection of performance degradation and identification of the current state. To accurately assess the health of a machine, this study introduces an FFT-based raw vibration data preprocessing and graph representation technique, which analyses changes in frequency bands to detect early degradation trends in vibration data that may appear normal. The approach proposes a methodology that utilizes a graph convolutional autoencoder trained using only normal data to extract health indicators using the differences in the vectors as degradation progresses. This approach has the advantage of using only normal data to detect subtle performance degradation early and effectively represent health indicators accordingly.</p>","PeriodicalId":51053,"journal":{"name":"Expert Systems","volume":"41 11","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-07-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1111/exsy.13689","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141871074","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Class imbalance and class overlap create difficulties in the training phase of the standard machine learning algorithm. Its performance is not well in minority classes, especially when there is a high class imbalance and significant class overlap. Recently it has been observed by researchers that, the joint effects of class overlap and imbalance are more harmful as compared to their direct impact. To handle these problems, many methods have been proposed by researchers in past years that can be broadly categorized as data-level, algorithm-level, ensemble learning, and hybrid methods. Existing data-level methods often suffer from problems like information loss and overfitting. To overcome these problems, we introduce a novel entropy-based hybrid sampling (EHS) method to handle class overlap in highly imbalanced datasets. The EHS eliminates less informative majority instances from the overlap region during the undersampling phase and regenerates high informative synthetic minority instances in the oversampling phase near the borderline. The proposed EHS achieved significant improvement in F1-score, G-mean, and AUC performance metrics value by DT, NB, and SVM classifiers as compared to well-established state-of-the-art methods. Classifiers performances are tested on 28 datasets with extreme ranges in imbalance and overlap.
{"title":"Entropy-based hybrid sampling (EHS) method to handle class overlap in highly imbalanced dataset","authors":"Anil Kumar, Dinesh Singh, Rama Shankar Yadav","doi":"10.1111/exsy.13679","DOIUrl":"10.1111/exsy.13679","url":null,"abstract":"<p>Class imbalance and class overlap create difficulties in the training phase of the standard machine learning algorithm. Its performance is not well in minority classes, especially when there is a high class imbalance and significant class overlap. Recently it has been observed by researchers that, the joint effects of class overlap and imbalance are more harmful as compared to their direct impact. To handle these problems, many methods have been proposed by researchers in past years that can be broadly categorized as data-level, algorithm-level, ensemble learning, and hybrid methods. Existing data-level methods often suffer from problems like information loss and overfitting. To overcome these problems, we introduce a novel entropy-based hybrid sampling (EHS) method to handle class overlap in highly imbalanced datasets. The EHS eliminates less informative majority instances from the overlap region during the undersampling phase and regenerates high informative synthetic minority instances in the oversampling phase near the borderline. The proposed EHS achieved significant improvement in F1-score, G-mean, and AUC performance metrics value by DT, NB, and SVM classifiers as compared to well-established state-of-the-art methods. Classifiers performances are tested on 28 datasets with extreme ranges in imbalance and overlap.</p>","PeriodicalId":51053,"journal":{"name":"Expert Systems","volume":"41 11","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-07-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141871073","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Human Activity Recognition (HAR) is a highly promising research area meant to automatically identify and interpret human behaviour using data received from sensors in various contexts. The potential uses of HAR are many, among them health care, sports coaching or monitoring the elderly or disabled. Nonetheless, there are numerous hurdles to be circumvented for HAR's precision and usefulness to be improved. One of the challenges is that there is no uniformity in data collection and annotation making it difficult to compare findings among different studies. Furthermore, more comprehensive datasets are necessary so as to include a wider range of human activities in different contexts while complex activities, which consist of multiple sub-activities, are still a challenge for recognition systems. Researchers have proposed new frontiers such as multi-modal sensor data fusion and deep learning approaches for enhancing HAR accuracy while addressing these issues. Also, we are seeing more non-traditional applications such as robotics and virtual reality/augmented world going forward with their use cases of HAR. This article offers an extensive review on the recent advances in HAR and highlights the major challenges facing this field as well as future opportunities for further researches.
人类活动识别(HAR)是一个极具发展前景的研究领域,其目的是在各种情况下利用传感器接收到的数据自动识别和解释人类行为。人类活动识别(HAR)的潜在用途很多,其中包括医疗保健、体育指导或监测老年人或残疾人。然而,要提高 HAR 的精确度和实用性,还有许多障碍需要克服。挑战之一是数据收集和注释不统一,因此很难比较不同研究的结果。此外,还需要更全面的数据集,以包括不同环境中更广泛的人类活动,而由多个子活动组成的复杂活动对识别系统来说仍是一个挑战。研究人员提出了一些新的前沿技术,如多模态传感器数据融合和深度学习方法,以便在解决这些问题的同时提高 HAR 的准确性。此外,我们还看到更多的非传统应用,如机器人和虚拟现实/增强世界,都在使用 HAR。本文对 HAR 的最新进展进行了广泛综述,并重点介绍了该领域面临的主要挑战以及未来进一步研究的机遇。
{"title":"Human activity recognition: A comprehensive review","authors":"Harmandeep Kaur, Veenu Rani, Munish Kumar","doi":"10.1111/exsy.13680","DOIUrl":"10.1111/exsy.13680","url":null,"abstract":"<p>Human Activity Recognition (HAR) is a highly promising research area meant to automatically identify and interpret human behaviour using data received from sensors in various contexts. The potential uses of HAR are many, among them health care, sports coaching or monitoring the elderly or disabled. Nonetheless, there are numerous hurdles to be circumvented for HAR's precision and usefulness to be improved. One of the challenges is that there is no uniformity in data collection and annotation making it difficult to compare findings among different studies. Furthermore, more comprehensive datasets are necessary so as to include a wider range of human activities in different contexts while complex activities, which consist of multiple sub-activities, are still a challenge for recognition systems. Researchers have proposed new frontiers such as multi-modal sensor data fusion and deep learning approaches for enhancing HAR accuracy while addressing these issues. Also, we are seeing more non-traditional applications such as robotics and virtual reality/augmented world going forward with their use cases of HAR. This article offers an extensive review on the recent advances in HAR and highlights the major challenges facing this field as well as future opportunities for further researches.</p>","PeriodicalId":51053,"journal":{"name":"Expert Systems","volume":"41 11","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-07-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141783662","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Afonso Baldo, Paulo J. S. Ferreira, João Mendes‐Moreira
With technological advancements, much data is being captured by sensors, smartphones, wearable devices, and so forth. These vast datasets are stored in data centres and utilized to forge data‐driven models for the condition monitoring of infrastructures and systems through future data mining tasks. However, these datasets often surpass the processing capabilities of traditional information systems and methodologies due to their significant size. Additionally, not all samples within these datasets contribute valuable information during the model training phase, leading to inefficiencies. The processing and training of Machine Learning algorithms become time‐consuming, and storing all the data demands excessive space, contributing to the Big Data challenge. In this paper, we propose two novel techniques to reduce large time‐series datasets into more compact versions without undermining the predictive performance of the resulting models. These methods also aim to decrease the time required for training the models and the storage space needed for the condensed datasets. We evaluated our techniques on five public datasets, employing three Machine Learning algorithms: Holt‐Winters, SARIMA, and LSTM. The outcomes indicate that for most of the datasets examined, our techniques maintain, and in several instances enhance, the forecasting accuracy of the models. Moreover, we significantly reduced the time required to train the Machine Learning algorithms employed.
{"title":"Sampling approaches to reduce very frequent seasonal time series","authors":"Afonso Baldo, Paulo J. S. Ferreira, João Mendes‐Moreira","doi":"10.1111/exsy.13690","DOIUrl":"https://doi.org/10.1111/exsy.13690","url":null,"abstract":"With technological advancements, much data is being captured by sensors, smartphones, wearable devices, and so forth. These vast datasets are stored in data centres and utilized to forge data‐driven models for the condition monitoring of infrastructures and systems through future data mining tasks. However, these datasets often surpass the processing capabilities of traditional information systems and methodologies due to their significant size. Additionally, not all samples within these datasets contribute valuable information during the model training phase, leading to inefficiencies. The processing and training of Machine Learning algorithms become time‐consuming, and storing all the data demands excessive space, contributing to the Big Data challenge. In this paper, we propose two novel techniques to reduce large time‐series datasets into more compact versions without undermining the predictive performance of the resulting models. These methods also aim to decrease the time required for training the models and the storage space needed for the condensed datasets. We evaluated our techniques on five public datasets, employing three Machine Learning algorithms: Holt‐Winters, SARIMA, and LSTM. The outcomes indicate that for most of the datasets examined, our techniques maintain, and in several instances enhance, the forecasting accuracy of the models. Moreover, we significantly reduced the time required to train the Machine Learning algorithms employed.","PeriodicalId":51053,"journal":{"name":"Expert Systems","volume":"168 1","pages":""},"PeriodicalIF":3.3,"publicationDate":"2024-07-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141783665","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Kyung‐yeul Kim, Young‐bo Yang, Mi‐ra Kim, Jihie Kim, Ji Su Park
When stress causes negative behaviours to emerge in our daily lives, it is important to intervene quickly and appropriately to control the negative problem behaviours. Questionnaires, a common method of information gathering, have the disadvantage that it is difficult to get the exact information needed due to defensive or insincere responses from subjects. As an alternative to these drawbacks, projective testing using pictures can provide the necessary information more accurately than questionnaires because the subject responds subconsciously and the direct experience expressed through pictures can be more accurate than questionnaires. Analysing hand‐drawn image data with the Wartegg Zeichen Test (WZT) is not easy. In this study, we used deep learning to analyse image data represented as pictures through WZT to predict early depression. We analyse the data of 54 people who were judged as early depression and 54 people without depression, and increase the number of people without depression to 100 and 500, and aim to study in unbalanced data. We use CNN and CNN‐SVM to analyse the drawing images of WZT's initial depression with deep learning and predict the outcome of depression. The results show that the initial depression is predicted with 92%–98% accuracy on the image data directly drawn by WZT. This is the first study to automatically analyse and predict early depression in WZT based on hand‐drawn image data using deep learning models. The extraction of features from WZT images by deep learning analysis is expected to create more research opportunities through the convergence of psychotherapy and Information and Communication Technology (ICT) technology, and is expected to have high growth potential.
{"title":"Predicting early depression in WZT drawing image based on deep learning","authors":"Kyung‐yeul Kim, Young‐bo Yang, Mi‐ra Kim, Jihie Kim, Ji Su Park","doi":"10.1111/exsy.13675","DOIUrl":"https://doi.org/10.1111/exsy.13675","url":null,"abstract":"When stress causes negative behaviours to emerge in our daily lives, it is important to intervene quickly and appropriately to control the negative problem behaviours. Questionnaires, a common method of information gathering, have the disadvantage that it is difficult to get the exact information needed due to defensive or insincere responses from subjects. As an alternative to these drawbacks, projective testing using pictures can provide the necessary information more accurately than questionnaires because the subject responds subconsciously and the direct experience expressed through pictures can be more accurate than questionnaires. Analysing hand‐drawn image data with the Wartegg Zeichen Test (WZT) is not easy. In this study, we used deep learning to analyse image data represented as pictures through WZT to predict early depression. We analyse the data of 54 people who were judged as early depression and 54 people without depression, and increase the number of people without depression to 100 and 500, and aim to study in unbalanced data. We use CNN and CNN‐SVM to analyse the drawing images of WZT's initial depression with deep learning and predict the outcome of depression. The results show that the initial depression is predicted with 92%–98% accuracy on the image data directly drawn by WZT. This is the first study to automatically analyse and predict early depression in WZT based on hand‐drawn image data using deep learning models. The extraction of features from WZT images by deep learning analysis is expected to create more research opportunities through the convergence of psychotherapy and Information and Communication Technology (ICT) technology, and is expected to have high growth potential.","PeriodicalId":51053,"journal":{"name":"Expert Systems","volume":"108 1","pages":""},"PeriodicalIF":3.3,"publicationDate":"2024-07-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141783664","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}