Bin Kong, Jing Qian, Pinhao Song, Jing Yang, Amir Hussain
Underwater images are often with biased colours and reduced contrast because of the absorption and scattering effects when light propagates in water. Such images with degradation cannot meet the needs of underwater operations. The main problem in classic underwater image restoration or enhancement methods is that they consume long calculation time, and often, the colour or contrast of the result images is still unsatisfied. Instead of using the complicated physical model of underwater imaging degradation, we propose a new method to deal with underwater images by imitating the colour constancy mechanism of human vision using double-opponency. Firstly, the original image is converted to the LMS space. Then the signals are linearly combined, and Gaussian convolutions are performed to imitate the function of receptive fields (RFs). Next, two RFs with different sizes work together to constitute the double-opponency response. Finally, the underwater light is estimated to correct the colours in the image. Further contrast stretching on the luminance is optional. Experiments show that the proposed method can obtain clarified underwater images with higher quality than before, and it spends significantly less time cost compared to other previously published typical methods.
{"title":"Underwater image clarifying based on human visual colour constancy using double-opponency","authors":"Bin Kong, Jing Qian, Pinhao Song, Jing Yang, Amir Hussain","doi":"10.1049/cit2.12260","DOIUrl":"10.1049/cit2.12260","url":null,"abstract":"<p>Underwater images are often with biased colours and reduced contrast because of the absorption and scattering effects when light propagates in water. Such images with degradation cannot meet the needs of underwater operations. The main problem in classic underwater image restoration or enhancement methods is that they consume long calculation time, and often, the colour or contrast of the result images is still unsatisfied. Instead of using the complicated physical model of underwater imaging degradation, we propose a new method to deal with underwater images by imitating the colour constancy mechanism of human vision using double-opponency. Firstly, the original image is converted to the LMS space. Then the signals are linearly combined, and Gaussian convolutions are performed to imitate the function of receptive fields (RFs). Next, two RFs with different sizes work together to constitute the double-opponency response. Finally, the underwater light is estimated to correct the colours in the image. Further contrast stretching on the luminance is optional. Experiments show that the proposed method can obtain clarified underwater images with higher quality than before, and it spends significantly less time cost compared to other previously published typical methods.</p>","PeriodicalId":46211,"journal":{"name":"CAAI Transactions on Intelligence Technology","volume":"9 3","pages":"632-648"},"PeriodicalIF":5.1,"publicationDate":"2023-07-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1049/cit2.12260","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"75163031","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Tereza Kucova, Michal Prauzek, Jaromir Konecny, Darius Andriukaitis, Mindaugas Zilys, Radek Martinek
Initiatives to minimise battery use, address sustainability, and reduce regular maintenance have driven the challenge to use alternative power sources to supply energy to devices deployed in Internet of Things (IoT) networks. As a key pillar of fifth generation (5G) and beyond 5G networks,IoT is estimated to reach 42 billion devices by the year 2025. Thermoelectric generators (TEGs) are solid state energy harvesters which reliably and renewably convert thermal energy into electrical energy. These devices are able to recover lost thermal energy, produce energy in extreme environments, generate electric power in remote areas, and power micro-sensors. Applying the state of the art, the authorspresent a comprehensive review of machine learning (ML) approaches applied in combination with TEG-powered IoT devices to manage and predict available energy. The application areas of TEG-driven IoT devices that exploit as a heat source the temperature differences found in the environment, biological structures, machines, and other technologies are summarised. Based on detailed research of the state of the art in TEG-powered devices, the authors investigated the research challenges, applied algorithms and application areas of this technology. The aims of the research were to devise new energy prediction and energy management systems based on ML methods, create supervised algorithms which better estimate incoming energy, and develop unsupervised and semi-supervised approaches which provide adaptive and dynamic operation. The review results indicate that TEGs are a suitable energy harvesting technology for low-power applications through their scalability, usability in ubiquitous temperature difference scenarios, and long operating lifetime. However, TEGs also have low energy efficiency (around 10%) and require a relatively constant heat source.
{"title":"Thermoelectric energy harvesting for internet of things devices using machine learning: A review","authors":"Tereza Kucova, Michal Prauzek, Jaromir Konecny, Darius Andriukaitis, Mindaugas Zilys, Radek Martinek","doi":"10.1049/cit2.12259","DOIUrl":"https://doi.org/10.1049/cit2.12259","url":null,"abstract":"<p>Initiatives to minimise battery use, address sustainability, and reduce regular maintenance have driven the challenge to use alternative power sources to supply energy to devices deployed in Internet of Things (IoT) networks. As a key pillar of fifth generation (5G) and beyond 5G networks,IoT is estimated to reach 42 billion devices by the year 2025. Thermoelectric generators (TEGs) are solid state energy harvesters which reliably and renewably convert thermal energy into electrical energy. These devices are able to recover lost thermal energy, produce energy in extreme environments, generate electric power in remote areas, and power micro-sensors. Applying the state of the art, the authorspresent a comprehensive review of machine learning (ML) approaches applied in combination with TEG-powered IoT devices to manage and predict available energy. The application areas of TEG-driven IoT devices that exploit as a heat source the temperature differences found in the environment, biological structures, machines, and other technologies are summarised. Based on detailed research of the state of the art in TEG-powered devices, the authors investigated the research challenges, applied algorithms and application areas of this technology. The aims of the research were to devise new energy prediction and energy management systems based on ML methods, create supervised algorithms which better estimate incoming energy, and develop unsupervised and semi-supervised approaches which provide adaptive and dynamic operation. The review results indicate that TEGs are a suitable energy harvesting technology for low-power applications through their scalability, usability in ubiquitous temperature difference scenarios, and long operating lifetime. However, TEGs also have low energy efficiency (around 10%) and require a relatively constant heat source.</p>","PeriodicalId":46211,"journal":{"name":"CAAI Transactions on Intelligence Technology","volume":"8 3","pages":"680-700"},"PeriodicalIF":5.1,"publicationDate":"2023-07-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1049/cit2.12259","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"50130220","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Urgent care clinics and emergency departments around the world periodically suffer from extended wait times beyond patient expectations due to surges in patient flows. The delays arising from inadequate staffing levels during these periods have been linked with adverse clinical outcomes. Previous research into forecasting patient flows has mostly used statistical techniques. These studies have also predominately focussed on short-term forecasts, which have limited practicality for the resourcing of medical personnel. This study joins an emerging body of work which seeks to explore the potential of machine learning algorithms to generate accurate forecasts of patient presentations. Our research uses datasets covering 10 years from two large urgent care clinics to develop long-term patient flow forecasts up to one quarter ahead using a range of state-of-the-art algorithms. A distinctive feature of this study is the use of eXplainable Artificial Intelligence (XAI) tools like Shapely and LIME that enable an in-depth analysis of the behaviour of the models, which would otherwise be uninterpretable. These analysis tools enabled us to explore the ability of the models to adapt to the volatility in patient demand during the COVID-19 pandemic lockdowns and to identify the most impactful variables, resulting in valuable insights into their performance. The results showed that a novel combination of advanced univariate models like Prophet as well as gradient boosting, into an ensemble, delivered the most accurate and consistent solutions on average. This approach generated improvements in the range of 16%–30% over the existing in-house methods for estimating the daily patient flows 90 days ahead.
{"title":"Forecasting patient demand at urgent care clinics using explainable machine learning","authors":"Teo Susnjak, Paula Maddigan","doi":"10.1049/cit2.12258","DOIUrl":"https://doi.org/10.1049/cit2.12258","url":null,"abstract":"<p>Urgent care clinics and emergency departments around the world periodically suffer from extended wait times beyond patient expectations due to surges in patient flows. The delays arising from inadequate staffing levels during these periods have been linked with adverse clinical outcomes. Previous research into forecasting patient flows has mostly used statistical techniques. These studies have also predominately focussed on short-term forecasts, which have limited practicality for the resourcing of medical personnel. This study joins an emerging body of work which seeks to explore the potential of machine learning algorithms to generate accurate forecasts of patient presentations. Our research uses datasets covering 10 years from two large urgent care clinics to develop long-term patient flow forecasts up to one quarter ahead using a range of state-of-the-art algorithms. A distinctive feature of this study is the use of eXplainable Artificial Intelligence (XAI) tools like Shapely and LIME that enable an in-depth analysis of the behaviour of the models, which would otherwise be uninterpretable. These analysis tools enabled us to explore the ability of the models to adapt to the volatility in patient demand during the COVID-19 pandemic lockdowns and to identify the most impactful variables, resulting in valuable insights into their performance. The results showed that a novel combination of advanced univariate models like Prophet as well as gradient boosting, into an ensemble, delivered the most accurate and consistent solutions on average. This approach generated improvements in the range of 16%–30% over the existing in-house methods for estimating the daily patient flows 90 days ahead.</p>","PeriodicalId":46211,"journal":{"name":"CAAI Transactions on Intelligence Technology","volume":"8 3","pages":"712-733"},"PeriodicalIF":5.1,"publicationDate":"2023-07-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1049/cit2.12258","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"50149509","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Yining Hua, J. Sui, H. Fang, Chuanping Hu, Dewei Yi
{"title":"Domain‐adapted driving scene understanding with uncertainty‐aware and diversified generative adversarial networks","authors":"Yining Hua, J. Sui, H. Fang, Chuanping Hu, Dewei Yi","doi":"10.1049/cit2.12257","DOIUrl":"https://doi.org/10.1049/cit2.12257","url":null,"abstract":"","PeriodicalId":46211,"journal":{"name":"CAAI Transactions on Intelligence Technology","volume":"11 1","pages":""},"PeriodicalIF":5.1,"publicationDate":"2023-07-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"84508332","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Some types of software systems, like event-based and non-deterministic ones, are usually specified as rules so that we can analyse the system behaviour by drawing inferences from firing the rules. However, when the fuzzy rules are used for the specification of non-deterministic behaviour and they contain a large number of variables, they constitute a complex form that is difficult to understand and infer. A solution is to visualise the system specification with the capability of automatic rule inference. In this study, by representing a high-level system specification, the authors visualise rule representation and firing using fuzzy coloured Petri-nets. Already, several fuzzy Petri-nets-based methods have been presented, but they either do not support a large number of rules and variables or do not consider significant cases like (a) the weight of the premise's propositions in the occurrence of the rule conclusion, (b) the weight of conclusion's proposition, (c) threshold values for premise and conclusion's propositions of the rule, and (d) the certainty factor (CF) for the rule or the conclusion's proposition. By considering cases (a)–(d), a wider variety of fuzzy rules are supported. The authors applied their model to the analysis of attacks against a part of a real secure water treatment system. In another real experiment, the authors applied the model to the two scenarios from their previous work and analysed the results.
{"title":"Fuzzy coloured petri nets-based method to analyse and verify the functionality of software","authors":"Mina Chavoshi, Seyed Morteza Babamir","doi":"10.1049/cit2.12251","DOIUrl":"https://doi.org/10.1049/cit2.12251","url":null,"abstract":"<p>Some types of software systems, like event-based and non-deterministic ones, are usually specified as rules so that we can analyse the system behaviour by drawing inferences from firing the rules. However, when the fuzzy rules are used for the specification of non-deterministic behaviour and they contain a large number of variables, they constitute a complex form that is difficult to understand and infer. A solution is to visualise the system specification with the capability of automatic rule inference. In this study, by representing a high-level system specification, the authors visualise rule representation and firing using <i>fuzzy coloured Petri-nets</i>. Already, several fuzzy Petri-nets-based methods have been presented, but they either do not support a large number of rules and variables or do not consider significant cases like (a) the weight of the premise's propositions in the occurrence of the rule conclusion, (b) the weight of conclusion's proposition, (c) threshold values for premise and conclusion's propositions of the rule, and (d) the certainty factor (CF) for the rule or the conclusion's proposition. By considering cases (a)–(d), a wider variety of fuzzy rules are supported. The authors applied their model to the analysis of attacks against a part of a real secure water treatment system. In another real experiment, the authors applied the model to the two scenarios from their previous work and analysed the results.</p>","PeriodicalId":46211,"journal":{"name":"CAAI Transactions on Intelligence Technology","volume":"8 3","pages":"863-879"},"PeriodicalIF":5.1,"publicationDate":"2023-07-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1049/cit2.12251","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"50134503","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Ruidong Yan, Penghui Li, Hongbo Gao, Jin Huang, Chengbo Wang
Disturbance observer-based control method has achieved good results in the car-following scenario of intelligent and connected vehicle (ICV). However, the gain of conventional extended disturbance observer (EDO)-based control method is usually set manually rather than adjusted adaptively according to real time traffic conditions, thus declining the car-following performance. To solve this problem, a car-following strategy of ICV using EDO adjusted by reinforcement learning is proposed. Different from the conventional method, the gain of proposed strategy can be adjusted by reinforcement learning to improve its estimation accuracy. Since the “equivalent disturbance” can be compensated by EDO to a great extent, the disturbance rejection ability of the car-following method will be improved significantly. Both Lyapunov approach and numerical simulations are carried out to verify the effectiveness of the proposed method.
基于扰动观测器的控制方法在智能网联汽车(ICV)的汽车跟随场景中取得了良好的效果。然而,传统的基于扩展扰动观测器(EDO)的控制方法的增益通常是手动设置的,而不是根据实时交通状况进行自适应调节,从而降低了汽车跟随性能。为解决这一问题,本文提出了一种通过强化学习调整 EDO 的 ICV 汽车跟随策略。与传统方法不同的是,所提策略的增益可通过强化学习进行调整,以提高其估计精度。由于 EDO 可以在很大程度上补偿 "等效扰动",因此汽车跟随方法的扰动抑制能力将得到显著提高。为了验证所提方法的有效性,我们采用了李雅普诺夫方法并进行了数值模拟。
{"title":"Car-following strategy of intelligent connected vehicle using extended disturbance observer adjusted by reinforcement learning","authors":"Ruidong Yan, Penghui Li, Hongbo Gao, Jin Huang, Chengbo Wang","doi":"10.1049/cit2.12252","DOIUrl":"10.1049/cit2.12252","url":null,"abstract":"<p>Disturbance observer-based control method has achieved good results in the car-following scenario of intelligent and connected vehicle (ICV). However, the gain of conventional extended disturbance observer (EDO)-based control method is usually set manually rather than adjusted adaptively according to real time traffic conditions, thus declining the car-following performance. To solve this problem, a car-following strategy of ICV using EDO adjusted by reinforcement learning is proposed. Different from the conventional method, the gain of proposed strategy can be adjusted by reinforcement learning to improve its estimation accuracy. Since the “equivalent disturbance” can be compensated by EDO to a great extent, the disturbance rejection ability of the car-following method will be improved significantly. Both Lyapunov approach and numerical simulations are carried out to verify the effectiveness of the proposed method.</p>","PeriodicalId":46211,"journal":{"name":"CAAI Transactions on Intelligence Technology","volume":"9 2","pages":"365-373"},"PeriodicalIF":5.1,"publicationDate":"2023-07-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1049/cit2.12252","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"77610280","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
M. Rehman, Arslan Shafique, Imdad Ullah Khan, Y. Ghadi, Jawad Ahmad, Mohammed S. Alshehri, Mimonah Al Qathrady, Majed Alhaisoni, Muhammad H. Zayyan
{"title":"An efficient deep learning model for brain tumour detection with privacy preservation","authors":"M. Rehman, Arslan Shafique, Imdad Ullah Khan, Y. Ghadi, Jawad Ahmad, Mohammed S. Alshehri, Mimonah Al Qathrady, Majed Alhaisoni, Muhammad H. Zayyan","doi":"10.1049/cit2.12254","DOIUrl":"https://doi.org/10.1049/cit2.12254","url":null,"abstract":"","PeriodicalId":46211,"journal":{"name":"CAAI Transactions on Intelligence Technology","volume":"1 1","pages":""},"PeriodicalIF":5.1,"publicationDate":"2023-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"89674267","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Explainable human‐in‐the‐loop healthcare image information quality assessment and selection","authors":"Yang Li, S. Ercişli","doi":"10.1049/cit2.12253","DOIUrl":"https://doi.org/10.1049/cit2.12253","url":null,"abstract":"","PeriodicalId":46211,"journal":{"name":"CAAI Transactions on Intelligence Technology","volume":"1 1","pages":""},"PeriodicalIF":5.1,"publicationDate":"2023-06-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"80399807","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Sparse representation is an effective data classification algorithm that depends on the known training samples to categorise the test sample. It has been widely used in various image classification tasks. Sparseness in sparse representation means that only a few of instances selected from all training samples can effectively convey the essential class-specific information of the test sample, which is very important for classification. For deformable images such as human faces, pixels at the same location of different images of the same subject usually have different intensities. Therefore, extracting features and correctly classifying such deformable objects is very hard. Moreover, the lighting, attitude and occlusion cause more difficulty. Considering the problems and challenges listed above, a novel image representation and classification algorithm is proposed. First, the authors’ algorithm generates virtual samples by a non-linear variation method. This method can effectively extract the low-frequency information of space-domain features of the original image, which is very useful for representing deformable objects. The combination of the original and virtual samples is more beneficial to improve the classification performance and robustness of the algorithm. Thereby, the authors’ algorithm calculates the expression coefficients of the original and virtual samples separately using the sparse representation principle and obtains the final score by a designed efficient score fusion scheme. The weighting coefficients in the score fusion scheme are set entirely automatically. Finally, the algorithm classifies the samples based on the final scores. The experimental results show that our method performs better classification than conventional sparse representation algorithms.
{"title":"Sparse representation scheme with enhanced medium pixel intensity for face recognition","authors":"Xuexue Zhang, Yongjun Zhang, Zewei Wang, Wei Long, Weihao Gao, Bob Zhang","doi":"10.1049/cit2.12247","DOIUrl":"10.1049/cit2.12247","url":null,"abstract":"<p>Sparse representation is an effective data classification algorithm that depends on the known training samples to categorise the test sample. It has been widely used in various image classification tasks. Sparseness in sparse representation means that only a few of instances selected from all training samples can effectively convey the essential class-specific information of the test sample, which is very important for classification. For deformable images such as human faces, pixels at the same location of different images of the same subject usually have different intensities. Therefore, extracting features and correctly classifying such deformable objects is very hard. Moreover, the lighting, attitude and occlusion cause more difficulty. Considering the problems and challenges listed above, a novel image representation and classification algorithm is proposed. First, the authors’ algorithm generates virtual samples by a non-linear variation method. This method can effectively extract the low-frequency information of space-domain features of the original image, which is very useful for representing deformable objects. The combination of the original and virtual samples is more beneficial to improve the classification performance and robustness of the algorithm. Thereby, the authors’ algorithm calculates the expression coefficients of the original and virtual samples separately using the sparse representation principle and obtains the final score by a designed efficient score fusion scheme. The weighting coefficients in the score fusion scheme are set entirely automatically. Finally, the algorithm classifies the samples based on the final scores. The experimental results show that our method performs better classification than conventional sparse representation algorithms.</p>","PeriodicalId":46211,"journal":{"name":"CAAI Transactions on Intelligence Technology","volume":"9 1","pages":"116-127"},"PeriodicalIF":5.1,"publicationDate":"2023-06-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1049/cit2.12247","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"83427094","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Appropriately characterising the mixed space–time relations of the contagion process caused by hybrid space and time factors remains the primary challenge in COVID-19 forecasting. However, in previous deep learning models for epidemic forecasting, spatial and temporal variations are captured separately. A unified model is developed to cover all spatio–temporal relations. However, this measure is insufficient for modelling the complex spatio–temporal relations of infectious disease transmission. A dynamic adaptive spatio–temporal graph network (DASTGN) is proposed based on attention mechanisms to improve prediction accuracy. In DASTGN, complex spatio–temporal relations are depicted by adaptively fusing the mixed space–time effects and dynamic space–time dependency structure. This dual-scale model considers the time-specific, space-specific, and direct effects of the propagation process at the fine-grained level. Furthermore, the model characterises impacts from various space–time neighbour blocks under time-varying interventions at the coarse-grained level. The performance comparisons on the three COVID-19 datasets reveal that DASTGN achieves state-of-the-art results with a maximum improvement of 17.092% in the root mean-square error and 11.563% in the mean absolute error. Experimental results indicate that the mechanisms of designing DASTGN can effectively detect some spreading characteristics of COVID-19. The spatio–temporal weight matrices learned in each proposed module reveal diffusion patterns in various scenarios. In conclusion, DASTGN has successfully captured the dynamic spatio–temporal variations of COVID-19, and considering multiple dynamic space–time relationships is essential in epidemic forecasting.
{"title":"Dynamic adaptive spatio–temporal graph network for COVID-19 forecasting","authors":"Xiaojun Pu, Jiaqi Zhu, Yunkun Wu, Chang Leng, Zitong Bo, Hongan Wang","doi":"10.1049/cit2.12238","DOIUrl":"10.1049/cit2.12238","url":null,"abstract":"<p>Appropriately characterising the mixed space–time relations of the contagion process caused by hybrid space and time factors remains the primary challenge in COVID-19 forecasting. However, in previous deep learning models for epidemic forecasting, spatial and temporal variations are captured separately. A unified model is developed to cover all spatio–temporal relations. However, this measure is insufficient for modelling the complex spatio–temporal relations of infectious disease transmission. A dynamic adaptive spatio–temporal graph network (DASTGN) is proposed based on attention mechanisms to improve prediction accuracy. In DASTGN, complex spatio–temporal relations are depicted by adaptively fusing the mixed space–time effects and dynamic space–time dependency structure. This dual-scale model considers the time-specific, space-specific, and direct effects of the propagation process at the fine-grained level. Furthermore, the model characterises impacts from various space–time neighbour blocks under time-varying interventions at the coarse-grained level. The performance comparisons on the three COVID-19 datasets reveal that DASTGN achieves state-of-the-art results with a maximum improvement of 17.092% in the root mean-square error and 11.563% in the mean absolute error. Experimental results indicate that the mechanisms of designing DASTGN can effectively detect some spreading characteristics of COVID-19. The spatio–temporal weight matrices learned in each proposed module reveal diffusion patterns in various scenarios. In conclusion, DASTGN has successfully captured the dynamic spatio–temporal variations of COVID-19, and considering multiple dynamic space–time relationships is essential in epidemic forecasting.</p>","PeriodicalId":46211,"journal":{"name":"CAAI Transactions on Intelligence Technology","volume":"9 3","pages":"769-786"},"PeriodicalIF":5.1,"publicationDate":"2023-06-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1049/cit2.12238","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"78016573","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}