Li He, Kaize Shi, Dingxian Wang, Xianzhi Wang, Guandong Xu
With the introduction of more recent deep learning models such as encoder-decoder, text generation frameworks have gained a lot of popularity. In Natural Language Generation (NLG), controlling the information and style of the output produced is a crucial and challenging task. The purpose of this paper is to develop informative and controllable text using social media language by incorporating topic knowledge into a keyword-to-text framework. A novel Topic-Controllable Key-to-Text (TC-K2T) generator that focuses on the issues of ignoring unordered keywords and utilising subject-controlled information from previous research is presented. TC-K2T is built on the framework of conditional language encoders. In order to guide the model to produce an informative and controllable language, the generator first inputs unordered keywords and uses subjects to simulate prior human knowledge. Using an additional probability term, the model increases the likelihood of topic words appearing in the generated text to bias the overall distribution. The proposed TC-K2T can produce more informative and controllable senescence, outperforming state-of-the-art models, according to empirical research on automatic evaluation metrics and human annotations.
{"title":"A topic-controllable keywords-to-text generator with knowledge base network","authors":"Li He, Kaize Shi, Dingxian Wang, Xianzhi Wang, Guandong Xu","doi":"10.1049/cit2.12280","DOIUrl":"10.1049/cit2.12280","url":null,"abstract":"<p>With the introduction of more recent deep learning models such as encoder-decoder, text generation frameworks have gained a lot of popularity. In Natural Language Generation (NLG), controlling the information and style of the output produced is a crucial and challenging task. The purpose of this paper is to develop informative and controllable text using social media language by incorporating topic knowledge into a keyword-to-text framework. A novel Topic-Controllable Key-to-Text (TC-K2T) generator that focuses on the issues of ignoring unordered keywords and utilising subject-controlled information from previous research is presented. TC-K2T is built on the framework of conditional language encoders. In order to guide the model to produce an informative and controllable language, the generator first inputs unordered keywords and uses subjects to simulate prior human knowledge. Using an additional probability term, the model increases the likelihood of topic words appearing in the generated text to bias the overall distribution. The proposed TC-K2T can produce more informative and controllable senescence, outperforming state-of-the-art models, according to empirical research on automatic evaluation metrics and human annotations.</p>","PeriodicalId":46211,"journal":{"name":"CAAI Transactions on Intelligence Technology","volume":"9 3","pages":"585-594"},"PeriodicalIF":5.1,"publicationDate":"2024-01-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1049/cit2.12280","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139530560","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
A novel method based on the cross-modality intersecting features of the palm-vein and the palmprint is proposed for identity verification. Capitalising on the unique geometrical relationship between the two biometric modalities, the cross-modality intersecting points provides a stable set of features for identity verification. To facilitate flexibility in template changes, a template transformation is proposed. While maintaining non-invertibility, the template transformation allows transformation sizes beyond that offered by the conventional means. Extensive experiments using three public palm databases are conducted to verify the effectiveness the proposed system for identity recognition.
{"title":"Extraction of intersecting palm-vein and palmprint features for cancellable identity verification","authors":"Jaekwon Lee, Beom-Seok Oh, Kar-Ann Toh","doi":"10.1049/cit2.12277","DOIUrl":"10.1049/cit2.12277","url":null,"abstract":"<p>A novel method based on the cross-modality intersecting features of the palm-vein and the palmprint is proposed for identity verification. Capitalising on the unique geometrical relationship between the two biometric modalities, the cross-modality intersecting points provides a stable set of features for identity verification. To facilitate flexibility in template changes, a template transformation is proposed. While maintaining non-invertibility, the template transformation allows transformation sizes beyond that offered by the conventional means. Extensive experiments using three public palm databases are conducted to verify the effectiveness the proposed system for identity recognition.</p>","PeriodicalId":46211,"journal":{"name":"CAAI Transactions on Intelligence Technology","volume":"9 1","pages":"69-86"},"PeriodicalIF":5.1,"publicationDate":"2024-01-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1049/cit2.12277","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139625816","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Nechirvan Asaad Zebari, Chira Nadheef Mohammed, Dilovan Asaad Zebari, Mazin Abed Mohammed, Diyar Qader Zeebaree, Haydar Abdulameer Marhoon, Karrar Hameed Abdulkareem, Seifedine Kadry, Wattana Viriyasitavat, Jan Nedoma, Radek Martinek
Detecting brain tumours is complex due to the natural variation in their location, shape, and intensity in images. While having accurate detection and segmentation of brain tumours would be beneficial, current methods still need to solve this problem despite the numerous available approaches. Precise analysis of Magnetic Resonance Imaging (MRI) is crucial for detecting, segmenting, and classifying brain tumours in medical diagnostics. Magnetic Resonance Imaging is a vital component in medical diagnosis, and it requires precise, efficient, careful, efficient, and reliable image analysis techniques. The authors developed a Deep Learning (DL) fusion model to classify brain tumours reliably. Deep Learning models require large amounts of training data to achieve good results, so the researchers utilised data augmentation techniques to increase the dataset size for training models. VGG16, ResNet50, and convolutional deep belief networks networks extracted deep features from MRI images. Softmax was used as the classifier, and the training set was supplemented with intentionally created MRI images of brain tumours in addition to the genuine ones. The features of two DL models were combined in the proposed model to generate a fusion model, which significantly increased classification accuracy. An openly accessible dataset from the internet was used to test the model's performance, and the experimental results showed that the proposed fusion model achieved a classification accuracy of 98.98%. Finally, the results were compared with existing methods, and the proposed model outperformed them significantly.
{"title":"A deep learning fusion model for accurate classification of brain tumours in Magnetic Resonance images","authors":"Nechirvan Asaad Zebari, Chira Nadheef Mohammed, Dilovan Asaad Zebari, Mazin Abed Mohammed, Diyar Qader Zeebaree, Haydar Abdulameer Marhoon, Karrar Hameed Abdulkareem, Seifedine Kadry, Wattana Viriyasitavat, Jan Nedoma, Radek Martinek","doi":"10.1049/cit2.12276","DOIUrl":"10.1049/cit2.12276","url":null,"abstract":"<p>Detecting brain tumours is complex due to the natural variation in their location, shape, and intensity in images. While having accurate detection and segmentation of brain tumours would be beneficial, current methods still need to solve this problem despite the numerous available approaches. Precise analysis of Magnetic Resonance Imaging (MRI) is crucial for detecting, segmenting, and classifying brain tumours in medical diagnostics. Magnetic Resonance Imaging is a vital component in medical diagnosis, and it requires precise, efficient, careful, efficient, and reliable image analysis techniques. The authors developed a Deep Learning (DL) fusion model to classify brain tumours reliably. Deep Learning models require large amounts of training data to achieve good results, so the researchers utilised data augmentation techniques to increase the dataset size for training models. VGG16, ResNet50, and convolutional deep belief networks networks extracted deep features from MRI images. Softmax was used as the classifier, and the training set was supplemented with intentionally created MRI images of brain tumours in addition to the genuine ones. The features of two DL models were combined in the proposed model to generate a fusion model, which significantly increased classification accuracy. An openly accessible dataset from the internet was used to test the model's performance, and the experimental results showed that the proposed fusion model achieved a classification accuracy of 98.98%. Finally, the results were compared with existing methods, and the proposed model outperformed them significantly.</p>","PeriodicalId":46211,"journal":{"name":"CAAI Transactions on Intelligence Technology","volume":"9 4","pages":"790-804"},"PeriodicalIF":8.4,"publicationDate":"2024-01-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1049/cit2.12276","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139536199","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
<p>With the development of artificial intelligence, remote sensing scene interpretation task has attracted extensive attention, which mainly includes scene classification, target detection, hyperspectral classification, and multi-modal analysis. The remote sensing scene interpretation has effectively promoted the development of the Earth observation field. It was the intention for this Special Issue to serve as a platform for the publication of the most recent research concepts from remote sensing image.</p><p>To recognise remote sensing scenes, several methods have been proposed to represent the scene image. The first paper (Zhang et al.) proposes a lightweight privacy-preserving recognition framework which diffuses the error between the encryption block and the original block to adjacent blocks which makes the transmission of high-resolution images more secure and efficient. The second paper (Ning et al.) introduces a knowledge distillation network for aerial scene recognition, which produces consistent predictions by distilling the predictive distribution between different scales. With the development of scene recognition task, its branch scene retrieval task also emerges. In this regard, the third paper (Yuan et al.) shows how to efficiently optimise the average accuracy to improve remote sensing image retrieval. This approach enables a more flexible optimisation strategy by involving positive post-samples, which provides a new way to improve the retrieval performance.</p><p>To detect targets, a series of advanced methods have been developed to improve detection accuracy and efficiency. The fourth paper (Zhang et al.) proposes an intelligent anchor learning strategy for arbitrary orientation target detection. The fifth paper (Ma et al.) focuses on infrared image detection of small and weak targets and proposes an efficient deep learning method. The sixth paper (Zhou et al.) proposes a convolutional transformer method based on spectral-spatial sequence features for hyperspectral image change detection. With the maturity of target detection techniques, researchers have begun to focus on more complex challenges, namely anomaly detection. In this subfield, the seventh paper (Wang et al.) provides a new solution for semi-supervised hyperspectral anomaly detection. It maps the raw spectrum into the fractional Fourier domain, thereby enhancing the distinguishability between background and anomaly. Meanwhile, the eighth paper (Zhao et al.) utilises a memory-enhanced self-encoder to improve the separation of anomaly samples from background in hyperspectral images. These studies demonstrate the rapid development in the target detection field, such as change detection and anomaly detection.</p><p>To classify hyperspectral images, the ninth paper (Liao et al.) shows how to integrate the features of convolutional neural networks and transformers to enhance the performance of hyperspectral image classification. This approach fully utilises the respective a
随着人工智能的发展,遥感场景解译任务受到了广泛的关注,主要包括场景分类、目标检测、高光谱分类和多模态分析。遥感场景解译有效地促进了对地观测领域的发展。本期特刊的目的是作为一个平台,发表遥感影像的最新研究概念。为了识别遥感场景,提出了几种场景图像表示方法。第一篇论文(Zhang et al.)提出了一种轻量级的隐私保护识别框架,该框架将加密块与原始块之间的错误扩散到相邻块,从而使高分辨率图像的传输更加安全高效。第二篇论文(Ning et al.)介绍了一种用于航景识别的知识蒸馏网络,该网络通过提取不同尺度之间的预测分布来产生一致的预测。随着场景识别任务的发展,其分支场景检索任务也应运而生。在这方面,第三篇论文(Yuan et al.)展示了如何有效地优化平均精度以提高遥感图像检索。该方法通过引入阳性后样本实现了更灵活的优化策略,为提高检索性能提供了一种新的方法。为了检测目标,人们开发了一系列先进的方法来提高检测精度和效率。第四篇论文(Zhang等人)提出了一种用于任意方向目标检测的智能锚点学习策略。第五篇论文(Ma et al.)专注于红外图像弱小目标的检测,提出了一种高效的深度学习方法。第六篇论文(Zhou et al.)提出了一种基于光谱空间序列特征的卷积变换方法,用于高光谱图像变化检测。随着目标检测技术的成熟,研究人员开始关注更复杂的挑战,即异常检测。在该子领域,第七篇论文(Wang et al.)为半监督高光谱异常检测提供了一种新的解决方案。它将原始光谱映射到分数傅里叶域,从而增强背景和异常之间的可区分性。同时,第八篇论文(Zhao et al.)利用记忆增强的自编码器来提高高光谱图像中异常样本与背景的分离。这些研究表明了变化检测和异常检测等目标检测领域的快速发展。为了对高光谱图像进行分类,第九篇论文(Liao et al.)展示了如何结合卷积神经网络和变压器的特点来提高高光谱图像的分类性能。该方法充分利用了卷积网络和变压器各自的优势,为高光谱图像的特征提取提供了全面的解决方案。此外,第十篇论文(Xie et al.)采用融合语义、空间和频谱特征的变压器网络,展示了多种信息类型的组合如何提高分类的准确性和鲁棒性。同时,第11篇论文(Ran et al.)将深度变压器建模与小样本学习相结合,解决了高光谱图像分类中的挑战,特别是在样本数量有限的情况下。该方法充分利用了少量样本中的信息,有效地提高了分类模型的泛化能力。多模态分析增加了观测地球表面和处理具有挑战性问题的能力。第12篇论文(Hong et al.)利用多光谱遥感数据和地理加权回归实验揭示了绿色基础设施布局对缓解城市热岛效应的重要性。第13篇论文(Zhang et al.)介绍了降水预测的多任务框架。将雷达回波图像与其他辅助任务相结合,提高了降水预报的精度和效率。最后,第十四篇论文(Zhang et al.)通过融合视觉和音频数据来改进机器人在动态环境中的自定位和环境感知。该融合方法在多机器人协同场景下具有良好的稳定性和重构性能。这些研究证明了多源数据分析在改善环境监测、预测和机器人导航等领域的潜力。我们感谢所有作者的投稿和所有审稿人的宝贵评论和意见。 我们希望这期特刊能够激发研究界在遥感场景解译方面的新成果。国家自然科学基金项目,资助/奖励号:62271484。
{"title":"Guest Editorial: Special issue on intelligence technology for remote sensing image","authors":"Xiangtao Zheng, Benoit Vozel, Danfeng Hong","doi":"10.1049/cit2.12275","DOIUrl":"https://doi.org/10.1049/cit2.12275","url":null,"abstract":"<p>With the development of artificial intelligence, remote sensing scene interpretation task has attracted extensive attention, which mainly includes scene classification, target detection, hyperspectral classification, and multi-modal analysis. The remote sensing scene interpretation has effectively promoted the development of the Earth observation field. It was the intention for this Special Issue to serve as a platform for the publication of the most recent research concepts from remote sensing image.</p><p>To recognise remote sensing scenes, several methods have been proposed to represent the scene image. The first paper (Zhang et al.) proposes a lightweight privacy-preserving recognition framework which diffuses the error between the encryption block and the original block to adjacent blocks which makes the transmission of high-resolution images more secure and efficient. The second paper (Ning et al.) introduces a knowledge distillation network for aerial scene recognition, which produces consistent predictions by distilling the predictive distribution between different scales. With the development of scene recognition task, its branch scene retrieval task also emerges. In this regard, the third paper (Yuan et al.) shows how to efficiently optimise the average accuracy to improve remote sensing image retrieval. This approach enables a more flexible optimisation strategy by involving positive post-samples, which provides a new way to improve the retrieval performance.</p><p>To detect targets, a series of advanced methods have been developed to improve detection accuracy and efficiency. The fourth paper (Zhang et al.) proposes an intelligent anchor learning strategy for arbitrary orientation target detection. The fifth paper (Ma et al.) focuses on infrared image detection of small and weak targets and proposes an efficient deep learning method. The sixth paper (Zhou et al.) proposes a convolutional transformer method based on spectral-spatial sequence features for hyperspectral image change detection. With the maturity of target detection techniques, researchers have begun to focus on more complex challenges, namely anomaly detection. In this subfield, the seventh paper (Wang et al.) provides a new solution for semi-supervised hyperspectral anomaly detection. It maps the raw spectrum into the fractional Fourier domain, thereby enhancing the distinguishability between background and anomaly. Meanwhile, the eighth paper (Zhao et al.) utilises a memory-enhanced self-encoder to improve the separation of anomaly samples from background in hyperspectral images. These studies demonstrate the rapid development in the target detection field, such as change detection and anomaly detection.</p><p>To classify hyperspectral images, the ninth paper (Liao et al.) shows how to integrate the features of convolutional neural networks and transformers to enhance the performance of hyperspectral image classification. This approach fully utilises the respective a","PeriodicalId":46211,"journal":{"name":"CAAI Transactions on Intelligence Technology","volume":"8 4","pages":"1164-1165"},"PeriodicalIF":5.1,"publicationDate":"2023-12-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ietresearch.onlinelibrary.wiley.com/doi/epdf/10.1049/cit2.12275","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138678873","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Few-shot image classification is the task of classifying novel classes using extremely limited labelled samples. To perform classification using the limited samples, one solution is to learn the feature alignment (FA) information between the labelled and unlabelled sample features. Most FA methods use the feature mean as the class prototype and calculate the correlation between prototype and unlabelled features to learn an alignment strategy. However, mean prototypes tend to degenerate informative features because spatial features at the same position may not be equally important for the final classification, leading to inaccurate correlation calculations. Therefore, the authors propose an effective intraclass FA strategy that aggregates semantically similar spatial features from an adaptive reference prototype in low-dimensional feature space to obtain an informative prototype feature map for precise correlation computation. Moreover, a dual correlation module to learn the hard and soft correlations was developed by the authors. This module combines the correlation information between the prototype and unlabelled features in both the original and learnable feature spaces, aiming to produce a comprehensive cross-correlation between the prototypes and unlabelled features. Using both FA and cross-attention modules, our model can maintain informative class features and capture important shared features for classification. Experimental results on three few-shot classification benchmarks show that the proposed method outperformed related methods and resulted in a 3% performance boost in the 1-shot setting by inserting the proposed module into the related methods.
{"title":"Learning feature alignment and dual correlation for few-shot image classification","authors":"Xilang Huang, Seon Han Choi","doi":"10.1049/cit2.12273","DOIUrl":"10.1049/cit2.12273","url":null,"abstract":"<p>Few-shot image classification is the task of classifying novel classes using extremely limited labelled samples. To perform classification using the limited samples, one solution is to learn the feature alignment (FA) information between the labelled and unlabelled sample features. Most FA methods use the feature mean as the class prototype and calculate the correlation between prototype and unlabelled features to learn an alignment strategy. However, mean prototypes tend to degenerate informative features because spatial features at the same position may not be equally important for the final classification, leading to inaccurate correlation calculations. Therefore, the authors propose an effective intraclass FA strategy that aggregates semantically similar spatial features from an adaptive reference prototype in low-dimensional feature space to obtain an informative prototype feature map for precise correlation computation. Moreover, a dual correlation module to learn the hard and soft correlations was developed by the authors. This module combines the correlation information between the prototype and unlabelled features in both the original and learnable feature spaces, aiming to produce a comprehensive cross-correlation between the prototypes and unlabelled features. Using both FA and cross-attention modules, our model can maintain informative class features and capture important shared features for classification. Experimental results on three few-shot classification benchmarks show that the proposed method outperformed related methods and resulted in a 3% performance boost in the 1-shot setting by inserting the proposed module into the related methods.</p>","PeriodicalId":46211,"journal":{"name":"CAAI Transactions on Intelligence Technology","volume":"9 2","pages":"303-318"},"PeriodicalIF":5.1,"publicationDate":"2023-12-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1049/cit2.12273","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138598869","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Green Infrastructure (GI) has garnered increasing attention from various regions due to its potential to mitigate urban heat island (UHI), which has been exacerbated by global climate change. This study focuses on the central area of Fuzhou city, one of the “furnace” cities, and aims to explore the correlation between the GI pattern and land surface temperature (LST) in the spring and autumn seasons. The research adopts a multiscale approach, starting from the urban scale and using urban geographic spatial characteristics, multispectral remote sensing data, and morphological spatial pattern analysis (MSPA). Significant MSPA elements were tested and combined with LST to conduct a geographic weighted regression (GWR) experiment. The findings reveal that the UHI in the central area of Fuzhou city has a spatial characteristic of “high temperature in the middle and low temperature around,” which is coupled with a “central scattered and peripheral concentrated” distribution of GI. This suggests that remote sensing data can effectively be utilised for UHI inversion. Additionally, the study finds that the complexity of GI, whether from the perspective of the overall GI pattern or the classification study based on the proportion of the core area, has an impact on the alleviation of UHI in both seasons. In conclusion, this study underscores the importance of a reasonable layout of urban green infrastructure for mitigating UHI.
{"title":"Exploring the spatiotemporal relationship between green infrastructure and urban heat island under multi-source remote sensing imagery: A case study of Fuzhou City","authors":"Tingting Hong, Xiaohui Huang, Guangjian Chen, Yiwei Yang, Lijia Chen","doi":"10.1049/cit2.12272","DOIUrl":"10.1049/cit2.12272","url":null,"abstract":"<p>Green Infrastructure (GI) has garnered increasing attention from various regions due to its potential to mitigate urban heat island (UHI), which has been exacerbated by global climate change. This study focuses on the central area of Fuzhou city, one of the “furnace” cities, and aims to explore the correlation between the GI pattern and land surface temperature (LST) in the spring and autumn seasons. The research adopts a multiscale approach, starting from the urban scale and using urban geographic spatial characteristics, multispectral remote sensing data, and morphological spatial pattern analysis (MSPA). Significant MSPA elements were tested and combined with LST to conduct a geographic weighted regression (GWR) experiment. The findings reveal that the UHI in the central area of Fuzhou city has a spatial characteristic of “high temperature in the middle and low temperature around,” which is coupled with a “central scattered and peripheral concentrated” distribution of GI. This suggests that remote sensing data can effectively be utilised for UHI inversion. Additionally, the study finds that the complexity of GI, whether from the perspective of the overall GI pattern or the classification study based on the proportion of the core area, has an impact on the alleviation of UHI in both seasons. In conclusion, this study underscores the importance of a reasonable layout of urban green infrastructure for mitigating UHI.</p>","PeriodicalId":46211,"journal":{"name":"CAAI Transactions on Intelligence Technology","volume":"8 4","pages":"1337-1349"},"PeriodicalIF":5.1,"publicationDate":"2023-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ietresearch.onlinelibrary.wiley.com/doi/epdf/10.1049/cit2.12272","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138604706","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Ki-Il Kim, Aswani Kumar Cherukuri, Xue Jun Li, Tanveer Ahmad, Muhammad Rafiq, Shehzad Ashraf Chaudhry
<p>The convergence of Internet of Things (IoT), vehicularad hoc network (VANET), and mobile ad hoc network relies on sensor networks to gather data from nodes or objects. These networks involve nodes, gateways, and anchors, operating on limited battery power, mainly used in broadcasting. IoT applications, like healthcare, smart cities, and transportation, often need position data and face challenges in delay sensitivity. Localisation is important in ITS and VANETs, influencing autonomous vehicles, collision warning systems, and road information dissemination. A robust localisation system, often combining GPS with techniques like Dead Reckoning and Image/Video Localisation, is essential for accuracy and security. Artificial intelligence (AI) integration, particularly in machine learning, enhances indoor wireless localisation effectiveness. Advancements in wireless communication (WSN, IoT, and massive MIMO) transform dense environments into programmable entities, but pose challenges in aligning self-learning AI with sensor tech for accuracy and budget considerations. We seek original research on sensor localisation, fusion, protocols, and positioning algorithms, inviting contributions from industry and academia to address these evolving challenges.</p><p>This special issue titled ‘Sensing, Communication, and Localization in WSN, IoT & VANET’ appears in the CAAI Transactions on Intelligence Technology. We encourage contributions addressing localisation accuracy, network coverage, upper and lower bounding, lane and vehicle detection, and related topics.</p><p>In the first paper, (Hamil et al.) explore how smartphone sensors and IoT devices aid in rescuing individuals during emergencies like fires in tall buildings. It introduces a pioneering Sensor Management and Data Fusion-Wireless Data Exchange fusion scheme, leveraging an evolutionary algorithm within complex multi-storey buildings. This scheme aims to diversify particle sets effectively, capturing the user's real-time state using wearable device sensors. The authors further explore how smartphones sensors utilise data for object movement alongside Bluetooth Low Energy beacon based localisation with the help of Sensor Management security and Data Fusion-Wireless Data Exchange scheme. The effectiveness of this scheme and its impact on a smartphone user's real-time state within indoor settings were assessed through various experiments in controlled environments.</p><p>In the second paper, (Khan J et al.) proposed a novel method to fine-tune alpha-beta filter parameters using a feed-forward backpropagation neural network. This model, comprising the alpha-beta filter as the core predictor and a feedforward artificial neural network as the learning element, uses temperature and humidity sensor data for precise predictions from noisy readings. By integrating the feed-forward backpropagation neural network significantly boosts prediction accuracy, slashing both roots mean square error (RMSE) and mea
物联网(IoT)、车载自组织网络(VANET)和移动自组织网络的融合依赖于传感器网络从节点或对象收集数据。这些网络包括节点、网关和锚,在有限的电池电量下运行,主要用于广播。物联网应用,如医疗保健、智慧城市和交通,通常需要位置数据,并面临延迟敏感性的挑战。定位在ITS和VANETs中非常重要,影响着自动驾驶汽车、碰撞预警系统和道路信息发布。一个强大的定位系统,通常结合GPS与诸如航位推算和图像/视频定位等技术,对于准确性和安全性至关重要。人工智能(AI)集成,特别是在机器学习方面,提高了室内无线定位的有效性。无线通信(WSN、物联网和大规模MIMO)的进步将密集环境转变为可编程实体,但在将自学习AI与传感器技术结合起来以实现准确性和预算考虑方面提出了挑战。我们寻求在传感器定位、融合、协议和定位算法方面的原创研究,邀请工业界和学术界的贡献来解决这些不断变化的挑战。本期特刊题为“WSN、IoT中的传感、通信和定位”。VANET出现在CAAI智能技术汇刊上。我们鼓励就定位精度、网络覆盖、上下边界、车道和车辆检测以及相关主题发表意见。在第一篇论文中,(Hamil et al.)探讨了智能手机传感器和物联网设备如何在高层建筑火灾等紧急情况下帮助救援人员。它引入了一种开创性的传感器管理和数据融合-无线数据交换融合方案,利用复杂多层建筑中的进化算法。该方案旨在有效地多样化粒子集,利用可穿戴设备传感器捕捉用户的实时状态。作者进一步探讨了智能手机传感器如何在传感器管理安全和数据融合无线数据交换方案的帮助下,利用基于蓝牙低功耗信标的本地化数据进行物体运动。通过在受控环境中的各种实验,评估了该方案的有效性及其对室内智能手机用户实时状态的影响。在第二篇论文中,(Khan J et al.)提出了一种使用前馈反向传播神经网络微调α - β滤波器参数的新方法。该模型由alpha-beta滤波器作为核心预测器和前馈人工神经网络作为学习元素组成,使用温度和湿度传感器数据从噪声读数中进行精确预测。通过整合前馈反向传播神经网络,显著提高了预测精度,降低了均方根误差(RMSE)和平均绝对误差(MAE)。在与传统方法(如α - β和卡尔曼滤波器)的实验中,所提出的模型表现优于传统方法,MAE提高了35.1%,RMSE提高了38.2%。在第三篇论文中,(Imtiaz等人)提出了一种存在翻转歧义的工业物联网本地化方案。为了减少IIoT网络中的定位误差估计,作者提出了一种新的贪婪锚点选择策略GSAP。本文提出了利用多维尺度进行初始位置估计的总体思路,减少了算法的收敛时间。推导了所提算法的Cramer - Rao下界表达式,以检验其最优性,并将结果与目前的技术水平进行比较。在第四篇论文中(Ismail等人)推导了单个EH中继下的NOMA窄带物联网网络。然而,窄带物联网设备的增长也导致了同信道干扰的增加,从而影响了NOMA的性能增强。在上行EH中继NOMA窄带物联网网络中,作者的目标是优化窄带物联网设备数据速率,同时满足其最低要求。考虑到设备能量、EH中继能量和数据缓存约束,该模型创建了一个复杂的室内定位框架,涉及功率、数据和时隙调度。这个模型提出了一个非凸优化挑战,没有一个直接的分析解决方案。通过仿真验证了该方法的有效性。这些改进使网络的能源效率提高了44.1%,数据速率比例公平提高了11.9%,频谱效率提高了55.4%。我们感谢所有作者的投稿和审稿人的宝贵反馈。我们希望这期特刊能在循环动态神经网络领域为研究界带来新的成果。
{"title":"Guest Editorial: Special issue on explainable AI empowered for indoor positioning and indoor navigation","authors":"Ki-Il Kim, Aswani Kumar Cherukuri, Xue Jun Li, Tanveer Ahmad, Muhammad Rafiq, Shehzad Ashraf Chaudhry","doi":"10.1049/cit2.12274","DOIUrl":"https://doi.org/10.1049/cit2.12274","url":null,"abstract":"<p>The convergence of Internet of Things (IoT), vehicularad hoc network (VANET), and mobile ad hoc network relies on sensor networks to gather data from nodes or objects. These networks involve nodes, gateways, and anchors, operating on limited battery power, mainly used in broadcasting. IoT applications, like healthcare, smart cities, and transportation, often need position data and face challenges in delay sensitivity. Localisation is important in ITS and VANETs, influencing autonomous vehicles, collision warning systems, and road information dissemination. A robust localisation system, often combining GPS with techniques like Dead Reckoning and Image/Video Localisation, is essential for accuracy and security. Artificial intelligence (AI) integration, particularly in machine learning, enhances indoor wireless localisation effectiveness. Advancements in wireless communication (WSN, IoT, and massive MIMO) transform dense environments into programmable entities, but pose challenges in aligning self-learning AI with sensor tech for accuracy and budget considerations. We seek original research on sensor localisation, fusion, protocols, and positioning algorithms, inviting contributions from industry and academia to address these evolving challenges.</p><p>This special issue titled ‘Sensing, Communication, and Localization in WSN, IoT & VANET’ appears in the CAAI Transactions on Intelligence Technology. We encourage contributions addressing localisation accuracy, network coverage, upper and lower bounding, lane and vehicle detection, and related topics.</p><p>In the first paper, (Hamil et al.) explore how smartphone sensors and IoT devices aid in rescuing individuals during emergencies like fires in tall buildings. It introduces a pioneering Sensor Management and Data Fusion-Wireless Data Exchange fusion scheme, leveraging an evolutionary algorithm within complex multi-storey buildings. This scheme aims to diversify particle sets effectively, capturing the user's real-time state using wearable device sensors. The authors further explore how smartphones sensors utilise data for object movement alongside Bluetooth Low Energy beacon based localisation with the help of Sensor Management security and Data Fusion-Wireless Data Exchange scheme. The effectiveness of this scheme and its impact on a smartphone user's real-time state within indoor settings were assessed through various experiments in controlled environments.</p><p>In the second paper, (Khan J et al.) proposed a novel method to fine-tune alpha-beta filter parameters using a feed-forward backpropagation neural network. This model, comprising the alpha-beta filter as the core predictor and a feedforward artificial neural network as the learning element, uses temperature and humidity sensor data for precise predictions from noisy readings. By integrating the feed-forward backpropagation neural network significantly boosts prediction accuracy, slashing both roots mean square error (RMSE) and mea","PeriodicalId":46211,"journal":{"name":"CAAI Transactions on Intelligence Technology","volume":"8 4","pages":"1101-1103"},"PeriodicalIF":5.1,"publicationDate":"2023-12-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ietresearch.onlinelibrary.wiley.com/doi/epdf/10.1049/cit2.12274","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138678834","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
In traditional secret image sharing schemes, a secret image is shared among shareholders who have the same position. But if the shareholders have two different positions, essential and non-essential, it is necessary to use essential secret image sharing schemes. In this article, a verifiable essential secret image sharing scheme based on HLRs is proposed. Shareholder's share consists of two parts. The first part is produced by the shareholders, which prevents the fraud of dealers. The second part is a shadow image that is produced by using HLRs and the first part of share. The verification of the first part of the shares is done for the first time by using multilinear and bilinear maps. Also, for verifying shadow images, Bloom Filters are used for the first time. The proposed scheme is more efficient than similar schemes, and for the first part of the shares, has formal security.
{"title":"A verifiable essential secret image sharing scheme based on HLRs (VESIS-(t, s, k, n))","authors":"Massoud Hadian Dehkordi, Seyed Taghi Farahi, Samaneh Mashhadi","doi":"10.1049/cit2.12271","DOIUrl":"10.1049/cit2.12271","url":null,"abstract":"<p>In traditional secret image sharing schemes, a secret image is shared among shareholders who have the same position. But if the shareholders have two different positions, essential and non-essential, it is necessary to use essential secret image sharing schemes. In this article, a verifiable essential secret image sharing scheme based on HLRs is proposed. Shareholder's share consists of two parts. The first part is produced by the shareholders, which prevents the fraud of dealers. The second part is a shadow image that is produced by using HLRs and the first part of share. The verification of the first part of the shares is done for the first time by using multilinear and bilinear maps. Also, for verifying shadow images, Bloom Filters are used for the first time. The proposed scheme is more efficient than similar schemes, and for the first part of the shares, has formal security.</p>","PeriodicalId":46211,"journal":{"name":"CAAI Transactions on Intelligence Technology","volume":"9 2","pages":"388-410"},"PeriodicalIF":5.1,"publicationDate":"2023-10-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1049/cit2.12271","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135618098","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Abstract When deploying mobile robots in real‐world scenarios, such as airports, train stations, hospitals, and schools, collisions with pedestrians are intolerable and catastrophic. Motion safety becomes one of the most fundamental requirements for mobile robots. However, until now, efficient and safe robot navigation in such dynamic environments is still an open problem. The critical reason is that the inconsistency between navigation efficiency and motion safety is greatly intensified by the high dynamics and uncertainties of pedestrians. To face the challenge, this paper proposes a safe deep reinforcement learning algorithm named Conflict‐Averse Safe Reinforcement Learning (CASRL) for autonomous robot navigation in dynamic environments. Specifically, it first separates the collision avoidance sub‐task from the overall navigation task and maintains a safety critic to evaluate the safety/risk of actions. Later, it constructs two task‐specific but model‐agnostic policy gradients for goal‐reaching and collision avoidance sub‐tasks to eliminate their mutual interference. Then, it further performs a conflict‐averse gradient manipulation to address the inconsistency between two sub‐tasks. Finally, extensive experiments are performed to evaluate the superiority of CASRL. Simulation results show an average 8.2% performance improvement over the vanilla baseline in eight groups of dynamic environments, which is further extended to 13.4% in the most challenging group. Besides, forty real‐world experiments fully illustrated that the CASRL could be successfully deployed on a real robot.
{"title":"A safe reinforcement learning approach for autonomous navigation of mobile robots in dynamic environments","authors":"Zhiqian Zhou, Junkai Ren, Zhiwen Zeng, Junhao Xiao, Xinglong Zhang, Xian Guo, Zongtan Zhou, Huimin Lu","doi":"10.1049/cit2.12269","DOIUrl":"https://doi.org/10.1049/cit2.12269","url":null,"abstract":"Abstract When deploying mobile robots in real‐world scenarios, such as airports, train stations, hospitals, and schools, collisions with pedestrians are intolerable and catastrophic. Motion safety becomes one of the most fundamental requirements for mobile robots. However, until now, efficient and safe robot navigation in such dynamic environments is still an open problem. The critical reason is that the inconsistency between navigation efficiency and motion safety is greatly intensified by the high dynamics and uncertainties of pedestrians. To face the challenge, this paper proposes a safe deep reinforcement learning algorithm named Conflict‐Averse Safe Reinforcement Learning (CASRL) for autonomous robot navigation in dynamic environments. Specifically, it first separates the collision avoidance sub‐task from the overall navigation task and maintains a safety critic to evaluate the safety/risk of actions. Later, it constructs two task‐specific but model‐agnostic policy gradients for goal‐reaching and collision avoidance sub‐tasks to eliminate their mutual interference. Then, it further performs a conflict‐averse gradient manipulation to address the inconsistency between two sub‐tasks. Finally, extensive experiments are performed to evaluate the superiority of CASRL. Simulation results show an average 8.2% performance improvement over the vanilla baseline in eight groups of dynamic environments, which is further extended to 13.4% in the most challenging group. Besides, forty real‐world experiments fully illustrated that the CASRL could be successfully deployed on a real robot.","PeriodicalId":46211,"journal":{"name":"CAAI Transactions on Intelligence Technology","volume":"62 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-10-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135146697","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Magnetic resonance (MR) imaging is a widely employed medical imaging technique that produces detailed anatomical images of the human body. The segmentation of MR images plays a crucial role in medical image analysis, as it enables accurate diagnosis, treatment planning, and monitoring of various diseases and conditions. Due to the lack of sufficient medical images, it is challenging to achieve an accurate segmentation, especially with the application of deep learning networks. The aim of this work is to study transfer learning from T1-weighted (T1-w) to T2-weighted (T2-w) MR sequences to enhance bone segmentation with minimal required computation resources. With the use of an excitation-based convolutional neural networks, four transfer learning mechanisms are proposed: transfer learning without fine tuning, open fine tuning, conservative fine tuning, and hybrid transfer learning. Moreover, a multi-parametric segmentation model is proposed using T2-w MR as an intensity-based augmentation technique. The novelty of this work emerges in the hybrid transfer learning approach that overcomes the overfitting issue and preserves the features of both modalities with minimal computation time and resources. The segmentation results are evaluated using 14 clinical 3D brain MR and CT images. The results reveal that hybrid transfer learning is superior for bone segmentation in terms of performance and computation time with DSCs of 0.5393 ± 0.0007. Although T2-w-based augmentation has no significant impact on the performance of T1-w MR segmentation, it helps in improving T2-w MR segmentation and developing a multi-sequences segmentation model.
{"title":"Transfer learning from T1-weighted to T2-weighted Magnetic resonance sequences for brain image segmentation","authors":"Imene Mecheter, Maysam Abbod, Habib Zaidi, Abbes Amira","doi":"10.1049/cit2.12270","DOIUrl":"10.1049/cit2.12270","url":null,"abstract":"<p>Magnetic resonance (MR) imaging is a widely employed medical imaging technique that produces detailed anatomical images of the human body. The segmentation of MR images plays a crucial role in medical image analysis, as it enables accurate diagnosis, treatment planning, and monitoring of various diseases and conditions. Due to the lack of sufficient medical images, it is challenging to achieve an accurate segmentation, especially with the application of deep learning networks. The aim of this work is to study transfer learning from T1-weighted (T1-w) to T2-weighted (T2-w) MR sequences to enhance bone segmentation with minimal required computation resources. With the use of an excitation-based convolutional neural networks, four transfer learning mechanisms are proposed: transfer learning without fine tuning, open fine tuning, conservative fine tuning, and hybrid transfer learning. Moreover, a multi-parametric segmentation model is proposed using T2-w MR as an intensity-based augmentation technique. The novelty of this work emerges in the hybrid transfer learning approach that overcomes the overfitting issue and preserves the features of both modalities with minimal computation time and resources. The segmentation results are evaluated using 14 clinical 3D brain MR and CT images. The results reveal that hybrid transfer learning is superior for bone segmentation in terms of performance and computation time with DSCs of 0.5393 ± 0.0007. Although T2-w-based augmentation has no significant impact on the performance of T1-w MR segmentation, it helps in improving T2-w MR segmentation and developing a multi-sequences segmentation model.</p>","PeriodicalId":46211,"journal":{"name":"CAAI Transactions on Intelligence Technology","volume":"9 1","pages":"26-39"},"PeriodicalIF":5.1,"publicationDate":"2023-10-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1049/cit2.12270","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135645979","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}