{"title":"特邀社论:生物医学中基于知识的深度学习系统","authors":"Yu-Dong Zhang, Juan Manuel Górriz","doi":"10.1049/cit2.12364","DOIUrl":null,"url":null,"abstract":"<p>Numerous healthcare procedures can be viewed as medical sector decisions. In the modern era, computers have become indispensable in the realm of medical decision-making. However, the common view of computers in the medical field typically extends only to applications that support doctors in diagnosing diseases. To more tightly intertwine computers with the biomedical sciences, professionals are now more frequently utilising knowledge-driven deep learning systems (KDLS) and their foundational technologies, especially in the domain of neuroimaging (NI).</p><p>Data for medical purposes can be sourced from a variety of imaging techniques, including but not limited to Computed Tomography (CT), Magnetic Resonance Imaging (MRI), Ultrasound, Single Photon Emission Computed Tomography (SPECT), Positron Emission Tomography (PET), Magnetic Particle Imaging (MPI), Electroencephalography (EEG), Magnetoencephalography (MEG), Optical Microscopy and Tomography, Photoacoustic Tomography, Electron Tomography, and Atomic Force Microscopy.</p><p>Historically, these imaging techniques have been analysed using traditional statistical methods, such as hypothesis testing or Bayesian inference, which often presuppose certain conditions that are not always met. An emerging solution is the implementation of machine learning (ML) within the context of KDLS, allowing for the empirical mapping of complex, multi-dimensional relationships within data sets.</p><p>The objective of this special issue is to showcase the latest advancements in the methodology of KDLS for evaluating functional connectivity, neurological disorders, and clinical neuroscience, such as conditions like Alzheimer's, Parkinson's, cerebrovascular accidents, brain tumours, epilepsy, multiple sclerosis, ALS, Autism Spectrum Disorder, and more. Additionally, the special issue seeks to elucidate the mechanisms behind the predictive capabilities of ML methods within KDLS for brain-related diseases and disorders.</p><p>We received an abundance of submissions, totalling more than 40, from over 10 countries. After a meticulous and rigorous peer review process, which employed a double-blind methodology, we ultimately selected eight outstanding papers for publication. This process ensured the highest standards of quality and impartiality in the selection.</p><p>In the article ‘A deep learning fusion model for accurate classification of brain tumours in Magnetic Resonance images’, Zebari et al. created a robust deep learning (DL) fusion model for accurate brain tumour classification. To enhance performance, they employed data augmentation to expand the training dataset. The model leveraged VGG16, ResNet50, and convolutional deep belief networks to extract features from MRI images using a softmax classifier. By fusing features from two DL models, the fusion model notably boosted classification precision. Tested with a publicly available dataset, it achieved a remarkable 98.98% accuracy rate, outperforming existing methods.</p><p>In the article ‘Knowledge-based deep learning system for classifying’, Dhaygude et al. proposed a deep three-dimensional convolutional neural network incorporating multi-task learning and attention mechanisms. An upgraded primary C3D network is utilised to create rougher low-level feature maps. It introduces a new convolution block that focuses on the structural aspects of the magnetic resonance imaging image and another block that extracts attention weights unique to certain pixel positions in the feature map and multiplies them with the feature map output. Then, several fully connected layers are used to achieve multi-task learning, generating three outputs, including the primary classification task. The other two outputs employ backpropagation during training to improve the primary classification job. Experimental findings show that the authors' proposed method outperforms current approaches for classifying AD, achieving enhanced classification accuracy and other indicators on the Alzheimer's disease neuroimaging initiative dataset. The authors demonstrated promise for future disease classification studies.</p><p>In the paper entitled ‘A novel medical image data protection scheme for smart healthcare system’, Rehman et al. presented a lightweight medical image encryption scheme utilising bit-plane decomposition and chaos theory. The results of the experiment yielded entropy, energy, and correlation values of 7.999, 0.0156, and 0.0001, respectively. This validated the effectiveness of the encryption system proposed in this paper, which offered high-quality encryption, a large key space, key sensitivity, and resistance to statistical attacks.</p><p>In the paper entitled ‘Image super-resolution via dynamic network’, Tian et al. presented a dynamic network for image super-resolution (DSRNet), which contained a residual enhancement block, wide enhancement block, feature refinement block, and construction block. The residual enhancement block was composed of a residual enhanced architecture to facilitate hierarchical features for image super-resolution. To enhance the robustness of the obtained super-resolution model for complex scenes, a wide enhancement block achieved a dynamic architecture to learn more robust information to enhance the applicability of an obtained super-resolution model for varying scenes. To prevent interference of components in a wide enhancement block, a refinement block utilised a stacked architecture to learn obtained features accurately. Also, a residual learning operation was embedded in the refinement block to prevent long-term dependency problems. Finally, a construction block was responsible for reconstructing high-quality images. Designed heterogeneous architecture could not only facilitate richer structural information but also be lightweight, which was suitable for mobile digital devices. Experimental results showed that our method is more competitive in performance, recovering time of image super-resolution, and complexity. The code of DSRNet can be obtained at https://github.com/hellloxiaotian/DSRNet.</p><p>In the paper ‘Improved organs at risk segmentation based on modified U-Net with self-attention and consistency regularisation’, Manko et al. presented a new approach to automatic OAR segmentation in the chest cavity in Computed Tomography (CT) images. The proposed approach is based on the modified U-Net architecture with the ResNe-34 encoder, the baseline adopted in this work. The new two-branch CS-SA U-Net architecture is proposed, which consists of two parallel U-Net models in which self-attention blocks with cosine similarity as query-key similarity function (CS-SA) blocks are inserted between the encoder and decoder, which enabled the use of consistency regularisation. The proposed solution demonstrates state-of-the-art performance for the problem of OAR segmentation in CT images on the publicly available SegTHOR benchmark dataset in terms of a Dice coefficient (oesophagus-0.8714, heart-0.9516, trachea-0.9286, and aorta-0.9510) and Hausdorff distance (oesophagus-0.2541, heart-0.1514, trachea-0.1722, and aorta-0.1114) and significantly outperforms the baseline. The current approach was demonstrated to be viable for improving the quality of OAR segmentation for radiotherapy planning.</p><p>In the paper ‘GAN-MD: A myocarditis detection using multi-channel’, Golilarz et al. suggested a novel technique for improving the classification precision of CNNs. Using generative adversarial networks (GANs) to create synthetic images for data augmentation, the authors addressed challenges such as mode collapse and unstable training. Incorporating a reconstruction loss into the GAN loss function requires the generator to produce images reflecting the discriminator features, thus enhancing the generated images' quality to more accurately replicate authentic data patterns. Moreover, combining this loss function with other regularisation methods, such as gradient penalty, has proven to improve the performance of diverse GAN models further. A significant challenge in myocarditis diagnosis was the imbalance of classification, where one class dominated the other. To mitigate this, the authors introduced a focal loss-based training method that effectively trains the model on the minority class samples. The GAN-MD approach, evaluated on the Z-Alizadeh Sani myocarditis dataset, achieved superior results (F-measure 86.2%; geometric mean 91.0%) compared with other deep learning models and traditional machine learning methods.</p><p>In the paper entitled ‘DeepGCN based on variable multi-graph and multimodal data for ASD diagnosis’, Liu et al. constructed a deep graph convolutional network (GCN) based on variable multi-graph and multimodal data (VMM-DGCN) for ASD diagnosis. Firstly, the functional connectivity matrix was constructed to extract primary features. Then, the authors constructed a variable multi-graph construction strategy to capture the multi-scale feature representations of each subject utilising convolutional filters with varying kernel sizes. Furthermore, the authors brought the non-imaging information into the feature representation at each scale. They constructed multiple population graphs based on multimodal data by fully considering the correlation between subjects. After extracting the deeper features of population graphs using the deep GCN (DeepGCN), the authors fused the node features of multiple subgraphs to perform node classification tasks for typical control and ASD patients. The proposed algorithm was evaluated on the autism brain imaging data exchange I (ABIDE I) dataset, achieving an accuracy of 91.62% and an area under the curve value of 95.74%. These results demonstrated its outstanding performance compared to other ASD diagnostic algorithms.</p><p>In the paper entitled ‘Safety control strategy of spinal lamina cutting based on force and cutting depth signals’, Zhang et al. innovatively believed the identification of the bone tissue status was regarded as a time series classification task. The classification algorithm LSTM-FCN was used to process fusion signals composed of force and cutting depth signals, thus accurately classifying the lamina bone tissue status. In addition, it was verified that the accuracy of the proposed method could reach 98.85% in identifying the state of porcine spinal laminectomy. The maximum penetration distance can be controlled within 0.6 mm, which is safe and can be used in practice.</p>","PeriodicalId":46211,"journal":{"name":"CAAI Transactions on Intelligence Technology","volume":"9 4","pages":"787-789"},"PeriodicalIF":8.4000,"publicationDate":"2024-08-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1049/cit2.12364","citationCount":"0","resultStr":"{\"title\":\"Guest Editorial: Knowledge-based deep learning system in bio-medicine\",\"authors\":\"Yu-Dong Zhang, Juan Manuel Górriz\",\"doi\":\"10.1049/cit2.12364\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p>Numerous healthcare procedures can be viewed as medical sector decisions. In the modern era, computers have become indispensable in the realm of medical decision-making. However, the common view of computers in the medical field typically extends only to applications that support doctors in diagnosing diseases. To more tightly intertwine computers with the biomedical sciences, professionals are now more frequently utilising knowledge-driven deep learning systems (KDLS) and their foundational technologies, especially in the domain of neuroimaging (NI).</p><p>Data for medical purposes can be sourced from a variety of imaging techniques, including but not limited to Computed Tomography (CT), Magnetic Resonance Imaging (MRI), Ultrasound, Single Photon Emission Computed Tomography (SPECT), Positron Emission Tomography (PET), Magnetic Particle Imaging (MPI), Electroencephalography (EEG), Magnetoencephalography (MEG), Optical Microscopy and Tomography, Photoacoustic Tomography, Electron Tomography, and Atomic Force Microscopy.</p><p>Historically, these imaging techniques have been analysed using traditional statistical methods, such as hypothesis testing or Bayesian inference, which often presuppose certain conditions that are not always met. An emerging solution is the implementation of machine learning (ML) within the context of KDLS, allowing for the empirical mapping of complex, multi-dimensional relationships within data sets.</p><p>The objective of this special issue is to showcase the latest advancements in the methodology of KDLS for evaluating functional connectivity, neurological disorders, and clinical neuroscience, such as conditions like Alzheimer's, Parkinson's, cerebrovascular accidents, brain tumours, epilepsy, multiple sclerosis, ALS, Autism Spectrum Disorder, and more. Additionally, the special issue seeks to elucidate the mechanisms behind the predictive capabilities of ML methods within KDLS for brain-related diseases and disorders.</p><p>We received an abundance of submissions, totalling more than 40, from over 10 countries. After a meticulous and rigorous peer review process, which employed a double-blind methodology, we ultimately selected eight outstanding papers for publication. This process ensured the highest standards of quality and impartiality in the selection.</p><p>In the article ‘A deep learning fusion model for accurate classification of brain tumours in Magnetic Resonance images’, Zebari et al. created a robust deep learning (DL) fusion model for accurate brain tumour classification. To enhance performance, they employed data augmentation to expand the training dataset. The model leveraged VGG16, ResNet50, and convolutional deep belief networks to extract features from MRI images using a softmax classifier. By fusing features from two DL models, the fusion model notably boosted classification precision. Tested with a publicly available dataset, it achieved a remarkable 98.98% accuracy rate, outperforming existing methods.</p><p>In the article ‘Knowledge-based deep learning system for classifying’, Dhaygude et al. proposed a deep three-dimensional convolutional neural network incorporating multi-task learning and attention mechanisms. An upgraded primary C3D network is utilised to create rougher low-level feature maps. It introduces a new convolution block that focuses on the structural aspects of the magnetic resonance imaging image and another block that extracts attention weights unique to certain pixel positions in the feature map and multiplies them with the feature map output. Then, several fully connected layers are used to achieve multi-task learning, generating three outputs, including the primary classification task. The other two outputs employ backpropagation during training to improve the primary classification job. Experimental findings show that the authors' proposed method outperforms current approaches for classifying AD, achieving enhanced classification accuracy and other indicators on the Alzheimer's disease neuroimaging initiative dataset. The authors demonstrated promise for future disease classification studies.</p><p>In the paper entitled ‘A novel medical image data protection scheme for smart healthcare system’, Rehman et al. presented a lightweight medical image encryption scheme utilising bit-plane decomposition and chaos theory. The results of the experiment yielded entropy, energy, and correlation values of 7.999, 0.0156, and 0.0001, respectively. This validated the effectiveness of the encryption system proposed in this paper, which offered high-quality encryption, a large key space, key sensitivity, and resistance to statistical attacks.</p><p>In the paper entitled ‘Image super-resolution via dynamic network’, Tian et al. presented a dynamic network for image super-resolution (DSRNet), which contained a residual enhancement block, wide enhancement block, feature refinement block, and construction block. The residual enhancement block was composed of a residual enhanced architecture to facilitate hierarchical features for image super-resolution. To enhance the robustness of the obtained super-resolution model for complex scenes, a wide enhancement block achieved a dynamic architecture to learn more robust information to enhance the applicability of an obtained super-resolution model for varying scenes. To prevent interference of components in a wide enhancement block, a refinement block utilised a stacked architecture to learn obtained features accurately. Also, a residual learning operation was embedded in the refinement block to prevent long-term dependency problems. Finally, a construction block was responsible for reconstructing high-quality images. Designed heterogeneous architecture could not only facilitate richer structural information but also be lightweight, which was suitable for mobile digital devices. Experimental results showed that our method is more competitive in performance, recovering time of image super-resolution, and complexity. The code of DSRNet can be obtained at https://github.com/hellloxiaotian/DSRNet.</p><p>In the paper ‘Improved organs at risk segmentation based on modified U-Net with self-attention and consistency regularisation’, Manko et al. presented a new approach to automatic OAR segmentation in the chest cavity in Computed Tomography (CT) images. The proposed approach is based on the modified U-Net architecture with the ResNe-34 encoder, the baseline adopted in this work. The new two-branch CS-SA U-Net architecture is proposed, which consists of two parallel U-Net models in which self-attention blocks with cosine similarity as query-key similarity function (CS-SA) blocks are inserted between the encoder and decoder, which enabled the use of consistency regularisation. The proposed solution demonstrates state-of-the-art performance for the problem of OAR segmentation in CT images on the publicly available SegTHOR benchmark dataset in terms of a Dice coefficient (oesophagus-0.8714, heart-0.9516, trachea-0.9286, and aorta-0.9510) and Hausdorff distance (oesophagus-0.2541, heart-0.1514, trachea-0.1722, and aorta-0.1114) and significantly outperforms the baseline. The current approach was demonstrated to be viable for improving the quality of OAR segmentation for radiotherapy planning.</p><p>In the paper ‘GAN-MD: A myocarditis detection using multi-channel’, Golilarz et al. suggested a novel technique for improving the classification precision of CNNs. Using generative adversarial networks (GANs) to create synthetic images for data augmentation, the authors addressed challenges such as mode collapse and unstable training. Incorporating a reconstruction loss into the GAN loss function requires the generator to produce images reflecting the discriminator features, thus enhancing the generated images' quality to more accurately replicate authentic data patterns. Moreover, combining this loss function with other regularisation methods, such as gradient penalty, has proven to improve the performance of diverse GAN models further. A significant challenge in myocarditis diagnosis was the imbalance of classification, where one class dominated the other. To mitigate this, the authors introduced a focal loss-based training method that effectively trains the model on the minority class samples. The GAN-MD approach, evaluated on the Z-Alizadeh Sani myocarditis dataset, achieved superior results (F-measure 86.2%; geometric mean 91.0%) compared with other deep learning models and traditional machine learning methods.</p><p>In the paper entitled ‘DeepGCN based on variable multi-graph and multimodal data for ASD diagnosis’, Liu et al. constructed a deep graph convolutional network (GCN) based on variable multi-graph and multimodal data (VMM-DGCN) for ASD diagnosis. Firstly, the functional connectivity matrix was constructed to extract primary features. Then, the authors constructed a variable multi-graph construction strategy to capture the multi-scale feature representations of each subject utilising convolutional filters with varying kernel sizes. Furthermore, the authors brought the non-imaging information into the feature representation at each scale. They constructed multiple population graphs based on multimodal data by fully considering the correlation between subjects. After extracting the deeper features of population graphs using the deep GCN (DeepGCN), the authors fused the node features of multiple subgraphs to perform node classification tasks for typical control and ASD patients. The proposed algorithm was evaluated on the autism brain imaging data exchange I (ABIDE I) dataset, achieving an accuracy of 91.62% and an area under the curve value of 95.74%. These results demonstrated its outstanding performance compared to other ASD diagnostic algorithms.</p><p>In the paper entitled ‘Safety control strategy of spinal lamina cutting based on force and cutting depth signals’, Zhang et al. innovatively believed the identification of the bone tissue status was regarded as a time series classification task. The classification algorithm LSTM-FCN was used to process fusion signals composed of force and cutting depth signals, thus accurately classifying the lamina bone tissue status. In addition, it was verified that the accuracy of the proposed method could reach 98.85% in identifying the state of porcine spinal laminectomy. The maximum penetration distance can be controlled within 0.6 mm, which is safe and can be used in practice.</p>\",\"PeriodicalId\":46211,\"journal\":{\"name\":\"CAAI Transactions on Intelligence Technology\",\"volume\":\"9 4\",\"pages\":\"787-789\"},\"PeriodicalIF\":8.4000,\"publicationDate\":\"2024-08-09\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://onlinelibrary.wiley.com/doi/epdf/10.1049/cit2.12364\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"CAAI Transactions on Intelligence Technology\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://onlinelibrary.wiley.com/doi/10.1049/cit2.12364\",\"RegionNum\":2,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"CAAI Transactions on Intelligence Technology","FirstCategoryId":"94","ListUrlMain":"https://onlinelibrary.wiley.com/doi/10.1049/cit2.12364","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
摘要
许多医疗保健程序都可以被视为医疗部门的决策。在现代,计算机已成为医疗决策领域不可或缺的工具。然而,通常人们对计算机在医疗领域的应用仅限于辅助医生诊断疾病。为了将计算机与生物医学更紧密地结合在一起,专业人士现在更频繁地使用知识驱动的深度学习系统(KDLS)及其基础技术,尤其是在神经成像(NI)领域。用于医疗目的的数据可以来自各种成像技术,包括但不限于计算机断层扫描(CT)、磁共振成像(MRI)、超声波、单光子发射计算机断层扫描(SPECT)、正电子发射计算机断层扫描(PET)、磁粉成像(MPI)、脑电图(EEG)、脑磁图(MEG)、光学显微镜和断层扫描、光声断层扫描、电子断层扫描和原子力显微镜。一直以来,这些成像技术都是采用传统的统计方法进行分析的,如假设检验或贝叶斯推理,而这些方法往往预先假定了某些条件,但这些条件并非总能得到满足。本特刊旨在展示 KDLS 在评估功能连接、神经系统疾病和临床神经科学方面的最新进展,例如阿尔茨海默病、帕金森病、脑血管意外、脑肿瘤、癫痫、多发性硬化、渐冻人症、自闭症等疾病。此外,该特刊还试图阐明 KDLS 中的 ML 方法对脑相关疾病和障碍的预测能力背后的机制。我们收到了来自 10 多个国家的大量投稿,共计 40 多篇。我们收到了来自 10 多个国家的大量投稿,共计 40 篇。经过细致严格的同行评审(采用双盲方法),我们最终选出了 8 篇优秀论文予以发表。在《用于磁共振图像中脑肿瘤精确分类的深度学习融合模型》一文中,Zebari 等人创建了一个强大的深度学习(DL)融合模型,用于精确的脑肿瘤分类。为了提高性能,他们采用了数据增强技术来扩展训练数据集。该模型利用 VGG16、ResNet50 和卷积深度信念网络,使用 softmax 分类器从 MRI 图像中提取特征。通过融合两个 DL 模型的特征,融合模型显著提高了分类精度。在《基于知识的深度学习分类系统》一文中,Dhaygude 等人提出了一种融合了多任务学习和注意力机制的深度三维卷积神经网络。他们利用升级后的初级 C3D 网络来创建更粗糙的底层特征图。它引入了一个新的卷积块,重点关注磁共振成像图像的结构方面,另一个卷积块则提取特征图中某些像素位置特有的注意力权重,并与特征图输出相乘。然后,使用多个全连接层实现多任务学习,产生三个输出,包括主要分类任务。另外两个输出在训练过程中采用反向传播,以改进主要分类工作。实验结果表明,作者提出的方法优于当前的 AD 分类方法,在阿尔茨海默病神经影像倡议数据集上实现了更高的分类准确率和其他指标。在题为 "A novel medical image data protection scheme for smart healthcare system "的论文中,Rehman 等人提出了一种利用位平面分解和混沌理论的轻量级医学图像加密方案。实验结果表明,该方案的熵值为 7.999,能量为 0.0156,相关性为 0.0001。在题为 "通过动态网络实现图像超分辨率 "的论文中,Tian 等人提出了一种用于图像超分辨率的动态网络(DSRNet),它包含残差增强块、宽增强块、特征细化块和构造块。
Guest Editorial: Knowledge-based deep learning system in bio-medicine
Numerous healthcare procedures can be viewed as medical sector decisions. In the modern era, computers have become indispensable in the realm of medical decision-making. However, the common view of computers in the medical field typically extends only to applications that support doctors in diagnosing diseases. To more tightly intertwine computers with the biomedical sciences, professionals are now more frequently utilising knowledge-driven deep learning systems (KDLS) and their foundational technologies, especially in the domain of neuroimaging (NI).
Data for medical purposes can be sourced from a variety of imaging techniques, including but not limited to Computed Tomography (CT), Magnetic Resonance Imaging (MRI), Ultrasound, Single Photon Emission Computed Tomography (SPECT), Positron Emission Tomography (PET), Magnetic Particle Imaging (MPI), Electroencephalography (EEG), Magnetoencephalography (MEG), Optical Microscopy and Tomography, Photoacoustic Tomography, Electron Tomography, and Atomic Force Microscopy.
Historically, these imaging techniques have been analysed using traditional statistical methods, such as hypothesis testing or Bayesian inference, which often presuppose certain conditions that are not always met. An emerging solution is the implementation of machine learning (ML) within the context of KDLS, allowing for the empirical mapping of complex, multi-dimensional relationships within data sets.
The objective of this special issue is to showcase the latest advancements in the methodology of KDLS for evaluating functional connectivity, neurological disorders, and clinical neuroscience, such as conditions like Alzheimer's, Parkinson's, cerebrovascular accidents, brain tumours, epilepsy, multiple sclerosis, ALS, Autism Spectrum Disorder, and more. Additionally, the special issue seeks to elucidate the mechanisms behind the predictive capabilities of ML methods within KDLS for brain-related diseases and disorders.
We received an abundance of submissions, totalling more than 40, from over 10 countries. After a meticulous and rigorous peer review process, which employed a double-blind methodology, we ultimately selected eight outstanding papers for publication. This process ensured the highest standards of quality and impartiality in the selection.
In the article ‘A deep learning fusion model for accurate classification of brain tumours in Magnetic Resonance images’, Zebari et al. created a robust deep learning (DL) fusion model for accurate brain tumour classification. To enhance performance, they employed data augmentation to expand the training dataset. The model leveraged VGG16, ResNet50, and convolutional deep belief networks to extract features from MRI images using a softmax classifier. By fusing features from two DL models, the fusion model notably boosted classification precision. Tested with a publicly available dataset, it achieved a remarkable 98.98% accuracy rate, outperforming existing methods.
In the article ‘Knowledge-based deep learning system for classifying’, Dhaygude et al. proposed a deep three-dimensional convolutional neural network incorporating multi-task learning and attention mechanisms. An upgraded primary C3D network is utilised to create rougher low-level feature maps. It introduces a new convolution block that focuses on the structural aspects of the magnetic resonance imaging image and another block that extracts attention weights unique to certain pixel positions in the feature map and multiplies them with the feature map output. Then, several fully connected layers are used to achieve multi-task learning, generating three outputs, including the primary classification task. The other two outputs employ backpropagation during training to improve the primary classification job. Experimental findings show that the authors' proposed method outperforms current approaches for classifying AD, achieving enhanced classification accuracy and other indicators on the Alzheimer's disease neuroimaging initiative dataset. The authors demonstrated promise for future disease classification studies.
In the paper entitled ‘A novel medical image data protection scheme for smart healthcare system’, Rehman et al. presented a lightweight medical image encryption scheme utilising bit-plane decomposition and chaos theory. The results of the experiment yielded entropy, energy, and correlation values of 7.999, 0.0156, and 0.0001, respectively. This validated the effectiveness of the encryption system proposed in this paper, which offered high-quality encryption, a large key space, key sensitivity, and resistance to statistical attacks.
In the paper entitled ‘Image super-resolution via dynamic network’, Tian et al. presented a dynamic network for image super-resolution (DSRNet), which contained a residual enhancement block, wide enhancement block, feature refinement block, and construction block. The residual enhancement block was composed of a residual enhanced architecture to facilitate hierarchical features for image super-resolution. To enhance the robustness of the obtained super-resolution model for complex scenes, a wide enhancement block achieved a dynamic architecture to learn more robust information to enhance the applicability of an obtained super-resolution model for varying scenes. To prevent interference of components in a wide enhancement block, a refinement block utilised a stacked architecture to learn obtained features accurately. Also, a residual learning operation was embedded in the refinement block to prevent long-term dependency problems. Finally, a construction block was responsible for reconstructing high-quality images. Designed heterogeneous architecture could not only facilitate richer structural information but also be lightweight, which was suitable for mobile digital devices. Experimental results showed that our method is more competitive in performance, recovering time of image super-resolution, and complexity. The code of DSRNet can be obtained at https://github.com/hellloxiaotian/DSRNet.
In the paper ‘Improved organs at risk segmentation based on modified U-Net with self-attention and consistency regularisation’, Manko et al. presented a new approach to automatic OAR segmentation in the chest cavity in Computed Tomography (CT) images. The proposed approach is based on the modified U-Net architecture with the ResNe-34 encoder, the baseline adopted in this work. The new two-branch CS-SA U-Net architecture is proposed, which consists of two parallel U-Net models in which self-attention blocks with cosine similarity as query-key similarity function (CS-SA) blocks are inserted between the encoder and decoder, which enabled the use of consistency regularisation. The proposed solution demonstrates state-of-the-art performance for the problem of OAR segmentation in CT images on the publicly available SegTHOR benchmark dataset in terms of a Dice coefficient (oesophagus-0.8714, heart-0.9516, trachea-0.9286, and aorta-0.9510) and Hausdorff distance (oesophagus-0.2541, heart-0.1514, trachea-0.1722, and aorta-0.1114) and significantly outperforms the baseline. The current approach was demonstrated to be viable for improving the quality of OAR segmentation for radiotherapy planning.
In the paper ‘GAN-MD: A myocarditis detection using multi-channel’, Golilarz et al. suggested a novel technique for improving the classification precision of CNNs. Using generative adversarial networks (GANs) to create synthetic images for data augmentation, the authors addressed challenges such as mode collapse and unstable training. Incorporating a reconstruction loss into the GAN loss function requires the generator to produce images reflecting the discriminator features, thus enhancing the generated images' quality to more accurately replicate authentic data patterns. Moreover, combining this loss function with other regularisation methods, such as gradient penalty, has proven to improve the performance of diverse GAN models further. A significant challenge in myocarditis diagnosis was the imbalance of classification, where one class dominated the other. To mitigate this, the authors introduced a focal loss-based training method that effectively trains the model on the minority class samples. The GAN-MD approach, evaluated on the Z-Alizadeh Sani myocarditis dataset, achieved superior results (F-measure 86.2%; geometric mean 91.0%) compared with other deep learning models and traditional machine learning methods.
In the paper entitled ‘DeepGCN based on variable multi-graph and multimodal data for ASD diagnosis’, Liu et al. constructed a deep graph convolutional network (GCN) based on variable multi-graph and multimodal data (VMM-DGCN) for ASD diagnosis. Firstly, the functional connectivity matrix was constructed to extract primary features. Then, the authors constructed a variable multi-graph construction strategy to capture the multi-scale feature representations of each subject utilising convolutional filters with varying kernel sizes. Furthermore, the authors brought the non-imaging information into the feature representation at each scale. They constructed multiple population graphs based on multimodal data by fully considering the correlation between subjects. After extracting the deeper features of population graphs using the deep GCN (DeepGCN), the authors fused the node features of multiple subgraphs to perform node classification tasks for typical control and ASD patients. The proposed algorithm was evaluated on the autism brain imaging data exchange I (ABIDE I) dataset, achieving an accuracy of 91.62% and an area under the curve value of 95.74%. These results demonstrated its outstanding performance compared to other ASD diagnostic algorithms.
In the paper entitled ‘Safety control strategy of spinal lamina cutting based on force and cutting depth signals’, Zhang et al. innovatively believed the identification of the bone tissue status was regarded as a time series classification task. The classification algorithm LSTM-FCN was used to process fusion signals composed of force and cutting depth signals, thus accurately classifying the lamina bone tissue status. In addition, it was verified that the accuracy of the proposed method could reach 98.85% in identifying the state of porcine spinal laminectomy. The maximum penetration distance can be controlled within 0.6 mm, which is safe and can be used in practice.
期刊介绍:
CAAI Transactions on Intelligence Technology is a leading venue for original research on the theoretical and experimental aspects of artificial intelligence technology. We are a fully open access journal co-published by the Institution of Engineering and Technology (IET) and the Chinese Association for Artificial Intelligence (CAAI) providing research which is openly accessible to read and share worldwide.