首页 > 最新文献

Health Information Science and Systems最新文献

英文 中文
DPD (DePression Detection) Net: a deep neural network for multimodal depression detection. DPD(抑郁检测)网络:用于多模态抑郁检测的深度神经网络。
IF 3.4 3区 医学 Q1 MEDICAL INFORMATICS Pub Date : 2024-11-12 eCollection Date: 2024-12-01 DOI: 10.1007/s13755-024-00311-9
Manlu He, Erwin M Bakker, Michael S Lew

Depression is one of the most prevalent mental conditions which could impair people's productivity and lead to severe consequences. The diagnosis of this disease is complex as it often relies on a physician's subjective interview-based screening. The aim of our work is to propose deep learning models for automatic depression detection by using different data modalities, which could assist in the diagnosis of depression. Current works on automatic depression detection mostly are tested on a single dataset, which might lack robustness, flexibility and scalability. To alleviate this problem, we design a novel Graph Neural Network-enhanced Transformer model named DePressionDetect Net (DPD Net) that leverages textual, audio and visual features and can work under two different application settings: the clinical setting and the social media setting. The model consists of a unimodal encoder module for encoding single modality, a multimodal encoder module for integrating the multimodal information, and a detection module for producing the final prediction. We also propose a model named DePressionDetect-with-EEG Net (DPD-E Net) to incorporate Electroencephalography (EEG) signals and speech data for depression detection. Experiments across four benchmark datasets show that DPD Net and DPD-E Net can outperform the state-of-the-art models on three datasets (i.e., E-DAIC dataset, Twitter depression dataset and MODMA dataset), and achieve competitive performance on the fourth one (i.e., D-vlog dataset). Ablation studies demonstrate the advantages of the proposed modules and the effectiveness of combining diverse modalities for automatic depression detection.

抑郁症是最常见的精神疾病之一,会损害人们的工作效率并导致严重后果。这种疾病的诊断非常复杂,因为它通常依赖于医生基于访谈的主观筛查。我们的工作旨在通过使用不同的数据模式,为抑郁症的自动检测提出深度学习模型,从而为抑郁症的诊断提供帮助。目前的抑郁症自动检测工作大多在单一数据集上进行测试,可能缺乏鲁棒性、灵活性和可扩展性。为了缓解这一问题,我们设计了一种名为 "抑郁检测网络"(DePressionDetect Net,DPD Net)的新型图神经网络增强变换器模型,该模型利用文本、音频和视觉特征,可在两种不同的应用环境下工作:临床环境和社交媒体环境。该模型由用于编码单一模态的单模态编码器模块、用于整合多模态信息的多模态编码器模块和用于生成最终预测结果的检测模块组成。我们还提出了一个名为 "DePressionDetect-with-EEG Net"(DPD-E Net)的模型,用于结合脑电图(EEG)信号和语音数据进行抑郁检测。四个基准数据集的实验表明,DPD Net 和 DPD-E Net 在三个数据集(即 E-DAIC 数据集、Twitter 抑郁症数据集和 MODMA 数据集)上的表现优于最先进的模型,并在第四个数据集(即 D-vlog 数据集)上取得了具有竞争力的性能。消融研究证明了所提模块的优势,以及结合多种模式进行抑郁自动检测的有效性。
{"title":"DPD (DePression Detection) Net: a deep neural network for multimodal depression detection.","authors":"Manlu He, Erwin M Bakker, Michael S Lew","doi":"10.1007/s13755-024-00311-9","DOIUrl":"10.1007/s13755-024-00311-9","url":null,"abstract":"<p><p>Depression is one of the most prevalent mental conditions which could impair people's productivity and lead to severe consequences. The diagnosis of this disease is complex as it often relies on a physician's subjective interview-based screening. The aim of our work is to propose deep learning models for automatic depression detection by using different data modalities, which could assist in the diagnosis of depression. Current works on automatic depression detection mostly are tested on a single dataset, which might lack robustness, flexibility and scalability. To alleviate this problem, we design a novel Graph Neural Network-enhanced Transformer model named DePressionDetect Net (DPD Net) that leverages textual, audio and visual features and can work under two different application settings: the clinical setting and the social media setting. The model consists of a unimodal encoder module for encoding single modality, a multimodal encoder module for integrating the multimodal information, and a detection module for producing the final prediction. We also propose a model named DePressionDetect-with-EEG Net (DPD-E Net) to incorporate Electroencephalography (EEG) signals and speech data for depression detection. Experiments across four benchmark datasets show that DPD Net and DPD-E Net can outperform the state-of-the-art models on three datasets (i.e., E-DAIC dataset, Twitter depression dataset and MODMA dataset), and achieve competitive performance on the fourth one (i.e., D-vlog dataset). Ablation studies demonstrate the advantages of the proposed modules and the effectiveness of combining diverse modalities for automatic depression detection.</p>","PeriodicalId":46312,"journal":{"name":"Health Information Science and Systems","volume":"12 1","pages":"53"},"PeriodicalIF":3.4,"publicationDate":"2024-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11557813/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142630203","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multiple feature selection based on an optimization strategy for causal analysis of health data. 基于优化策略的多重特征选择,用于健康数据的因果分析。
IF 4.7 3区 医学 Q1 MEDICAL INFORMATICS Pub Date : 2024-11-12 eCollection Date: 2024-12-01 DOI: 10.1007/s13755-024-00312-8
Ruichen Cong, Ou Deng, Shoji Nishimura, Atsushi Ogihara, Qun Jin

Purpose: Recent advancements in information technology and wearable devices have revolutionized healthcare through health data analysis. Identifying significant relationships in complex health data enhances healthcare and public health strategies. In health analytics, causal graphs are important for investigating the relationships among health features. However, they face challenges owing to the large number of features, complexity, and computational demands. Feature selection methods are useful for addressing these challenges. In this paper, we present a framework for multiple feature selection based on an optimization strategy for causal analysis of health data.

Methods: We select multiple health features based on an optimization strategy. First, we define a Weighted Total Score (WTS) index to assess the feature importance after the combination of different feature selection methods. To explore an optimal set of weights for each method, we design a multiple feature selection algorithm integrated with the greedy algorithm. The features are then ranked according to their WTS, enabling selection of the most important ones. After that, causal graphs are constructed based on the selected features, and the statistical significance of the paths is assessed. Furthermore, evaluation experiments are conducted on an experiment dataset collected for this study and an open dataset for diabetes.

Results: The results demonstrate that our approach outperforms baseline models by reducing the number of features while improving model performance. Moreover, the statistical significance of the relationships between features uncovered through causal graphs is validated for both datasets.

Conclusion: By using the proposed framework for multiple feature selection based on an optimization strategy for causal analysis, the number of features is reduced and the causal relationships are uncovered and validated.

目的信息技术和可穿戴设备的最新进展通过健康数据分析彻底改变了医疗保健。从复杂的健康数据中找出重要的关系,有助于加强医疗保健和公共卫生战略。在健康分析中,因果图对于研究健康特征之间的关系非常重要。然而,由于特征数量大、复杂性高和计算要求高,它们面临着挑战。特征选择方法有助于应对这些挑战。在本文中,我们提出了一个基于优化策略的多特征选择框架,用于健康数据的因果分析:我们根据优化策略选择多个健康特征。首先,我们定义了一个加权总分(WTS)指数,用于评估不同特征选择方法组合后的特征重要性。为了探索每种方法的最佳权重集,我们设计了一种与贪婪算法相结合的多重特征选择算法。然后根据 WTS 对特征进行排序,从而选出最重要的特征。然后,根据所选特征构建因果图,并评估路径的统计意义。此外,我们还在为本研究收集的实验数据集和糖尿病公开数据集上进行了评估实验:结果表明,我们的方法在提高模型性能的同时减少了特征数量,从而优于基线模型。此外,通过因果图揭示的特征间关系的统计意义在两个数据集上都得到了验证:结论:通过使用基于因果分析优化策略的多特征选择框架,减少了特征数量,揭示并验证了因果关系。
{"title":"Multiple feature selection based on an optimization strategy for causal analysis of health data.","authors":"Ruichen Cong, Ou Deng, Shoji Nishimura, Atsushi Ogihara, Qun Jin","doi":"10.1007/s13755-024-00312-8","DOIUrl":"10.1007/s13755-024-00312-8","url":null,"abstract":"<p><strong>Purpose: </strong>Recent advancements in information technology and wearable devices have revolutionized healthcare through health data analysis. Identifying significant relationships in complex health data enhances healthcare and public health strategies. In health analytics, causal graphs are important for investigating the relationships among health features. However, they face challenges owing to the large number of features, complexity, and computational demands. Feature selection methods are useful for addressing these challenges. In this paper, we present a framework for multiple feature selection based on an optimization strategy for causal analysis of health data.</p><p><strong>Methods: </strong>We select multiple health features based on an optimization strategy. First, we define a Weighted Total Score (WTS) index to assess the feature importance after the combination of different feature selection methods. To explore an optimal set of weights for each method, we design a multiple feature selection algorithm integrated with the greedy algorithm. The features are then ranked according to their WTS, enabling selection of the most important ones. After that, causal graphs are constructed based on the selected features, and the statistical significance of the paths is assessed. Furthermore, evaluation experiments are conducted on an experiment dataset collected for this study and an open dataset for diabetes.</p><p><strong>Results: </strong>The results demonstrate that our approach outperforms baseline models by reducing the number of features while improving model performance. Moreover, the statistical significance of the relationships between features uncovered through causal graphs is validated for both datasets.</p><p><strong>Conclusion: </strong>By using the proposed framework for multiple feature selection based on an optimization strategy for causal analysis, the number of features is reduced and the causal relationships are uncovered and validated.</p>","PeriodicalId":46312,"journal":{"name":"Health Information Science and Systems","volume":"12 1","pages":"52"},"PeriodicalIF":4.7,"publicationDate":"2024-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11554952/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142630210","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Machine-learning-based prediction of cardiovascular events for hyperlipidemia population with lipid variability and remnant cholesterol as biomarkers. 基于机器学习的高脂血症人群心血管事件预测,以血脂变异性和残余胆固醇为生物标志物。
IF 3.4 3区 医学 Q1 MEDICAL INFORMATICS Pub Date : 2024-11-11 eCollection Date: 2024-12-01 DOI: 10.1007/s13755-024-00310-w
Zhenzhen Du, Shuang Wang, Ouzhou Yang, Juan He, Yujie Yang, Jing Zheng, Honglei Zhao, Yunpeng Cai

Purpose: Dyslipidemia poses a significant risk for the progression to cardiovascular diseases. Despite the identification of numerous risk factors and the proposal of various risk scales, there is still an urgent need for effective predictive models for the onset of cardiovascular diseases in the hyperlipidemic population, which are essential for the prevention of CVD.

Methods: We carried out a retrospective cohort study with 23,548 hyperlipidemia patients in Shenzhen Health Information Big Data Platform, including 11,723 CVD onset cases in a 3-year follow-up. The population was randomly divided into 70% as an independent training dataset and remaining 30% as test set. Four distinct machine-learning algorithms were implemented on the training dataset with the aim of developing highly accurate predictive models, and their performance was subsequently benchmarked against conventional risk assessment scales. An ablation study was also carried out to analyze the impact of individual risk factors to model performance.

Results: The non-linear algorithm, LightGBM, excelled in forecasting the incidence of cardiovascular disease within 3 years, achieving an area under the 'receiver operating characteristic curve' (AUROC) of 0.883. This performance surpassed that of the conventional logistic regression model, which had an AUROC of 0.725, on identical datasets. Concurrently, in direct comparative analyses, machine-learning approaches have notably outperformed the three traditional risk assessment methods within their respective applicable populations. These include the Framingham cardiovascular disease risk score, 2019 ESC/EAS guidelines for the management of dyslipidemia and the 2016 Chinese recommendations for the management of dyslipidemia in adults. Further analysis of risk factors showed that the variability of blood lipid levels and remnant cholesterol played an important role in indicating an increased risk of CVD.

Conclusions: We have shown that the application of machine-learning techniques significantly enhances the precision of cardiovascular risk forecasting among hyperlipidemic patients, addressing the critical issue of disease prediction's heterogeneity and non-linearity. Furthermore, some recently-suggested biomarkers, including blood lipid variability and remnant cholesterol are also important predictors of cardiovascular events, suggesting the importance of continuous lipid monitoring and healthcare profiling through big data platforms.

目的:血脂异常是引发心血管疾病的重要风险因素。尽管已经发现了许多危险因素,并提出了各种风险量表,但仍迫切需要建立有效的高脂血症人群心血管疾病发病预测模型,这对预防心血管疾病至关重要:我们在深圳市健康信息大数据平台上对 23,548 名高脂血症患者进行了回顾性队列研究,其中包括 3 年随访的 11,723 例心血管疾病发病病例。研究对象被随机分为70%作为独立的训练数据集,其余30%作为测试集。在训练数据集上实施了四种不同的机器学习算法,目的是开发出高度准确的预测模型,随后将其性能与传统的风险评估量表进行比较。此外,还进行了一项消融研究,以分析个别风险因素对模型性能的影响:结果:非线性算法 LightGBM 在预测 3 年内心血管疾病发病率方面表现出色,"接收者操作特征曲线 "下面积(AUROC)达到 0.883。在相同的数据集上,其性能超过了传统的逻辑回归模型,后者的接受者操作特征曲线下面积为 0.725。同时,在直接比较分析中,机器学习方法在各自适用人群中的表现明显优于三种传统风险评估方法。这些方法包括弗雷明汉心血管疾病风险评分、2019年ESC/EAS血脂异常管理指南和2016年中国成人血脂异常管理建议。对风险因素的进一步分析表明,血脂水平和残余胆固醇的变异性在表明心血管疾病风险增加方面起着重要作用:我们的研究表明,机器学习技术的应用大大提高了高脂血症患者心血管风险预测的准确性,解决了疾病预测的异质性和非线性这一关键问题。此外,最近提出的一些生物标志物,包括血脂变异性和残余胆固醇,也是心血管事件的重要预测指标,这表明通过大数据平台进行连续血脂监测和医疗保健分析的重要性。
{"title":"Machine-learning-based prediction of cardiovascular events for hyperlipidemia population with lipid variability and remnant cholesterol as biomarkers.","authors":"Zhenzhen Du, Shuang Wang, Ouzhou Yang, Juan He, Yujie Yang, Jing Zheng, Honglei Zhao, Yunpeng Cai","doi":"10.1007/s13755-024-00310-w","DOIUrl":"10.1007/s13755-024-00310-w","url":null,"abstract":"<p><strong>Purpose: </strong>Dyslipidemia poses a significant risk for the progression to cardiovascular diseases. Despite the identification of numerous risk factors and the proposal of various risk scales, there is still an urgent need for effective predictive models for the onset of cardiovascular diseases in the hyperlipidemic population, which are essential for the prevention of CVD.</p><p><strong>Methods: </strong>We carried out a retrospective cohort study with 23,548 hyperlipidemia patients in Shenzhen Health Information Big Data Platform, including 11,723 CVD onset cases in a 3-year follow-up. The population was randomly divided into 70% as an independent training dataset and remaining 30% as test set. Four distinct machine-learning algorithms were implemented on the training dataset with the aim of developing highly accurate predictive models, and their performance was subsequently benchmarked against conventional risk assessment scales. An ablation study was also carried out to analyze the impact of individual risk factors to model performance.</p><p><strong>Results: </strong>The non-linear algorithm, LightGBM, excelled in forecasting the incidence of cardiovascular disease within 3 years, achieving an area under the 'receiver operating characteristic curve' (AUROC) of 0.883. This performance surpassed that of the conventional logistic regression model, which had an AUROC of 0.725, on identical datasets. Concurrently, in direct comparative analyses, machine-learning approaches have notably outperformed the three traditional risk assessment methods within their respective applicable populations. These include the Framingham cardiovascular disease risk score, 2019 ESC/EAS guidelines for the management of dyslipidemia and the 2016 Chinese recommendations for the management of dyslipidemia in adults. Further analysis of risk factors showed that the variability of blood lipid levels and remnant cholesterol played an important role in indicating an increased risk of CVD.</p><p><strong>Conclusions: </strong>We have shown that the application of machine-learning techniques significantly enhances the precision of cardiovascular risk forecasting among hyperlipidemic patients, addressing the critical issue of disease prediction's heterogeneity and non-linearity. Furthermore, some recently-suggested biomarkers, including blood lipid variability and remnant cholesterol are also important predictors of cardiovascular events, suggesting the importance of continuous lipid monitoring and healthcare profiling through big data platforms.</p>","PeriodicalId":46312,"journal":{"name":"Health Information Science and Systems","volume":"12 1","pages":"51"},"PeriodicalIF":3.4,"publicationDate":"2024-11-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11551092/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142630206","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Machine learning approach to flare-up detection and clustering in chronic obstructive pulmonary disease (COPD) patients. 慢性阻塞性肺病(COPD)患者发作检测和聚类的机器学习方法。
IF 3.4 3区 医学 Q1 MEDICAL INFORMATICS Pub Date : 2024-10-23 eCollection Date: 2024-12-01 DOI: 10.1007/s13755-024-00308-4
Ramón Rueda, Esteban Fabello, Tatiana Silva, Samuel Genzor, Jan Mizera, Ladislav Stanke

Purpose: Chronic obstructive pulmonary disease (COPD) is a prevalent and preventable condition that typically worsens over time. Acute exacerbations of COPD significantly impact disease progression, underscoring the importance of prevention efforts. This observational study aimed to achieve two main objectives: (1) identify patients at risk of exacerbations using an ensemble of clustering algorithms, and (2) classify patients into distinct clusters based on disease severity.

Methods: Data from portable medical devices were analyzed post-hoc using hyperparameter optimization with Self-Organizing Maps (SOM), Density-Based Spatial Clustering of Applications with Noise (DBSCAN), Isolation Forest, and Support Vector Machine (SVM) algorithms, to detect flare-ups. Principal Component Analysis (PCA) followed by KMeans clustering was applied to categorize patients by severity.

Results: 25 patients were included within the study population, data from 17 patients had the required reliability. Five patients were identified in the highest deterioration group, with one clinically confirmed exacerbation accurately detected by our ensemble algorithm. Then, PCA and KMeans clustering grouped patients into three clusters based on severity: Cluster 0 started with the least severe characteristics but experienced decline, Cluster 1 consistently showed the most severe characteristics, and Cluster 2 showed slight improvement.

Conclusion: Our approach effectively identified patients at risk of exacerbations and classified them by disease severity. Although promising, the approach would need to be verified on a larger sample with a larger number of recorded clinically verified exacerbations.

目的:慢性阻塞性肺病(COPD)是一种可预防的常见疾病,通常会随着时间的推移而恶化。慢性阻塞性肺疾病的急性加重会严重影响疾病的进展,因此预防工作尤为重要。这项观察性研究旨在实现两个主要目标:(1) 使用聚类算法组合识别有恶化风险的患者;(2) 根据疾病严重程度将患者划分为不同的群组:使用自组织图(SOM)、基于密度的噪声应用空间聚类(DBSCAN)、隔离森林(Isolation Forest)和支持向量机(SVM)算法进行超参数优化,对便携式医疗设备的数据进行事后分析,以检测病情恶化。采用主成分分析法(PCA)和 KMeans 聚类法对患者的严重程度进行分类。有五名患者被确定为病情恶化程度最高的一组,我们的集合算法准确检测出了一名临床确诊的病情恶化患者。然后,PCA 和 KMeans 聚类法根据严重程度将患者分为三组:第 0 组开始时特征最不严重,但病情有所恶化;第 1 组持续表现出最严重的特征;第 2 组病情略有好转:我们的方法能有效识别有病情加重风险的患者,并根据病情严重程度对他们进行分类。虽然这种方法很有前景,但还需要在更大的样本中进行验证,并记录更多临床验证的病情加重情况。
{"title":"Machine learning approach to flare-up detection and clustering in chronic obstructive pulmonary disease (COPD) patients.","authors":"Ramón Rueda, Esteban Fabello, Tatiana Silva, Samuel Genzor, Jan Mizera, Ladislav Stanke","doi":"10.1007/s13755-024-00308-4","DOIUrl":"10.1007/s13755-024-00308-4","url":null,"abstract":"<p><strong>Purpose: </strong>Chronic obstructive pulmonary disease (COPD) is a prevalent and preventable condition that typically worsens over time. Acute exacerbations of COPD significantly impact disease progression, underscoring the importance of prevention efforts. This observational study aimed to achieve two main objectives: (1) identify patients at risk of exacerbations using an ensemble of clustering algorithms, and (2) classify patients into distinct clusters based on disease severity.</p><p><strong>Methods: </strong>Data from portable medical devices were analyzed post-hoc using hyperparameter optimization with Self-Organizing Maps (SOM), Density-Based Spatial Clustering of Applications with Noise (DBSCAN), Isolation Forest, and Support Vector Machine (SVM) algorithms, to detect flare-ups. Principal Component Analysis (PCA) followed by KMeans clustering was applied to categorize patients by severity.</p><p><strong>Results: </strong>25 patients were included within the study population, data from 17 patients had the required reliability. Five patients were identified in the highest deterioration group, with one clinically confirmed exacerbation accurately detected by our ensemble algorithm. Then, PCA and KMeans clustering grouped patients into three clusters based on severity: Cluster 0 started with the least severe characteristics but experienced decline, Cluster 1 consistently showed the most severe characteristics, and Cluster 2 showed slight improvement.</p><p><strong>Conclusion: </strong>Our approach effectively identified patients at risk of exacerbations and classified them by disease severity. Although promising, the approach would need to be verified on a larger sample with a larger number of recorded clinically verified exacerbations.</p>","PeriodicalId":46312,"journal":{"name":"Health Information Science and Systems","volume":"12 1","pages":"50"},"PeriodicalIF":3.4,"publicationDate":"2024-10-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11499475/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142516717","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Explainable federated learning scheme for secure healthcare data sharing. 用于安全共享医疗数据的可解释联合学习方案。
IF 3.4 3区 医学 Q1 MEDICAL INFORMATICS Pub Date : 2024-09-13 eCollection Date: 2024-12-01 DOI: 10.1007/s13755-024-00306-6
Liutao Zhao, Haoran Xie, Lin Zhong, Yujue Wang

Artificial intelligence has immense potential for applications in smart healthcare. Nowadays, a large amount of medical data collected by wearable or implantable devices has been accumulated in Body Area Networks. Unlocking the value of this data can better explore the applications of artificial intelligence in the smart healthcare field. To utilize these dispersed data, this paper proposes an innovative Federated Learning scheme, focusing on the challenges of explainability and security in smart healthcare. In the proposed scheme, the federated modeling process and explainability analysis are independent of each other. By introducing post-hoc explanation techniques to analyze the global model, the scheme avoids the performance degradation caused by pursuing explainability while understanding the mechanism of the model. In terms of security, firstly, a fair and efficient client private gradient evaluation method is introduced for explainable evaluation of gradient contributions, quantifying client contributions in federated learning and filtering the impact of low-quality data. Secondly, to address the privacy issues of medical health data collected by wireless Body Area Networks, a multi-server model is proposed to solve the secure aggregation problem in federated learning. Furthermore, by employing homomorphic secret sharing and homomorphic hashing techniques, a non-interactive, verifiable secure aggregation protocol is proposed, ensuring that client data privacy is protected and the correctness of the aggregation results is maintained even in the presence of up to t colluding malicious servers. Experimental results demonstrate that the proposed scheme's explainability is consistent with that of centralized training scenarios and shows competitive performance in terms of security and efficiency.

Graphical abstract:

人工智能在智能医疗领域的应用潜力巨大。如今,由可穿戴或植入式设备收集的大量医疗数据已在体域网络中积累起来。挖掘这些数据的价值可以更好地探索人工智能在智能医疗领域的应用。为了利用这些分散的数据,本文提出了一种创新的联盟学习方案,重点关注智能医疗领域中可解释性和安全性的挑战。在所提出的方案中,联合建模过程和可解释性分析是相互独立的。通过引入事后解释技术来分析全局模型,该方案避免了在理解模型机制的同时追求可解释性而导致的性能下降。在安全性方面,首先,针对梯度贡献的可解释性评估,引入了一种公平高效的客户端私有梯度评估方法,量化了联合学习中的客户端贡献,过滤了低质量数据的影响。其次,针对无线体域网收集的医疗健康数据的隐私问题,提出了一种多服务器模型,以解决联合学习中的安全聚合问题。此外,通过采用同态秘密共享和同态散列技术,提出了一种非交互式、可验证的安全聚合协议,确保客户端数据隐私得到保护,即使存在多达 t 个恶意串通的服务器,也能保持聚合结果的正确性。实验结果表明,所提方案的可解释性与集中式训练方案一致,并在安全性和效率方面表现出了竞争力:
{"title":"Explainable federated learning scheme for secure healthcare data sharing.","authors":"Liutao Zhao, Haoran Xie, Lin Zhong, Yujue Wang","doi":"10.1007/s13755-024-00306-6","DOIUrl":"10.1007/s13755-024-00306-6","url":null,"abstract":"<p><p>Artificial intelligence has immense potential for applications in smart healthcare. Nowadays, a large amount of medical data collected by wearable or implantable devices has been accumulated in Body Area Networks. Unlocking the value of this data can better explore the applications of artificial intelligence in the smart healthcare field. To utilize these dispersed data, this paper proposes an innovative Federated Learning scheme, focusing on the challenges of explainability and security in smart healthcare. In the proposed scheme, the federated modeling process and explainability analysis are independent of each other. By introducing post-hoc explanation techniques to analyze the global model, the scheme avoids the performance degradation caused by pursuing explainability while understanding the mechanism of the model. In terms of security, firstly, a fair and efficient client private gradient evaluation method is introduced for explainable evaluation of gradient contributions, quantifying client contributions in federated learning and filtering the impact of low-quality data. Secondly, to address the privacy issues of medical health data collected by wireless Body Area Networks, a multi-server model is proposed to solve the secure aggregation problem in federated learning. Furthermore, by employing homomorphic secret sharing and homomorphic hashing techniques, a non-interactive, verifiable secure aggregation protocol is proposed, ensuring that client data privacy is protected and the correctness of the aggregation results is maintained even in the presence of up to <i>t</i> colluding malicious servers. Experimental results demonstrate that the proposed scheme's explainability is consistent with that of centralized training scenarios and shows competitive performance in terms of security and efficiency.</p><p><strong>Graphical abstract: </strong></p>","PeriodicalId":46312,"journal":{"name":"Health Information Science and Systems","volume":"12 1","pages":"49"},"PeriodicalIF":3.4,"publicationDate":"2024-09-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11399375/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142298293","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Comorbidity progression analysis: patient stratification and comorbidity prediction using temporal comorbidity network. 合并症进展分析:利用时间合并症网络对患者进行分层和合并症预测。
IF 3.4 3区 医学 Q1 MEDICAL INFORMATICS Pub Date : 2024-09-12 eCollection Date: 2024-12-01 DOI: 10.1007/s13755-024-00307-5
Ye Liang, Chonghui Guo, Hailin Li

Objective: The study aims to identify distinct population-specific comorbidity progression patterns, timely detect potential comorbidities, and gain better understanding of the progression of comorbid conditions among patients.

Methods: This work presents a comorbidity progression analysis framework that utilizes temporal comorbidity networks (TCN) for patient stratification and comorbidity prediction. We propose a TCN construction approach that utilizes longitudinal, temporal diagnosis data of patients to construct their TCN. Subsequently, we employ the TCN for patient stratification by conducting preliminary analysis, and typical prescription analysis to uncover potential comorbidity progression patterns in different patient groups. Finally, we propose an innovative comorbidity prediction method by utilizing the distance-matched temporal comorbidity network (TCN-DM). This method identifies similar patients with disease prevalence and disease transition patterns and combines their diagnosis information with that of the current patient to predict potential comorbidity at the patient's next visit.

Results: This study validated the capability of the framework using a real-world dataset MIMIC-III, with heart failure (HF) as interested disease to investigate comorbidity progression in HF patients. With TCN, this study can identify four significant distinctive HF subgroups, revealing the progression of comorbidities in patients. Furthermore, compared to other methods, TCN-DM demonstrated better predictive performance with F1-Score values ranging from 0.454 to 0.612, showcasing its superiority.

Conclusions: This study can identify comorbidity patterns for individuals and population, and offer promising prediction for future comorbidity developments in patients.

研究目的本研究旨在识别特定人群的独特合并症进展模式,及时发现潜在合并症,并更好地了解患者合并症的进展情况:本研究提出了一种合并症进展分析框架,该框架利用时间合并症网络(TCN)对患者进行分层和合并症预测。我们提出了一种 TCN 构建方法,利用患者的纵向时间诊断数据来构建他们的 TCN。随后,我们通过进行初步分析和典型处方分析,利用 TCN 对患者进行分层,从而发现不同患者群体中潜在的合并症进展模式。最后,我们利用距离匹配时间合并症网络(TCN-DM)提出了一种创新的合并症预测方法。该方法可识别具有疾病流行和疾病转变模式的类似患者,并将其诊断信息与当前患者的诊断信息相结合,以预测患者下次就诊时的潜在合并症:本研究利用真实世界数据集 MIMIC-III,以心力衰竭(HF)为相关疾病,对该框架的能力进行了验证,以调查 HF 患者的合并症进展情况。通过 TCN,本研究可以识别出四个明显的 HF 亚组,揭示出患者合并症的进展情况。此外,与其他方法相比,TCN-DM 的预测性能更好,F1-Score 值从 0.454 到 0.612 不等,显示了其优越性:本研究可识别个人和人群的合并症模式,并为预测患者未来的合并症发展提供了前景。
{"title":"Comorbidity progression analysis: patient stratification and comorbidity prediction using temporal comorbidity network.","authors":"Ye Liang, Chonghui Guo, Hailin Li","doi":"10.1007/s13755-024-00307-5","DOIUrl":"10.1007/s13755-024-00307-5","url":null,"abstract":"<p><strong>Objective: </strong>The study aims to identify distinct population-specific comorbidity progression patterns, timely detect potential comorbidities, and gain better understanding of the progression of comorbid conditions among patients.</p><p><strong>Methods: </strong>This work presents a comorbidity progression analysis framework that utilizes temporal comorbidity networks (TCN) for patient stratification and comorbidity prediction. We propose a TCN construction approach that utilizes longitudinal, temporal diagnosis data of patients to construct their TCN. Subsequently, we employ the TCN for patient stratification by conducting preliminary analysis, and typical prescription analysis to uncover potential comorbidity progression patterns in different patient groups. Finally, we propose an innovative comorbidity prediction method by utilizing the distance-matched temporal comorbidity network (TCN-DM). This method identifies similar patients with disease prevalence and disease transition patterns and combines their diagnosis information with that of the current patient to predict potential comorbidity at the patient's next visit.</p><p><strong>Results: </strong>This study validated the capability of the framework using a real-world dataset MIMIC-III, with heart failure (HF) as interested disease to investigate comorbidity progression in HF patients. With TCN, this study can identify four significant distinctive HF subgroups, revealing the progression of comorbidities in patients. Furthermore, compared to other methods, TCN-DM demonstrated better predictive performance with F1-Score values ranging from 0.454 to 0.612, showcasing its superiority.</p><p><strong>Conclusions: </strong>This study can identify comorbidity patterns for individuals and population, and offer promising prediction for future comorbidity developments in patients.</p>","PeriodicalId":46312,"journal":{"name":"Health Information Science and Systems","volume":"12 1","pages":"48"},"PeriodicalIF":3.4,"publicationDate":"2024-09-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11393239/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142298292","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Explainable depression symptom detection in social media. 社交媒体中可解释的抑郁症状检测。
IF 4.7 3区 医学 Q1 MEDICAL INFORMATICS Pub Date : 2024-09-06 eCollection Date: 2024-12-01 DOI: 10.1007/s13755-024-00303-9
Eliseo Bao, Anxo Pérez, Javier Parapar

Users of social platforms often perceive these sites as supportive spaces to post about their mental health issues. Those conversations contain important traces about individuals' health risks. Recently, researchers have exploited this online information to construct mental health detection models, which aim to identify users at risk on platforms like Twitter, Reddit or Facebook. Most of these models are focused on achieving good classification results, ignoring the explainability and interpretability of the decisions. Recent research has pointed out the importance of using clinical markers, such as the use of symptoms, to improve trust in the computational models by health professionals. In this paper, we introduce transformer-based architectures designed to detect and explain the appearance of depressive symptom markers in user-generated content from social media. We present two approaches: (i) train a model to classify, and another one to explain the classifier's decision separately and (ii) unify the two tasks simultaneously within a single model. Additionally, for this latter manner, we also investigated the performance of recent conversational Large Language Models (LLMs) utilizing both in-context learning and finetuning. Our models provide natural language explanations, aligning with validated symptoms, thus enabling clinicians to interpret the decisions more effectively. We evaluate our approaches using recent symptom-focused datasets, using both offline metrics and expert-in-the-loop evaluations to assess the quality of our models' explanations. Our findings demonstrate that it is possible to achieve good classification results while generating interpretable symptom-based explanations.

社交平台的用户通常将这些网站视为发布心理健康问题的支持性空间。这些对话包含了有关个人健康风险的重要痕迹。最近,研究人员利用这些在线信息构建了心理健康检测模型,旨在识别 Twitter、Reddit 或 Facebook 等平台上的风险用户。这些模型大多专注于实现良好的分类结果,而忽略了决策的可解释性和可解读性。最近的研究指出,使用临床标记(如使用症状)来提高医疗专业人员对计算模型的信任度非常重要。在本文中,我们介绍了基于转换器的架构,旨在检测和解释社交媒体用户生成内容中出现的抑郁症状标记。我们提出了两种方法:(i) 分别训练一个模型来分类,另一个模型来解释分类器的决定;(ii) 在一个模型中同时统一这两项任务。此外,对于后一种方式,我们还利用上下文学习和微调研究了近期会话大语言模型(LLM)的性能。我们的模型提供自然语言解释,并与经过验证的症状保持一致,从而使临床医生能够更有效地解释决定。我们使用最近的症状数据集对我们的方法进行了评估,使用离线度量和专家在环评估来评估模型解释的质量。我们的研究结果表明,在生成可解释的基于症状的解释的同时实现良好的分类结果是可能的。
{"title":"Explainable depression symptom detection in social media.","authors":"Eliseo Bao, Anxo Pérez, Javier Parapar","doi":"10.1007/s13755-024-00303-9","DOIUrl":"10.1007/s13755-024-00303-9","url":null,"abstract":"<p><p>Users of social platforms often perceive these sites as supportive spaces to post about their mental health issues. Those conversations contain important traces about individuals' health risks. Recently, researchers have exploited this online information to construct mental health detection models, which aim to identify users at risk on platforms like Twitter, Reddit or Facebook. Most of these models are focused on achieving good classification results, ignoring the explainability and interpretability of the decisions. Recent research has pointed out the importance of using clinical markers, such as the use of symptoms, to improve trust in the computational models by health professionals. In this paper, we introduce transformer-based architectures designed to detect and explain the appearance of depressive symptom markers in user-generated content from social media. We present two approaches: (i) train a model to classify, and another one to explain the classifier's decision separately and (ii) unify the two tasks simultaneously within a single model. Additionally, for this latter manner, we also investigated the performance of recent conversational Large Language Models (LLMs) utilizing both in-context learning and finetuning. Our models provide natural language explanations, aligning with validated symptoms, thus enabling clinicians to interpret the decisions more effectively. We evaluate our approaches using recent symptom-focused datasets, using both offline metrics and expert-in-the-loop evaluations to assess the quality of our models' explanations. Our findings demonstrate that it is possible to achieve good classification results while generating interpretable symptom-based explanations.</p>","PeriodicalId":46312,"journal":{"name":"Health Information Science and Systems","volume":"12 1","pages":"47"},"PeriodicalIF":4.7,"publicationDate":"2024-09-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11379836/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142156307","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A lightweight network based on multi-feature pseudo-color mapping for arrhythmia recognition. 基于多特征伪彩色映射的轻量级心律失常识别网络。
IF 3.4 3区 医学 Q1 MEDICAL INFORMATICS Pub Date : 2024-09-04 eCollection Date: 2024-12-01 DOI: 10.1007/s13755-024-00304-8
Yijun Ma, Junyan Li, Jinbiao Zhang, Jilin Wang, Guozhen Sun, Yatao Zhang

Heartbeats classification is a crucial tool for arrhythmia diagnosis. In this study, a multi-feature pseudo-color mapping (MfPc Mapping) was proposed, and a lightweight FlexShuffleNet was designed to classify heartbeats. MfPc Mapping converts one-dimensional (1-D) electrocardiogram (ECG) recordings into corresponding two-dimensional (2-D) multi-feature RGB graphs, and it offers good excellent interpretability and data visualization. FlexShuffleNet is a lightweight network that can be adapted to classification tasks of varying complexity by tuning hyperparameters. The method has three steps. The first step is data preprocessing, which includes de-noising the raw ECG recordings, removing baseline drift, extracting heartbeats, and performing data balancing, the second step is transforming the heartbeats using MfPc Mapping. Finally, the FlexShuffleNet is employed to classify heartbeats into 14 categories. This study was evaluated on the test set of the MIT-BIH arrhythmia database (MIT/BIH DB), and it yielded the results i.e., accuracy of 99.77%, sensitivity of 94.60%, precision of 89.83% and specificity of 99.85% and F1-score of 0.9125 in 14-category classification task. Additionally, validation on Shandong Province Hospital database (SPH DB) yielded the results i.e., accuracy of 92.08%, sensitivity of 93.63%, precision of 91.25% and specificity of 99.85% and F1-score of 0.9315. The results show the satisfied performance of the proposed method.

心跳分类是心律失常诊断的重要工具。本研究提出了一种多特征伪彩色映射(MfPc Mapping),并设计了一个轻量级的 FlexShuffleNet 来对心跳进行分类。MfPc Mapping 可将一维(1-D)心电图(ECG)记录转换成相应的二维(2-D)多特征 RGB 图形,具有良好的可解释性和数据可视化。FlexShuffleNet 是一种轻量级网络,可通过调整超参数适应不同复杂度的分类任务。该方法分为三个步骤。第一步是数据预处理,包括对原始心电图记录进行去噪、去除基线漂移、提取心搏和进行数据平衡;第二步是使用 MfPc 映射转换心搏。最后,使用 FlexShuffleNet 将心跳分为 14 类。这项研究在 MIT-BIH 心律失常数据库(MIT/BIH DB)的测试集上进行了评估,结果显示,在 14 类分类任务中,准确率为 99.77%,灵敏度为 94.60%,精确度为 89.83%,特异性为 99.85%,F1 分数为 0.9125。此外,山东省医院数据库(SPH DB)的验证结果为:准确率 92.08%,灵敏度 93.63%,精确度 91.25%,特异性 99.85%,F1 分数 0.9315。这些结果表明,拟议方法的性能令人满意。
{"title":"A lightweight network based on multi-feature pseudo-color mapping for arrhythmia recognition.","authors":"Yijun Ma, Junyan Li, Jinbiao Zhang, Jilin Wang, Guozhen Sun, Yatao Zhang","doi":"10.1007/s13755-024-00304-8","DOIUrl":"10.1007/s13755-024-00304-8","url":null,"abstract":"<p><p>Heartbeats classification is a crucial tool for arrhythmia diagnosis. In this study, a multi-feature pseudo-color mapping (MfPc Mapping) was proposed, and a lightweight FlexShuffleNet was designed to classify heartbeats. MfPc Mapping converts one-dimensional (1-D) electrocardiogram (ECG) recordings into corresponding two-dimensional (2-D) multi-feature RGB graphs, and it offers good excellent interpretability and data visualization. FlexShuffleNet is a lightweight network that can be adapted to classification tasks of varying complexity by tuning hyperparameters. The method has three steps. The first step is data preprocessing, which includes de-noising the raw ECG recordings, removing baseline drift, extracting heartbeats, and performing data balancing, the second step is transforming the heartbeats using MfPc Mapping. Finally, the FlexShuffleNet is employed to classify heartbeats into 14 categories. This study was evaluated on the test set of the MIT-BIH arrhythmia database (MIT/BIH DB), and it yielded the results i.e., accuracy of 99.77%, sensitivity of 94.60%, precision of 89.83% and specificity of 99.85% and F1-score of 0.9125 in 14-category classification task. Additionally, validation on Shandong Province Hospital database (SPH DB) yielded the results i.e., accuracy of 92.08%, sensitivity of 93.63%, precision of 91.25% and specificity of 99.85% and F1-score of 0.9315. The results show the satisfied performance of the proposed method.</p>","PeriodicalId":46312,"journal":{"name":"Health Information Science and Systems","volume":"12 1","pages":"46"},"PeriodicalIF":3.4,"publicationDate":"2024-09-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11371975/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142141311","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Tree hole rescue: an AI approach for suicide risk detection and online suicide intervention. 树洞救援:一种用于自杀风险检测和在线自杀干预的人工智能方法。
IF 3.4 3区 医学 Q1 MEDICAL INFORMATICS Pub Date : 2024-09-03 eCollection Date: 2024-12-01 DOI: 10.1007/s13755-024-00298-3
Zhisheng Huang, Qing Hu

Adolescent suicide has become an important social issue of general concern. Many young people express their suicidal feelings and intentions through online social media, e.g., Twitter, Microblog. The "tree hole" is the Chinese name for places on the Web where people post secrets. It provides the possibility of using Artificial Intelligence and big data technology to detect the posts where someone express the suicidal signal from those "tree hole" social media. We have developed the Web-based intelligent agents (i.e., AI-based programs) which can monitor the "tree hole" websites in Microblog every day by using knowledge graph technology. We have organized Tree-hole Rescue Team, which consists of more than 1000 volunteers, to carry out suicide rescue intervention according to the daily monitoring notifications. From 2018 to 2023, Tree-hole Rescue Team has prevented more than 6600 suicides. A few thousands of people have been saved within those 6 years. In this paper, we present the basic technology of Web-based Tree Hole intelligent agents and elaborate how the intelligent agents can discover suicide attempts and issue corresponding monitoring notifications and how the volunteers of Tree Hole Rescue Team can conduct online suicide intervention. This research also shows that the knowledge graph approach can be used for the semantic analysis on social media.

青少年自杀已成为人们普遍关注的重要社会问题。许多青少年通过网络社交媒体,如微博、微信等,表达自己的自杀情绪和意向。树洞 "是网络上人们发布秘密的地方的中文名称。这为利用人工智能和大数据技术从这些 "树洞 "社交媒体中检测出有人表达自杀信号的帖子提供了可能。我们开发了基于网络的智能代理(即基于人工智能的程序),利用知识图谱技术每天监测微博中的 "树洞 "网站。我们组织了由 1000 多名志愿者组成的 "树洞救援队",根据每天的监测通知进行自杀救援干预。从 2018 年到 2023 年,树洞救援队已经阻止了 6600 多起自杀事件。在这 6 年中,有数千人被拯救。本文介绍了基于网络的树洞智能代理的基本技术,阐述了智能代理如何发现自杀企图并发出相应的监测通知,以及树洞救援队的志愿者如何进行在线自杀干预。这项研究还表明,知识图谱方法可用于社交媒体的语义分析。
{"title":"Tree hole rescue: an AI approach for suicide risk detection and online suicide intervention.","authors":"Zhisheng Huang, Qing Hu","doi":"10.1007/s13755-024-00298-3","DOIUrl":"10.1007/s13755-024-00298-3","url":null,"abstract":"<p><p>Adolescent suicide has become an important social issue of general concern. Many young people express their suicidal feelings and intentions through online social media, e.g., Twitter, Microblog. The \"tree hole\" is the Chinese name for places on the Web where people post secrets. It provides the possibility of using Artificial Intelligence and big data technology to detect the posts where someone express the suicidal signal from those \"tree hole\" social media. We have developed the Web-based intelligent agents (i.e., AI-based programs) which can monitor the \"tree hole\" websites in Microblog every day by using knowledge graph technology. We have organized Tree-hole Rescue Team, which consists of more than 1000 volunteers, to carry out suicide rescue intervention according to the daily monitoring notifications. From 2018 to 2023, Tree-hole Rescue Team has prevented more than 6600 suicides. A few thousands of people have been saved within those 6 years. In this paper, we present the basic technology of Web-based Tree Hole intelligent agents and elaborate how the intelligent agents can discover suicide attempts and issue corresponding monitoring notifications and how the volunteers of Tree Hole Rescue Team can conduct online suicide intervention. This research also shows that the knowledge graph approach can be used for the semantic analysis on social media.</p>","PeriodicalId":46312,"journal":{"name":"Health Information Science and Systems","volume":"12 1","pages":"45"},"PeriodicalIF":3.4,"publicationDate":"2024-09-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11371955/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142141312","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Convolutional neural network framework for EEG-based ADHD diagnosis in children. 基于脑电图的儿童多动症诊断卷积神经网络框架。
IF 3.4 3区 医学 Q1 MEDICAL INFORMATICS Pub Date : 2024-08-31 eCollection Date: 2024-12-01 DOI: 10.1007/s13755-024-00305-7
Umaisa Hassan, Amit Singhal

Purpose: Attention-deficit hyperactivity disorder (ADHD) stands as a significant psychiatric and neuro-developmental disorder with global prevalence. The prevalence of ADHD among school children in India is estimated to range from 5% to 8%. However, certain studies have reported higher prevalence rates, reaching as high as 11%. Utilizing electroencephalography (EEG) signals for the early detection and classification of ADHD in children is crucial.

Methods: In this study, we introduce a CNN architecture characterized by its simplicity, comprising solely two convolutional layers. Our approach involves pre-processing EEG signals through a band-pass filter and segmenting them into 5-s frames. Following this, the frames undergo normalization and canonical correlation analysis. Subsequently, the proposed CNN architecture is employed for training and testing purposes.

Results: Our methodology yields remarkable results, with 100% accuracy, sensitivity, and specificity when utilizing the complete 19-channel EEG signals for diagnosing ADHD in children. However, employing the entire set of EEG channels presents challenges related to the computational complexity. Therefore, we investigate the feasibility of using only frontal brain EEG channels for ADHD detection, which yields an accuracy of 99.08%.

Conclusions: The proposed method yields high accuracy and is easy to implement, hence, it has the potential for widespread practical deployment to diagnose ADHD.

目的:注意力缺陷多动障碍(ADHD)是一种严重的精神和神经发育障碍,在全球普遍存在。据估计,注意力缺陷多动障碍在印度学龄儿童中的发病率为 5%至 8%。不过,某些研究报告称,发病率更高,达到 11%。利用脑电图(EEG)信号对儿童多动症进行早期检测和分类至关重要:在本研究中,我们介绍了一种 CNN 架构,其特点是简单,仅由两个卷积层组成。我们的方法包括通过带通滤波器预处理脑电信号,并将其分割成 5 秒钟的帧。然后,对这些帧进行归一化处理和典型相关分析。随后,提出的 CNN 架构被用于训练和测试目的:我们的方法效果显著,在利用完整的 19 通道脑电信号诊断儿童多动症时,准确率、灵敏度和特异性均达到 100%。然而,使用整套脑电图通道会带来计算复杂性方面的挑战。因此,我们研究了仅使用大脑额叶脑电图通道进行多动症检测的可行性,其准确率高达 99.08%:结论:所提出的方法准确率高且易于实施,因此有可能在实际应用中广泛用于诊断多动症。
{"title":"Convolutional neural network framework for EEG-based ADHD diagnosis in children.","authors":"Umaisa Hassan, Amit Singhal","doi":"10.1007/s13755-024-00305-7","DOIUrl":"10.1007/s13755-024-00305-7","url":null,"abstract":"<p><strong>Purpose: </strong>Attention-deficit hyperactivity disorder (ADHD) stands as a significant psychiatric and neuro-developmental disorder with global prevalence. The prevalence of ADHD among school children in India is estimated to range from 5% to 8%. However, certain studies have reported higher prevalence rates, reaching as high as 11%. Utilizing electroencephalography (EEG) signals for the early detection and classification of ADHD in children is crucial.</p><p><strong>Methods: </strong>In this study, we introduce a CNN architecture characterized by its simplicity, comprising solely two convolutional layers. Our approach involves pre-processing EEG signals through a band-pass filter and segmenting them into 5-s frames. Following this, the frames undergo normalization and canonical correlation analysis. Subsequently, the proposed CNN architecture is employed for training and testing purposes.</p><p><strong>Results: </strong>Our methodology yields remarkable results, with 100% accuracy, sensitivity, and specificity when utilizing the complete 19-channel EEG signals for diagnosing ADHD in children. However, employing the entire set of EEG channels presents challenges related to the computational complexity. Therefore, we investigate the feasibility of using only frontal brain EEG channels for ADHD detection, which yields an accuracy of 99.08%.</p><p><strong>Conclusions: </strong>The proposed method yields high accuracy and is easy to implement, hence, it has the potential for widespread practical deployment to diagnose ADHD.</p>","PeriodicalId":46312,"journal":{"name":"Health Information Science and Systems","volume":"12 1","pages":"44"},"PeriodicalIF":3.4,"publicationDate":"2024-08-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11365922/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142120855","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Health Information Science and Systems
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1