Pub Date : 2025-05-14eCollection Date: 2025-07-01DOI: 10.1007/s13534-025-00479-3
Gayathiri R R, Arya Bhardwaj, R Pradeep Kumar, Bala Chakravarthy Neelapu, Kunal Pal, J Sivaraman
Detecting Left Ventricular Systolic Dysfunction (LVSD) is crucial for counteracting heart failure progression. While Electrocardiograms (ECG) are widely used, their standalone diagnostic accuracy is insufficient. Integrating Artificial Intelligence (AI) with ECG analysis offers a promising approach to enhance precision. A systematic review was conducted to assess AI-enabled ECG for LVSD detection. Of 394 initial studies, 19 qualified for the systematic review, with 17 incorporated into meta-analysis. Study quality was gauged using QUADAS-2. Univariate meta-analysis, Spearman correlation, and bivariate meta-analyses were performed, along with publication bias assessment. The pooled sensitivity and specificity for AI-enabled ECG models were 86.9% and 84.4%, respectively. Studies with an ejection fraction (EF) threshold of 35% had the highest sensitivity, while those with 50% showed lower sensitivity and specificity. A weak positive Spearman correlation was found across all studies (ρ = 0.374, p = 0.066). A strong positive correlation for externally validated studies (ρ = 0.696, p = 0.008), and a weak negative correlation for test-only studies, indicated a threshold effect. Hierarchical summary receiver operating characteristic curve showed diagnostic robustness for studies with a 40% EF threshold; however, it showed a lack of real-world generalizability for test-only studies. AI-enabled ECG models show strong diagnostic potential for severe LVSD but remain limited for mild cases. External validation is essential for robustness and generalizability. Future research should enhance diagnostic accuracy for mild LVSD and address publication bias to optimize AI-based tools.
检测左心室收缩功能障碍(LVSD)是对抗心力衰竭进展的关键。虽然心电图(ECG)被广泛使用,但其单独诊断的准确性不足。将人工智能(AI)与ECG分析相结合提供了一种很有前途的提高精度的方法。我们进行了一项系统评价,以评估人工智能心电图对LVSD的检测效果。在394项初始研究中,19项符合系统评价,17项纳入荟萃分析。使用QUADAS-2评估研究质量。进行单因素荟萃分析、Spearman相关分析和双因素荟萃分析,并进行发表偏倚评估。人工智能心电图模型的敏感性和特异性分别为86.9%和84.4%。射血分数(EF)阈值为35%的研究灵敏度最高,而阈值为50%的研究灵敏度和特异性较低。在所有研究中均发现微弱的正Spearman相关(ρ = 0.374, p = 0.066)。外部验证的研究有很强的正相关(ρ = 0.696, p = 0.008),仅测试的研究有弱的负相关,表明存在阈值效应。分级汇总受者工作特征曲线对具有40% EF阈值的研究具有诊断稳健性;然而,它显示出对仅限测试的研究缺乏现实世界的普遍性。人工智能支持的心电图模型显示出对严重LVSD的强大诊断潜力,但对轻度病例的诊断仍然有限。外部验证对于鲁棒性和泛化性至关重要。未来的研究应提高轻度LVSD的诊断准确性,解决发表偏倚问题,优化基于人工智能的工具。
{"title":"Artificial intelligence in ECG-based diagnosis of low left ventricular ejection fraction: a systematic review and meta-analysis.","authors":"Gayathiri R R, Arya Bhardwaj, R Pradeep Kumar, Bala Chakravarthy Neelapu, Kunal Pal, J Sivaraman","doi":"10.1007/s13534-025-00479-3","DOIUrl":"https://doi.org/10.1007/s13534-025-00479-3","url":null,"abstract":"<p><p>Detecting Left Ventricular Systolic Dysfunction (LVSD) is crucial for counteracting heart failure progression. While Electrocardiograms (ECG) are widely used, their standalone diagnostic accuracy is insufficient. Integrating Artificial Intelligence (AI) with ECG analysis offers a promising approach to enhance precision. A systematic review was conducted to assess AI-enabled ECG for LVSD detection. Of 394 initial studies, 19 qualified for the systematic review, with 17 incorporated into meta-analysis. Study quality was gauged using QUADAS-2. Univariate meta-analysis, Spearman correlation, and bivariate meta-analyses were performed, along with publication bias assessment. The pooled sensitivity and specificity for AI-enabled ECG models were 86.9% and 84.4%, respectively. Studies with an ejection fraction (EF) threshold of 35% had the highest sensitivity, while those with 50% showed lower sensitivity and specificity. A weak positive Spearman correlation was found across all studies (ρ = 0.374, <i>p</i> = 0.066). A strong positive correlation for externally validated studies (ρ = 0.696, <i>p</i> = 0.008), and a weak negative correlation for test-only studies, indicated a threshold effect. Hierarchical summary receiver operating characteristic curve showed diagnostic robustness for studies with a 40% EF threshold; however, it showed a lack of real-world generalizability for test-only studies. AI-enabled ECG models show strong diagnostic potential for severe LVSD but remain limited for mild cases. External validation is essential for robustness and generalizability. Future research should enhance diagnostic accuracy for mild LVSD and address publication bias to optimize AI-based tools.</p>","PeriodicalId":46898,"journal":{"name":"Biomedical Engineering Letters","volume":"15 4","pages":"661-676"},"PeriodicalIF":3.2,"publicationDate":"2025-05-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12229370/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144585253","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
The main purpose of this study is to develop a method that can objectively evaluate the intestinal motility of patients with functional gastrointestinal disorders through non-invasive means. The research question focuses on whether the asymmetry in electronic stethoscope (ES) arrangements can enhance the accuracy of bowel sound (BS) source localization, which is crucial for detailed assessments of intestinal motility. This study introduced a new index called the angle-based asymmetry degree ([Formula: see text]), derived from the interior angles of the quadrilateral formed by the ESs, to quantitatively evaluate the asymmetry of four-ES-based arrangement patterns. We conducted simulations in an abdominal acoustic environment to compare the effects of symmetric and asymmetric ES arrangements on BS source localization accuracy. The influence of different [Formula: see text] values on localization performance was also analyzed under various signal-to-noise ratio ([Formula: see text]) conditions. The simulations revealed that BS source localization accuracy improved as the [Formula: see text] increased. Asymmetric ES arrangements significantly enhanced the localization accuracy compared to conventional symmetric arrangements, even in environments with high levels of noise. Additionally, various ES arrangement patterns corresponding to different [Formula: see text] values demonstrated improvements in localization accuracy. The study concludes that using asymmetric ES arrangements based on the newly proposed [Formula: see text] index substantially improves BS source localization accuracy. These findings suggest that asymmetry in ES placements can be a critical factor in enhancing non-invasive evaluations of intestinal motility, thereby contributing to the development of more effective BS source localization technologies. The results hold promise for practical applications in diagnosing and managing functional gastrointestinal disorders.
{"title":"Sensor arrangement strategy for effective bowel sound source localization.","authors":"Kenji Takawaki, Takeyuki Haraguchi, Takahiro Emoto","doi":"10.1007/s13534-025-00476-6","DOIUrl":"https://doi.org/10.1007/s13534-025-00476-6","url":null,"abstract":"<p><p>The main purpose of this study is to develop a method that can objectively evaluate the intestinal motility of patients with functional gastrointestinal disorders through non-invasive means. The research question focuses on whether the asymmetry in electronic stethoscope (ES) arrangements can enhance the accuracy of bowel sound (BS) source localization, which is crucial for detailed assessments of intestinal motility. This study introduced a new index called the angle-based asymmetry degree ([Formula: see text]), derived from the interior angles of the quadrilateral formed by the ESs, to quantitatively evaluate the asymmetry of four-ES-based arrangement patterns. We conducted simulations in an abdominal acoustic environment to compare the effects of symmetric and asymmetric ES arrangements on BS source localization accuracy. The influence of different [Formula: see text] values on localization performance was also analyzed under various signal-to-noise ratio ([Formula: see text]) conditions. The simulations revealed that BS source localization accuracy improved as the [Formula: see text] increased. Asymmetric ES arrangements significantly enhanced the localization accuracy compared to conventional symmetric arrangements, even in environments with high levels of noise. Additionally, various ES arrangement patterns corresponding to different [Formula: see text] values demonstrated improvements in localization accuracy. The study concludes that using asymmetric ES arrangements based on the newly proposed [Formula: see text] index substantially improves BS source localization accuracy. These findings suggest that asymmetry in ES placements can be a critical factor in enhancing non-invasive evaluations of intestinal motility, thereby contributing to the development of more effective BS source localization technologies. The results hold promise for practical applications in diagnosing and managing functional gastrointestinal disorders.</p>","PeriodicalId":46898,"journal":{"name":"Biomedical Engineering Letters","volume":"15 4","pages":"763-772"},"PeriodicalIF":3.2,"publicationDate":"2025-05-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12229448/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144585257","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-05-03eCollection Date: 2025-07-01DOI: 10.1007/s13534-025-00477-5
Jung-Hwan Kim, Ho-Seung Cha, Chang-Hwan Im
Recognition of human emotions holds great potential for various daily-life applications. With the increasing interest in virtual reality (VR) technologies, numerous studies have proposed new approaches to integrating emotion recognition into VR environments. However, despite recent advancements, camera-based emotion-recognition technology faces critical limitations due to the physical obstruction caused by head-mounted displays (HMDs). Facial electromyography (fEMG) offers a promising alternative for human emotion-recognition in VR environments, as electrodes can be readily embedded in the padding of commercial HMD devices. However, conventional fEMG-based emotion recognition approaches, although not yet developed for VR applications, require lengthy and tedious calibration sessions. These sessions typically involve collecting fEMG data during the presentation of audio-visual stimuli for eliciting specific emotions. We trained a machine learning classifier using fEMG data acquired while users intentionally made posed facial expressions. This approach simplifies the traditionally time-consuming calibration process, making it less burdensome for users. The proposed method was validated using 20 participants who made posed facial expressions for calibration and then watched emotion-evoking video clips for validation. The results demonstrated the effectiveness of our method in classifying high- and low-valence states, achieving a macro F1-score of 88.20%. This underscores the practicality and efficiency of the proposed method. To the best of our knowledge, this is the first study to successfully build an fEMG-based emotion-recognition model using posed facial expressions. This approach paves the way for developing user-friendly interface technologies in VR-immersive environments.
{"title":"Facial electromyogram-based emotion recognition for virtual reality applications using machine learning classifiers trained on posed expressions.","authors":"Jung-Hwan Kim, Ho-Seung Cha, Chang-Hwan Im","doi":"10.1007/s13534-025-00477-5","DOIUrl":"https://doi.org/10.1007/s13534-025-00477-5","url":null,"abstract":"<p><p>Recognition of human emotions holds great potential for various daily-life applications. With the increasing interest in virtual reality (VR) technologies, numerous studies have proposed new approaches to integrating emotion recognition into VR environments. However, despite recent advancements, camera-based emotion-recognition technology faces critical limitations due to the physical obstruction caused by head-mounted displays (HMDs). Facial electromyography (fEMG) offers a promising alternative for human emotion-recognition in VR environments, as electrodes can be readily embedded in the padding of commercial HMD devices. However, conventional fEMG-based emotion recognition approaches, although not yet developed for VR applications, require lengthy and tedious calibration sessions. These sessions typically involve collecting fEMG data during the presentation of audio-visual stimuli for eliciting specific emotions. We trained a machine learning classifier using fEMG data acquired while users intentionally made posed facial expressions. This approach simplifies the traditionally time-consuming calibration process, making it less burdensome for users. The proposed method was validated using 20 participants who made posed facial expressions for calibration and then watched emotion-evoking video clips for validation. The results demonstrated the effectiveness of our method in classifying high- and low-valence states, achieving a macro F1-score of 88.20%. This underscores the practicality and efficiency of the proposed method. To the best of our knowledge, this is the first study to successfully build an fEMG-based emotion-recognition model using posed facial expressions. This approach paves the way for developing user-friendly interface technologies in VR-immersive environments.</p>","PeriodicalId":46898,"journal":{"name":"Biomedical Engineering Letters","volume":"15 4","pages":"773-783"},"PeriodicalIF":3.2,"publicationDate":"2025-05-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12229427/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144585255","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-04-25eCollection Date: 2025-07-01DOI: 10.1007/s13534-025-00474-8
Shuangping Tan, Tong Zhang, Youfeng Deng, Zhimin Nie, Xiali Wu, Xinyue Yan, Xiaojuan Zhang, Huike Yi, Xianci Song, Jun Li
Accurate segmentation of pulmonary nodule based on computed tomography (CT) images is of great significance for the diagnosis and treatment of lung cancer. However, the current popular segmentation algorithms usually do not involve expert knowledge from radiologists, thereby carrying the risk of failing to produce generalizable and trustworthy models. In this study, we develop an expert-informed active pulmonary nodule segmentation method that iteratively optimize a deep segmentation model using an active learning scheme. The uncertainties from both intermediate segmentation results and correction inputs from radiologists are combined effectively. Interactive graph interfaces are developed to enable online corrections, significantly facilitating the integration of expert knowledge from radiologists. Evaluation results on the Luna16 dataset demonstrate that the proposed approach significantly promotes the segmentation performance of pulmonary nodules. The proposed method can effectively incorporate expert knowledge of multiple radiologists into deep segmentation algorithms, which not only promote the segmentation performance, but also enhance the validity, reliability, and generalizability of computer-aided diagnosis methods.
{"title":"A study for expert-informed active pulmonary nodule segmentation.","authors":"Shuangping Tan, Tong Zhang, Youfeng Deng, Zhimin Nie, Xiali Wu, Xinyue Yan, Xiaojuan Zhang, Huike Yi, Xianci Song, Jun Li","doi":"10.1007/s13534-025-00474-8","DOIUrl":"https://doi.org/10.1007/s13534-025-00474-8","url":null,"abstract":"<p><p>Accurate segmentation of pulmonary nodule based on computed tomography (CT) images is of great significance for the diagnosis and treatment of lung cancer. However, the current popular segmentation algorithms usually do not involve expert knowledge from radiologists, thereby carrying the risk of failing to produce generalizable and trustworthy models. In this study, we develop an expert-informed active pulmonary nodule segmentation method that iteratively optimize a deep segmentation model using an active learning scheme. The uncertainties from both intermediate segmentation results and correction inputs from radiologists are combined effectively. Interactive graph interfaces are developed to enable online corrections, significantly facilitating the integration of expert knowledge from radiologists. Evaluation results on the Luna16 dataset demonstrate that the proposed approach significantly promotes the segmentation performance of pulmonary nodules. The proposed method can effectively incorporate expert knowledge of multiple radiologists into deep segmentation algorithms, which not only promote the segmentation performance, but also enhance the validity, reliability, and generalizability of computer-aided diagnosis methods.</p>","PeriodicalId":46898,"journal":{"name":"Biomedical Engineering Letters","volume":"15 4","pages":"735-748"},"PeriodicalIF":3.2,"publicationDate":"2025-04-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12229428/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144585251","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-04-24eCollection Date: 2025-07-01DOI: 10.1007/s13534-025-00473-9
Yunendah Nur Fuadah, Ki Moo Lim
This systematic review examines the transformative impact of artificial intelligence (AI), including machine learning (ML) and deep learning (DL), on cardiovascular signal analysis, focusing on electrocardiograms (ECG), phonocardiograms (PCG), and photoplethysmograms (PPG). It evaluates state-of-the-art methodologies that enhance diagnostic accuracy and predictive analytics by leveraging AI-driven systems. A wide range of public and private datasets are assessed, with attention to their strengths and limitations in supporting cardiovascular diagnostics. Key preprocessing techniques, such as noise reduction, signal normalization, and artifact mitigation, are explored for their role in improving signal quality. The review also highlights feature extraction methods, from time-domain and frequency-domain analyses to advanced morphological and spectral approaches, which contribute to robust classifier performance. Traditional ML models, such as K-Nearest Neighbor (KNN), Support Vector Machines (SVM) and Random Forests (RF), are compared with advanced DL architectures, including Convolutional Neural Networks (CNNs), Long Short-Term Memory networks (LSTMs), and transfer learning models for detecting cardiovascular diseases. Despite these advancements, challenges such as dataset heterogeneity, preprocessing variability, and computational complexity persist, hindering clinical adoption. The review underscores the importance of large-scale, diverse datasets, multi-modal signal integration, and explainable AI to foster clinical trust and facilitate ethical deployment. By addressing these challenges, this review highlights the potential of AI to revolutionize cardiovascular healthcare through early diagnosis, wearable technology, and real-time decision support, paving the way for precision medicine and improved patient outcomes.
{"title":"Advances in cardiovascular signal analysis with future directions: a review of machine learning and deep learning models for cardiovascular disease classification based on ECG, PCG, and PPG signals.","authors":"Yunendah Nur Fuadah, Ki Moo Lim","doi":"10.1007/s13534-025-00473-9","DOIUrl":"https://doi.org/10.1007/s13534-025-00473-9","url":null,"abstract":"<p><p>This systematic review examines the transformative impact of artificial intelligence (AI), including machine learning (ML) and deep learning (DL), on cardiovascular signal analysis, focusing on electrocardiograms (ECG), phonocardiograms (PCG), and photoplethysmograms (PPG). It evaluates state-of-the-art methodologies that enhance diagnostic accuracy and predictive analytics by leveraging AI-driven systems. A wide range of public and private datasets are assessed, with attention to their strengths and limitations in supporting cardiovascular diagnostics. Key preprocessing techniques, such as noise reduction, signal normalization, and artifact mitigation, are explored for their role in improving signal quality. The review also highlights feature extraction methods, from time-domain and frequency-domain analyses to advanced morphological and spectral approaches, which contribute to robust classifier performance. Traditional ML models, such as K-Nearest Neighbor (KNN), Support Vector Machines (SVM) and Random Forests (RF), are compared with advanced DL architectures, including Convolutional Neural Networks (CNNs), Long Short-Term Memory networks (LSTMs), and transfer learning models for detecting cardiovascular diseases. Despite these advancements, challenges such as dataset heterogeneity, preprocessing variability, and computational complexity persist, hindering clinical adoption. The review underscores the importance of large-scale, diverse datasets, multi-modal signal integration, and explainable AI to foster clinical trust and facilitate ethical deployment. By addressing these challenges, this review highlights the potential of AI to revolutionize cardiovascular healthcare through early diagnosis, wearable technology, and real-time decision support, paving the way for precision medicine and improved patient outcomes.</p>","PeriodicalId":46898,"journal":{"name":"Biomedical Engineering Letters","volume":"15 4","pages":"619-660"},"PeriodicalIF":3.2,"publicationDate":"2025-04-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12229450/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144585252","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-04-21eCollection Date: 2025-07-01DOI: 10.1007/s13534-025-00475-7
Hyunwook Kang, Jin Woo Choi, Byung Hyung Kim
Electroencephalogram (EEG) signal is receiving much attention from recent studies since it is highly associated with intrinsic emotion. However, EEG signals contain underlying factors of variations across different sessions of the same subject, which make it difficult to learn temporal relationships between successive time steps. To disentangle invariant features, we propose a feature re-weighting mechanism on the extracted EEG features for temporal sequence modeling. Based on this method, our proposed model, called Convolutional Channel Modulator for Transformer and LSTM networks (CCMTL), extracts emotion-related inter-channel correlations using convolution operations and emphasizes important features by generating a channel attention map. This attention map is then used to perform matrix multiplication on the extracted features, which helps the subsequent Transformer to focus on important affective features. Furthermore, the sequential temporal modeling enhances the overall model's capability to understand temporal relationships both in global and sequential contexts. Experimental settings on public EEG emotion datasets demonstrate the superiority of the proposed CCMTL, surpassing six state-of-the-art models. Our code is publicly available at https://github.com/affctivai/CCMTL.
{"title":"Convolutional channel modulator for transformer and LSTM networks in EEG-based emotion recognition.","authors":"Hyunwook Kang, Jin Woo Choi, Byung Hyung Kim","doi":"10.1007/s13534-025-00475-7","DOIUrl":"https://doi.org/10.1007/s13534-025-00475-7","url":null,"abstract":"<p><p>Electroencephalogram (EEG) signal is receiving much attention from recent studies since it is highly associated with intrinsic emotion. However, EEG signals contain underlying factors of variations across different sessions of the same subject, which make it difficult to learn temporal relationships between successive time steps. To disentangle invariant features, we propose a feature re-weighting mechanism on the extracted EEG features for temporal sequence modeling. Based on this method, our proposed model, called Convolutional Channel Modulator for Transformer and LSTM networks (CCMTL), extracts emotion-related inter-channel correlations using convolution operations and emphasizes important features by generating a channel attention map. This attention map is then used to perform matrix multiplication on the extracted features, which helps the subsequent Transformer to focus on important affective features. Furthermore, the sequential temporal modeling enhances the overall model's capability to understand temporal relationships both in global and sequential contexts. Experimental settings on public EEG emotion datasets demonstrate the superiority of the proposed CCMTL, surpassing six state-of-the-art models. Our code is publicly available at https://github.com/affctivai/CCMTL.</p>","PeriodicalId":46898,"journal":{"name":"Biomedical Engineering Letters","volume":"15 4","pages":"749-761"},"PeriodicalIF":3.2,"publicationDate":"2025-04-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12229436/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144585254","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-04-09eCollection Date: 2025-05-01DOI: 10.1007/s13534-025-00472-w
Yong-Hoon Jeong, Young-Seok Choi
Remote photoplethysmography (rPPG) offers significant potential for health monitoring and emotional analysis through non-contact physiological measurement from facial videos. However, noise remains a crucial challenge, limiting the generalizability of current rPPG methods. This paper introduces Diffusion-Phys, a novel framework using diffusion models for robust heart rate (HR) estimation from facial videos. Diffusion-Phys employs Multi-scale Spatial-Temporal Maps (MSTmaps) to preprocess input data and introduces Gaussian noise to simulate real-world conditions. The model is trained using a denoising network for accurate HR estimation. Experimental evaluations on the VIPL-HR, UBFC-rPPG and PURE datasets demonstrate that Diffusion-Phys achieves comparable or superior performance to state-of-the-art methods, with lower computational complexity. These results highlight the effectiveness of explicitly addressing noise through diffusion modeling, improving the reliability and generalization of non-contact physiological measurement systems.
{"title":"Diffusion-Phys: noise-robust heart rate estimation from facial videos via diffusion models.","authors":"Yong-Hoon Jeong, Young-Seok Choi","doi":"10.1007/s13534-025-00472-w","DOIUrl":"10.1007/s13534-025-00472-w","url":null,"abstract":"<p><p>Remote photoplethysmography (rPPG) offers significant potential for health monitoring and emotional analysis through non-contact physiological measurement from facial videos. However, noise remains a crucial challenge, limiting the generalizability of current rPPG methods. This paper introduces Diffusion-Phys, a novel framework using diffusion models for robust heart rate (HR) estimation from facial videos. Diffusion-Phys employs Multi-scale Spatial-Temporal Maps (MSTmaps) to preprocess input data and introduces Gaussian noise to simulate real-world conditions. The model is trained using a denoising network for accurate HR estimation. Experimental evaluations on the VIPL-HR, UBFC-rPPG and PURE datasets demonstrate that Diffusion-Phys achieves comparable or superior performance to state-of-the-art methods, with lower computational complexity. These results highlight the effectiveness of explicitly addressing noise through diffusion modeling, improving the reliability and generalization of non-contact physiological measurement systems.</p>","PeriodicalId":46898,"journal":{"name":"Biomedical Engineering Letters","volume":"15 3","pages":"575-585"},"PeriodicalIF":3.2,"publicationDate":"2025-04-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12011667/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143990201","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-04-05eCollection Date: 2025-05-01DOI: 10.1007/s13534-025-00471-x
Chang-Hee Han, Seong-Uk Kim, Kyung-Soo Lim, Young-Jin Jung, Sangho Lee, Sung Hoon Kim, Han-Jeong Hwang
The development of conductive polymer-based dry electrodes with high conductivity is promising for practical applications in daily life due to their biocompatibility, flexibility, lightweight, and comfort. The objective of this study is to demonstrate the feasibility of using a novel silicone-based dry electrode for measuring various bioelectrical signals.The silicone-based electrode, manufactured using an optimized polymer matrix, combines high conductivity with flexibility, ensuring superior wearability and reliable bioelectrical signal monitoring. To evaluate its performance, we compared its impedance and flexibility with those of a commercial electrode. Additionally, its compatibility for measuring biological signals was assessed through performance comparisons across various bioelectrical signals. Fourteen healthy participants performed three experimental paradigms: (1) eyes closed and open to measure alpha electroencephalography (EEG) as well as resting-state electrocardiography (ECG), (2) eye blinking to measure electrooculography (EOG), and (3) wrist movement to measure electromyography (EMG). All bioelectrical signals were measured simultaneously using both the silicone-based dry electrode and a commercial dry electrode. The performance of the silicone-based dry electrode was evaluated by comparing the signal quality of both electrodes. The silicone-based dry electrode exhibited lower electrical impedance (39.43 kΩ on average, p = 0.0058) and greater flexibility (Young's modulus: silicone 1.51 ± 0.10 MPa vs. commercial 2.46 ± 0.38 MPa) compared to the commercial dry electrode. Overall, there were minimal differences in signal quality between the two electrodes: i) EEG (α power SNR: silicone 1.39 ± 0.34 vs. commercial 1.36 ± 0.29), ii) ECG (R-peak recall: 99.20 ± 2.50%, correlation coefficient: 0.96 ± 0.08), iii) EOG (eye blink recall: 100.00%, correlation coefficient: 0.98 ± 0.03), and iv) EMG (no significant difference in SNR values). These findings indicate that the developed electrode not only ensures superior flexibility but also maintains compatible electrical properties for measuring various bioelectrical signals.
{"title":"Evaluation of a silicone-based flexible dry electrode for measuring human bioelectrical signals.","authors":"Chang-Hee Han, Seong-Uk Kim, Kyung-Soo Lim, Young-Jin Jung, Sangho Lee, Sung Hoon Kim, Han-Jeong Hwang","doi":"10.1007/s13534-025-00471-x","DOIUrl":"10.1007/s13534-025-00471-x","url":null,"abstract":"<p><p>The development of conductive polymer-based dry electrodes with high conductivity is promising for practical applications in daily life due to their biocompatibility, flexibility, lightweight, and comfort. The objective of this study is to demonstrate the feasibility of using a novel silicone-based dry electrode for measuring various bioelectrical signals.The silicone-based electrode, manufactured using an optimized polymer matrix, combines high conductivity with flexibility, ensuring superior wearability and reliable bioelectrical signal monitoring. To evaluate its performance, we compared its impedance and flexibility with those of a commercial electrode. Additionally, its compatibility for measuring biological signals was assessed through performance comparisons across various bioelectrical signals. Fourteen healthy participants performed three experimental paradigms: (1) eyes closed and open to measure alpha electroencephalography (EEG) as well as resting-state electrocardiography (ECG), (2) eye blinking to measure electrooculography (EOG), and (3) wrist movement to measure electromyography (EMG). All bioelectrical signals were measured simultaneously using both the silicone-based dry electrode and a commercial dry electrode. The performance of the silicone-based dry electrode was evaluated by comparing the signal quality of both electrodes. The silicone-based dry electrode exhibited lower electrical impedance (39.43 kΩ on average, <i>p</i> = 0.0058) and greater flexibility (Young's modulus: silicone 1.51 ± 0.10 MPa vs. commercial 2.46 ± 0.38 MPa) compared to the commercial dry electrode. Overall, there were minimal differences in signal quality between the two electrodes: i) EEG (α power SNR: silicone 1.39 ± 0.34 vs. commercial 1.36 ± 0.29), ii) ECG (R-peak recall: 99.20 ± 2.50%, correlation coefficient: 0.96 ± 0.08), iii) EOG (eye blink recall: 100.00%, correlation coefficient: 0.98 ± 0.03), and iv) EMG (no significant difference in SNR values). These findings indicate that the developed electrode not only ensures superior flexibility but also maintains compatible electrical properties for measuring various bioelectrical signals.</p>","PeriodicalId":46898,"journal":{"name":"Biomedical Engineering Letters","volume":"15 3","pages":"563-574"},"PeriodicalIF":3.2,"publicationDate":"2025-04-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12011699/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144020137","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-03-28eCollection Date: 2025-05-01DOI: 10.1007/s13534-025-00470-y
Ji Heon Lim, Ji Wook Choi, Na Yeon Kim, Taewook Kang, Bong Geun Chung
Three-dimensional (3D) tumor spheroid models closely mimic in vivo tumor environment and play a vital role in studying oncological research. Despite their significance, the existing methods for analyzing 3D tumor spheroids often suffer from limitations, including low throughput, high cost, and insufficient resolution. To address these challenges, we developed a portable imaging system for the real-time sensing and quantitative analysis of the 3D tumor spheroids. The system integrated the seamless workflow of spheroid generation, cell morphology tracking, and drug screening. The spheroid generation was successfully characterized using MCF-7 breast cancer cells by optimizing cell concentration (5-20 × 106 cells/mL), incubation time (24-96 h) and microwell diameter (400-600 μm). A custom-written algorithm was developed for automated analysis of spheroids, exhibiting high sensitivity (98.99%) and specificity (98.21%). Confusion matrices and receiver operating characteristic curve analysis further confirmed the robustness of the algorithm with an area under the curve value of 93.75% and an equal error rate of 0.79%. Following the characterization, the real-time sensing of spheroid generation and the response of spheroids to drug treatment were successfully demonstrated. Furthermore, the live/dead assays with chemotherapy provided a detailed insight into the efficacy and cytotoxic effects of the drug, demonstrating a significant dose-dependent decrease in a spheroid viability. Therefore, our system offers considerable potential for enhancing drug development processes and personalized treatment strategies, thereby contributing to more effective cancer therapies.
Supplementary information: The online version contains supplementary material available at 10.1007/s13534-025-00470-y.
{"title":"Real-time monitoring and quantitative analysis of 3D tumor spheroids using portable cellular imaging system.","authors":"Ji Heon Lim, Ji Wook Choi, Na Yeon Kim, Taewook Kang, Bong Geun Chung","doi":"10.1007/s13534-025-00470-y","DOIUrl":"10.1007/s13534-025-00470-y","url":null,"abstract":"<p><p>Three-dimensional (3D) tumor spheroid models closely mimic in vivo tumor environment and play a vital role in studying oncological research. Despite their significance, the existing methods for analyzing 3D tumor spheroids often suffer from limitations, including low throughput, high cost, and insufficient resolution. To address these challenges, we developed a portable imaging system for the real-time sensing and quantitative analysis of the 3D tumor spheroids. The system integrated the seamless workflow of spheroid generation, cell morphology tracking, and drug screening. The spheroid generation was successfully characterized using MCF-7 breast cancer cells by optimizing cell concentration (5-20 × 10<sup>6</sup> cells/mL), incubation time (24-96 h) and microwell diameter (400-600 μm). A custom-written algorithm was developed for automated analysis of spheroids, exhibiting high sensitivity (98.99%) and specificity (98.21%). Confusion matrices and receiver operating characteristic curve analysis further confirmed the robustness of the algorithm with an area under the curve value of 93.75% and an equal error rate of 0.79%. Following the characterization, the real-time sensing of spheroid generation and the response of spheroids to drug treatment were successfully demonstrated. Furthermore, the live/dead assays with chemotherapy provided a detailed insight into the efficacy and cytotoxic effects of the drug, demonstrating a significant dose-dependent decrease in a spheroid viability. Therefore, our system offers considerable potential for enhancing drug development processes and personalized treatment strategies, thereby contributing to more effective cancer therapies.</p><p><strong>Supplementary information: </strong>The online version contains supplementary material available at 10.1007/s13534-025-00470-y.</p>","PeriodicalId":46898,"journal":{"name":"Biomedical Engineering Letters","volume":"15 3","pages":"549-561"},"PeriodicalIF":3.2,"publicationDate":"2025-03-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12011688/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144041732","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Human-computer interaction (HCI) focuses on designing efficient and intuitive interactions between humans and computer systems. Recent advancements have utilized multimodal approaches, such as electroencephalography (EEG)-based systems combined with other biosignals, along with deep learning to enhance performance and reliability. However, no systematic review has consolidated findings on EEG-based multimodal HCI systems. This review examined 124 studies published from 2016 to 2024, retrieved from the Web of Science database, focusing on hybrid EEG-based multimodal HCI systems employing deep learning. The keywords used for evaluation were as follows: 'Deep Learning' AND 'EEG' AND ('fNIRS' OR 'NIRS' OR 'MEG' OR 'fMRI' OR 'EOG' OR 'EMG' OR 'ECG' OR 'PPG' OR 'GSR'). Main topics explored are: (1) types of biosignals used with EEG, (2) neural network architectures, (3) fusion strategies, (4) system performance, and (5) target applications. Frequently paired signals, such as EOG, EMG, and fNIRS, effectively complement EEG by addressing its limitations. Convolutional neural networks are extensively employed for spatio-temporal-spectral feature extraction, with early and intermediate fusion strategies being the most commonly used. Applications, such as sleep stage classification, emotion recognition, and mental state decoding, have shown notable performance improvements. Despite these advancements, challenges remain, including the lack of real-time online systems, difficulties in signal synchronization, limited data availability, and insufficient explainable AI (XAI) methods to interpret signal interactions. Emerging solutions, such as portable systems, lightweight deep learning models, and data augmentation techniques, offer promising pathways to address these issues. This review highlights the potential of EEG-based multimodal HCI systems and emphasizes the need for advancements in real-time interaction, fusion algorithms, and XAI to enhance their adaptability, interpretability, and reliability.
人机交互(HCI)侧重于设计人与计算机系统之间有效和直观的交互。最近的进展是利用多模式方法,例如基于脑电图(EEG)的系统与其他生物信号相结合,以及深度学习来提高性能和可靠性。然而,目前还没有系统性的综述对基于脑电图的多模态HCI系统的研究结果进行整合。本综述从Web of Science数据库中检索了2016年至2024年发表的124项研究,重点关注采用深度学习的基于脑电图的混合多模态HCI系统。用于评估的关键词如下:“深度学习”和“脑电图”和(“fNIRS”或“NIRS”或“MEG”或“fMRI”或“EOG”或“EMG”或“ECG”或“PPG”或“GSR”)。主要探讨的主题有:(1)脑电图使用的生物信号类型,(2)神经网络架构,(3)融合策略,(4)系统性能,以及(5)目标应用。频繁配对的信号,如EOG、EMG和fNIRS,通过解决EEG的局限性,有效地补充了EEG。卷积神经网络广泛用于时空光谱特征提取,其中早期和中期融合策略是最常用的。睡眠阶段分类、情绪识别和精神状态解码等应用已经显示出显著的性能改善。尽管取得了这些进步,但挑战依然存在,包括缺乏实时在线系统、信号同步困难、数据可用性有限以及解释信号相互作用的可解释人工智能(XAI)方法不足。新兴的解决方案,如便携式系统、轻量级深度学习模型和数据增强技术,为解决这些问题提供了有希望的途径。这篇综述强调了基于脑电图的多模态HCI系统的潜力,并强调需要在实时交互、融合算法和XAI方面取得进展,以增强其适应性、可解释性和可靠性。
{"title":"A review of hybrid EEG-based multimodal human-computer interfaces using deep learning: applications, advances, and challenges.","authors":"Hyung-Tak Lee, Miseon Shim, Xianghong Liu, Hye-Ran Cheon, Sang-Gyu Kim, Chang-Hee Han, Han-Jeong Hwang","doi":"10.1007/s13534-025-00469-5","DOIUrl":"https://doi.org/10.1007/s13534-025-00469-5","url":null,"abstract":"<p><p>Human-computer interaction (HCI) focuses on designing efficient and intuitive interactions between humans and computer systems. Recent advancements have utilized multimodal approaches, such as electroencephalography (EEG)-based systems combined with other biosignals, along with deep learning to enhance performance and reliability. However, no systematic review has consolidated findings on EEG-based multimodal HCI systems. This review examined 124 studies published from 2016 to 2024, retrieved from the Web of Science database, focusing on hybrid EEG-based multimodal HCI systems employing deep learning. The keywords used for evaluation were as follows: 'Deep Learning' AND 'EEG' AND ('fNIRS' OR 'NIRS' OR 'MEG' OR 'fMRI' OR 'EOG' OR 'EMG' OR 'ECG' OR 'PPG' OR 'GSR'). Main topics explored are: (1) types of biosignals used with EEG, (2) neural network architectures, (3) fusion strategies, (4) system performance, and (5) target applications. Frequently paired signals, such as EOG, EMG, and fNIRS, effectively complement EEG by addressing its limitations. Convolutional neural networks are extensively employed for spatio-temporal-spectral feature extraction, with early and intermediate fusion strategies being the most commonly used. Applications, such as sleep stage classification, emotion recognition, and mental state decoding, have shown notable performance improvements. Despite these advancements, challenges remain, including the lack of real-time online systems, difficulties in signal synchronization, limited data availability, and insufficient explainable AI (XAI) methods to interpret signal interactions. Emerging solutions, such as portable systems, lightweight deep learning models, and data augmentation techniques, offer promising pathways to address these issues. This review highlights the potential of EEG-based multimodal HCI systems and emphasizes the need for advancements in real-time interaction, fusion algorithms, and XAI to enhance their adaptability, interpretability, and reliability.</p>","PeriodicalId":46898,"journal":{"name":"Biomedical Engineering Letters","volume":"15 4","pages":"587-618"},"PeriodicalIF":3.2,"publicationDate":"2025-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12229323/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144585250","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}