Pub Date : 2024-03-21DOI: 10.1109/JTEHM.2024.3368106
Takhellambam Gautam Meitei;Wei-Chun Chang;Pou-Leng Cheong;Yi-Min Wang;Chia-Wei Sun
Osteoporosis is a prevalent chronic disease worldwide, particularly affecting the aging population. The gold standard diagnostic tool for osteoporosis is Dual-energy X-ray Absorptiometry (DXA). However, the expensive cost of the DXA machine and the need for skilled professionals to operate it restrict its accessibility to the general public. This paper builds upon previous research and proposes a novel approach for rapidly screening bone density. The method involves utilizing near-infrared light to capture local body information within the human body. Deep learning techniques are employed to analyze the obtained data and extract meaningful insights related to bone density. Our initial prediction, utilizing multi-linear regression, demonstrated a strong correlation (r = 0.98, p-value = 0.003**) with the measured Bone Mineral Density (BMD) obtained from Dual-energy X-ray Absorptiometry (DXA). This indicates a highly significant relationship between the predicted values and the actual BMD measurements. A deep learning-based algorithm is applied to analyze the underlying information further to predict bone density at the wrist, hip, and spine. The prediction of bone densities in the hip and spine holds significant importance due to their status as gold-standard sites for assessing an individual’s bone density. Our prediction rate had an error margin below 10% for the wrist and below 20% for the hip and spine bone density.
骨质疏松症是一种全球流行的慢性疾病,对老龄人口的影响尤为严重。骨质疏松症的金标准诊断工具是双能 X 射线吸收测定法(DXA)。然而,DXA 仪器价格昂贵,而且需要熟练的专业人员操作,这限制了普通大众对它的使用。本文在以往研究的基础上,提出了一种快速筛查骨密度的新方法。该方法利用近红外线捕捉人体局部信息。利用深度学习技术来分析获得的数据,并提取与骨密度相关的有意义的见解。我们利用多线性回归进行的初步预测显示,该预测与通过双能 X 射线吸收仪(DXA)测量的骨密度(BMD)之间存在很强的相关性(r = 0.98,p 值 = 0.003**)。这表明预测值与实际 BMD 测量值之间存在着非常显著的关系。应用基于深度学习的算法进一步分析基础信息,以预测手腕、髋部和脊柱的骨密度。由于髋部和脊柱是评估个人骨密度的黄金标准部位,因此预测这两个部位的骨密度具有重要意义。我们对腕部骨密度的预测误差率低于 10%,对髋部和脊柱骨密度的预测误差率低于 20%。
{"title":"A Study on Intelligent Optical Bone Densitometry","authors":"Takhellambam Gautam Meitei;Wei-Chun Chang;Pou-Leng Cheong;Yi-Min Wang;Chia-Wei Sun","doi":"10.1109/JTEHM.2024.3368106","DOIUrl":"10.1109/JTEHM.2024.3368106","url":null,"abstract":"Osteoporosis is a prevalent chronic disease worldwide, particularly affecting the aging population. The gold standard diagnostic tool for osteoporosis is Dual-energy X-ray Absorptiometry (DXA). However, the expensive cost of the DXA machine and the need for skilled professionals to operate it restrict its accessibility to the general public. This paper builds upon previous research and proposes a novel approach for rapidly screening bone density. The method involves utilizing near-infrared light to capture local body information within the human body. Deep learning techniques are employed to analyze the obtained data and extract meaningful insights related to bone density. Our initial prediction, utilizing multi-linear regression, demonstrated a strong correlation (r = 0.98, p-value = 0.003**) with the measured Bone Mineral Density (BMD) obtained from Dual-energy X-ray Absorptiometry (DXA). This indicates a highly significant relationship between the predicted values and the actual BMD measurements. A deep learning-based algorithm is applied to analyze the underlying information further to predict bone density at the wrist, hip, and spine. The prediction of bone densities in the hip and spine holds significant importance due to their status as gold-standard sites for assessing an individual’s bone density. Our prediction rate had an error margin below 10% for the wrist and below 20% for the hip and spine bone density.","PeriodicalId":54255,"journal":{"name":"IEEE Journal of Translational Engineering in Health and Medicine-Jtehm","volume":"12 ","pages":"401-412"},"PeriodicalIF":3.4,"publicationDate":"2024-03-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10477504","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140202996","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Background: CHIVID is a telemedicine solution developed under tight time constraints that assists Thai healthcare practitioners in monitoring non-severe COVID-19 patients in isolation programs during crises. It assesses patient health and notifies healthcare practitioners of high-risk scenarios through a chatbot. The system was designed to integrate with the famous Thai messaging app LINE, reducing development time and enhancing user-friendliness, and the system allowed patients to upload a pulse oximeter image automatically processed by the PACMAN function to extract oxygen saturation and heart rate values to reduce patient input errors. Methods: This article describes the proposed system and presents a mixed-methods study that evaluated the system’s performance by collecting survey responses from 70 healthcare practitioners and analyzing 14,817 patient records. Results: Approximately 71.4% of healthcare practitioners use the system more than twice daily, with the majority managing 1–10 patients, while 11.4% handle over 101 patients. The progress note is a function that healthcare practitioners most frequently use and are satisfied with. Regarding patient data, 58.9%(8,724/14,817) are male, and 49.7%(7,367/14,817) within the 18 to 34 age range. The average length of isolation was 7.6 days, and patients submitted progress notes twice daily on average. Notably, individuals aged 18 to 34 demonstrated the highest utilization rates for the PACMAN function. Furthermore, most patients, totaling over 95.52%(14,153/14,817), were discharged normally. Conclusion: The findings indicate that CHIVID could be one of the telemedicine solutions for hospitals with patient overflow and healthcare practitioners unfamiliar with telemedicine technology to improve patient care during a critical crisis. Clinical and Translational Impact Statement— CHIVID’s success arises from seamlessly integrating telemedicine into third-party application within a limited timeframe and effectively using clinical decision support systems to address challenges during the COVID-19 crisis.
{"title":"CHIVID: A Rapid Deployment of Community and Home Isolation During COVID-19 Pandemics","authors":"Parpada Piamjinda;Chiraphat Boonnag;Piyalitt Ittichaiwong;Seandee Rattanasonrerk;Kanyakorn Veerakanjana;Khanita Duangchaemkarn;Warissara Limpornchitwilai;Kamonwan Thanontip;Napasara Asawalertsak;Thitikorn Kaewlee;Theerawit Wilaiprasitporn","doi":"10.1109/JTEHM.2024.3377258","DOIUrl":"10.1109/JTEHM.2024.3377258","url":null,"abstract":"Background: CHIVID is a telemedicine solution developed under tight time constraints that assists Thai healthcare practitioners in monitoring non-severe COVID-19 patients in isolation programs during crises. It assesses patient health and notifies healthcare practitioners of high-risk scenarios through a chatbot. The system was designed to integrate with the famous Thai messaging app LINE, reducing development time and enhancing user-friendliness, and the system allowed patients to upload a pulse oximeter image automatically processed by the PACMAN function to extract oxygen saturation and heart rate values to reduce patient input errors. Methods: This article describes the proposed system and presents a mixed-methods study that evaluated the system’s performance by collecting survey responses from 70 healthcare practitioners and analyzing 14,817 patient records. Results: Approximately 71.4% of healthcare practitioners use the system more than twice daily, with the majority managing 1–10 patients, while 11.4% handle over 101 patients. The progress note is a function that healthcare practitioners most frequently use and are satisfied with. Regarding patient data, 58.9%(8,724/14,817) are male, and 49.7%(7,367/14,817) within the 18 to 34 age range. The average length of isolation was 7.6 days, and patients submitted progress notes twice daily on average. Notably, individuals aged 18 to 34 demonstrated the highest utilization rates for the PACMAN function. Furthermore, most patients, totaling over 95.52%(14,153/14,817), were discharged normally. Conclusion: The findings indicate that CHIVID could be one of the telemedicine solutions for hospitals with patient overflow and healthcare practitioners unfamiliar with telemedicine technology to improve patient care during a critical crisis. Clinical and Translational Impact Statement— CHIVID’s success arises from seamlessly integrating telemedicine into third-party application within a limited timeframe and effectively using clinical decision support systems to address challenges during the COVID-19 crisis.","PeriodicalId":54255,"journal":{"name":"IEEE Journal of Translational Engineering in Health and Medicine-Jtehm","volume":"12 ","pages":"390-400"},"PeriodicalIF":3.4,"publicationDate":"2024-03-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10472531","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140126520","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Acoustic features extracted from speech can help with the diagnosis of neurological diseases and monitoring of symptoms over time. Temporal segmentation of audio signals into individual words is an important pre-processing step needed prior to extracting acoustic features. Machine learning techniques could be used to automate speech segmentation via automatic speech recognition (ASR) and sequence to sequence alignment. While state-of-the-art ASR models achieve good performance on healthy speech, their performance significantly drops when evaluated on dysarthric speech. Fine-tuning ASR models on impaired speech can improve performance in dysarthric individuals, but it requires representative clinical data, which is difficult to collect and may raise privacy concerns. This study explores the feasibility of using two augmentation methods to increase ASR performance on dysarthric speech: 1) healthy individuals varying their speaking rate and loudness (as is often used in assessments of pathological speech); 2) synthetic speech with variations in speaking rate and accent (to ensure more diverse vocal representations and fairness). Experimental evaluations showed that fine-tuning a pre-trained ASR model with data from these two sources outperformed a model fine-tuned only on real clinical data and matched the performance of a model fine-tuned on the combination of real clinical data and synthetic speech. When evaluated on held-out acoustic data from 24 individuals with various neurological diseases, the best performing model achieved an average word error rate of 5.7% and a mean correct count accuracy of 94.4%. In segmenting the data into individual words, a mean intersection-over-union of 89.2% was obtained against manual parsing (ground truth). It can be concluded that emulated and synthetic augmentations can significantly reduce the need for real clinical data of dysarthric speech when fine-tuning ASR models and, in turn, for speech segmentation.
从语音中提取声学特征有助于诊断神经系统疾病和监测症状的变化。将音频信号按时间分割成单个单词是提取声学特征前所需的重要预处理步骤。机器学习技术可用于通过自动语音识别(ASR)和序列对序列配准自动进行语音分割。虽然最先进的 ASR 模型在健康语音上取得了良好的性能,但在评估听力障碍语音时,其性能却明显下降。在受损语音上对 ASR 模型进行微调可以提高发育障碍患者的性能,但这需要有代表性的临床数据,而这些数据很难收集,而且可能会引起隐私方面的担忧。本研究探讨了使用两种增强方法提高肢体运动障碍语音的 ASR 性能的可行性:1) 改变健康人的说话速度和响度(病理语音评估中常用的方法);2) 改变说话速度和口音的合成语音(以确保更多样化的声音表现和公平性)。实验评估结果表明,利用这两种来源的数据对预先训练好的 ASR 模型进行微调,其效果优于仅根据真实临床数据进行微调的模型,并且与根据真实临床数据和合成语音组合进行微调的模型效果相当。在对 24 名患有各种神经系统疾病的患者的语音数据进行评估时,表现最好的模型的平均单词错误率为 5.7%,平均正确计数准确率为 94.4%。在将数据分割成单个单词时,与人工解析(地面实况)相比,平均交叉-重合率达到 89.2%。可以得出这样的结论:在微调 ASR 模型时,仿真和合成增强可以大大减少对真实临床语音数据的需求,进而减少对语音分段的需求。
{"title":"Improving Dysarthric Speech Segmentation With Emulated and Synthetic Augmentation","authors":"Saeid Alavi Naeini;Leif Simmatis;Deniz Jafari;Yana Yunusova;Babak Taati","doi":"10.1109/JTEHM.2024.3375323","DOIUrl":"10.1109/JTEHM.2024.3375323","url":null,"abstract":"Acoustic features extracted from speech can help with the diagnosis of neurological diseases and monitoring of symptoms over time. Temporal segmentation of audio signals into individual words is an important pre-processing step needed prior to extracting acoustic features. Machine learning techniques could be used to automate speech segmentation via automatic speech recognition (ASR) and sequence to sequence alignment. While state-of-the-art ASR models achieve good performance on healthy speech, their performance significantly drops when evaluated on dysarthric speech. Fine-tuning ASR models on impaired speech can improve performance in dysarthric individuals, but it requires representative clinical data, which is difficult to collect and may raise privacy concerns. This study explores the feasibility of using two augmentation methods to increase ASR performance on dysarthric speech: 1) healthy individuals varying their speaking rate and loudness (as is often used in assessments of pathological speech); 2) synthetic speech with variations in speaking rate and accent (to ensure more diverse vocal representations and fairness). Experimental evaluations showed that fine-tuning a pre-trained ASR model with data from these two sources outperformed a model fine-tuned only on real clinical data and matched the performance of a model fine-tuned on the combination of real clinical data and synthetic speech. When evaluated on held-out acoustic data from 24 individuals with various neurological diseases, the best performing model achieved an average word error rate of 5.7% and a mean correct count accuracy of 94.4%. In segmenting the data into individual words, a mean intersection-over-union of 89.2% was obtained against manual parsing (ground truth). It can be concluded that emulated and synthetic augmentations can significantly reduce the need for real clinical data of dysarthric speech when fine-tuning ASR models and, in turn, for speech segmentation.","PeriodicalId":54255,"journal":{"name":"IEEE Journal of Translational Engineering in Health and Medicine-Jtehm","volume":"12 ","pages":"382-389"},"PeriodicalIF":3.4,"publicationDate":"2024-03-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10464345","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140105378","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Objective: Blood circulation is an important indicator of wound healing. In this study, a tissue oxygen saturation detecting (TOSD) system that is based on multispectral imaging (MSI) is proposed to quantify the degree of tissue oxygen saturation (StO2) in cutaneous tissues. Methods: A wound segmentation algorithm is used to segment automatically wound and skin areas, eliminating the need for manual labeling and applying adaptive tissue optics. Animal experiments were conducted on six mice in which they were observed seven times, once every two days. The TOSD system illuminated cutaneous tissues with two wavelengths of light - red ( $mathrm {lambda } = 660$