Pub Date : 2023-01-01DOI: 10.1016/j.ibmed.2023.100107
Adrito Das , Danyal Z. Khan , John G. Hanrahan , Hani J. Marcus , Danail Stoyanov
Operation notes are a crucial component of patient care. However, writing them manually is prone to human error, particularly in high pressured clinical environments. Automatic generation of operation notes from video recordings can alleviate some of the administrative burdens, improve accuracy, and provide additional information. To achieve this for endoscopic pituitary surgery, 27-steps were identified via expert consensus. Then, for the 97-videos recorded for this study, a timestamp of each step was annotated by an expert surgeon. To automatically determine whether a step is present in a video, a three-stage architecture was created. Firstly, for each step, a convolution neural network was used for binary image classification on each frame of a video. Secondly, for each step, the binary frame classifications were passed to a discriminator for binary video classification. Thirdly, for each video, the binary video classifications were passed to an accumulator for multi-label step classification. The architecture was trained on 77-videos, and tested on 20-videos, where a 0.80 weighted-F1 score was achieved. The classifications were inputted into a clinically based predefined template, and further enriched with additional video analytics. This work therefore demonstrates automatic generation of operative notes from surgical videos is feasible, and can assist surgeons during documentation.
{"title":"Automatic generation of operation notes in endoscopic pituitary surgery videos using workflow recognition","authors":"Adrito Das , Danyal Z. Khan , John G. Hanrahan , Hani J. Marcus , Danail Stoyanov","doi":"10.1016/j.ibmed.2023.100107","DOIUrl":"https://doi.org/10.1016/j.ibmed.2023.100107","url":null,"abstract":"<div><p>Operation notes are a crucial component of patient care. However, writing them manually is prone to human error, particularly in high pressured clinical environments. Automatic generation of operation notes from video recordings can alleviate some of the administrative burdens, improve accuracy, and provide additional information. To achieve this for endoscopic pituitary surgery, 27-steps were identified via expert consensus. Then, for the 97-videos recorded for this study, a timestamp of each step was annotated by an expert surgeon. To automatically determine whether a step is present in a video, a three-stage architecture was created. Firstly, for each step, a convolution neural network was used for binary image classification on each frame of a video. Secondly, for each step, the binary frame classifications were passed to a discriminator for binary video classification. Thirdly, for each video, the binary video classifications were passed to an accumulator for multi-label step classification. The architecture was trained on 77-videos, and tested on 20-videos, where a 0.80 weighted-<em>F</em><sub>1</sub> score was achieved. The classifications were inputted into a clinically based predefined template, and further enriched with additional video analytics. This work therefore demonstrates automatic generation of operative notes from surgical videos is feasible, and can assist surgeons during documentation.</p></div>","PeriodicalId":73399,"journal":{"name":"Intelligence-based medicine","volume":"8 ","pages":"Article 100107"},"PeriodicalIF":0.0,"publicationDate":"2023-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49869238","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-01-01DOI: 10.1016/j.ibmed.2023.100108
Shahrzad Moinian , Nyoman D. Kurniawan , Shekhar S. Chandra , Viktor Vegh , David C. Reutens
A primary challenge for in vivo kidney magnetic resonance imaging (MRI) is the presence of different types of involuntary physiological motion, affecting the diagnostic utility of acquired images due to severe motion artifacts. Existing prospective and retrospective motion correction methods remain ineffective when dealing with complex large amplitude nonrigid motion artifacts. Here, we introduce an unsupervised deep learning-based image to image translation method between motion-affected and motion-free image domains, for correction of rigid-body, respiratory and nonrigid motion artifacts in vivo kidney MRI.
High resolution (i.e., 156 × 156 × 370 μm) ex vivo 3 Tesla MRI scans of 13 porcine kidneys (because of their anatomical homology to human kidney) were conducted using a 3D T2-weighted turbo spin echo sequence. Rigid-body, respiratory and nonrigid motion-affected images were then simulated using the magnitude-only ex vivo motion-free image set. Each 2D coronal slice of motion-affected and motion-free image volume was then divided into patches of 128 × 128 for training the model. We proposed to add normalised cross-correlation loss to cycle consistency generative adversarial network structure (NCC-CycleGAN), to enforce edge alignment between motion-corrected and motion-free image domains.
Our NCC-CycleGAN motion correction model demonstrated high performance with an in-tissue structural similarity index measure of 0.77 ± 0.08, peak signal-to-noise ratio of 26.67 ± 3.44 and learned perceptual image patch similarity of 0.173 ± 0.05 between the reconstructed motion-corrected and ground truth motion-free images. This corresponds to a significant respective average improvement of 34%, 23% and 39% (p < 0.05; paired t-test) for the three metrics to correct the three different types of simulated motion artifacts.
We demonstrated the feasibility of developing an unsupervised deep learning-based method for efficient automated retrospective kidney MRI motion correction, while preserving microscopic tissue structures in high resolution imaging.
{"title":"An unsupervised deep learning-based image translation method for retrospective motion correction of high resolution kidney MRI","authors":"Shahrzad Moinian , Nyoman D. Kurniawan , Shekhar S. Chandra , Viktor Vegh , David C. Reutens","doi":"10.1016/j.ibmed.2023.100108","DOIUrl":"https://doi.org/10.1016/j.ibmed.2023.100108","url":null,"abstract":"<div><p>A primary challenge for <em>in vivo</em> kidney magnetic resonance imaging (MRI) is the presence of different types of involuntary physiological motion, affecting the diagnostic utility of acquired images due to severe motion artifacts. Existing prospective and retrospective motion correction methods remain ineffective when dealing with complex large amplitude nonrigid motion artifacts. Here, we introduce an unsupervised deep learning-based image to image translation method between motion-affected and motion-free image domains, for correction of rigid-body, respiratory and nonrigid motion artifacts in <em>vivo</em> kidney MRI.</p><p>High resolution (i.e., 156 × 156 × 370 μm) <em>ex vivo</em> 3 Tesla MRI scans of 13 porcine kidneys (because of their anatomical homology to human kidney) were conducted using a 3D T2-weighted turbo spin echo sequence. Rigid-body, respiratory and nonrigid motion-affected images were then simulated using the magnitude-only <em>ex vivo</em> motion-free image set. Each 2D coronal slice of motion-affected and motion-free image volume was then divided into patches of 128 × 128 for training the model. We proposed to add normalised cross-correlation loss to cycle consistency generative adversarial network structure (NCC-CycleGAN), to enforce edge alignment between motion-corrected and motion-free image domains.</p><p>Our NCC-CycleGAN motion correction model demonstrated high performance with an in-tissue structural similarity index measure of 0.77 ± 0.08, peak signal-to-noise ratio of 26.67 ± 3.44 and learned perceptual image patch similarity of 0.173 ± 0.05 between the reconstructed motion-corrected and ground truth motion-free images. This corresponds to a significant respective average improvement of 34%, 23% and 39% (p < 0.05; paired <em>t</em>-test) for the three metrics to correct the three different types of simulated motion artifacts.</p><p>We demonstrated the feasibility of developing an unsupervised deep learning-based method for efficient automated retrospective kidney MRI motion correction, while preserving microscopic tissue structures in high resolution imaging.</p></div>","PeriodicalId":73399,"journal":{"name":"Intelligence-based medicine","volume":"8 ","pages":"Article 100108"},"PeriodicalIF":0.0,"publicationDate":"2023-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49869240","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-01-01DOI: 10.1016/j.ibmed.2023.100113
Mahmoud Elmahdy, Ronnie Sebro
The use of artificial intelligence (AI) programs in healthcare and medicine has steadily increased over the past decade. One major challenge affecting the use of AI programs is that the results of AI programs are sometimes not replicable, meaning that the performance of the AI program is substantially different in the external testing dataset when compared to its performance in the training or validation datasets. This often happens when the external testing dataset is very different from the training or validation datasets. Sex, ethnicity, and race are some of the most important biological and social determinants of health, and are important factors that may differ between training, validation, and external testing datasets, and may contribute to the lack of reproducibility of AI programs. We reviewed over 28,000 original research articles published in the three journals with the highest impact factors in each of 16 medical specialties between 2019 and 2022, to evaluate how often the sex, ethnic, and racial compositions of the datasets used to develop AI algorithms were reported. We also reviewed all currently used AI reporting guidelines, to evaluate which guidelines recommend specific reporting of sex, ethnicity, and race. We find that only 42.47 % (338/797) of articles reported sex, 1.4 % (12/831) reported ethnicity, and 7.3 % (61/831) reported race. When sex was reported, approximately 55.8 % of the study participants were female, and when ethnicity was reported, only 6.2 % of the study participants were Hispanic/Latino. When race was reported, only 29.4 % of study participants were non-White. Most AI guidelines (93.3 %; 14/15) also did not recommend reporting sex, ethnicity, and race. To have fair and ethnical AI, it is important that the sex, ethnic, and racial compositions of the datasets used to develop the AI program are known.
{"title":"Sex, ethnicity, and race data are often unreported in artificial intelligence and machine learning studies in medicine","authors":"Mahmoud Elmahdy, Ronnie Sebro","doi":"10.1016/j.ibmed.2023.100113","DOIUrl":"https://doi.org/10.1016/j.ibmed.2023.100113","url":null,"abstract":"<div><p>The use of artificial intelligence (AI) programs in healthcare and medicine has steadily increased over the past decade. One major challenge affecting the use of AI programs is that the results of AI programs are sometimes not replicable, meaning that the performance of the AI program is substantially different in the external testing dataset when compared to its performance in the training or validation datasets. This often happens when the external testing dataset is very different from the training or validation datasets. Sex, ethnicity, and race are some of the most important biological and social determinants of health, and are important factors that may differ between training, validation, and external testing datasets, and may contribute to the lack of reproducibility of AI programs. We reviewed over 28,000 original research articles published in the three journals with the highest impact factors in each of 16 medical specialties between 2019 and 2022, to evaluate how often the sex, ethnic, and racial compositions of the datasets used to develop AI algorithms were reported. We also reviewed all currently used AI reporting guidelines, to evaluate which guidelines recommend specific reporting of sex, ethnicity, and race. We find that only 42.47 % (338/797) of articles reported sex, 1.4 % (12/831) reported ethnicity, and 7.3 % (61/831) reported race. When sex was reported, approximately 55.8 % of the study participants were female, and when ethnicity was reported, only 6.2 % of the study participants were Hispanic/Latino. When race was reported, only 29.4 % of study participants were non-White. Most AI guidelines (93.3 %; 14/15) also did not recommend reporting sex, ethnicity, and race. To have fair and ethnical AI, it is important that the sex, ethnic, and racial compositions of the datasets used to develop the AI program are known.</p></div>","PeriodicalId":73399,"journal":{"name":"Intelligence-based medicine","volume":"8 ","pages":"Article 100113"},"PeriodicalIF":0.0,"publicationDate":"2023-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2666521223000273/pdfft?md5=070012db350ebd8eac9219c40819eaa8&pid=1-s2.0-S2666521223000273-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"92045506","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-01-01DOI: 10.1016/j.ibmed.2023.100123
Amjed Al Fahoum, Ala’a Zyout
The scientific literature on depression detection using electroencephalogram (EEG) signals is extensive and offers numerous innovative approaches. However, these existing state-of-the-art (SOTA) have limitations that hinder their overall efficacy. They rely significantly on datasets with limited scope and accessibility, which introduces potential biases and diminishes generalizability. In addition, they concentrate on analyzing a single dataset, potentially overlooking the inherent variability and complexity of EEG patterns associated with depression. Moreover, certain SOTA methods employ deep learning architectures with exponential time complexity, resulting in computationally intensive and time-consuming training procedures. Therefore, their practicability and applicability in real-world scenarios are compromised. To address these limitations, a novel integrated methodology that combines the advantages of phase space reconstruction and deep neural networks is proposed. It employs publicly available EEG datasets, mitigating the inherent biases of exclusive data sources. Moreover, the method incorporates reconstructed phase space analysis, a feature engineering technique that captures more accurately the complex EEG patterns associated with depression. Simultaneously, the incorporation of a deep neural network component guarantees optimal efficiency and accurate, seamless classification. Using publicly available datasets, cross-dataset validation, and a novel combination of reconstructed phase space analysis and deep neural networks, the proposed method circumvents the shortcomings of current state-of-the-art (SOTA) approaches. This innovation represents a significant advance in enhancing the accuracy of depression detection and provides the base for EEG-based depression assessment tools applicable to real-world settings. The findings of the study provide a more robust and efficient model, which increases classification precision and decreases computing burden. The study findings layout the foundation for scalable, accessible mental health solutions, identification of the pathological deficits in affected brain tissues, and demonstrate the potential of technology-driven approaches to support and guide depressed individuals and enhance mental health outcomes.
{"title":"Early detection of neurological abnormalities using a combined phase space reconstruction and deep learning approach","authors":"Amjed Al Fahoum, Ala’a Zyout","doi":"10.1016/j.ibmed.2023.100123","DOIUrl":"https://doi.org/10.1016/j.ibmed.2023.100123","url":null,"abstract":"<div><p>The scientific literature on depression detection using electroencephalogram (EEG) signals is extensive and offers numerous innovative approaches. However, these existing state-of-the-art (SOTA) have limitations that hinder their overall efficacy. They rely significantly on datasets with limited scope and accessibility, which introduces potential biases and diminishes generalizability. In addition, they concentrate on analyzing a single dataset, potentially overlooking the inherent variability and complexity of EEG patterns associated with depression. Moreover, certain SOTA methods employ deep learning architectures with exponential time complexity, resulting in computationally intensive and time-consuming training procedures. Therefore, their practicability and applicability in real-world scenarios are compromised. To address these limitations, a novel integrated methodology that combines the advantages of phase space reconstruction and deep neural networks is proposed. It employs publicly available EEG datasets, mitigating the inherent biases of exclusive data sources. Moreover, the method incorporates reconstructed phase space analysis, a feature engineering technique that captures more accurately the complex EEG patterns associated with depression. Simultaneously, the incorporation of a deep neural network component guarantees optimal efficiency and accurate, seamless classification. Using publicly available datasets, cross-dataset validation, and a novel combination of reconstructed phase space analysis and deep neural networks, the proposed method circumvents the shortcomings of current state-of-the-art (SOTA) approaches. This innovation represents a significant advance in enhancing the accuracy of depression detection and provides the base for EEG-based depression assessment tools applicable to real-world settings. The findings of the study provide a more robust and efficient model, which increases classification precision and decreases computing burden. The study findings layout the foundation for scalable, accessible mental health solutions, identification of the pathological deficits in affected brain tissues, and demonstrate the potential of technology-driven approaches to support and guide depressed individuals and enhance mental health outcomes.</p></div>","PeriodicalId":73399,"journal":{"name":"Intelligence-based medicine","volume":"8 ","pages":"Article 100123"},"PeriodicalIF":0.0,"publicationDate":"2023-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2666521223000376/pdfft?md5=49e91661bcf14318d233d3bb140064a2&pid=1-s2.0-S2666521223000376-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138466610","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-01-01DOI: 10.1016/j.ibmed.2023.100112
Yoon-Seop Kim , Min Woong Kim , Je Seop Lee , Hee Seung Kang , Erdenebayar Urtnasan , Jung Woo Lee , Ji Hun Kim
Background
In military settings, determining whether a patient with abdominal pain requires emergency care can be challenging due to the absence or inexperience of medical staff. Misjudging the severity of abdominal pain can lead to delayed treatment or unnecessary transfers, both of which consume valuable resources. Therefore, our aim was to develop an artificial intelligence model capable of classifying the urgency of abdominal pain cases, taking into account patient characteristics.
Methods
We collected structured and unstructured data from patients with abdominal pain visiting South Korean military hospital emergency rooms between January 2015 and 2020. After excluding patients with missing values, 20,432 patients were enrolled. Structured data consisted of age, sex, vital signs, past medical history, and symptoms, while unstructured data included preprocessed free text descriptions of chief complaints and present illness. Patients were divided into training, validation, and test datasets in an 8:1:1 ratio. Using structured data, we developed four conventional machine learning models and a novel mixed model, which combined one of the best performing machine learning models with emergency medical knowledge. And we also created a deep learning model using both structured and unstructured data.
Results
Xgboost demonstrated the highest performance among the six models, with an area under the precision-recall curve (AUPRC) score of 0.61. The other five models achieved AUPRC scores as follows: logistic regression (0.24), decision tree (0.22), multi-layer perceptron (0.21), deep neural network (0.58), and mixed model (0.58).
Conclusion
This study is the first to develop an AI model for identifying emergency cases of abdominal pain in a military setting. With more balanced and better-structured datasets, clinically significant AI model could be developed based on the findings of this study.
{"title":"Development of an artificial intelligence model for triage in a military emergency department: Focusing on abdominal pain in soldiers","authors":"Yoon-Seop Kim , Min Woong Kim , Je Seop Lee , Hee Seung Kang , Erdenebayar Urtnasan , Jung Woo Lee , Ji Hun Kim","doi":"10.1016/j.ibmed.2023.100112","DOIUrl":"https://doi.org/10.1016/j.ibmed.2023.100112","url":null,"abstract":"<div><h3>Background</h3><p>In military settings, determining whether a patient with abdominal pain requires emergency care can be challenging due to the absence or inexperience of medical staff. Misjudging the severity of abdominal pain can lead to delayed treatment or unnecessary transfers, both of which consume valuable resources. Therefore, our aim was to develop an artificial intelligence model capable of classifying the urgency of abdominal pain cases, taking into account patient characteristics.</p></div><div><h3>Methods</h3><p>We collected structured and unstructured data from patients with abdominal pain visiting South Korean military hospital emergency rooms between January 2015 and 2020. After excluding patients with missing values, 20,432 patients were enrolled. Structured data consisted of age, sex, vital signs, past medical history, and symptoms, while unstructured data included preprocessed free text descriptions of chief complaints and present illness. Patients were divided into training, validation, and test datasets in an 8:1:1 ratio. Using structured data, we developed four conventional machine learning models and a novel mixed model, which combined one of the best performing machine learning models with emergency medical knowledge. And we also created a deep learning model using both structured and unstructured data.</p></div><div><h3>Results</h3><p>Xgboost demonstrated the highest performance among the six models, <u>with an area under the precision-recall curve (AUPRC) score of 0.61. The other five models achieved AUPRC scores as follows: logistic regression (0.24), decision tree (0.22), multi-layer perceptron (0.21), deep neural network (0.58), and mixed model (0.58).</u></p></div><div><h3>Conclusion</h3><p>This study is the first to develop an AI model for identifying emergency cases of abdominal pain in a military setting. With more balanced and better-structured datasets, clinically significant AI model could be developed based on the findings of this study.</p></div>","PeriodicalId":73399,"journal":{"name":"Intelligence-based medicine","volume":"8 ","pages":"Article 100112"},"PeriodicalIF":0.0,"publicationDate":"2023-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49869181","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-01-01DOI: 10.1016/j.ibmed.2023.100096
Ean S. Bett , Timothy C. Frommeyer , Tejaswini Reddy , James “Ty” Johnson
Background
Electronic health records (EHR) were implemented to improve patient care, reduce healthcare disparities, engage patients and families, improve care coordination, and maintain privacy and security. Unfortunately, the mandated use of EHR has also resulted in significantly increased clerical and administrative burden, with physicians spending an estimated three-fourths of their daily time interacting with the EHR, which negatively affects within-clinic processes and contributes to burnout. In-room scribes are associated with improvement in all aspects of physician satisfaction and increased productivity, though less is known about the use of other technologies such as Google Glass (GG), Natural Language Processing (NLP) and Machine-Based Learning (MBL) systems. Given the need to decrease administrative burden on clinicians, particularly in the utilization of the EHR, there is a need to explore the intersection between varying degrees of technology in the clinical encounter and their ability to meet the aforementioned goals of the EHR.
Aims
The primary aim is to determine predictors of overall perception of care dependent on varying mechanisms used for documentation and medical decision-making in a routine clinical encounter. Secondary aims include comparing the perception of individual vignettes based on demographics of the participants and investigating any differences in perception questions by demographics of the participants.
Methods
Video vignettes were shown to 498 OhioHealth Physician Group patients and to ResearchMatch volunteers during a 15-month period following IRB approval. Data included a baseline survey to gather demographic and background characteristics and then a perceptual survey where patients rated the physician in the video on 5 facets using a 1 to 5 Likert scale. The analysis included summarizing data of all continuous and categorical variables as well as overall perceptions analyzed using multivariate linear regression with perception score as the outcome variable.
Results
Univariate modeling identified sex, education, and type of technology as three factors that were statistically significantly related to the overall perception score. Males had higher scores than females (p = 0.03) and those with lower education had higher scores (p < 0.001). In addition, the physician documenting outside of the room encounter had statistically significantly higher overall perception scores (mean = 22.2, p < 0.001) and the physician documenting in the room encounter had statistically significantly lower overall perception scores (mean = 15.3, p < 0.001) when compared to the other vignettes. Multivariable modeling identified all three of the univariably significant factors as independent factors related to overall perception score. Specifically, high school education had higher scores than associate/bachelor education (LSM = 21.6 vs.
实施电子健康记录(EHR)是为了改善患者护理,减少医疗保健差距,吸引患者和家庭参与,改善护理协调,并维护隐私和安全。不幸的是,电子病历的强制使用也导致了文书和行政负担的显著增加,医生每天花费大约四分之三的时间与电子病历互动,这对诊所内的流程产生了负面影响,并导致了职业倦怠。尽管人们对谷歌Glass (GG)、自然语言处理(NLP)和机器学习(MBL)系统等其他技术的使用知之甚少,但室内誊写员与医生满意度的提高和工作效率的提高有关。考虑到需要减轻临床医生的行政负担,特别是在电子病历的使用方面,有必要探索临床遇到的不同程度的技术与他们实现上述电子病历目标的能力之间的交集。目的主要目的是确定在常规临床遇到的不同机制中,依赖于文件和医疗决策的总体护理感知的预测因子。次要目的包括根据参与者的人口统计数据比较个人小插曲的感知,并根据参与者的人口统计数据调查感知问题的任何差异。方法在IRB批准后的15个月期间,向498名俄亥俄健康医师组患者和ResearchMatch志愿者播放视频片段。数据包括一项收集人口统计和背景特征的基线调查,然后是一项感性调查,其中患者使用1到5的李克特量表从5个方面对视频中的医生进行评分。分析包括总结所有连续变量和分类变量的数据,以及使用以感知评分为结果变量的多元线性回归分析总体感知。结果单变量模型确定性别、教育程度和技术类型是与总体感知得分有统计学显著相关的三个因素。男性得分高于女性(p = 0.03),受教育程度较低者得分较高(p <0.001)。此外,记录房间外遭遇的医生有统计学上显著更高的总体感知得分(平均= 22.2,p <0.001),在病房就诊的医生的总体感知得分在统计学上显著降低(平均= 15.3,p <0.001),与其他小插曲相比。多变量模型确定了所有三个不可变显著因素作为与整体感知得分相关的独立因素。具体而言,高中教育的得分高于副学士/学士教育(LSM = 21.6 vs. 19.9, p = 0.0002),高于硕士/高等教育(LSM = 21.6 vs. 19.5, p <0.0001)。在个人感知得分上,各组之间没有发现差异。男性在“医生向患者清楚地解释了诊断和治疗”和“医生真诚可靠”方面得分高于女性。高中学历的人在所有五项个人感知得分上都高于副学士/学士和硕士。结论:研究发现,性别、教育程度和技术类型是常规临床遭遇中用于记录和医疗决策的不同技术的总体感知的重要指标。重要的是,描述与电子病历互动最少的小插图获得了最积极的总体感知得分,而描述医生在互动过程中利用电子病历的小插图获得了最不积极的总体感知得分。这表明,只要患者在互动过程中感到参与,他们最重视医生的充分关注,而对区分数据转录和医疗决策的后勤工作感觉不那么强烈。因此,作者建议最大限度地将面对面的时间整合到临床接触中,允许在医患互动中增加个人注意力的感知。
{"title":"Assessment of patient perceptions of technology and the use of machine-based learning in a clinical encounter","authors":"Ean S. Bett , Timothy C. Frommeyer , Tejaswini Reddy , James “Ty” Johnson","doi":"10.1016/j.ibmed.2023.100096","DOIUrl":"https://doi.org/10.1016/j.ibmed.2023.100096","url":null,"abstract":"<div><h3>Background</h3><p>Electronic health records (EHR) were implemented to improve patient care, reduce healthcare disparities, engage patients and families, improve care coordination, and maintain privacy and security. Unfortunately, the mandated use of EHR has also resulted in significantly increased clerical and administrative burden, with physicians spending an estimated three-fourths of their daily time interacting with the EHR, which negatively affects within-clinic processes and contributes to burnout. In-room scribes are associated with improvement in all aspects of physician satisfaction and increased productivity, though less is known about the use of other technologies such as Google Glass (GG), Natural Language Processing (NLP) and Machine-Based Learning (MBL) systems. Given the need to decrease administrative burden on clinicians, particularly in the utilization of the EHR, there is a need to explore the intersection between varying degrees of technology in the clinical encounter and their ability to meet the aforementioned goals of the EHR.</p></div><div><h3>Aims</h3><p>The primary aim is to determine predictors of overall perception of care dependent on varying mechanisms used for documentation and medical decision-making in a routine clinical encounter. Secondary aims include comparing the perception of individual vignettes based on demographics of the participants and investigating any differences in perception questions by demographics of the participants.</p></div><div><h3>Methods</h3><p>Video vignettes were shown to 498 OhioHealth Physician Group patients and to ResearchMatch volunteers during a 15-month period following IRB approval. Data included a baseline survey to gather demographic and background characteristics and then a perceptual survey where patients rated the physician in the video on 5 facets using a 1 to 5 Likert scale. The analysis included summarizing data of all continuous and categorical variables as well as overall perceptions analyzed using multivariate linear regression with perception score as the outcome variable.</p></div><div><h3>Results</h3><p>Univariate modeling identified sex, education, and type of technology as three factors that were statistically significantly related to the overall perception score. Males had higher scores than females (p = 0.03) and those with lower education had higher scores (p < 0.001). In addition, the physician documenting outside of the room encounter had statistically significantly higher overall perception scores (mean = 22.2, p < 0.001) and the physician documenting in the room encounter had statistically significantly lower overall perception scores (mean = 15.3, p < 0.001) when compared to the other vignettes. Multivariable modeling identified all three of the univariably significant factors as independent factors related to overall perception score. Specifically, high school education had higher scores than associate/bachelor education (LSM = 21.6 vs. ","PeriodicalId":73399,"journal":{"name":"Intelligence-based medicine","volume":"7 ","pages":"Article 100096"},"PeriodicalIF":0.0,"publicationDate":"2023-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49857365","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-01-01DOI: 10.1016/j.ibmed.2023.100109
Jesse Ehrenfeld
The American Medical Association's latest survey on digital health trends showed that adoption of digital tools has grown significantly in the past 3–4 yrs, among all physicians, regardless of gender, specialty or age. This article examines digital health trends, including AI, and the AMA's role in ensuring that physicians are actively involved in the creation of new technologies and innovations in medicine. Physicians understand the potential for new digital tools to address health disparities for patients and streamline our workflow better than anyone.
{"title":"Physician leadership in the new era of AI and digital health tools","authors":"Jesse Ehrenfeld","doi":"10.1016/j.ibmed.2023.100109","DOIUrl":"https://doi.org/10.1016/j.ibmed.2023.100109","url":null,"abstract":"<div><p>The American Medical Association's latest survey on digital health trends showed that adoption of digital tools has grown significantly in the past 3–4 yrs, among all physicians, regardless of gender, specialty or age. This article examines digital health trends, including AI, and the AMA's role in ensuring that physicians are actively involved in the creation of new technologies and innovations in medicine. Physicians understand the potential for new digital tools to address health disparities for patients and streamline our workflow better than anyone.</p></div>","PeriodicalId":73399,"journal":{"name":"Intelligence-based medicine","volume":"8 ","pages":"Article 100109"},"PeriodicalIF":0.0,"publicationDate":"2023-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49869239","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-01-01DOI: 10.1016/j.ibmed.2023.100120
Luigi De Angelis , Francesco Baglivo , Guglielmo Arzilli , Leonardo Calamita , Paolo Ferragina , Gaetano Pierpaolo Privitera , Caterina Rizzo
{"title":"Hospital-acquired infections surveillance and prevention: using Natural Language Processing to analyze unstructured text of hospital discharge letters for surgical site infections identification and risk-stratification.","authors":"Luigi De Angelis , Francesco Baglivo , Guglielmo Arzilli , Leonardo Calamita , Paolo Ferragina , Gaetano Pierpaolo Privitera , Caterina Rizzo","doi":"10.1016/j.ibmed.2023.100120","DOIUrl":"https://doi.org/10.1016/j.ibmed.2023.100120","url":null,"abstract":"","PeriodicalId":73399,"journal":{"name":"Intelligence-based medicine","volume":"8 ","pages":"Article 100120"},"PeriodicalIF":0.0,"publicationDate":"2023-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2666521223000340/pdfft?md5=4d1203c98598473fd6b2a278e08e5243&pid=1-s2.0-S2666521223000340-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138558703","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-01-01DOI: 10.1016/j.ibmed.2023.100124
Anwesha Mohanty, Alistair Sutherland, Marija Bezbradica, Hossein Javidnia
Within the realm of medical diagnosis, deep learning techniques have revolutionized the way diseases are identified and studied. However, a persistent challenge has been data scarcity for many disease categories. One primary reason for this is issues related to patient privacy and copyright constraints on medical datasets. To address this, our research explores the use of synthetic data generation, focusing on Rhinophyma, a subclass of Rosacea. Our novel approach uses 3D parametric modeling to create synthetic images of Rhinophyma, addressing the data scarcity problem. Through this method, we generated 20,000 images representing 2000 distinct anatomical deformations of Rhinophyma. This research not only showcases the potential of using 3D parametric modeling for Rhinophyma but hints at its applicability for other diseases with anatomical abnormalities. With just 30 % of this synthetic dataset, we achieved a remarkable 95 % recall in classifying 220 real-world Rhinophyma images. The performance of our classification model is further validated using GradCAM visualisation. Our findings underscore the potential of such techniques to propel medical research and develop superior deep learning diagnostic models when only limited real-world images are available.
{"title":"Rhi3DGen: Analyzing Rhinophyma using 3D face models and synthetic data","authors":"Anwesha Mohanty, Alistair Sutherland, Marija Bezbradica, Hossein Javidnia","doi":"10.1016/j.ibmed.2023.100124","DOIUrl":"https://doi.org/10.1016/j.ibmed.2023.100124","url":null,"abstract":"<div><p>Within the realm of medical diagnosis, deep learning techniques have revolutionized the way diseases are identified and studied. However, a persistent challenge has been data scarcity for many disease categories. One primary reason for this is issues related to patient privacy and copyright constraints on medical datasets. To address this, our research explores the use of synthetic data generation, focusing on <em>Rhinophyma</em>, a subclass of Rosacea. Our novel approach uses 3D parametric modeling to create synthetic images of <em>Rhinophyma</em>, addressing the data scarcity problem. Through this method, we generated 20,000 images representing 2000 distinct anatomical deformations of <em>Rhinophyma</em>. This research not only showcases the potential of using 3D parametric modeling for <em>Rhinophyma</em> but hints at its applicability for other diseases with anatomical abnormalities. With just 30 % of this synthetic dataset, we achieved a remarkable 95 % recall in classifying 220 real-world <em>Rhinophyma</em> images. The performance of our classification model is further validated using GradCAM visualisation. Our findings underscore the potential of such techniques to propel medical research and develop superior deep learning diagnostic models when only limited real-world images are available.</p></div>","PeriodicalId":73399,"journal":{"name":"Intelligence-based medicine","volume":"8 ","pages":"Article 100124"},"PeriodicalIF":0.0,"publicationDate":"2023-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2666521223000388/pdfft?md5=7db4c86694fa7c487ae887d65e0fc36c&pid=1-s2.0-S2666521223000388-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138465808","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
This work investigates the performance of different machine learning (ML) methods for classifying postmenopausal osteoporosis Thai patients. Our dataset contains 377 samples compiled retrospectively using the medical records of a Thai woman in the postmenopause stage from the obstetrics and gynecology clinic, Ramathibodi Hospital, Bangkok, Thailand. Missing data imputation, feature selection, and handling imbalanced techniques are independently applied as pre-processing approaches. The performance of different ML algorithms, including k-nearest neighbors (k-NN), neural network (NN), naïve Bayesian (NB), Bayesian network (BN), support vector machine (SVM), random forest (RF), and decision tree (DT), is compared between the pre-processed and original data. The results demonstrate that different ML algorithms combined with pre-processing techniques achieve varying results. In terms of accuracy, the three best-performing methods are the NN, NB, and RF models when a wrapper approach is used with an appropriate learner. In terms of specificity, the DT model achieves the best performance when the synthetic minority oversampling technique method is applied. When feature selection techniques are applied, the k-NN, BN, and SVM algorithms obtain the best sensitivity, whereas the NN shows the best area under the curve. Overall, in comparison with the original dataset, the pre-processed approaches improved model performance. Therefore, proper pre-processing techniques should be considered when developing ML classifiers to identify the best appropriate model.
{"title":"Machine learning's performance in classifying postmenopausal osteoporosis Thai patients","authors":"Kittisak Thawnashom , Pornsarp Pornsawad , Bunjira Makond","doi":"10.1016/j.ibmed.2023.100099","DOIUrl":"https://doi.org/10.1016/j.ibmed.2023.100099","url":null,"abstract":"<div><p>This work investigates the performance of different machine learning (ML) methods for classifying postmenopausal osteoporosis Thai patients. Our dataset contains 377 samples compiled retrospectively using the medical records of a Thai woman in the postmenopause stage from the obstetrics and gynecology clinic, Ramathibodi Hospital, Bangkok, Thailand. Missing data imputation, feature selection, and handling imbalanced techniques are independently applied as pre-processing approaches. The performance of different ML algorithms, including <em>k</em>-nearest neighbors (<em>k</em>-NN), neural network (NN), naïve Bayesian (NB), Bayesian network (BN), support vector machine (SVM), random forest (RF), and decision tree (DT), is compared between the pre-processed and original data. The results demonstrate that different ML algorithms combined with pre-processing techniques achieve varying results. In terms of accuracy, the three best-performing methods are the NN, NB, and RF models when a wrapper approach is used with an appropriate learner. In terms of specificity, the DT model achieves the best performance when the synthetic minority oversampling technique method is applied. When feature selection techniques are applied, the <em>k</em>-NN, BN, and SVM algorithms obtain the best sensitivity, whereas the NN shows the best area under the curve. Overall, in comparison with the original dataset, the pre-processed approaches improved model performance. Therefore, proper pre-processing techniques should be considered when developing ML classifiers to identify the best appropriate model.</p></div>","PeriodicalId":73399,"journal":{"name":"Intelligence-based medicine","volume":"7 ","pages":"Article 100099"},"PeriodicalIF":0.0,"publicationDate":"2023-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49857363","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}