首页 > 最新文献

Methods of Information in Medicine最新文献

英文 中文
Automated Information Extraction from Unstructured Hematopathology Reports to Support Response Assessment in Myeloproliferative Neoplasms. 从非结构化血液病报告中自动提取信息以支持骨髓增生性肿瘤的反应评估。
IF 1.3 4区 医学 Q3 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2024-12-01 Epub Date: 2025-04-17 DOI: 10.1055/a-2590-6456
Spencer Krichevsky, Evan T Sholle, Prakash M Adekkanattu, Sajjad Abedian, Madhu Ouseph, Elwood Taylor, Ghaith Abu-Zeinah, Diana Jaber, Claudia Sosner, Marika M Cusick, Niamh Savage, Richard T Silver, Joseph M Scandura, Thomas R Campion

Assessing treatment response in patients with myeloproliferative neoplasms is difficult because data components exist in unstructured bone marrow pathology (hematopathology) reports, which require specialized, manual annotation, and interpretation. Although natural language processing (NLP) has been successfully implemented for the extraction of features from solid tumor reports, little is known about its application to hematopathology.An open-source NLP framework called Leo was implemented to parse document segments and extract concept phrases utilized for assessing responses in myeloproliferative neoplasms. A reference standard was generated through the manual review of hematopathology notes.Compared with a reference standard (n = 300 reports), our NLP method extracted features such as aspirate myeloblasts (F1 = 98%) and biopsy reticulin fibrosis (F1 = 93%) with high accuracy. However, other values, such as myeloblasts from the biopsy (F1 = 6%) and via flow cytometry (F1 = 8%), were affected by sparsity representative of reporting conventions. The four features with the highest clinical importance were extracted with F1 scores exceeding 90%. Whereas manual annotation of 300 reports required 30 hours of staff effort, automated NLP required 3.5 hours of runtime for 34,301 reports.To the best of our knowledge, this is among the first studies to demonstrate the application of NLP to hematopathology for clinical feature extraction. The approach may inform efforts at other institutions, and the code is available at https://github.com/wcmc-research-informatics/BmrExtractor.

评估骨髓增生性肿瘤患者的治疗反应是困难的,因为数据成分存在于非结构化的骨髓病理学(血液病理学)报告中,这些报告需要专门的手工注释和解释。虽然自然语言处理(NLP)已经成功地用于实体肿瘤报告的特征提取,但其在血液病理学中的应用尚不清楚。实现了一个名为Leo的开源NLP框架,用于解析文档片段并提取用于评估骨髓增殖性肿瘤反应的概念短语。参考标准是通过手工检查血液病记录生成的。与参考标准(n = 300份报告)相比,我们的NLP方法提取吸出性成髓细胞(F1 = 98%)和活检网状蛋白纤维化(F1 = 93%)等特征的准确性较高。然而,其他值,如来自活检(F1 = 6%)和流式细胞术(F1 = 8%)的成髓细胞,受到报告惯例的稀疏性代表的影响。提取临床重要性最高的4个特征,F1评分超过90%。手动注释300个报告需要30小时的工作时间,而自动NLP需要3.5小时的运行时间来处理34,301个报告。据我们所知,这是第一个将自然语言处理应用于血液病理学临床特征提取的研究。该方法可以为其他机构的工作提供信息,代码可在https://github.com/wcmc-research-informatics/BmrExtractor上获得。
{"title":"Automated Information Extraction from Unstructured Hematopathology Reports to Support Response Assessment in Myeloproliferative Neoplasms.","authors":"Spencer Krichevsky, Evan T Sholle, Prakash M Adekkanattu, Sajjad Abedian, Madhu Ouseph, Elwood Taylor, Ghaith Abu-Zeinah, Diana Jaber, Claudia Sosner, Marika M Cusick, Niamh Savage, Richard T Silver, Joseph M Scandura, Thomas R Campion","doi":"10.1055/a-2590-6456","DOIUrl":"10.1055/a-2590-6456","url":null,"abstract":"<p><p>Assessing treatment response in patients with myeloproliferative neoplasms is difficult because data components exist in unstructured bone marrow pathology (hematopathology) reports, which require specialized, manual annotation, and interpretation. Although natural language processing (NLP) has been successfully implemented for the extraction of features from solid tumor reports, little is known about its application to hematopathology.An open-source NLP framework called Leo was implemented to parse document segments and extract concept phrases utilized for assessing responses in myeloproliferative neoplasms. A reference standard was generated through the manual review of hematopathology notes.Compared with a reference standard (<i>n</i> = 300 reports), our NLP method extracted features such as aspirate myeloblasts (F1 = 98%) and biopsy reticulin fibrosis (F1 = 93%) with high accuracy. However, other values, such as myeloblasts from the biopsy (F1 = 6%) and via flow cytometry (F1 = 8%), were affected by sparsity representative of reporting conventions. The four features with the highest clinical importance were extracted with F1 scores exceeding 90%. Whereas manual annotation of 300 reports required 30 hours of staff effort, automated NLP required 3.5 hours of runtime for 34,301 reports.To the best of our knowledge, this is among the first studies to demonstrate the application of NLP to hematopathology for clinical feature extraction. The approach may inform efforts at other institutions, and the code is available at https://github.com/wcmc-research-informatics/BmrExtractor.</p>","PeriodicalId":49822,"journal":{"name":"Methods of Information in Medicine","volume":" ","pages":"176-182"},"PeriodicalIF":1.3,"publicationDate":"2024-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144054744","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Harnessing Advanced Machine Learning Techniques for Microscopic Vessel Segmentation in Pulmonary Fibrosis Using Novel Hierarchical Phase-Contrast Tomography Images. 利用先进的机器学习技术在肺纤维化中使用新型分层相衬断层扫描(HiP-CT)图像进行显微血管分割。
IF 1.3 4区 医学 Q3 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2024-09-01 Epub Date: 2025-02-18 DOI: 10.1055/a-2540-8166
Pardeep Vasudev, Mehran Azimbagirad, Shahab Aslani, Moucheng Xu, Yufei Wang, Robert Chapman, Hannah Coleman, Christopher Werlein, Claire Walsh, Peter Lee, Paul Tafforeau, Joseph Jacob

Background:  Fibrotic lung disease is a progressive illness that causes scarring and ultimately respiratory failure, with irreversible damage by the time it is diagnosed on computed tomography imaging. Recent research postulates the role of the lung vasculature on the pathogenesis of the disease. With the recent development of high-resolution hierarchical phase-contrast tomography (HiP-CT), we have the potential to understand and detect changes in the lungs long before conventional imaging. However, to gain quantitative insight into vascular changes you first need to be able to segment the vessels before further downstream analysis can be conducted. Aside from this, HiP-CT generates large-volume, high-resolution data which is time-consuming and expensive to label.

Objectives:  This project aims to qualitatively assess the latest machine learning methods for vessel segmentation in HiP-CT data to enable label propagation as the first step for imaging biomarker discovery, with the goal to identify early-stage interstitial lung disease amenable to treatment, before fibrosis begins.

Methods:  Semisupervised learning (SSL) has become a growing method to tackle sparsely labeled datasets due to its leveraging of unlabeled data. In this study, we will compare two SSL methods; Seg PL, based on pseudo-labeling, and MisMatch, using consistency regularization against state-of-the-art supervised learning method, nnU-Net, on vessel segmentation in sparsely labeled lung HiP-CT data.

Results:  On initial experimentation, both MisMatch and SegPL showed promising performance on qualitative review. In comparison with supervised learning, both MisMatch and SegPL showed better out-of-distribution performance within the same sample (different vessel morphology and texture vessels), though supervised learning provided more consistent segmentations for well-represented labels in the limited annotations.

Conclusion:  Further quantitative research is required to better assess the generalizability of these findings, though they show promising first steps toward leveraging this novel data to tackle fibrotic lung disease.

背景:纤维化性肺病是一种进行性疾病,可导致瘢痕形成并最终导致呼吸衰竭,在计算机断层成像诊断时具有不可逆转的损害。最近的研究假设肺血管系统在疾病发病机制中的作用,并且随着高分辨率分层相衬断层扫描(HiP-CT)的最新发展,我们有可能在传统成像之前很久就了解和检测肺部的变化。然而,为了获得对血管变化的定量了解,首先需要能够在进行进一步的下游分析之前对血管进行分段。除此之外,HiP-CT产生的数据量大,分辨率高,耗时且标签成本高。目的:本项目旨在定性评估HiP-CT数据中用于血管分割的最新机器学习方法,使标签传播成为成像生物标志物发现的第一步,目标是在纤维化开始之前识别适合治疗的早期间质性肺疾病。方法:半监督学习已成为一种日益增长的方法来处理稀疏标记的数据集,由于其利用未标记的数据。在本研究中,我们将比较两种半监督学习方法;Seg PL,基于伪标记和错配,在nnU-Net中使用一致性正则化对抗最先进的监督学习方法,在稀疏标记的肺HiP-CT数据中进行血管分割。结果:在初始实验中,MisMatch和SegPL均表现出良好的定性评价性能。与监督学习相比,在同一样本(不同的血管形态和纹理血管)中,MisMatch和SegPL都表现出更好的非分布性能,尽管监督学习为有限注释中表现良好的标签提供了更一致的分割。结论:需要进一步的定量研究来更好地评估这些发现的普遍性,尽管它们在利用这些新数据治疗纤维化肺疾病方面迈出了有希望的第一步。
{"title":"Harnessing Advanced Machine Learning Techniques for Microscopic Vessel Segmentation in Pulmonary Fibrosis Using Novel Hierarchical Phase-Contrast Tomography Images.","authors":"Pardeep Vasudev, Mehran Azimbagirad, Shahab Aslani, Moucheng Xu, Yufei Wang, Robert Chapman, Hannah Coleman, Christopher Werlein, Claire Walsh, Peter Lee, Paul Tafforeau, Joseph Jacob","doi":"10.1055/a-2540-8166","DOIUrl":"10.1055/a-2540-8166","url":null,"abstract":"<p><strong>Background: </strong> Fibrotic lung disease is a progressive illness that causes scarring and ultimately respiratory failure, with irreversible damage by the time it is diagnosed on computed tomography imaging. Recent research postulates the role of the lung vasculature on the pathogenesis of the disease. With the recent development of high-resolution hierarchical phase-contrast tomography (HiP-CT), we have the potential to understand and detect changes in the lungs long before conventional imaging. However, to gain quantitative insight into vascular changes you first need to be able to segment the vessels before further downstream analysis can be conducted. Aside from this, HiP-CT generates large-volume, high-resolution data which is time-consuming and expensive to label.</p><p><strong>Objectives: </strong> This project aims to qualitatively assess the latest machine learning methods for vessel segmentation in HiP-CT data to enable label propagation as the first step for imaging biomarker discovery, with the goal to identify early-stage interstitial lung disease amenable to treatment, before fibrosis begins.</p><p><strong>Methods: </strong> Semisupervised learning (SSL) has become a growing method to tackle sparsely labeled datasets due to its leveraging of unlabeled data. In this study, we will compare two SSL methods; Seg PL, based on pseudo-labeling, and MisMatch, using consistency regularization against state-of-the-art supervised learning method, nnU-Net, on vessel segmentation in sparsely labeled lung HiP-CT data.</p><p><strong>Results: </strong> On initial experimentation, both MisMatch and SegPL showed promising performance on qualitative review. In comparison with supervised learning, both MisMatch and SegPL showed better out-of-distribution performance within the same sample (different vessel morphology and texture vessels), though supervised learning provided more consistent segmentations for well-represented labels in the limited annotations.</p><p><strong>Conclusion: </strong> Further quantitative research is required to better assess the generalizability of these findings, though they show promising first steps toward leveraging this novel data to tackle fibrotic lung disease.</p>","PeriodicalId":49822,"journal":{"name":"Methods of Information in Medicine","volume":" ","pages":"97-108"},"PeriodicalIF":1.3,"publicationDate":"2024-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12133326/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143450734","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Toward a National Information Model for Medication Orders in Sweden. 瑞典药品订单的国家信息模型。
IF 1.3 4区 医学 Q3 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2024-09-01 Epub Date: 2025-02-27 DOI: 10.1055/a-2546-4092
Sofie Holmeland, Tobias Blomberg, Andreas Mårtensson, Sabine Koch

Semantic interoperability among health information systems (HISs), in particular electronic health records (EHRs), is crucial for informed healthcare decisions and access to vital health data by the patient. However, inconsistent medication information and limited health data exchange contribute to medication errors worldwide. Although Sweden offers various solutions for health information exchange, there is a limitation in the exchange of medication orders and a lack of understanding the structure of medication orders among EHRs, highlighting the need for further exploration of the structure of medication orders.This study aims to develop a common information model of medication orders for EHRs to be used in the Swedish context.An explorative qualitative design study was conducted. Documents and reference models of how medication orders are structured were collected, and semi-structured interviews were conducted with five purposefully selected participants with insight into how medication orders are structured in EHRs in Sweden. Data were analyzed using information needs analysis, information structure analysis, and code systems, classifications, and terminology analysis.The following information areas were identified for a medication order: medication, medication indication, way of administration, medication order details, and dosage. These information areas were conceptualized into a Unified Modeling Language Class Diagram information model with defined classes, attributes, and data types. The resulting information model provides a representation of how medication orders are depicted in EHRs in Sweden and is aligned with existing national information models such as the National Medication List, while still providing additional information related to medication order details.The developed information model could potentially provide a national standardized model for medication orders, contributing to enhanced semantic interoperability and improving data exchange across various HISs. This could enhance data consistency, reducing the risk of medication errors and thereby improving patient safety.

背景:卫生信息系统(HISs)之间的语义互操作性,特别是电子健康记录(EHRs),对于知情的医疗保健决策和患者获取重要健康数据至关重要。然而,不一致的用药信息和有限的健康数据交换导致了世界范围内的用药错误。虽然瑞典为卫生信息交换提供了各种解决方案,但医嘱的交换存在局限性,并且对电子病历之间的医嘱结构缺乏了解,这突出表明需要进一步探索医嘱结构。目的:本研究的目的是开发一个共同的信息模型的药物订单的电子病历在瑞典的情况下使用。方法:进行探索性定性设计研究。收集了关于医嘱结构的文件和参考模型,并对五名有目的地选择的参与者进行了半结构化访谈,以深入了解瑞典电子病历中医嘱的结构。使用信息需求分析、信息结构分析、代码系统、分类和术语分析对数据进行分析。结果:在给药单中确定了以下信息区域:药物、给药指征、给药方式、给药单详细信息和剂量。这些信息区域被概念化到一个已开发的统一建模语言类图信息模型中,该模型具有已定义的类、属性和数据类型。由此产生的信息模型提供了瑞典电子病历中如何描述药物订单的表示,并与现有的国家信息模型(如国家药物清单)保持一致,同时仍然提供与药物订单详细信息相关的附加信息。结论:所建立的信息模型可为全国药品医嘱信息提供标准化模型,有助于提高医嘱信息的语义互操作性,改善医嘱信息在不同医疗机构间的数据交换。这可以增强数据一致性,减少用药错误的风险,从而提高患者安全。
{"title":"Toward a National Information Model for Medication Orders in Sweden.","authors":"Sofie Holmeland, Tobias Blomberg, Andreas Mårtensson, Sabine Koch","doi":"10.1055/a-2546-4092","DOIUrl":"10.1055/a-2546-4092","url":null,"abstract":"<p><p>Semantic interoperability among health information systems (HISs), in particular electronic health records (EHRs), is crucial for informed healthcare decisions and access to vital health data by the patient. However, inconsistent medication information and limited health data exchange contribute to medication errors worldwide. Although Sweden offers various solutions for health information exchange, there is a limitation in the exchange of medication orders and a lack of understanding the structure of medication orders among EHRs, highlighting the need for further exploration of the structure of medication orders.This study aims to develop a common information model of medication orders for EHRs to be used in the Swedish context.An explorative qualitative design study was conducted. Documents and reference models of how medication orders are structured were collected, and semi-structured interviews were conducted with five purposefully selected participants with insight into how medication orders are structured in EHRs in Sweden. Data were analyzed using information needs analysis, information structure analysis, and code systems, classifications, and terminology analysis.The following information areas were identified for a medication order: medication, medication indication, way of administration, medication order details, and dosage. These information areas were conceptualized into a Unified Modeling Language Class Diagram information model with defined classes, attributes, and data types. The resulting information model provides a representation of how medication orders are depicted in EHRs in Sweden and is aligned with existing national information models such as the National Medication List, while still providing additional information related to medication order details.The developed information model could potentially provide a national standardized model for medication orders, contributing to enhanced semantic interoperability and improving data exchange across various HISs. This could enhance data consistency, reducing the risk of medication errors and thereby improving patient safety.</p>","PeriodicalId":49822,"journal":{"name":"Methods of Information in Medicine","volume":" ","pages":"109-121"},"PeriodicalIF":1.3,"publicationDate":"2024-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12133328/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143525129","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Large-Scale Integration of DICOM Metadata into HL7-FHIR for Medical Research. 医学研究中DICOM元数据大规模集成HL7-FHIR。
IF 1.3 4区 医学 Q3 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2024-09-01 Epub Date: 2025-04-15 DOI: 10.1055/a-2521-4250
Alexa Iancu, Johannes Bauer, Matthias S May, Hans-Ulrich Prokosch, Arnd Dörfler, Michael Uder, Lorenz A Kapsner

Background:  The current gap between the availability of routine imaging data and its provisioning for medical research hinders the utilization of radiological information for secondary purposes. To address this, the German Medical Informatics Initiative (MII) has established frameworks for harmonizing and integrating clinical data across institutions, including the integration of imaging data into research repositories, which can be expanded to routine imaging data.

Objectives:  This project aims to address this gap by developing a large-scale data processing pipeline to extract, convert, and pseudonymize DICOM (Digital Imaging and Communications in Medicine) metadata into "ImagingStudy" Fast Healthcare Interoperability Resources (FHIR) and integrate them into research repositories for secondary use.

Methods:  The data processing pipeline was developed, implemented, and tested at the Data Integration Center of the University Hospital Erlangen. It leverages existing open-source solutions and integrates seamlessly into the hospital's research IT infrastructure. The pipeline automates the extraction, conversion, and pseudonymization processes, ensuring compliance with both local and MII data protection standards. A large-scale evaluation was conducted using the imaging studies acquired by two departments at University Hospital Erlangen within 1 year. Attributes such as modality, examined body region, laterality, and the number of series and instances were analyzed to assess the quality and availability of the metadata.

Results:  Once established, the pipeline processed a substantial dataset comprising over 150,000 DICOM studies within an operational period of 26 days. Data analysis revealed significant heterogeneity and incompleteness in certain attributes, particularly the DICOM tag "Body Part Examined." Despite these challenges, the pipeline successfully generated valid and standardized FHIR, providing a robust basis for future research.

Conclusion:  We demonstrated the setup and test of a large-scale end-to-end data processing pipeline that transforms DICOM imaging metadata directly from clinical routine into the Health Level 7-FHIR format, pseudonymizes the resources, and stores them in an FHIR server. We showcased that the derived FHIRs offer numerous research opportunities, for example, feasibility assessments within Bavarian and Germany-wide research infrastructures. Insights from this study highlight the need to extend the "ImagingStudy" FHIR with additional attributes and refine their use within the German MII.

背景:目前常规影像学数据的可用性与医学研究的提供之间存在差距,这阻碍了放射学信息用于次要目的的利用。为了解决这个问题,德国医学信息学倡议(MII)已经建立了协调和整合跨机构临床数据的框架,包括将成像数据集成到研究存储库中,这可以扩展到常规成像数据。目标:该项目旨在通过开发大规模数据处理管道来解决这一差距,该管道将DICOM(医学数字成像和通信)元数据提取、转换和假名化为“ImagingStudy”快速医疗互操作性资源(FHIR),并将其集成到研究存储库中供二次使用。方法:在埃尔兰根大学医院数据集成中心开发、实施和测试数据处理管道。它利用现有的开源解决方案,并无缝集成到医院的研究It基础设施中。该管道自动化了提取、转换和假名化过程,确保符合本地和MII数据保护标准。利用埃尔兰根大学医院两个科室在一年内获得的影像学研究进行了大规模评估。分析了模态、检查的主体区域、横向性以及系列和实例的数量等属性,以评估元数据的质量和可用性。结果:一旦建立,该管道在26天的运行期内处理了包含超过150,000个DICOM研究的大量数据集。数据分析揭示了某些属性的显著异质性和不完整性,特别是DICOM标签“身体部位检查”。尽管存在这些挑战,但该管道成功地产生了有效和标准化的FHIR,为未来的研究提供了坚实的基础。结论:我们演示了一个大规模端到端数据处理管道的设置和测试,该管道将DICOM成像元数据直接从临床常规转换为Health Level 7-FHIR格式,对资源进行假名化,并将其存储在FHIR服务器中。我们展示了衍生的fhir提供了许多研究机会,例如,在巴伐利亚和德国范围内的研究基础设施内进行可行性评估。本研究的见解强调了将“ImagingStudy”FHIR扩展为附加属性并改进其在德国MII中的使用的必要性。
{"title":"Large-Scale Integration of DICOM Metadata into HL7-FHIR for Medical Research.","authors":"Alexa Iancu, Johannes Bauer, Matthias S May, Hans-Ulrich Prokosch, Arnd Dörfler, Michael Uder, Lorenz A Kapsner","doi":"10.1055/a-2521-4250","DOIUrl":"10.1055/a-2521-4250","url":null,"abstract":"<p><strong>Background: </strong> The current gap between the availability of routine imaging data and its provisioning for medical research hinders the utilization of radiological information for secondary purposes. To address this, the German Medical Informatics Initiative (MII) has established frameworks for harmonizing and integrating clinical data across institutions, including the integration of imaging data into research repositories, which can be expanded to routine imaging data.</p><p><strong>Objectives: </strong> This project aims to address this gap by developing a large-scale data processing pipeline to extract, convert, and pseudonymize DICOM (Digital Imaging and Communications in Medicine) metadata into \"ImagingStudy\" Fast Healthcare Interoperability Resources (FHIR) and integrate them into research repositories for secondary use.</p><p><strong>Methods: </strong> The data processing pipeline was developed, implemented, and tested at the Data Integration Center of the University Hospital Erlangen. It leverages existing open-source solutions and integrates seamlessly into the hospital's research IT infrastructure. The pipeline automates the extraction, conversion, and pseudonymization processes, ensuring compliance with both local and MII data protection standards. A large-scale evaluation was conducted using the imaging studies acquired by two departments at University Hospital Erlangen within 1 year. Attributes such as modality, examined body region, laterality, and the number of series and instances were analyzed to assess the quality and availability of the metadata.</p><p><strong>Results: </strong> Once established, the pipeline processed a substantial dataset comprising over 150,000 DICOM studies within an operational period of 26 days. Data analysis revealed significant heterogeneity and incompleteness in certain attributes, particularly the DICOM tag \"Body Part Examined.\" Despite these challenges, the pipeline successfully generated valid and standardized FHIR, providing a robust basis for future research.</p><p><strong>Conclusion: </strong> We demonstrated the setup and test of a large-scale end-to-end data processing pipeline that transforms DICOM imaging metadata directly from clinical routine into the Health Level 7-FHIR format, pseudonymizes the resources, and stores them in an FHIR server. We showcased that the derived FHIRs offer numerous research opportunities, for example, feasibility assessments within Bavarian and Germany-wide research infrastructures. Insights from this study highlight the need to extend the \"ImagingStudy\" FHIR with additional attributes and refine their use within the German MII.</p>","PeriodicalId":49822,"journal":{"name":"Methods of Information in Medicine","volume":" ","pages":"77-84"},"PeriodicalIF":1.3,"publicationDate":"2024-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12133321/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143993767","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Alternative Strategies to Generate Class Activation Maps Supporting AI-based Advice in Vertebral Fracture Detection in X-ray Images. 生成类激活图的备选策略,支持基于人工智能的x射线图像椎体骨折检测建议。
IF 1.3 4区 医学 Q3 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2024-09-01 Epub Date: 2025-06-03 DOI: 10.1055/a-2562-2163
Samuele Pe, Lorenzo Famiglini, Enrico Gallazzi, Chandra Bortolotto, Luisa Carone, Andrea Cisarri, Alberto Salina, Lorenzo Preda, Riccardo Bellazzi, Federico Cabitza, Enea Parimbelli

Balancing artificial intelligence (AI) support with appropriate human oversight is challenging, with associated risks such as algorithm aversion and technology dominance. Research areas like eXplainable AI (XAI) and Frictional AI aim to address these challenges. Studies have shown that presenting XAI explanations as "juxtaposed evidence" supporting contrasting classifications, rather than just providing predictions, can be beneficial.This study aimed to design and compare multiple pipelines for generating juxtaposed evidence in the form of class activation maps (CAMs) that highlight areas of interest in a fracture detection task with X-ray images.We designed three pipelines to generate such evidence. The pipelines are based on a fracture detection task from 630 thoraco-lumbar X-ray images (48% of which contained fractures). The first, a single-model approach, uses an algorithm of the Grad-CAM family applied to a ResNeXt-50 network trained through transfer learning. The second, a dual-model approach, employs two networks-one optimized for sensitivity and the other for specificity-providing targeted explanations for positive and negative cases. The third, a generative approach, leverages autoencoders to create activation maps from feature tensors, extracted from the raw images. Each approach produced two versions of activation maps: AM3-as we termed it-which captures fine-grained, low-level features, and AM4, highlighting high-level, aggregated features. We conducted a validation study by comparing the generated maps with binary ground-truth masks derived from a consensus of four clinician annotators, identifying the actual locations of fractures in a subset of positive cases.HiResCAM proved to be the best performing Grad-CAM variant and was used in both the single- and dual-model strategies. The generative approach demonstrated the greatest overlap with the clinicians' assessments, indicating its ability to align with human expertise.The results highlight the potential of Judicial AI to enhance diagnostic decision-making and foster a synergistic collaboration between humans and AI.

平衡人工智能(AI)支持与适当的人类监督是具有挑战性的,伴随着算法厌恶和技术主导等相关风险。可解释AI (eXplainable AI, XAI)和Frictional AI等研究领域旨在解决这些挑战。研究表明,将XAI解释作为支持对比分类的“并列证据”,而不仅仅是提供预测,可能是有益的。本研究旨在设计和比较多个管道,以类激活图(CAMs)的形式生成并列的证据,这些证据突出了x射线图像裂缝检测任务中感兴趣的区域。我们设计了三个管道来生成这样的证据。管道是基于630张胸腰椎x射线图像的骨折检测任务(其中48%包含骨折)。第一种是单模型方法,将Grad-CAM系列算法应用于通过迁移学习训练的ResNeXt-50网络。第二种是双模型方法,采用两个网络——一个优化敏感性,另一个优化特异性——为积极和消极的案例提供有针对性的解释。第三种是生成方法,利用自编码器从原始图像中提取的特征张量创建激活图。每种方法都产生了两个版本的激活图:am3(我们称之为am3),它捕获细粒度的、低级别的特性;AM4,突出高级的、聚合的特性。我们进行了一项验证研究,将生成的图谱与四位临床医生注释者一致得出的二元基真掩模进行比较,确定了一部分阳性病例的实际骨折位置。HiResCAM被证明是表现最好的Grad-CAM变体,可用于单模型和双模型策略。生成式方法与临床医生的评估有最大的重叠,表明它有能力与人类的专业知识保持一致。研究结果强调了司法人工智能在增强诊断决策和促进人类与人工智能之间协同合作方面的潜力。
{"title":"Alternative Strategies to Generate Class Activation Maps Supporting AI-based Advice in Vertebral Fracture Detection in X-ray Images.","authors":"Samuele Pe, Lorenzo Famiglini, Enrico Gallazzi, Chandra Bortolotto, Luisa Carone, Andrea Cisarri, Alberto Salina, Lorenzo Preda, Riccardo Bellazzi, Federico Cabitza, Enea Parimbelli","doi":"10.1055/a-2562-2163","DOIUrl":"10.1055/a-2562-2163","url":null,"abstract":"<p><p>Balancing artificial intelligence (AI) support with appropriate human oversight is challenging, with associated risks such as <i>algorithm aversion</i> and <i>technology dominance</i>. Research areas like eXplainable AI (XAI) and Frictional AI aim to address these challenges. Studies have shown that presenting XAI explanations as \"juxtaposed evidence\" supporting contrasting classifications, rather than just providing predictions, can be beneficial.This study aimed to design and compare multiple pipelines for generating juxtaposed evidence in the form of class activation maps (CAMs) that highlight areas of interest in a fracture detection task with X-ray images.We designed three pipelines to generate such evidence. The pipelines are based on a fracture detection task from 630 thoraco-lumbar X-ray images (48% of which contained fractures). The first, a <i>single-model</i> approach, uses an algorithm of the Grad-CAM family applied to a ResNeXt-50 network trained through transfer learning. The second, a <i>dual-model</i> approach, employs two networks-one optimized for sensitivity and the other for specificity-providing targeted explanations for positive and negative cases. The third, a <i>generative</i> approach, leverages autoencoders to create activation maps from feature tensors, extracted from the raw images. Each approach produced two versions of activation maps: AM3-as we termed it-which captures fine-grained, low-level features, and AM4, highlighting high-level, aggregated features. We conducted a validation study by comparing the generated maps with binary ground-truth masks derived from a consensus of four clinician annotators, identifying the actual locations of fractures in a subset of positive cases.HiResCAM proved to be the best performing Grad-CAM variant and was used in both the single- and dual-model strategies. The generative approach demonstrated the greatest overlap with the clinicians' assessments, indicating its ability to align with human expertise.The results highlight the potential of Judicial AI to enhance diagnostic decision-making and foster a synergistic collaboration between humans and AI.</p>","PeriodicalId":49822,"journal":{"name":"Methods of Information in Medicine","volume":"63 3-04","pages":"122-136"},"PeriodicalIF":1.3,"publicationDate":"2024-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12133327/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144217397","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Unlocking the Potential of the Next Generation: Methods of Information in Medicine Student Paper Award 2024. 解锁下一代的潜力:信息方法在医学学生论文奖2024。
IF 1.3 4区 医学 Q3 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2024-09-01 Epub Date: 2025-06-03 DOI: 10.1055/a-2588-7805
Sabine Koch, John H Holmes, Lucia Sacchi
{"title":"Unlocking the Potential of the Next Generation: Methods of Information in Medicine Student Paper Award 2024.","authors":"Sabine Koch, John H Holmes, Lucia Sacchi","doi":"10.1055/a-2588-7805","DOIUrl":"https://doi.org/10.1055/a-2588-7805","url":null,"abstract":"","PeriodicalId":49822,"journal":{"name":"Methods of Information in Medicine","volume":"63 3-04","pages":"63-65"},"PeriodicalIF":1.3,"publicationDate":"2024-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144217398","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Leveraging Guideline-Based Clinical Decision Support Systems with Large Language Models: A Case Study with Breast Cancer. 利用基于指南的临床决策支持系统与大语言模型:乳腺癌的案例研究。
IF 1.3 4区 医学 Q3 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2024-09-01 Epub Date: 2025-01-29 DOI: 10.1055/a-2528-4299
Solène Delourme, Akram Redjdal, Jacques Bouaud, Brigitte Seroussi

Background:  Multidisciplinary tumor boards (MTBs) have been established in most countries to allow experts collaboratively determine the best treatment decisions for cancer patients. However, MTBs often face challenges such as case overload, which can compromise MTB decision quality. Clinical decision support systems (CDSSs) have been introduced to assist clinicians in this process. Despite their potential, CDSSs are still underutilized in routine practice. The emergence of large language models (LLMs), such as ChatGPT, offers new opportunities to improve the efficiency and usability of traditional CDSSs.

Objectives:  OncoDoc2 is a guideline-based CDSS developed using a documentary approach and applied to breast cancer management. This study aims to evaluate the potential of LLMs, used as question-answering (QA) systems, to improve the usability of OncoDoc2 across different prompt engineering techniques (PETs).

Methods:  Data extracted from breast cancer patient summaries (BCPSs), together with questions formulated by OncoDoc2, were used to create prompts for various LLMs, and several PETs were designed and tested. Using a sample of 200 randomized BCPSs, LLMs and PETs were initially compared with regard to their responses to OncoDoc2 questions using classic metrics (accuracy, precision, recall, and F1 score). Best performing LLMs and PETs were further assessed by comparing the therapeutic recommendations generated by OncoDoc2, based on LLM inputs, to those provided by MTB clinicians using OncoDoc2. Finally, the best performing method was validated using a new sample of 30 randomized BCPSs.

Results:  The combination of Mistral and OpenChat models under the enhanced Zero-Shot PET showed the best performance as a question-answering system. This approach gets a precision of 60.16%, a recall of 54.18%, an F1 score of 56.59%, and an accuracy of 75.57% on the validation set of 30 BCPSs. However, this approach yielded poor results as a CDSS, with only 16.67% of the recommendations generated by OncoDoc2 based on LLM inputs matching the gold standard.

Conclusion:  All the criteria in the OncoDoc2 decision tree are crucial for capturing the uniqueness of each patient. Any deviation from a criterion alters the recommendations generated. Despite achieving a good accuracy rate of 75.57%, LLMs still face challenges in reliably understanding complex medical contexts and be effective as CDSSs.

背景:大多数国家都成立了多学科肿瘤委员会(MTBs),以便专家们共同为癌症患者做出最佳治疗决定。然而,多学科肿瘤委员会经常面临病例超载等挑战,这可能会影响多学科肿瘤委员会的决策质量。临床决策支持系统(CDSS)的出现就是为了在这一过程中为临床医生提供帮助。尽管 CDSS 具有潜力,但在常规实践中仍未得到充分利用。大型语言模型(LLM)(如 ChatGPT)的出现为提高传统临床决策支持系统(CDSS)的效率和可用性提供了新的机遇:目的:OncoDoc2 是一种基于指南的 CDSS,采用文档方法开发,适用于乳腺癌管理。本研究旨在评估作为问题解答(QA)系统使用的 LLMs 在不同提示工程技术(PET)下提高 OncoDoc2 可用性的潜力:方法:从乳腺癌患者摘要(BCPS)中提取的数据与 OncoDoc2 提出的问题一起用于创建各种 LLM 的提示,并设计和测试了几种 PET。利用 200 份随机 BCPS 样本,使用经典指标(准确度、精确度、召回率和 F1 分数)对 LLM 和 PET 对 OncoDoc2 问题的回答进行了初步比较。通过比较 OncoDoc2 根据 LLM 输入生成的治疗建议和 MTB 临床医生使用 OncoDoc2 提供的治疗建议,进一步评估了表现最佳的 LLM 和 PET。最后,使用新的 30 个随机 BCPS 样本验证了性能最佳的方法:结果:Mistral 和 OpenChat 模型在增强的零点 PET 下的组合作为问题解答系统表现最佳。在 30 个 BCPS 的验证集上,该方法的精确度为 60.16%,召回率为 54.18%,F1 分数为 56.59%,准确率为 75.57%。然而,作为 CDSS,这种方法的结果并不理想,OncoDoc2 基于 LLM 输入生成的建议中只有 16.67% 与黄金标准相匹配:OncoDoc2决策树中的所有标准对于捕捉每位患者的独特性都至关重要。与标准的任何偏差都会改变生成的建议。尽管实现了 75.57% 的良好准确率,但 LLM 在可靠地理解复杂的医疗环境并有效地用作 CDSS 方面仍面临挑战。
{"title":"Leveraging Guideline-Based Clinical Decision Support Systems with Large Language Models: A Case Study with Breast Cancer.","authors":"Solène Delourme, Akram Redjdal, Jacques Bouaud, Brigitte Seroussi","doi":"10.1055/a-2528-4299","DOIUrl":"10.1055/a-2528-4299","url":null,"abstract":"<p><strong>Background: </strong> Multidisciplinary tumor boards (MTBs) have been established in most countries to allow experts collaboratively determine the best treatment decisions for cancer patients. However, MTBs often face challenges such as case overload, which can compromise MTB decision quality. Clinical decision support systems (CDSSs) have been introduced to assist clinicians in this process. Despite their potential, CDSSs are still underutilized in routine practice. The emergence of large language models (LLMs), such as ChatGPT, offers new opportunities to improve the efficiency and usability of traditional CDSSs.</p><p><strong>Objectives: </strong> OncoDoc2 is a guideline-based CDSS developed using a documentary approach and applied to breast cancer management. This study aims to evaluate the potential of LLMs, used as question-answering (QA) systems, to improve the usability of OncoDoc2 across different prompt engineering techniques (PETs).</p><p><strong>Methods: </strong> Data extracted from breast cancer patient summaries (BCPSs), together with questions formulated by OncoDoc2, were used to create prompts for various LLMs, and several PETs were designed and tested. Using a sample of 200 randomized BCPSs, LLMs and PETs were initially compared with regard to their responses to OncoDoc2 questions using classic metrics (accuracy, precision, recall, and F1 score). Best performing LLMs and PETs were further assessed by comparing the therapeutic recommendations generated by OncoDoc2, based on LLM inputs, to those provided by MTB clinicians using OncoDoc2. Finally, the best performing method was validated using a new sample of 30 randomized BCPSs.</p><p><strong>Results: </strong> The combination of Mistral and OpenChat models under the enhanced Zero-Shot PET showed the best performance as a question-answering system. This approach gets a precision of 60.16%, a recall of 54.18%, an F1 score of 56.59%, and an accuracy of 75.57% on the validation set of 30 BCPSs. However, this approach yielded poor results as a CDSS, with only 16.67% of the recommendations generated by OncoDoc2 based on LLM inputs matching the gold standard.</p><p><strong>Conclusion: </strong> All the criteria in the OncoDoc2 decision tree are crucial for capturing the uniqueness of each patient. Any deviation from a criterion alters the recommendations generated. Despite achieving a good accuracy rate of 75.57%, LLMs still face challenges in reliably understanding complex medical contexts and be effective as CDSSs.</p>","PeriodicalId":49822,"journal":{"name":"Methods of Information in Medicine","volume":" ","pages":"85-96"},"PeriodicalIF":1.3,"publicationDate":"2024-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12133322/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143069247","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
The Significance of Information Quality for the Secondary Use of the Information in the National Health Care Quality Registers in Finland. 信息质量对芬兰国家医疗保健质量登记信息二次利用的重要性。
IF 1.3 4区 医学 Q3 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2024-09-01 Epub Date: 2025-01-08 DOI: 10.1055/a-2511-7866
Anna Frondelius, Ulla-Mari Kinnunen, Vesa Jormanainen

Background:  The aim of the national health care quality registers is to monitor, assess, and improve the quality of care. The information utilized in quality registers must be of high quality to ensure that the information produced by the registers is reliable and useful. In Finland, one of the key sources of information for the quality registers is the national Kanta services.

Objectives:  The objective of the study was to increase understanding of the significance of information quality for the secondary use of the information in the national health care quality registers and to provide information on whether the information quality of the national Kanta services supports the information needs of the national quality registers, and how information quality should be developed.

Methods:  The research data were collected by interviewing six experts responsible for national health care quality registers, and it was analyzed using theory-driven qualitative content analysis based on the DeLone and McLean model.

Results:  Based on the results, the relevance of the information in the Kanta services met the information needs of the national quality registers. However, due to the limited amount of structured information and deficiencies in the completeness of the information, relevant information could not be fully utilized. Deficiencies in information quality posed challenges in information retrieval and hindered drawing conclusions in reporting. Challenges in information quality did not diminish the intention to use the information when information was considered relevant. Solutions to improve information quality included structuring, development of documentation practices, patient information systems and quality assurance, as well as collaboration among stakeholders.

Conclusion:  The Kanta services' information is relevant for the national health care quality registers, but developing the quality of the information, especially in terms of structures and completeness, is the key to fully enabling the secondary use of this information.

背景:国家卫生保健质量登记的目的是监测、评估和提高卫生保健质量。在质量寄存器中使用的信息必须是高质量的,以确保由寄存器产生的信息是可靠和有用的。在芬兰,质量登记信息的主要来源之一是国家坎塔服务。本研究的目的是提高对信息质量对国家卫生保健质量登记册中信息二次使用的重要性的认识,并提供关于国家Kanta服务的信息质量是否支持国家质量登记册的信息需求以及应如何开发信息质量的信息。方法通过对6名国家卫生保健质量登记机构专家的访谈收集研究数据,采用基于DeLone和McLean模型的理论驱动定性内容分析方法进行分析。结果调查结果表明,坎塔服务信息的相关性满足了国家质量登记机构的信息需求。然而,由于结构化信息的数量有限,信息的完整性不足,相关信息不能得到充分利用。信息质量的不足给信息检索带来了挑战,也阻碍了报告结论的得出。信息质量方面的挑战并没有削弱人们在认为信息相关时使用这些信息的意愿。改善信息质量的解决方案包括结构、文档实践的开发、患者信息系统和质量保证,以及利益相关者之间的合作。结论Kanta服务信息与国家卫生保健质量登记相关,但提高信息质量,特别是信息结构和完整性,是充分利用这些信息的关键。
{"title":"The Significance of Information Quality for the Secondary Use of the Information in the National Health Care Quality Registers in Finland.","authors":"Anna Frondelius, Ulla-Mari Kinnunen, Vesa Jormanainen","doi":"10.1055/a-2511-7866","DOIUrl":"10.1055/a-2511-7866","url":null,"abstract":"<p><strong>Background: </strong> The aim of the national health care quality registers is to monitor, assess, and improve the quality of care. The information utilized in quality registers must be of high quality to ensure that the information produced by the registers is reliable and useful. In Finland, one of the key sources of information for the quality registers is the national Kanta services.</p><p><strong>Objectives: </strong> The objective of the study was to increase understanding of the significance of information quality for the secondary use of the information in the national health care quality registers and to provide information on whether the information quality of the national Kanta services supports the information needs of the national quality registers, and how information quality should be developed.</p><p><strong>Methods: </strong> The research data were collected by interviewing six experts responsible for national health care quality registers, and it was analyzed using theory-driven qualitative content analysis based on the DeLone and McLean model.</p><p><strong>Results: </strong> Based on the results, the relevance of the information in the Kanta services met the information needs of the national quality registers. However, due to the limited amount of structured information and deficiencies in the completeness of the information, relevant information could not be fully utilized. Deficiencies in information quality posed challenges in information retrieval and hindered drawing conclusions in reporting. Challenges in information quality did not diminish the intention to use the information when information was considered relevant. Solutions to improve information quality included structuring, development of documentation practices, patient information systems and quality assurance, as well as collaboration among stakeholders.</p><p><strong>Conclusion: </strong> The Kanta services' information is relevant for the national health care quality registers, but developing the quality of the information, especially in terms of structures and completeness, is the key to fully enabling the secondary use of this information.</p>","PeriodicalId":49822,"journal":{"name":"Methods of Information in Medicine","volume":" ","pages":"66-76"},"PeriodicalIF":1.3,"publicationDate":"2024-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12133325/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142957856","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
The Completeness of the Operating Room Data. 手术室数据的完整性。
IF 1.3 4区 医学 Q3 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2024-09-01 Epub Date: 2025-03-26 DOI: 10.1055/a-2566-7958
Päivi Nurmela, Minna Mykkänen, Ulla-Mari Kinnunen

In the operating theater, a large collection of data are collected at each surgical visit. Some of these data are patient information, some is related to resource management, which is linked to hospital finances. Poor quality data lead to poor decisions, impacting patient safety and the continuity of care.The study aimed at evaluating the completeness of the data documented within surgical operations. Based on the results, the goal is to improve data quality and to identify improvement ideas of data management.The study was a quantitative evaluation of 33,684 surgical visits, focusing on data omissions. The organization identified 58 operating room data variables related to visits, procedures, resources, and personnel. Data completeness was evaluated for 36 variables, excluding 47 visits that were missing the "Complete" flag. Data preprocessing was done using Python and Pandas, with pseudonymization of personnel names. Data were analyzed using the R programming language. Data omissions were coded as "1" for missing values and "0" for others. Summary variables were created to indicate the number of personnel and procedure and data omissions per visit.The average completeness of the operating room data was 98%, which is considered excellent. However, seven variables-the start and end date and time of anesthesia, the type of treatment, personnel group, and assistant information-had completeness below the 95% target level. A total of 34% of the surgical visits contained at least one data omission. In the yearly comparison, the completeness values of variables were statistically significantly higher in 2022 compared with 2023.By ensuring existing quality assurance practices, verifying internal data maintenance and verifying and standardizing documenting practices the organization can achieve net benefits through improved data completeness, thus enhancing patient records, financial information, and management. Improved data quality will also benefit national and international registers.

在手术室,每次就诊都会收集大量的数据。其中一些数据是患者信息,一些是与资源管理相关的,这与医院财务有关。低质量的数据导致糟糕的决策,影响患者安全和护理的连续性。本研究旨在评估外科手术中记录的数据的完整性,并根据结果提高数据质量,确定数据管理改进思路。方法对33,684例外科就诊进行定量评价,重点分析数据遗漏。该组织确定了58个手术室数据变量,涉及访问、程序、资源和人员。对36个变量的数据完整性进行了评估,排除了47个缺少“完整”标志的访问。数据预处理使用Python和Pandas完成,并对人员姓名进行了假名化处理。数据分析使用R编程语言完成。数据遗漏用“1”表示缺失值,用“0”表示其他值。创建了摘要变量来表示人员和程序的数量,以及每次访问的数据遗漏。结果手术室资料的平均完整性为98%,为优等。然而,麻醉开始和结束日期和时间、治疗类型、人员分组和辅助信息等7个变量的完整性低于95%的目标水平。34%的外科就诊至少有一个数据遗漏。在年度比较中,变量的完备性值在2022年明显高于2023年。结论通过确保现有的质量保证实践,验证内部数据维护,验证和标准化文档实践,组织可以通过提高数据完整性,增强患者记录,财务信息和管理来实现净效益。数据质量的提高也将使国家和国际登记处受益。关键词:数据,患者数据,质量,卫生信息系统,手术室。
{"title":"The Completeness of the Operating Room Data.","authors":"Päivi Nurmela, Minna Mykkänen, Ulla-Mari Kinnunen","doi":"10.1055/a-2566-7958","DOIUrl":"10.1055/a-2566-7958","url":null,"abstract":"<p><p>In the operating theater, a large collection of data are collected at each surgical visit. Some of these data are patient information, some is related to resource management, which is linked to hospital finances. Poor quality data lead to poor decisions, impacting patient safety and the continuity of care.The study aimed at evaluating the completeness of the data documented within surgical operations. Based on the results, the goal is to improve data quality and to identify improvement ideas of data management.The study was a quantitative evaluation of 33,684 surgical visits, focusing on data omissions. The organization identified 58 operating room data variables related to visits, procedures, resources, and personnel. Data completeness was evaluated for 36 variables, excluding 47 visits that were missing the \"Complete\" flag. Data preprocessing was done using Python and Pandas, with pseudonymization of personnel names. Data were analyzed using the R programming language. Data omissions were coded as \"1\" for missing values and \"0\" for others. Summary variables were created to indicate the number of personnel and procedure and data omissions per visit.The average completeness of the operating room data was 98%, which is considered excellent. However, seven variables-the start and end date and time of anesthesia, the type of treatment, personnel group, and assistant information-had completeness below the 95% target level. A total of 34% of the surgical visits contained at least one data omission. In the yearly comparison, the completeness values of variables were statistically significantly higher in 2022 compared with 2023.By ensuring existing quality assurance practices, verifying internal data maintenance and verifying and standardizing documenting practices the organization can achieve net benefits through improved data completeness, thus enhancing patient records, financial information, and management. Improved data quality will also benefit national and international registers.</p>","PeriodicalId":49822,"journal":{"name":"Methods of Information in Medicine","volume":" ","pages":"137-144"},"PeriodicalIF":1.3,"publicationDate":"2024-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12133324/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143732739","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Deep Learning for Predicting Progression of Patellofemoral Osteoarthritis Based on Lateral Knee Radiographs, Demographic Data, and Symptomatic Assessments. 基于膝关节外侧X光片、人口统计学数据和症状评估的深度学习预测髌骨骨关节炎的进展情况
IF 1.3 4区 医学 Q3 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2024-05-01 Epub Date: 2024-04-11 DOI: 10.1055/a-2305-2115
Neslihan Bayramoglu, Martin Englund, Ida K Haugen, Muneaki Ishijima, Simo Saarakkala

Objective: In this study, we propose a novel framework that utilizes deep learning and attention mechanisms to predict the radiographic progression of patellofemoral osteoarthritis (PFOA) over a period of 7 years.

Material and methods: This study included subjects (1,832 subjects, 3,276 knees) from the baseline of the Multicenter Osteoarthritis Study (MOST). Patellofemoral joint regions of interest were identified using an automated landmark detection tool (BoneFinder) on lateral knee X-rays. An end-to-end deep learning method was developed for predicting PFOA progression based on imaging data in a five-fold cross-validation setting. To evaluate the performance of the models, a set of baselines based on known risk factors were developed and analyzed using gradient boosting machine (GBM). Risk factors included age, sex, body mass index, and Western Ontario and McMaster Universities Arthritis Index score, and the radiographic osteoarthritis stage of the tibiofemoral joint (Kellgren and Lawrence [KL] score). Finally, to increase predictive power, we trained an ensemble model using both imaging and clinical data.

Results: Among the individual models, the performance of our deep convolutional neural network attention model achieved the best performance with an area under the receiver operating characteristic curve (AUC) of 0.856 and average precision (AP) of 0.431, slightly outperforming the deep learning approach without attention (AUC = 0.832, AP = 0.4) and the best performing reference GBM model (AUC = 0.767, AP = 0.334). The inclusion of imaging data and clinical variables in an ensemble model allowed statistically more powerful prediction of PFOA progression (AUC = 0.865, AP = 0.447), although the clinical significance of this minor performance gain remains unknown. The spatial attention module improved the predictive performance of the backbone model, and the visual interpretation of attention maps focused on the joint space and the regions where osteophytes typically occur.

Conclusion: This study demonstrated the potential of machine learning models to predict the progression of PFOA using imaging and clinical variables. These models could be used to identify patients who are at high risk of progression and prioritize them for new treatments. However, even though the accuracy of the models were excellent in this study using the MOST dataset, they should be still validated using external patient cohorts in the future.

目的:在本研究中,我们提出了一种新的框架,利用深度学习和注意力机制来预测髌骨骨关节炎(PFOA)在七年内的放射学进展:在这项研究中,我们提出了一个新颖的框架,利用深度学习和注意力机制来预测七年内髌股骨关节炎(PFOA)的放射学进展:本研究纳入了多中心骨关节炎研究(MOST)基线的受试者(1832名受试者,3276个膝关节)。使用膝关节侧位 X 光片上的自动地标检测工具(BoneFinder)确定髌股关节感兴趣区。开发了一种端到端的深度学习方法,用于在 5 倍交叉验证设置中根据成像数据预测 PFOA 的进展。为了评估模型的性能,使用梯度提升机(GBM)开发并分析了一组基于已知风险因素的基线。风险因素包括年龄、性别、体重指数(BMI)和 WOMAC 评分,以及胫股关节的放射骨关节炎分期(KL 评分)。最后,为了提高预测能力,我们利用影像学和临床数据训练了一个集合模型:在单个模型中,我们的深度卷积神经网络注意力模型性能最佳,AUC 为 0.856,AP 为 0.431;略优于无注意力的深度学习方法(AUC=0.832,AP=0.4)和性能最佳的参考 GBM 模型(AUC=0.767,AP=0.334)。在一个集合模型中加入成像数据和临床变量后,对 PFOA 进展的预测在统计学上更为有力(AUC=0.865,AP=0.447),但这一微小的性能提升的临床意义仍不得而知。空间注意力模块提高了骨干模型的预测性能,注意力图的可视化解读侧重于关节空间和骨质增生的典型发生区域:本研究证明了机器学习模型利用成像和临床变量预测 PFOA 进展的潜力。这些模型可用于识别病情恶化风险较高的患者,并优先选择新的治疗方法。不过,尽管本研究中使用 MOST 数据集的模型准确性很高,但今后仍应使用外部患者队列对其进行验证。
{"title":"Deep Learning for Predicting Progression of Patellofemoral Osteoarthritis Based on Lateral Knee Radiographs, Demographic Data, and Symptomatic Assessments.","authors":"Neslihan Bayramoglu, Martin Englund, Ida K Haugen, Muneaki Ishijima, Simo Saarakkala","doi":"10.1055/a-2305-2115","DOIUrl":"10.1055/a-2305-2115","url":null,"abstract":"<p><strong>Objective: </strong>In this study, we propose a novel framework that utilizes deep learning and attention mechanisms to predict the radiographic progression of patellofemoral osteoarthritis (PFOA) over a period of 7 years.</p><p><strong>Material and methods: </strong>This study included subjects (1,832 subjects, 3,276 knees) from the baseline of the Multicenter Osteoarthritis Study (MOST). Patellofemoral joint regions of interest were identified using an automated landmark detection tool (BoneFinder) on lateral knee X-rays. An end-to-end deep learning method was developed for predicting PFOA progression based on imaging data in a five-fold cross-validation setting. To evaluate the performance of the models, a set of baselines based on known risk factors were developed and analyzed using gradient boosting machine (GBM). Risk factors included age, sex, body mass index, and Western Ontario and McMaster Universities Arthritis Index score, and the radiographic osteoarthritis stage of the tibiofemoral joint (Kellgren and Lawrence [KL] score). Finally, to increase predictive power, we trained an ensemble model using both imaging and clinical data.</p><p><strong>Results: </strong>Among the individual models, the performance of our deep convolutional neural network attention model achieved the best performance with an area under the receiver operating characteristic curve (AUC) of 0.856 and average precision (AP) of 0.431, slightly outperforming the deep learning approach without attention (AUC = 0.832, AP = 0.4) and the best performing reference GBM model (AUC = 0.767, AP = 0.334). The inclusion of imaging data and clinical variables in an ensemble model allowed statistically more powerful prediction of PFOA progression (AUC = 0.865, AP = 0.447), although the clinical significance of this minor performance gain remains unknown. The spatial attention module improved the predictive performance of the backbone model, and the visual interpretation of attention maps focused on the joint space and the regions where osteophytes typically occur.</p><p><strong>Conclusion: </strong>This study demonstrated the potential of machine learning models to predict the progression of PFOA using imaging and clinical variables. These models could be used to identify patients who are at high risk of progression and prioritize them for new treatments. However, even though the accuracy of the models were excellent in this study using the MOST dataset, they should be still validated using external patient cohorts in the future.</p>","PeriodicalId":49822,"journal":{"name":"Methods of Information in Medicine","volume":" ","pages":"1-10"},"PeriodicalIF":1.3,"publicationDate":"2024-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11495941/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140854286","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Methods of Information in Medicine
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1