首页 > 最新文献

Journal of endourology最新文献

英文 中文
The Rise of AI in Endourology and Robotic Surgery. 人工智能在腔内泌尿学和机器人手术中的崛起。
IF 2.9 2区 医学 Q1 UROLOGY & NEPHROLOGY Pub Date : 2024-08-01 DOI: 10.1089/end.2024.32789.pd
Prokar Dasgupta, Ashok Hemal, Glenn Preminger, Roger Sur
{"title":"The Rise of AI in Endourology and Robotic Surgery.","authors":"Prokar Dasgupta, Ashok Hemal, Glenn Preminger, Roger Sur","doi":"10.1089/end.2024.32789.pd","DOIUrl":"https://doi.org/10.1089/end.2024.32789.pd","url":null,"abstract":"","PeriodicalId":15723,"journal":{"name":"Journal of endourology","volume":null,"pages":null},"PeriodicalIF":2.9,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141909807","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Artificial Intelligence for Urology Research: The Holy Grail of Data Science or Pandora's Box of Misinformation? 人工智能用于泌尿学研究:数据科学的圣杯还是错误信息的潘多拉魔盒?
IF 2.9 2区 医学 Q1 UROLOGY & NEPHROLOGY Pub Date : 2024-08-01 Epub Date: 2024-05-15 DOI: 10.1089/end.2023.0703
Ryan M Blake, Johnathan A Khusid

Introduction: Artificial intelligence tools such as the large language models (LLMs) Bard and ChatGPT have generated significant research interest. Utilization of these LLMs to study the epidemiology of a target population could benefit urologists. We investigated whether Bard and ChatGPT can perform a large-scale calculation of the incidence and prevalence of kidney stone disease. Materials and Methods: We obtained reference values from two published studies, which used the National Health and Nutrition Examination Survey (NHANES) database to calculate the prevalence and incidence of kidney stone disease. We then tested the capability of Bard and ChatGPT to perform similar calculations using two different methods. First, we instructed the LLMs to access the data sets and independently perform the calculation. Second, we instructed the interfaces to generate a customized computer code, which could perform the calculation on downloaded data sets. Results: While ChatGPT denied the ability to access and perform calculations on the NHANES database, Bard intermittently claimed the ability to do so. Bard provided either accurate results or inaccurate and inconsistent results. For example, Bard's "calculations" for the incidence of kidney stones from 2015 to 2018 were 2.1% (95% CI 1.5-2.7), 1.75% (95% CI 1.6-1.9), and 0.8% (95% CI 0.7-0.9), while the published number was 2.1% (95% CI 1.5-2.7). Bard provided discrete mathematical details of its calculations, however, when prompted further, admitted to having obtained the numbers from online sources, including our chosen reference articles, rather than from a de novo calculation. Both LLMs were able to produce a code (Python) to use on the downloaded NHANES data sets, however, these would not readily execute. Conclusions: ChatGPT and Bard are currently incapable of performing epidemiologic calculations and lack transparency and accountability. Caution should be used, particularly with Bard, as claims of its capabilities were convincingly misleading, and results were inconsistent.

引言 大型语言模型(LLMs)Bard 和 ChatGPT 等人工智能工具引起了人们极大的研究兴趣。利用这些大型语言模型研究目标人群的流行病学可使泌尿科医生受益匪浅。我们研究了 Bard 和 ChatGPT 能否大规模计算肾结石病的发病率和流行率。材料与方法 我们从两项已发表的研究中获得了参考值,这两项研究使用了美国国家健康与营养调查(NHANES)数据库来计算肾结石病的患病率和发病率。然后,我们用两种不同的方法测试了 Bard 和 ChatGPT 进行类似计算的能力。首先,我们指示 LLMs 访问数据集并独立进行计算。其次,我们指示界面生成定制的计算机代码,以便在下载的数据集上执行计算。结果 虽然 ChatGPT 否认自己有能力访问 NHANES 数据库并进行计算,但 Bard 却断断续续地声称自己有能力这样做。Bard 要么提供了准确的结果,要么提供了不准确和不一致的结果。例如,巴德公司对 2015-2018 年肾结石发病率的 "计算结果 "分别为 2.1%(95% CI:1.5-2.7)、1.75%(95% CI:1.6-1.9)和 0.8%(95% CI 0.7-0.9),而公布的数字为 2.1%(95% CI 1.5-2.7)。Bard 提供了其计算的离散数学细节,但在进一步询问时,Bard 承认这些数字是从网上来源获得的,包括我们选择的参考文献,而不是重新计算的结果。两位 LLM 都能在下载的 NHANES 数据集上生成代码(Python),但这些代码并不容易执行。结论 ChatGPT 和 Bard 目前无法进行流行病学计算,缺乏透明度和问责制。应谨慎使用 ChatGPT 和 Bard,尤其是 Bard,因为对其功能的宣称具有令人信服的误导性,而且结果也不一致。
{"title":"Artificial Intelligence for Urology Research: The Holy Grail of Data Science or Pandora's Box of Misinformation?","authors":"Ryan M Blake, Johnathan A Khusid","doi":"10.1089/end.2023.0703","DOIUrl":"10.1089/end.2023.0703","url":null,"abstract":"<p><p><b><i>Introduction:</i></b> Artificial intelligence tools such as the large language models (LLMs) Bard and ChatGPT have generated significant research interest. Utilization of these LLMs to study the epidemiology of a target population could benefit urologists. We investigated whether Bard and ChatGPT can perform a large-scale calculation of the incidence and prevalence of kidney stone disease. <b><i>Materials and Methods:</i></b> We obtained reference values from two published studies, which used the National Health and Nutrition Examination Survey (NHANES) database to calculate the prevalence and incidence of kidney stone disease. We then tested the capability of Bard and ChatGPT to perform similar calculations using two different methods. First, we instructed the LLMs to access the data sets and independently perform the calculation. Second, we instructed the interfaces to generate a customized computer code, which could perform the calculation on downloaded data sets. <b><i>Results:</i></b> While ChatGPT denied the ability to access and perform calculations on the NHANES database, Bard intermittently claimed the ability to do so. Bard provided either accurate results or inaccurate and inconsistent results. For example, Bard's \"calculations\" for the incidence of kidney stones from 2015 to 2018 were 2.1% (95% CI 1.5-2.7), 1.75% (95% CI 1.6-1.9), and 0.8% (95% CI 0.7-0.9), while the published number was 2.1% (95% CI 1.5-2.7). Bard provided discrete mathematical details of its calculations, however, when prompted further, admitted to having obtained the numbers from online sources, including our chosen reference articles, rather than from a <i>de novo</i> calculation. Both LLMs were able to produce a code (Python) to use on the downloaded NHANES data sets, however, these would not readily execute. <b><i>Conclusions:</i></b> ChatGPT and Bard are currently incapable of performing epidemiologic calculations and lack transparency and accountability. Caution should be used, particularly with Bard, as claims of its capabilities were convincingly misleading, and results were inconsistent.</p>","PeriodicalId":15723,"journal":{"name":"Journal of endourology","volume":null,"pages":null},"PeriodicalIF":2.9,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140305822","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Development of UroSAM: A Machine Learning Model to Automatically Identify Kidney Stone Composition from Endoscopic Video. 开发 UroSAM:从内窥镜视频中自动识别肾结石成分的机器学习模型。
IF 2.9 2区 医学 Q1 UROLOGY & NEPHROLOGY Pub Date : 2024-08-01 Epub Date: 2024-05-31 DOI: 10.1089/end.2023.0740
Jixuan Leng, Junfei Liu, Galen Cheng, Haohan Wang, Scott Quarrier, Jiebo Luo, Rajat Jain

Introduction: Chemical composition analysis is important in prevention counseling for kidney stone disease. Advances in laser technology have made dusting techniques more prevalent, but this offers no consistent way to collect enough material to send for chemical analysis, leading many to forgo this test. We developed a novel machine learning (ML) model to effectively assess stone composition based on intraoperative endoscopic video data. Methods: Two endourologists performed ureteroscopy for kidney stones ≥ 10 mm. Representative videos were recorded intraoperatively. Individual frames were extracted from the videos, and the stone was outlined by human tracing. An ML model, UroSAM, was built and trained to automatically identify kidney stones in the images and predict the majority stone composition as follows: calcium oxalate monohydrate (COM), dihydrate (COD), calcium phosphate (CAP), or uric acid (UA). UroSAM was built on top of the publicly available Segment Anything Model (SAM) and incorporated a U-Net convolutional neural network (CNN). Discussion: A total of 78 ureteroscopy videos were collected; 50 were used for the model after exclusions (32 COM, 8 COD, 8 CAP, 2 UA). The ML model segmented the images with 94.77% precision. Dice coefficient (0.9135) and Intersection over Union (0.8496) confirmed good segmentation performance of the ML model. A video-wise evaluation demonstrated 60% correct classification of stone composition. Subgroup analysis showed correct classification in 84.4% of COM videos. A post hoc adaptive threshold technique was used to mitigate biasing of the model toward COM because of data imbalance; this improved the overall correct classification to 62% while improving the classification of COD, CAP, and UA videos. Conclusions: This study demonstrates the effective development of UroSAM, an ML model that precisely identifies kidney stones from natural endoscopic video data. More high-quality video data will improve the performance of the model in classifying the majority stone composition.

导言 化学成分分析对于肾结石疾病的预防咨询非常重要。激光技术的进步使除尘技术更加普及,但这种方法无法始终如一地收集足够的材料送去进行化学分析,导致许多人放弃了这项检查。我们开发了一种新型机器学习(ML)模型,可根据术中内窥镜视频数据有效评估结石成分。方法 两名内镜医师对≥ 10 毫米的肾结石进行输尿管镜检查。术中录制了具有代表性的视频。从视频中提取单帧图像,并通过人工追踪勾勒出结石的轮廓。建立并训练了一个 ML 模型 UroSAM,用于自动识别图像中的肾结石,并预测结石的主要成分:一水草酸钙 (COM)、二水草酸钙 (COD)、磷酸钙 (CAP) 或尿酸 (UA)。UroSAM 建立在公开可用的分段任意模型 (SAM) 基础上,并结合了 U-Net 卷积神经网络 (CNN)。讨论 共收集了 78 个输尿管镜检查视频,在排除了 32 个 COM、8 个 COD、8 个 CAP 和 2 个 UA 之后,有 50 个视频被用于该模型。ML 模型分割图像的精确度为 94.77%。Dice coefficient(0.9135)和 Intersection over Union(0.8496)证实了 ML 模型良好的分割性能。视频评估显示,对结石成分的正确分类率为 60%。分组分析表明,84.4% 的 COM 视频分类正确。采用了一种事后自适应阈值技术,以减轻由于数据不平衡而导致的模型对 COM 的偏差--这将整体正确分类率提高到了 62%,同时改善了 COD、CAP 和 UA 视频的分类。结论 本研究证明了 UroSAM 的成功开发,这是一种能从自然内窥镜视频数据中精确识别肾结石的 ML 模型。更多高质量的视频数据将提高该模型对大多数结石成分进行分类的性能。
{"title":"Development of UroSAM: A Machine Learning Model to Automatically Identify Kidney Stone Composition from Endoscopic Video.","authors":"Jixuan Leng, Junfei Liu, Galen Cheng, Haohan Wang, Scott Quarrier, Jiebo Luo, Rajat Jain","doi":"10.1089/end.2023.0740","DOIUrl":"10.1089/end.2023.0740","url":null,"abstract":"<p><p><b><i>Introduction:</i></b> Chemical composition analysis is important in prevention counseling for kidney stone disease. Advances in laser technology have made dusting techniques more prevalent, but this offers no consistent way to collect enough material to send for chemical analysis, leading many to forgo this test. We developed a novel machine learning (ML) model to effectively assess stone composition based on intraoperative endoscopic video data. <b><i>Methods:</i></b> Two endourologists performed ureteroscopy for kidney stones ≥ 10 mm. Representative videos were recorded intraoperatively. Individual frames were extracted from the videos, and the stone was outlined by human tracing. An ML model, UroSAM, was built and trained to automatically identify kidney stones in the images and predict the majority stone composition as follows: calcium oxalate monohydrate (COM), dihydrate (COD), calcium phosphate (CAP), or uric acid (UA). UroSAM was built on top of the publicly available Segment Anything Model (SAM) and incorporated a U-Net convolutional neural network (CNN). <b><i>Discussion:</i></b> A total of 78 ureteroscopy videos were collected; 50 were used for the model after exclusions (32 COM, 8 COD, 8 CAP, 2 UA). The ML model segmented the images with 94.77% precision. Dice coefficient (0.9135) and Intersection over Union (0.8496) confirmed good segmentation performance of the ML model. A video-wise evaluation demonstrated 60% correct classification of stone composition. Subgroup analysis showed correct classification in 84.4% of COM videos. A <i>post hoc</i> adaptive threshold technique was used to mitigate biasing of the model toward COM because of data imbalance; this improved the overall correct classification to 62% while improving the classification of COD, CAP, and UA videos. <b><i>Conclusions:</i></b> This study demonstrates the effective development of UroSAM, an ML model that precisely identifies kidney stones from natural endoscopic video data. More high-quality video data will improve the performance of the model in classifying the majority stone composition.</p>","PeriodicalId":15723,"journal":{"name":"Journal of endourology","volume":null,"pages":null},"PeriodicalIF":2.9,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140957832","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
ChatGPT in Urology: Bridging Knowledge and Practice for Tomorrow's Healthcare, a Comprehensive Review. ChatGPT in Urology:为未来的医疗保健架起知识与实践的桥梁,全面回顾。
IF 2.9 2区 医学 Q1 UROLOGY & NEPHROLOGY Pub Date : 2024-08-01 Epub Date: 2024-07-03 DOI: 10.1089/end.2023.0700
Catalina Solano, Nick Tarazona, Gabriela Prieto Angarita, Andrea Ascencio Medina, Saralia Ruiz, Valentina Melo Pedroza, Olivier Traxer

Background: Among emerging AI technologies, Chat-Generative Pre-Trained Transformer (ChatGPT) emerges as a notable language model, uniquely developed through artificial intelligence research. Its proven versatility across various domains, from language translation to healthcare data processing, underscores its promise within medical documentation, diagnostics, research, and education. The current comprehensive review aimed to investigate the utility of ChatGPT in urology education and practice and to highlight its potential limitations. Methods: The authors conducted a comprehensive literature review of the use of ChatGPT and its applications in urology education, research, and practice. Through a systematic review of the literature, with a search strategy using databases, such as PubMed and Embase, we analyzed the advantages and limitations of using ChatGPT in urology and evaluated its potential impact. Results: A total of 78 records were eligible for inclusion. The benefits of ChatGPT were frequently cited across various contexts. In educational/academic benefits mentioned in 21 records (87.5%), ChatGPT showed the ability to assist urologists by offering precise information and responding to inquiries derived from patient data analysis, thereby supporting decision making; in 18 records (75%), advantages comprised personalized medicine, predictive capabilities for disease risks and outcomes, streamlining clinical workflows and improved diagnostics. Nevertheless, apprehensions were expressed regarding potential misinformation, underscoring the necessity for human supervision to guarantee patient safety and address ethical concerns. Conclusion: The potential applications of ChatGPT hold the capacity to bring about transformative changes in urology education, research, and practice. AI technology can serve as a useful tool to augment human intelligence; however, it is essential to use it in a responsible and ethical manner.

背景介绍在新兴的人工智能技术中,ChatGPT(聊天生成预训练转换器)是一种通过人工智能研究独特开发的语言模型。从语言翻译到医疗保健数据处理,它在各个领域的通用性已得到证实,这凸显了它在医疗文档、诊断、研究和教育领域的前景。本综述旨在研究 ChatGPT 在泌尿外科教育和实践中的实用性,并强调其潜在的局限性。 方法:作者对 ChatGPT 的使用及其在泌尿外科教育、研究和实践中的应用进行了全面的文献综述。通过使用 PubMed 和 Embase 等数据库的搜索策略对文献进行系统性回顾,我们分析了在泌尿外科中使用 ChatGPT 的优势和局限性,并评估了其潜在影响。 结果共有 78 条记录符合纳入条件。ChatGPT 的优点在各种情况下都被频繁提及。21 份记录(87.5%)提到了教育/学术方面的益处,18 份记录(75%)提到了 ChatGPT 的优势,包括个性化医疗、疾病风险和结果的预测能力、简化临床工作流程、改进诊断。然而,也有人对潜在的错误信息表示担忧,强调了人工监督的必要性,以保证患者安全并解决伦理问题:ChatGPT 的潜在应用有能力为泌尿外科教育、研究和实践带来变革。人工智能技术可以作为增强人类智能的有用工具,但必须以负责任和合乎道德的方式使用它。
{"title":"ChatGPT in Urology: Bridging Knowledge and Practice for Tomorrow's Healthcare, a Comprehensive Review.","authors":"Catalina Solano, Nick Tarazona, Gabriela Prieto Angarita, Andrea Ascencio Medina, Saralia Ruiz, Valentina Melo Pedroza, Olivier Traxer","doi":"10.1089/end.2023.0700","DOIUrl":"10.1089/end.2023.0700","url":null,"abstract":"<p><p><b><i>Background:</i></b> Among emerging AI technologies, Chat-Generative Pre-Trained Transformer (ChatGPT) emerges as a notable language model, uniquely developed through artificial intelligence research. Its proven versatility across various domains, from language translation to healthcare data processing, underscores its promise within medical documentation, diagnostics, research, and education. The current comprehensive review aimed to investigate the utility of ChatGPT in urology education and practice and to highlight its potential limitations. <b><i>Methods:</i></b> The authors conducted a comprehensive literature review of the use of ChatGPT and its applications in urology education, research, and practice. Through a systematic review of the literature, with a search strategy using databases, such as PubMed and Embase, we analyzed the advantages and limitations of using ChatGPT in urology and evaluated its potential impact. <b><i>Results:</i></b> A total of 78 records were eligible for inclusion. The benefits of ChatGPT were frequently cited across various contexts. In educational/academic benefits mentioned in 21 records (87.5%), ChatGPT showed the ability to assist urologists by offering precise information and responding to inquiries derived from patient data analysis, thereby supporting decision making; in 18 records (75%), advantages comprised personalized medicine, predictive capabilities for disease risks and outcomes, streamlining clinical workflows and improved diagnostics. Nevertheless, apprehensions were expressed regarding potential misinformation, underscoring the necessity for human supervision to guarantee patient safety and address ethical concerns. <b><i>Conclusion:</i></b> The potential applications of ChatGPT hold the capacity to bring about transformative changes in urology education, research, and practice. AI technology can serve as a useful tool to augment human intelligence; however, it is essential to use it in a responsible and ethical manner.</p>","PeriodicalId":15723,"journal":{"name":"Journal of endourology","volume":null,"pages":null},"PeriodicalIF":2.9,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141317401","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Novel Machine-Learning Algorithm to Predict Stone Recurrence with 24-Hour Urine Data. 利用 24 小时尿液数据预测结石复发的新型机器学习算法
IF 2.9 2区 医学 Q1 UROLOGY & NEPHROLOGY Pub Date : 2024-08-01 DOI: 10.1089/end.2023.0457
Kevin Shee, Andrew W Liu, Carter Chan, Heiko Yang, Wilson Sui, Manoj Desai, Sunita Ho, Thomas Chi, Marshall L Stoller

Objectives: The absence of predictive markers for kidney stone recurrence poses a challenge for the clinical management of stone disease. The unpredictability of stone events is also a significant limitation for clinical trials, where many patients must be enrolled to obtain sufficient stone events for analysis. In this study, we sought to use machine learning methods to identify a novel algorithm to predict stone recurrence. Subjects/Patients and Methods: Patients enrolled in the Registry for Stones of the Kidney and Ureter (ReSKU), a registry of nephrolithiasis patients collected between 2015-2020, with at least one prospectively collected 24-hour urine test (Litholink 24-hour urine test; Labcorp) were included in the training set. A validation set was obtained from chart review of stone patients not enrolled in ReSKU with 24-hour urine data. Stone events were defined as either an office visit where a patient reports symptomatic passage of stones or a surgical procedure for stone removal. Seven prediction classification methods were evaluated. Predictive analyses and receiver operator characteristics (ROC) curve generation were performed in R. Results: A training set of 423 kidney stone patients with stone event data and 24-hour urine samples were trained using the prediction classification methods. The highest performing prediction model was a Logistic Regression with ElasticNet machine learning model (area under curve [AUC] = 0.65). Restricting analysis to high confidence predictions significantly improved model accuracy (AUC = 0.82). The prediction model was validated on a validation set of 172 stone patients with stone event data and 24-hour urine samples. Prediction accuracy in the validation set demonstrated moderate discriminative ability (AUC = 0.64). Repeat modeling was performed with four of the highest scoring features, and ROC analyses demonstrated minimal loss in accuracy (AUC = 0.63). Conclusion: Machine-learning models based on 24-hour urine data can predict stone recurrences with a moderate degree of accuracy.

目的:缺乏肾结石复发的预测标志物给结石病的临床治疗带来了挑战。结石事件的不可预测性也严重限制了临床试验的进行,因为临床试验必须招募许多患者才能获得足够的结石事件进行分析。在本研究中,我们试图利用机器学习方法找出一种预测结石复发的新算法。研究对象/患者和方法:训练集包括加入肾结石和输尿管结石登记处(ReSKU)的患者,该登记处收集了2015-2020年间的肾结石患者,至少有一次前瞻性收集的24小时尿检(Litholink 24小时尿检;Labcorp)。验证集来自未加入 ReSKU 且有 24 小时尿液数据的结石患者的病历审查。结石事件被定义为患者报告无症状排石的门诊就诊或手术取石。对七种预测分类方法进行了评估。结果:使用预测分类方法对包含结石事件数据和 24 小时尿样的 423 名肾结石患者的训练集进行了训练。性能最高的预测模型是带有 ElasticNet 的逻辑回归机器学习模型(曲线下面积 [AUC] = 0.65)。将分析范围限制在高置信度预测上可显著提高模型的准确性(AUC = 0.82)。该预测模型在由 172 名结石患者组成的验证集上进行了验证,验证集包含结石事件数据和 24 小时尿样。验证集的预测准确率显示出中等程度的判别能力(AUC = 0.64)。对四个得分最高的特征进行了重复建模,ROC 分析表明准确性损失最小(AUC = 0.63)。结论基于 24 小时尿液数据的机器学习模型能够以中等程度的准确性预测结石复发。
{"title":"A Novel Machine-Learning Algorithm to Predict Stone Recurrence with 24-Hour Urine Data.","authors":"Kevin Shee, Andrew W Liu, Carter Chan, Heiko Yang, Wilson Sui, Manoj Desai, Sunita Ho, Thomas Chi, Marshall L Stoller","doi":"10.1089/end.2023.0457","DOIUrl":"https://doi.org/10.1089/end.2023.0457","url":null,"abstract":"<p><p><b><i>Objectives:</i></b> The absence of predictive markers for kidney stone recurrence poses a challenge for the clinical management of stone disease. The unpredictability of stone events is also a significant limitation for clinical trials, where many patients must be enrolled to obtain sufficient stone events for analysis. In this study, we sought to use machine learning methods to identify a novel algorithm to predict stone recurrence. <b><i>Subjects/Patients and Methods:</i></b> Patients enrolled in the Registry for Stones of the Kidney and Ureter (ReSKU), a registry of nephrolithiasis patients collected between 2015-2020, with at least one prospectively collected 24-hour urine test (Litholink 24-hour urine test; Labcorp) were included in the training set. A validation set was obtained from chart review of stone patients not enrolled in ReSKU with 24-hour urine data. Stone events were defined as either an office visit where a patient reports symptomatic passage of stones or a surgical procedure for stone removal. Seven prediction classification methods were evaluated. Predictive analyses and receiver operator characteristics (ROC) curve generation were performed in R. <b><i>Results:</i></b> A training set of 423 kidney stone patients with stone event data and 24-hour urine samples were trained using the prediction classification methods. The highest performing prediction model was a Logistic Regression with ElasticNet machine learning model (area under curve [AUC] = 0.65). Restricting analysis to high confidence predictions significantly improved model accuracy (AUC = 0.82). The prediction model was validated on a validation set of 172 stone patients with stone event data and 24-hour urine samples. Prediction accuracy in the validation set demonstrated moderate discriminative ability (AUC = 0.64). Repeat modeling was performed with four of the highest scoring features, and ROC analyses demonstrated minimal loss in accuracy (AUC = 0.63). <b><i>Conclusion:</i></b> Machine-learning models based on 24-hour urine data can predict stone recurrences with a moderate degree of accuracy.</p>","PeriodicalId":15723,"journal":{"name":"Journal of endourology","volume":null,"pages":null},"PeriodicalIF":2.9,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141909806","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Trends of "Artificial Intelligence, Machine Learning, Virtual Reality, and Radiomics in Urolithiasis" over the Last 30 Years (1994-2023) as Published in the Literature (PubMed): A Comprehensive Review. 发表在文献(PubMed)上的“人工智能、机器学习、虚拟现实和泌尿系结石放射组学”在过去30年(1994-2023)的趋势:综合综述。
IF 2.9 2区 医学 Q1 UROLOGY & NEPHROLOGY Pub Date : 2024-08-01 DOI: 10.1089/end.2023.0263
Carlotta Nedbal, Clara Cerrato, Victoria Jahrreiss, Amelia Pietropaolo, Andrea Benedetto Galosi, Daniele Castellani, Bhaskar Kumar Somani

Purpose: To analyze the bibliometric publication trend on the application of "Artificial Intelligence (AI) and its subsets (Machine Learning-ML, Virtual reality-VR, Radiomics) in Urolithiasis" over 3 decades. We looked at the publication trends associated with AI and stone disease, including both clinical and surgical applications, and training in endourology. Methods: Through a MeshTerms research on PubMed, we performed a comprehensive review from 1994-2023 for all published articles on "AI, ML, VR, and Radiomics." Articles were then divided into three categories as follows: A-Clinical (Nonsurgical), B-Clinical (Surgical), and C-Training articles, and articles were then assigned to following three periods: Period-1 (1994-2003), Period-2 (2004-2013), and Period-3 (2014-2023). Results: A total of 343 articles were noted (Groups A-129, B-163, and C-51), and trends increased from Period-1 to Period-2 at 123% (p = 0.009) and to period-3 at 453% (p = 0.003). This increase from Period-2 to Period-3 for groups A, B, and C was 476% (p = 0.019), 616% (0.001), and 185% (p < 0.001), respectively. Group A articles included rise in articles on "stone characteristics" (+2100%; p = 0.011), "renal function" (p = 0.002), "stone diagnosis" (+192%), "prediction of stone passage" (+400%), and "quality of life" (+1000%). Group B articles included rise in articles on "URS" (+2650%, p = 0.008), "PCNL"(+600%, p = 0.001), and "SWL" (+650%, p = 0.018). Articles on "Targeting" (+453%, p < 0.001), "Outcomes" (+850%, p = 0.013), and "Technological Innovation" (p = 0.0311) had rising trends. Group C articles included rise in articles on "PCNL" (+300%, p = 0.039) and "URS" (+188%, p = 0.003). Conclusion: Publications on AI and its subset areas for urolithiasis have seen an exponential increase over the last decade, with an increase in surgical and nonsurgical clinical areas, as well as in training. Future AI related growth in the field of endourology and urolithiasis is likely to improve training, patient centered decision-making, and clinical outcomes.

目的:分析近30年来“人工智能(AI)及其子集(机器学习ML、虚拟现实VR、放射组学)在泌尿系结石中的应用”的文献计量学出版趋势。我们研究了与人工智能和结石疾病相关的出版趋势,包括临床和外科应用,以及腔内泌尿外科的培训。方法:通过在PubMed上进行MeshTerms研究,我们对1994-2023年间发表的所有关于“AI、ML、VR和放射组学”的论文进行了全面综述。然后将论文分为三类:A临床(非手术)、B临床(手术)和C训练论文,然后将文章分为三个时期:第1期(1994-2003)、第2期(2004-2013)和第3期(2014-2023)。结果:共记录了343篇论文(A-129、B-163和C-51组),从第1期到第2期的趋势增加了123%(p=0.009),A、B和C组从2期增加到3期为476%(p=0.019),616%(0.001)和185%。
{"title":"Trends of \"Artificial Intelligence, Machine Learning, Virtual Reality, and Radiomics in Urolithiasis\" over the Last 30 Years (1994-2023) as Published in the Literature (PubMed): A Comprehensive Review.","authors":"Carlotta Nedbal, Clara Cerrato, Victoria Jahrreiss, Amelia Pietropaolo, Andrea Benedetto Galosi, Daniele Castellani, Bhaskar Kumar Somani","doi":"10.1089/end.2023.0263","DOIUrl":"10.1089/end.2023.0263","url":null,"abstract":"<p><p><b><i>Purpose:</i></b> To analyze the bibliometric publication trend on the application of \"Artificial Intelligence (AI) and its subsets (Machine Learning-ML, Virtual reality-VR, Radiomics) in Urolithiasis\" over 3 decades. We looked at the publication trends associated with AI and stone disease, including both clinical and surgical applications, and training in endourology. <b><i>Methods:</i></b> Through a MeshTerms research on PubMed, we performed a comprehensive review from 1994-2023 for all published articles on \"AI, ML, VR, and Radiomics.\" Articles were then divided into three categories as follows: A-Clinical (Nonsurgical), B-Clinical (Surgical), and C-Training articles, and articles were then assigned to following three periods: Period-1 (1994-2003), Period-2 (2004-2013), and Period-3 (2014-2023). <b><i>Results:</i></b> A total of 343 articles were noted (Groups A-129, B-163, and C-51), and trends increased from Period-1 to Period-2 at 123% (<i>p</i> = 0.009) and to period-3 at 453% (<i>p</i> = 0.003). This increase from Period-2 to Period-3 for groups A, B, and C was 476% (<i>p</i> = 0.019), 616% (0.001), and 185% (<i>p</i> < 0.001), respectively. Group A articles included rise in articles on \"stone characteristics\" (+2100%; <i>p</i> = 0.011), \"renal function\" (<i>p</i> = 0.002), \"stone diagnosis\" (+192%), \"prediction of stone passage\" (+400%), and \"quality of life\" (+1000%). Group B articles included rise in articles on \"URS\" (+2650%, <i>p</i> = 0.008), \"PCNL\"(+600%, <i>p</i> = 0.001), and \"SWL\" (+650%, <i>p</i> = 0.018). Articles on \"Targeting\" (+453%, <i>p</i> < 0.001), \"Outcomes\" (+850%, <i>p</i> = 0.013), and \"Technological Innovation\" (<i>p</i> = 0.0311) had rising trends. Group C articles included rise in articles on \"PCNL\" (+300%, <i>p</i> = 0.039) and \"URS\" (+188%, <i>p</i> = 0.003). <b><i>Conclusion:</i></b> Publications on AI and its subset areas for urolithiasis have seen an exponential increase over the last decade, with an increase in surgical and nonsurgical clinical areas, as well as in training. Future AI related growth in the field of endourology and urolithiasis is likely to improve training, patient centered decision-making, and clinical outcomes.</p>","PeriodicalId":15723,"journal":{"name":"Journal of endourology","volume":null,"pages":null},"PeriodicalIF":2.9,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"54229309","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Evaluation of the Current Status of Artificial Intelligence for Endourology Patient Education: A Blind Comparison of ChatGPT and Google Bard Against Traditional Information Resources. 人工智能在腔内泌尿科患者教育中的应用现状评估:ChatGPT 和 Google Bard 与传统信息资源的盲比。
IF 2.9 2区 医学 Q1 UROLOGY & NEPHROLOGY Pub Date : 2024-08-01 Epub Date: 2024-05-17 DOI: 10.1089/end.2023.0696
Christopher Connors, Kavita Gupta, Johnathan A Khusid, Raymond Khargi, Alan J Yaghoubian, Micah Levy, Blair Gallante, William Atallah, Mantu Gupta

Introduction: Artificial intelligence (AI) platforms such as ChatGPT and Bard are increasingly utilized to answer patient health care questions. We present the first study to blindly evaluate AI-generated responses to common endourology patient questions against official patient education materials. Methods: Thirty-two questions and answers spanning kidney stones, ureteral stents, benign prostatic hyperplasia (BPH), and upper tract urothelial carcinoma were extracted from official Urology Care Foundation (UCF) patient education documents. The same questions were input into ChatGPT 4.0 and Bard, limiting responses to within ±10% of the word count of the corresponding UCF response to ensure fair comparison. Six endourologists blindly evaluated responses from each platform using Likert scales for accuracy, clarity, comprehensiveness, and patient utility. Reviewers identified which response they believed was not AI generated. Finally, Flesch-Kincaid Reading Grade Level formulas assessed the readability of each platform response. Ratings were compared using analysis of variance (ANOVA) and chi-square tests. Results: ChatGPT responses were rated the highest across all categories, including accuracy, comprehensiveness, clarity, and patient utility, while UCF answers were consistently scored the lowest, all p < 0.01. A subanalysis revealed that this trend was consistent across question categories (i.e., kidney stones, BPH, etc.). However, AI-generated responses were more likely to be classified at an advanced reading level, while UCF responses showed improved readability (college or higher reading level: ChatGPT = 100%, Bard = 66%, and UCF = 19%), p < 0.001. When asked to identify which answer was not AI generated, 54.2% of responses indicated ChatGPT, 26.6% indicated Bard, and only 19.3% correctly identified it as the UCF response. Conclusions: In a blind evaluation, AI-generated responses from ChatGPT and Bard surpassed the quality of official patient education materials in endourology, suggesting that current AI platforms are already a reliable resource for basic urologic care information. AI-generated responses do, however, tend to require a higher reading level, which may limit their applicability to a broader audience.

导言 ChatGPT 和 Bard 等人工智能(AI)平台越来越多地被用于回答患者的医疗保健问题。我们首次进行了一项研究,对照官方患者教育材料,对人工智能生成的常见腔内泌尿科患者问题的回答进行了盲评。方法 从官方泌尿外科护理基金会(UCF)患者教育文件中提取了 32 个问题和答案,涉及肾结石、输尿管支架、良性前列腺增生症和UTUC。同样的问题被输入到 ChatGPT 4.0 和 Bard 中,为确保公平比较,回答字数限制在相应 UCF 回答字数的  10% 以内。六位内科专家使用李克特量表对每个平台上的回答进行盲评,评估内容包括准确性、清晰度、全面性和对患者的实用性。评审人员确定了他们认为不是人工智能生成的回复。最后,Flesch-Kincaid 阅读等级公式评估了每个平台回复的可读性。评分采用方差分析和齐次方检验进行比较。结果 在包括准确性、全面性、清晰度和患者实用性在内的所有类别中,ChatGPT 回答的评分最高,而 UCF 回答的评分一直最低,所有 p 均为 0。
{"title":"Evaluation of the Current Status of Artificial Intelligence for Endourology Patient Education: A Blind Comparison of ChatGPT and Google Bard Against Traditional Information Resources.","authors":"Christopher Connors, Kavita Gupta, Johnathan A Khusid, Raymond Khargi, Alan J Yaghoubian, Micah Levy, Blair Gallante, William Atallah, Mantu Gupta","doi":"10.1089/end.2023.0696","DOIUrl":"10.1089/end.2023.0696","url":null,"abstract":"<p><p><b><i>Introduction:</i></b> Artificial intelligence (AI) platforms such as ChatGPT and Bard are increasingly utilized to answer patient health care questions. We present the first study to blindly evaluate AI-generated responses to common endourology patient questions against official patient education materials. <b><i>Methods:</i></b> Thirty-two questions and answers spanning kidney stones, ureteral stents, benign prostatic hyperplasia (BPH), and upper tract urothelial carcinoma were extracted from official Urology Care Foundation (UCF) patient education documents. The same questions were input into ChatGPT 4.0 and Bard, limiting responses to within ±10% of the word count of the corresponding UCF response to ensure fair comparison. Six endourologists blindly evaluated responses from each platform using Likert scales for accuracy, clarity, comprehensiveness, and patient utility. Reviewers identified which response they believed was not AI generated. Finally, Flesch-Kincaid Reading Grade Level formulas assessed the readability of each platform response. Ratings were compared using analysis of variance (ANOVA) and chi-square tests. <b><i>Results:</i></b> ChatGPT responses were rated the highest across all categories, including accuracy, comprehensiveness, clarity, and patient utility, while UCF answers were consistently scored the lowest, all <i>p</i> < 0.01. A subanalysis revealed that this trend was consistent across question categories (i.e., kidney stones, BPH, etc.). However, AI-generated responses were more likely to be classified at an advanced reading level, while UCF responses showed improved readability (college or higher reading level: ChatGPT = 100%, Bard = 66%, and UCF = 19%), <i>p</i> < 0.001. When asked to identify which answer was not AI generated, 54.2% of responses indicated ChatGPT, 26.6% indicated Bard, and only 19.3% correctly identified it as the UCF response. <b><i>Conclusions:</i></b> In a blind evaluation, AI-generated responses from ChatGPT and Bard surpassed the quality of official patient education materials in endourology, suggesting that current AI platforms are already a reliable resource for basic urologic care information. AI-generated responses do, however, tend to require a higher reading level, which may limit their applicability to a broader audience.</p>","PeriodicalId":15723,"journal":{"name":"Journal of endourology","volume":null,"pages":null},"PeriodicalIF":2.9,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140028206","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Automated Analysis of Split Kidney Function from CT Scans Using Deep Learning and Delta Radiomics. 利用深度学习和德尔塔放射组学自动分析 CT 扫描中的分肾功能。
IF 2.9 2区 医学 Q1 UROLOGY & NEPHROLOGY Pub Date : 2024-08-01 Epub Date: 2024-05-16 DOI: 10.1089/end.2023.0488
Ramon Luis Correa-Medero, Jiwoong Jeong, Bhavik Patel, Imon Banerjee, Haidar Abdul-Muhsin

Background: Differential kidney function assessment is an important part of preoperative evaluation of various urological interventions. It is obtained through dedicated nuclear medical imaging and is not yet implemented through conventional Imaging. Objective: We assess if differential kidney function can be obtained through evaluation of contrast-enhanced computed tomography(CT) using a combination of deep learning and (2D and 3D) radiomic features. Methods: All patients who underwent kidney nuclear scanning at Mayo Clinic sites between 2018-2022 were collected. CT scans of the kidneys were obtained within a 3-month interval before or after the nuclear scans were extracted. Patients who underwent a urological or radiological intervention within this time frame were excluded. A segmentation model was used to segment both kidneys. 2D and 3D radiomics features were extracted and compared between the two kidneys to compute delta radiomics and assess its ability to predict differential kidney function. Performance was reported using receiver operating characteristics, sensitivity, and specificity. Results: Studies from Arizona & Rochester formed our internal dataset (n = 1,159). Studies from Florida were separately processed as an external test set to validate generalizability. We obtained 323 studies from our internal sites and 39 studies from external sites. The best results were obtained by a random forest model trained on 3D delta radiomics features. This model achieved an area under curve (AUC) of 0.85 and 0.81 on internal and external test sets, while specificity and sensitivity were 0.84,0.68 on the internal set, 0.70, and 0.65 on the external set. Conclusion: This proposed automated pipeline can derive important differential kidney function information from contrast-enhanced CT and reduce the need for dedicated nuclear scans for early-stage differential kidney functional assessment. Clinical Impact: We establish a machine learning methodology for assessing differential kidney function from routine CT without the need for expensive and radioactive nuclear medicine scans.

背景 肾功能鉴别评估是各种泌尿外科手术术前评估的重要组成部分。它是通过专门的核医学成像获得的,尚未通过常规成像实现:我们利用深度学习与(二维和三维)放射学特征相结合的方法,评估能否通过对比增强计算机断层扫描(CT)评估获得不同的肾功能。方法 收集2018-2022年间在梅奥诊所接受肾脏核扫描的所有患者。肾脏的 CT 扫描是在提取核扫描之前或之后的三个月间隔内获得的。在此时间段内接受过泌尿外科或放射介入治疗的患者被排除在外。使用分割模型对双肾进行分割。提取二维和三维放射组学特征并在两个肾脏之间进行比较,以计算delta放射组学并评估其预测不同肾功能的能力。使用接收器操作特性、灵敏度和特异性报告结果:亚利桑那州和罗切斯特的研究构成了我们的内部数据集(n=1,159)。来自佛罗里达州的研究作为外部测试集单独处理,以验证可推广性。我们从内部网站获得了 323 项研究,从外部网站获得了 39 项研究。根据三维三角放射组学特征训练的随机森林模型取得了最佳结果。该模型在内部和外部测试集上的AUC分别为0.85和0.81,特异性和灵敏度在内部测试集上分别为0.84和0.68,在外部测试集上分别为0.70和0.65:结论:这一拟议的自动化管道可从对比增强 CT 中获取重要的肾功能差异信息,并减少早期肾功能差异评估对专用核扫描的需求:临床影响:我们建立了一种机器学习方法,可从常规 CT 中评估肾功能差异,而无需进行昂贵的放射性核医学扫描。
{"title":"Automated Analysis of Split Kidney Function from CT Scans Using Deep Learning and Delta Radiomics.","authors":"Ramon Luis Correa-Medero, Jiwoong Jeong, Bhavik Patel, Imon Banerjee, Haidar Abdul-Muhsin","doi":"10.1089/end.2023.0488","DOIUrl":"10.1089/end.2023.0488","url":null,"abstract":"<p><p><b><i>Background:</i></b> Differential kidney function assessment is an important part of preoperative evaluation of various urological interventions. It is obtained through dedicated nuclear medical imaging and is not yet implemented through conventional Imaging. <b><i>Objective:</i></b> We assess if differential kidney function can be obtained through evaluation of contrast-enhanced computed tomography(CT) using a combination of deep learning and (2D and 3D) radiomic features. <b><i>Methods:</i></b> All patients who underwent kidney nuclear scanning at Mayo Clinic sites between 2018-2022 were collected. CT scans of the kidneys were obtained within a 3-month interval before or after the nuclear scans were extracted. Patients who underwent a urological or radiological intervention within this time frame were excluded. A segmentation model was used to segment both kidneys. 2D and 3D radiomics features were extracted and compared between the two kidneys to compute delta radiomics and assess its ability to predict differential kidney function. Performance was reported using receiver operating characteristics, sensitivity, and specificity. <b><i>Results:</i></b> Studies from Arizona & Rochester formed our internal dataset (<i>n</i> = 1,159). Studies from Florida were separately processed as an external test set to validate generalizability. We obtained 323 studies from our internal sites and 39 studies from external sites. The best results were obtained by a random forest model trained on 3D delta radiomics features. This model achieved an area under curve (AUC) of 0.85 and 0.81 on internal and external test sets, while specificity and sensitivity were 0.84,0.68 on the internal set, 0.70, and 0.65 on the external set. <b><i>Conclusion:</i></b> This proposed automated pipeline can derive important differential kidney function information from contrast-enhanced CT and reduce the need for dedicated nuclear scans for early-stage differential kidney functional assessment. <b><i>Clinical Impact:</i></b> We establish a machine learning methodology for assessing differential kidney function from routine CT without the need for expensive and radioactive nuclear medicine scans.</p>","PeriodicalId":15723,"journal":{"name":"Journal of endourology","volume":null,"pages":null},"PeriodicalIF":2.9,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140863086","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Evaluating the Performance of ChatGPT in Urology: A Comparative Study of Knowledge Interpretation and Patient Guidance. 评估 ChatGPT 在泌尿外科中的表现:知识解释与患者指导的比较研究。
IF 2.9 2区 医学 Q1 UROLOGY & NEPHROLOGY Pub Date : 2024-08-01 Epub Date: 2024-05-30 DOI: 10.1089/end.2023.0413
Bahadır Şahin, Yunus Emre Genç, Kader Doğan, Tarık Emre Şener, Çağrı Akın Şekerci, Yılören Tanıdır, Selçuk Yücel, Tufan Tarcan, Haydar Kamil Çam

Background/Aim: To evaluate the performance of Chat Generative Pre-trained Transformer (ChatGPT), a large language model trained by Open artificial intelligence. Materials and Methods: This study has three main steps to evaluate the effectiveness of ChatGPT in the urologic field. The first step involved 35 questions from our institution's experts, who have at least 10 years of experience in their fields. The responses of ChatGPT versions were qualitatively compared with the responses of urology residents to the same questions. The second step assesses the reliability of ChatGPT versions in answering current debate topics. The third step was to assess the reliability of ChatGPT versions in providing medical recommendations and directives to patients' commonly asked questions during the outpatient and inpatient clinic. Results: In the first step, version 4 provided correct answers to 25 questions out of 35 while version 3.5 provided only 19 (71.4% vs 54%). It was observed that residents in their last year of education in our clinic also provided a mean of 25 correct answers, and 4th year residents provided a mean of 19.3 correct responses. The second step involved evaluating the response of both versions to debate situations in urology, and it was found that both versions provided variable and inappropriate results. In the last step, both versions had a similar success rate in providing recommendations and guidance to patients based on expert ratings. Conclusion: The difference between the two versions of the 35 questions in the first step of the study was thought to be due to the improvement of ChatGPT's literature and data synthesis abilities. It may be a logical approach to use ChatGPT versions to inform the nonhealth care providers' questions with quick and safe answers but should not be used to as a diagnostic tool or make a choice among different treatment modalities.

背景/目的:评估由开放人工智能训练的大型语言模型--聊天生成预训练转换器(ChatGPT)的性能。材料与方法:本研究通过三个主要步骤来评估 ChatGPT 在泌尿科领域的有效性。第一步是由我们机构的专家提出 35 个问题,这些专家在各自领域至少有 10 年的经验。ChatGPT 版本的回答与泌尿科住院医师对相同问题的回答进行了定性比较。第二步是评估 ChatGPT 版本在回答当前辩论话题时的可靠性。第三步是评估 ChatGPT 版本在门诊和住院期间针对患者常见问题提供医疗建议和指示的可靠性。结果:第一步,在 35 个问题中,第 4 版提供了 25 个问题的正确答案,而第 3.5 版仅提供了 19 个问题的正确答案(71.4% 对 54%)。据观察,在本诊所接受最后一年教育的住院医师也平均提供了 25 个正确答案,而四年级住院医师平均提供了 19.3 个正确答案。第二步是评估两个版本对泌尿科辩论情况的反应,结果发现两个版本都提供了不同的、不恰当的结果。最后一步,两个版本在根据专家评分向患者提供建议和指导方面的成功率相似。结论研究第一步中两个版本 35 个问题之间的差异被认为是由于 ChatGPT 文献和数据综合能力的提高。使用 ChatGPT 版本为非医疗服务提供者的问题提供快速、安全的答案可能是一种合理的方法,但不应被用作诊断工具或在不同的治疗方式中做出选择。
{"title":"Evaluating the Performance of ChatGPT in Urology: A Comparative Study of Knowledge Interpretation and Patient Guidance.","authors":"Bahadır Şahin, Yunus Emre Genç, Kader Doğan, Tarık Emre Şener, Çağrı Akın Şekerci, Yılören Tanıdır, Selçuk Yücel, Tufan Tarcan, Haydar Kamil Çam","doi":"10.1089/end.2023.0413","DOIUrl":"10.1089/end.2023.0413","url":null,"abstract":"<p><p><b><i>Background/Aim:</i></b> To evaluate the performance of Chat Generative Pre-trained Transformer (ChatGPT), a large language model trained by Open artificial intelligence. <b><i>Materials and Methods:</i></b> This study has three main steps to evaluate the effectiveness of ChatGPT in the urologic field. The first step involved 35 questions from our institution's experts, who have at least 10 years of experience in their fields. The responses of ChatGPT versions were qualitatively compared with the responses of urology residents to the same questions. The second step assesses the reliability of ChatGPT versions in answering current debate topics. The third step was to assess the reliability of ChatGPT versions in providing medical recommendations and directives to patients' commonly asked questions during the outpatient and inpatient clinic. <b><i>Results:</i></b> In the first step, version 4 provided correct answers to 25 questions out of 35 while version 3.5 provided only 19 (71.4% <i>vs</i> 54%). It was observed that residents in their last year of education in our clinic also provided a mean of 25 correct answers, and 4th year residents provided a mean of 19.3 correct responses. The second step involved evaluating the response of both versions to debate situations in urology, and it was found that both versions provided variable and inappropriate results. In the last step, both versions had a similar success rate in providing recommendations and guidance to patients based on expert ratings. <b><i>Conclusion:</i></b> The difference between the two versions of the 35 questions in the first step of the study was thought to be due to the improvement of ChatGPT's literature and data synthesis abilities. It may be a logical approach to use ChatGPT versions to inform the nonhealth care providers' questions with quick and safe answers but should not be used to as a diagnostic tool or make a choice among different treatment modalities.</p>","PeriodicalId":15723,"journal":{"name":"Journal of endourology","volume":null,"pages":null},"PeriodicalIF":2.9,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141179798","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Predictive Modeling of Urinary Stone Composition Using Machine Learning and Clinical Data: Implications for Treatment Strategies and Pathophysiological Insights. 利用机器学习和临床数据对尿路结石组成进行预测建模:对治疗策略和病理生理学见解的影响。
IF 2.9 2区 医学 Q1 UROLOGY & NEPHROLOGY Pub Date : 2024-08-01 Epub Date: 2024-05-30 DOI: 10.1089/end.2023.0446
John A Chmiel, Gerrit A Stuivenberg, Jennifer F W Wong, Linda Nott, Jeremy P Burton, Hassan Razvi, Jennifer Bjazevic

Purpose: Preventative strategies and surgical treatments for urolithiasis depend on stone composition. However, stone composition is often unknown until the stone is passed or surgically managed. Given that stone composition likely reflects the physiological parameters during its formation, we used clinical data from stone formers to predict stone composition. Materials and Methods: Data on stone composition, 24-hour urine, serum biochemistry, patient demographics, and medical history were prospectively collected from 777 kidney stone patients. Data were used to train gradient boosted machine and logistic regression models to distinguish calcium vs noncalcium, calcium oxalate monohydrate vs dihydrate, and calcium oxalate vs calcium phosphate vs uric acid stone types. Model performance was evaluated using the kappa score, and the influence of each predictor variable was assessed. Results: The calcium vs noncalcium model differentiated stone types with a kappa of 0.5231. The most influential predictors were 24-hour urine calcium, blood urate, and phosphate. The calcium oxalate monohydrate vs dihydrate model is the first of its kind and could discriminate stone types with a kappa of 0.2042. The key predictors were 24-hour urine urea, calcium, and oxalate. The multiclass model had a kappa of 0.3023 and the top predictors were age and 24-hour urine calcium and creatinine. Conclusions: Clinical data can be leveraged with machine learning algorithms to predict stone composition, which may help urologists determine stone type and guide their management plan before stone treatment. Investigating the most influential predictors of each classifier may improve the understanding of key clinical features of urolithiasis and shed light on pathophysiology.

目的:尿石症的预防策略和手术治疗取决于结石的组成。然而,在结石排出或手术处理之前,通常不知道结石的成分。鉴于结石成分可能反映其形成过程中的生理参数,我们使用来自结石患者的临床数据来预测结石成分。材料与方法:前瞻性收集777例肾结石患者的结石组成、24小时尿液、血清生化、患者人口统计学和病史。数据用于训练梯度增强机器和逻辑回归模型,以区分钙与非钙,一水草酸钙与二水草酸钙,草酸钙与磷酸钙与尿酸结石类型。采用kappa评分评估模型性能,并评估各预测变量的影响。结果:钙与非钙模型成功区分了结石类型,kappa为0.5231。影响最大的预测因子是24小时尿钙、血尿酸和磷酸盐。草酸钙一水/二水模型是同类模型中第一个,kappa值为0.2042,可以区分结石类型。关键预测指标是24小时尿尿素、钙和草酸盐。多类模型kappa为0.3023,预测因子为年龄、24小时尿钙和肌酐。结论:临床数据可以与机器学习算法相结合来预测结石成分,这可能有助于泌尿科医生确定结石类型并指导他们在结石治疗前的管理计划。研究每种分类中最具影响力的预测因子可以提高对尿石症关键临床特征的理解,并阐明其病理生理学。
{"title":"Predictive Modeling of Urinary Stone Composition Using Machine Learning and Clinical Data: Implications for Treatment Strategies and Pathophysiological Insights.","authors":"John A Chmiel, Gerrit A Stuivenberg, Jennifer F W Wong, Linda Nott, Jeremy P Burton, Hassan Razvi, Jennifer Bjazevic","doi":"10.1089/end.2023.0446","DOIUrl":"10.1089/end.2023.0446","url":null,"abstract":"<p><p><b><i>Purpose:</i></b> Preventative strategies and surgical treatments for urolithiasis depend on stone composition. However, stone composition is often unknown until the stone is passed or surgically managed. Given that stone composition likely reflects the physiological parameters during its formation, we used clinical data from stone formers to predict stone composition. <b><i>Materials and Methods:</i></b> Data on stone composition, 24-hour urine, serum biochemistry, patient demographics, and medical history were prospectively collected from 777 kidney stone patients. Data were used to train gradient boosted machine and logistic regression models to distinguish calcium <i>vs</i> noncalcium, calcium oxalate monohydrate <i>vs</i> dihydrate, and calcium oxalate <i>vs</i> calcium phosphate <i>vs</i> uric acid stone types. Model performance was evaluated using the kappa score, and the influence of each predictor variable was assessed. <b><i>Results:</i></b> The calcium <i>vs</i> noncalcium model differentiated stone types with a kappa of 0.5231. The most influential predictors were 24-hour urine calcium, blood urate, and phosphate. The calcium oxalate monohydrate <i>vs</i> dihydrate model is the first of its kind and could discriminate stone types with a kappa of 0.2042. The key predictors were 24-hour urine urea, calcium, and oxalate. The multiclass model had a kappa of 0.3023 and the top predictors were age and 24-hour urine calcium and creatinine. <b><i>Conclusions:</i></b> Clinical data can be leveraged with machine learning algorithms to predict stone composition, which may help urologists determine stone type and guide their management plan before stone treatment. Investigating the most influential predictors of each classifier may improve the understanding of key clinical features of urolithiasis and shed light on pathophysiology.</p>","PeriodicalId":15723,"journal":{"name":"Journal of endourology","volume":null,"pages":null},"PeriodicalIF":2.9,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"136397694","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Journal of endourology
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1