人工智能时代脑电图教育的未来。

IF 6.6 1区 医学 Q1 CLINICAL NEUROLOGY Epilepsia Pub Date : 2025-03-04 DOI:10.1111/epi.18326
John R. McLaren, Doyle Yuan, Sándor Beniczky, M. Brandon Westover, Fábio A. Nascimento
{"title":"人工智能时代脑电图教育的未来。","authors":"John R. McLaren,&nbsp;Doyle Yuan,&nbsp;Sándor Beniczky,&nbsp;M. Brandon Westover,&nbsp;Fábio A. Nascimento","doi":"10.1111/epi.18326","DOIUrl":null,"url":null,"abstract":"<p>Artificial intelligence (AI), the technology that enables computers to simulate human problem-solving capabilities, is rapidly evolving. In the field of epilepsy, AI's application has already demonstrated potential for improved quality, cost, and access to patient care.<span><sup>1</sup></span> Although these advancements are exciting, they also pose several questions for the epilepsy community that are imperative to address now, before the imminent implementation of AI in clinical practice. As academics and educators, one question we are often asked is whether, or to what degree, electroencephalography (EEG) interpretation should be taught to the next generations of (human) neurologists. To address this in a meaningful way, we first examine our current predicament and then anticipate the road ahead. In doing so, we argue for a system that carefully integrates AI-based algorithms into the workflow of expert human EEG interpretation, which will ultimately necessitate a change in the way we educate our trainees and the next generation of electroencephalographers.</p><p>EEG is the tool most often used in the diagnostic evaluation of individuals with suspected epilepsy and frequently employed during sleep studies, surgeries, and in neurocritical care settings; therefore, accurate and reliable EEG interpretation is essential for optimal care of a variety of patients. In real-world practice, EEG misinterpretation does occur—either through over-calling normal EEG patterns as abnormal, or under-calling abnormal findings as benign, both of which can negatively impact patient care and outcomes. EEG “over-calling” (i.e., false-positive errors) can lead to epilepsy misdiagnosis with resultant unnecessary driving restrictions, employment difficulties, overprescription of anti-seizure medications, unwarranted surgery, inappropriate prognostication, other forms of comorbid stigma and social marginalization, as well as notable negative effects to health care systems. Meanwhile, “under-calling” (i.e., false-negative errors) can result in missed opportunities to prevent seizure-related injury, negative cognitive sequelae, and, in some circumstances, death.<span><sup>2-8</sup></span> Additional implications are well documented in patients without epilepsy but where EEG is still used.<span><sup>9</sup></span></p><p>Unfortunately, there are significant gaps in the current landscape of EEG education. For both adult and pediatric neurology training programs, there is limited exposure to EEG, subpar quality teaching, a paucity of objective competencies, and significant inter-program variability.<span><sup>10, 11</sup></span> As such, many neurology graduates leave training without feeling confident in their EEG-reading capabilities.<span><sup>12, 13</sup></span></p><p>This presents a substantial problem in clinical practice, as a large portion of EEG studies across the world—including the United States and many European countries—are interpreted by general neurologists without additional EEG/epilepsy training.<span><sup>14, 15</sup></span></p><p>With so much room for improvement, it is easy to see why AI-centered EEG interpretation holds promise. Several automated AI algorithms have shown high accuracy in identifying EEG patterns including interictal epileptiform discharges (IEDs), interictal slowing, seizures, and findings on the ictal-interictal continuum. Some of these algorithms have been validated in routine and critical care EEGs, achieving human expert level performance with improved efficiency and consistency.<span><sup>16-24</sup></span> As such, it is more than plausible that the future will involve appreciable adoption of these technologies. However, we need to consider the larger implications of implementation and have the foresight to address them in a meaningful way through proper education.</p><p>First, one must recognize that the high accuracy and efficiency of a tool does not always lead to equivalent levels of credibility. Human, or in this case, patient, perception is an important variable that changes an otherwise straightforward equation. The credibility of AI systems, regardless of their accuracy, remains generally quite low. In large global surveys on perceptions of this new technology, most people are wary about trusting AI and have a low or moderate acceptance of it.<span><sup>25</sup></span> Due to the inherent legal liabilities and ethical considerations<span><sup>26</sup></span> involved in misdiagnosis and mismanagement, which have not yet been fully elucidated, it is implausible that fully autonomous AI would be applied to the diagnostic or therapeutic process until the community at large takes time to thoughtfully address these issues (e.g., through the establishment of minimum standards).</p><p>Furthermore, the many limitations of AI must be acknowledged. Even the highest quality AI models are primarily data-driven algorithms that are trained and validated on finite, historical datasets with expert input. These datasets, if not properly designed, can perpetuate discriminatory practices, give rise to “hallucinations” (false predictions),<span><sup>27</sup></span> and create suboptimal behavioral change, all of which could impact patient care. As there is not yet a complete, automated EEG interpretation system available, we must carefully consider the datasets on which these AI algorithms were trained and validated to ensure that they are applied in the appropriate context. For example, SPaRCNET<span><sup>16</sup></span> was developed to read critical care EEG, whereas SCORE-AI<span><sup>18, 19</sup></span> was developed to read outpatient/routine EEG; hence, a one size fits all approach to interpretation would be imprudent without more rigorous investigation. We also must acknowledge that many seizure types have markedly different electrographic signals (e.g., neonatal seizures, epileptic spasms, atonic seizures, myoclonic seizures, and so on) for which AI algorithms have not been developed and will require trained human readers to recognize these patterns along with corresponding video of the events, if available, at least until more sophisticated models are developed. As such, the most plausible scenario, and the gold standard we should aim for, is a complementary “hybrid” model<span><sup>20</sup></span>—one in which augmentation with AI could allow experts to make higher quality decisions more efficiently than those not using the technology, while maintaining oversight and human credibility that will best serve our patients, particularly in cases where the algorithms fall short.</p><p>Over the past few years, examples of some hybrid approaches have been demonstrated that have the potential to improve interpretation and promote better, more-efficient reading skills. In a comparison between three fully automated AI algorithms and a human-supervised AI model applying an operational definition to detect IEDs, the specificity of the fully automated approaches was too low for clinical implementation, despite the high sensitivity. Meanwhile the human-supervised approach significantly increased the specificity, maintained good sensitivity and accuracy, and decreased the time burden of review compared to conventional visual analysis. In another study, a more “interpretable” deep learning model that accurately classifies six patterns of potentially harmful EEG activity was applied to EEG recordings in the intensive care unit (ICU) setting and led to significant pattern classification accuracy improvement by human readers using this assisted technology.<span><sup>28</sup></span> It is important to note that this performance was also significantly better than that of a corresponding uninterpretable “black-box” model, demonstrating further promise for AI–human collaboration but also highlighting the importance of design and context for the end user.</p><p>Additional applications that have not yet been described or validated should also make interpretation more efficient and education more focused. For example, applying appropriate AI algorithms with high sensitivities will allow trainees to focus less on the more time-consuming aspects of interpretation, such as scrolling through long, multiday continuous EEG files, and instead focus on interpreting snippets flagged as “concerning” by AI. Similarly, rapid-response EEG systems with high negative predictive value could be used to triage which EEG recordings require more detailed review overnight and which could wait until morning, freeing up considerable clinical burden on trainees. Better automated detection algorithms could be applied in ICU settings to reduce “alarm fatigue,” a longstanding issue that has been well documented to affect clinician concentration and diagnostic accuracy, leading to burnout and reduced quality of life.<span><sup>29-32</sup></span> In resource-limited settings, where there are either no epileptologists, or general neurologists without proper exposure to EEG reading during training, AI implementation could help mitigate these deficits. Although these algorithms can run locally on a tablet without internet connection, we also acknowledge that many low- and middle-income areas lack the important high-tech infrastructure for easy adoption,<span><sup>33</sup></span> such as the resources for machinery, personnel, quality control, and cloud storage, all of which can contribute to an increased carbon footprint and may limit feasibility of use.<span><sup>26</sup></span> Finally, in challenging or equivocal EEG cases, AI interpretation could offer a “second opinion” for readers, with certain models validated to perform closer to the consensus of a group of experts than the average of the individual expert.<span><sup>18</sup></span></p><p>In an ideal world, we will not just be teaching trainees how to integrate AI-based EEG interpretation into practice, but we will be using AI to help us learn as well. In a study where a machine learning model was applied to EEG studies of critically ill patients of various etiologies, the convolutional neural network was able to identify novel EEG signatures to predict future clinical outcomes.<span><sup>34</sup></span> Visual analysis showed that the machine application learned EEG patterns typically recognized by human experts but also suggested new criteria that could serve as important electrographic biomarkers to improve care, demonstrating more potential novel applications for AI–human collaboration.</p><p>This is all to say that, although there are still important challenges we must address prior to a larger rollout, we advocate for academic programs to implement a curriculum that includes the operation, implementation, and oversight of AI in clinical practice. To create a roadmap for EEG and epilepsy education in the time of AI, we must simultaneously acknowledge that our current educational system would benefit from higher quality teaching, more consistent exposure to EEG during training, and standardized competencies that prepare our neurologists to perform well,<span><sup>10, 11, 35</sup></span> and also that better trained readers will help ensure the highest quality care in a future “hybrid” model of AI augmentation. We know from our colleagues in cardiology that inexperienced readers may not have the confidence to override the automated output of ECG studies<span><sup>36</sup></span>; therefore, to reduce false-positive and false-negative errors of AI interpreted EEG results, we will need to continue to produce well-trained, confident human readers. In the same vein, to ensure optimal AI output, we must acknowledge that the data fed into the algorithms will be pulled from human interpretation; therefore, a byproduct of high-quality EEG education should be an improved AI.</p><p>We propose that—once these AI-based algorithms are commercially available for EEG interpretation—all trainees should have meaningful exposure to this emerging technology. There should be clear parameters for where and when this technology can be used, including discussions on the ethical, practical, and equity issues surrounding implementation and the establishment of minimal acceptable standards. We may not need to teach our trainees or patients how a given algorithm works from a technical perspective, but all clinicians using this technology should be able to explain the clinical implication of these results to the patient; therefore, any educational curriculum must include modules on both proper interpretation and communication. For oversight, particularly as we continue to refine these technologies and train newer models, we will need dedicated teaching on the limitations of the technology and how humans best operate in this environment. Finally, continuing education on both the capabilities and limitations of AI will be essential for all neurologists not in active training programs to effectively incorporate these tools into their practice, for we believe AI should not be a tool available just for the technologically savvy, but for everyone in the community willing to learn how to safely use it. If we can implement these changes, rather than resist the sea change ahead, the future of epilepsy will be brighter than ever.</p><p>Dr. McLaren, Dr. Yuan, Dr. Beniczky, and Dr. Nascimento have no conflicts of interest. Dr. Westover has received support from grants from the National Institutes of Health (NIH; RF1AG064312, RF1NS120947, R01AG073410, R01HL161253, R01NS126282, R01AG073598, R01NS131347, and R01NS130119), and the National Science Foundation (NSF; 2014431). Dr. Westover is a co-founder and scientific advisor of, and consultant to Beacon Biosignals, and he has a personal equity interest in the company. He also receives royalties for authoring “Pocket Neurology” from Wolters Kluwer and “Atlas of Intensive Care Quantitative EEG” by Demos Medical.</p><p>We confirm that we have read the Journal's position on issues involved in ethical publication and affirm that this report is consistent with those guidelines.</p>","PeriodicalId":11768,"journal":{"name":"Epilepsia","volume":"66 6","pages":"1838-1842"},"PeriodicalIF":6.6000,"publicationDate":"2025-03-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1111/epi.18326","citationCount":"0","resultStr":"{\"title\":\"The future of EEG education in the era of artificial intelligence\",\"authors\":\"John R. McLaren,&nbsp;Doyle Yuan,&nbsp;Sándor Beniczky,&nbsp;M. Brandon Westover,&nbsp;Fábio A. Nascimento\",\"doi\":\"10.1111/epi.18326\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p>Artificial intelligence (AI), the technology that enables computers to simulate human problem-solving capabilities, is rapidly evolving. In the field of epilepsy, AI's application has already demonstrated potential for improved quality, cost, and access to patient care.<span><sup>1</sup></span> Although these advancements are exciting, they also pose several questions for the epilepsy community that are imperative to address now, before the imminent implementation of AI in clinical practice. As academics and educators, one question we are often asked is whether, or to what degree, electroencephalography (EEG) interpretation should be taught to the next generations of (human) neurologists. To address this in a meaningful way, we first examine our current predicament and then anticipate the road ahead. In doing so, we argue for a system that carefully integrates AI-based algorithms into the workflow of expert human EEG interpretation, which will ultimately necessitate a change in the way we educate our trainees and the next generation of electroencephalographers.</p><p>EEG is the tool most often used in the diagnostic evaluation of individuals with suspected epilepsy and frequently employed during sleep studies, surgeries, and in neurocritical care settings; therefore, accurate and reliable EEG interpretation is essential for optimal care of a variety of patients. In real-world practice, EEG misinterpretation does occur—either through over-calling normal EEG patterns as abnormal, or under-calling abnormal findings as benign, both of which can negatively impact patient care and outcomes. EEG “over-calling” (i.e., false-positive errors) can lead to epilepsy misdiagnosis with resultant unnecessary driving restrictions, employment difficulties, overprescription of anti-seizure medications, unwarranted surgery, inappropriate prognostication, other forms of comorbid stigma and social marginalization, as well as notable negative effects to health care systems. Meanwhile, “under-calling” (i.e., false-negative errors) can result in missed opportunities to prevent seizure-related injury, negative cognitive sequelae, and, in some circumstances, death.<span><sup>2-8</sup></span> Additional implications are well documented in patients without epilepsy but where EEG is still used.<span><sup>9</sup></span></p><p>Unfortunately, there are significant gaps in the current landscape of EEG education. For both adult and pediatric neurology training programs, there is limited exposure to EEG, subpar quality teaching, a paucity of objective competencies, and significant inter-program variability.<span><sup>10, 11</sup></span> As such, many neurology graduates leave training without feeling confident in their EEG-reading capabilities.<span><sup>12, 13</sup></span></p><p>This presents a substantial problem in clinical practice, as a large portion of EEG studies across the world—including the United States and many European countries—are interpreted by general neurologists without additional EEG/epilepsy training.<span><sup>14, 15</sup></span></p><p>With so much room for improvement, it is easy to see why AI-centered EEG interpretation holds promise. Several automated AI algorithms have shown high accuracy in identifying EEG patterns including interictal epileptiform discharges (IEDs), interictal slowing, seizures, and findings on the ictal-interictal continuum. Some of these algorithms have been validated in routine and critical care EEGs, achieving human expert level performance with improved efficiency and consistency.<span><sup>16-24</sup></span> As such, it is more than plausible that the future will involve appreciable adoption of these technologies. However, we need to consider the larger implications of implementation and have the foresight to address them in a meaningful way through proper education.</p><p>First, one must recognize that the high accuracy and efficiency of a tool does not always lead to equivalent levels of credibility. Human, or in this case, patient, perception is an important variable that changes an otherwise straightforward equation. The credibility of AI systems, regardless of their accuracy, remains generally quite low. In large global surveys on perceptions of this new technology, most people are wary about trusting AI and have a low or moderate acceptance of it.<span><sup>25</sup></span> Due to the inherent legal liabilities and ethical considerations<span><sup>26</sup></span> involved in misdiagnosis and mismanagement, which have not yet been fully elucidated, it is implausible that fully autonomous AI would be applied to the diagnostic or therapeutic process until the community at large takes time to thoughtfully address these issues (e.g., through the establishment of minimum standards).</p><p>Furthermore, the many limitations of AI must be acknowledged. Even the highest quality AI models are primarily data-driven algorithms that are trained and validated on finite, historical datasets with expert input. These datasets, if not properly designed, can perpetuate discriminatory practices, give rise to “hallucinations” (false predictions),<span><sup>27</sup></span> and create suboptimal behavioral change, all of which could impact patient care. As there is not yet a complete, automated EEG interpretation system available, we must carefully consider the datasets on which these AI algorithms were trained and validated to ensure that they are applied in the appropriate context. For example, SPaRCNET<span><sup>16</sup></span> was developed to read critical care EEG, whereas SCORE-AI<span><sup>18, 19</sup></span> was developed to read outpatient/routine EEG; hence, a one size fits all approach to interpretation would be imprudent without more rigorous investigation. We also must acknowledge that many seizure types have markedly different electrographic signals (e.g., neonatal seizures, epileptic spasms, atonic seizures, myoclonic seizures, and so on) for which AI algorithms have not been developed and will require trained human readers to recognize these patterns along with corresponding video of the events, if available, at least until more sophisticated models are developed. As such, the most plausible scenario, and the gold standard we should aim for, is a complementary “hybrid” model<span><sup>20</sup></span>—one in which augmentation with AI could allow experts to make higher quality decisions more efficiently than those not using the technology, while maintaining oversight and human credibility that will best serve our patients, particularly in cases where the algorithms fall short.</p><p>Over the past few years, examples of some hybrid approaches have been demonstrated that have the potential to improve interpretation and promote better, more-efficient reading skills. In a comparison between three fully automated AI algorithms and a human-supervised AI model applying an operational definition to detect IEDs, the specificity of the fully automated approaches was too low for clinical implementation, despite the high sensitivity. Meanwhile the human-supervised approach significantly increased the specificity, maintained good sensitivity and accuracy, and decreased the time burden of review compared to conventional visual analysis. In another study, a more “interpretable” deep learning model that accurately classifies six patterns of potentially harmful EEG activity was applied to EEG recordings in the intensive care unit (ICU) setting and led to significant pattern classification accuracy improvement by human readers using this assisted technology.<span><sup>28</sup></span> It is important to note that this performance was also significantly better than that of a corresponding uninterpretable “black-box” model, demonstrating further promise for AI–human collaboration but also highlighting the importance of design and context for the end user.</p><p>Additional applications that have not yet been described or validated should also make interpretation more efficient and education more focused. For example, applying appropriate AI algorithms with high sensitivities will allow trainees to focus less on the more time-consuming aspects of interpretation, such as scrolling through long, multiday continuous EEG files, and instead focus on interpreting snippets flagged as “concerning” by AI. Similarly, rapid-response EEG systems with high negative predictive value could be used to triage which EEG recordings require more detailed review overnight and which could wait until morning, freeing up considerable clinical burden on trainees. Better automated detection algorithms could be applied in ICU settings to reduce “alarm fatigue,” a longstanding issue that has been well documented to affect clinician concentration and diagnostic accuracy, leading to burnout and reduced quality of life.<span><sup>29-32</sup></span> In resource-limited settings, where there are either no epileptologists, or general neurologists without proper exposure to EEG reading during training, AI implementation could help mitigate these deficits. Although these algorithms can run locally on a tablet without internet connection, we also acknowledge that many low- and middle-income areas lack the important high-tech infrastructure for easy adoption,<span><sup>33</sup></span> such as the resources for machinery, personnel, quality control, and cloud storage, all of which can contribute to an increased carbon footprint and may limit feasibility of use.<span><sup>26</sup></span> Finally, in challenging or equivocal EEG cases, AI interpretation could offer a “second opinion” for readers, with certain models validated to perform closer to the consensus of a group of experts than the average of the individual expert.<span><sup>18</sup></span></p><p>In an ideal world, we will not just be teaching trainees how to integrate AI-based EEG interpretation into practice, but we will be using AI to help us learn as well. In a study where a machine learning model was applied to EEG studies of critically ill patients of various etiologies, the convolutional neural network was able to identify novel EEG signatures to predict future clinical outcomes.<span><sup>34</sup></span> Visual analysis showed that the machine application learned EEG patterns typically recognized by human experts but also suggested new criteria that could serve as important electrographic biomarkers to improve care, demonstrating more potential novel applications for AI–human collaboration.</p><p>This is all to say that, although there are still important challenges we must address prior to a larger rollout, we advocate for academic programs to implement a curriculum that includes the operation, implementation, and oversight of AI in clinical practice. To create a roadmap for EEG and epilepsy education in the time of AI, we must simultaneously acknowledge that our current educational system would benefit from higher quality teaching, more consistent exposure to EEG during training, and standardized competencies that prepare our neurologists to perform well,<span><sup>10, 11, 35</sup></span> and also that better trained readers will help ensure the highest quality care in a future “hybrid” model of AI augmentation. We know from our colleagues in cardiology that inexperienced readers may not have the confidence to override the automated output of ECG studies<span><sup>36</sup></span>; therefore, to reduce false-positive and false-negative errors of AI interpreted EEG results, we will need to continue to produce well-trained, confident human readers. In the same vein, to ensure optimal AI output, we must acknowledge that the data fed into the algorithms will be pulled from human interpretation; therefore, a byproduct of high-quality EEG education should be an improved AI.</p><p>We propose that—once these AI-based algorithms are commercially available for EEG interpretation—all trainees should have meaningful exposure to this emerging technology. There should be clear parameters for where and when this technology can be used, including discussions on the ethical, practical, and equity issues surrounding implementation and the establishment of minimal acceptable standards. We may not need to teach our trainees or patients how a given algorithm works from a technical perspective, but all clinicians using this technology should be able to explain the clinical implication of these results to the patient; therefore, any educational curriculum must include modules on both proper interpretation and communication. For oversight, particularly as we continue to refine these technologies and train newer models, we will need dedicated teaching on the limitations of the technology and how humans best operate in this environment. Finally, continuing education on both the capabilities and limitations of AI will be essential for all neurologists not in active training programs to effectively incorporate these tools into their practice, for we believe AI should not be a tool available just for the technologically savvy, but for everyone in the community willing to learn how to safely use it. If we can implement these changes, rather than resist the sea change ahead, the future of epilepsy will be brighter than ever.</p><p>Dr. McLaren, Dr. Yuan, Dr. Beniczky, and Dr. Nascimento have no conflicts of interest. Dr. Westover has received support from grants from the National Institutes of Health (NIH; RF1AG064312, RF1NS120947, R01AG073410, R01HL161253, R01NS126282, R01AG073598, R01NS131347, and R01NS130119), and the National Science Foundation (NSF; 2014431). Dr. Westover is a co-founder and scientific advisor of, and consultant to Beacon Biosignals, and he has a personal equity interest in the company. He also receives royalties for authoring “Pocket Neurology” from Wolters Kluwer and “Atlas of Intensive Care Quantitative EEG” by Demos Medical.</p><p>We confirm that we have read the Journal's position on issues involved in ethical publication and affirm that this report is consistent with those guidelines.</p>\",\"PeriodicalId\":11768,\"journal\":{\"name\":\"Epilepsia\",\"volume\":\"66 6\",\"pages\":\"1838-1842\"},\"PeriodicalIF\":6.6000,\"publicationDate\":\"2025-03-04\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://onlinelibrary.wiley.com/doi/epdf/10.1111/epi.18326\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Epilepsia\",\"FirstCategoryId\":\"3\",\"ListUrlMain\":\"https://onlinelibrary.wiley.com/doi/10.1111/epi.18326\",\"RegionNum\":1,\"RegionCategory\":\"医学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"CLINICAL NEUROLOGY\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Epilepsia","FirstCategoryId":"3","ListUrlMain":"https://onlinelibrary.wiley.com/doi/10.1111/epi.18326","RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"CLINICAL NEUROLOGY","Score":null,"Total":0}
引用次数: 0

摘要

人工智能(AI)是一种使计算机能够模拟人类解决问题能力的技术,它正在迅速发展。在癫痫领域,人工智能的应用已经显示出改善质量、成本和患者护理可及性的潜力尽管这些进展令人兴奋,但它们也为癫痫界提出了几个问题,在即将在临床实践中实施人工智能之前,现在必须解决这些问题。作为学者和教育工作者,我们经常被问到的一个问题是,是否应该向下一代(人类)神经学家教授脑电图(EEG)解释,或者应该在多大程度上教授。为了以一种有意义的方式解决这个问题,我们首先审视我们目前的困境,然后预测未来的道路。在此过程中,我们主张建立一个系统,将基于人工智能的算法仔细整合到专家人类脑电图解释的工作流程中,这最终将需要改变我们教育学员和下一代脑电图学家的方式。EEG是最常用于疑似癫痫患者的诊断评估的工具,在睡眠研究、手术和神经危重症护理环境中经常使用;因此,准确可靠的脑电图解释对于各种患者的最佳护理至关重要。在现实世界的实践中,脑电图的误读确实会发生——要么将正常的脑电图模式过度称为异常,要么将异常结果低估为良性,这两种情况都会对患者的护理和结果产生负面影响。脑电图“过度诊断”(即假阳性错误)可导致癫痫误诊,从而导致不必要的驾驶限制、就业困难、抗癫痫药物的过度处方、无根据的手术、不适当的预测、其他形式的共病污名和社会边缘化,以及对卫生保健系统的显著负面影响。同时,“呼叫不足”(即假阴性错误)可能导致错过预防癫痫相关损伤的机会、负面认知后遗症,并在某些情况下导致死亡。2-8在没有癫痫但仍在使用脑电图的患者中有详细记录的其他影响。不幸的是,目前脑电图教育领域存在着显著的差距。对于成人和儿童神经学培训项目,脑电图暴露有限,教学质量欠佳,缺乏客观能力,项目间差异显著。10,11因此,许多神经学毕业生在培训结束后对自己的脑电图阅读能力没有信心。12,13这在临床实践中提出了一个实质性的问题,因为世界上大部分的脑电图研究——包括美国和许多欧洲国家——都是由普通神经科医生在没有额外的脑电图/癫痫训练的情况下进行的。14,15由于有如此多的改进空间,很容易看出为什么以人工智能为中心的脑电图解释有希望。一些自动化人工智能算法在识别脑电图模式方面显示出很高的准确性,包括间歇癫痫样放电(ied)、间歇减慢、癫痫发作和间歇-间歇连续体的发现。其中一些算法已在常规和重症监护脑电图中得到验证,在提高效率和一致性的情况下达到了人类专家水平的表现。16-24因此,这些技术在未来被大量采用是非常可能的。然而,我们需要考虑实施的更大影响,并有远见,通过适当的教育以有意义的方式解决这些问题。首先,人们必须认识到,一个工具的高准确性和高效率并不总是导致同等水平的可信度。人类,或者在这个例子中,病人,感知是一个重要的变量,它改变了一个简单的等式。人工智能系统的可信度,无论其准确性如何,总体上仍然很低。在对这项新技术看法的大型全球调查中,大多数人对信任人工智能持谨慎态度,对它的接受程度很低或中等由于尚未完全阐明的误诊和管理不善所涉及的固有法律责任和伦理考虑,在整个社会花时间仔细解决这些问题(例如,通过建立最低标准)之前,将完全自主的人工智能应用于诊断或治疗过程是不可能的。此外,必须承认人工智能的许多局限性。即使是最高质量的人工智能模型也主要是数据驱动的算法,这些算法是在有限的历史数据集和专家输入上进行训练和验证的。这些数据集如果设计不当,可能会延续歧视性做法,产生“幻觉”(错误的预测),27并造成次优行为改变,所有这些都可能影响患者护理。 由于目前还没有一个完整的自动脑电图解释系统,我们必须仔细考虑这些人工智能算法训练和验证的数据集,以确保它们在适当的环境中应用。例如,SPaRCNET16被开发用于读取重症监护EEG,而score - ai18,19被开发用于读取门诊/常规EEG;因此,如果没有更严格的调查,一刀切的解释方法将是轻率的。我们还必须承认,许多癫痫类型具有明显不同的电信号(例如,新生儿癫痫发作,癫痫痉挛,无张力癫痫发作,肌阵挛性癫痫发作等),人工智能算法尚未开发出来,并且需要训练有素的人类读者识别这些模式以及相应的事件视频,如果有的话,至少要等到更复杂的模型开发出来。因此,最合理的情况,也是我们应该追求的黄金标准,是一种互补的“混合”模型,在这种模型中,人工智能的增强可以让专家比那些不使用这项技术的人更有效地做出更高质量的决策,同时保持监督和人类的可信度,这将最好地为我们的病人服务,尤其是在算法不足的情况下。在过去的几年里,一些混合方法的例子已经被证明有可能改善口译和促进更好、更有效的阅读技能。在三种全自动人工智能算法和应用操作定义检测简易爆炸装置的人工监督人工智能模型之间的比较中,全自动方法的特异性太低,无法用于临床实施,尽管灵敏度很高。与传统的视觉分析相比,人工监督方法显著提高了特异性,保持了良好的灵敏度和准确性,减少了审查的时间负担。在另一项研究中,一种更“可解释”的深度学习模型被应用于重症监护病房(ICU)的脑电图记录,该模型可以准确分类六种潜在有害的脑电图活动模式,并通过人类读者使用这种辅助技术显著提高了模式分类的准确性值得注意的是,这种表现也明显优于相应的不可解释的“黑盒”模型,这表明了人工智能与人类合作的进一步前景,但也强调了设计和上下文对最终用户的重要性。尚未描述或验证的其他应用程序也应该使解释更有效,教育更集中。例如,应用适当的高灵敏度人工智能算法将使受训者减少对更耗时的解释方面的关注,例如滚动浏览长时间,连续多日的脑电图文件,而是专注于解释被人工智能标记为“关注”的片段。同样,具有高阴性预测值的快速反应脑电图系统可以用于分类哪些脑电图记录需要在夜间进行更详细的检查,哪些可以等到早上,从而减轻了实习生的相当大的临床负担。更好的自动检测算法可以应用于ICU环境,以减少“警报疲劳”,这是一个长期存在的问题,已经有充分的证据表明,它会影响临床医生的注意力和诊断准确性,导致倦怠和生活质量下降。29-32在资源有限的环境中,既没有癫痫学家,也没有在训练期间适当接触脑电图读数的普通神经病学家,人工智能的实施可以帮助减轻这些缺陷。虽然这些算法可以在没有互联网连接的平板电脑上本地运行,但我们也承认,许多低收入和中等收入地区缺乏易于采用的重要高科技基础设施33,例如机器、人员、质量控制和云存储资源,所有这些都可能导致碳足迹增加,并可能限制使用的可行性26最后,在具有挑战性或模棱两可的脑电图病例中,人工智能解释可以为读者提供“第二意见”,某些模型经过验证,比单个专家的平均水平更接近一组专家的共识。在一个理想的世界里,我们不仅会教学员如何将基于人工智能的脑电图解释融入实践,而且还会使用人工智能来帮助我们学习。在一项研究中,机器学习模型被应用于各种病因的危重病人的脑电图研究,卷积神经网络能够识别新的脑电图特征来预测未来的临床结果。 视觉分析表明,机器应用程序学习了通常由人类专家识别的脑电图模式,但也提出了新的标准,可以作为重要的电图生物标志物来改善护理,展示了人工智能-人类合作的更多潜在新应用。这就是说,尽管在更大规模的推广之前,我们仍然必须解决一些重要的挑战,但我们主张学术项目实施一门课程,包括人工智能在临床实践中的操作、实施和监督。为了在人工智能时代为脑电图和癫痫教育制定路线图,我们必须同时承认,我们目前的教育系统将受益于更高质量的教学,在培训期间更一致地接触脑电图,以及使我们的神经科医生做好准备的标准化能力,10,11,35,而且训练有素的读者将有助于确保在未来人工智能增强的“混合”模型中提供最高质量的护理。我们从心脏病学的同事那里了解到,没有经验的读者可能没有信心推翻心电图研究的自动输出36;因此,为了减少人工智能解释的脑电图结果的假阳性和假阴性错误,我们需要继续培养训练有素、自信的人类读者。同样,为了确保最佳的人工智能输出,我们必须承认,输入算法的数据将来自人类的解释;因此,高质量脑电图教育的副产品应该是改进的人工智能。我们建议,一旦这些基于人工智能的算法可用于EEG解释,所有学员都应该有意义地接触到这种新兴技术。应该对何时何地可以使用这项技术有明确的参数,包括讨论围绕实施和建立最低可接受标准的道德、实际和公平问题。我们可能不需要从技术角度教我们的学员或患者如何使用给定的算法,但所有使用这项技术的临床医生都应该能够向患者解释这些结果的临床意义;因此,任何教育课程都必须包括适当的口译和沟通模块。在监督方面,特别是在我们继续完善这些技术和培训新模型的过程中,我们需要专门教授技术的局限性,以及人类如何在这种环境中最好地运作。最后,关于人工智能的能力和局限性的继续教育对于所有没有积极培训计划的神经科医生有效地将这些工具纳入他们的实践是必不可少的,因为我们相信人工智能不应该只是技术精通的工具,而是社区中愿意学习如何安全使用它的每个人。如果我们能够实施这些变化,而不是抗拒前方的巨大变化,癫痫的未来将比以往任何时候都更加光明。麦克拉伦、袁医生、贝尼茨基医生和纳西门托医生没有利益冲突。Westover博士获得了美国国立卫生研究院(NIH;RF1AG064312、RF1NS120947、R01AG073410、R01HL161253、R01NS126282、R01AG073598、R01NS131347、R01NS130119),国家科学基金(NSF;2014431)。Westover博士是Beacon Biosignals的联合创始人、科学顾问和顾问,他拥有该公司的个人股权。他还因Wolters Kluwer出版的《口袋神经学》和Demos Medical出版的《重症监护定量脑电图图谱》而获得版税。我们确认,我们已经阅读了《华尔街日报》关于出版伦理问题的立场,并确认本报告符合这些准则。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
The future of EEG education in the era of artificial intelligence

Artificial intelligence (AI), the technology that enables computers to simulate human problem-solving capabilities, is rapidly evolving. In the field of epilepsy, AI's application has already demonstrated potential for improved quality, cost, and access to patient care.1 Although these advancements are exciting, they also pose several questions for the epilepsy community that are imperative to address now, before the imminent implementation of AI in clinical practice. As academics and educators, one question we are often asked is whether, or to what degree, electroencephalography (EEG) interpretation should be taught to the next generations of (human) neurologists. To address this in a meaningful way, we first examine our current predicament and then anticipate the road ahead. In doing so, we argue for a system that carefully integrates AI-based algorithms into the workflow of expert human EEG interpretation, which will ultimately necessitate a change in the way we educate our trainees and the next generation of electroencephalographers.

EEG is the tool most often used in the diagnostic evaluation of individuals with suspected epilepsy and frequently employed during sleep studies, surgeries, and in neurocritical care settings; therefore, accurate and reliable EEG interpretation is essential for optimal care of a variety of patients. In real-world practice, EEG misinterpretation does occur—either through over-calling normal EEG patterns as abnormal, or under-calling abnormal findings as benign, both of which can negatively impact patient care and outcomes. EEG “over-calling” (i.e., false-positive errors) can lead to epilepsy misdiagnosis with resultant unnecessary driving restrictions, employment difficulties, overprescription of anti-seizure medications, unwarranted surgery, inappropriate prognostication, other forms of comorbid stigma and social marginalization, as well as notable negative effects to health care systems. Meanwhile, “under-calling” (i.e., false-negative errors) can result in missed opportunities to prevent seizure-related injury, negative cognitive sequelae, and, in some circumstances, death.2-8 Additional implications are well documented in patients without epilepsy but where EEG is still used.9

Unfortunately, there are significant gaps in the current landscape of EEG education. For both adult and pediatric neurology training programs, there is limited exposure to EEG, subpar quality teaching, a paucity of objective competencies, and significant inter-program variability.10, 11 As such, many neurology graduates leave training without feeling confident in their EEG-reading capabilities.12, 13

This presents a substantial problem in clinical practice, as a large portion of EEG studies across the world—including the United States and many European countries—are interpreted by general neurologists without additional EEG/epilepsy training.14, 15

With so much room for improvement, it is easy to see why AI-centered EEG interpretation holds promise. Several automated AI algorithms have shown high accuracy in identifying EEG patterns including interictal epileptiform discharges (IEDs), interictal slowing, seizures, and findings on the ictal-interictal continuum. Some of these algorithms have been validated in routine and critical care EEGs, achieving human expert level performance with improved efficiency and consistency.16-24 As such, it is more than plausible that the future will involve appreciable adoption of these technologies. However, we need to consider the larger implications of implementation and have the foresight to address them in a meaningful way through proper education.

First, one must recognize that the high accuracy and efficiency of a tool does not always lead to equivalent levels of credibility. Human, or in this case, patient, perception is an important variable that changes an otherwise straightforward equation. The credibility of AI systems, regardless of their accuracy, remains generally quite low. In large global surveys on perceptions of this new technology, most people are wary about trusting AI and have a low or moderate acceptance of it.25 Due to the inherent legal liabilities and ethical considerations26 involved in misdiagnosis and mismanagement, which have not yet been fully elucidated, it is implausible that fully autonomous AI would be applied to the diagnostic or therapeutic process until the community at large takes time to thoughtfully address these issues (e.g., through the establishment of minimum standards).

Furthermore, the many limitations of AI must be acknowledged. Even the highest quality AI models are primarily data-driven algorithms that are trained and validated on finite, historical datasets with expert input. These datasets, if not properly designed, can perpetuate discriminatory practices, give rise to “hallucinations” (false predictions),27 and create suboptimal behavioral change, all of which could impact patient care. As there is not yet a complete, automated EEG interpretation system available, we must carefully consider the datasets on which these AI algorithms were trained and validated to ensure that they are applied in the appropriate context. For example, SPaRCNET16 was developed to read critical care EEG, whereas SCORE-AI18, 19 was developed to read outpatient/routine EEG; hence, a one size fits all approach to interpretation would be imprudent without more rigorous investigation. We also must acknowledge that many seizure types have markedly different electrographic signals (e.g., neonatal seizures, epileptic spasms, atonic seizures, myoclonic seizures, and so on) for which AI algorithms have not been developed and will require trained human readers to recognize these patterns along with corresponding video of the events, if available, at least until more sophisticated models are developed. As such, the most plausible scenario, and the gold standard we should aim for, is a complementary “hybrid” model20—one in which augmentation with AI could allow experts to make higher quality decisions more efficiently than those not using the technology, while maintaining oversight and human credibility that will best serve our patients, particularly in cases where the algorithms fall short.

Over the past few years, examples of some hybrid approaches have been demonstrated that have the potential to improve interpretation and promote better, more-efficient reading skills. In a comparison between three fully automated AI algorithms and a human-supervised AI model applying an operational definition to detect IEDs, the specificity of the fully automated approaches was too low for clinical implementation, despite the high sensitivity. Meanwhile the human-supervised approach significantly increased the specificity, maintained good sensitivity and accuracy, and decreased the time burden of review compared to conventional visual analysis. In another study, a more “interpretable” deep learning model that accurately classifies six patterns of potentially harmful EEG activity was applied to EEG recordings in the intensive care unit (ICU) setting and led to significant pattern classification accuracy improvement by human readers using this assisted technology.28 It is important to note that this performance was also significantly better than that of a corresponding uninterpretable “black-box” model, demonstrating further promise for AI–human collaboration but also highlighting the importance of design and context for the end user.

Additional applications that have not yet been described or validated should also make interpretation more efficient and education more focused. For example, applying appropriate AI algorithms with high sensitivities will allow trainees to focus less on the more time-consuming aspects of interpretation, such as scrolling through long, multiday continuous EEG files, and instead focus on interpreting snippets flagged as “concerning” by AI. Similarly, rapid-response EEG systems with high negative predictive value could be used to triage which EEG recordings require more detailed review overnight and which could wait until morning, freeing up considerable clinical burden on trainees. Better automated detection algorithms could be applied in ICU settings to reduce “alarm fatigue,” a longstanding issue that has been well documented to affect clinician concentration and diagnostic accuracy, leading to burnout and reduced quality of life.29-32 In resource-limited settings, where there are either no epileptologists, or general neurologists without proper exposure to EEG reading during training, AI implementation could help mitigate these deficits. Although these algorithms can run locally on a tablet without internet connection, we also acknowledge that many low- and middle-income areas lack the important high-tech infrastructure for easy adoption,33 such as the resources for machinery, personnel, quality control, and cloud storage, all of which can contribute to an increased carbon footprint and may limit feasibility of use.26 Finally, in challenging or equivocal EEG cases, AI interpretation could offer a “second opinion” for readers, with certain models validated to perform closer to the consensus of a group of experts than the average of the individual expert.18

In an ideal world, we will not just be teaching trainees how to integrate AI-based EEG interpretation into practice, but we will be using AI to help us learn as well. In a study where a machine learning model was applied to EEG studies of critically ill patients of various etiologies, the convolutional neural network was able to identify novel EEG signatures to predict future clinical outcomes.34 Visual analysis showed that the machine application learned EEG patterns typically recognized by human experts but also suggested new criteria that could serve as important electrographic biomarkers to improve care, demonstrating more potential novel applications for AI–human collaboration.

This is all to say that, although there are still important challenges we must address prior to a larger rollout, we advocate for academic programs to implement a curriculum that includes the operation, implementation, and oversight of AI in clinical practice. To create a roadmap for EEG and epilepsy education in the time of AI, we must simultaneously acknowledge that our current educational system would benefit from higher quality teaching, more consistent exposure to EEG during training, and standardized competencies that prepare our neurologists to perform well,10, 11, 35 and also that better trained readers will help ensure the highest quality care in a future “hybrid” model of AI augmentation. We know from our colleagues in cardiology that inexperienced readers may not have the confidence to override the automated output of ECG studies36; therefore, to reduce false-positive and false-negative errors of AI interpreted EEG results, we will need to continue to produce well-trained, confident human readers. In the same vein, to ensure optimal AI output, we must acknowledge that the data fed into the algorithms will be pulled from human interpretation; therefore, a byproduct of high-quality EEG education should be an improved AI.

We propose that—once these AI-based algorithms are commercially available for EEG interpretation—all trainees should have meaningful exposure to this emerging technology. There should be clear parameters for where and when this technology can be used, including discussions on the ethical, practical, and equity issues surrounding implementation and the establishment of minimal acceptable standards. We may not need to teach our trainees or patients how a given algorithm works from a technical perspective, but all clinicians using this technology should be able to explain the clinical implication of these results to the patient; therefore, any educational curriculum must include modules on both proper interpretation and communication. For oversight, particularly as we continue to refine these technologies and train newer models, we will need dedicated teaching on the limitations of the technology and how humans best operate in this environment. Finally, continuing education on both the capabilities and limitations of AI will be essential for all neurologists not in active training programs to effectively incorporate these tools into their practice, for we believe AI should not be a tool available just for the technologically savvy, but for everyone in the community willing to learn how to safely use it. If we can implement these changes, rather than resist the sea change ahead, the future of epilepsy will be brighter than ever.

Dr. McLaren, Dr. Yuan, Dr. Beniczky, and Dr. Nascimento have no conflicts of interest. Dr. Westover has received support from grants from the National Institutes of Health (NIH; RF1AG064312, RF1NS120947, R01AG073410, R01HL161253, R01NS126282, R01AG073598, R01NS131347, and R01NS130119), and the National Science Foundation (NSF; 2014431). Dr. Westover is a co-founder and scientific advisor of, and consultant to Beacon Biosignals, and he has a personal equity interest in the company. He also receives royalties for authoring “Pocket Neurology” from Wolters Kluwer and “Atlas of Intensive Care Quantitative EEG” by Demos Medical.

We confirm that we have read the Journal's position on issues involved in ethical publication and affirm that this report is consistent with those guidelines.

求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
Epilepsia
Epilepsia 医学-临床神经学
CiteScore
10.90
自引率
10.70%
发文量
319
审稿时长
2-4 weeks
期刊介绍: Epilepsia is the leading, authoritative source for innovative clinical and basic science research for all aspects of epilepsy and seizures. In addition, Epilepsia publishes critical reviews, opinion pieces, and guidelines that foster understanding and aim to improve the diagnosis and treatment of people with seizures and epilepsy.
期刊最新文献
Mental health of children with epilepsy in Ukraine during the war. An n-of-1 gene-directed drug repurposing trial for an ultrarare genetic condition. Rehabilitation of cognition and psychosocial well-being in epilepsy: Results of a randomized waiting list-controlled trial. Dual role of spreading depolarization in an epileptic focus. Spatiotemporal dynamics of seizure networks: Progressive posterior hippocampal recruitment in mesial temporal lobe epilepsy.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1