首页 > 最新文献

Proceedings. IEEE International Symposium on Computer-Based Medical Systems最新文献

英文 中文
Automated Deep Learning Approach for Post-Operative Neonatal Pain Detection and Prediction through Physiological Signals. 基于生理信号的新生儿术后疼痛检测与预测的自动深度学习方法。
Pub Date : 2025-06-01 Epub Date: 2025-07-04 DOI: 10.1109/cbms65348.2025.00164
Jacqueline Hausmann, Jiayi Wang, Marcia Kneusel, Stephanie Prescott, Peter R Mouton, Yu Sun, Dmitry Goldgof

It is well-known that severe pain and powerful pain medications cause short- and long-term damage to the developing nervous system of newborns. Caregivers routinely use physiological vital signs [Heart Rate (HR), Respiration Rate (RR), Oxygen Saturation (SR)] to monitor post-surgical pain in the Neonatal Intensive Care Unit (NICU). Here we present a novel approach that combines continuous, non-invasive monitoring of these vital signs and Computer Vision/Deep Learning to make automatic neonate pain detection with an accuracy of 74% AUC, 67.59% mAP. Further, we report for the first time our Early Pain Detection (EPD) approach that explores prediction of the time to onset of post-surgical pain in neonates. Our EPD can alert NICU workers to postoperative neonatal pain about 5 to 10 minutes prior to pain onset. In addition to alleviating the need for intermittent pain assessments by busy NICU nurses via long-term observation, our EPD approach creates a time window prior to pain onset for the use of less harmful pain mitigation strategies. Through effective pain mitigation prior to spinal sensitization, EPD could minimize or eliminate severe post-surgical pain and the consequential need for powerful analgesics in post-surgical neonates.

众所周知,剧烈疼痛和强效止痛药会对新生儿发育中的神经系统造成短期和长期的损害。在新生儿重症监护病房(NICU),照例使用生理生命体征[心率(HR)、呼吸频率(RR)、血氧饱和度(SR)]监测术后疼痛。在这里,我们提出了一种新的方法,结合了对这些生命体征的连续、无创监测和计算机视觉/深度学习,使新生儿疼痛自动检测具有74% AUC和67.59% mAP的准确性。此外,我们首次报道了我们的早期疼痛检测(EPD)方法,该方法探索了新生儿术后疼痛发作时间的预测。我们的EPD可以在新生儿疼痛发作前5到10分钟提醒新生儿新生儿术后疼痛。除了通过长期观察减轻繁忙的NICU护士对间歇性疼痛评估的需求外,我们的EPD方法在疼痛发作之前创建了一个时间窗口,用于使用危害较小的疼痛缓解策略。通过脊髓致敏前的有效疼痛缓解,EPD可以减少或消除术后新生儿严重的术后疼痛和对强效镇痛药的需求。
{"title":"Automated Deep Learning Approach for Post-Operative Neonatal Pain Detection and Prediction through Physiological Signals.","authors":"Jacqueline Hausmann, Jiayi Wang, Marcia Kneusel, Stephanie Prescott, Peter R Mouton, Yu Sun, Dmitry Goldgof","doi":"10.1109/cbms65348.2025.00164","DOIUrl":"10.1109/cbms65348.2025.00164","url":null,"abstract":"<p><p>It is well-known that severe pain and powerful pain medications cause short- and long-term damage to the developing nervous system of newborns. Caregivers routinely use physiological vital signs [Heart Rate (HR), Respiration Rate (RR), Oxygen Saturation (SR)] to monitor post-surgical pain in the Neonatal Intensive Care Unit (NICU). Here we present a novel approach that combines continuous, non-invasive monitoring of these vital signs and Computer Vision/Deep Learning to make automatic neonate pain detection with an accuracy of 74% AUC, 67.59% mAP. Further, we report for the first time our Early Pain Detection (EPD) approach that explores prediction of the time to onset of post-surgical pain in neonates. Our EPD can alert NICU workers to postoperative neonatal pain about 5 to 10 minutes prior to pain onset. In addition to alleviating the need for intermittent pain assessments by busy NICU nurses via long-term observation, our EPD approach creates a time window prior to pain onset for the use of less harmful pain mitigation strategies. Through effective pain mitigation prior to spinal sensitization, EPD could minimize or eliminate severe post-surgical pain and the consequential need for powerful analgesics in post-surgical neonates.</p>","PeriodicalId":74567,"journal":{"name":"Proceedings. IEEE International Symposium on Computer-Based Medical Systems","volume":"2025 ","pages":"801-806"},"PeriodicalIF":0.0,"publicationDate":"2025-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12444759/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145115183","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Few-Shot Prompting with Vision Language Model for Pain Classification in Infant Cry Sounds. 基于视觉语言模型的短时提示婴儿哭声疼痛分类。
Pub Date : 2025-06-01 Epub Date: 2025-07-04 DOI: 10.1109/cbms65348.2025.00174
Anthony McCofie, Abhiram Kandiyana, Peter R Mouton, Yu Sun, Dmitry Goldgof

Accurately detecting pain in infants remains a complex challenge. Conventional deep neural networks used for analyzing infant cry sounds typically demand large labeled datasets, substantial computational power, and often lack interpretability. In this work, we introduce a novel approach that leverages OpenAI's vision-language model, GPT-4(V), combined with mel spectrogram-based representations of infant cries through prompting. This prompting strategy significantly reduces the dependence on large training datasets while enhancing transparency and interpretability. Using the USF-MNPAD-II dataset, our method achieves an accuracy of 83.33% with only 16 training samples, in contrast to the 4,914 samples required in the baseline model. To our knowledge, this represents the first application of few-shot prompting with vision-language models such as GPT-4o for infant pain classification.

准确地检测婴儿的疼痛仍然是一个复杂的挑战。用于分析婴儿哭声的传统深度神经网络通常需要大量的标记数据集,大量的计算能力,并且往往缺乏可解释性。在这项工作中,我们引入了一种新的方法,利用OpenAI的视觉语言模型GPT-4(V),结合基于mel谱图的婴儿哭声提示表示。这种提示策略大大减少了对大型训练数据集的依赖,同时提高了透明度和可解释性。使用USF-MNPAD-II数据集,我们的方法仅使用16个训练样本就实现了83.33%的准确率,而基线模型需要4,914个样本。据我们所知,这代表了第一次使用视觉语言模型(如gpt - 40)进行婴儿疼痛分类的几次提示。
{"title":"Few-Shot Prompting with Vision Language Model for Pain Classification in Infant Cry Sounds.","authors":"Anthony McCofie, Abhiram Kandiyana, Peter R Mouton, Yu Sun, Dmitry Goldgof","doi":"10.1109/cbms65348.2025.00174","DOIUrl":"10.1109/cbms65348.2025.00174","url":null,"abstract":"<p><p>Accurately detecting pain in infants remains a complex challenge. Conventional deep neural networks used for analyzing infant cry sounds typically demand large labeled datasets, substantial computational power, and often lack interpretability. In this work, we introduce a novel approach that leverages OpenAI's vision-language model, GPT-4(V), combined with mel spectrogram-based representations of infant cries through prompting. This prompting strategy significantly reduces the dependence on large training datasets while enhancing transparency and interpretability. Using the USF-MNPAD-II dataset, our method achieves an accuracy of 83.33% with only 16 training samples, in contrast to the 4,914 samples required in the baseline model. To our knowledge, this represents the first application of few-shot prompting with vision-language models such as GPT-4o for infant pain classification.</p>","PeriodicalId":74567,"journal":{"name":"Proceedings. IEEE International Symposium on Computer-Based Medical Systems","volume":"2025 ","pages":"857-862"},"PeriodicalIF":0.0,"publicationDate":"2025-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12444757/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145115158","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Artificial Intelligence Assurance in Head and Neck Surgery: Now and Next. 头颈外科的人工智能保障:现在和未来。
Pub Date : 2025-06-01 Epub Date: 2025-07-04 DOI: 10.1109/cbms65348.2025.00195
Yuansan Liu, Sudanthi Wijewickrema, Bridget Copson, Jean-Marc Gerard, Sameer Antani

Artificial intelligence (AI) is making significant advances toward becoming a well-established and promise-bearing technology in various medical domains such as screening, diagnostics, and biopharma research. However, its state remains relatively nascent in surgery and surgical therapeutics. This presents an opportunity for leveraging ongoing rapid advances in AI technology and the increasing availability of large, diverse datasets to pave the way for their use in these domains. Expanding the use of AI to include various processes in surgery-related workflows could provide several benefits, such as greater assurance for reduced errors, better assistance to surgeons, and overall improved patient outcomes. To encourage further research in surgical AI, this article summarizes the state-of-the-art in AI assurance in various aspects of a patient's timeline when undergoing head and neck surgeries, including diagnostics, preoperative considerations, intraoperative guidance, and postoperative and outcome predictions. The work aims to highlight gaps in the state-of-the-art and identify opportunities for the computer-based medical systems community to encourage future research and development on the subject.

人工智能(AI)正在取得重大进展,在筛查、诊断和生物制药研究等各个医疗领域成为一项成熟且有前景的技术。然而,它在外科和外科治疗中仍处于相对初级的状态。这为利用人工智能技术的持续快速发展和大型多样化数据集的日益可用性提供了机会,为它们在这些领域的使用铺平了道路。将人工智能的使用扩大到包括手术相关工作流程中的各种流程可以提供几个好处,例如更大程度地保证减少错误,更好地帮助外科医生,并整体改善患者的治疗效果。为了鼓励外科人工智能的进一步研究,本文总结了人工智能在患者接受头颈部手术时各个方面的最新保证,包括诊断、术前考虑、术中指导、术后和结果预测。这项工作旨在突出最先进的差距,并为基于计算机的医疗系统界确定机会,以鼓励未来在这一主题上的研究和开发。
{"title":"Artificial Intelligence Assurance in Head and Neck Surgery: Now and Next.","authors":"Yuansan Liu, Sudanthi Wijewickrema, Bridget Copson, Jean-Marc Gerard, Sameer Antani","doi":"10.1109/cbms65348.2025.00195","DOIUrl":"https://doi.org/10.1109/cbms65348.2025.00195","url":null,"abstract":"<p><p>Artificial intelligence (AI) is making significant advances toward becoming a well-established and promise-bearing technology in various medical domains such as screening, diagnostics, and biopharma research. However, its state remains relatively nascent in surgery and surgical therapeutics. This presents an opportunity for leveraging ongoing rapid advances in AI technology and the increasing availability of large, diverse datasets to pave the way for their use in these domains. Expanding the use of AI to include various processes in surgery-related workflows could provide several benefits, such as greater assurance for reduced errors, better assistance to surgeons, and overall improved patient outcomes. To encourage further research in surgical AI, this article summarizes the state-of-the-art in AI assurance in various aspects of a patient's timeline when undergoing head and neck surgeries, including diagnostics, preoperative considerations, intraoperative guidance, and postoperative and outcome predictions. The work aims to highlight gaps in the state-of-the-art and identify opportunities for the computer-based medical systems community to encourage future research and development on the subject.</p>","PeriodicalId":74567,"journal":{"name":"Proceedings. IEEE International Symposium on Computer-Based Medical Systems","volume":"2025 ","pages":"977-982"},"PeriodicalIF":0.0,"publicationDate":"2025-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12369650/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144981925","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
The Hidden Threat of Hallucinations in Binary Chest X-ray Pneumonia Classification. 肺炎胸片二元分型中幻觉的潜在威胁。
Pub Date : 2025-06-01 Epub Date: 2025-07-04 DOI: 10.1109/cbms65348.2025.00138
Sivaramakrishnan Rajaraman, Zhaohui Liang, Niccolo Marini, Zhiyun Xue, Sameer Antani

Hallucination in deep learning (DL) classification, where DL models yield confidently erroneous predictions remains a pressing concern. This study investigates whether binary classifiers are truly learning disease-specific features when distinguishing overlapping radiological presentations among pneumonia subtypes on chest X-ray (CXR) images. Specifically, we evaluate if uncertainty measure is a valuable tool in classifying signs of different pathogen-specific subtypes of pneumonia. We evaluated two binary classifiers to classify bacterial pneumonia and viral pneumonia, respectively, from normal CXRs. A third classifier explored the ability to distinguish bacterial from viral pneumonia presentation to highlight our concern regarding the observed hallucinations in the former cases. Our comprehensive analysis computes the Matthews Correlation Coefficient and prediction entropy metrics on a pediatric CXR dataset and reveals that the normal/bacterial and normal/viral classifiers consistently and confidently misclassify the unseen pneumonia subtype to their respective disease class. These findings expose a critical limitation concerning the tendency of binary classifiers to hallucinate by relying on general pneumonia indicators rather than pathogen-specific patterns, thereby challenging their utility in clinical workflows.

深度学习(DL)分类中的幻觉,其中DL模型产生自信的错误预测仍然是一个紧迫的问题。本研究探讨了二元分类器在区分胸部x线(CXR)图像上肺炎亚型的重叠放射表现时是否真正了解疾病特异性特征。具体来说,我们评估不确定性测量是否是分类不同病原体特异性肺炎亚型体征的有价值的工具。我们评估了两种二元分类器,分别将细菌性肺炎和病毒性肺炎与正常cxr进行分类。第三种分类探讨了区分细菌性和病毒性肺炎表现的能力,以突出我们对前一种病例中观察到的幻觉的关注。我们的综合分析计算了儿童CXR数据集上的马修斯相关系数和预测熵指标,并揭示了正常/细菌和正常/病毒分类器一致且自信地将未见的肺炎亚型错误分类为各自的疾病类别。这些发现揭示了二元分类器依赖于一般肺炎指标而不是病原体特异性模式而产生幻觉的关键局限性,从而挑战了它们在临床工作流程中的实用性。
{"title":"The Hidden Threat of Hallucinations in Binary Chest X-ray Pneumonia Classification.","authors":"Sivaramakrishnan Rajaraman, Zhaohui Liang, Niccolo Marini, Zhiyun Xue, Sameer Antani","doi":"10.1109/cbms65348.2025.00138","DOIUrl":"https://doi.org/10.1109/cbms65348.2025.00138","url":null,"abstract":"<p><p>Hallucination in deep learning (DL) classification, where DL models yield confidently erroneous predictions remains a pressing concern. This study investigates whether binary classifiers are truly learning disease-specific features when distinguishing overlapping radiological presentations among pneumonia subtypes on chest X-ray (CXR) images. Specifically, we evaluate if uncertainty measure is a valuable tool in classifying signs of different pathogen-specific subtypes of pneumonia. We evaluated two binary classifiers to classify bacterial pneumonia and viral pneumonia, respectively, from normal CXRs. A third classifier explored the ability to distinguish bacterial from viral pneumonia presentation to highlight our concern regarding the observed hallucinations in the former cases. Our comprehensive analysis computes the Matthews Correlation Coefficient and prediction entropy metrics on a pediatric CXR dataset and reveals that the normal/bacterial and normal/viral classifiers consistently and confidently misclassify the unseen pneumonia subtype to their respective disease class. These findings expose a critical limitation concerning the tendency of binary classifiers to hallucinate by relying on general pneumonia indicators rather than pathogen-specific patterns, thereby challenging their utility in clinical workflows.</p>","PeriodicalId":74567,"journal":{"name":"Proceedings. IEEE International Symposium on Computer-Based Medical Systems","volume":"2025 ","pages":"668-673"},"PeriodicalIF":0.0,"publicationDate":"2025-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12369649/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144981943","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Enhancing Concept-Based Explanation with Vision-Language Models. 用视觉语言模型增强基于概念的解释。
Pub Date : 2024-06-01 Epub Date: 2024-07-25 DOI: 10.1109/CBMS61543.2024.00044
Imran Hossain, Ghada Zamzmi, Peter Mouton, Yu Sun, Dmitry Goldgof

Although concept-based approaches are widely used to explain a model's behavior and assess the contributions of different concepts in decision-making, identifying relevant concepts can be challenging for non-experts. This paper introduces a novel method that simplifies concept selection by leveraging the capabilities of a state-of-the-art large Vision-Language Model (VLM). Our method employs a VLM to select textual concepts that describe the classes in the target dataset. We then transform these influential textual concepts into human-readable image concepts using a text-to-image model. This process allows us to explain the targeted network in a post-hoc manner. Further, we use directional derivatives and concept activation vectors to quantify the importance of the generated concepts. We evaluate our method on a neonatal pain classification task, analyzing the sensitivity of the model's output for the generated concepts. The results demonstrate that the VLM not only generates coherent and meaningful concepts that are easily understandable by non-experts but also achieves performance comparable to that of natural image concepts without the need for additional annotation costs.

尽管基于概念的方法被广泛用于解释模型的行为和评估决策中不同概念的贡献,但识别相关概念对于非专家来说可能是具有挑战性的。本文介绍了一种利用最先进的大型视觉语言模型(VLM)简化概念选择的新方法。我们的方法使用VLM来选择描述目标数据集中类的文本概念。然后,我们使用文本到图像模型将这些有影响力的文本概念转换为人类可读的图像概念。这个过程使我们能够以事后的方式解释目标网络。此外,我们使用方向导数和概念激活向量来量化生成的概念的重要性。我们在新生儿疼痛分类任务上评估了我们的方法,分析了模型输出对生成概念的敏感性。结果表明,VLM不仅生成了非专家易于理解的连贯且有意义的概念,而且在不需要额外注释成本的情况下,达到了与自然图像概念相当的性能。
{"title":"Enhancing Concept-Based Explanation with Vision-Language Models.","authors":"Imran Hossain, Ghada Zamzmi, Peter Mouton, Yu Sun, Dmitry Goldgof","doi":"10.1109/CBMS61543.2024.00044","DOIUrl":"10.1109/CBMS61543.2024.00044","url":null,"abstract":"<p><p>Although concept-based approaches are widely used to explain a model's behavior and assess the contributions of different concepts in decision-making, identifying relevant concepts can be challenging for non-experts. This paper introduces a novel method that simplifies concept selection by leveraging the capabilities of a state-of-the-art large Vision-Language Model (VLM). Our method employs a VLM to select textual concepts that describe the classes in the target dataset. We then transform these influential textual concepts into human-readable image concepts using a text-to-image model. This process allows us to explain the targeted network in a post-hoc manner. Further, we use directional derivatives and concept activation vectors to quantify the importance of the generated concepts. We evaluate our method on a neonatal pain classification task, analyzing the sensitivity of the model's output for the generated concepts. The results demonstrate that the VLM not only generates coherent and meaningful concepts that are easily understandable by non-experts but also achieves performance comparable to that of natural image concepts without the need for additional annotation costs.</p>","PeriodicalId":74567,"journal":{"name":"Proceedings. IEEE International Symposium on Computer-Based Medical Systems","volume":"2024 ","pages":"219-224"},"PeriodicalIF":0.0,"publicationDate":"2024-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12458896/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145152251","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Prediction of MRI-Induced Power Absorption in Patients with DBS Leads. 脑起搏器导联患者mri诱导的能量吸收预测。
Pub Date : 2024-06-01 Epub Date: 2024-07-05 DOI: 10.1109/cbms61543.2024.00087
Yalcin Tur, Jasmine Vu, Selam Waktola, Alpay Medetalibeyoglu, Laleh Golestanirad, Ulas Bagci

The interaction between deep brain stimulation (DBS) systems and magnetic resonance imaging (MRI) can induce tissue heating in patients. While electromagnetic (EM) simulations can be used to estimate the specific absorption rate (SAR) values in the presence of an implanted DBS system, they are computationally expensive. To address this drawback, we predict local SAR values in the tips of DBS leads with machine learning based efficient algorithms, specifically XgBoost and deep learning. We significantly outperformed the previous state of the art, and adapted new machine learning models based on Residual Networks family as well as XgBoost models. We observed that already extracted limited features are better suited for ensemble learning via XgBoost than deep networks due the small-data regime. Although we conclude that boosting gradient algorithm is more suitable for this non-linear regression problem due to structured nature of the data and small data regime, we found that width plays a more critical role than depth in network design and it has a strong potential for future research. Our experimental results, using a dataset of 260 instances that are patient-derived and artificial, reached an outstanding RMSE of 17.8 W/kg with XgBoost, 78 W/kg with deep networks, given that the previous study on this problem reached a state-of-the-art root mean square error value (RMSE) of 168 W/kg.

脑深部刺激(DBS)系统与磁共振成像(MRI)之间的相互作用可以诱导患者的组织加热。虽然电磁(EM)模拟可以用来估计植入DBS系统时的比吸收率(SAR)值,但它们的计算成本很高。为了解决这一缺陷,我们使用基于机器学习的高效算法,特别是XgBoost和深度学习,来预测DBS引线尖端的局部SAR值。我们大大超越了之前的技术水平,并采用了基于残余网络家族和XgBoost模型的新机器学习模型。我们观察到,由于小数据机制,已经提取的有限特征比深度网络更适合通过XgBoost进行集成学习。尽管我们得出结论,由于数据的结构化性质和小数据体系,增强梯度算法更适合于这种非线性回归问题,但我们发现宽度在网络设计中比深度起着更关键的作用,并且它具有强大的未来研究潜力。我们的实验结果使用了260个患者衍生的人工数据集,XgBoost的RMSE为17.8 W/kg,深度网络的RMSE为78 W/kg,而之前对该问题的研究达到了最先进的均方根误差值(RMSE)为168 W/kg。
{"title":"Prediction of MRI-Induced Power Absorption in Patients with DBS Leads.","authors":"Yalcin Tur, Jasmine Vu, Selam Waktola, Alpay Medetalibeyoglu, Laleh Golestanirad, Ulas Bagci","doi":"10.1109/cbms61543.2024.00087","DOIUrl":"10.1109/cbms61543.2024.00087","url":null,"abstract":"<p><p>The interaction between deep brain stimulation (DBS) systems and magnetic resonance imaging (MRI) can induce tissue heating in patients. While electromagnetic (EM) simulations can be used to estimate the specific absorption rate (SAR) values in the presence of an implanted DBS system, they are computationally expensive. To address this drawback, we predict local SAR values in the tips of DBS leads with machine learning based efficient algorithms, specifically XgBoost and deep learning. We significantly outperformed the previous state of the art, and adapted new machine learning models based on Residual Networks family as well as XgBoost models. We observed that already extracted limited features are better suited for ensemble learning via XgBoost than deep networks due the small-data regime. Although we conclude that boosting gradient algorithm is more suitable for this non-linear regression problem due to structured nature of the data and small data regime, we found that width plays a more critical role than depth in network design and it has a strong potential for future research. Our experimental results, using a dataset of 260 instances that are patient-derived and artificial, reached an outstanding RMSE of 17.8 W/kg with XgBoost, 78 W/kg with deep networks, given that the previous study on this problem reached a state-of-the-art root mean square error value (RMSE) of 168 W/kg.</p>","PeriodicalId":74567,"journal":{"name":"Proceedings. IEEE International Symposium on Computer-Based Medical Systems","volume":"2024 ","pages":"490-495"},"PeriodicalIF":0.0,"publicationDate":"2024-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12477686/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145202461","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Automated Design of Task-Dedicated Illumination with Particle Swarm Optimization 基于粒子群优化的任务专用照明自动设计
Pub Date : 2023-01-01 DOI: 10.1109/CBMS58004.2023.00254
Austin Ryan English
{"title":"Automated Design of Task-Dedicated Illumination with Particle Swarm Optimization","authors":"Austin Ryan English","doi":"10.1109/CBMS58004.2023.00254","DOIUrl":"https://doi.org/10.1109/CBMS58004.2023.00254","url":null,"abstract":"","PeriodicalId":74567,"journal":{"name":"Proceedings. IEEE International Symposium on Computer-Based Medical Systems","volume":"9 1","pages":"416-421"},"PeriodicalIF":0.0,"publicationDate":"2023-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"78809764","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Video Capsule Endoscopy Classification using Focal Modulation Guided Convolutional Neural Network. 基于焦调制引导卷积神经网络的视频胶囊内窥镜分类。
Pub Date : 2022-07-01 DOI: 10.1109/CBMS55023.2022.00064
Abhishek Srivastava, Nikhil Kumar Tomar, Ulas Bagci, Debesh Jha

Video capsule endoscopy is a hot topic in computer vision and medicine. Deep learning can have a positive impact on the future of video capsule endoscopy technology. It can improve the anomaly detection rate, reduce physicians' time for screening, and aid in real-world clinical analysis. Computer-Aided diagnosis (CADx) classification system for video capsule endoscopy has shown a great promise for further improvement. For example, detection of cancerous polyp and bleeding can lead to swift medical response and improve the survival rate of the patients. To this end, an automated CADx system must have high throughput and decent accuracy. In this study, we propose FocalConvNet, a focal modulation network integrated with lightweight convolutional layers for the classification of small bowel anatomical landmarks and luminal findings. FocalConvNet leverages focal modulation to attain global context and allows global-local spatial interactions throughout the forward pass. Moreover, the convolutional block with its intrinsic inductive/learning bias and capacity to extract hierarchical features allows our FocalConvNet to achieve favourable results with high throughput. We compare our FocalConvNet with other state-of-the-art (SOTA) on Kvasir-Capsule, a large-scale VCE dataset with 44,228 frames with 13 classes of different anomalies. We achieved the weighted F1-score, recall and Matthews correlation coefficient (MCC) of 0.6734, 0.6373 and 0.2974, respectively, outperforming SOTA methodologies. Further, we obtained the highest throughput of 148.02 images/second rate to establish the potential of FocalConvNet in a real-time clinical environment. The code of the proposed FocalConvNet is available at https://github.com/NoviceMAn-prog/FocalConvNet.

视频胶囊内窥镜是计算机视觉和医学领域的研究热点。深度学习可以对视频胶囊内窥镜技术的未来产生积极的影响。它可以提高异常检出率,减少医生的筛查时间,并有助于现实世界的临床分析。视频胶囊内窥镜的计算机辅助诊断(CADx)分类系统有很大的发展前景。例如,发现癌性息肉和出血可以迅速做出医疗反应,提高患者的存活率。为此,自动化CADx系统必须具有高吞吐量和良好的精度。在这项研究中,我们提出了FocalConvNet,这是一个集成了轻量级卷积层的焦点调制网络,用于小肠解剖标志和腔内发现的分类。FocalConvNet利用焦点调制来获得全局上下文,并允许在整个前传过程中进行全局-局部空间交互。此外,卷积块具有其固有的归纳/学习偏差和提取分层特征的能力,使我们的FocalConvNet能够以高吞吐量获得良好的结果。我们将我们的FocalConvNet与Kvasir-Capsule上的其他先进技术(SOTA)进行了比较,Kvasir-Capsule是一个大型VCE数据集,具有44,228帧和13类不同的异常。我们获得了加权f1得分,召回率和马修斯相关系数(MCC)分别为0.6734,0.6373和0.2974,优于SOTA方法。此外,我们获得了148.02张图像/秒的最高吞吐量,以确定FocalConvNet在实时临床环境中的潜力。所提出的FocalConvNet的代码可在https://github.com/NoviceMAn-prog/FocalConvNet上获得。
{"title":"Video Capsule Endoscopy Classification using Focal Modulation Guided Convolutional Neural Network.","authors":"Abhishek Srivastava,&nbsp;Nikhil Kumar Tomar,&nbsp;Ulas Bagci,&nbsp;Debesh Jha","doi":"10.1109/CBMS55023.2022.00064","DOIUrl":"https://doi.org/10.1109/CBMS55023.2022.00064","url":null,"abstract":"<p><p>Video capsule endoscopy is a hot topic in computer vision and medicine. Deep learning can have a positive impact on the future of video capsule endoscopy technology. It can improve the anomaly detection rate, reduce physicians' time for screening, and aid in real-world clinical analysis. Computer-Aided diagnosis (CADx) classification system for video capsule endoscopy has shown a great promise for further improvement. For example, detection of cancerous polyp and bleeding can lead to swift medical response and improve the survival rate of the patients. To this end, an automated CADx system must have high throughput and decent accuracy. In this study, we propose <i>FocalConvNet</i>, a focal modulation network integrated with lightweight convolutional layers for the classification of small bowel anatomical landmarks and luminal findings. FocalConvNet leverages focal modulation to attain global context and allows global-local spatial interactions throughout the forward pass. Moreover, the convolutional block with its intrinsic inductive/learning bias and capacity to extract hierarchical features allows our FocalConvNet to achieve favourable results with high throughput. We compare our FocalConvNet with other state-of-the-art (SOTA) on Kvasir-Capsule, a large-scale VCE dataset with 44,228 frames with 13 classes of different anomalies. We achieved the weighted F1-score, recall and Matthews correlation coefficient (MCC) of 0.6734, 0.6373 and 0.2974, respectively, outperforming SOTA methodologies. Further, we obtained the highest throughput of 148.02 images/second rate to establish the potential of FocalConvNet in a real-time clinical environment. The code of the proposed FocalConvNet is available at https://github.com/NoviceMAn-prog/FocalConvNet.</p>","PeriodicalId":74567,"journal":{"name":"Proceedings. IEEE International Symposium on Computer-Based Medical Systems","volume":"2022 ","pages":"323-328"},"PeriodicalIF":0.0,"publicationDate":"2022-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9914988/pdf/nihms-1871537.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"10708367","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Automatic Polyp Segmentation with Multiple Kernel Dilated Convolution Network. 基于多核扩展卷积网络的息肉自动分割。
Pub Date : 2022-07-01 DOI: 10.1109/CBMS55023.2022.00063
Nikhil Kumar Tomar, Abhishek Srivastava, Ulas Bagci, Debesh Jha

The detection and removal of precancerous polyps through colonoscopy is the primary technique for the prevention of colorectal cancer worldwide. However, the miss rate of colorectal polyp varies significantly among the endoscopists. It is well known that a computer-aided diagnosis (CAD) system can assist endoscopists in detecting colon polyps and minimize the variation among endoscopists. In this study, we introduce a novel deep learning architecture, named MKDCNet, for automatic polyp segmentation robust to significant changes in polyp data distribution. MKDCNet is simply an encoder-decoder neural network that uses the pre-trained ResNet50 as the encoder and novel multiple kernel dilated convolution (MKDC) block that expands the field of view to learn more robust and heterogeneous representation. Extensive experiments on four publicly available polyp datasets and cell nuclei dataset show that the proposed MKDCNet outperforms the state-of-the-art methods when trained and tested on the same dataset as well when tested on unseen polyp datasets from different distributions. With rich results, we demonstrated the robustness of the proposed architecture. From an efficiency perspective, our algorithm can process at ( 45) frames per second on RTX 3090 GPU. MKDCNet can be a strong benchmark for building real-time systems for clinical colonoscopies. The code of the proposed MKDCNet is available at https://github.com/nikhilroxtomar/MKDCNet.

通过结肠镜检查发现和切除癌前息肉是世界范围内预防结直肠癌的主要技术。然而,内镜医师对结直肠息肉的漏诊率差异很大。众所周知,计算机辅助诊断(CAD)系统可以帮助内窥镜医师发现结肠息肉,并最大限度地减少内窥镜医师之间的差异。在本研究中,我们引入了一种名为MKDCNet的新型深度学习架构,用于对息肉数据分布的显著变化进行自动息肉分割。MKDCNet是一个简单的编码器-解码器神经网络,它使用预训练的ResNet50作为编码器和新的多核扩展卷积(MKDC)块,扩展视野以学习更鲁棒和异构的表示。在四个公开可用的息肉数据集和细胞核数据集上进行的大量实验表明,所提出的MKDCNet在同一数据集上训练和测试以及在来自不同分布的未见过的息肉数据集上测试时都优于最先进的方法。通过丰富的结果,我们证明了所提出体系结构的鲁棒性。从效率的角度来看,我们的算法可以在RTX 3090 GPU上以每秒(≈45)帧的速度处理。MKDCNet可以成为构建临床结肠镜检查实时系统的有力基准。建议的MKDCNet的代码可在https://github.com/nikhilroxtomar/MKDCNet上获得。
{"title":"Automatic Polyp Segmentation with Multiple Kernel Dilated Convolution Network.","authors":"Nikhil Kumar Tomar,&nbsp;Abhishek Srivastava,&nbsp;Ulas Bagci,&nbsp;Debesh Jha","doi":"10.1109/CBMS55023.2022.00063","DOIUrl":"https://doi.org/10.1109/CBMS55023.2022.00063","url":null,"abstract":"<p><p>The detection and removal of precancerous polyps through colonoscopy is the primary technique for the prevention of colorectal cancer worldwide. However, the miss rate of colorectal polyp varies significantly among the endoscopists. It is well known that a computer-aided diagnosis (CAD) system can assist endoscopists in detecting colon polyps and minimize the variation among endoscopists. In this study, we introduce a novel deep learning architecture, named MKDCNet, for automatic polyp segmentation robust to significant changes in polyp data distribution. MKDCNet is simply an encoder-decoder neural network that uses the pre-trained <i>ResNet50</i> as the encoder and novel <i>multiple kernel dilated convolution (MKDC)</i> block that expands the field of view to learn more robust and heterogeneous representation. Extensive experiments on four publicly available polyp datasets and cell nuclei dataset show that the proposed MKDCNet outperforms the state-of-the-art methods when trained and tested on the same dataset as well when tested on unseen polyp datasets from different distributions. With rich results, we demonstrated the robustness of the proposed architecture. From an efficiency perspective, our algorithm can process at (<i>≈</i> 45) frames per second on RTX 3090 GPU. MKDCNet can be a strong benchmark for building real-time systems for clinical colonoscopies. The code of the proposed MKDCNet is available at https://github.com/nikhilroxtomar/MKDCNet.</p>","PeriodicalId":74567,"journal":{"name":"Proceedings. IEEE International Symposium on Computer-Based Medical Systems","volume":"2022 ","pages":"317-322"},"PeriodicalIF":0.0,"publicationDate":"2022-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9921313/pdf/nihms-1871530.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"10708366","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Mental Health Ubiquitous Monitoring: Detecting Context-Enriched Sociability Patterns Through Complex Event Processing 心理健康无所不在监测:通过复杂事件处理检测情境丰富的社交模式
Pub Date : 2020-07-01 DOI: 10.1109/CBMS49503.2020.00052
I. Moura, Francisco Silva, L. Coutinho, A. Teles
Traditionally, the process of monitoring and evaluating social behavior related to mental health has based on self-reported information, which is limited by the subjective character of responses and by various cognitive biases. Today, however, computational methods can use ubiquitous devices to monitor social behaviors related to mental health rather than relying on self-reports. Therefore, these technologies can be used to identify the routine of social activities, which enables the recognition of abnormal behaviors that may be indicative of mental disorders. In this paper, we present a solution for detecting context-enriched sociability patterns. Specifically, we introduced an algorithm capable of recognizing the social routine of monitored people. To implement the proposed algorithm, it was used a set of Complex Event Processing (CEP) rules, which allow the continuous processing of the social data stream derived from ubiquitous devices. The experiments performed indicated that the proposed solution is capable of detecting sociability patterns similar to a batch algorithm and demonstrated that context-based recognition provides a better understanding of social routine.
传统上,监测和评估与心理健康有关的社会行为的过程是基于自我报告的信息,这受到反应的主观特征和各种认知偏见的限制。然而,今天,计算方法可以使用无处不在的设备来监测与心理健康相关的社会行为,而不是依赖于自我报告。因此,这些技术可以用来识别日常的社会活动,从而能够识别可能表明精神障碍的异常行为。在本文中,我们提出了一种检测上下文丰富的社交模式的解决方案。具体来说,我们介绍了一种能够识别被监控人员的社交日常的算法。为了实现所提出的算法,该算法使用了一套复杂事件处理(CEP)规则,该规则允许对来自无处不在的设备的社交数据流进行连续处理。实验表明,所提出的解决方案能够检测类似于批处理算法的社交模式,并证明基于上下文的识别能够更好地理解社交常规。
{"title":"Mental Health Ubiquitous Monitoring: Detecting Context-Enriched Sociability Patterns Through Complex Event Processing","authors":"I. Moura, Francisco Silva, L. Coutinho, A. Teles","doi":"10.1109/CBMS49503.2020.00052","DOIUrl":"https://doi.org/10.1109/CBMS49503.2020.00052","url":null,"abstract":"Traditionally, the process of monitoring and evaluating social behavior related to mental health has based on self-reported information, which is limited by the subjective character of responses and by various cognitive biases. Today, however, computational methods can use ubiquitous devices to monitor social behaviors related to mental health rather than relying on self-reports. Therefore, these technologies can be used to identify the routine of social activities, which enables the recognition of abnormal behaviors that may be indicative of mental disorders. In this paper, we present a solution for detecting context-enriched sociability patterns. Specifically, we introduced an algorithm capable of recognizing the social routine of monitored people. To implement the proposed algorithm, it was used a set of Complex Event Processing (CEP) rules, which allow the continuous processing of the social data stream derived from ubiquitous devices. The experiments performed indicated that the proposed solution is capable of detecting sociability patterns similar to a batch algorithm and demonstrated that context-based recognition provides a better understanding of social routine.","PeriodicalId":74567,"journal":{"name":"Proceedings. IEEE International Symposium on Computer-Based Medical Systems","volume":"19 1","pages":"239-244"},"PeriodicalIF":0.0,"publicationDate":"2020-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"74119568","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
期刊
Proceedings. IEEE International Symposium on Computer-Based Medical Systems
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1