Lingwei Li, Tongtong Liu, Peng Wang, Lianzheng Su, Lei Wang, Xinmiao Wang, Chidao Chen
{"title":"基于多感知对比学习的卵巢肿瘤CT自动分类。","authors":"Lingwei Li, Tongtong Liu, Peng Wang, Lianzheng Su, Lei Wang, Xinmiao Wang, Chidao Chen","doi":"10.1007/s00261-025-04879-y","DOIUrl":null,"url":null,"abstract":"<div><p>Ovarian cancer is among the most common malignant tumours in women worldwide, and early identification is essential for enhancing patient survival chances. The development of automated and trustworthy diagnostic techniques is necessary because traditional CT picture processing mostly depends on the subjective assessment of radiologists, which can result in variability. Deep learning approaches in medical image analysis have advanced significantly, particularly showing considerable promise in the automatic categorisation of ovarian tumours. This research presents an automated diagnostic approach for ovarian tumour CT images utilising supervised contrastive learning and a Multiple Perception Encoder (MP Encoder). The approach incorporates T-Pro technology to augment data diversity and simulates semantic perturbations to increase the model’s generalisation capability. The incorporation of Multi-Scale Perception Module (MSP Module) and Multi-Attention Module (MA Module) enhances the model’s sensitivity to the intricate morphology and subtle characteristics of ovarian tumours, resulting in improved classification accuracy and robustness, ultimately achieving an average classification accuracy of 98.43%. Experimental results indicate the method’s exceptional efficacy in ovarian tumour classification, particularly in cases involving tumours with intricate morphology or worse picture quality, thereby markedly enhancing classification accuracy. This advanced deep learning framework proficiently tackles the complexities of ovarian tumour CT image interpretation, offering clinicians enhanced diagnostic support and aiding in the optimisation of early detection and treatment strategies for ovarian cancer.</p></div>","PeriodicalId":7126,"journal":{"name":"Abdominal Radiology","volume":"50 9","pages":"4342 - 4358"},"PeriodicalIF":2.2000,"publicationDate":"2025-03-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Multiple perception contrastive learning for automated ovarian tumor classification in CT images\",\"authors\":\"Lingwei Li, Tongtong Liu, Peng Wang, Lianzheng Su, Lei Wang, Xinmiao Wang, Chidao Chen\",\"doi\":\"10.1007/s00261-025-04879-y\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><p>Ovarian cancer is among the most common malignant tumours in women worldwide, and early identification is essential for enhancing patient survival chances. The development of automated and trustworthy diagnostic techniques is necessary because traditional CT picture processing mostly depends on the subjective assessment of radiologists, which can result in variability. Deep learning approaches in medical image analysis have advanced significantly, particularly showing considerable promise in the automatic categorisation of ovarian tumours. This research presents an automated diagnostic approach for ovarian tumour CT images utilising supervised contrastive learning and a Multiple Perception Encoder (MP Encoder). The approach incorporates T-Pro technology to augment data diversity and simulates semantic perturbations to increase the model’s generalisation capability. The incorporation of Multi-Scale Perception Module (MSP Module) and Multi-Attention Module (MA Module) enhances the model’s sensitivity to the intricate morphology and subtle characteristics of ovarian tumours, resulting in improved classification accuracy and robustness, ultimately achieving an average classification accuracy of 98.43%. Experimental results indicate the method’s exceptional efficacy in ovarian tumour classification, particularly in cases involving tumours with intricate morphology or worse picture quality, thereby markedly enhancing classification accuracy. This advanced deep learning framework proficiently tackles the complexities of ovarian tumour CT image interpretation, offering clinicians enhanced diagnostic support and aiding in the optimisation of early detection and treatment strategies for ovarian cancer.</p></div>\",\"PeriodicalId\":7126,\"journal\":{\"name\":\"Abdominal Radiology\",\"volume\":\"50 9\",\"pages\":\"4342 - 4358\"},\"PeriodicalIF\":2.2000,\"publicationDate\":\"2025-03-12\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Abdominal Radiology\",\"FirstCategoryId\":\"3\",\"ListUrlMain\":\"https://link.springer.com/article/10.1007/s00261-025-04879-y\",\"RegionNum\":3,\"RegionCategory\":\"医学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q2\",\"JCRName\":\"RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Abdominal Radiology","FirstCategoryId":"3","ListUrlMain":"https://link.springer.com/article/10.1007/s00261-025-04879-y","RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING","Score":null,"Total":0}
引用次数: 0
摘要
卵巢癌是全世界妇女中最常见的恶性肿瘤之一,早期发现对于提高患者的生存机会至关重要。由于传统的CT图像处理主要依赖于放射科医生的主观评估,这可能导致变化,因此开发自动化和可靠的诊断技术是必要的。医学图像分析中的深度学习方法已经取得了显着进展,特别是在卵巢肿瘤的自动分类中显示出相当大的前景。本研究提出了一种利用监督对比学习和多感知编码器(MP编码器)对卵巢肿瘤CT图像进行自动诊断的方法。该方法结合了T-Pro技术来增加数据多样性,并模拟语义扰动以增加模型的泛化能力。多尺度感知模块(Multi-Scale Perception Module, MSP Module)和多注意模块(Multi-Attention Module, MA Module)的结合增强了模型对卵巢肿瘤复杂形态和微妙特征的敏感性,提高了分类精度和鲁棒性,最终平均分类准确率达到98.43%。实验结果表明,该方法在卵巢肿瘤分类中具有特殊的功效,特别是在肿瘤形态复杂或图像质量较差的情况下,从而显着提高了分类精度。这种先进的深度学习框架熟练地解决了卵巢肿瘤CT图像解释的复杂性,为临床医生提供了增强的诊断支持,并帮助优化卵巢癌的早期检测和治疗策略。
Multiple perception contrastive learning for automated ovarian tumor classification in CT images
Ovarian cancer is among the most common malignant tumours in women worldwide, and early identification is essential for enhancing patient survival chances. The development of automated and trustworthy diagnostic techniques is necessary because traditional CT picture processing mostly depends on the subjective assessment of radiologists, which can result in variability. Deep learning approaches in medical image analysis have advanced significantly, particularly showing considerable promise in the automatic categorisation of ovarian tumours. This research presents an automated diagnostic approach for ovarian tumour CT images utilising supervised contrastive learning and a Multiple Perception Encoder (MP Encoder). The approach incorporates T-Pro technology to augment data diversity and simulates semantic perturbations to increase the model’s generalisation capability. The incorporation of Multi-Scale Perception Module (MSP Module) and Multi-Attention Module (MA Module) enhances the model’s sensitivity to the intricate morphology and subtle characteristics of ovarian tumours, resulting in improved classification accuracy and robustness, ultimately achieving an average classification accuracy of 98.43%. Experimental results indicate the method’s exceptional efficacy in ovarian tumour classification, particularly in cases involving tumours with intricate morphology or worse picture quality, thereby markedly enhancing classification accuracy. This advanced deep learning framework proficiently tackles the complexities of ovarian tumour CT image interpretation, offering clinicians enhanced diagnostic support and aiding in the optimisation of early detection and treatment strategies for ovarian cancer.
期刊介绍:
Abdominal Radiology seeks to meet the professional needs of the abdominal radiologist by publishing clinically pertinent original, review and practice related articles on the gastrointestinal and genitourinary tracts and abdominal interventional and radiologic procedures. Case reports are generally not accepted unless they are the first report of a new disease or condition, or part of a special solicited section.
Reasons to Publish Your Article in Abdominal Radiology:
· Official journal of the Society of Abdominal Radiology (SAR)
· Published in Cooperation with:
European Society of Gastrointestinal and Abdominal Radiology (ESGAR)
European Society of Urogenital Radiology (ESUR)
Asian Society of Abdominal Radiology (ASAR)
· Efficient handling and Expeditious review
· Author feedback is provided in a mentoring style
· Global readership
· Readers can earn CME credits