Pub Date : 2024-05-01Epub Date: 2024-08-22DOI: 10.1109/ISBI56570.2024.10635349
Sepehr Salem Ghahfarokhi, Tyrell To, Julie Jorns, Tina Yen, Bing Yu, Dong Hye Ye
Data limitation is a significant challenge in applying deep learning to medical images. Recently, the diffusion probabilistic model (DPM) has shown the potential to generate high-quality images by converting Gaussian random noise into realistic images. In this paper, we apply the DPM to augment the deep ultraviolet fluorescence (DUV) image dataset with an aim to improve breast cancer classification for intra-operative margin assessment. For classification, we divide the whole surface DUV image into small patches and extract convolutional features for each patch by utilizing the pre-trained ResNet. Then, we feed them into an XGBoost classifier for patch-level decisions and then fuse them with a regional importance map computed by Grad-CAM++ for whole surface-level prediction. Our experimental results show that augmenting the training dataset with the DPM significantly improves breast cancer detection performance in DUV images, increasing accuracy from 93% to 97%, compared to using Affine transformations and ProGAN.
{"title":"DEEP LEARNING FOR AUTOMATED DETECTION OF BREAST CANCER IN DEEP ULTRAVIOLET FLUORESCENCE IMAGES WITH DIFFUSION PROBABILISTIC MODEL.","authors":"Sepehr Salem Ghahfarokhi, Tyrell To, Julie Jorns, Tina Yen, Bing Yu, Dong Hye Ye","doi":"10.1109/ISBI56570.2024.10635349","DOIUrl":"10.1109/ISBI56570.2024.10635349","url":null,"abstract":"<p><p>Data limitation is a significant challenge in applying deep learning to medical images. Recently, the diffusion probabilistic model (DPM) has shown the potential to generate high-quality images by converting Gaussian random noise into realistic images. In this paper, we apply the DPM to augment the deep ultraviolet fluorescence (DUV) image dataset with an aim to improve breast cancer classification for intra-operative margin assessment. For classification, we divide the whole surface DUV image into small patches and extract convolutional features for each patch by utilizing the pre-trained ResNet. Then, we feed them into an XGBoost classifier for patch-level decisions and then fuse them with a regional importance map computed by Grad-CAM++ for whole surface-level prediction. Our experimental results show that augmenting the training dataset with the DPM significantly improves breast cancer detection performance in DUV images, increasing accuracy from 93% to 97%, compared to using Affine transformations and ProGAN.</p>","PeriodicalId":74566,"journal":{"name":"Proceedings. IEEE International Symposium on Biomedical Imaging","volume":"2024 ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12045284/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143999914","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-05-01Epub Date: 2024-08-22DOI: 10.1109/isbi56570.2024.10635855
Toygan Kilic, Jürgen Herrler, Patrick Liebig, Ömer Burak Demirel, Armin Nagel, Mingyi Hong, Georgios B Giannakis, Kamil Ugurbil, Mehmet Akçakaya
Parallel transmission (pTx) is an important technique for reducing transmit field inhomogeneities at ultrahigh-field (UHF) MRI. pTx typically involves solving an optimization problem for radiofrequency pulse design, with hard constraints on specific-absorption rate (SAR) and/or power, which may be time-consuming. In this work, we propose a novel approach towards incorporating hard constraints to physics-driven neural networks. Our method unrolls an extension of the log-barrier method, where the central path problems are solved via the gradient descent method whose optimal step sizes are learned with a neural network. Results indicate that our method is substantially faster compared to traditional convex optimization techniques, while achieving similar performance.
{"title":"TOWARDS FAST HARD-CONSTRAINED PARALLEL TRANSMIT DESIGN IN ULTRAHIGH FIELD MRI WITH PHYSICS-DRIVEN NEURAL NETWORKS.","authors":"Toygan Kilic, Jürgen Herrler, Patrick Liebig, Ömer Burak Demirel, Armin Nagel, Mingyi Hong, Georgios B Giannakis, Kamil Ugurbil, Mehmet Akçakaya","doi":"10.1109/isbi56570.2024.10635855","DOIUrl":"10.1109/isbi56570.2024.10635855","url":null,"abstract":"<p><p>Parallel transmission (pTx) is an important technique for reducing transmit field inhomogeneities at ultrahigh-field (UHF) MRI. pTx typically involves solving an optimization problem for radiofrequency pulse design, with hard constraints on specific-absorption rate (SAR) and/or power, which may be time-consuming. In this work, we propose a novel approach towards incorporating hard constraints to physics-driven neural networks. Our method unrolls an extension of the log-barrier method, where the central path problems are solved via the gradient descent method whose optimal step sizes are learned with a neural network. Results indicate that our method is substantially faster compared to traditional convex optimization techniques, while achieving similar performance.</p>","PeriodicalId":74566,"journal":{"name":"Proceedings. IEEE International Symposium on Biomedical Imaging","volume":"2024 ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11736015/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143017999","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-05-01Epub Date: 2024-08-22DOI: 10.1109/isbi56570.2024.10635403
Xiaofeng Liu, Fangxu Xing, Hanna Gaggin, C-C Jay Kuo, Georges El Fakhri, Jonghye Woo
The off-the-shelf model for unsupervised domain adaptation (OSUDA) has been introduced to protect patient data privacy and intellectual property of the source domain without access to the labeled source domain data. Yet, an off-the-shelf diagnosis model, deliberately compromised by backdoor attacks during the source domain training phase, can function as a parasite-host, disseminating the backdoor to the target domain model during the OSUDA stage. Because of limitations in accessing or controlling the source domain training data, OSUDA can make the target domain model highly vulnerable and susceptible to prominent attacks. To sidestep this issue, we propose to quantify the channel-wise backdoor sensitivity via a Lipschitz constant and, explicitly, eliminate the backdoor infection by overwriting the backdoor-related channel kernels with random initialization. Furthermore, we propose to employ an auxiliary model with a full source model to ensure accurate pseudo-labeling, taking into account the controllable, clean target training data in OSUDA. We validate our framework using a multi-center, multi-vendor, and multi-disease (M&M) cardiac dataset. Our findings suggest that the target model is susceptible to backdoor attacks during OSUDA, and our defense mechanism effectively mitigates the infection of target domain victims.
{"title":"EXPLORING BACKDOOR ATTACKS IN OFF-THE-SHELF UNSUPERVISED DOMAIN ADAPTATION FOR SECURING CARDIAC MRI-BASED DIAGNOSIS.","authors":"Xiaofeng Liu, Fangxu Xing, Hanna Gaggin, C-C Jay Kuo, Georges El Fakhri, Jonghye Woo","doi":"10.1109/isbi56570.2024.10635403","DOIUrl":"10.1109/isbi56570.2024.10635403","url":null,"abstract":"<p><p>The off-the-shelf model for unsupervised domain adaptation (OSUDA) has been introduced to protect patient data privacy and intellectual property of the source domain without access to the labeled source domain data. Yet, an off-the-shelf diagnosis model, deliberately compromised by backdoor attacks during the source domain training phase, can function as a parasite-host, disseminating the backdoor to the target domain model during the OSUDA stage. Because of limitations in accessing or controlling the source domain training data, OSUDA can make the target domain model highly vulnerable and susceptible to prominent attacks. To sidestep this issue, we propose to quantify the channel-wise backdoor sensitivity via a Lipschitz constant and, explicitly, eliminate the backdoor infection by overwriting the backdoor-related channel kernels with random initialization. Furthermore, we propose to employ an auxiliary model with a full source model to ensure accurate pseudo-labeling, taking into account the controllable, clean target training data in OSUDA. We validate our framework using a multi-center, multi-vendor, and multi-disease (M&M) cardiac dataset. Our findings suggest that the target model is susceptible to backdoor attacks during OSUDA, and our defense mechanism effectively mitigates the infection of target domain victims.</p>","PeriodicalId":74566,"journal":{"name":"Proceedings. IEEE International Symposium on Biomedical Imaging","volume":"2024 ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11483644/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142482683","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-05-01Epub Date: 2024-08-22DOI: 10.1109/isbi56570.2024.10635204
Gianfranco Cortés, Yue Yu, Robin Chen, Melissa Armstrong, David Vaillancourt, Baba C Vemuri
Diffusion MRI (dMRI) has shown significant promise in capturing subtle changes in neural microstructure caused by neurodegenerative disorders. In this paper, we propose a novel end-to-end compound architecture for processing raw dMRI data. It consists of a 3D convolutional kernel network (CKN) that extracts macro-architectural features across voxels and a gauge equivariant Volterra network (GEVNet) on the sphere that extracts micro-architectural features from within voxels. The use of higher order convolutions enables our architecture to model spatially extended nonlinear interactions across the applied diffusion-sensitizing magnetic field gradients. The compound network is globally equivariant to 3D translations and locally equivariant to 3D rotations. We demonstrate the efficacy of our model on the classification of neurodegenerative disorders.
{"title":"HIGHER ORDER GAUGE EQUIVARIANT CONVOLUTIONS FOR NEURODEGENERATIVE DISORDER CLASSIFICATION.","authors":"Gianfranco Cortés, Yue Yu, Robin Chen, Melissa Armstrong, David Vaillancourt, Baba C Vemuri","doi":"10.1109/isbi56570.2024.10635204","DOIUrl":"10.1109/isbi56570.2024.10635204","url":null,"abstract":"<p><p>Diffusion MRI (dMRI) has shown significant promise in capturing subtle changes in neural microstructure caused by neurodegenerative disorders. In this paper, we propose a novel end-to-end compound architecture for processing raw dMRI data. It consists of a 3D convolutional kernel network (CKN) that extracts macro-architectural features across voxels and a gauge equivariant Volterra network (GEVNet) on the sphere that extracts micro-architectural features from within voxels. The use of higher order convolutions enables our architecture to model spatially extended nonlinear interactions across the applied diffusion-sensitizing magnetic field gradients. The compound network is globally equivariant to 3D translations and locally equivariant to 3D rotations. We demonstrate the efficacy of our model on the classification of neurodegenerative disorders.</p>","PeriodicalId":74566,"journal":{"name":"Proceedings. IEEE International Symposium on Biomedical Imaging","volume":"2024 ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11610404/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142775550","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-05-01Epub Date: 2024-08-22DOI: 10.1109/isbi56570.2024.10635881
Qisheng He, Nicholas Summerfield, Ming Dong, Carri Glide-Hurst
In medical image segmentation, although multi-modality training is possible, clinical translation is challenged by the limited availability of all image types for a given patient. Different from typical segmentation models, modality-agnostic (MAG) learning trains a single model based on all available modalities but remains input-agnostic, allowing a single model to produce accurate segmentation given any modality combinations. In this paper, we propose a novel frame-work, MAG learning through Multi-modality Self-distillation (MAG-MS), for medical image segmentation. MAG-MS distills knowledge from the fusion of multiple modalities and applies it to enhance representation learning for individual modalities. This makes it an adaptable and efficient solution for handling limited modalities during testing scenarios. Our extensive experiments on benchmark datasets demonstrate its superior segmentation accuracy, MAG robustness, and efficiency than the current state-of-the-art methods.
{"title":"MODALITY-AGNOSTIC LEARNING FOR MEDICAL IMAGE SEGMENTATION USING MULTI-MODALITY SELF-DISTILLATION.","authors":"Qisheng He, Nicholas Summerfield, Ming Dong, Carri Glide-Hurst","doi":"10.1109/isbi56570.2024.10635881","DOIUrl":"10.1109/isbi56570.2024.10635881","url":null,"abstract":"<p><p>In medical image segmentation, although multi-modality training is possible, clinical translation is challenged by the limited availability of all image types for a given patient. Different from typical segmentation models, modality-agnostic (MAG) learning trains a single model based on all available modalities but remains input-agnostic, allowing a single model to produce accurate segmentation given any modality combinations. In this paper, we propose a novel frame-work, MAG learning through Multi-modality Self-distillation (MAG-MS), for medical image segmentation. MAG-MS distills knowledge from the fusion of multiple modalities and applies it to enhance representation learning for individual modalities. This makes it an adaptable and efficient solution for handling limited modalities during testing scenarios. Our extensive experiments on benchmark datasets demonstrate its superior segmentation accuracy, MAG robustness, and efficiency than the current state-of-the-art methods.</p>","PeriodicalId":74566,"journal":{"name":"Proceedings. IEEE International Symposium on Biomedical Imaging","volume":"2024 ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11673955/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142904143","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-05-01Epub Date: 2024-08-22DOI: 10.1109/ISBI56570.2024.10635723
Yuexi Du, Regina J Hooley, John Lewin, Nicha C Dvornek
Digital Breast Tomosynthesis (DBT) is a widely used medical imaging modality for breast cancer screening and diagnosis, offering higher spatial resolution and greater detail through its 3D-like breast volume imaging capability. However, the increased data volume also introduces pronounced data imbalance challenges, where only a small fraction of the volume contains suspicious tissue. This further exacerbates the data imbalance due to the case-level distribution in real-world data and leads to learning a trivial classification model that only predicts the majority class. To address this, we propose a novel method using view-level contrastive Self-supervised Initialization and Fine-Tuning for identifying abnormal DBT images, namely SIFT-DBT. We further introduce a patch-level multi-instance learning method to preserve spatial resolution. The proposed method achieves 92.69% volume-wise AUC on an evaluation of 970 unique studies.
{"title":"SIFT-DBT: SELF-SUPERVISED INITIALIZATION AND FINE-TUNING FOR IMBALANCED DIGITAL BREAST TOMOSYNTHESIS IMAGE CLASSIFICATION.","authors":"Yuexi Du, Regina J Hooley, John Lewin, Nicha C Dvornek","doi":"10.1109/ISBI56570.2024.10635723","DOIUrl":"10.1109/ISBI56570.2024.10635723","url":null,"abstract":"<p><p>Digital Breast Tomosynthesis (DBT) is a widely used medical imaging modality for breast cancer screening and diagnosis, offering higher spatial resolution and greater detail through its 3D-like breast volume imaging capability. However, the increased data volume also introduces pronounced data imbalance challenges, where only a small fraction of the volume contains suspicious tissue. This further exacerbates the data imbalance due to the case-level distribution in real-world data and leads to learning a trivial classification model that only predicts the majority class. To address this, we propose a novel method using view-level contrastive <b>S</b>elf-supervised <b>I</b>nitialization and <b>F</b>ine-<b>T</b>uning for identifying abnormal <b>DBT</b> images, namely <b>SIFT-DBT</b>. We further introduce a patch-level multi-instance learning method to preserve spatial resolution. The proposed method achieves 92.69% volume-wise AUC on an evaluation of 970 unique studies.</p>","PeriodicalId":74566,"journal":{"name":"Proceedings. IEEE International Symposium on Biomedical Imaging","volume":"2024 ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11386909/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142302996","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-05-01Epub Date: 2024-08-22DOI: 10.1109/isbi56570.2024.10635851
Nishchal Sapkota, Yejia Zhang, Susan M Motch Perrine, Yuhan Hsi, Sirui Li, Meng Wu, Greg Holmes, Abdul R Abdulai, Ethylin W Jabs, Joan T Richtsmeier, Danny Z Chen
Studying the morphological development of cartilaginous and osseous structures is critical to the early detection of life-threatening skeletal dysmorphology. Embryonic cartilage undergoes rapid structural changes within hours, introducing biological variations and morphological shifts that limit the generalization of deep learning-based segmentation models that infer across multiple embryonic age groups. Obtaining individual models for each age group is expensive and less effective, while direct transfer (predicting an age unseen during training) suffers a potential performance drop due to morphological shifts. We propose a novel Transformer-based segmentation model with improved biological priors that better distills morphologically diverse information through conditional mechanisms. This enables a single model to accurately predict cartilage across multiple age groups. Experiments on the mice cartilage dataset show the superiority of our new model compared to other competitive segmentation models. Additional studies on a separate mice cartilage dataset with a distinct mutation show that our model generalizes well and effectively captures age-based cartilage morphology patterns. Code is available in GitHub.
{"title":"CONUNETR: A CONDITIONAL TRANSFORMER NETWORK FOR 3D MICRO-CT EMBRYONIC CARTILAGE SEGMENTATION.","authors":"Nishchal Sapkota, Yejia Zhang, Susan M Motch Perrine, Yuhan Hsi, Sirui Li, Meng Wu, Greg Holmes, Abdul R Abdulai, Ethylin W Jabs, Joan T Richtsmeier, Danny Z Chen","doi":"10.1109/isbi56570.2024.10635851","DOIUrl":"10.1109/isbi56570.2024.10635851","url":null,"abstract":"<p><p>Studying the morphological development of cartilaginous and osseous structures is critical to the early detection of life-threatening skeletal dysmorphology. Embryonic cartilage undergoes rapid structural changes within hours, introducing biological variations and morphological shifts that limit the generalization of deep learning-based segmentation models that infer across multiple embryonic age groups. Obtaining individual models for each age group is expensive and less effective, while direct transfer (predicting an age unseen during training) suffers a potential performance drop due to morphological shifts. We propose a novel Transformer-based segmentation model with improved biological priors that better distills morphologically diverse information through conditional mechanisms. This enables a single model to accurately predict cartilage across multiple age groups. Experiments on the mice cartilage dataset show the superiority of our new model compared to other competitive segmentation models. Additional studies on a separate mice cartilage dataset with a distinct mutation show that our model generalizes well and effectively captures age-based cartilage morphology patterns. Code is available in GitHub.</p>","PeriodicalId":74566,"journal":{"name":"Proceedings. IEEE International Symposium on Biomedical Imaging","volume":"2024 ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12476089/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145187530","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-04-01Epub Date: 2023-09-01DOI: 10.1109/isbi53787.2023.10230534
Jiong Wu, Yong Fan
Convolutional neural networks (CNNs) have been widely used to build deep learning models for medical image registration, but manually designed network architectures are not necessarily optimal. This paper presents a hierarchical NAS framework (HNAS-Reg), consisting of both convolutional operation search and network topology search, to identify the optimal network architecture for deformable medical image registration. To mitigate the computational overhead and memory constraints, a partial channel strategy is utilized without losing optimization quality. Experiments on three datasets, consisting of 636 T1-weighted magnetic resonance images (MRIs), have demonstrated that the proposal method can build a deep learning model with improved image registration accuracy and reduced model size, compared with state-of-the-art image registration approaches, including one representative traditional approach and two unsupervised learning-based approaches.
{"title":"HNAS-Reg: Hierarchical Neural Architecture Search for Deformable Medical Image Registration.","authors":"Jiong Wu, Yong Fan","doi":"10.1109/isbi53787.2023.10230534","DOIUrl":"10.1109/isbi53787.2023.10230534","url":null,"abstract":"<p><p>Convolutional neural networks (CNNs) have been widely used to build deep learning models for medical image registration, but manually designed network architectures are not necessarily optimal. This paper presents a hierarchical NAS framework (HNAS-Reg), consisting of both convolutional operation search and network topology search, to identify the optimal network architecture for deformable medical image registration. To mitigate the computational overhead and memory constraints, a partial channel strategy is utilized without losing optimization quality. Experiments on three datasets, consisting of 636 T1-weighted magnetic resonance images (MRIs), have demonstrated that the proposal method can build a deep learning model with improved image registration accuracy and reduced model size, compared with state-of-the-art image registration approaches, including one representative traditional approach and two unsupervised learning-based approaches.</p>","PeriodicalId":74566,"journal":{"name":"Proceedings. IEEE International Symposium on Biomedical Imaging","volume":"2023 ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2023-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10544790/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"41172564","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-04-01Epub Date: 2023-09-01DOI: 10.1109/isbi53787.2023.10230597
Zhengbo Zhou, Jun Luo, Dooman Arefan, Gene Kitamura, Shandong Wu
Curriculum learning is a learning method that trains models in a meaningful order from easier to harder samples. A key here is to devise automatic and objective difficulty measures of samples. In the medical domain, previous work applied domain knowledge from human experts to qualitatively assess classification difficulty of medical images to guide curriculum learning, which requires extra annotation efforts, relies on subjective human experience, and may introduce bias. In this work, we propose a new automated curriculum learning technique using the variance of gradients (VoG) to compute an objective difficulty measure of samples and evaluated its effects on elbow fracture classification from X-ray images. Specifically, we used VoG as a metric to rank each sample in terms of the classification difficulty, where high VoG scores indicate more difficult cases for classification, to guide the curriculum training process We compared the proposed technique to a baseline (without curriculum learning), a previous method that used human annotations on classification difficulty, and anti-curriculum learning. Our experiment results showed comparable and higher performance for the binary and multi-class bone fracture classification tasks.
{"title":"Human Not in the Loop: Objective Sample Difficulty Measures for Curriculum Learning.","authors":"Zhengbo Zhou, Jun Luo, Dooman Arefan, Gene Kitamura, Shandong Wu","doi":"10.1109/isbi53787.2023.10230597","DOIUrl":"10.1109/isbi53787.2023.10230597","url":null,"abstract":"<p><p>Curriculum learning is a learning method that trains models in a meaningful order from easier to harder samples. A key here is to devise automatic and objective difficulty measures of samples. In the medical domain, previous work applied domain knowledge from human experts to qualitatively assess classification difficulty of medical images to guide curriculum learning, which requires extra annotation efforts, relies on subjective human experience, and may introduce bias. In this work, we propose a new automated curriculum learning technique using the variance of gradients (VoG) to compute an objective difficulty measure of samples and evaluated its effects on elbow fracture classification from X-ray images. Specifically, we used VoG as a metric to rank each sample in terms of the classification difficulty, where high VoG scores indicate more difficult cases for classification, to guide the curriculum training process We compared the proposed technique to a baseline (without curriculum learning), a previous method that used human annotations on classification difficulty, and anti-curriculum learning. Our experiment results showed comparable and higher performance for the binary and multi-class bone fracture classification tasks.</p>","PeriodicalId":74566,"journal":{"name":"Proceedings. IEEE International Symposium on Biomedical Imaging","volume":"2023 ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2023-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10602195/pdf/nihms-1891600.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"54232739","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
We develop deep clustering survival machines to simultaneously predict survival information and characterize data heterogeneity that is not typically modeled by conventional survival analysis methods. By modeling timing information of survival data generatively with a mixture of parametric distributions, referred to as expert distributions, our method learns weights of the expert distributions for individual instances based on their features discriminatively such that each instance's survival information can be characterized by a weighted combination of the learned expert distributions. Extensive experiments on both real and synthetic datasets have demonstrated that our method is capable of obtaining promising clustering results and competitive time-to-event predicting performance.
{"title":"Deep Clustering Survival Machines with Interpretable Expert Distributions.","authors":"Bojian Hou, Hongming Li, Zhicheng Jiao, Zhen Zhou, Hao Zheng, Yong Fan","doi":"10.1109/isbi53787.2023.10230844","DOIUrl":"10.1109/isbi53787.2023.10230844","url":null,"abstract":"<p><p>We develop deep clustering survival machines to simultaneously predict survival information and characterize data heterogeneity that is not typically modeled by conventional survival analysis methods. By modeling timing information of survival data <i>generatively</i> with a mixture of parametric distributions, referred to as expert distributions, our method learns weights of the expert distributions for individual instances based on their features <i>discriminatively</i> such that each instance's survival information can be characterized by a weighted combination of the learned expert distributions. Extensive experiments on both real and synthetic datasets have demonstrated that our method is capable of obtaining promising clustering results and competitive time-to-event predicting performance.</p>","PeriodicalId":74566,"journal":{"name":"Proceedings. IEEE International Symposium on Biomedical Imaging","volume":"2023 ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2023-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10544788/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"41167287","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}