Zhuotong Cai, Tianyi Zeng, Eléonore V Lieffrig, Jiazhen Zhang, Fuyao Chen, Takuya Toyonaga, Chenyu You, Jingmin Xin, Nanning Zheng, Yihuan Lu, James S Duncan, John A Onofrey
{"title":"Cross-Attention for Improved Motion Correction in Brain PET.","authors":"Zhuotong Cai, Tianyi Zeng, Eléonore V Lieffrig, Jiazhen Zhang, Fuyao Chen, Takuya Toyonaga, Chenyu You, Jingmin Xin, Nanning Zheng, Yihuan Lu, James S Duncan, John A Onofrey","doi":"10.1007/978-3-031-44858-4_4","DOIUrl":null,"url":null,"abstract":"<p><p>Head movement during long scan sessions degrades the quality of reconstruction in positron emission tomography (PET) and introduces artifacts, which limits clinical diagnosis and treatment. Recent deep learning-based motion correction work utilized raw PET list-mode data and hardware motion tracking (HMT) to learn head motion in a supervised manner. However, motion prediction results were not robust to testing subjects outside the training data domain. In this paper, we integrate a cross-attention mechanism into the supervised deep learning network to improve motion correction across test subjects. Specifically, cross-attention learns the spatial correspondence between the reference images and moving images to explicitly focus the model on the most correlative inherent information - the head region the motion correction. We validate our approach on brain PET data from two different scanners: HRRT without time of flight (ToF) and mCT with ToF. Compared with traditional and deep learning benchmarks, our network improved the performance of motion correction by 58% and 26% in translation and rotation, respectively, in multi-subject testing in HRRT studies. In mCT studies, our approach improved performance by 66% and 64% for translation and rotation, respectively. Our results demonstrate that cross-attention has the potential to improve the quality of brain PET image reconstruction without the dependence on HMT. All code will be released on GitHub: https://github.com/OnofreyLab/dl_hmc_attention_mlcn2023.</p>","PeriodicalId":510900,"journal":{"name":"Machine learning in clinical neuroimaging : 6th international workshop, MLCN 2023, held in conjunction with MICCAI 2023, Vancouver, BC, Canada, October 8, 2023, proceedings. MLCN (Workshop) (6th : 2023 : Vancouver, B.C.)","volume":"14312 ","pages":"34-45"},"PeriodicalIF":0.0000,"publicationDate":"2023-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10758996/pdf/","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Machine learning in clinical neuroimaging : 6th international workshop, MLCN 2023, held in conjunction with MICCAI 2023, Vancouver, BC, Canada, October 8, 2023, proceedings. MLCN (Workshop) (6th : 2023 : Vancouver, B.C.)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1007/978-3-031-44858-4_4","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Head movement during long scan sessions degrades the quality of reconstruction in positron emission tomography (PET) and introduces artifacts, which limits clinical diagnosis and treatment. Recent deep learning-based motion correction work utilized raw PET list-mode data and hardware motion tracking (HMT) to learn head motion in a supervised manner. However, motion prediction results were not robust to testing subjects outside the training data domain. In this paper, we integrate a cross-attention mechanism into the supervised deep learning network to improve motion correction across test subjects. Specifically, cross-attention learns the spatial correspondence between the reference images and moving images to explicitly focus the model on the most correlative inherent information - the head region the motion correction. We validate our approach on brain PET data from two different scanners: HRRT without time of flight (ToF) and mCT with ToF. Compared with traditional and deep learning benchmarks, our network improved the performance of motion correction by 58% and 26% in translation and rotation, respectively, in multi-subject testing in HRRT studies. In mCT studies, our approach improved performance by 66% and 64% for translation and rotation, respectively. Our results demonstrate that cross-attention has the potential to improve the quality of brain PET image reconstruction without the dependence on HMT. All code will be released on GitHub: https://github.com/OnofreyLab/dl_hmc_attention_mlcn2023.