{"title":"Improving reliability of movement assessment in Parkinson's disease using computer vision-based automated severity estimation.","authors":"Jinyu Xu, Xin Xu, Xudong Guo, Zezhi Li, Boya Dong, Chen Qi, Chunhui Yang, Dong Zhou, Jiali Wang, Lu Song, Ping He, Shanshan Kong, Shuchang Zheng, Sichao Fu, Wei Xie, Xuan Liu, Ya Cao, Yilin Liu, Yiqing Qiu, Zhiyuan Zheng, Fei Yang, Jing Gan, Xi Wu","doi":"10.1177/1877718X241312605","DOIUrl":null,"url":null,"abstract":"<p><strong>Background: </strong>Clinical assessments of motor symptoms rely on observations and subjective judgments against standardized scales, leading to variability due to confounders. Improving inter-rater agreement is essential for effective disease management.</p><p><strong>Objective: </strong>We developed an objective rating system for Parkinson's disease (PD) that integrates computer vision (CV) and machine learning to correct potential discrepancies among raters while providing the basis for model performance to gain professional acceptance.</p><p><strong>Methods: </strong>A prospective PD cohort (n = 128) were recruited from multi-centers. Motor examination videos were recorded using an android tablet with CV-based software following the MDS-UPDRS Part-III instructions. Videos included facial, upper- and lower-limb movements, arising from a chair, standing, and walking. Fifteen certified clinicians were recruited from multi-centers. For each video, five clinicians were randomly selected to independently rate the severity of motor symptoms, validate the videos and movement variables (MovVars). Machine learning algorithms were applied for automated rating and feature importance analysis. Inter-rater agreement among human raters and the agreement between artificial intelligence (AI)-generated ratings and expert consensus were calculated.</p><p><strong>Results: </strong>For all validated videos (n = 1024), AI-based ratings showed an average absolute accuracy of 69.63% and an average acceptable accuracy of 98.78% against the clinician consensus. The mean absolute error between the AI-based scores and clinician consensus was 0.32, outperforming the inter-rater variability (0.65), potentially due to the combined utilization of diverse MovVars.</p><p><strong>Conclusions: </strong>The algorithm enabled accurate video-based evaluation of mild motor symptom severity. AI-assisted assessment improved the inter-rater agreement, demonstrating the practical value of CV-based tools in screening, diagnosing, and treating movement disorders.</p>","PeriodicalId":16660,"journal":{"name":"Journal of Parkinson's disease","volume":" ","pages":"1877718X241312605"},"PeriodicalIF":4.0000,"publicationDate":"2025-02-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Parkinson's disease","FirstCategoryId":"3","ListUrlMain":"https://doi.org/10.1177/1877718X241312605","RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"NEUROSCIENCES","Score":null,"Total":0}
引用次数: 0
Abstract
Background: Clinical assessments of motor symptoms rely on observations and subjective judgments against standardized scales, leading to variability due to confounders. Improving inter-rater agreement is essential for effective disease management.
Objective: We developed an objective rating system for Parkinson's disease (PD) that integrates computer vision (CV) and machine learning to correct potential discrepancies among raters while providing the basis for model performance to gain professional acceptance.
Methods: A prospective PD cohort (n = 128) were recruited from multi-centers. Motor examination videos were recorded using an android tablet with CV-based software following the MDS-UPDRS Part-III instructions. Videos included facial, upper- and lower-limb movements, arising from a chair, standing, and walking. Fifteen certified clinicians were recruited from multi-centers. For each video, five clinicians were randomly selected to independently rate the severity of motor symptoms, validate the videos and movement variables (MovVars). Machine learning algorithms were applied for automated rating and feature importance analysis. Inter-rater agreement among human raters and the agreement between artificial intelligence (AI)-generated ratings and expert consensus were calculated.
Results: For all validated videos (n = 1024), AI-based ratings showed an average absolute accuracy of 69.63% and an average acceptable accuracy of 98.78% against the clinician consensus. The mean absolute error between the AI-based scores and clinician consensus was 0.32, outperforming the inter-rater variability (0.65), potentially due to the combined utilization of diverse MovVars.
Conclusions: The algorithm enabled accurate video-based evaluation of mild motor symptom severity. AI-assisted assessment improved the inter-rater agreement, demonstrating the practical value of CV-based tools in screening, diagnosing, and treating movement disorders.
期刊介绍:
The Journal of Parkinson''s Disease (JPD) publishes original research in basic science, translational research and clinical medicine in Parkinson’s disease in cooperation with the Journal of Alzheimer''s Disease. It features a first class Editorial Board and provides rigorous peer review and rapid online publication.