Andy Huang, Kyle Hall, C. Watson, Seyed Reza Shahamiri
{"title":"语言障碍说话者可理解性自动评估综述","authors":"Andy Huang, Kyle Hall, C. Watson, Seyed Reza Shahamiri","doi":"10.1109/sped53181.2021.9587400","DOIUrl":null,"url":null,"abstract":"Automated dysarthria intelligibility assessment offers the opportunity to develop reliable, low-cost, and scalable tools, which help to solve current shortcomings of manual and subjective intelligibility assessments. This paper reviews the literature regarding automated intelligibility assessment, identifying the highest performing published models and concluding on promising avenues for further research. Our review shows that most of the existing work were able to achieve very high accuracies. However, we have found that most of these studies validated their models using speech samples of the same speakers used in training, making their results less generalizable. Furthermore, there is a lack of study on how well these models perform on speakers from different datasets or different microphone setups. This lack of generalizability has implications to the real-life application of these models. Future research directions could include the use of more robust methods of validation such as using unseen speakers, as well as incorporating speakers from different datasets. This would provide confidence that the models are generalized and therefore allow them to be used in real-world clinical practice.","PeriodicalId":193702,"journal":{"name":"2021 International Conference on Speech Technology and Human-Computer Dialogue (SpeD)","volume":"85 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"4","resultStr":"{\"title\":\"A Review of Automated Intelligibility Assessment for Dysarthric Speakers\",\"authors\":\"Andy Huang, Kyle Hall, C. Watson, Seyed Reza Shahamiri\",\"doi\":\"10.1109/sped53181.2021.9587400\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Automated dysarthria intelligibility assessment offers the opportunity to develop reliable, low-cost, and scalable tools, which help to solve current shortcomings of manual and subjective intelligibility assessments. This paper reviews the literature regarding automated intelligibility assessment, identifying the highest performing published models and concluding on promising avenues for further research. Our review shows that most of the existing work were able to achieve very high accuracies. However, we have found that most of these studies validated their models using speech samples of the same speakers used in training, making their results less generalizable. Furthermore, there is a lack of study on how well these models perform on speakers from different datasets or different microphone setups. This lack of generalizability has implications to the real-life application of these models. Future research directions could include the use of more robust methods of validation such as using unseen speakers, as well as incorporating speakers from different datasets. This would provide confidence that the models are generalized and therefore allow them to be used in real-world clinical practice.\",\"PeriodicalId\":193702,\"journal\":{\"name\":\"2021 International Conference on Speech Technology and Human-Computer Dialogue (SpeD)\",\"volume\":\"85 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2021-10-13\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"4\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2021 International Conference on Speech Technology and Human-Computer Dialogue (SpeD)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/sped53181.2021.9587400\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 International Conference on Speech Technology and Human-Computer Dialogue (SpeD)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/sped53181.2021.9587400","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
A Review of Automated Intelligibility Assessment for Dysarthric Speakers
Automated dysarthria intelligibility assessment offers the opportunity to develop reliable, low-cost, and scalable tools, which help to solve current shortcomings of manual and subjective intelligibility assessments. This paper reviews the literature regarding automated intelligibility assessment, identifying the highest performing published models and concluding on promising avenues for further research. Our review shows that most of the existing work were able to achieve very high accuracies. However, we have found that most of these studies validated their models using speech samples of the same speakers used in training, making their results less generalizable. Furthermore, there is a lack of study on how well these models perform on speakers from different datasets or different microphone setups. This lack of generalizability has implications to the real-life application of these models. Future research directions could include the use of more robust methods of validation such as using unseen speakers, as well as incorporating speakers from different datasets. This would provide confidence that the models are generalized and therefore allow them to be used in real-world clinical practice.