Pub Date : 2024-09-26Epub Date: 2023-08-23DOI: 10.1044/2023_JSLHR-22-00622
Jonathan L Preston, Nicole F Caballero, Megan C Leece, Dongliang Wang, Benedette M Herbst, Nina R Benway
Purpose: This study examines how ultrasound biofeedback and intensive treatment distribution affect speech sound generalization during an evidence-based treatment, Speech Motor Chaining, for children with persisting speech errors associated with childhood apraxia of speech (CAS).
Method: In a 2 × 2 factorial randomized controlled trial, children ages 9-17 years meeting CAS criteria were randomized to receive (a) a distributed treatment (20 sessions twice weekly over 10 weeks) or intensive treatment (20 hr in 5 weeks, with 10 hr in Week 1) and (b) treatment with or without biofeedback. Due to the COVID pandemic, some participants were randomized to distributed/intensive telepractice treatment only. The primary outcome was percent target sounds correct on untreated phrases (i.e., generalization) at the 10-week time point. More than 50,000 narrow phonetic transcriptions were analyzed.
Results: Forty-eight participants completed treatment. Intensive treatment significantly increased generalization at all time points. The effect of biofeedback was significant at 5 weeks from the start of treatment but not significant at the primary 10-week time point. However, when comparing each group immediately after their 20 hr of treatment finished, generalization was significantly greater in intensive over distributed treatment and greater in ultrasound over no-ultrasound treatment (with a significant interaction favoring intensive treatment with ultrasound). Only the advantage of intensive treatment remained significant 5 weeks after groups finished treatment. There was no significant difference between face-to-face and telepractice modalities.
Conclusions: When the number of treatment hours is fixed, an intensive schedule of Speech Motor Chaining facilitated greater improvement than a distributed schedule. Ultrasound biofeedback initially accelerated learning, but the benefits may dissipate as treatment continues or after it ends.
{"title":"A Randomized Controlled Trial of Treatment Distribution and Biofeedback Effects on Speech Production in School-Age Children With Apraxia of Speech.","authors":"Jonathan L Preston, Nicole F Caballero, Megan C Leece, Dongliang Wang, Benedette M Herbst, Nina R Benway","doi":"10.1044/2023_JSLHR-22-00622","DOIUrl":"10.1044/2023_JSLHR-22-00622","url":null,"abstract":"<p><strong>Purpose: </strong>This study examines how ultrasound biofeedback and intensive treatment distribution affect speech sound generalization during an evidence-based treatment, Speech Motor Chaining, for children with persisting speech errors associated with childhood apraxia of speech (CAS).</p><p><strong>Method: </strong>In a 2 × 2 factorial randomized controlled trial, children ages 9-17 years meeting CAS criteria were randomized to receive (a) a distributed treatment (20 sessions twice weekly over 10 weeks) or intensive treatment (20 hr in 5 weeks, with 10 hr in Week 1) and (b) treatment with or without biofeedback. Due to the COVID pandemic, some participants were randomized to distributed/intensive telepractice treatment only. The primary outcome was percent target sounds correct on untreated phrases (i.e., generalization) at the 10-week time point. More than 50,000 narrow phonetic transcriptions were analyzed.</p><p><strong>Results: </strong>Forty-eight participants completed treatment. Intensive treatment significantly increased generalization at all time points. The effect of biofeedback was significant at 5 weeks from the start of treatment but not significant at the primary 10-week time point. However, when comparing each group immediately after their 20 hr of treatment finished, generalization was significantly greater in intensive over distributed treatment and greater in ultrasound over no-ultrasound treatment (with a significant interaction favoring intensive treatment with ultrasound). Only the advantage of intensive treatment remained significant 5 weeks after groups finished treatment. There was no significant difference between face-to-face and telepractice modalities.</p><p><strong>Conclusions: </strong>When the number of treatment hours is fixed, an intensive schedule of Speech Motor Chaining facilitated greater improvement than a distributed schedule. Ultrasound biofeedback initially accelerated learning, but the benefits may dissipate as treatment continues or after it ends.</p>","PeriodicalId":51254,"journal":{"name":"Journal of Speech Language and Hearing Research","volume":null,"pages":null},"PeriodicalIF":2.2,"publicationDate":"2024-09-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"10061459","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-26DOI: 10.1044/2024_JSLHR-24-00122
Mark Hasegawa-Johnson, Xiuwen Zheng, Heejin Kim, Clarion Mendes, Meg Dickinson, Erik Hege, Chris Zwilling, Marie Moore Channell, Laura Mattie, Heather Hodges, Lorraine Ramig, Mary Bellard, Mike Shebanek, Leda Sarι, Kaustubh Kalgaonkar, David Frerichs, Jeffrey P Bigham, Leah Findlater, Colin Lea, Sarah Herrlinger, Peter Korn, Shadi Abou-Zahra, Rus Heywood, Katrin Tomanek, Bob MacDonald
Purpose: The Speech Accessibility Project (SAP) intends to facilitate research and development in automatic speech recognition (ASR) and other machine learning tasks for people with speech disabilities. The purpose of this article is to introduce this project as a resource for researchers, including baseline analysis of the first released data package.
Method: The project aims to facilitate ASR research by collecting, curating, and distributing transcribed U.S. English speech from people with speech and/or language disabilities. Participants record speech from their place of residence by connecting their personal computer, cell phone, and assistive devices, if needed, to the SAP web portal. All samples are manually transcribed, and 30 per participant are annotated using differential diagnostic pattern dimensions. For purposes of ASR experiments, the participants have been randomly assigned to a training set, a development set for controlled testing of a trained ASR, and a test set to evaluate ASR error rate.
Results: The SAP 2023-10-05 Data Package contains the speech of 211 people with dysarthria as a correlate of Parkinson's disease, and the associated test set contains 42 additional speakers. A baseline ASR, with a word error rate of 3.4% for typical speakers, transcribes test speech with a word error rate of 36.3%. Fine-tuning reduces the word error rate to 23.7%.
Conclusions: Preliminary findings suggest that a large corpus of dysarthric and dysphonic speech has the potential to significantly improve speech technology for people with disabilities. By providing these data to researchers, the SAP intends to significantly accelerate research into accessible speech technology.
目的:语音无障碍项目(SAP)旨在促进自动语音识别(ASR)和其他机器学习任务的研究和开发,为语音残疾人士提供便利。本文旨在介绍该项目,将其作为研究人员的资源,包括对首次发布的数据包进行基线分析:该项目旨在通过收集、整理和发布语音和/或语言残障人士转录的美国英语语音来促进 ASR 研究。参与者将个人电脑、手机和辅助设备(如需要)连接到 SAP 门户网站,在居住地录制语音。所有样本均由人工转录,并使用差异诊断模式维度对每位参与者的 30 个样本进行注释。为了进行 ASR 实验,参与者被随机分配到一个训练集、一个用于对训练过的 ASR 进行控制测试的开发集和一个用于评估 ASR 错误率的测试集:SAP 2023-10-05 数据包包含 211 名与帕金森病相关的构音障碍患者的语音,相关测试集包含另外 42 名发言者的语音。基线 ASR 对典型说话者的词错误率为 3.4%,而转录测试语音的词错误率为 36.3%。微调后,词错误率降低到 23.7%:初步研究结果表明,庞大的发音障碍和发音困难语音语料库有可能极大地改进面向残疾人的语音技术。通过向研究人员提供这些数据,SAP 打算大大加快无障碍语音技术的研究。补充材料:https://doi.org/10.23641/asha.27078079。
{"title":"Community-Supported Shared Infrastructure in Support of Speech Accessibility.","authors":"Mark Hasegawa-Johnson, Xiuwen Zheng, Heejin Kim, Clarion Mendes, Meg Dickinson, Erik Hege, Chris Zwilling, Marie Moore Channell, Laura Mattie, Heather Hodges, Lorraine Ramig, Mary Bellard, Mike Shebanek, Leda Sarι, Kaustubh Kalgaonkar, David Frerichs, Jeffrey P Bigham, Leah Findlater, Colin Lea, Sarah Herrlinger, Peter Korn, Shadi Abou-Zahra, Rus Heywood, Katrin Tomanek, Bob MacDonald","doi":"10.1044/2024_JSLHR-24-00122","DOIUrl":"https://doi.org/10.1044/2024_JSLHR-24-00122","url":null,"abstract":"<p><strong>Purpose: </strong>The Speech Accessibility Project (SAP) intends to facilitate research and development in automatic speech recognition (ASR) and other machine learning tasks for people with speech disabilities. The purpose of this article is to introduce this project as a resource for researchers, including baseline analysis of the first released data package.</p><p><strong>Method: </strong>The project aims to facilitate ASR research by collecting, curating, and distributing transcribed U.S. English speech from people with speech and/or language disabilities. Participants record speech from their place of residence by connecting their personal computer, cell phone, and assistive devices, if needed, to the SAP web portal. All samples are manually transcribed, and 30 per participant are annotated using differential diagnostic pattern dimensions. For purposes of ASR experiments, the participants have been randomly assigned to a training set, a development set for controlled testing of a trained ASR, and a test set to evaluate ASR error rate.</p><p><strong>Results: </strong>The SAP 2023-10-05 Data Package contains the speech of 211 people with dysarthria as a correlate of Parkinson's disease, and the associated test set contains 42 additional speakers. A baseline ASR, with a word error rate of 3.4% for typical speakers, transcribes test speech with a word error rate of 36.3%. Fine-tuning reduces the word error rate to 23.7%.</p><p><strong>Conclusions: </strong>Preliminary findings suggest that a large corpus of dysarthric and dysphonic speech has the potential to significantly improve speech technology for people with disabilities. By providing these data to researchers, the SAP intends to significantly accelerate research into accessible speech technology.</p><p><strong>Supplemental material: </strong>https://doi.org/10.23641/asha.27078079.</p>","PeriodicalId":51254,"journal":{"name":"Journal of Speech Language and Hearing Research","volume":null,"pages":null},"PeriodicalIF":2.2,"publicationDate":"2024-09-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142332056","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-25DOI: 10.1044/2024_JSLHR-23-00727
Jarrad H Van Stan, Robert E Hillman, Carol Krusemark, Jason Muise, Tara Stadelman-Cohen, Daryush D Mehta, Dagmar Sternad
Purpose: Floating ball voice therapy (FBVT) is a voice-controlled virtual environment based on a common treatment component across multiple evidence-based therapies: improved vocal efficiency (target) via practicing voicing with modified resonance and airflow (ingredient). This study preliminarily tested FBVT's effects on outcomes and the potential for its novel variability metrics to predict individual patient generalization.
Method: Ten patients with nonphonotraumatic vocal hyperfunction (NPVH) practiced FBVT for 10 days. Outcomes were assessed by a vocal efficiency ratio, a validated NPVH index, the patient-reported Voice-Related Quality of Life (V-RQOL), and forced-choice auditory judgments of overall severity. Exploration in early practice (Day 1) was estimated by how the patient's two-dimensional variability (mean airflow and intensity) related to error (difference between the patient-produced and normative vocal efficiency ratio). Generalization from the game to spontaneous speech was evaluated using the validated NPVH index.
Results: Ten days of FBVT were associated with improved vocal efficiency (Cohen's d = 1.3), NPVH index (d = -1.1), V-RQOL total score (d = 0.9), and overall severity (odds ratio = 2.5). Patients who generalized on Day 10 exhibited airflow/intensity exploration that was more aligned with the error gradient on Day 1 (d = 0.6-1.2).
Conclusions: A relatively small dosage of FBVT (i.e., 10 practice sessions) was associated with multiple improved voice therapy outcomes. The FBVT variability metrics on Practice Day 1 demonstrated strong potential to predict which patients generalized to connected speech. Future work can more thoroughly evaluate effects on outcomes and characterizing the quality of vocal exploration with a larger patient population.
{"title":"Floating Ball Voice Therapy: Preliminary Effects on Outcomes and Predicting Individual Patient Differences in Generalization.","authors":"Jarrad H Van Stan, Robert E Hillman, Carol Krusemark, Jason Muise, Tara Stadelman-Cohen, Daryush D Mehta, Dagmar Sternad","doi":"10.1044/2024_JSLHR-23-00727","DOIUrl":"https://doi.org/10.1044/2024_JSLHR-23-00727","url":null,"abstract":"<p><strong>Purpose: </strong>Floating ball voice therapy (FBVT) is a voice-controlled virtual environment based on a common treatment component across multiple evidence-based therapies: improved vocal efficiency (target) via practicing voicing with modified resonance and airflow (ingredient). This study preliminarily tested FBVT's effects on outcomes and the potential for its novel variability metrics to predict individual patient generalization.</p><p><strong>Method: </strong>Ten patients with nonphonotraumatic vocal hyperfunction (NPVH) practiced FBVT for 10 days. Outcomes were assessed by a vocal efficiency ratio, a validated NPVH index, the patient-reported Voice-Related Quality of Life (V-RQOL), and forced-choice auditory judgments of overall severity. Exploration in early practice (Day 1) was estimated by how the patient's two-dimensional variability (mean airflow and intensity) related to error (difference between the patient-produced and normative vocal efficiency ratio). Generalization from the game to spontaneous speech was evaluated using the validated NPVH index.</p><p><strong>Results: </strong>Ten days of FBVT were associated with improved vocal efficiency (Cohen's <i>d</i> = 1.3), NPVH index (<i>d</i> = -1.1), V-RQOL total score (<i>d</i> = 0.9), and overall severity (odds ratio = 2.5). Patients who generalized on Day 10 exhibited airflow/intensity exploration that was more aligned with the error gradient on Day 1 (<i>d</i> = 0.6-1.2).</p><p><strong>Conclusions: </strong>A relatively small dosage of FBVT (i.e., 10 practice sessions) was associated with multiple improved voice therapy outcomes. The FBVT variability metrics on Practice Day 1 demonstrated strong potential to predict which patients generalized to connected speech. Future work can more thoroughly evaluate effects on outcomes and characterizing the quality of vocal exploration with a larger patient population.</p><p><strong>Supplemental material: </strong>https://doi.org/10.23641/asha.27040873.</p>","PeriodicalId":51254,"journal":{"name":"Journal of Speech Language and Hearing Research","volume":null,"pages":null},"PeriodicalIF":2.2,"publicationDate":"2024-09-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142331984","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-25DOI: 10.1044/2024_JSLHR-23-00655
Emilia C Lew, Anastasia Sares, Annie C Gilbert, Yue Zhang, Alexandre Lehmann, Mickael Deroche
Purpose: Greater recognition of the impact of hearing loss on cognitive functions has led speech/hearing clinics to focus more on auditory memory outcomes. Typically evaluated by scoring participants' recall on a list of unrelated words after they have heard the list read out loud, this method implies pitch and timing variations across words. Here, we questioned whether these variations could impact performance differentially in one language or another.
Method: In a series of online studies evaluating auditory short-term memory in normally hearing adults, we examined how pitch patterns (Experiment 1), timing patterns (Experiment 2), and interactions between the two (Experiment 3) affected free recall of words, cued recall of forgotten words, and mental demand. Note that visual memory was never directly tested; written words were only used after auditory encoding in the cued recall part. Studies were administered in both French and English, always conducted with native listeners.
Result: Confirming prior work, grouping mechanisms facilitated free recall, but not cued recall (the latter being only affected by longer presentation time) or ratings of mental demand. Critically, grouping by pitch provided more benefit for French than for English listeners, while grouping by time was equally beneficial in both languages.
Conclusion: Pitch is more useful to French- than to English-speaking listeners for encoding spoken words in short-term memory, perhaps due to the syllable-based versus stress-based rhythms inherent to each language.
{"title":"Differences Between French and English in the Use of Suprasegmental Cues for the Short-Term Recall of Word Lists.","authors":"Emilia C Lew, Anastasia Sares, Annie C Gilbert, Yue Zhang, Alexandre Lehmann, Mickael Deroche","doi":"10.1044/2024_JSLHR-23-00655","DOIUrl":"https://doi.org/10.1044/2024_JSLHR-23-00655","url":null,"abstract":"<p><strong>Purpose: </strong>Greater recognition of the impact of hearing loss on cognitive functions has led speech/hearing clinics to focus more on auditory memory outcomes. Typically evaluated by scoring participants' recall on a list of unrelated words after they have heard the list read out loud, this method implies pitch and timing variations across words. Here, we questioned whether these variations could impact performance differentially in one language or another.</p><p><strong>Method: </strong>In a series of online studies evaluating auditory short-term memory in normally hearing adults, we examined how pitch patterns (Experiment 1), timing patterns (Experiment 2), and interactions between the two (Experiment 3) affected free recall of words, cued recall of forgotten words, and mental demand. Note that visual memory was never directly tested; written words were only used after auditory encoding in the cued recall part. Studies were administered in both French and English, always conducted with native listeners.</p><p><strong>Result: </strong>Confirming prior work, grouping mechanisms facilitated free recall, but not cued recall (the latter being only affected by longer presentation time) or ratings of mental demand. Critically, grouping by pitch provided more benefit for French than for English listeners, while grouping by time was equally beneficial in both languages.</p><p><strong>Conclusion: </strong>Pitch is more useful to French- than to English-speaking listeners for encoding spoken words in short-term memory, perhaps due to the syllable-based versus stress-based rhythms inherent to each language.</p><p><strong>Supplemental material: </strong>https://doi.org/10.23641/asha.27048328.</p>","PeriodicalId":51254,"journal":{"name":"Journal of Speech Language and Hearing Research","volume":null,"pages":null},"PeriodicalIF":2.2,"publicationDate":"2024-09-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142331983","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-23DOI: 10.1044/2024_JSLHR-24-00247
Nina Bausek, Robert J Arnold
This Letter to the Editor was created in response to the article titled, "Instrumental Assessment of Aero-Resistive Expiratory Muscle Strength Rehabilitation Devices" by Dietsch et al. (2024). The article aims to compare six expiratory muscle strength training (EMST) devices and investigates their minimum trigger pressure, variability across the settings, and stability. The models tested include five positive expiratory pressure (PEP) devices, Acapella, TheraPEP, Threshold PEP, EMST75, and EMST150. It also includes a combined inspiratory and expiratory muscle training device, the Breather. We have several concerns about the integrity of results presented regarding the Breather as presented in the article. These include the heterogeneity of device sample used, as well as inadequate methodology and the experimental setup.
{"title":"Letter to the Editor Regarding \"Instrumental Assessment of Aero-Resistive Expiratory Muscle Strength Rehabilitation Devices\".","authors":"Nina Bausek, Robert J Arnold","doi":"10.1044/2024_JSLHR-24-00247","DOIUrl":"https://doi.org/10.1044/2024_JSLHR-24-00247","url":null,"abstract":"<p><p>This Letter to the Editor was created in response to the article titled, \"Instrumental Assessment of Aero-Resistive Expiratory Muscle Strength Rehabilitation Devices\" by Dietsch et al. (2024). The article aims to compare six expiratory muscle strength training (EMST) devices and investigates their minimum trigger pressure, variability across the settings, and stability. The models tested include five positive expiratory pressure (PEP) devices, Acapella, TheraPEP, Threshold PEP, EMST75, and EMST150. It also includes a combined inspiratory and expiratory muscle training device, the Breather. We have several concerns about the integrity of results presented regarding the Breather as presented in the article. These include the heterogeneity of device sample used, as well as inadequate methodology and the experimental setup.</p>","PeriodicalId":51254,"journal":{"name":"Journal of Speech Language and Hearing Research","volume":null,"pages":null},"PeriodicalIF":2.2,"publicationDate":"2024-09-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142309082","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-23DOI: 10.1044/2024_JSLHR-24-00429
Angela M Dietsch, Rahul Krishnamurthy, Kelsey Young, Steven M Barlow
{"title":"Response to the Letter to the Editor Regarding \"Instrumental Assessment of Aero-Resistive Expiratory Muscle Strength Rehabilitation Devices\".","authors":"Angela M Dietsch, Rahul Krishnamurthy, Kelsey Young, Steven M Barlow","doi":"10.1044/2024_JSLHR-24-00429","DOIUrl":"10.1044/2024_JSLHR-24-00429","url":null,"abstract":"","PeriodicalId":51254,"journal":{"name":"Journal of Speech Language and Hearing Research","volume":null,"pages":null},"PeriodicalIF":2.2,"publicationDate":"2024-09-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142309083","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-20DOI: 10.1044/2024_JSLHR-23-00583
Núria Montagut, Sergi Borrego-Écija, Jorge Herrero, Magdalena Castellví, Mircea Balasa, Albert Lladó, Stephanie M Grasso, Raquel Sánchez-Valle
Purpose: Primary progressive aphasia (PPA) is a neurodegenerative disorder characterized by worsening of speech and/or language. Script training intervention promotes automatized speech production via repeated practice of scripted content. This study evaluated the acceptability, feasibility, and effects of a modified version of Video-Implemented Script Training for Aphasia (VISTA) in the three PPA variants and compared outcomes by intervention modality (teletherapy vs. in person).
Method: Thirteen bilingual (Spanish-Catalan) participants were included (semantic variant, n = 5; logopenic variant, n = 5; nonfluent/agrammatic variant, n = 3; teletherapy, n = 7). Using a nonrandomized design, intervention was administered in participants' dominant language. Participants were trained on an individualized script twice per week, over 8 weeks. Performance on measures related to script accuracy, content, and subjective ratings of production quality was evaluated at baseline, immediately post, and at 3 and 6 months post-intervention.
Results: No significant differences were observed on the basis of intervention modality. Participants demonstrated significant improvements from pre- to post-intervention in script production, synonym production, keywords, and global quality on the trained script. Maintenance was observed when comparing performance at post-intervention relative to 3- and 6-month follow-up for script and synonym production. Significant improvement in production quality of the untrained topic was observed following intervention. Different patterns of benefit were observed by PPA variant.
Conclusions: Modified VISTA was acceptable and effective across the three PPA variants, as evidenced by improvements on a broader array of outcome measures than those previously reported. Findings also provide further support for provision for teletherapy in individuals with PPA.
{"title":"Effects of Modified Video-Implemented Script Training for Aphasia in the Three Variants of Primary Progressive Aphasia.","authors":"Núria Montagut, Sergi Borrego-Écija, Jorge Herrero, Magdalena Castellví, Mircea Balasa, Albert Lladó, Stephanie M Grasso, Raquel Sánchez-Valle","doi":"10.1044/2024_JSLHR-23-00583","DOIUrl":"10.1044/2024_JSLHR-23-00583","url":null,"abstract":"<p><strong>Purpose: </strong>Primary progressive aphasia (PPA) is a neurodegenerative disorder characterized by worsening of speech and/or language. Script training intervention promotes automatized speech production via repeated practice of scripted content. This study evaluated the acceptability, feasibility, and effects of a modified version of Video-Implemented Script Training for Aphasia (VISTA) in the three PPA variants and compared outcomes by intervention modality (teletherapy vs. in person).</p><p><strong>Method: </strong>Thirteen bilingual (Spanish-Catalan) participants were included (semantic variant, <i>n</i> = 5; logopenic variant, <i>n</i> = 5; nonfluent/agrammatic variant, <i>n</i> = 3; teletherapy, <i>n</i> = 7). Using a nonrandomized design, intervention was administered in participants' dominant language. Participants were trained on an individualized script twice per week, over 8 weeks. Performance on measures related to script accuracy, content, and subjective ratings of production quality was evaluated at baseline, immediately post, and at 3 and 6 months post-intervention.</p><p><strong>Results: </strong>No significant differences were observed on the basis of intervention modality. Participants demonstrated significant improvements from pre- to post-intervention in script production, synonym production, keywords, and global quality on the trained script. Maintenance was observed when comparing performance at post-intervention relative to 3- and 6-month follow-up for script and synonym production. Significant improvement in production quality of the untrained topic was observed following intervention. Different patterns of benefit were observed by PPA variant.</p><p><strong>Conclusions: </strong>Modified VISTA was acceptable and effective across the three PPA variants, as evidenced by improvements on a broader array of outcome measures than those previously reported. Findings also provide further support for provision for teletherapy in individuals with PPA.</p><p><strong>Supplemental material: </strong>https://doi.org/10.23641/asha.26999326.</p>","PeriodicalId":51254,"journal":{"name":"Journal of Speech Language and Hearing Research","volume":null,"pages":null},"PeriodicalIF":2.2,"publicationDate":"2024-09-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142300394","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-20DOI: 10.1044/2024_JSLHR-22-00562
Inmaculada Fajardo, Nadina Gómez-Merino, Antonio Ferrer, Isabel R Rodríguez-Ortiz
Purpose: The aim of the study was to analyze how face masks influence speech perception and time spent looking at the speaker's mouth and eyes by adults with and without hearing loss.
Method: Twenty participants with hearing loss and 20 without were asked to repeat Spanish words presented in various conditions, including different types of face masks (no mask, transparent window mask, and opaque mask FFP2) and presentation modes (audiovisual, video only, and audio only). Recognition accuracy and the percentage of time looking at the speaker's eyes and mouth (dwell time) were measured.
Results: In the audiovisual condition, participants with hearing loss had significantly better word recognition scores when the speaker wore no mask compared to when they wore an opaque face mask. However, there were no differences between the transparent mask and no mask conditions. For those with typical hearing, the type of face mask did not affect speech recognition. Audiovisual presentation consistently improved speech recognition for participants with hearing loss across all face mask conditions, but for those with typical hearing, it only improved compared to video-only mode. These participants demonstrated a ceiling effect in audiovisual and audio-only modes. Regarding eye movement patterns, participants spent less time looking at the speaker's mouth and more time at the eyes when the speaker wore an opaque mask compared to no mask or a transparent mask.
Conclusion: The use of transparent face masks (ClearMask-type model) is recommended in contexts where face masks are still used (hospitals) to prevent the hindering effect of opaque masks (FFP2-type model) in speech perception among people with hearing loss, provided that any fogging of the window of the transparent mask is controlled by wiping it off as needed and the light is in front of the speaker to minimize shadows.
{"title":"Hearing What You Can't See: Influence of Face Masks on Speech Perception and Eye Movement by Adults With Hearing Loss.","authors":"Inmaculada Fajardo, Nadina Gómez-Merino, Antonio Ferrer, Isabel R Rodríguez-Ortiz","doi":"10.1044/2024_JSLHR-22-00562","DOIUrl":"https://doi.org/10.1044/2024_JSLHR-22-00562","url":null,"abstract":"<p><strong>Purpose: </strong>The aim of the study was to analyze how face masks influence speech perception and time spent looking at the speaker's mouth and eyes by adults with and without hearing loss.</p><p><strong>Method: </strong>Twenty participants with hearing loss and 20 without were asked to repeat Spanish words presented in various conditions, including different types of face masks (no mask, transparent window mask, and opaque mask FFP2) and presentation modes (audiovisual, video only, and audio only). Recognition accuracy and the percentage of time looking at the speaker's eyes and mouth (dwell time) were measured.</p><p><strong>Results: </strong>In the audiovisual condition, participants with hearing loss had significantly better word recognition scores when the speaker wore no mask compared to when they wore an opaque face mask. However, there were no differences between the transparent mask and no mask conditions. For those with typical hearing, the type of face mask did not affect speech recognition. Audiovisual presentation consistently improved speech recognition for participants with hearing loss across all face mask conditions, but for those with typical hearing, it only improved compared to video-only mode. These participants demonstrated a ceiling effect in audiovisual and audio-only modes. Regarding eye movement patterns, participants spent less time looking at the speaker's mouth and more time at the eyes when the speaker wore an opaque mask compared to no mask or a transparent mask.</p><p><strong>Conclusion: </strong>The use of transparent face masks (ClearMask-type model) is recommended in contexts where face masks are still used (hospitals) to prevent the hindering effect of opaque masks (FFP2-type model) in speech perception among people with hearing loss, provided that any fogging of the window of the transparent mask is controlled by wiping it off as needed and the light is in front of the speaker to minimize shadows.</p>","PeriodicalId":51254,"journal":{"name":"Journal of Speech Language and Hearing Research","volume":null,"pages":null},"PeriodicalIF":2.2,"publicationDate":"2024-09-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142300395","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-20DOI: 10.1044/2024_JSLHR-23-00591
Ben Maassen, Hayo Terband
Background: Children with speech sound disorders (SSD) form a heterogeneous group, with respect to severity, etiology, proximal causes, speech error characteristics, and response to treatment. Infants develop speech and language in interaction with neurological maturation and general perceptual, motoric, and cognitive skills in a social-emotional context.
Purpose: After a brief introduction into psycholinguistic models of speech production and levels of causation, in this review article, we present an in-depth overview of mechanisms and processes, and the dynamics thereof, which are crucial in typical speech development. These basic mechanisms and processes are: (a) neurophysiological motor refinement, that is, the maturational articulatory mechanisms that drive babbling and the more differentiated production of larger speech patterns; (b) sensorimotor integration, which forms the steering function from phonetics to phonology; and (c) motor hierarchy and articulatory phonology describing the gestural organization of syllables, which underlie fluent speech production. These dynamics have consequences for the diagnosis and further analysis of SSD in children. We argue that current diagnostic classification systems do not do justice to the multilevel, multifactorial, and interactive character of the underlying mechanisms and processes. This is illustrated by a recent Dutch study yielding distinct performance profiles among children with SSD, which allows for a dimensional interpretation of underlying processing deficits.
Conclusions: Analyses of mainstream treatments with respect to the treatment goals and the speech mechanisms addressed show that treatment programs are quite transparent in their aims and approach and how they contribute to remediating specific deficits or mechanisms. Recent studies into clinical reasoning reveal that the clinical challenge for speech-language pathologists is how to select the most appropriate treatment at the most appropriate time for each individual child with SSD. We argue that a process-oriented approach has merits as compared to categorical diagnostics as a toolbox to aid in the interpretation of the speech profile in terms of underlying deficits and to connect these to a specific intervention approach and treatment target.
{"title":"Toward Process-Oriented, Dimensional Approaches for Diagnosis and Treatment of Speech Sound Disorders in Children: Position Statement and Future Perspectives.","authors":"Ben Maassen, Hayo Terband","doi":"10.1044/2024_JSLHR-23-00591","DOIUrl":"https://doi.org/10.1044/2024_JSLHR-23-00591","url":null,"abstract":"<p><strong>Background: </strong>Children with speech sound disorders (SSD) form a heterogeneous group, with respect to severity, etiology, proximal causes, speech error characteristics, and response to treatment. Infants develop speech and language in interaction with neurological maturation and general perceptual, motoric, and cognitive skills in a social-emotional context.</p><p><strong>Purpose: </strong>After a brief introduction into psycholinguistic models of speech production and levels of causation, in this review article, we present an in-depth overview of mechanisms and processes, and the dynamics thereof, which are crucial in typical speech development. These basic mechanisms and processes are: (a) neurophysiological motor refinement, that is, the maturational articulatory mechanisms that drive babbling and the more differentiated production of larger speech patterns; (b) sensorimotor integration, which forms the steering function from phonetics to phonology; and (c) motor hierarchy and articulatory phonology describing the gestural organization of syllables, which underlie fluent speech production. These dynamics have consequences for the diagnosis and further analysis of SSD in children. We argue that current diagnostic classification systems do not do justice to the multilevel, multifactorial, and interactive character of the underlying mechanisms and processes. This is illustrated by a recent Dutch study yielding distinct performance profiles among children with SSD, which allows for a dimensional interpretation of underlying processing deficits.</p><p><strong>Conclusions: </strong>Analyses of mainstream treatments with respect to the treatment goals and the speech mechanisms addressed show that treatment programs are quite transparent in their aims and approach and how they contribute to remediating specific deficits or mechanisms. Recent studies into clinical reasoning reveal that the clinical challenge for speech-language pathologists is how to select the most appropriate treatment at the most appropriate time for each individual child with SSD. We argue that a process-oriented approach has merits as compared to categorical diagnostics as a toolbox to aid in the interpretation of the speech profile in terms of underlying deficits and to connect these to a specific intervention approach and treatment target.</p>","PeriodicalId":51254,"journal":{"name":"Journal of Speech Language and Hearing Research","volume":null,"pages":null},"PeriodicalIF":2.2,"publicationDate":"2024-09-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142300396","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-20DOI: 10.1044/2024_JSLHR-23-00769
Elizabeth Schoen Simmons, Rhea Paul
Purpose: Typically developing toddlers extract patterns from their input to add words to their spoken lexicons, yet some evidence suggests that late talkers leverage the statistical regularities of the ambient language differently than do peers. Using the extended statistical learning account, we sought to compare lexical-level statistical features of spoken vocabularies between late talkers and two typically developing comparison groups.
Method: MacArthur-Bates Communicative Developmental Inventories American English Words and Sentences (N = 1,636) were extracted from Wordbank, a database of CDIs. Inventories were divided into three groups: (a) a late talker group (n = 202); (b) a typically developing age-matched group (n = 1,238); and (c) a younger, typically developing group (n = 196) matched to the late talkers on expressive language. Neighborhood density and word frequency were calculated for each word produced by each participant and standardized to z scores. Mixed-effects models were used to evaluate group differences.
Results: The late talker and younger, language-matched groups' spoken vocabularies consist, on standardized average, of words from denser phonological neighborhoods and words higher in frequency of occurrence in parent-child speech, compared to older, typically developing toddlers.
Conclusions: These findings provide support for the extended statistical learning account. Late talkers appear to generally be extracting and using similar patterns from their language input as do younger toddlers with similar levels of expressive vocabulary. This suggests that late talkers may be following a delayed, not deviant, trajectory of expressive language growth.
{"title":"Are Late Talkers Just Late? Neighborhood Density and Word Frequency Properties of Late Talkers' Spoken Vocabularies.","authors":"Elizabeth Schoen Simmons, Rhea Paul","doi":"10.1044/2024_JSLHR-23-00769","DOIUrl":"https://doi.org/10.1044/2024_JSLHR-23-00769","url":null,"abstract":"<p><strong>Purpose: </strong>Typically developing toddlers extract patterns from their input to add words to their spoken lexicons, yet some evidence suggests that late talkers leverage the statistical regularities of the ambient language differently than do peers. Using the extended statistical learning account, we sought to compare lexical-level statistical features of spoken vocabularies between late talkers and two typically developing comparison groups.</p><p><strong>Method: </strong>MacArthur-Bates Communicative Developmental Inventories American English Words and Sentences (<i>N</i> = 1,636) were extracted from Wordbank, a database of CDIs. Inventories were divided into three groups: (a) a late talker group (<i>n</i> = 202); (b) a typically developing age-matched group (<i>n</i> = 1,238); and (c) a younger, typically developing group (<i>n</i> = 196) matched to the late talkers on expressive language. Neighborhood density and word frequency were calculated for each word produced by each participant and standardized to <i>z</i> scores. Mixed-effects models were used to evaluate group differences.</p><p><strong>Results: </strong>The late talker and younger, language-matched groups' spoken vocabularies consist, on standardized average, of words from denser phonological neighborhoods and words higher in frequency of occurrence in parent-child speech, compared to older, typically developing toddlers.</p><p><strong>Conclusions: </strong>These findings provide support for the extended statistical learning account. Late talkers appear to generally be extracting and using similar patterns from their language input as do younger toddlers with similar levels of expressive vocabulary. This suggests that late talkers may be following a delayed, not deviant, trajectory of expressive language growth.</p>","PeriodicalId":51254,"journal":{"name":"Journal of Speech Language and Hearing Research","volume":null,"pages":null},"PeriodicalIF":2.2,"publicationDate":"2024-09-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142300393","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}