Pub Date : 2024-03-01Epub Date: 2023-03-26DOI: 10.1177/00238309231156615
Misaki Kato, Melissa M Baese-Berk
Previous research has shown that native listeners benefit from clearly produced speech, as well as from predictable semantic context when these enhancements are delivered in native speech. However, it is unclear whether native listeners benefit from acoustic and semantic enhancements differently when listening to other varieties of speech, including non-native speech. The current study examines to what extent native English listeners benefit from acoustic and semantic cues present in native and non-native English speech. Native English listeners transcribed sentence final words that were of different levels of semantic predictability, produced in plain- or clear-speaking styles by Native English talkers and by native Mandarin talkers of higher- and lower-proficiency in English. The perception results demonstrated that listeners benefited from semantic cues in higher- and lower-proficiency talkers' speech (i.e., transcribed speech more accurately), but not from acoustic cues, even though higher-proficiency talkers did make substantial acoustic enhancements from plain to clear speech. The current results suggest that native listeners benefit more robustly from semantic cues than from acoustic cues when those cues are embedded in non-native speech.
{"title":"The Effects of Acoustic and Semantic Enhancements on Perception of Native and Non-Native Speech.","authors":"Misaki Kato, Melissa M Baese-Berk","doi":"10.1177/00238309231156615","DOIUrl":"10.1177/00238309231156615","url":null,"abstract":"<p><p>Previous research has shown that native listeners benefit from clearly produced speech, as well as from predictable semantic context when these enhancements are delivered in native speech. However, it is unclear whether native listeners benefit from acoustic and semantic enhancements differently when listening to other varieties of speech, including non-native speech. The current study examines to what extent native English listeners benefit from acoustic and semantic cues present in native and non-native English speech. Native English listeners transcribed sentence final words that were of different levels of semantic predictability, produced in plain- or clear-speaking styles by Native English talkers and by native Mandarin talkers of higher- and lower-proficiency in English. The perception results demonstrated that listeners benefited from semantic cues in higher- and lower-proficiency talkers' speech (i.e., transcribed speech more accurately), but not from acoustic cues, even though higher-proficiency talkers did make substantial acoustic enhancements from plain to clear speech. The current results suggest that native listeners benefit more robustly from semantic cues than from acoustic cues when those cues are embedded in non-native speech.</p>","PeriodicalId":51255,"journal":{"name":"Language and Speech","volume":" ","pages":"40-71"},"PeriodicalIF":1.8,"publicationDate":"2024-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9177266","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"文学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-01Epub Date: 2023-03-16DOI: 10.1177/00238309231152492
Claire Moore-Cantwell, Joe Pater, Robert Staubs, Benjamin Zobel, Lisa Sanders
The experimental study of artificial language learning has become a widely used means of investigating the predictions of theories of language learning and representation. Although much is now known about the generalizations that learners make from various kinds of data, relatively little is known about how those representations affect speech processing. This paper presents an event-related potential (ERP) study of brain responses to violations of lab-learned phonotactics. Novel words that violated a learned phonotactic constraint elicited a larger Late Positive Component (LPC) than novel words that satisfied it. Similar LPCs have been found for violations of natively acquired linguistic structure, as well as for violations of other types of abstract generalizations, such as musical structure. We argue that lab-learned phonotactic generalizations are represented abstractly and affect the evaluation of speech in a manner that is similar to natively acquired syntactic and phonological rules.
{"title":"Violations of Lab-Learned Phonological Patterns Elicit a Late Positive Component.","authors":"Claire Moore-Cantwell, Joe Pater, Robert Staubs, Benjamin Zobel, Lisa Sanders","doi":"10.1177/00238309231152492","DOIUrl":"10.1177/00238309231152492","url":null,"abstract":"<p><p>The experimental study of artificial language learning has become a widely used means of investigating the predictions of theories of language learning and representation. Although much is now known about the generalizations that learners make from various kinds of data, relatively little is known about how those representations affect speech processing. This paper presents an event-related potential (ERP) study of brain responses to violations of lab-learned phonotactics. Novel words that violated a learned phonotactic constraint elicited a larger Late Positive Component (LPC) than novel words that satisfied it. Similar LPCs have been found for violations of natively acquired linguistic structure, as well as for violations of other types of abstract generalizations, such as musical structure. We argue that lab-learned phonotactic generalizations are represented abstractly and affect the evaluation of speech in a manner that is similar to natively acquired syntactic and phonological rules.</p>","PeriodicalId":51255,"journal":{"name":"Language and Speech","volume":" ","pages":"19-39"},"PeriodicalIF":1.8,"publicationDate":"2024-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9476291","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"文学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-12-29DOI: 10.1177/00238309231215355
Ander Beristain
Segment-to-segment timing overlap between Vowel-Nasal gestures in /VN/ sequences varies cross-linguistically. However, how bilinguals may adjust those timing gestures is still unanswered. Regarding timing strategies in a second language (L2), research finds that native (L1) strategies can be partially transferred to the L2, and that higher L2 proficiency promotes a more successful phonetic performance. My goal is to answer whether bilingual speakers can adjust their L1 coarticulatory settings in their L2 and to observe whether their L2 accentedness plays a role in ultimate attainment. Ten native speakers of Spanish (L1Sp) who were highly proficient L2 English speakers participated in Spanish and English read-aloud tasks. A control group of 16 L1 English speakers undertook the English experiment. Aerodynamic data were collected using pressure transducers. Each participant produced tokens with nasalized vowels in CVN# words and oral vowels in CV(CV) words. Four linguistically trained judges (two per target language) evaluated a set of pseudo-randomized sentences produced by the participants containing words with nasalized vowels and rated the speech on a 1 (heavily accented) to 9 (native-like) Likert-type scale. Measurements for onset and degree of overall nasality were obtained. Results indicate the L1Sp group can accommodate gestural timing strategies cross-linguistically as they exhibit an earlier nasality onset and increment nasality proportion in L2 English in a native-like manner. In addition, a positive correlation between greater vowel nasality degree and native-like accentedness in the L2 was found, suggesting L2 timing settings might be specified in higher spoken proficiency levels.
{"title":"Gestural Timing Patterns of Nasality in Highly Proficient Spanish Learners of English: Aerodynamic Evidence.","authors":"Ander Beristain","doi":"10.1177/00238309231215355","DOIUrl":"https://doi.org/10.1177/00238309231215355","url":null,"abstract":"<p><p>Segment-to-segment timing overlap between Vowel-Nasal gestures in /VN/ sequences varies cross-linguistically. However, how bilinguals may adjust those timing gestures is still unanswered. Regarding timing strategies in a second language (L2), research finds that native (L1) strategies can be partially transferred to the L2, and that higher L2 proficiency promotes a more successful phonetic performance. My goal is to answer whether bilingual speakers can adjust their L1 coarticulatory settings in their L2 and to observe whether their L2 accentedness plays a role in ultimate attainment. Ten native speakers of Spanish (L1Sp) who were highly proficient L2 English speakers participated in Spanish and English read-aloud tasks. A control group of 16 L1 English speakers undertook the English experiment. Aerodynamic data were collected using pressure transducers. Each participant produced tokens with nasalized vowels in CVN# words and oral vowels in CV(CV) words. Four linguistically trained judges (two per target language) evaluated a set of pseudo-randomized sentences produced by the participants containing words with nasalized vowels and rated the speech on a 1 (heavily accented) to 9 (native-like) Likert-type scale. Measurements for onset and degree of overall nasality were obtained. Results indicate the L1Sp group can accommodate gestural timing strategies cross-linguistically as they exhibit an earlier nasality onset and increment nasality proportion in L2 English in a native-like manner. In addition, a positive correlation between greater vowel nasality degree and native-like accentedness in the L2 was found, suggesting L2 timing settings might be specified in higher spoken proficiency levels.</p>","PeriodicalId":51255,"journal":{"name":"Language and Speech","volume":" ","pages":"238309231215355"},"PeriodicalIF":1.8,"publicationDate":"2023-12-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139059054","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"文学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-12-01Epub Date: 2022-12-07DOI: 10.1177/00238309221127029
Katarzyna Bromberek-Dyzman, Joanna Kowalik, Anna Pękacz
Irony comprehension involves understanding implicit attitudes communicated on top of the explicit, literal meaning. Because of the double-binding stemming from the explicit-implicit incongruity, irony is assumed to be cognitively taxing when communicated in the native language (NL), and even more so in the foreign language (FL). Prior studies investigating irony comprehension in the FL indicated that irony may be processed with similar speed and accuracy in the foreign and the native language, and that irony comprehension efficiency in the former might be lower. Building on these findings, this study investigates if the tasks participants perform when reading irony in their FL affect the efficiency of irony comprehension. We invited 150 advanced Polish users of English to take part in one of two tasks. Participants were asked to read 3- to 5-sentence-long-scenarios, ending in literal (Literal Praise, Literal Criticism) or ironic comments (Ironic Praise, Ironic Criticism) and to either (1) make true/false judgments (T/F task; N = 83), or to (2) judge the emotional value in emotive decision task (EDT; N = 67). A full spectrum of verbal irony employed to communicate ironic praise and criticism, and their literal equivalents: literal praise and literal criticism, was tested. A three-way mixed ANOVA with 2× Valence (Praise, Criticism), 2× Mode (Literal, Ironic) as within-subject repeated-measures, and 2× Task (True/False, Emotive Decision) as a between-subject independent measure, revealed that efficiency of irony comprehension was differently modified by the two tasks. Therefore, we argue that irony comprehension in English as a foreign language is task-constrained.
{"title":"Task Effects in Irony Comprehension in English as a Foreign Language.","authors":"Katarzyna Bromberek-Dyzman, Joanna Kowalik, Anna Pękacz","doi":"10.1177/00238309221127029","DOIUrl":"10.1177/00238309221127029","url":null,"abstract":"<p><p>Irony comprehension involves understanding implicit attitudes communicated on top of the explicit, literal meaning. Because of the double-binding stemming from the explicit-implicit incongruity, irony is assumed to be cognitively taxing when communicated in the native language (NL), and even more so in the foreign language (FL). Prior studies investigating irony comprehension in the FL indicated that irony may be processed with similar speed and accuracy in the foreign and the native language, and that irony comprehension efficiency in the former might be lower. Building on these findings, this study investigates if the tasks participants perform when reading irony in their FL affect the efficiency of irony comprehension. We invited 150 advanced Polish users of English to take part in one of two tasks. Participants were asked to read 3- to 5-sentence-long-scenarios, ending in literal (Literal Praise, Literal Criticism) or ironic comments (Ironic Praise, Ironic Criticism) and to either (1) make true/false judgments (T/F task; <i>N</i> = 83), or to (2) judge the emotional value in emotive decision task (EDT; <i>N</i> = 67). A full spectrum of verbal irony employed to communicate ironic praise and criticism, and their literal equivalents: literal praise and literal criticism, was tested. A three-way mixed ANOVA with 2× Valence (Praise, Criticism), 2× Mode (Literal, Ironic) as within-subject repeated-measures, and 2× Task (True/False, Emotive Decision) as a between-subject independent measure, revealed that efficiency of irony comprehension was differently modified by the two tasks. Therefore, we argue that irony comprehension in English as a foreign language is task-constrained.</p>","PeriodicalId":51255,"journal":{"name":"Language and Speech","volume":" ","pages":"833-850"},"PeriodicalIF":1.8,"publicationDate":"2023-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10666523/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"10361246","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"文学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-12-01Epub Date: 2023-01-21DOI: 10.1177/00238309221140910
Katsura Aoyama, Lingzi Hong, James E Flege, Reiko Akahane-Yamada, Tsuneo Yamada
The primary purpose of this research report was to investigate the relationships between acoustic characteristics and perceived intelligibility for native Japanese speakers' productions of American English liquids. This report was based on a reanalysis of intelligibility scores and acoustic analyses that were reported in two previous studies. We examined which acoustic parameters were associated with higher perceived intelligibility scores for their productions of /l/ and /ɹ/ in American English, and whether Japanese speakers' productions of the two liquids were acoustically differentiated from each other. Results demonstrated that the second formant (F2) was strongly correlated with the perceived intelligibility scores for the Japanese adults' productions. Results also demonstrated that the Japanese adults' and children's productions of /l/ and /ɹ/ were indeed differentiated by some acoustic parameters including the third formant (F3). In addition, some changes occurred in the Japanese children's productions over the course of 1 year. Overall, the present report shows that Japanese speakers of American English may be making a distinction between /l/ and /ɹ/ in production, although the distinctions are made in a different way compared with native English speakers' productions. These findings have implications for setting realistic goals for improving intelligibility of English /l/ and /ɹ/ for Japanese speakers, as well as theoretical advancement of second-language speech learning.
{"title":"Relationships Between Acoustic Characteristics and Intelligibility Scores: A Reanalysis of Japanese Speakers' Productions of American English Liquids.","authors":"Katsura Aoyama, Lingzi Hong, James E Flege, Reiko Akahane-Yamada, Tsuneo Yamada","doi":"10.1177/00238309221140910","DOIUrl":"10.1177/00238309221140910","url":null,"abstract":"<p><p>The primary purpose of this research report was to investigate the relationships between acoustic characteristics and perceived intelligibility for native Japanese speakers' productions of American English liquids. This report was based on a reanalysis of intelligibility scores and acoustic analyses that were reported in two previous studies. We examined which acoustic parameters were associated with higher perceived intelligibility scores for their productions of /l/ and /ɹ/ in American English, and whether Japanese speakers' productions of the two liquids were acoustically differentiated from each other. Results demonstrated that the second formant (F2) was strongly correlated with the perceived intelligibility scores for the Japanese adults' productions. Results also demonstrated that the Japanese adults' and children's productions of /l/ and /ɹ/ were indeed differentiated by some acoustic parameters including the third formant (F3). In addition, some changes occurred in the Japanese children's productions over the course of 1 year. Overall, the present report shows that Japanese speakers of American English may be making a distinction between /l/ and /ɹ/ in production, although the distinctions are made in a different way compared with native English speakers' productions. These findings have implications for setting realistic goals for improving intelligibility of English /l/ and /ɹ/ for Japanese speakers, as well as theoretical advancement of second-language speech learning.</p>","PeriodicalId":51255,"journal":{"name":"Language and Speech","volume":" ","pages":"1030-1045"},"PeriodicalIF":1.8,"publicationDate":"2023-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9103813","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"文学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-12-01Epub Date: 2022-12-27DOI: 10.1177/00238309221137326
Nicholas Henriksen, Shayna Greenley, Amber Galvano
The "hyper-variation" present in rhotic sounds makes them particularly apt for sociophonetic research. This paper investigates the variable realization of the voiced alveolar-trill phoneme /r/ through an acoustic analysis of unscripted speech produced by 80 speakers of Spanish. Although the most common phonetic variant of /r/ contained two lingual constrictions, we find substantial inter-speaker variation in our data, ranging from zero to five lingual contacts. The results demonstrate that the variation in Spanish results from a systematic interaction of factors, deriving from well-documented processes of consonantal lenition (e.g., weakening in unstressed syllables) in addition to processes inherent to the trill's articulation (e.g., high-vowel antagonism). Importantly, speaker sex displayed the strongest effect among all the predictors, which leads us to consider the role of sociolinguistic factors, in addition to possible biomechanical differences, on /r/ production. We contextualize the findings within a literature that theorizes rhotic consonants as a single class of sounds despite remarkable patterns of cross-language and speaker-specific variation.
{"title":"Sociophonetic Investigation of the Spanish Alveolar Trill /r/ in Two Canonical-Trill Varieties.","authors":"Nicholas Henriksen, Shayna Greenley, Amber Galvano","doi":"10.1177/00238309221137326","DOIUrl":"10.1177/00238309221137326","url":null,"abstract":"<p><p>The \"hyper-variation\" present in rhotic sounds makes them particularly apt for sociophonetic research. This paper investigates the variable realization of the voiced alveolar-trill phoneme /r/ through an acoustic analysis of unscripted speech produced by 80 speakers of Spanish. Although the most common phonetic variant of /r/ contained two lingual constrictions, we find substantial inter-speaker variation in our data, ranging from zero to five lingual contacts. The results demonstrate that the variation in Spanish results from a systematic interaction of factors, deriving from well-documented processes of consonantal lenition (e.g., weakening in unstressed syllables) in addition to processes inherent to the trill's articulation (e.g., high-vowel antagonism). Importantly, speaker sex displayed the strongest effect among all the predictors, which leads us to consider the role of sociolinguistic factors, in addition to possible biomechanical differences, on /r/ production. We contextualize the findings within a literature that theorizes rhotic consonants as a single class of sounds despite remarkable patterns of cross-language and speaker-specific variation.</p>","PeriodicalId":51255,"journal":{"name":"Language and Speech","volume":" ","pages":"896-934"},"PeriodicalIF":1.8,"publicationDate":"2023-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"10805610","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"文学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-12-01Epub Date: 2022-12-08DOI: 10.1177/00238309221133836
Chandan R Narayan
The relationship between speaking rate and burst amplitude was investigated in plosives with differing oro-laryngeal timing: long-lag voice-onset time (VOT) (North American English) and short-lag VOT (Indian Tamil). Burst amplitude (reflecting both intraoral pressure and flow geometry of the oral channel) was hypothesized to decrease in pre-vocalic plosive syllables with the increase in speaking rate, which imposes temporal constraints on both intraoral pressure buildup behind the oral occlusion and respiratory air flow. The results showed that decreased vowel duration (which is associated with increased speaking rate) led to decreased burst amplitude in both short- and long-lag plosives. Aggregate models of bilabial and velar plosives (found in both languages) suggested lower burst amplitudes in short-lag stops. Place-of-articulation effects in both languages were consistent with models of stop consonant acoustics, and place interactions with vowel duration were most apparent with long-lag English stops. The results are discussed in terms of speaking rate and language-internal forces, contributing to burst amplitude variation and their implications for speech perception and potential to affect lenition phenomena.
{"title":"Speaking Rate, Oro-Laryngeal Timing, and Place of Articulation Effects on Burst Amplitude: Evidence From English and Tamil.","authors":"Chandan R Narayan","doi":"10.1177/00238309221133836","DOIUrl":"10.1177/00238309221133836","url":null,"abstract":"<p><p>The relationship between speaking rate and burst amplitude was investigated in plosives with differing oro-laryngeal timing: long-lag voice-onset time (VOT) (North American English) and short-lag VOT (Indian Tamil). Burst amplitude (reflecting both intraoral pressure and flow geometry of the oral channel) was hypothesized to decrease in pre-vocalic plosive syllables with the increase in speaking rate, which imposes temporal constraints on both intraoral pressure buildup behind the oral occlusion and respiratory air flow. The results showed that decreased vowel duration (which is associated with increased speaking rate) led to decreased burst amplitude in both short- and long-lag plosives. Aggregate models of bilabial and velar plosives (found in both languages) suggested lower burst amplitudes in short-lag stops. Place-of-articulation effects in both languages were consistent with models of stop consonant acoustics, and place interactions with vowel duration were most apparent with long-lag English stops. The results are discussed in terms of speaking rate and language-internal forces, contributing to burst amplitude variation and their implications for speech perception and potential to affect lenition phenomena.</p>","PeriodicalId":51255,"journal":{"name":"Language and Speech","volume":" ","pages":"851-869"},"PeriodicalIF":1.8,"publicationDate":"2023-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10666501/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"10369671","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"文学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-12-01Epub Date: 2023-01-12DOI: 10.1177/00238309221142526
Rebecca L Morley, Bridget J Smith
The voicing effect is among the most studied and robust of phonetic phenomena. Yet there remains a lack of consensus on why vowels preceding voiced obstruents should be longer than vowels preceding voiceless obstruents. In this paper we provide an analysis of the voicing effect in a corpus of natural speech, and using production data from a metronome-timed word repetition study. From this evidence, as well as the existing literature, we conclude that vowel duration differences follow from consonant duration differences. The characteristic voicing effect in English is largely limited to words of especially long duration, and preceding vowel duration does not reliably cue obstruent voicing under the following circumstances: when obstruent voicing or duration cues conflict; for lax or unstressed vowels; and for most conversational speech. We show that this behavior can be modeled using a competing-constraints framework, where all segments resist expanding or compressing past a preferred duration. Inherent segment elasticity determines the degree of resistance, but segment duration is ultimately determined by the interaction of these segmental constraints with constraints on the distribution of the lengthening force within the syllable, and how closely target durations are matched. This account of the voicing effect has a number of implications for phonological theory, especially the central role that the concept of prominence plays in the analysis of underlying features.
{"title":"A Reanalysis of the Voicing Effect in English: With Implications for Featural Specification.","authors":"Rebecca L Morley, Bridget J Smith","doi":"10.1177/00238309221142526","DOIUrl":"10.1177/00238309221142526","url":null,"abstract":"<p><p>The voicing effect is among the most studied and robust of phonetic phenomena. Yet there remains a lack of consensus on why vowels preceding voiced obstruents should be longer than vowels preceding voiceless obstruents. In this paper we provide an analysis of the voicing effect in a corpus of natural speech, and using production data from a metronome-timed word repetition study. From this evidence, as well as the existing literature, we conclude that vowel duration differences follow from consonant duration differences. The characteristic voicing effect in English is largely limited to words of especially long duration, and preceding vowel duration does <i>not</i> reliably cue obstruent voicing under the following circumstances: when obstruent voicing or duration cues conflict; for lax or unstressed vowels; and for most conversational speech. We show that this behavior can be modeled using a competing-constraints framework, where all segments resist expanding or compressing past a preferred duration. Inherent segment elasticity determines the degree of resistance, but segment duration is ultimately determined by the interaction of these segmental constraints with constraints on the distribution of the lengthening force within the syllable, and how closely target durations are matched. This account of the voicing effect has a number of implications for phonological theory, especially the central role that the concept of prominence plays in the analysis of underlying features.</p>","PeriodicalId":51255,"journal":{"name":"Language and Speech","volume":" ","pages":"935-973"},"PeriodicalIF":1.8,"publicationDate":"2023-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"10511687","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"文学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-12-01Epub Date: 2023-02-12DOI: 10.1177/00238309231153355
Wenli Liu, Xiaoguang Pan, Xiang Zhou
Empirical evidence and theoretical models suggest that phonetic category perception involves two stages of auditory and phonetic processing. However, few studies examined the time course of these two processing stages. With brief stop consonant segments as context stimuli, this study examined the temporal dynamics of stop consonant perception by varying the inter-stimulus interval between context and target stimuli. The results suggest that phonetic category activation of stop consonants may appear before 100 ms of processing time. Furthermore, the activation of phonetic categories resulted in contrast context effects on identifying the target stop continuum; the auditory processing of stop consonants resulted in a different context effect from those caused by phonetic category activation. The findings provide further evidence for the two-stage model of speech perception and reveal the time course of auditory and phonetic processing.
{"title":"The Temporal Dynamics of Stop Consonant Perception: Evidence from Context Effects.","authors":"Wenli Liu, Xiaoguang Pan, Xiang Zhou","doi":"10.1177/00238309231153355","DOIUrl":"10.1177/00238309231153355","url":null,"abstract":"<p><p>Empirical evidence and theoretical models suggest that phonetic category perception involves two stages of auditory and phonetic processing. However, few studies examined the time course of these two processing stages. With brief stop consonant segments as context stimuli, this study examined the temporal dynamics of stop consonant perception by varying the inter-stimulus interval between context and target stimuli. The results suggest that phonetic category activation of stop consonants may appear before 100 ms of processing time. Furthermore, the activation of phonetic categories resulted in contrast context effects on identifying the target stop continuum; the auditory processing of stop consonants resulted in a different context effect from those caused by phonetic category activation. The findings provide further evidence for the two-stage model of speech perception and reveal the time course of auditory and phonetic processing.</p>","PeriodicalId":51255,"journal":{"name":"Language and Speech","volume":" ","pages":"1046-1055"},"PeriodicalIF":1.8,"publicationDate":"2023-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"10701039","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"文学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-12-01Epub Date: 2023-01-21DOI: 10.1177/00238309221143062
Keke Yu, Yacong Zhou, Linjun Zhang, Li Li, Ping Li, Ruiming Wang
Previous studies have suggested the effect of linguistic information on voice perception (e.g., the language-familiarity effect [LFE]). However, it remains unclear which type of specific information in speech contributes to voice perception, including acoustic, phonological, lexical, and semantic information. It is also underexamined whether the roles of these different types of information are modulated by the experimental paradigm (speaker discrimination vs. speaker identification). In this study, we conducted two experiments to investigate these issues regarding LFEs. Experiment 1 examined the roles of acoustic and phonological information in speaker discrimination and identification with forward and time-reversed Mandarin and Indonesian sentences. Experiment 2 further identified the roles of phonological, lexical, and semantic information with forward, word-scrambled, and reconstructed (consisting of pseudo-Mandarin words) Mandarin and forward Indonesian sentences. For Mandarin-only participants, in Experiment 1, speaker discrimination was more accurate for forward than reversed sentences, but there was no LFE in either sentence. Speaker identification was also more accurate for forward than reversed sentences, whereas there was an LFE for forward sentences. In Experiment 2, speaker discrimination was better for word-scrambled than reconstructed Mandarin sentences. Speaker identification was more accurate for forward and word-scrambled Mandarin sentences but less accurate for Mandarin reconstructed and forward Indonesian sentences. In general, the pattern of the results for Indonesian learners was the same as that for Mandarin-only speakers. These results suggest that different kinds of information support speaker discrimination and identification in native and unfamiliar languages. The LFE in speaker identification depends on both phonological and lexical information.
{"title":"How Different Types of Linguistic Information Impact Voice Perception: Evidence From the Language-Familiarity Effect.","authors":"Keke Yu, Yacong Zhou, Linjun Zhang, Li Li, Ping Li, Ruiming Wang","doi":"10.1177/00238309221143062","DOIUrl":"10.1177/00238309221143062","url":null,"abstract":"<p><p>Previous studies have suggested the effect of linguistic information on voice perception (e.g., the language-familiarity effect [LFE]). However, it remains unclear which type of specific information in speech contributes to voice perception, including acoustic, phonological, lexical, and semantic information. It is also underexamined whether the roles of these different types of information are modulated by the experimental paradigm (speaker discrimination vs. speaker identification). In this study, we conducted two experiments to investigate these issues regarding LFEs. Experiment 1 examined the roles of acoustic and phonological information in speaker discrimination and identification with forward and time-reversed Mandarin and Indonesian sentences. Experiment 2 further identified the roles of phonological, lexical, and semantic information with forward, word-scrambled, and reconstructed (consisting of pseudo-Mandarin words) Mandarin and forward Indonesian sentences. For Mandarin-only participants, in Experiment 1, speaker discrimination was more accurate for forward than reversed sentences, but there was no LFE in either sentence. Speaker identification was also more accurate for forward than reversed sentences, whereas there was an LFE for forward sentences. In Experiment 2, speaker discrimination was better for word-scrambled than reconstructed Mandarin sentences. Speaker identification was more accurate for forward and word-scrambled Mandarin sentences but less accurate for Mandarin reconstructed and forward Indonesian sentences. In general, the pattern of the results for Indonesian learners was the same as that for Mandarin-only speakers. These results suggest that different kinds of information support speaker discrimination and identification in native and unfamiliar languages. The LFE in speaker identification depends on both phonological and lexical information.</p>","PeriodicalId":51255,"journal":{"name":"Language and Speech","volume":" ","pages":"1007-1029"},"PeriodicalIF":1.8,"publicationDate":"2023-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9103814","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"文学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}