Towards Greater Conceptual Clarity in Complexity and Difficulty: A Commentary on “Complexity and Difficulty in Second Language Acquisition: A Theoretical and Methodological Overview”

IF 4.2 1区 文学 Q1 EDUCATION & EDUCATIONAL RESEARCH Language Learning Pub Date : 2024-10-22 DOI:10.1111/lang.12688
Xiaofei Lu
{"title":"Towards Greater Conceptual Clarity in Complexity and Difficulty: A Commentary on “Complexity and Difficulty in Second Language Acquisition: A Theoretical and Methodological Overview”","authors":"Xiaofei Lu","doi":"10.1111/lang.12688","DOIUrl":null,"url":null,"abstract":"<p>The conceptual review article by Bulté, Housen, and Pallotti constitutes a much-needed effort to disentangle the conceptual and methodological messiness surrounding complexity and difficulty, two key constructs in second language (L2) acquisition research. Seeking clear conceptual distinction between these constructs, Bulté et al. interpret complexity as structural properties of linguistic items and difficulty as the cognitive costs associated with acquiring and using such items. They further propose a small set of core measures for researchers to routinely use so as to promote replicability and knowledge accumulation. In this commentary, I offer some thoughts on their discussion of the definitions and measurements of these constructs.</p><p>The inclusion and exclusion of “sophistication” in different definitions of “complexity” (e.g., Kyle et al., <span>2021</span>; Ortega, <span>2003</span>), along with the introduction of the terms “absolute complexity” and “relative complexity” (e.g., Housen &amp; Simoens, <span>2016</span>), contributed to the terminological confusion surrounding complexity. I concur with Bulté et al.’s narrow interpretation of complexity as equivalent to absolute complexity that excludes sophistication. Meanwhile, an explicit discussion of (a) the precise relationship among the terms “difficulty,” “sophistication,” and “relative complexity” and (b) the future status of the terms “absolute complexity” and “relative complexity” would provide greater terminological clarity for the field.</p><p>Bulté et al. touch upon two important debates on the definition and measurement of complexity. First, they define complexity independently from notions of register-/genre-based adequacy. Register/genre variation research has yielded valuable insights into the linguistic characteristics of different registers/genres. However, the relative frequencies of linguistic items in different registers/genres should not form the basis for including or excluding them in analyzing the complexity of texts of a specific register/genre. This is because while a linguistic item may be less frequent in one register/genre than in others, it may nevertheless play an indispensable role in that register/genre, and this specific role should be analyzed register- or genre-internally. Second, they believe that the fine-grained approach to complexity analysis is complementary with a more holistic approach. In my view, all complexity measures are on a scale of granularity, and the criticism that holistic measures (e.g., dependent clauses per clause) lack specificity and interpretability, which is sometimes cited without thorough consideration, similarly applies to most fine-grained measures (e.g., relative clauses per 1,000 words). For example, just as there are different types of dependent clauses, there are also different types of relative clauses, each of which may in turn have subtypes. In fact, granularity could go all the way down to lexicalization. Holistic measures provide an efficient way to assess high-level complexity (e.g., the emergence or overall level of finite subordination or noun modification). As L2 writing teachers and researchers may be interested in complexity at different levels of granularity in different contexts and for different purposes, it is important to recognize the complementarity of holistic and fine-grained approaches to complexity analysis, similar to that of holistic and analytical rating scales used in writing assessments.</p><p>Bulté et al. do not consider the sentence as a unit of analysis but consider the T-unit or AS-unit as the largest syntactic unit of analysis. This recommendation opens up an interesting discussion. First, the sentence is an intuitively useful unit of information organization in writing. Second, complexification by clausal coordination is known to be especially important for beginning learners (e.g., Ortega, <span>2003</span>). Third, by disregarding punctuated sentence fragments in our analysis, we may miss the opportunity to capture intentional, appropriate uses of such fragments and potentially unintended, erroneous uses in learner texts. Fourth, the T-unit may not be an appropriate or the largest unit of analysis in all languages. For example, the topic–comment unit has been argued to be more appropriate than the T-unit for analyzing complexity in Chinese, and a terminal topic–comment unit could correspond to two or more T-units (e.g., Hu et al., <span>2022</span>; Yu, <span>2021</span>).</p><p>For syntactic complexity measurement, Bulté et al. recommend using mean words per phrase and mean phrases per clause as two of the six core constitutional measures. To promote replicability, it would seem necessary to agree on a consistent approach to identifying and counting phrases (e.g., how many phrases are there in <i>read a book about ancient China</i>?). For lexical complexity measurement, Bulté et al. recommend lemmatizing inflectional word forms but leave the treatment of derivational forms open. They propose mean word length as the core constitutional measure of this construct. However, without defining lexical complexity as word family complexity, mean morphemes per word and mean length of (root) morphemes might better align with their effort to capture constitutional complexity in syntactic complexity measurement, as words are constituted by morphemes, which are then represented by letters or characters in written form.</p><p>Bulté et al. recognize the normalized rate of occurrence of linguistic forms as a valid approach to calculating text-level complexity. Measures of this type are known to facilitate the comparison of the rate of occurrence of linguistic forms across texts. However, they prioritize the frequency of occurrence of linguistic forms over their internal complexity. More importantly, they do not consider the distribution of linguistic forms across the larger units in which they occur and thus cannot account for the structural organization of the text. Texts are not organized by <i>n</i>-word chunks and rarely display homogeneous dispersion of linguistic forms. Rather, specific linguistic forms occur within larger units, both syntactic units immediately containing them and larger units realizing discursive acts/functions, that contribute to the overall text structure. Therefore, measures based on normalized rate of occurrence are best used in combination with other types of measures that capture the internal complexity of linguistic forms and their distribution in larger units.</p><p>Bulté et al. argue against the consideration of word meanings in analyzing the complexity or difficulty of L2 production based on the concern that it is difficult to determine the meanings of a linguistic item in learner texts. While this concern is warranted, the consideration of word meanings may be necessary in measuring difficulty. For example, based on production and interview data, Liu and Lu (<span>2020</span>) reported that their Chinese-speaking L2 English learners frequently used some but not other meanings of specific words (e.g., <i>composition</i> meaning a piece of writing vs. the way in which something is formed by its parts) due to their knowledge gaps. Lu and Hu (<span>2022</span>) further argued that different meanings of polysemous words may represent different levels of sophistication. They showed that it is possible to pinpoint the meanings of most polysemous words in learner texts and that sense-aware lexical sophistication indices can better predict learner proficiency than form-based ones. The cases in which the meanings of polysemous words are unclear (e.g., due to incorrect or creative usage) could constitute another useful dimension of analysis, as learner development involves learning not only to express more sophisticated meanings but also to express meanings unambiguously. Regarding the concern for accurate word sense disambiguation, large language models could now be fine-tuned to improve performance, and manual analysis can also be involved.</p><p>In sum, I am thankful for Bulté et al.’s contribution towards resolving the conceptual and methodological controversies around the constructs of complexity and difficulty, and I have raised a few further issues to be considered. Importantly, the complex and multidimensional nature of these constructs necessitates a set of complementary measures of different types, at different levels of granularity, and with sensitivity to features of different languages.</p>","PeriodicalId":51371,"journal":{"name":"Language Learning","volume":"75 2","pages":"594-598"},"PeriodicalIF":4.2000,"publicationDate":"2024-10-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1111/lang.12688","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Language Learning","FirstCategoryId":"98","ListUrlMain":"https://onlinelibrary.wiley.com/doi/10.1111/lang.12688","RegionNum":1,"RegionCategory":"文学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"EDUCATION & EDUCATIONAL RESEARCH","Score":null,"Total":0}
引用次数: 0

Abstract

The conceptual review article by Bulté, Housen, and Pallotti constitutes a much-needed effort to disentangle the conceptual and methodological messiness surrounding complexity and difficulty, two key constructs in second language (L2) acquisition research. Seeking clear conceptual distinction between these constructs, Bulté et al. interpret complexity as structural properties of linguistic items and difficulty as the cognitive costs associated with acquiring and using such items. They further propose a small set of core measures for researchers to routinely use so as to promote replicability and knowledge accumulation. In this commentary, I offer some thoughts on their discussion of the definitions and measurements of these constructs.

The inclusion and exclusion of “sophistication” in different definitions of “complexity” (e.g., Kyle et al., 2021; Ortega, 2003), along with the introduction of the terms “absolute complexity” and “relative complexity” (e.g., Housen & Simoens, 2016), contributed to the terminological confusion surrounding complexity. I concur with Bulté et al.’s narrow interpretation of complexity as equivalent to absolute complexity that excludes sophistication. Meanwhile, an explicit discussion of (a) the precise relationship among the terms “difficulty,” “sophistication,” and “relative complexity” and (b) the future status of the terms “absolute complexity” and “relative complexity” would provide greater terminological clarity for the field.

Bulté et al. touch upon two important debates on the definition and measurement of complexity. First, they define complexity independently from notions of register-/genre-based adequacy. Register/genre variation research has yielded valuable insights into the linguistic characteristics of different registers/genres. However, the relative frequencies of linguistic items in different registers/genres should not form the basis for including or excluding them in analyzing the complexity of texts of a specific register/genre. This is because while a linguistic item may be less frequent in one register/genre than in others, it may nevertheless play an indispensable role in that register/genre, and this specific role should be analyzed register- or genre-internally. Second, they believe that the fine-grained approach to complexity analysis is complementary with a more holistic approach. In my view, all complexity measures are on a scale of granularity, and the criticism that holistic measures (e.g., dependent clauses per clause) lack specificity and interpretability, which is sometimes cited without thorough consideration, similarly applies to most fine-grained measures (e.g., relative clauses per 1,000 words). For example, just as there are different types of dependent clauses, there are also different types of relative clauses, each of which may in turn have subtypes. In fact, granularity could go all the way down to lexicalization. Holistic measures provide an efficient way to assess high-level complexity (e.g., the emergence or overall level of finite subordination or noun modification). As L2 writing teachers and researchers may be interested in complexity at different levels of granularity in different contexts and for different purposes, it is important to recognize the complementarity of holistic and fine-grained approaches to complexity analysis, similar to that of holistic and analytical rating scales used in writing assessments.

Bulté et al. do not consider the sentence as a unit of analysis but consider the T-unit or AS-unit as the largest syntactic unit of analysis. This recommendation opens up an interesting discussion. First, the sentence is an intuitively useful unit of information organization in writing. Second, complexification by clausal coordination is known to be especially important for beginning learners (e.g., Ortega, 2003). Third, by disregarding punctuated sentence fragments in our analysis, we may miss the opportunity to capture intentional, appropriate uses of such fragments and potentially unintended, erroneous uses in learner texts. Fourth, the T-unit may not be an appropriate or the largest unit of analysis in all languages. For example, the topic–comment unit has been argued to be more appropriate than the T-unit for analyzing complexity in Chinese, and a terminal topic–comment unit could correspond to two or more T-units (e.g., Hu et al., 2022; Yu, 2021).

For syntactic complexity measurement, Bulté et al. recommend using mean words per phrase and mean phrases per clause as two of the six core constitutional measures. To promote replicability, it would seem necessary to agree on a consistent approach to identifying and counting phrases (e.g., how many phrases are there in read a book about ancient China?). For lexical complexity measurement, Bulté et al. recommend lemmatizing inflectional word forms but leave the treatment of derivational forms open. They propose mean word length as the core constitutional measure of this construct. However, without defining lexical complexity as word family complexity, mean morphemes per word and mean length of (root) morphemes might better align with their effort to capture constitutional complexity in syntactic complexity measurement, as words are constituted by morphemes, which are then represented by letters or characters in written form.

Bulté et al. recognize the normalized rate of occurrence of linguistic forms as a valid approach to calculating text-level complexity. Measures of this type are known to facilitate the comparison of the rate of occurrence of linguistic forms across texts. However, they prioritize the frequency of occurrence of linguistic forms over their internal complexity. More importantly, they do not consider the distribution of linguistic forms across the larger units in which they occur and thus cannot account for the structural organization of the text. Texts are not organized by n-word chunks and rarely display homogeneous dispersion of linguistic forms. Rather, specific linguistic forms occur within larger units, both syntactic units immediately containing them and larger units realizing discursive acts/functions, that contribute to the overall text structure. Therefore, measures based on normalized rate of occurrence are best used in combination with other types of measures that capture the internal complexity of linguistic forms and their distribution in larger units.

Bulté et al. argue against the consideration of word meanings in analyzing the complexity or difficulty of L2 production based on the concern that it is difficult to determine the meanings of a linguistic item in learner texts. While this concern is warranted, the consideration of word meanings may be necessary in measuring difficulty. For example, based on production and interview data, Liu and Lu (2020) reported that their Chinese-speaking L2 English learners frequently used some but not other meanings of specific words (e.g., composition meaning a piece of writing vs. the way in which something is formed by its parts) due to their knowledge gaps. Lu and Hu (2022) further argued that different meanings of polysemous words may represent different levels of sophistication. They showed that it is possible to pinpoint the meanings of most polysemous words in learner texts and that sense-aware lexical sophistication indices can better predict learner proficiency than form-based ones. The cases in which the meanings of polysemous words are unclear (e.g., due to incorrect or creative usage) could constitute another useful dimension of analysis, as learner development involves learning not only to express more sophisticated meanings but also to express meanings unambiguously. Regarding the concern for accurate word sense disambiguation, large language models could now be fine-tuned to improve performance, and manual analysis can also be involved.

In sum, I am thankful for Bulté et al.’s contribution towards resolving the conceptual and methodological controversies around the constructs of complexity and difficulty, and I have raised a few further issues to be considered. Importantly, the complex and multidimensional nature of these constructs necessitates a set of complementary measures of different types, at different levels of granularity, and with sensitivity to features of different languages.

查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
在 "复杂性与难度 "中实现更清晰的概念:评述 "第二语言习得中的复杂性和难度:理论与方法概述
这篇由bult<s:1>、Housen和Pallotti撰写的概念性综述文章为理清第二语言习得研究中围绕复杂性和难度这两个关键概念的概念和方法混乱做出了急需的努力。为了在这些构念之间寻找清晰的概念区别,bult<e:1>等人将复杂性解释为语言项目的结构属性,将难度解释为与获取和使用这些项目相关的认知成本。他们进一步提出了一组可供研究人员日常使用的核心措施,以促进可复制性和知识积累。在这篇评论中,我对他们对这些构造的定义和度量的讨论提供了一些想法。在“复杂性”的不同定义中包含和排除“复杂性”(例如,Kyle等人,2021;Ortega, 2003),同时引入了术语“绝对复杂性”和“相对复杂性”(例如,Housen &amp;Simoens, 2016),导致了围绕复杂性的术语混淆。我同意bult<s:1>等人对复杂性的狭义解释,即等同于排除复杂性的绝对复杂性。同时,明确讨论(a)术语“难度”、“复杂性”和“相对复杂性”之间的确切关系,以及(b)术语“绝对复杂性”和“相对复杂性”的未来地位,将为该领域的术语提供更大的清晰度。bult<s:1>等人谈到了关于复杂性的定义和度量的两个重要争论。首先,他们独立于基于寄存器/类型的充分性概念来定义复杂性。语域/体裁变异研究对不同语域/体裁的语言特征提供了有价值的见解。然而,不同语域/体裁中语言项目的相对频率不应成为分析特定语域/体裁文本复杂性时包括或排除它们的依据。这是因为,虽然一个语言项目可能在一个语域/体裁中比在其他语域/体裁中更少出现,但它可能在该语域/体裁中发挥着不可或缺的作用,这种特定的作用应该在语域或体裁内部进行分析。其次,他们认为细粒度的复杂性分析方法与更全面的方法是互补的。在我看来,所有的复杂性度量都是在粒度尺度上的,批评整体度量(例如,每个子句的从属子句)缺乏专一性和可解释性,这有时被引用而没有经过彻底的考虑,同样适用于大多数细粒度度量(例如,每1000个单词的相对子句)。例如,正如有不同类型的从属分句一样,也有不同类型的关系分句,每一种关系分句又可能有子类型。事实上,粒度可以一直到词汇化。整体测量提供了一种评估高级复杂性的有效方法(例如,有限从属关系或名词修饰的出现或总体水平)。由于第二语言写作教师和研究人员可能对不同背景和不同目的下不同粒度水平的复杂性感兴趣,因此认识到整体和细粒度方法对复杂性分析的互补性是很重要的,类似于写作评估中使用的整体和分析评级量表。bult<e:1>等人没有将句子视为分析单位,而是将t单位或as单位视为最大的句法分析单位。这一建议引发了一场有趣的讨论。首先,句子是写作中直观有用的信息组织单位。其次,众所周知,小句协调的复杂性对初学者尤其重要(例如,Ortega, 2003)。第三,如果在我们的分析中忽略加标点的句子片段,我们可能会错过捕捉这些片段在学习者文本中有意的、适当的使用和潜在的无意的、错误的使用的机会。第四,t单元可能不是所有语言中合适的或最大的分析单元。例如,有人认为话题-评论单位比t -单位更适合分析中文的复杂性,一个终端话题-评论单位可以对应两个或多个t -单位(例如,Hu et al., 2022;余,2021)。对于句法复杂性的测量,bult<e:1>等人建议使用平均词每短语和平均短语每条款作为六个核心宪法措施中的两个。为了促进可复制性,似乎有必要就识别和计数短语的一致方法达成一致(例如,阅读一本关于古代中国的书中有多少短语?)对于词汇复杂性测量,bult<e:1>等人建议对屈折词形进行词法化,但对衍生词形的处理保持开放。 他们提出平均词长作为该构式的核心构成度量。然而,如果不将词汇复杂性定义为词族复杂性,那么每个单词的平均语素和(根)语素的平均长度可能会更好地与句法复杂性测量中捕获构成复杂性的努力相一致,因为单词是由语素组成的,然后由书面形式的字母或字符表示。bult<e:1>等人认为语言形式的规范化出现率是计算文本级复杂性的有效方法。这种类型的测量是已知的,以促进跨文本的语言形式的出现率的比较。然而,他们优先考虑语言形式的出现频率,而不是其内部复杂性。更重要的是,他们没有考虑语言形式在更大的单位中的分布,因此不能解释文本的结构组织。文本不是由n个词块组织的,很少显示语言形式的均匀分散。相反,特定的语言形式出现在更大的单位中,既包括直接包含它们的句法单位,也包括实现话语行为/功能的更大单位,这有助于整体文本结构。因此,基于标准化出现率的度量最好与其他类型的度量结合使用,这些度量可以捕获语言形式的内部复杂性及其在更大单位中的分布。bult<e:1>等人反对在分析第二语言生成的复杂性或难度时考虑词义,他们认为很难确定学习者文本中语言项目的含义。虽然这种担心是合理的,但在衡量难度时,考虑单词的含义可能是必要的。例如,基于生产和访谈数据,Liu和Lu(2020)报告说,由于他们的知识差距,说汉语的第二语言英语学习者经常使用特定单词的某些含义,而不是其他含义(例如,作文的意思是一篇文章,而不是由其各部分组成的方式)。Lu和Hu(2022)进一步认为,多义词的不同含义可能代表不同的复杂程度。他们表明,在学习者文本中,大多数多义词的意思都是有可能精确定位的,而且与基于形式的词汇复杂性指数相比,感知意义的词汇复杂性指数能更好地预测学习者的熟练程度。多义词含义不明确的情况(例如,由于不正确或创造性的用法)可以构成另一个有用的分析维度,因为学习者的发展不仅包括学习表达更复杂的含义,还包括学习明确地表达含义。考虑到对准确的词义消歧的关注,现在可以对大型语言模型进行微调以提高性能,并且还可以进行手动分析。总之,我很感谢bult<s:1>等人对解决围绕复杂性和难度的概念和方法上的争议所做的贡献,并且我提出了一些需要考虑的进一步问题。重要的是,这些结构的复杂性和多维性需要一组不同类型、不同粒度级别和对不同语言特性敏感的互补度量。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
Language Learning
Language Learning Multiple-
CiteScore
9.10
自引率
15.90%
发文量
65
期刊介绍: Language Learning is a scientific journal dedicated to the understanding of language learning broadly defined. It publishes research articles that systematically apply methods of inquiry from disciplines including psychology, linguistics, cognitive science, educational inquiry, neuroscience, ethnography, sociolinguistics, sociology, and anthropology. It is concerned with fundamental theoretical issues in language learning such as child, second, and foreign language acquisition, language education, bilingualism, literacy, language representation in mind and brain, culture, cognition, pragmatics, and intergroup relations. A subscription includes one or two annual supplements, alternating among a volume from the Language Learning Cognitive Neuroscience Series, the Currents in Language Learning Series or the Language Learning Special Issue Series.
期刊最新文献
Effects of Early Language Stays Abroad on the Foreign Language Proficiency of German High School Students Effective When Distinctive: The Role of Phonetic Similarity in Nested Dependency Learning Across Preschool Years What Is the Ideal Time to Provide Corrective Feedback? An Approximate Replication of Li, Zhu, and Ellis (2016) Hedgehog Pillows and Squirrel Plates: Priming Semantic Structure in Children's Comprehension Bimodal Input in Processing and Learning of Novel Vocabulary During Reading: An Eye‐Tracking Study
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1