Zur Operationalisierung literaturwissenschaftlicher Begriffe in der algorithmischen Textanalyse. Eine Annäherung über Norbert Altenhofers hermeneutische Modellinterpretation von KleistsDas Erdbeben in Chili

IF 0.6 0 LITERARY THEORY & CRITICISM Journal of Literary Theory Pub Date : 2021-11-06 DOI:10.1515/jlt-2021-2008
A. Pichler, Nils Reiter
{"title":"Zur Operationalisierung literaturwissenschaftlicher Begriffe in der algorithmischen Textanalyse. Eine Annäherung über Norbert Altenhofers hermeneutische Modellinterpretation von KleistsDas Erdbeben in Chili","authors":"A. Pichler, Nils Reiter","doi":"10.1515/jlt-2021-2008","DOIUrl":null,"url":null,"abstract":"Abstract The present article discusses and reflects on possible ways of operationalizing the terminology of traditional literary studies for use in computational literary studies. By »operationalization«, we mean the development of a method for tracing a (theoretical) term back to text-surface phenomena; this is done explicitly and in a rule-based manner, involving a series of substeps. This procedure is presented in detail using as a concrete example Norbert Altenhofer’s »model interpretation« (Modellinterpretation) of Heinrich von Kleist’s The Earthquake in Chile. In the process, we develop a multi-stage operation – reflected upon throughout in terms of its epistemological implications – that is based on a rational-hermeneutic reconstruction of Altenhofer’s interpretation, which focuses on »mysteriousness« (Rätselhaftigkeit), a concept from everyday language. As we go on to demonstrate, when trying to operationalize this term, one encounters numerous difficulties, which is owing to the fact that Altenhofer’s use of it is underspecified in a number of ways. Thus, for instance, and contrary to Altenhofer’s suggestion, Kleist’s sentences containing »relativizing or perspectivizing phrases such as ›it seemed‹ or ›it was as if‹« (Altenhofer 2007, 45) do by no means, when analyzed linguistically, suggest a questioning or challenge of the events narrated, since the unreal quality of those German sentences only relates to the comparison in the subordinate clause, not to the respective main clause. Another indicator central to Altenhofer’s ascription of »mysteriousness« is his concept of a »complete facticity« (lückenlose Faktizität) which »does not seem to leave anything ›open‹« (Altenhofer 2007, 45). Again, the precise designation of what exactly qualifies facticity as »complete« is left open, since Kleist’s novella does indeed select for portrayal certain phenomena and actions within the narrated world (and not others). The degree of factuality in Kleist’s text may be higher than it is in other texts, but it is by no means »complete«. In the context of Altenhofer’s interpretation, »complete facticity« may be taken to mean a narrative mode in which terrible events are reported using conspicuously sober and at times drastic language. Following the critical reconstruction of Altenhofer’s use of terminology, the central terms and their relationship to one another are first explicated (in natural language), which already necessitates intensive conceptual work. We do so implementing a hierarchical understanding of the terms discussed: the definition of one term uses other terms which also need to be defined and operationalized. In accordance with the requirements of computational text analysis, this hierarchy of terms should end in »directly measurable« terms – i. e., in terms that can be clearly identified on the surface of the text. This, however, leads to the question of whether (and, if so, on the basis of which theoretical assumptions) the terminology of literary studies may be traced back in this way to text-surface phenomena. Following the pragmatic as well as the theoretical discussion of this complex of questions, we indicate ways by which such definitions may be converted into manual or automatic recognition. In the case of manual recognition, the paradigm of annotation – as established and methodologically reflected in (computational) linguistics – will be useful, and a well-controlled annotation process will help to further clarify the terms in question. The primary goal, however, is to establish a recognition rule by which individuals may intersubjectively and reliably identify instances of the term in question in a given text. While it is true that in applying this method to literary studies, new challenges arise – such as the question of the validity and reliability of the annotations –, these challenges are at present being researched intensively in the field of computational literary studies, which has resulted in a large and growing body of research to draw on. In terms of computer-aided recognition, we examine, by way of example, two distinct approaches: 1) The kind of operationalization which is guided by precedent definitions and annotation rules benefits from the fact that each of its steps is transparent, may be validated and interpreted, and that existing tools from computational linguistics can be integrated into the process. In the scenario used here, these would be tools for recognizing and assigning character speech, for the resolution of coreference and the assessment of events; all of these, in turn, may be based on either machine learning, prescribed rules or dictionaries. 2) In recent years, so-called end-to-end systems have become popular which, with the help of neural networks, »infer« target terms directly from a numerical representation of the data. These systems achieve superior results in many areas. However, their lack of transparency also raises new questions, especially with regard to the interpretation of results. Finally, we discuss options for quality assurance and draw a first conclusion. Since numerous decisions have to be made in the course of operationalization, and these, in practice, are often pragmatically justified, the question quickly arises as to how »good« a given operationalization actually is. And since the tools borrowed from computational linguistics (especially the so-called inter-annotator agreement) can only partially be transferred to computational literary studies and, moreover, objective standards for the quality of a given implementation will be difficult to find, it ultimately falls to the community of researchers and scholars to decide, based on their research standards, which operationalizations they accept. At the same time, operationalization is the central link between the computer sciences and literary studies, as well as being a necessary component for a large part of the research done in computational literary studies. The advantage of a conscious, deliberate and reflective operationalization practice lies not only in the fact that it can be used to achieve reliable quantitative results (or that a certain lack of reliability at least is a known factor); it also lies in its facilitation of interdisciplinary cooperation: in the course of operationalization, concrete sets of data are discussed, as are the methods for analysing them, which taken together minimizes the risk of misunderstandings, »false friends« and of an unproductive exchange more generally.","PeriodicalId":42872,"journal":{"name":"Journal of Literary Theory","volume":null,"pages":null},"PeriodicalIF":0.6000,"publicationDate":"2021-11-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"2","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Literary Theory","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1515/jlt-2021-2008","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"0","JCRName":"LITERARY THEORY & CRITICISM","Score":null,"Total":0}
引用次数: 2

Abstract

Abstract The present article discusses and reflects on possible ways of operationalizing the terminology of traditional literary studies for use in computational literary studies. By »operationalization«, we mean the development of a method for tracing a (theoretical) term back to text-surface phenomena; this is done explicitly and in a rule-based manner, involving a series of substeps. This procedure is presented in detail using as a concrete example Norbert Altenhofer’s »model interpretation« (Modellinterpretation) of Heinrich von Kleist’s The Earthquake in Chile. In the process, we develop a multi-stage operation – reflected upon throughout in terms of its epistemological implications – that is based on a rational-hermeneutic reconstruction of Altenhofer’s interpretation, which focuses on »mysteriousness« (Rätselhaftigkeit), a concept from everyday language. As we go on to demonstrate, when trying to operationalize this term, one encounters numerous difficulties, which is owing to the fact that Altenhofer’s use of it is underspecified in a number of ways. Thus, for instance, and contrary to Altenhofer’s suggestion, Kleist’s sentences containing »relativizing or perspectivizing phrases such as ›it seemed‹ or ›it was as if‹« (Altenhofer 2007, 45) do by no means, when analyzed linguistically, suggest a questioning or challenge of the events narrated, since the unreal quality of those German sentences only relates to the comparison in the subordinate clause, not to the respective main clause. Another indicator central to Altenhofer’s ascription of »mysteriousness« is his concept of a »complete facticity« (lückenlose Faktizität) which »does not seem to leave anything ›open‹« (Altenhofer 2007, 45). Again, the precise designation of what exactly qualifies facticity as »complete« is left open, since Kleist’s novella does indeed select for portrayal certain phenomena and actions within the narrated world (and not others). The degree of factuality in Kleist’s text may be higher than it is in other texts, but it is by no means »complete«. In the context of Altenhofer’s interpretation, »complete facticity« may be taken to mean a narrative mode in which terrible events are reported using conspicuously sober and at times drastic language. Following the critical reconstruction of Altenhofer’s use of terminology, the central terms and their relationship to one another are first explicated (in natural language), which already necessitates intensive conceptual work. We do so implementing a hierarchical understanding of the terms discussed: the definition of one term uses other terms which also need to be defined and operationalized. In accordance with the requirements of computational text analysis, this hierarchy of terms should end in »directly measurable« terms – i. e., in terms that can be clearly identified on the surface of the text. This, however, leads to the question of whether (and, if so, on the basis of which theoretical assumptions) the terminology of literary studies may be traced back in this way to text-surface phenomena. Following the pragmatic as well as the theoretical discussion of this complex of questions, we indicate ways by which such definitions may be converted into manual or automatic recognition. In the case of manual recognition, the paradigm of annotation – as established and methodologically reflected in (computational) linguistics – will be useful, and a well-controlled annotation process will help to further clarify the terms in question. The primary goal, however, is to establish a recognition rule by which individuals may intersubjectively and reliably identify instances of the term in question in a given text. While it is true that in applying this method to literary studies, new challenges arise – such as the question of the validity and reliability of the annotations –, these challenges are at present being researched intensively in the field of computational literary studies, which has resulted in a large and growing body of research to draw on. In terms of computer-aided recognition, we examine, by way of example, two distinct approaches: 1) The kind of operationalization which is guided by precedent definitions and annotation rules benefits from the fact that each of its steps is transparent, may be validated and interpreted, and that existing tools from computational linguistics can be integrated into the process. In the scenario used here, these would be tools for recognizing and assigning character speech, for the resolution of coreference and the assessment of events; all of these, in turn, may be based on either machine learning, prescribed rules or dictionaries. 2) In recent years, so-called end-to-end systems have become popular which, with the help of neural networks, »infer« target terms directly from a numerical representation of the data. These systems achieve superior results in many areas. However, their lack of transparency also raises new questions, especially with regard to the interpretation of results. Finally, we discuss options for quality assurance and draw a first conclusion. Since numerous decisions have to be made in the course of operationalization, and these, in practice, are often pragmatically justified, the question quickly arises as to how »good« a given operationalization actually is. And since the tools borrowed from computational linguistics (especially the so-called inter-annotator agreement) can only partially be transferred to computational literary studies and, moreover, objective standards for the quality of a given implementation will be difficult to find, it ultimately falls to the community of researchers and scholars to decide, based on their research standards, which operationalizations they accept. At the same time, operationalization is the central link between the computer sciences and literary studies, as well as being a necessary component for a large part of the research done in computational literary studies. The advantage of a conscious, deliberate and reflective operationalization practice lies not only in the fact that it can be used to achieve reliable quantitative results (or that a certain lack of reliability at least is a known factor); it also lies in its facilitation of interdisciplinary cooperation: in the course of operationalization, concrete sets of data are discussed, as are the methods for analysing them, which taken together minimizes the risk of misunderstandings, »false friends« and of an unproductive exchange more generally.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
用算法文字分析中的文学论文的操作方法这里有一个关于
摘要本文讨论并反思了在计算文学研究中使用传统文学研究术语的可能方法。通过“操作化”,我们指的是开发一种方法,将(理论)术语追溯回文本表面现象;这是以基于规则的方式显式完成的,涉及一系列子步骤。以Norbert Altenhofer对Heinrich von Kleist的《智利地震》的“模型解释”(modelinterpretation)为例,详细介绍了这一过程。在这个过程中,我们发展了一个多阶段的操作——从其认识论的意义上反映出来——这是基于对Altenhofer的解释的理性解释学重建,其重点是“神秘”(Rätselhaftigkeit),这是一个来自日常语言的概念。正如我们接下来所展示的,当试图操作这个术语时,人们会遇到许多困难,这是由于Altenhofer对它的使用在许多方面都没有明确规定。因此,例如,与Altenhofer的建议相反,克莱斯特的句子包含“相对化或透视化”的短语,如“it seems”或“it was as if”(Altenhofer 2007, 45),当从语言学上分析时,绝不意味着对所叙述的事件提出质疑或挑战,因为这些德语句子的不真实性质只与从句中的比较有关,而不是与各自的主句有关。Altenhofer对“神秘性”的归属的另一个核心指标是他的“完全事实性”(l<s:1> ckenlose Faktizität)的概念,“似乎没有留下任何“开放”(Altenhofer 2007,45)。再一次,关于什么是“完整”的真实性的确切定义是开放的,因为克莱斯特的中篇小说确实选择了描述所叙述的世界中的某些现象和行为(而不是其他)。克莱斯特的文本中的真实性程度可能高于其他文本,但它绝不是“完整的”。在Altenhofer的解释中,“完全的真实性”可以被理解为一种叙事模式,在这种模式中,可怕的事件被用明显清醒的,有时甚至是激烈的语言报道。在Altenhofer对术语使用的批判性重建之后,首先(用自然语言)解释了中心术语及其彼此之间的关系,这已经需要密集的概念工作。我们这样做是为了实现对所讨论的术语的层次理解:一个术语的定义使用其他术语,这些术语也需要定义和操作化。根据计算文本分析的要求,这个术语层次应该以“直接可测量的”术语结束。,这些术语可以在文本的表面上清楚地识别出来。然而,这导致了一个问题,即文学研究的术语是否(如果是,基于哪些理论假设)可以以这种方式追溯到文本表面现象。在对这一复杂问题的语用和理论讨论之后,我们指出了将这些定义转换为手动或自动识别的方法。在人工识别的情况下,注释的范例——在(计算)语言学中建立和方法上的反映——将是有用的,一个控制良好的注释过程将有助于进一步澄清所讨论的术语。然而,主要目标是建立一个识别规则,通过该规则,个人可以主观地和可靠地识别给定文本中有关术语的实例。虽然在将这种方法应用于文学研究时确实会出现新的挑战-例如注释的有效性和可靠性问题-这些挑战目前正在计算文学研究领域进行深入研究,这导致了大量且不断增长的研究机构可供借鉴。在计算机辅助识别方面,我们通过示例来研究两种不同的方法:1)由先例定义和注释规则指导的操作化类型受益于其每个步骤都是透明的,可以验证和解释,并且计算语言学的现有工具可以集成到该过程中。在这里使用的场景中,这些工具将用于识别和分配角色语音、解决相互参照和评估事件;反过来,所有这些都可能基于机器学习、规定规则或字典。2)近年来,所谓的端到端系统变得流行起来,它在神经网络的帮助下,直接从数据的数字表示中“推断”目标术语。这些系统在许多领域取得了优异的效果。 但是,它们缺乏透明度也提出了新的问题,特别是在对结果的解释方面。最后,我们讨论了质量保证的选择,并得出了第一个结论。由于在操作化过程中必须做出许多决策,而这些决策在实践中通常是合理的,因此很快就会出现一个问题,即给定的操作化实际上有多“好”。而且,由于从计算语言学借来的工具(尤其是所谓的注释者间协议)只能部分地转移到计算文学研究中,而且,很难找到给定实现质量的客观标准,因此最终取决于研究人员和学者社区,根据他们的研究标准来决定他们接受哪些操作化。同时,操作化是计算机科学和文学研究之间的中心环节,也是计算文学研究中大部分研究的必要组成部分。有意识的、深思熟虑的和反思性的操作化做法的优点不仅在于它可以用来获得可靠的定量结果(或者至少某种程度上缺乏可靠性是一个已知的因素);它还在于促进跨学科合作:在操作过程中,讨论了具体的数据集,以及分析它们的方法,这些数据集结合在一起,最大限度地减少了误解、“假朋友”和更普遍的非生产性交流的风险。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
Journal of Literary Theory
Journal of Literary Theory LITERARY THEORY & CRITICISM-
自引率
0.00%
发文量
19
期刊最新文献
Die Autonomie der Literatur auf dem Prüfstand. Bourdieus feldtheoretischer Ansatz als Alternative zu soziologistischen Kurzschlüssen Experiencing Literary Audiobooks: A Framework for Theoretical and Empirical Investigations of the Auditory Reception of Literature Autor und Subjekt im lyrischen Gedicht: Rezension und Neukonzeption einer Theorie der lyrischen Persona Die Literaturautonomie im deutschen Rechtssystem. Grenzen, Widersprüche und literaturtheoretische Potenziale Ästhetische Autonomie zwischen Ethik und Ästhetik
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1