{"title":"文学控制论:矛的尖","authors":"N. Katherine Hayles","doi":"10.1353/nlh.2023.a907175","DOIUrl":null,"url":null,"abstract":"Literary Cybernetics:The Point (of the Spear) N. Katherine Hayles (bio) Cybernetics and literary studies are on a collision course that will transform what it means to read, to write, and to be human. The essays on \"literary cybernetics\" in this issue touch on many different ways in which this phrase can be interpreted, but for all their rich variety, they do not entirely capture either the urgency of our present situation or the inevitability of the coming transformations, even if we do not know and cannot reliably predict exactly what new forms will emerge, postcollision. At the pointy end of the spear driving into the heart of literary studies are the large language models (LLMs) created by rich tech companies and increasingly available to the general public such as OpenAI's GPT-3, -4, and ChatGPT (Generative Pretrained Transformer, versions -3 and -4), Google's LaMDA (Language Model for Dialogue Applications), and Google's BERT (Bidirectional Encoder Representations from Transformers). GPT-3, for example, was trained on forty-five terabytes of human-authored texts, mostly scraped from the web (a terabyte of data would fill about 570 million pages).1 Reading for GPT-4 is very different than for a human child learning to decode letters. Words are broken into tokens (generally word fragments of about four letters) and transformed into vectors processed through its ninety-six layers of neurons. This generates probability matrixes which are then processed further through a software function such as Softmax and output as words. The outputs are probabilistic projections of what the next word (or series of words) in a sequence would be.2 These models instantiate many of the cybernetic concepts discussed in this volume's essays, including the qualities emphasized in Paul Jaussen's \"The Art of Distinction.\" They include recursivity (outputs are fed back into the model as inputs) and environment/system distinctions. Originally the training set of texts constitutes an LLM's environment, but as the model learns, assumptions in the data set are absorbed into the continuously adjusted weights of the different neuron layers; after training, the model's interactions with human interlocutors constitute another kind of environment/system distinction. So intricate are the model's instantiations of these and other [End Page 1289] cybernetic ideas that the models can legitimately be called the ultimate cybernetic machines. The textual outputs of GPT-4 are several orders of magnitude more sophisticated than those produced by chatbots such as Siri and earlier algorithmic text-programs. GPT-4's texts are not only syntactically correct and semantically coherent; they often also demonstrate dazzlingly complex turns of rhetoric and logic. Indeed, they are often so good that they cannot be reliably distinguished from human-authored texts. From the technical description above, it would be quite surprising to discover that GPT-4 can discern and reproduce literary styles as diverse as the King James Bible and Mark Twain, and even more surprising to find that it can also identify and reproduce high-level literary qualities such as genre. Essays written by GPT-3 have been submitted to college professors for evaluation as if they were written by undergraduate students and have received excellent grades (A and A-), along with positive comments.3 As Lea Pao notes in \"Ways of Cybernetic Thinking,\" one problem that immediately leaps to mind for educators is the chaos this will cause for humanities courses, where the typical way in which students demonstrate their abilities to synthesize and interpret what they have learned is by having them write essays on assigned topics. As disruptive as this looming deluge of plagiarism may be, in a philosophical sense it is the least of the transformative effects. More fundamental are debates over what kinds of meaning machine-produced texts can have, and indeed if they can be said to have any meaning at all, beyond what human readers project into them. For literary studies, the blurring of boundaries between human cognition and machine learning means that assumptions about the uniqueness of the human ability to learn, use, and manipulate symbolic abstractions and languages are immediately drawn into question.4 Already machine-generated texts are ubiquitous in formulaic prose such as sports...","PeriodicalId":19150,"journal":{"name":"New Literary History","volume":"197 1","pages":"0"},"PeriodicalIF":0.8000,"publicationDate":"2023-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Literary Cybernetics: The Point (of the Spear)\",\"authors\":\"N. Katherine Hayles\",\"doi\":\"10.1353/nlh.2023.a907175\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Literary Cybernetics:The Point (of the Spear) N. Katherine Hayles (bio) Cybernetics and literary studies are on a collision course that will transform what it means to read, to write, and to be human. The essays on \\\"literary cybernetics\\\" in this issue touch on many different ways in which this phrase can be interpreted, but for all their rich variety, they do not entirely capture either the urgency of our present situation or the inevitability of the coming transformations, even if we do not know and cannot reliably predict exactly what new forms will emerge, postcollision. At the pointy end of the spear driving into the heart of literary studies are the large language models (LLMs) created by rich tech companies and increasingly available to the general public such as OpenAI's GPT-3, -4, and ChatGPT (Generative Pretrained Transformer, versions -3 and -4), Google's LaMDA (Language Model for Dialogue Applications), and Google's BERT (Bidirectional Encoder Representations from Transformers). GPT-3, for example, was trained on forty-five terabytes of human-authored texts, mostly scraped from the web (a terabyte of data would fill about 570 million pages).1 Reading for GPT-4 is very different than for a human child learning to decode letters. Words are broken into tokens (generally word fragments of about four letters) and transformed into vectors processed through its ninety-six layers of neurons. This generates probability matrixes which are then processed further through a software function such as Softmax and output as words. The outputs are probabilistic projections of what the next word (or series of words) in a sequence would be.2 These models instantiate many of the cybernetic concepts discussed in this volume's essays, including the qualities emphasized in Paul Jaussen's \\\"The Art of Distinction.\\\" They include recursivity (outputs are fed back into the model as inputs) and environment/system distinctions. Originally the training set of texts constitutes an LLM's environment, but as the model learns, assumptions in the data set are absorbed into the continuously adjusted weights of the different neuron layers; after training, the model's interactions with human interlocutors constitute another kind of environment/system distinction. So intricate are the model's instantiations of these and other [End Page 1289] cybernetic ideas that the models can legitimately be called the ultimate cybernetic machines. The textual outputs of GPT-4 are several orders of magnitude more sophisticated than those produced by chatbots such as Siri and earlier algorithmic text-programs. GPT-4's texts are not only syntactically correct and semantically coherent; they often also demonstrate dazzlingly complex turns of rhetoric and logic. Indeed, they are often so good that they cannot be reliably distinguished from human-authored texts. From the technical description above, it would be quite surprising to discover that GPT-4 can discern and reproduce literary styles as diverse as the King James Bible and Mark Twain, and even more surprising to find that it can also identify and reproduce high-level literary qualities such as genre. Essays written by GPT-3 have been submitted to college professors for evaluation as if they were written by undergraduate students and have received excellent grades (A and A-), along with positive comments.3 As Lea Pao notes in \\\"Ways of Cybernetic Thinking,\\\" one problem that immediately leaps to mind for educators is the chaos this will cause for humanities courses, where the typical way in which students demonstrate their abilities to synthesize and interpret what they have learned is by having them write essays on assigned topics. As disruptive as this looming deluge of plagiarism may be, in a philosophical sense it is the least of the transformative effects. More fundamental are debates over what kinds of meaning machine-produced texts can have, and indeed if they can be said to have any meaning at all, beyond what human readers project into them. For literary studies, the blurring of boundaries between human cognition and machine learning means that assumptions about the uniqueness of the human ability to learn, use, and manipulate symbolic abstractions and languages are immediately drawn into question.4 Already machine-generated texts are ubiquitous in formulaic prose such as sports...\",\"PeriodicalId\":19150,\"journal\":{\"name\":\"New Literary History\",\"volume\":\"197 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.8000,\"publicationDate\":\"2023-03-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"New Literary History\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1353/nlh.2023.a907175\",\"RegionNum\":2,\"RegionCategory\":\"文学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"0\",\"JCRName\":\"LITERATURE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"New Literary History","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1353/nlh.2023.a907175","RegionNum":2,"RegionCategory":"文学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"0","JCRName":"LITERATURE","Score":null,"Total":0}
引用次数: 0
摘要
N.凯瑟琳·海尔斯(生物)控制论和文学研究正处于碰撞的过程中,这将改变人们阅读、写作和做人的意义。本期关于“文学控制论”的文章触及了许多不同的方式来解释这个短语,但尽管它们种类繁多,它们并没有完全捕捉到我们当前形势的紧迫性,也没有完全捕捉到即将到来的变革的必然性,即使我们不知道也不能可靠地准确预测碰撞后会出现什么样的新形式。文学研究的核心是由富有的科技公司创建的大型语言模型(llm),这些模型越来越多地为公众所使用,比如OpenAI的GPT-3、-4和ChatGPT(生成预训练的变形金刚,版本-3和-4),谷歌的LaMDA(对话应用的语言模型)和谷歌的BERT(变形金刚的双向编码器表示)。例如,GPT-3是在45太字节的人类撰写的文本上进行训练的,这些文本大多是从网络上抓取的(1太字节的数据将填满大约5.7亿页)GPT-4的阅读与人类儿童学习解码字母的阅读非常不同。单词被分解成符号(通常是大约四个字母的单词片段),并通过它的96层神经元转换成向量。这将生成概率矩阵,然后通过软件功能(如Softmax)进一步处理并输出为单词。输出是序列中下一个单词(或一系列单词)的概率预测这些模型实例化了本卷文章中讨论的许多控制论概念,包括保罗·詹森(Paul Jaussen)的《区分的艺术》(the Art of Distinction)中强调的品质。它们包括递归性(输出作为输入反馈到模型中)和环境/系统差异。最初文本的训练集构成了LLM的环境,但随着模型的学习,数据集中的假设被吸收到不断调整的不同神经元层的权重中;经过训练后,模型与人类对话者的交互构成了另一种环境/系统区分。这些和其他控制论思想的模型实例是如此复杂,以至于这些模型可以被合理地称为终极控制论机器。GPT-4的文本输出比Siri等聊天机器人和早期的算法文本程序产生的文本要复杂好几个数量级。GPT-4的文本不仅句法正确,语义连贯;他们还经常展示出令人眼花缭乱的复杂修辞和逻辑转折。事实上,它们通常非常好,以至于无法可靠地将它们与人类撰写的文本区分开来。从上面的技术描述中,我们会很惊讶地发现GPT-4可以识别和复制像钦差版圣经和马克吐温这样多样化的文学风格,更令人惊讶的是它还可以识别和复制像流派这样的高级文学品质。2 . GPT-3的论文已经像本科生一样提交给大学教授进行评估,并获得了优异的成绩(A和A-),并得到了积极的评价正如Lea Pao在《控制论思维的方式》(Ways of Cybernetic Thinking)一书中所指出的那样,教育工作者立即想到的一个问题是,这将给人文学科课程带来混乱,在人文学科课程中,学生展示他们综合和解释所学知识的能力的典型方式是让他们就指定的主题写文章。尽管这种若隐若现的抄袭浪潮可能具有破坏性,但从哲学意义上讲,它是最不具变革性的影响。更根本的争论是,机器生成的文本可以有什么样的意义,实际上,如果它们可以说有任何意义,超出了人类读者投射到它们身上的意义。对于文学研究来说,人类认知和机器学习之间界限的模糊意味着,关于人类学习、使用和操纵符号抽象和语言的能力的独特性的假设立即被纳入了问题机器生成的文本在体育等公式化的文体中已经无处不在……
Literary Cybernetics:The Point (of the Spear) N. Katherine Hayles (bio) Cybernetics and literary studies are on a collision course that will transform what it means to read, to write, and to be human. The essays on "literary cybernetics" in this issue touch on many different ways in which this phrase can be interpreted, but for all their rich variety, they do not entirely capture either the urgency of our present situation or the inevitability of the coming transformations, even if we do not know and cannot reliably predict exactly what new forms will emerge, postcollision. At the pointy end of the spear driving into the heart of literary studies are the large language models (LLMs) created by rich tech companies and increasingly available to the general public such as OpenAI's GPT-3, -4, and ChatGPT (Generative Pretrained Transformer, versions -3 and -4), Google's LaMDA (Language Model for Dialogue Applications), and Google's BERT (Bidirectional Encoder Representations from Transformers). GPT-3, for example, was trained on forty-five terabytes of human-authored texts, mostly scraped from the web (a terabyte of data would fill about 570 million pages).1 Reading for GPT-4 is very different than for a human child learning to decode letters. Words are broken into tokens (generally word fragments of about four letters) and transformed into vectors processed through its ninety-six layers of neurons. This generates probability matrixes which are then processed further through a software function such as Softmax and output as words. The outputs are probabilistic projections of what the next word (or series of words) in a sequence would be.2 These models instantiate many of the cybernetic concepts discussed in this volume's essays, including the qualities emphasized in Paul Jaussen's "The Art of Distinction." They include recursivity (outputs are fed back into the model as inputs) and environment/system distinctions. Originally the training set of texts constitutes an LLM's environment, but as the model learns, assumptions in the data set are absorbed into the continuously adjusted weights of the different neuron layers; after training, the model's interactions with human interlocutors constitute another kind of environment/system distinction. So intricate are the model's instantiations of these and other [End Page 1289] cybernetic ideas that the models can legitimately be called the ultimate cybernetic machines. The textual outputs of GPT-4 are several orders of magnitude more sophisticated than those produced by chatbots such as Siri and earlier algorithmic text-programs. GPT-4's texts are not only syntactically correct and semantically coherent; they often also demonstrate dazzlingly complex turns of rhetoric and logic. Indeed, they are often so good that they cannot be reliably distinguished from human-authored texts. From the technical description above, it would be quite surprising to discover that GPT-4 can discern and reproduce literary styles as diverse as the King James Bible and Mark Twain, and even more surprising to find that it can also identify and reproduce high-level literary qualities such as genre. Essays written by GPT-3 have been submitted to college professors for evaluation as if they were written by undergraduate students and have received excellent grades (A and A-), along with positive comments.3 As Lea Pao notes in "Ways of Cybernetic Thinking," one problem that immediately leaps to mind for educators is the chaos this will cause for humanities courses, where the typical way in which students demonstrate their abilities to synthesize and interpret what they have learned is by having them write essays on assigned topics. As disruptive as this looming deluge of plagiarism may be, in a philosophical sense it is the least of the transformative effects. More fundamental are debates over what kinds of meaning machine-produced texts can have, and indeed if they can be said to have any meaning at all, beyond what human readers project into them. For literary studies, the blurring of boundaries between human cognition and machine learning means that assumptions about the uniqueness of the human ability to learn, use, and manipulate symbolic abstractions and languages are immediately drawn into question.4 Already machine-generated texts are ubiquitous in formulaic prose such as sports...
期刊介绍:
New Literary History focuses on questions of theory, method, interpretation, and literary history. Rather than espousing a single ideology or intellectual framework, it canvasses a wide range of scholarly concerns. By examining the bases of criticism, the journal provokes debate on the relations between literary and cultural texts and present needs. A major international forum for scholarly exchange, New Literary History has received six awards from the Council of Editors of Learned Journals.