大型语言模型基础设施管理说明

Q2 Computer Science First Monday Pub Date : 2024-02-11 DOI:10.5210/fm.v29i2.13567
Lara Dal Molin
{"title":"大型语言模型基础设施管理说明","authors":"Lara Dal Molin","doi":"10.5210/fm.v29i2.13567","DOIUrl":null,"url":null,"abstract":"This paper draws on information infrastructures (IIs) in science and technology studies (STS), as well as on feminist STS scholarship and contemporary critical accounts of digital technologies, to build an initial mapping of the infrastructural mechanisms and implications of large language models (LLMs). Through a comparison with discriminatory machine learning (ML) systems and a case study on gender bias, I present LLMs as contested artefacts with categorising and performative capabilities. This paper suggests that generative systems do not tangibly depart from traditional, discriminative counterparts in terms of their underlying probabilistic mechanisms, and therefore both technologies can be theorised as infrastructures of categorisation. However, LLMs additionally retain performative capabilities through their linguistic outputs. Here, I outline the intuition behind this phenomenon, which I refer to as “language as infrastructure”. While traditional, discriminative systems “disappear” into larger IIs, the hype surrounding generative technologies presents an opportunity to scrutinise these artefacts, to alter their computational mechanisms and introduce governance measures]. I illustrate this thesis through Sharma’s formulation of “broken machine”, and suggest dataset curation and participatory design as governance mechanisms that can partly address downstream harms in LLMs (Barocas, et al., 2023).","PeriodicalId":38833,"journal":{"name":"First Monday","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2024-02-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Notes towards infrastructure governance for large language models\",\"authors\":\"Lara Dal Molin\",\"doi\":\"10.5210/fm.v29i2.13567\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"This paper draws on information infrastructures (IIs) in science and technology studies (STS), as well as on feminist STS scholarship and contemporary critical accounts of digital technologies, to build an initial mapping of the infrastructural mechanisms and implications of large language models (LLMs). Through a comparison with discriminatory machine learning (ML) systems and a case study on gender bias, I present LLMs as contested artefacts with categorising and performative capabilities. This paper suggests that generative systems do not tangibly depart from traditional, discriminative counterparts in terms of their underlying probabilistic mechanisms, and therefore both technologies can be theorised as infrastructures of categorisation. However, LLMs additionally retain performative capabilities through their linguistic outputs. Here, I outline the intuition behind this phenomenon, which I refer to as “language as infrastructure”. While traditional, discriminative systems “disappear” into larger IIs, the hype surrounding generative technologies presents an opportunity to scrutinise these artefacts, to alter their computational mechanisms and introduce governance measures]. I illustrate this thesis through Sharma’s formulation of “broken machine”, and suggest dataset curation and participatory design as governance mechanisms that can partly address downstream harms in LLMs (Barocas, et al., 2023).\",\"PeriodicalId\":38833,\"journal\":{\"name\":\"First Monday\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-02-11\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"First Monday\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.5210/fm.v29i2.13567\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q2\",\"JCRName\":\"Computer Science\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"First Monday","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.5210/fm.v29i2.13567","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"Computer Science","Score":null,"Total":0}
引用次数: 0

摘要

本文借鉴了科技研究(STS)中的信息基础架构(IIs),以及女权主义科技研究学术成果和当代数字技术批判性论述,初步描绘了大型语言模型(LLMs)的基础架构机制和影响。通过与歧视性机器学习(ML)系统和性别偏见案例研究的比较,我将 LLMs 表述为具有分类和执行能力的有争议的人工制品。本文认为,生成系统在其基本概率机制方面与传统的歧视性系统并无明显区别,因此这两种技术都可以被理论化为分类的基础设施。不过,LLM 还通过其语言输出保留了表演能力。在此,我将概述这一现象背后的直觉,并将其称为 "作为基础设施的语言"。当传统的辨别系统 "消失 "在更大的 IIs 中时,围绕生成技术的炒作提供了一个审视这些人工制品、改变其计算机制并引入治理措施的机会]。我通过夏尔马提出的 "坏掉的机器 "来说明这一论点,并建议将数据集整理和参与式设计作为治理机制,以部分解决 LLM 中的下游危害(Barocas, et al.)
本文章由计算机程序翻译,如有差异,请以英文原文为准。
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
Notes towards infrastructure governance for large language models
This paper draws on information infrastructures (IIs) in science and technology studies (STS), as well as on feminist STS scholarship and contemporary critical accounts of digital technologies, to build an initial mapping of the infrastructural mechanisms and implications of large language models (LLMs). Through a comparison with discriminatory machine learning (ML) systems and a case study on gender bias, I present LLMs as contested artefacts with categorising and performative capabilities. This paper suggests that generative systems do not tangibly depart from traditional, discriminative counterparts in terms of their underlying probabilistic mechanisms, and therefore both technologies can be theorised as infrastructures of categorisation. However, LLMs additionally retain performative capabilities through their linguistic outputs. Here, I outline the intuition behind this phenomenon, which I refer to as “language as infrastructure”. While traditional, discriminative systems “disappear” into larger IIs, the hype surrounding generative technologies presents an opportunity to scrutinise these artefacts, to alter their computational mechanisms and introduce governance measures]. I illustrate this thesis through Sharma’s formulation of “broken machine”, and suggest dataset curation and participatory design as governance mechanisms that can partly address downstream harms in LLMs (Barocas, et al., 2023).
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
First Monday
First Monday Computer Science-Computer Networks and Communications
CiteScore
2.20
自引率
0.00%
发文量
86
期刊介绍: First Monday is one of the first openly accessible, peer–reviewed journals on the Internet, solely devoted to the Internet. Since its start in May 1996, First Monday has published 1,035 papers in 164 issues; these papers were written by 1,316 different authors. In addition, eight special issues have appeared. The most recent special issue was entitled A Web site with a view — The Third World on First Monday and it was edited by Eduardo Villanueva Mansilla. First Monday is indexed in Communication Abstracts, Computer & Communications Security Abstracts, DoIS, eGranary Digital Library, INSPEC, Information Science & Technology Abstracts, LISA, PAIS, and other services.
期刊最新文献
French-speaking photo models communication: A comparison across platforms and profiles, a possible evolution Angry sharing: Exploring the influence of Facebook reactions on political post sharing Everyday positivity: An appraisal analysis of online identity in food blogs Tweeting on thin ice: Scientists in dialogic climate change communication with the public Education runs quickly violence runs slowly: An analysis of closed captioning speed and reading level in children’s television franchises
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1