{"title":"ChatGPT破坏了人类的反身性、科学责任和负责任的管理研究","authors":"Dirk Lindebaum, Peter Fleming","doi":"10.1111/1467-8551.12781","DOIUrl":null,"url":null,"abstract":"<p>With ChatGPT being promoted to and by academics for writing scholarly articles more effectively, we ask what kind of knowledge does ChatGPT produce, what this means for our reflexivity as <i>responsible</i> management educators/researchers, and how an absence of reflexivity disqualifies us from shaping management knowledge in <i>responsible</i> ways. We urgently need to grasp what makes human knowledge distinct compared with knowledge generated by ChatGPT <i>et al.</i> Thus, we first explain how ChatGPT operates and unpack its intrinsic epistemological limitations. Using high-probability choices that are derivative, ChatGPT has <i>no stake</i> in the knowledge it produces and is thus likely prone to offering irresponsible outputs. By contrast, genuine human thinking—embodied in a contingent socio-cultural setting—uses low-probability choices both ‘inside’ and ‘outside’ the box of training data, making it creative, contextual and committed. We conclude that the use of ChatGPT is wholly incompatible with scientific responsibility and responsible management.</p>","PeriodicalId":48342,"journal":{"name":"British Journal of Management","volume":"35 2","pages":"566-575"},"PeriodicalIF":4.5000,"publicationDate":"2023-11-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1111/1467-8551.12781","citationCount":"0","resultStr":"{\"title\":\"ChatGPT Undermines Human Reflexivity, Scientific Responsibility and Responsible Management Research\",\"authors\":\"Dirk Lindebaum, Peter Fleming\",\"doi\":\"10.1111/1467-8551.12781\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p>With ChatGPT being promoted to and by academics for writing scholarly articles more effectively, we ask what kind of knowledge does ChatGPT produce, what this means for our reflexivity as <i>responsible</i> management educators/researchers, and how an absence of reflexivity disqualifies us from shaping management knowledge in <i>responsible</i> ways. We urgently need to grasp what makes human knowledge distinct compared with knowledge generated by ChatGPT <i>et al.</i> Thus, we first explain how ChatGPT operates and unpack its intrinsic epistemological limitations. Using high-probability choices that are derivative, ChatGPT has <i>no stake</i> in the knowledge it produces and is thus likely prone to offering irresponsible outputs. By contrast, genuine human thinking—embodied in a contingent socio-cultural setting—uses low-probability choices both ‘inside’ and ‘outside’ the box of training data, making it creative, contextual and committed. We conclude that the use of ChatGPT is wholly incompatible with scientific responsibility and responsible management.</p>\",\"PeriodicalId\":48342,\"journal\":{\"name\":\"British Journal of Management\",\"volume\":\"35 2\",\"pages\":\"566-575\"},\"PeriodicalIF\":4.5000,\"publicationDate\":\"2023-11-20\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://onlinelibrary.wiley.com/doi/epdf/10.1111/1467-8551.12781\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"British Journal of Management\",\"FirstCategoryId\":\"91\",\"ListUrlMain\":\"https://onlinelibrary.wiley.com/doi/10.1111/1467-8551.12781\",\"RegionNum\":2,\"RegionCategory\":\"管理学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"BUSINESS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"British Journal of Management","FirstCategoryId":"91","ListUrlMain":"https://onlinelibrary.wiley.com/doi/10.1111/1467-8551.12781","RegionNum":2,"RegionCategory":"管理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"BUSINESS","Score":null,"Total":0}
ChatGPT Undermines Human Reflexivity, Scientific Responsibility and Responsible Management Research
With ChatGPT being promoted to and by academics for writing scholarly articles more effectively, we ask what kind of knowledge does ChatGPT produce, what this means for our reflexivity as responsible management educators/researchers, and how an absence of reflexivity disqualifies us from shaping management knowledge in responsible ways. We urgently need to grasp what makes human knowledge distinct compared with knowledge generated by ChatGPT et al. Thus, we first explain how ChatGPT operates and unpack its intrinsic epistemological limitations. Using high-probability choices that are derivative, ChatGPT has no stake in the knowledge it produces and is thus likely prone to offering irresponsible outputs. By contrast, genuine human thinking—embodied in a contingent socio-cultural setting—uses low-probability choices both ‘inside’ and ‘outside’ the box of training data, making it creative, contextual and committed. We conclude that the use of ChatGPT is wholly incompatible with scientific responsibility and responsible management.
期刊介绍:
The British Journal of Management provides a valuable outlet for research and scholarship on management-orientated themes and topics. It publishes articles of a multi-disciplinary and interdisciplinary nature as well as empirical research from within traditional disciplines and managerial functions. With contributions from around the globe, the journal includes articles across the full range of business and management disciplines. A subscription to British Journal of Management includes International Journal of Management Reviews, also published on behalf of the British Academy of Management.