What's in the black box? How algorithmic knowledge promotes corrective and restrictive actions to counter misinformation in the USA, the UK, South Korea and Mexico
{"title":"What's in the black box? How algorithmic knowledge promotes corrective and restrictive actions to counter misinformation in the USA, the UK, South Korea and Mexico","authors":"Myojung Chung","doi":"10.1108/intr-07-2022-0578","DOIUrl":null,"url":null,"abstract":"PurposeWhile there has been a growing call for insights on algorithms given their impact on what people encounter on social media, it remains unknown how enhanced algorithmic knowledge serves as a countermeasure to problematic information flow. To fill this gap, this study aims to investigate how algorithmic knowledge predicts people's attitudes and behaviors regarding misinformation through the lens of the third-person effect.Design/methodology/approachFour national surveys in the USA (N = 1,415), the UK (N = 1,435), South Korea (N = 1,798) and Mexico (N = 784) were conducted between April and September 2021. The survey questionnaire measured algorithmic knowledge, perceived influence of misinformation on self and others, intention to take corrective actions, support for government regulation and content moderation. Collected data were analyzed using multigroup SEM.FindingsResults indicate that algorithmic knowledge was associated with presumed influence of misinformation on self and others to different degrees. Presumed media influence on self was a strong predictor of intention to take actions to correct misinformation, while presumed media influence on others was a strong predictor of support for government-led platform regulation and platform-led content moderation. There were nuanced but noteworthy differences in the link between presumed media influence and behavioral responses across the four countries studied.Originality/valueThese findings are relevant for grasping the role of algorithmic knowledge in countering rampant misinformation on social media, as well as for expanding US-centered extant literature by elucidating the distinctive views regarding social media algorithms and misinformation in four countries.","PeriodicalId":54925,"journal":{"name":"Internet Research","volume":" ","pages":""},"PeriodicalIF":5.9000,"publicationDate":"2023-08-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Internet Research","FirstCategoryId":"94","ListUrlMain":"https://doi.org/10.1108/intr-07-2022-0578","RegionNum":3,"RegionCategory":"管理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"BUSINESS","Score":null,"Total":0}
引用次数: 0
Abstract
PurposeWhile there has been a growing call for insights on algorithms given their impact on what people encounter on social media, it remains unknown how enhanced algorithmic knowledge serves as a countermeasure to problematic information flow. To fill this gap, this study aims to investigate how algorithmic knowledge predicts people's attitudes and behaviors regarding misinformation through the lens of the third-person effect.Design/methodology/approachFour national surveys in the USA (N = 1,415), the UK (N = 1,435), South Korea (N = 1,798) and Mexico (N = 784) were conducted between April and September 2021. The survey questionnaire measured algorithmic knowledge, perceived influence of misinformation on self and others, intention to take corrective actions, support for government regulation and content moderation. Collected data were analyzed using multigroup SEM.FindingsResults indicate that algorithmic knowledge was associated with presumed influence of misinformation on self and others to different degrees. Presumed media influence on self was a strong predictor of intention to take actions to correct misinformation, while presumed media influence on others was a strong predictor of support for government-led platform regulation and platform-led content moderation. There were nuanced but noteworthy differences in the link between presumed media influence and behavioral responses across the four countries studied.Originality/valueThese findings are relevant for grasping the role of algorithmic knowledge in countering rampant misinformation on social media, as well as for expanding US-centered extant literature by elucidating the distinctive views regarding social media algorithms and misinformation in four countries.
期刊介绍:
This wide-ranging interdisciplinary journal looks at the social, ethical, economic and political implications of the internet. Recent issues have focused on online and mobile gaming, the sharing economy, and the dark side of social media.