Rosina O Weber, Adam J Johs, Prateek Goel, João Marques Silva
Researchers focusing on how artificial intelligence (AI) methods explain their decisions often discuss controversies and limitations. Some even assert that most publications offer little to no valuable contributions. In this article, we substantiate the claim that explainable AI (XAI) is in trouble by describing and illustrating four problems: the disagreements on the scope of XAI, the lack of definitional cohesion, precision, and adoption, the issues with motivations for XAI research, and limited and inconsistent evaluations. As we delve into their potential underlying sources, our analysis finds these problems seem to originate from AI researchers succumbing to the pitfalls of interdisciplinarity or from insufficient scientific rigor. Analyzing these potential factors, we discuss the literature at times coming across unexplored research questions. Hoping to alleviate existing problems, we make recommendations on precautions against the challenges of interdisciplinarity and propose directions in support of scientific rigor.
{"title":"XAI is in trouble","authors":"Rosina O Weber, Adam J Johs, Prateek Goel, João Marques Silva","doi":"10.1002/aaai.12184","DOIUrl":"https://doi.org/10.1002/aaai.12184","url":null,"abstract":"<p>Researchers focusing on how artificial intelligence (AI) methods explain their decisions often discuss controversies and limitations. Some even assert that most publications offer little to no valuable contributions. In this article, we substantiate the claim that explainable AI (XAI) is in trouble by describing and illustrating four problems: the disagreements on the scope of XAI, the lack of definitional cohesion, precision, and adoption, the issues with motivations for XAI research, and limited and inconsistent evaluations. As we delve into their potential underlying sources, our analysis finds these problems seem to originate from AI researchers succumbing to the pitfalls of interdisciplinarity or from insufficient scientific rigor. Analyzing these potential factors, we discuss the literature at times coming across unexplored research questions. Hoping to alleviate existing problems, we make recommendations on precautions against the challenges of interdisciplinarity and propose directions in support of scientific rigor.</p>","PeriodicalId":7854,"journal":{"name":"Ai Magazine","volume":"45 3","pages":"300-316"},"PeriodicalIF":2.5,"publicationDate":"2024-07-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/aaai.12184","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142324510","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
The European Union Parliament passed the EU AI Act in 2024, which is an important milestone towards the world's first comprehensive AI law to formally take effect. Although this is a significant achievement, the real work begins with putting these rules into action, a journey filled with challenges and opportunities. This perspective article reviews recent interdisciplinary research aimed at facilitating the implementation of the prohibited AI practices outlined in the EU AI Act. It also explores the necessary future efforts to effectively enforce the banning of those prohibited practices across the EU market and the challenges associated with such enforcement. Addressing these future tasks and challenges calls for the establishment of an interdisciplinary governance framework. This framework may contain a workflow that can identify the necessary expertise and coordinate experts’ collaboration at different stages of AI governance. Additionally, it involves developing and implementing a set of compliance and ethical safeguards to ensure effective management and supervision of AI practices.
{"title":"Implementation of the EU AI act calls for interdisciplinary governance","authors":"Huixin Zhong","doi":"10.1002/aaai.12183","DOIUrl":"10.1002/aaai.12183","url":null,"abstract":"<p>The European Union Parliament passed the EU AI Act in 2024, which is an important milestone towards the world's first comprehensive AI law to formally take effect. Although this is a significant achievement, the real work begins with putting these rules into action, a journey filled with challenges and opportunities. This perspective article reviews recent interdisciplinary research aimed at facilitating the implementation of the prohibited AI practices outlined in the EU AI Act. It also explores the necessary future efforts to effectively enforce the banning of those prohibited practices across the EU market and the challenges associated with such enforcement. Addressing these future tasks and challenges calls for the establishment of an interdisciplinary governance framework. This framework may contain a workflow that can identify the necessary expertise and coordinate experts’ collaboration at different stages of AI governance. Additionally, it involves developing and implementing a set of compliance and ethical safeguards to ensure effective management and supervision of AI practices.</p>","PeriodicalId":7854,"journal":{"name":"Ai Magazine","volume":"45 3","pages":"333-337"},"PeriodicalIF":2.5,"publicationDate":"2024-07-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/aaai.12183","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141821161","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
The current literature on AI-advised decision making—involving explainable AI systems advising human decision makers—presents a series of inconclusive and confounding results. To synthesize these findings, we propose a simple theory that elucidates the frequent failure of AI explanations to engender appropriate reliance and complementary decision making performance. In contrast to other common desiderata, for example, interpretability or spelling out the AI's reasoning process, we argue that explanations are only useful to the extent that they allow a human decision maker to verify the correctness of the AI's prediction. Prior studies find in many decision making contexts that AI explanations do not facilitate such verification. Moreover, most tasks fundamentally do not allow easy verification, regardless of explanation method, limiting the potential benefit of any type of explanation. We also compare the objective of complementary performance with that of appropriate reliance, decomposing the latter into the notions of outcome-graded and strategy-graded reliance.
{"title":"In search of verifiability: Explanations rarely enable complementary performance in AI-advised decision making","authors":"Raymond Fok, Daniel S. Weld","doi":"10.1002/aaai.12182","DOIUrl":"https://doi.org/10.1002/aaai.12182","url":null,"abstract":"<p>The current literature on AI-advised decision making—involving explainable AI systems advising human decision makers—presents a series of inconclusive and confounding results. To synthesize these findings, we propose a simple theory that elucidates the frequent failure of AI explanations to engender appropriate reliance and complementary decision making performance. In contrast to other common desiderata, for example, interpretability or spelling out the AI's reasoning process, we argue that explanations are only useful to the extent that they <i>allow a human decision maker to verify the correctness of the AI's prediction</i>. Prior studies find in many decision making contexts that AI explanations <i>do not</i> facilitate such verification. Moreover, most tasks fundamentally do not allow easy verification, regardless of explanation method, limiting the potential benefit of any type of explanation. We also compare the objective of complementary performance with that of appropriate reliance, decomposing the latter into the notions of outcome-graded and strategy-graded reliance.</p>","PeriodicalId":7854,"journal":{"name":"Ai Magazine","volume":"45 3","pages":"317-332"},"PeriodicalIF":2.5,"publicationDate":"2024-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/aaai.12182","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142324408","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
<p>Artificial intelligence (AI) is impacting and has the potential to upend entire business models and structures. The adoption of such new technologies to support newsgathering processes is established practice for newsrooms. For AI specifically, we are seeing a new era of AI-assisted journalism emerge with trust in the AI-driven analyses and accuracy of results as core tenets.</p><p>In Part I of this position paper, we discuss the contributions of six recently published research papers co-authored by Bloomberg's Artificial Intelligence Engineering team that show the intricacies of training AI models for reliable newsgathering processes. The papers investigate (a) the creation of models for updated headline generation, showing that headline generation models benefit from access to the past state of the article, (b) sequentially controlled text generation, which is a novel task and we show that in general, more structured awareness results in higher control accuracy and grammatical coherence, (c) chart summarization, which looks into identifying the key message and generating sentences that describe salient information in the multimodal documents, (d) a semistructured natural language inference task to develop a framework for data augmentation for tabular inference, (e) the introduction of a human-annotated dataset (ENTSUM) for controllable summarization with a focus on named entities as the aspect to control, and (f) a novel defense mechanism against adversarial attacks (ATINTER). We also examine Bloomberg's research work, building its own internal, not-for-commercial-use large language model, BloombergGPT, and training it with the goal of demonstrating support for a wide range of tasks within the financial industry.</p><p>In Part II, we analyze the evolution of automation tasks in the Bloomberg newsroom that led to the creation of Bloomberg's News Innovation Lab. Technology-assisted content creation has been a reality at Bloomberg News for nearly a decade and has evolved from rules-based headline generation from structured files to the constant exploration of potential ways to assist story creation and storytelling in the financial domain. The Lab now oversees the operation of hundreds of software bots that create semi- and fully automated stories of financial relevance, providing journalists with depth in terms of data and analysis, speed in terms of reacting to breaking news, and transparency to corners of the financial world where data investigation is a gigantic undertaking. The Lab recently introduced new tools that provide journalists with the ability to explore automation on demand while it continues to experiment with ways to assist story production.</p><p>In Part III, we conceptually discuss the transformative impact that generative AI can have in any newsroom, along with considerations about the technology's shortcomings in its current state of development. As with any revolutionary new technology, as well as with exciting research op
{"title":"A new era of AI-assisted journalism at Bloomberg","authors":"Claudia Quinonez, Edgar Meij","doi":"10.1002/aaai.12181","DOIUrl":"10.1002/aaai.12181","url":null,"abstract":"<p>Artificial intelligence (AI) is impacting and has the potential to upend entire business models and structures. The adoption of such new technologies to support newsgathering processes is established practice for newsrooms. For AI specifically, we are seeing a new era of AI-assisted journalism emerge with trust in the AI-driven analyses and accuracy of results as core tenets.</p><p>In Part I of this position paper, we discuss the contributions of six recently published research papers co-authored by Bloomberg's Artificial Intelligence Engineering team that show the intricacies of training AI models for reliable newsgathering processes. The papers investigate (a) the creation of models for updated headline generation, showing that headline generation models benefit from access to the past state of the article, (b) sequentially controlled text generation, which is a novel task and we show that in general, more structured awareness results in higher control accuracy and grammatical coherence, (c) chart summarization, which looks into identifying the key message and generating sentences that describe salient information in the multimodal documents, (d) a semistructured natural language inference task to develop a framework for data augmentation for tabular inference, (e) the introduction of a human-annotated dataset (ENTSUM) for controllable summarization with a focus on named entities as the aspect to control, and (f) a novel defense mechanism against adversarial attacks (ATINTER). We also examine Bloomberg's research work, building its own internal, not-for-commercial-use large language model, BloombergGPT, and training it with the goal of demonstrating support for a wide range of tasks within the financial industry.</p><p>In Part II, we analyze the evolution of automation tasks in the Bloomberg newsroom that led to the creation of Bloomberg's News Innovation Lab. Technology-assisted content creation has been a reality at Bloomberg News for nearly a decade and has evolved from rules-based headline generation from structured files to the constant exploration of potential ways to assist story creation and storytelling in the financial domain. The Lab now oversees the operation of hundreds of software bots that create semi- and fully automated stories of financial relevance, providing journalists with depth in terms of data and analysis, speed in terms of reacting to breaking news, and transparency to corners of the financial world where data investigation is a gigantic undertaking. The Lab recently introduced new tools that provide journalists with the ability to explore automation on demand while it continues to experiment with ways to assist story production.</p><p>In Part III, we conceptually discuss the transformative impact that generative AI can have in any newsroom, along with considerations about the technology's shortcomings in its current state of development. As with any revolutionary new technology, as well as with exciting research op","PeriodicalId":7854,"journal":{"name":"Ai Magazine","volume":"45 2","pages":"187-199"},"PeriodicalIF":0.9,"publicationDate":"2024-06-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/aaai.12181","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141382493","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Correcting misinformation is a complex task, influenced by various psychological, social, and technical factors. Most research evaluation methods for identifying effective correction approaches tend to rely on either crowdsourcing, questionnaires, lab-based simulations, or hypothetical scenarios. However, the translation of these methods and findings into real-world settings, where individuals willingly and freely disseminate misinformation, remains largely unexplored. Consequently, we lack a comprehensive understanding of how individuals who share misinformation in natural online environments would respond to corrective interventions. In this study, we explore the effectiveness of corrective messaging on 3898 users who shared misinformation on Twitter/X over 2 years. We designed and deployed a bot to automatically identify individuals who share misinformation and subsequently alert them to related fact-checks in various message formats. Our analysis shows that only a small minority of users react positively to the corrective messages, with most users either ignoring them or reacting negatively. Nevertheless, we also found that more active users were proportionally more likely to react positively to corrections and we observed that different message tones made particular user groups more likely to react to the bot.
{"title":"Exploring the impact of automated correction of misinformation in social media","authors":"Grégoire Burel, Mohammadali Tavakoli, Harith Alani","doi":"10.1002/aaai.12180","DOIUrl":"10.1002/aaai.12180","url":null,"abstract":"<p>Correcting misinformation is a complex task, influenced by various psychological, social, and technical factors. Most research evaluation methods for identifying effective correction approaches tend to rely on either crowdsourcing, questionnaires, lab-based simulations, or hypothetical scenarios. However, the translation of these methods and findings into real-world settings, where individuals willingly and freely disseminate misinformation, remains largely unexplored. Consequently, we lack a comprehensive understanding of how individuals who share misinformation in natural online environments would respond to corrective interventions. In this study, we explore the effectiveness of corrective messaging on 3898 users who shared misinformation on Twitter/X over 2 years. We designed and deployed a bot to automatically identify individuals who share misinformation and subsequently alert them to related fact-checks in various message formats. Our analysis shows that only a small minority of users react positively to the corrective messages, with most users either ignoring them or reacting negatively. Nevertheless, we also found that more active users were proportionally more likely to react positively to corrections and we observed that different message tones made particular user groups more likely to react to the bot.</p>","PeriodicalId":7854,"journal":{"name":"Ai Magazine","volume":"45 2","pages":"227-245"},"PeriodicalIF":0.9,"publicationDate":"2024-06-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/aaai.12180","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141387863","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Andreas L. Opdahl, Natali Helberger, Nicholas Diakopoulos
<p>In a time of rising populism and strategic disinformation, quality journalism has become more important than ever. Trusted and high-quality media outlets are needed to provide accurate information to the public in order to protect public safety and wellbeing while supporting the information needs of citizens in order to promote healthy liberal democracies. But quality journalism is also under pressure due to competition for attention from new information channels, declining trust in institutions, and dwindling resources to support the information needs of local communities while there are simultaneously new resource demands to mitigate the impacts of mis- and disinformation. Given this challenging context, how can Artificial Intelligence (AI) support the provision of quality information for society?</p><p>This special issue therefore examines how ongoing advances in AI, including Machine Learning (ML), and generative AI such as Large Language Models (LLMs), can be harnessed to support efficient production and distribution of high-quality news. It takes a broad outlook on the area, including articles that deal with uses and implications of AI in all stages of news production and dissemination, from gathering and analyzing information to creating, presenting, or recommending news content, while also dealing with an onslaught of mis- and disinformation in the broader online information ecosystem. It also discusses AI on different levels, from individual news production tasks, through organizational transformations and ramifications, to societal and economic conditions and consequences. A common red thread throughout the articles is that AI has great transformational potential, also in the media sector, but the factors driving and enabling such transformations are not only technological. Such factors also very much pertain to the broader organizational, infrastructure and economic context, and successful alignment of the different actors along the value chain, including media users.</p><p>The articles presented here offer an optimistic picture of how quality information and the media ecosystem might evolve in positive ways in light of the technological change driven by AI. And while critical approaches and research are by all means warranted such that professional ethical commitments are maintained, we hope this collection at least provides some ideas and inspiration for technologists and other stakeholders to engage further with how to orient their work towards addressing problems, seeking fruitful cooperations with the different stakeholders along the value chain and providing benefits to support quality media production.</p><p>Next, we outline the six articles in the collection providing a brief summary of each to orient the reader.</p><p>LLMs and other generative AI technologies are ushering in a new phase of disruption in the news industry that may affect news production and consumption as well as distribution. David Caswell, in his paper <i>
在民粹主义和战略性虚假信息抬头的时代,高质量的新闻报道比以往任何时候都更加重要。我们需要可信赖的高质量媒体为公众提供准确的信息,以保护公众的安全和福祉,同时支持公民的信息需求,以促进健康的自由民主。但是,由于新的信息渠道争夺注意力、对机构的信任度下降、支持当地社区信息需求的资源不断减少,同时为减轻错误信息和虚假信息的影响又提出了新的资源需求,高质量的新闻报道也面临着压力。因此,本特刊将探讨如何利用机器学习(ML)和大型语言模型(LLMs)等生成式人工智能等人工智能领域的最新进展,支持高质量新闻的高效制作和传播。它以广阔的视角审视这一领域,收录的文章涉及人工智能在新闻制作和传播各个阶段的应用和影响,从收集和分析信息到创建、呈现或推荐新闻内容,同时还应对更广泛的在线信息生态系统中大量的错误信息和虚假信息。文章还从不同层面讨论了人工智能,从个人新闻制作任务,到组织变革和影响,再到社会和经济条件及后果。整篇文章的一条共同主线是,人工智能具有巨大的变革潜力,在媒体领域也是如此,但推动和促成这种变革的因素不仅仅是技术。这里介绍的文章乐观地描绘了在人工智能技术变革的推动下,优质信息和媒体生态系统如何以积极的方式发展。我们希望本文集至少能为技术专家和其他利益相关者提供一些想法和启发,使他们能进一步思考如何将自己的工作导向解决问题、寻求与价值链上不同利益相关者的富有成效的合作,并为支持优质媒体生产提供益处。接下来,我们将概述文集中的六篇文章,并对每篇文章进行简要概述,以便为读者指明方向。LLM 和其他生成式人工智能技术正在为新闻行业带来一个新的颠覆阶段,可能会影响新闻生产、消费和传播。大卫-卡斯维尔(David Caswell)在他的论文《受众、自动化与人工智能:从结构化新闻到语言模型》中认为,像英国广播公司(BBC)这样的大型新闻机构已经为这种转变做好了准备,因为之前已经利用结构化技术实现了个性化内容自动化工作流程的创新。在财经新闻领域,人工智能正在重塑新闻业,并促进人工智能辅助新闻流程进入一个新时代,而这必须以信任和准确性为基础。Claudia Quinonez 和 Edgar Meij 的论文《A New Era of AI-Assisted Journalism at Bloomberg》(彭博社的人工智能辅助新闻新时代)举例说明了彭博社如何在更新标题生成和可控文本摘要等任务中探索人工智能模型。作者还讨论了彭博社新闻编辑室的自动化问题,在那里,软件机器人自动进行新闻创作,以实现更快速、更深入的财经报道。论文还探讨了生成式人工智能对新闻业的广泛影响,强调严格的准确性标准对财经受众至关重要。面对广告和受众收入的下降,许多地方新闻机构正在探索如何利用人工智能来应对经济压力,提高价值创造能力。Bartosz Wilczek、Mario Haim 和 Neil Thurman 在他们的论文《用人工智能改造地方新闻价值链》中概述了人工智能在地方新闻中的潜力,指出了人工智能最能带来益处的领域。他们还讨论了地方新闻编辑室在整个价值创造链中面临的具体实施挑战,包括资源限制,并提出了克服这些挑战的策略。 Nava Tintarev、Martijn Willemsen 和 Bart P Knijnenburg 的论文《衡量在新闻推荐中增加透明度和控制的益处》解释了向用户提供解释如何帮助他们理解某些新闻条目被推荐的原因,并使他们的阅读习惯符合个人目标,如知识扩展和观点多样性。作者认为,需要在实时推荐环境中进行更真实的评估,以评估解释性干预对用户行为的实际影响。由于心理、社会和技术因素,纠正错误信息涉及复杂的挑战。Gregoire Burel、Mohammadali Tavakoli 和 Harith Alani 在论文《探索自动纠正社交媒体中错误信息的影响》中指出,人工智能驱动的纠正方法在现实环境中的有效性还没有得到充分研究。他们研究了分享错误信息的用户对不同类型的机器人生成的社交媒体纠正信息的反应,为如何制定纠正信息以及针对哪类用户提供了新的理解。在社会层面,人工智能正在改变经济结构和新闻机构的融资方式。在本特刊的最后一篇论文《人工智能经济中的新闻业务》中,Helle Sjøvaag 探讨了人工智能对新闻行业竞争、并购和 IT 能力的影响,并讨论了人工智能如何影响新闻业的传统业务模式。其目的是为理解新闻业在数据驱动和人工智能推动的平台经济中的经济前景提供一个词汇表。
{"title":"Guest Editorial: AI and the news","authors":"Andreas L. Opdahl, Natali Helberger, Nicholas Diakopoulos","doi":"10.1002/aaai.12179","DOIUrl":"https://doi.org/10.1002/aaai.12179","url":null,"abstract":"<p>In a time of rising populism and strategic disinformation, quality journalism has become more important than ever. Trusted and high-quality media outlets are needed to provide accurate information to the public in order to protect public safety and wellbeing while supporting the information needs of citizens in order to promote healthy liberal democracies. But quality journalism is also under pressure due to competition for attention from new information channels, declining trust in institutions, and dwindling resources to support the information needs of local communities while there are simultaneously new resource demands to mitigate the impacts of mis- and disinformation. Given this challenging context, how can Artificial Intelligence (AI) support the provision of quality information for society?</p><p>This special issue therefore examines how ongoing advances in AI, including Machine Learning (ML), and generative AI such as Large Language Models (LLMs), can be harnessed to support efficient production and distribution of high-quality news. It takes a broad outlook on the area, including articles that deal with uses and implications of AI in all stages of news production and dissemination, from gathering and analyzing information to creating, presenting, or recommending news content, while also dealing with an onslaught of mis- and disinformation in the broader online information ecosystem. It also discusses AI on different levels, from individual news production tasks, through organizational transformations and ramifications, to societal and economic conditions and consequences. A common red thread throughout the articles is that AI has great transformational potential, also in the media sector, but the factors driving and enabling such transformations are not only technological. Such factors also very much pertain to the broader organizational, infrastructure and economic context, and successful alignment of the different actors along the value chain, including media users.</p><p>The articles presented here offer an optimistic picture of how quality information and the media ecosystem might evolve in positive ways in light of the technological change driven by AI. And while critical approaches and research are by all means warranted such that professional ethical commitments are maintained, we hope this collection at least provides some ideas and inspiration for technologists and other stakeholders to engage further with how to orient their work towards addressing problems, seeking fruitful cooperations with the different stakeholders along the value chain and providing benefits to support quality media production.</p><p>Next, we outline the six articles in the collection providing a brief summary of each to orient the reader.</p><p>LLMs and other generative AI technologies are ushering in a new phase of disruption in the news industry that may affect news production and consumption as well as distribution. David Caswell, in his paper <i>","PeriodicalId":7854,"journal":{"name":"Ai Magazine","volume":"45 2","pages":"172-173"},"PeriodicalIF":0.9,"publicationDate":"2024-05-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/aaai.12179","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141430188","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
With their advertising and audience revenues in decline, local news organizations have been experiencing comparatively high degrees of disruption in recent years. Artificial Intelligence (AI) offers opportunities for local news organizations to better cope with the economic challenges they face. However, local news organizations need to carefully prioritize where AI will create the most value. After all, they serve customers in the audience and advertising markets, with external effects on society. At the same time, they are limited by scarce resources, which constrains the implementation of AI. Therefore, based on Porter's value chain, this article pursues two goals. First, drawing on previous research, we provide a systematic overview of activities for which local news organizations see the biggest potential of AI to create value. Moreover, we highlight promising AI use cases based on benchmarking with national news organizations. Second, we discuss local news organizations’ challenges in implementing AI and how they might overcome such obstacles.
{"title":"Transforming the value chain of local journalism with artificial intelligence","authors":"Bartosz Wilczek, Mario Haim, Neil Thurman","doi":"10.1002/aaai.12174","DOIUrl":"https://doi.org/10.1002/aaai.12174","url":null,"abstract":"<p>With their advertising and audience revenues in decline, local news organizations have been experiencing comparatively high degrees of disruption in recent years. Artificial Intelligence (AI) offers opportunities for local news organizations to better cope with the economic challenges they face. However, local news organizations need to carefully prioritize where AI will create the most value. After all, they serve customers in the audience and advertising markets, with external effects on society. At the same time, they are limited by scarce resources, which constrains the implementation of AI. Therefore, based on Porter's value chain, this article pursues two goals. First, drawing on previous research, we provide a systematic overview of activities for which local news organizations see the biggest potential of AI to create value. Moreover, we highlight promising AI use cases based on benchmarking with national news organizations. Second, we discuss local news organizations’ challenges in implementing AI and how they might overcome such obstacles.</p>","PeriodicalId":7854,"journal":{"name":"Ai Magazine","volume":"45 2","pages":"200-211"},"PeriodicalIF":0.9,"publicationDate":"2024-05-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/aaai.12174","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141430214","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
The advance of machine learning (ML) systems in real-world scenarios usually expects safe deployment in high-stake applications (e.g., medical diagnosis) for critical decision-making process. To this end, provable robustness of ML is usually required to measure and understand how reliable the deployed ML system is and how trustworthy their predictions can be. Many studies have been done to enhance the robustness in recent years from different angles, such as variance-regularized robust objective functions and conformal prediction (CP) for uncertainty quantification on testing data. Although these tools provably improve the robustness of ML model, there is still an inevitable gap to integrate them into an end-to-end deployment. For example, robust objectives usually require carefully designed optimization algorithms, while CP treats ML models as black boxes. This paper is a brief introduction to our recent research focusing on filling this gap. Specifically, for learning robust objectives, we designed sample-efficient stochastic optimization algorithms that achieves the optimal (or faster compared to existing algorithms) convergence rates. Moreover, for CP-based uncertainty quantification, we established a framework to analyze the expected prediction set size (smaller size means more efficiency) of CP methods in both standard and adversarial settings. This paper elaborates the key challenges and our exploration towards efficient algorithms with details of background methods, notions for robustness measure, concepts of algorithmic efficiency, our proposed algorithms and results. All of them further motivate our future research on risk-aware ML that can be critical for AI–human collaborative systems. The future work mainly targets designing conformal robust objectives and their efficient optimization algorithms.
机器学习(ML)系统在现实世界场景中的发展,通常期望在关键决策过程中的高风险应用(如医疗诊断)中安全部署。为此,通常需要证明 ML 的鲁棒性,以衡量和了解部署的 ML 系统的可靠性及其预测的可信度。近年来,人们从不同角度对增强鲁棒性进行了许多研究,例如方差规则化鲁棒目标函数和用于测试数据不确定性量化的保形预测(CP)。虽然这些工具都能有效提高 ML 模型的鲁棒性,但要将它们集成到端到端的部署中,仍存在不可避免的差距。例如,稳健目标通常需要精心设计的优化算法,而 CP 则将 ML 模型视为黑盒。本文简要介绍了我们最近为填补这一空白而开展的研究。具体来说,针对鲁棒目标的学习,我们设计了样本效率高的随机优化算法,以达到最佳收敛率(或与现有算法相比更快的收敛率)。此外,对于基于 CP 的不确定性量化,我们建立了一个框架,用于分析标准和对抗环境下 CP 方法的预期预测集规模(规模越小效率越高)。本文通过详细介绍背景方法、鲁棒性度量概念、算法效率概念、我们提出的算法和结果,阐述了关键挑战和我们对高效算法的探索。所有这些都进一步激发了我们对风险感知人工智能的未来研究,这对人工智能与人类协作系统至关重要。未来工作的主要目标是设计保形鲁棒目标及其高效优化算法。
{"title":"Improve robustness of machine learning via efficient optimization and conformal prediction","authors":"Yan Yan","doi":"10.1002/aaai.12173","DOIUrl":"10.1002/aaai.12173","url":null,"abstract":"<p>The advance of machine learning (ML) systems in real-world scenarios usually expects safe deployment in high-stake applications (e.g., medical diagnosis) for critical decision-making process. To this end, provable robustness of ML is usually required to measure and understand how reliable the deployed ML system is and how trustworthy their predictions can be. Many studies have been done to enhance the robustness in recent years from different angles, such as variance-regularized robust objective functions and conformal prediction (CP) for uncertainty quantification on testing data. Although these tools provably improve the robustness of ML model, there is still an inevitable gap to integrate them into an <i>end-to-end</i> deployment. For example, robust objectives usually require carefully designed optimization algorithms, while CP treats ML models as black boxes. This paper is a brief introduction to our recent research focusing on filling this gap. Specifically, for learning robust objectives, we designed sample-efficient stochastic optimization algorithms that achieves the optimal (or faster compared to existing algorithms) convergence rates. Moreover, for CP-based uncertainty quantification, we established a framework to analyze the expected prediction set size (smaller size means more efficiency) of CP methods in both standard and adversarial settings. This paper elaborates the key challenges and our exploration towards efficient algorithms with details of background methods, notions for robustness measure, concepts of algorithmic efficiency, our proposed algorithms and results. All of them further motivate our future research on risk-aware ML that can be critical for AI–human collaborative systems. The future work mainly targets designing conformal robust objectives and their efficient optimization algorithms.</p>","PeriodicalId":7854,"journal":{"name":"Ai Magazine","volume":"45 2","pages":"270-279"},"PeriodicalIF":0.9,"publicationDate":"2024-05-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/aaai.12173","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141128836","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Nava Tintarev, Bart P. Knijnenburg, Martijn C. Willemsen
Personalized news experiences powered by recommender systems permeate our lives and have the potential to influence not only our opinions, but also our decisions. At the same time, the content and viewpoints contained within news recommendations are driven by multiple factors, including both personalization and editorial selection. Explanations could help users gain a better understanding of the factors contributing to the news items selected for them to read. Indeed, recent works show that explanations are essential for users of news recommenders to understand their consumption preferences and set intentions in line with their goals, such as goals for knowledge development and increased diversity of content or viewpoints. We give examples of such works on explanation and interactive interface interventions which have been effective in influencing readers' consumption intentions and behaviors in news recommendations. However, the state-of-the-art in news recommender systems currently fall short in terms of evaluating such interventions in live systems, limiting our ability to measure their true impact on user behavior and opinions. To help understand the true benefit of these interfaces, we therefore call for improving the realism of studies for news.
{"title":"Measuring the benefit of increased transparency and control in news recommendation","authors":"Nava Tintarev, Bart P. Knijnenburg, Martijn C. Willemsen","doi":"10.1002/aaai.12171","DOIUrl":"10.1002/aaai.12171","url":null,"abstract":"<p>Personalized news experiences powered by recommender systems permeate our lives and have the potential to influence not only our opinions, but also our decisions. At the same time, the content and viewpoints contained within news recommendations are driven by multiple factors, including both personalization and editorial selection. Explanations could help users gain a better understanding of the factors contributing to the news items selected for them to read. Indeed, recent works show that explanations are essential for users of news recommenders to understand their consumption preferences and set intentions in line with their goals, such as goals for knowledge development and increased diversity of content or viewpoints. We give examples of such works on explanation and interactive interface interventions which have been effective in influencing readers' consumption intentions and behaviors in news recommendations. However, the state-of-the-art in news recommender systems currently fall short in terms of evaluating such interventions in live systems, limiting our ability to measure their true impact on user behavior and opinions. To help understand the true benefit of these interfaces, we therefore call for improving the realism of studies for news.</p>","PeriodicalId":7854,"journal":{"name":"Ai Magazine","volume":"45 2","pages":"212-226"},"PeriodicalIF":0.9,"publicationDate":"2024-04-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/aaai.12171","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140693533","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
This article considers the impact of AI on the economy and financing of journalism organizations. AI has structural implications on the news media beyond the practice of journalism and the management of news as a process. AI also shifts the premises of competition, competitive advantage, mergers and acquisitions, and IT capabilities in the news industries. Not least, it fundamentally challenges journalism's traditional business model. Considered hereunder is the two-sided market model, journalism's traditional platform function, its network effects, and its public good characteristics. The aim of the article is, thus, to reconceptualize core economic features of the news industries in the context of AI to provide a vocabulary with which to assess the economic future of journalism in a data-driven platform economy.
本文探讨了人工智能对新闻机构的经济和融资的影响。人工智能对新闻媒体的结构性影响超出了新闻实践和新闻管理这一过程。人工智能还改变了新闻行业的竞争前提、竞争优势、并购和 IT 能力。更重要的是,它从根本上挑战了新闻业的传统商业模式。本文主要探讨了双面市场模式、新闻业传统的平台功能、网络效应及其公益性特征。因此,本文旨在重新认识人工智能背景下新闻业的核心经济特征,为评估数据驱动的平台经济中新闻业的经济前景提供一个词汇表。
{"title":"The business of news in the AI economy","authors":"Helle Sjøvaag","doi":"10.1002/aaai.12172","DOIUrl":"10.1002/aaai.12172","url":null,"abstract":"<p>This article considers the impact of AI on the economy and financing of journalism organizations. AI has structural implications on the news media beyond the practice of journalism and the management of news as a process. AI also shifts the premises of competition, competitive advantage, mergers and acquisitions, and IT capabilities in the news industries. Not least, it fundamentally challenges journalism's traditional business model. Considered hereunder is the two-sided market model, journalism's traditional platform function, its network effects, and its public good characteristics. The aim of the article is, thus, to reconceptualize core economic features of the news industries in the context of AI to provide a vocabulary with which to assess the economic future of journalism in a data-driven platform economy.</p>","PeriodicalId":7854,"journal":{"name":"Ai Magazine","volume":"45 2","pages":"246-255"},"PeriodicalIF":0.9,"publicationDate":"2024-04-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/aaai.12172","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140725116","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}