首页 > 最新文献

Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems最新文献

英文 中文
Train as you Fight: Evaluating Authentic Cybersecurity Training in Cyber Ranges 在战斗中训练:评估网络范围内的真实网络安全培训
Pub Date : 2023-04-19 DOI: 10.1145/3544548.3581046
M. Glas, Manfred Vielberth, Guenther Pernul
Humans can play a decisive role in detecting and mitigating cyber attacks if they possess sufficient cybersecurity skills and knowledge. Realizing this potential requires effective cybersecurity training. Cyber range exercises (CRXs) represent a novel form of cybersecurity training in which trainees can experience realistic cyber attacks in authentic environments. Although evaluation is undeniably essential for any learning environment, it has been widely neglected in CRX research. Addressing this issue, we propose a taxonomy-based framework to facilitate a comprehensive and structured evaluation of CRXs. To demonstrate the applicability and potential of the framework, we instantiate it to evaluate Iceberg CRX, a training we recently developed to improve cybersecurity education at our university. For this matter, we conducted a user study with 50 students to identify both strengths and weaknesses of the CRX.
如果拥有足够的网络安全技能和知识,人类可以在发现和减轻网络攻击方面发挥决定性作用。实现这一潜力需要有效的网络安全培训。网络靶场演习(CRXs)是一种新颖的网络安全培训形式,学员可以在真实的环境中体验真实的网络攻击。虽然评价在任何学习环境中都是必不可少的,但在CRX研究中却被广泛忽视。针对这一问题,我们提出了一个基于分类法的框架,以促进对crx的全面和结构化评估。为了证明该框架的适用性和潜力,我们实例化了它来评估冰山CRX,这是我们最近开发的一项培训,旨在改善我们大学的网络安全教育。为此,我们对50名学生进行了用户研究,以确定CRX的优点和缺点。
{"title":"Train as you Fight: Evaluating Authentic Cybersecurity Training in Cyber Ranges","authors":"M. Glas, Manfred Vielberth, Guenther Pernul","doi":"10.1145/3544548.3581046","DOIUrl":"https://doi.org/10.1145/3544548.3581046","url":null,"abstract":"Humans can play a decisive role in detecting and mitigating cyber attacks if they possess sufficient cybersecurity skills and knowledge. Realizing this potential requires effective cybersecurity training. Cyber range exercises (CRXs) represent a novel form of cybersecurity training in which trainees can experience realistic cyber attacks in authentic environments. Although evaluation is undeniably essential for any learning environment, it has been widely neglected in CRX research. Addressing this issue, we propose a taxonomy-based framework to facilitate a comprehensive and structured evaluation of CRXs. To demonstrate the applicability and potential of the framework, we instantiate it to evaluate Iceberg CRX, a training we recently developed to improve cybersecurity education at our university. For this matter, we conducted a user study with 50 students to identify both strengths and weaknesses of the CRX.","PeriodicalId":314098,"journal":{"name":"Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems","volume":"38 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127204117","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
OK Google, Let's Learn: Using Voice User Interfaces for Informal Self-Regulated Learning of Health Topics among Younger and Older Adults 好了,谷歌,让我们学习:使用语音用户界面在年轻人和老年人中进行非正式的自我调节的健康主题学习
Pub Date : 2023-04-19 DOI: 10.1145/3544548.3581507
Smit Desai, Jessie Chin
In this paper, we present Health Buddy, a voice agent integrated into commercially available Voice User Interfaces (VUIs) to support informal self-regulated learning (SRL) of health-related topics through multiple learning strategies and examine the efficacy of Health Buddy on learning outcomes for younger and older adults. We conducted a mixed-factorial-design experiment with 26 younger and 25 older adults, assigned to three SRL strategies (within-subjects): monologue, dialogue-based scaffolding building, and conceptual diagramming. We found that while younger adults benefit more from scaffolding building and conceptual diagramming, both younger and older adults showed equivalent learning outcomes. Furthermore, interaction fluency (operationalized by the number of conversational breakdowns) was associated with learning outcomes regardless of age. While older adults did not experience less fluent conversations, interaction fluency affected their technology acceptance toward VUIs more than younger ones. Our study discusses age-related learning differences and has implications for designing VUI-based learning programs for older adults.
在本文中,我们介绍了一个集成在商业语音用户界面(VUIs)中的语音代理Health Buddy,该语音代理通过多种学习策略支持健康相关主题的非正式自我调节学习(SRL),并研究了Health Buddy对年轻人和老年人学习结果的影响。我们对26名年轻人和25名老年人进行了混合因子设计实验,分配了三种SRL策略(在受试者中):独白、基于对话的脚手架构建和概念图解。我们发现,虽然年轻人从搭建脚手架和概念图中受益更多,但年轻人和老年人的学习成果都是一样的。此外,无论年龄大小,互动流畅性(通过会话中断的次数进行操作)都与学习结果相关。虽然老年人的会话流畅度并没有降低,但与年轻人相比,互动流畅度对他们接受ui技术的影响更大。我们的研究讨论了与年龄相关的学习差异,并对设计基于ui的老年人学习程序具有启示意义。
{"title":"OK Google, Let's Learn: Using Voice User Interfaces for Informal Self-Regulated Learning of Health Topics among Younger and Older Adults","authors":"Smit Desai, Jessie Chin","doi":"10.1145/3544548.3581507","DOIUrl":"https://doi.org/10.1145/3544548.3581507","url":null,"abstract":"In this paper, we present Health Buddy, a voice agent integrated into commercially available Voice User Interfaces (VUIs) to support informal self-regulated learning (SRL) of health-related topics through multiple learning strategies and examine the efficacy of Health Buddy on learning outcomes for younger and older adults. We conducted a mixed-factorial-design experiment with 26 younger and 25 older adults, assigned to three SRL strategies (within-subjects): monologue, dialogue-based scaffolding building, and conceptual diagramming. We found that while younger adults benefit more from scaffolding building and conceptual diagramming, both younger and older adults showed equivalent learning outcomes. Furthermore, interaction fluency (operationalized by the number of conversational breakdowns) was associated with learning outcomes regardless of age. While older adults did not experience less fluent conversations, interaction fluency affected their technology acceptance toward VUIs more than younger ones. Our study discusses age-related learning differences and has implications for designing VUI-based learning programs for older adults.","PeriodicalId":314098,"journal":{"name":"Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127377383","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Towards Intermediated Workflows for Hybrid Telemedicine 面向混合远程医疗的中间工作流
Pub Date : 2023-04-19 DOI: 10.1145/3544548.3580653
K. Bhat, Neha Kumar, Karthik Shamanna, Nipun Kwatra, Mohit Jain
The growing platformization of health has spurred new avenues for healthcare access and reinvigorated telemedicine as a viable pathway to care. Telemedicine adoption during the COVID-19 pandemic has surfaced barriers to patient-centered care that call for attention. Our work extends current Human-Computer Interaction (HCI) research on telemedicine and the challenges to remote care, and investigates the scope for enhancing remote care seeking and provision through telemedicine workflows involving intermediation. Our study, focused on the urban Indian context, involved providing doctors with videos of remote clinical examinations to aid in telemedicine. We present a qualitative evaluation of this modified telemedicine experience, highlighting how workflows involving intermediation could bridge existing gaps in telemedicine, and how their acceptance among doctors could shift interaction dynamics between doctors and patients. We conclude by discussing the implications of such telemedicine workflows on patient-centered care and the future of care work.
日益增长的健康平台化刺激了获得医疗保健的新途径,并使远程医疗重新焕发活力,成为一种可行的护理途径。在COVID-19大流行期间,远程医疗的采用已经暴露出以患者为中心的护理的障碍,需要引起注意。我们的工作扩展了当前远程医疗的人机交互(HCI)研究和远程医疗的挑战,并研究了通过涉及中介的远程医疗工作流程加强远程医疗寻求和提供的范围。我们的研究主要集中在印度的城市背景下,包括向医生提供远程临床检查的视频,以帮助远程医疗。我们对这种改进后的远程医疗体验进行了定性评估,强调了涉及中介的工作流程如何能够弥合远程医疗中现有的差距,以及医生对它们的接受如何能够改变医生和患者之间的互动动态。最后,我们讨论了这种远程医疗工作流程对以患者为中心的护理和护理工作的未来的影响。
{"title":"Towards Intermediated Workflows for Hybrid Telemedicine","authors":"K. Bhat, Neha Kumar, Karthik Shamanna, Nipun Kwatra, Mohit Jain","doi":"10.1145/3544548.3580653","DOIUrl":"https://doi.org/10.1145/3544548.3580653","url":null,"abstract":"The growing platformization of health has spurred new avenues for healthcare access and reinvigorated telemedicine as a viable pathway to care. Telemedicine adoption during the COVID-19 pandemic has surfaced barriers to patient-centered care that call for attention. Our work extends current Human-Computer Interaction (HCI) research on telemedicine and the challenges to remote care, and investigates the scope for enhancing remote care seeking and provision through telemedicine workflows involving intermediation. Our study, focused on the urban Indian context, involved providing doctors with videos of remote clinical examinations to aid in telemedicine. We present a qualitative evaluation of this modified telemedicine experience, highlighting how workflows involving intermediation could bridge existing gaps in telemedicine, and how their acceptance among doctors could shift interaction dynamics between doctors and patients. We conclude by discussing the implications of such telemedicine workflows on patient-centered care and the future of care work.","PeriodicalId":314098,"journal":{"name":"Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems","volume":"18 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127459909","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Interactive AR Applications for Nonspeaking Autistic People? - A Usability Study 为不会说话的自闭症患者提供交互式AR应用程序?-可用性研究
Pub Date : 2023-04-19 DOI: 10.1145/3544548.3580721
A. Nazari, Ali Shahidi, Kate M. Kaufman, Julia E Bondi, Lorans Alabood, Vikram K. Jaswal, Diwakar Krishnamurthy, Mea Wang
About one-third of autistic people are nonspeaking, and most are never provided access to an effective alternative to speech. Thoughtfully designed AR applications could provide members of this population with structured learning opportunities, including training on skills that underlie alternative forms of communication. A fundamental step toward creating such opportunities, however, is to investigate nonspeaking autistic people’s ability to tolerate a head-mounted AR device and to interact with virtual objects. We present the first study to examine the usability of an interactive AR-based application by this population. We recruited 17 nonspeaking autistic subjects to play a HoloLens 2 game we developed that involved holographic animations and buttons. Almost all subjects tolerated the device long enough to begin the game, and most completed increasingly challenging tasks that involved pressing holographic buttons. Based on the results, we discuss best practice design and process recommendations. Our findings contradict prevailing assumptions about nonspeaking autistic people and thus open up exciting possibilities for AR-based solutions for this understudied and underserved population.
大约三分之一的自闭症患者不会说话,而且大多数人从来没有获得过有效的替代语言的机会。精心设计的AR应用程序可以为这一群体的成员提供结构化的学习机会,包括对潜在的其他交流形式的技能培训。然而,创造这种机会的一个基本步骤是调查不会说话的自闭症患者对头戴式增强现实设备的耐受能力,以及与虚拟物体互动的能力。我们提出了第一项研究,以检查这一人群的交互式ar应用程序的可用性。我们招募了17名不会说话的自闭症受试者来玩我们开发的HoloLens 2游戏,其中包含全息动画和按键。几乎所有的实验对象都能忍受这个设备足够长的时间来开始游戏,而且大多数人都完成了越来越具有挑战性的任务,其中包括按全息按钮。基于结果,我们讨论了最佳实践设计和流程建议。我们的发现反驳了关于不会说话的自闭症患者的普遍假设,从而为这一研究不足、服务不足的人群提供了基于ar的解决方案,这是令人兴奋的可能性。
{"title":"Interactive AR Applications for Nonspeaking Autistic People? - A Usability Study","authors":"A. Nazari, Ali Shahidi, Kate M. Kaufman, Julia E Bondi, Lorans Alabood, Vikram K. Jaswal, Diwakar Krishnamurthy, Mea Wang","doi":"10.1145/3544548.3580721","DOIUrl":"https://doi.org/10.1145/3544548.3580721","url":null,"abstract":"About one-third of autistic people are nonspeaking, and most are never provided access to an effective alternative to speech. Thoughtfully designed AR applications could provide members of this population with structured learning opportunities, including training on skills that underlie alternative forms of communication. A fundamental step toward creating such opportunities, however, is to investigate nonspeaking autistic people’s ability to tolerate a head-mounted AR device and to interact with virtual objects. We present the first study to examine the usability of an interactive AR-based application by this population. We recruited 17 nonspeaking autistic subjects to play a HoloLens 2 game we developed that involved holographic animations and buttons. Almost all subjects tolerated the device long enough to begin the game, and most completed increasingly challenging tasks that involved pressing holographic buttons. Based on the results, we discuss best practice design and process recommendations. Our findings contradict prevailing assumptions about nonspeaking autistic people and thus open up exciting possibilities for AR-based solutions for this understudied and underserved population.","PeriodicalId":314098,"journal":{"name":"Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems","volume":"43 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123271674","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Improving Automatic Summarization for Browsing Longform Spoken Dialog 改进浏览长篇口语对话的自动摘要
Pub Date : 2023-04-19 DOI: 10.1145/3544548.3581339
Daniel Li, Thomas Chen, Alec Zadikian, Albert Tung, Lydia B. Chilton
Longform spoken dialog delivers rich streams of informative content through podcasts, interviews, debates, and meetings. While production of this medium has grown tremendously, spoken dialog remains challenging to consume as listening is slower than reading and difficult to skim or navigate relative to text. Recent systems leveraging automatic speech recognition (ASR) and automatic summarization allow users to better browse speech data and forage for information of interest. However, these systems intake disfluent speech which causes automatic summarization to yield readability, adequacy, and accuracy problems. To improve navigability and browsability of speech, we present three training agnostic post-processing techniques that address dialog concerns of readability, coherence, and adequacy. We integrate these improvements with user interfaces which communicate estimated summary metrics to aid user browsing heuristics. Quantitative evaluation metrics show a 19% improvement in summary quality. We discuss how summarization technologies can help people browse longform audio in trustworthy and readable ways.
长篇口语对话通过播客、访谈、辩论和会议提供丰富的信息流。虽然这种媒介的生产已经有了巨大的增长,但口语对话仍然具有挑战性,因为听力比阅读慢,而且相对于文本难以略读或导航。利用自动语音识别(ASR)和自动摘要的最新系统允许用户更好地浏览语音数据并搜索感兴趣的信息。然而,这些系统吸收不流利的语音,导致自动摘要产生可读性,充分性和准确性问题。为了提高语音的可导航性和可浏览性,我们提出了三种训练不可知论的后处理技术,以解决对话的可读性、连贯性和充分性问题。我们将这些改进与用户界面集成在一起,用户界面可以传达估计的汇总指标,以帮助用户进行启发式浏览。定量评估指标显示总结质量提高了19%。我们讨论了摘要技术如何帮助人们以可信和可读的方式浏览长格式音频。
{"title":"Improving Automatic Summarization for Browsing Longform Spoken Dialog","authors":"Daniel Li, Thomas Chen, Alec Zadikian, Albert Tung, Lydia B. Chilton","doi":"10.1145/3544548.3581339","DOIUrl":"https://doi.org/10.1145/3544548.3581339","url":null,"abstract":"Longform spoken dialog delivers rich streams of informative content through podcasts, interviews, debates, and meetings. While production of this medium has grown tremendously, spoken dialog remains challenging to consume as listening is slower than reading and difficult to skim or navigate relative to text. Recent systems leveraging automatic speech recognition (ASR) and automatic summarization allow users to better browse speech data and forage for information of interest. However, these systems intake disfluent speech which causes automatic summarization to yield readability, adequacy, and accuracy problems. To improve navigability and browsability of speech, we present three training agnostic post-processing techniques that address dialog concerns of readability, coherence, and adequacy. We integrate these improvements with user interfaces which communicate estimated summary metrics to aid user browsing heuristics. Quantitative evaluation metrics show a 19% improvement in summary quality. We discuss how summarization technologies can help people browse longform audio in trustworthy and readable ways.","PeriodicalId":314098,"journal":{"name":"Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems","volume":"31 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123451587","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Governor: Turning Open Government Data Portals into Interactive Databases 州长:将开放的政府数据门户转变为交互式数据库
Pub Date : 2023-04-19 DOI: 10.1145/3544548.3580868
Chang Liu, Arif Usta, J. Zhao, S. Salihoglu
The launch of open governmental data portals (OGDPs) has popularized the open data movement of last decade. Although the amount of data in OGDPs is increasing, their functionalities are limited to finding datasets with titles/descriptions and downloading the actual files. This hinders the end users, especially those without technical skills, to find the open data tables and make use of them. We present Governor, an open-sourced[17] web application developed to make OGDPs more accessible to end users by facilitating searching actual records in the tables, previewing them directly without downloading, and suggesting joinable and unionable tables to users based on their latest working tables. Governor also manages the provenance of integrated tables allowing users and their collaborators to easily trace back to the original tables in OGDP. We evaluate Governor with a two-part user study and the results demonstrate its value and effectiveness in finding and integrating tables in OGDP.
开放政府数据门户(ogdp)的推出使过去十年的开放数据运动得到普及。尽管ogdp中的数据量正在增加,但它们的功能仅限于查找带有标题/描述的数据集和下载实际文件。这阻碍了最终用户,特别是那些没有技术技能的用户,查找和使用开放数据表。我们提出了一个开源的web应用程序Governor[17],通过方便搜索表中的实际记录,无需下载即可直接预览,并根据用户最新的工作表向用户建议可合并和可联合的表,从而使最终用户更容易访问ogdp。总督还管理集成表的来源,允许用户及其协作者轻松地追溯到OGDP中的原始表。我们通过两部分的用户研究来评估总督,结果证明了它在查找和整合OGDP表格方面的价值和有效性。
{"title":"Governor: Turning Open Government Data Portals into Interactive Databases","authors":"Chang Liu, Arif Usta, J. Zhao, S. Salihoglu","doi":"10.1145/3544548.3580868","DOIUrl":"https://doi.org/10.1145/3544548.3580868","url":null,"abstract":"The launch of open governmental data portals (OGDPs) has popularized the open data movement of last decade. Although the amount of data in OGDPs is increasing, their functionalities are limited to finding datasets with titles/descriptions and downloading the actual files. This hinders the end users, especially those without technical skills, to find the open data tables and make use of them. We present Governor, an open-sourced[17] web application developed to make OGDPs more accessible to end users by facilitating searching actual records in the tables, previewing them directly without downloading, and suggesting joinable and unionable tables to users based on their latest working tables. Governor also manages the provenance of integrated tables allowing users and their collaborators to easily trace back to the original tables in OGDP. We evaluate Governor with a two-part user study and the results demonstrate its value and effectiveness in finding and integrating tables in OGDP.","PeriodicalId":314098,"journal":{"name":"Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems","volume":"32 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123484511","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Why Johnny Can’t Prompt: How Non-AI Experts Try (and Fail) to Design LLM Prompts 为什么约翰尼不能提示:非ai专家如何尝试(和失败)设计LLM提示
Pub Date : 2023-04-19 DOI: 10.1145/3544548.3581388
J. Zamfirescu-Pereira, Richmond Y. Wong, Bjoern Hartmann, Qiang Yang
Pre-trained large language models (“LLMs”) like GPT-3 can engage in fluent, multi-turn instruction-taking out-of-the-box, making them attractive materials for designing natural language interactions. Using natural language to steer LLM outputs (“prompting”) has emerged as an important design technique potentially accessible to non-AI-experts. Crafting effective prompts can be challenging, however, and prompt-based interactions are brittle. Here, we explore whether non-AI-experts can successfully engage in “end-user prompt engineering” using a design probe—a prototype LLM-based chatbot design tool supporting development and systematic evaluation of prompting strategies. Ultimately, our probe participants explored prompt designs opportunistically, not systematically, and struggled in ways echoing end-user programming systems and interactive machine learning systems. Expectations stemming from human-to-human instructional experiences, and a tendency to overgeneralize, were barriers to effective prompt design. These findings have implications for non-AI-expert-facing LLM-based tool design and for improving LLM-and-prompt literacy among programmers and the public, and present opportunities for further research.
像GPT-3这样的预训练大型语言模型(“llm”)可以进行流畅的、多回合的开箱即用的教学,使它们成为设计自然语言交互的有吸引力的材料。使用自然语言来引导法学硕士输出(“提示”)已经成为一种重要的设计技术,非人工智能专家也可以使用。然而,制作有效的提示是具有挑战性的,基于提示的交互是脆弱的。在这里,我们探讨了非人工智能专家是否可以使用设计探针成功地参与“终端用户提示工程”。设计探针是一种基于llm的原型聊天机器人设计工具,支持提示策略的开发和系统评估。最终,我们的调查参与者机会主义地而不是系统地探索了提示设计,并以与最终用户编程系统和交互式机器学习系统相呼应的方式进行了挣扎。源于人与人之间教学经验的期望,以及过度一般化的倾向,是有效提示设计的障碍。这些发现对非面向人工智能专家的基于法学硕士的工具设计和提高程序员和公众的法学硕士素养具有启示意义,并为进一步研究提供了机会。
{"title":"Why Johnny Can’t Prompt: How Non-AI Experts Try (and Fail) to Design LLM Prompts","authors":"J. Zamfirescu-Pereira, Richmond Y. Wong, Bjoern Hartmann, Qiang Yang","doi":"10.1145/3544548.3581388","DOIUrl":"https://doi.org/10.1145/3544548.3581388","url":null,"abstract":"Pre-trained large language models (“LLMs”) like GPT-3 can engage in fluent, multi-turn instruction-taking out-of-the-box, making them attractive materials for designing natural language interactions. Using natural language to steer LLM outputs (“prompting”) has emerged as an important design technique potentially accessible to non-AI-experts. Crafting effective prompts can be challenging, however, and prompt-based interactions are brittle. Here, we explore whether non-AI-experts can successfully engage in “end-user prompt engineering” using a design probe—a prototype LLM-based chatbot design tool supporting development and systematic evaluation of prompting strategies. Ultimately, our probe participants explored prompt designs opportunistically, not systematically, and struggled in ways echoing end-user programming systems and interactive machine learning systems. Expectations stemming from human-to-human instructional experiences, and a tendency to overgeneralize, were barriers to effective prompt design. These findings have implications for non-AI-expert-facing LLM-based tool design and for improving LLM-and-prompt literacy among programmers and the public, and present opportunities for further research.","PeriodicalId":314098,"journal":{"name":"Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems","volume":"27 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126782060","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 83
Colaroid: A Literate Programming Approach for Authoring Explorable Multi-Stage Tutorials Colaroid:编写可探索多阶段教程的识字编程方法
Pub Date : 2023-04-19 DOI: 10.1145/3544548.3581525
A. Wang, Andrew Head, Ashley Ge Zhang, Steve Oney, Christopher Brooks
Multi-stage programming tutorials are key learning resources for programmers, using progressive incremental steps to teach them how to build larger software systems. A good multi-stage tutorial describes the code clearly, explains the rationale and code changes for each step, and allows readers to experiment as they work through the tutorial. In practice, it is time-consuming for authors to create tutorials with these attributes. In this paper, we introduce Colaroid, an interactive authoring tool for creating high quality multi-stage tutorials. Colaroid tutorials are augmented computational notebooks, where snippets and outputs represent a snapshot of a project, with source code differences highlighted, complete source code context for each snippet, and the ability to load and tinker with any stage of the project in a linked IDE. In two laboratory studies, we found Colaroid makes it easy to create multi-stage tutorials, while offering advantages to readers compared to video and web-based tutorials.
多阶段编程教程是程序员的关键学习资源,使用渐进的增量步骤来教他们如何构建更大的软件系统。一个好的多阶段教程清晰地描述了代码,解释了每个步骤的基本原理和代码更改,并允许读者在学习教程时进行实验。在实践中,作者创建带有这些属性的教程非常耗时。在本文中,我们介绍Colaroid,一个交互式创作工具,用于创建高质量的多阶段教程。Colaroid教程是增强的计算笔记本,其中代码片段和输出代表项目的快照,突出显示源代码差异,每个代码片段的完整源代码上下文,以及在链接的IDE中加载和修改项目任何阶段的能力。在两项实验室研究中,我们发现Colaroid可以很容易地创建多阶段教程,同时与视频和网络教程相比,它为读者提供了优势。
{"title":"Colaroid: A Literate Programming Approach for Authoring Explorable Multi-Stage Tutorials","authors":"A. Wang, Andrew Head, Ashley Ge Zhang, Steve Oney, Christopher Brooks","doi":"10.1145/3544548.3581525","DOIUrl":"https://doi.org/10.1145/3544548.3581525","url":null,"abstract":"Multi-stage programming tutorials are key learning resources for programmers, using progressive incremental steps to teach them how to build larger software systems. A good multi-stage tutorial describes the code clearly, explains the rationale and code changes for each step, and allows readers to experiment as they work through the tutorial. In practice, it is time-consuming for authors to create tutorials with these attributes. In this paper, we introduce Colaroid, an interactive authoring tool for creating high quality multi-stage tutorials. Colaroid tutorials are augmented computational notebooks, where snippets and outputs represent a snapshot of a project, with source code differences highlighted, complete source code context for each snippet, and the ability to load and tinker with any stage of the project in a linked IDE. In two laboratory studies, we found Colaroid makes it easy to create multi-stage tutorials, while offering advantages to readers compared to video and web-based tutorials.","PeriodicalId":314098,"journal":{"name":"Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122375784","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Slide4N: Creating Presentation Slides from Computational Notebooks with Human-AI Collaboration Slide4N:通过人机协作从计算笔记本中创建演示幻灯片
Pub Date : 2023-04-19 DOI: 10.1145/3544548.3580753
Fengjie Wang, Xuye Liu, Oujing Liu, Ali Neshati, Tengfei Ma, Min Zhu, J. Zhao
Data scientists often have to use other presentation tools (e.g., Microsoft PowerPoint) to create slides to communicate their analysis obtained using computational notebooks. Much tedious and repetitive work is needed to transfer the routines of notebooks (e.g., code, plots) to the presentable contents on slides (e.g., bullet points, figures). We propose a human-AI collaborative approach and operationalize it within Slide4N, an interactive AI assistant for data scientists to create slides from computational notebooks. Slide4N leverages advanced natural language processing techniques to distill key information from user-selected notebook cells and then renders them in appropriate slide layouts. The tool also provides intuitive interactions that allow further refinement and customization of the generated slides. We evaluated Slide4N with a two-part user study, where participants appreciated this human-AI collaborative approach compared to fully-manual or fully-automatic methods. The results also indicate the usefulness and effectiveness of Slide4N in slide creation tasks from notebooks.
数据科学家通常不得不使用其他的演示工具(例如,微软的PowerPoint)来创建幻灯片,以交流他们使用计算笔记本获得的分析。将笔记本上的常规内容(如代码、图表)转换为幻灯片上可展示的内容(如要点、数字)需要进行大量乏味和重复的工作。我们提出了一种人类与人工智能的协作方法,并在Slide4N中进行了操作,Slide4N是一种交互式人工智能助手,用于数据科学家从计算笔记本中创建幻灯片。Slide4N利用先进的自然语言处理技术,从用户选择的笔记本单元格中提取关键信息,然后将其呈现在适当的幻灯片布局中。该工具还提供直观的交互,允许进一步改进和定制生成的幻灯片。我们通过两部分的用户研究来评估Slide4N,与全手动或全自动方法相比,参与者更欣赏这种人类-人工智能协作方法。结果还表明Slide4N在从笔记本中创建幻灯片任务中的有用性和有效性。
{"title":"Slide4N: Creating Presentation Slides from Computational Notebooks with Human-AI Collaboration","authors":"Fengjie Wang, Xuye Liu, Oujing Liu, Ali Neshati, Tengfei Ma, Min Zhu, J. Zhao","doi":"10.1145/3544548.3580753","DOIUrl":"https://doi.org/10.1145/3544548.3580753","url":null,"abstract":"Data scientists often have to use other presentation tools (e.g., Microsoft PowerPoint) to create slides to communicate their analysis obtained using computational notebooks. Much tedious and repetitive work is needed to transfer the routines of notebooks (e.g., code, plots) to the presentable contents on slides (e.g., bullet points, figures). We propose a human-AI collaborative approach and operationalize it within Slide4N, an interactive AI assistant for data scientists to create slides from computational notebooks. Slide4N leverages advanced natural language processing techniques to distill key information from user-selected notebook cells and then renders them in appropriate slide layouts. The tool also provides intuitive interactions that allow further refinement and customization of the generated slides. We evaluated Slide4N with a two-part user study, where participants appreciated this human-AI collaborative approach compared to fully-manual or fully-automatic methods. The results also indicate the usefulness and effectiveness of Slide4N in slide creation tasks from notebooks.","PeriodicalId":314098,"journal":{"name":"Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems","volume":"37 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122778009","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Transcending the “Male Code”: Implicit Masculine Biases in NLP Contexts 超越“男性密码”:NLP语境中的内隐男性偏见
Pub Date : 2023-04-19 DOI: 10.1145/3544548.3581017
Katie Seaborn, S. Chandra, Thibault Fabre
Critical scholarship has elevated the problem of gender bias in data sets used to train virtual assistants (VAs). Most work has focused on explicit biases in language, especially against women, girls, femme-identifying people, and genderqueer folk; implicit associations through word embeddings; and limited models of gender and masculinities, especially toxic masculinities, conflation of sex and gender, and a sex/gender binary framing of the masculine as diametric to the feminine. Yet, we must also interrogate how masculinities are “coded” into language and the assumption of “male” as the linguistic default: implicit masculine biases. To this end, we examined two natural language processing (NLP) data sets. We found that when gendered language was present, so were gender biases and especially masculine biases. Moreover, these biases related in nuanced ways to the NLP context. We offer a new dictionary called AVA that covers ambiguous associations between gendered language and the language of VAs.
批判性的学术研究提高了用于训练虚拟助理(VAs)的数据集中的性别偏见问题。大多数研究都集中在语言上的明显偏见,尤其是针对女性、女孩、女性认同者和性别酷儿群体的偏见;词嵌入的内隐联想;性别和男子气概的有限模型,特别是有毒的男子气概,性和性别的合并,以及性别/性别二元框架,男性与女性截然相反。然而,我们也必须质问男性气质是如何被“编码”到语言中的,以及“男性”作为语言默认值的假设:隐性的男性偏见。为此,我们研究了两个自然语言处理(NLP)数据集。我们发现,当性别语言出现时,性别偏见,尤其是男性偏见也会出现。此外,这些偏见以微妙的方式与NLP环境相关。我们提供了一个名为AVA的新词典,涵盖了性别语言和VAs语言之间的模糊关联。
{"title":"Transcending the “Male Code”: Implicit Masculine Biases in NLP Contexts","authors":"Katie Seaborn, S. Chandra, Thibault Fabre","doi":"10.1145/3544548.3581017","DOIUrl":"https://doi.org/10.1145/3544548.3581017","url":null,"abstract":"Critical scholarship has elevated the problem of gender bias in data sets used to train virtual assistants (VAs). Most work has focused on explicit biases in language, especially against women, girls, femme-identifying people, and genderqueer folk; implicit associations through word embeddings; and limited models of gender and masculinities, especially toxic masculinities, conflation of sex and gender, and a sex/gender binary framing of the masculine as diametric to the feminine. Yet, we must also interrogate how masculinities are “coded” into language and the assumption of “male” as the linguistic default: implicit masculine biases. To this end, we examined two natural language processing (NLP) data sets. We found that when gendered language was present, so were gender biases and especially masculine biases. Moreover, these biases related in nuanced ways to the NLP context. We offer a new dictionary called AVA that covers ambiguous associations between gendered language and the language of VAs.","PeriodicalId":314098,"journal":{"name":"Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems","volume":"48 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122837494","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
期刊
Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1