首页 > 最新文献

Proceedings of the 28th International Conference on Intelligent User Interfaces最新文献

英文 中文
Pearl: A Technology Probe for Machine-Assisted Reflection on Personal Data Pearl:个人数据的机器辅助反射技术探索
Pub Date : 2023-03-27 DOI: 10.1145/3581641.3584054
Matthew Jörke, Yasaman S. Sefidgar, Talie Massachi, Jina Suh, Gonzalo A. Ramos
Reflection on one’s personal data can be an effective tool for supporting wellbeing. However, current wellbeing reflection support tools tend to offer a one-size-fits-all approach, ignoring the diversity of people’s wellbeing goals and their agency in the self-reflection process. In this work, we identify an opportunity to help people work toward their wellbeing goals by empowering them to reflect on their data on their own terms. Through a formative study, we inform the design and implementation of Pearl, a workplace wellbeing reflection support tool that allows users to explore their personal data in relation to their wellbeing goal. Pearl is a calendar-based interactive machine teaching system that allows users to visualize data sources and tag regions of interest on their calendar. In return, the system provides insights about these tags that can be saved to a reflection journal. We used Pearl as a technology probe with 12 participants without data science expertise and found that all participants successfully gained insights into their workplace wellbeing. In our analysis, we discuss how Pearl’s capabilities facilitate insights, the role of machine assistance in the self-reflection process, and the data sources that participants found most insightful. We conclude with design dimensions for intelligent reflection support systems as inspiration for future work.
对个人数据的反思可以成为支持幸福的有效工具。然而,目前的幸福反思支持工具倾向于提供一种一刀切的方法,忽视了人们的幸福目标和他们在自我反思过程中的作用的多样性。在这项工作中,我们发现了一个机会,让人们能够以自己的方式反思自己的数据,从而帮助他们朝着自己的幸福目标努力。通过一项形成性研究,我们为Pearl的设计和实施提供了信息,Pearl是一种工作场所健康反思支持工具,允许用户探索与他们的健康目标相关的个人数据。Pearl是一个基于日历的交互式机器教学系统,它允许用户可视化数据源,并在日历上标记感兴趣的区域。作为回报,系统提供了关于这些标签的见解,这些见解可以保存到反射日志中。我们使用Pearl作为技术探针,对12名没有数据科学专业知识的参与者进行了调查,发现所有参与者都成功地了解了他们的工作场所幸福感。在我们的分析中,我们讨论了Pearl的功能如何促进洞察力,机器在自我反思过程中的辅助作用,以及参与者发现最具洞察力的数据源。我们总结了智能反射支持系统的设计尺寸,作为未来工作的灵感。
{"title":"Pearl: A Technology Probe for Machine-Assisted Reflection on Personal Data","authors":"Matthew Jörke, Yasaman S. Sefidgar, Talie Massachi, Jina Suh, Gonzalo A. Ramos","doi":"10.1145/3581641.3584054","DOIUrl":"https://doi.org/10.1145/3581641.3584054","url":null,"abstract":"Reflection on one’s personal data can be an effective tool for supporting wellbeing. However, current wellbeing reflection support tools tend to offer a one-size-fits-all approach, ignoring the diversity of people’s wellbeing goals and their agency in the self-reflection process. In this work, we identify an opportunity to help people work toward their wellbeing goals by empowering them to reflect on their data on their own terms. Through a formative study, we inform the design and implementation of Pearl, a workplace wellbeing reflection support tool that allows users to explore their personal data in relation to their wellbeing goal. Pearl is a calendar-based interactive machine teaching system that allows users to visualize data sources and tag regions of interest on their calendar. In return, the system provides insights about these tags that can be saved to a reflection journal. We used Pearl as a technology probe with 12 participants without data science expertise and found that all participants successfully gained insights into their workplace wellbeing. In our analysis, we discuss how Pearl’s capabilities facilitate insights, the role of machine assistance in the self-reflection process, and the data sources that participants found most insightful. We conclude with design dimensions for intelligent reflection support systems as inspiration for future work.","PeriodicalId":118159,"journal":{"name":"Proceedings of the 28th International Conference on Intelligent User Interfaces","volume":"106 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-03-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121100724","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
It Seems Smart, but It Acts Stupid: Development of Trust in AI Advice in a Repeated Legal Decision-Making Task 看似聪明,实则愚蠢:在重复的法律决策任务中发展对人工智能建议的信任
Pub Date : 2023-03-27 DOI: 10.1145/3581641.3584058
Patricia K. Kahr, G. Rooks, M. Willemsen, Chris C. P. Snijders
Humans increasingly interact with AI systems, and successful interactions rely on individuals trusting such systems (when appropriate). Considering that trust is fragile and often cannot be restored quickly, we focus on how trust develops over time in a human-AI-interaction scenario. In a 2x2 between-subject experiment, we test how model accuracy (high vs. low) and type of explanation (human-like vs. not) affect trust in AI over time. We study a complex decision-making task in which individuals estimate jail time for 20 criminal law cases with AI advice. Results show that trust is significantly higher for high-accuracy models. Also, behavioral trust does not decline, and subjective trust even increases significantly with high accuracy. Human-like explanations did not generally affect trust but boosted trust in high-accuracy models.
人类与人工智能系统的互动越来越多,而成功的互动依赖于个人对这些系统的信任(在适当的时候)。考虑到信任是脆弱的,往往无法迅速恢复,我们关注的是在人类与人工智能交互的场景中,信任是如何随着时间的推移而发展的。在受试者之间的2x2实验中,我们测试了模型准确性(高vs低)和解释类型(类人vs非类人)如何随着时间的推移影响对AI的信任。我们研究了一个复杂的决策任务,在这个任务中,个人根据人工智能的建议估计20个刑事法律案件的监禁时间。结果表明,高精度模型的信任度显著提高。行为信任也没有下降,主观信任甚至显著增加,准确率高。类似人类的解释通常不会影响信任,但会提高对高精度模型的信任。
{"title":"It Seems Smart, but It Acts Stupid: Development of Trust in AI Advice in a Repeated Legal Decision-Making Task","authors":"Patricia K. Kahr, G. Rooks, M. Willemsen, Chris C. P. Snijders","doi":"10.1145/3581641.3584058","DOIUrl":"https://doi.org/10.1145/3581641.3584058","url":null,"abstract":"Humans increasingly interact with AI systems, and successful interactions rely on individuals trusting such systems (when appropriate). Considering that trust is fragile and often cannot be restored quickly, we focus on how trust develops over time in a human-AI-interaction scenario. In a 2x2 between-subject experiment, we test how model accuracy (high vs. low) and type of explanation (human-like vs. not) affect trust in AI over time. We study a complex decision-making task in which individuals estimate jail time for 20 criminal law cases with AI advice. Results show that trust is significantly higher for high-accuracy models. Also, behavioral trust does not decline, and subjective trust even increases significantly with high accuracy. Human-like explanations did not generally affect trust but boosted trust in high-accuracy models.","PeriodicalId":118159,"journal":{"name":"Proceedings of the 28th International Conference on Intelligent User Interfaces","volume":"32 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-03-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115988761","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Lessons Learned from Designing and Evaluating CLAICA: A Continuously Learning AI Cognitive Assistant 设计和评估CLAICA的经验教训:一个持续学习的人工智能认知助手
Pub Date : 2023-03-27 DOI: 10.1145/3581641.3584042
Samuel Kernan Freire, E. Niforatos, Chaofan Wang, Santiago Ruiz-Arenas, Mina Foosherian, S. Wellsandt, A. Bozzon
Learning to operate a complex system, such as an agile production line, can be a daunting task. The high variability in products and frequent reconfigurations make it difficult to keep documentation up-to-date and share new knowledge amongst factory workers. We introduce CLAICA, a Continuously Learning AI Cognitive Assistant that supports workers in the aforementioned scenario. CLAICA learns from (experienced) workers, formalizes new knowledge, stores it in a knowledge base, along with contextual information, and shares it when relevant. We conducted a user study with 83 participants who performed eight knowledge exchange tasks with CLAICA, completed a survey, and provided qualitative feedback. Our results provide a deeper understanding of how prior training, context expertise, and interaction modality affect the user experience of cognitive assistants. We draw on our results to elicit design and evaluation guidelines for cognitive assistants that support knowledge exchange in fast-paced and demanding environments, such as an agile production line.
学习操作一个复杂的系统,比如一条灵活的生产线,可能是一项艰巨的任务。产品的高度可变性和频繁的重新配置使得保持文档更新和在工厂工人之间分享新知识变得困难。我们介绍了CLAICA,这是一款持续学习的人工智能认知助手,可以为上述场景中的工作人员提供支持。CLAICA向(有经验的)工人学习,将新知识形式化,将其与上下文信息一起存储在知识库中,并在相关时共享。我们对83名参与者进行了一项用户研究,他们通过CLAICA完成了8项知识交换任务,完成了一项调查,并提供了定性反馈。我们的研究结果更深入地了解了先前的培训、上下文专业知识和交互方式如何影响认知助手的用户体验。我们利用我们的结果来引出认知助手的设计和评估指南,这些助手支持在快节奏和苛刻的环境中进行知识交换,例如敏捷生产线。
{"title":"Lessons Learned from Designing and Evaluating CLAICA: A Continuously Learning AI Cognitive Assistant","authors":"Samuel Kernan Freire, E. Niforatos, Chaofan Wang, Santiago Ruiz-Arenas, Mina Foosherian, S. Wellsandt, A. Bozzon","doi":"10.1145/3581641.3584042","DOIUrl":"https://doi.org/10.1145/3581641.3584042","url":null,"abstract":"Learning to operate a complex system, such as an agile production line, can be a daunting task. The high variability in products and frequent reconfigurations make it difficult to keep documentation up-to-date and share new knowledge amongst factory workers. We introduce CLAICA, a Continuously Learning AI Cognitive Assistant that supports workers in the aforementioned scenario. CLAICA learns from (experienced) workers, formalizes new knowledge, stores it in a knowledge base, along with contextual information, and shares it when relevant. We conducted a user study with 83 participants who performed eight knowledge exchange tasks with CLAICA, completed a survey, and provided qualitative feedback. Our results provide a deeper understanding of how prior training, context expertise, and interaction modality affect the user experience of cognitive assistants. We draw on our results to elicit design and evaluation guidelines for cognitive assistants that support knowledge exchange in fast-paced and demanding environments, such as an agile production line.","PeriodicalId":118159,"journal":{"name":"Proceedings of the 28th International Conference on Intelligent User Interfaces","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-03-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116858357","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Pragmatic Communication with Embodied Agents 具身行为人的语用交际
Pub Date : 2023-03-27 DOI: 10.1145/3581641.3584101
J. Chai
With the emergence of a new generation of embodied AI agents (e.g., cognitive robots), it has become increasingly important to empower these agents with the ability to learn and collaborate with humans through language communication. Despite recent advances, language communication in embodied AI still faces many challenges. Human language not only needs to ground to agents’ perception and action but also needs to facilitate collaboration between humans and agents. To address these challenges, I will introduce several efforts in my lab that study pragmatic communication with embodied agents. I will talk about how language use is shaped by shared experience and knowledge (i.e., common ground) and how collaborative effort is important to mediate perceptual differences and handle exceptions. I will discuss task learning by following language instructions and highlight the need for neuro-symbolic representations for situation awareness and transparency. I will further present explicit modeling of partners’ goals, beliefs, and abilities (i.e., theory of mind) and discuss its role in language communication for situated collaborative tasks.
随着新一代具身人工智能代理(如认知机器人)的出现,赋予这些代理通过语言交流学习和与人类合作的能力变得越来越重要。尽管最近取得了一些进展,但嵌入人工智能的语言交流仍然面临许多挑战。人类语言不仅需要基于智能体的感知和行为,还需要促进人类与智能体之间的协作。为了应对这些挑战,我将在我的实验室中介绍几项研究与具身代理的语用交流的努力。我将讨论如何通过共享经验和知识(即共同基础)来塑造语言的使用,以及协作努力对于调解感知差异和处理异常的重要性。我将通过遵循语言指令来讨论任务学习,并强调对情境感知和透明度的神经符号表征的需求。我将进一步展示合作伙伴的目标、信念和能力(即心智理论)的明确模型,并讨论其在情境协作任务的语言交流中的作用。
{"title":"Pragmatic Communication with Embodied Agents","authors":"J. Chai","doi":"10.1145/3581641.3584101","DOIUrl":"https://doi.org/10.1145/3581641.3584101","url":null,"abstract":"With the emergence of a new generation of embodied AI agents (e.g., cognitive robots), it has become increasingly important to empower these agents with the ability to learn and collaborate with humans through language communication. Despite recent advances, language communication in embodied AI still faces many challenges. Human language not only needs to ground to agents’ perception and action but also needs to facilitate collaboration between humans and agents. To address these challenges, I will introduce several efforts in my lab that study pragmatic communication with embodied agents. I will talk about how language use is shaped by shared experience and knowledge (i.e., common ground) and how collaborative effort is important to mediate perceptual differences and handle exceptions. I will discuss task learning by following language instructions and highlight the need for neuro-symbolic representations for situation awareness and transparency. I will further present explicit modeling of partners’ goals, beliefs, and abilities (i.e., theory of mind) and discuss its role in language communication for situated collaborative tasks.","PeriodicalId":118159,"journal":{"name":"Proceedings of the 28th International Conference on Intelligent User Interfaces","volume":"13 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-03-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130806824","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Don’t fail me! The Level 5 Autonomous Driving Information Dilemma regarding Transparency and User Experience 别让我失望!关于透明度和用户体验的5级自动驾驶信息困境
Pub Date : 2023-03-27 DOI: 10.1145/3581641.3584085
Tobias Schneider, J. Hois, Alischa Rosenstein, Sandra Metzl, Ansgar R. S. Gerlicher, Sabiha Ghellal, Steve Love
Autonomous vehicles can behave unexpectedly, as automated systems that rely on data-driven machine learning have shown to infer false predictions or misclassifications, e.g., due to stickers on traffic signs, and thus fail in some situations. In critical situations, system designs must guarantee safety and reliability. However, in non-critical situations, the possibility of failures resulting in unexpected behaviour should be considered, as they negatively impact the passenger’s user experience and acceptance. We analyse if an interactive conversational user interface can mitigate negative experiences when interacting with imperfect artificial intelligence systems. In our quantitative interactive online survey (N=113) and comparative qualitative Wizard of Oz study (N=8), users were able to interact with an autonomous SAE level 5 driving simulation. Our findings demonstrate that increased transparency improves user experience and acceptance. Furthermore, we show that additional information in failure scenarios can lead to an information dilemma and should be implemented carefully.
自动驾驶汽车的行为可能出乎意料,因为依赖于数据驱动的机器学习的自动化系统已经显示出错误的预测或错误的分类,例如,由于交通标志上的贴纸,因此在某些情况下会失败。在紧急情况下,系统设计必须保证安全性和可靠性。然而,在非关键情况下,应考虑故障导致意外行为的可能性,因为它们会对乘客的用户体验和接受度产生负面影响。我们分析了当与不完善的人工智能系统交互时,交互式会话用户界面是否可以减轻负面体验。在我们的定量互动在线调查(N=113)和比较定性的Wizard of Oz研究(N=8)中,用户能够与自动驾驶SAE 5级驾驶模拟进行互动。我们的研究结果表明,增加透明度可以改善用户体验和接受度。此外,我们还表明,故障场景中的附加信息可能导致信息困境,应谨慎实现。
{"title":"Don’t fail me! The Level 5 Autonomous Driving Information Dilemma regarding Transparency and User Experience","authors":"Tobias Schneider, J. Hois, Alischa Rosenstein, Sandra Metzl, Ansgar R. S. Gerlicher, Sabiha Ghellal, Steve Love","doi":"10.1145/3581641.3584085","DOIUrl":"https://doi.org/10.1145/3581641.3584085","url":null,"abstract":"Autonomous vehicles can behave unexpectedly, as automated systems that rely on data-driven machine learning have shown to infer false predictions or misclassifications, e.g., due to stickers on traffic signs, and thus fail in some situations. In critical situations, system designs must guarantee safety and reliability. However, in non-critical situations, the possibility of failures resulting in unexpected behaviour should be considered, as they negatively impact the passenger’s user experience and acceptance. We analyse if an interactive conversational user interface can mitigate negative experiences when interacting with imperfect artificial intelligence systems. In our quantitative interactive online survey (N=113) and comparative qualitative Wizard of Oz study (N=8), users were able to interact with an autonomous SAE level 5 driving simulation. Our findings demonstrate that increased transparency improves user experience and acceptance. Furthermore, we show that additional information in failure scenarios can lead to an information dilemma and should be implemented carefully.","PeriodicalId":118159,"journal":{"name":"Proceedings of the 28th International Conference on Intelligent User Interfaces","volume":"35 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-03-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115591543","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
TimToShape: Supporting Practice of Musical Instruments by Visualizing Timbre with 2D Shapes based on Crossmodal Correspondences TimToShape:通过基于交叉模态对应的2D图形可视化音色来支持乐器实践
Pub Date : 2023-03-27 DOI: 10.1145/3581641.3584053
Kota Arai, Yutaro Hirao, Takuji Narumi, Tomohiko Nakamura, Shinnosuke Takamichi, Shigeo Yoshida
Timbre is high-dimensional and sensuous, making it difficult for musical-instrument learners to improve their timbre. Although some systems exist to improve timbre, they require expert labeling for timbre evaluation; however, solely visualizing the results of unsupervised learning lacks the intuitiveness of feedback because human perception is not considered. Therefore, we employ crossmodal correspondences for intuitive visualization of the timbre. We designed TimToShape, a system that visualizes timbre with 2D shapes based on the user’s input of timbre–shape correspondences. TimToShape generates a shape morphed by linear interpolation according to the timbre’s position in the latent space, which is obtained by unsupervised learning with a variational autoencoder (VAE). We confirmed that people perceived shapes generated by TimToShape to correspond more to timbre than randomly generated shapes. Furthermore, a user study of six violin players revealed that TimToShape was well-received in terms of visual clarity and interpretability.
音色是高维的、感性的,这给乐器学习者提高音色带来了困难。虽然存在一些系统来改善音色,但它们需要专家标记来评估音色;然而,仅仅将无监督学习的结果可视化缺乏反馈的直观性,因为没有考虑人的感知。因此,我们采用跨模对应来直观地可视化音色。我们设计了TimToShape,这是一个基于用户输入音色形状对应关系的二维音色可视化系统。TimToShape根据音色在潜在空间中的位置,通过线性插值生成一个变形的形状,该形状是通过变分自编码器(VAE)的无监督学习获得的。我们证实,人们认为TimToShape生成的形状比随机生成的形状更符合音色。此外,一项针对六位小提琴手的用户研究显示,TimToShape在视觉清晰度和可解释性方面广受欢迎。
{"title":"TimToShape: Supporting Practice of Musical Instruments by Visualizing Timbre with 2D Shapes based on Crossmodal Correspondences","authors":"Kota Arai, Yutaro Hirao, Takuji Narumi, Tomohiko Nakamura, Shinnosuke Takamichi, Shigeo Yoshida","doi":"10.1145/3581641.3584053","DOIUrl":"https://doi.org/10.1145/3581641.3584053","url":null,"abstract":"Timbre is high-dimensional and sensuous, making it difficult for musical-instrument learners to improve their timbre. Although some systems exist to improve timbre, they require expert labeling for timbre evaluation; however, solely visualizing the results of unsupervised learning lacks the intuitiveness of feedback because human perception is not considered. Therefore, we employ crossmodal correspondences for intuitive visualization of the timbre. We designed TimToShape, a system that visualizes timbre with 2D shapes based on the user’s input of timbre–shape correspondences. TimToShape generates a shape morphed by linear interpolation according to the timbre’s position in the latent space, which is obtained by unsupervised learning with a variational autoencoder (VAE). We confirmed that people perceived shapes generated by TimToShape to correspond more to timbre than randomly generated shapes. Furthermore, a user study of six violin players revealed that TimToShape was well-received in terms of visual clarity and interpretability.","PeriodicalId":118159,"journal":{"name":"Proceedings of the 28th International Conference on Intelligent User Interfaces","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-03-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123995533","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Follow the Successful Herd: Towards Explanations for Improved Use and Mental Models of Natural Language Systems 跟随成功的羊群:对自然语言系统的改进使用和心理模型的解释
Pub Date : 2023-03-27 DOI: 10.1145/3581641.3584088
Michelle Brachman, Qian Pan, H. Do, Casey Dugan, Arunima Chaudhary, James M. Johnson, Priyanshu Rai, T. Chakraborti, T. Gschwind, Jim Laredo, Christoph Miksovic, P. Scotton, Kartik Talamadupula, Gegi Thomas
While natural language systems continue improving, they are still imperfect. If a user has a better understanding of how a system works, they may be able to better accomplish their goals even in imperfect systems. We explored whether explanations can support effective authoring of natural language utterances and how those explanations impact users’ mental models in the context of a natural language system that generates small programs. Through an online study (n=252), we compared two main types of explanations: 1) system-focused, which provide information about how the system processes utterances and matches terms to a knowledge base, and 2) social, which provide information about how other users have successfully interacted with the system. Our results indicate that providing social suggestions of terms to add to an utterance helped users to repair and generate correct flows more than system-focused explanations or social recommendations of words to modify. We also found that participants commonly understood some mechanisms of the natural language system, such as the matching of terms to a knowledge base, but they often lacked other critical knowledge, such as how the system handled structuring and ordering. Based on these findings, we make design recommendations for supporting interactions with and understanding of natural language systems.
虽然自然语言系统在不断改进,但它们仍然不完美。如果用户对系统如何工作有了更好的理解,即使在不完美的系统中,他们也可以更好地完成目标。我们探讨了解释是否可以支持自然语言话语的有效创作,以及在生成小程序的自然语言系统的背景下,这些解释如何影响用户的心理模型。通过一项在线研究(n=252),我们比较了两种主要类型的解释:1)以系统为中心的,它提供了关于系统如何处理话语并将术语与知识库相匹配的信息;2)社会的,它提供了关于其他用户如何成功与系统交互的信息。我们的研究结果表明,提供词汇的社会建议来添加到话语中,比以系统为中心的解释或词汇的社会建议来修改更能帮助用户修复和生成正确的流程。我们还发现,参与者通常理解自然语言系统的一些机制,例如术语与知识库的匹配,但他们通常缺乏其他关键知识,例如系统如何处理结构和排序。基于这些发现,我们提出了支持与自然语言系统交互和理解的设计建议。
{"title":"Follow the Successful Herd: Towards Explanations for Improved Use and Mental Models of Natural Language Systems","authors":"Michelle Brachman, Qian Pan, H. Do, Casey Dugan, Arunima Chaudhary, James M. Johnson, Priyanshu Rai, T. Chakraborti, T. Gschwind, Jim Laredo, Christoph Miksovic, P. Scotton, Kartik Talamadupula, Gegi Thomas","doi":"10.1145/3581641.3584088","DOIUrl":"https://doi.org/10.1145/3581641.3584088","url":null,"abstract":"While natural language systems continue improving, they are still imperfect. If a user has a better understanding of how a system works, they may be able to better accomplish their goals even in imperfect systems. We explored whether explanations can support effective authoring of natural language utterances and how those explanations impact users’ mental models in the context of a natural language system that generates small programs. Through an online study (n=252), we compared two main types of explanations: 1) system-focused, which provide information about how the system processes utterances and matches terms to a knowledge base, and 2) social, which provide information about how other users have successfully interacted with the system. Our results indicate that providing social suggestions of terms to add to an utterance helped users to repair and generate correct flows more than system-focused explanations or social recommendations of words to modify. We also found that participants commonly understood some mechanisms of the natural language system, such as the matching of terms to a knowledge base, but they often lacked other critical knowledge, such as how the system handled structuring and ordering. Based on these findings, we make design recommendations for supporting interactions with and understanding of natural language systems.","PeriodicalId":118159,"journal":{"name":"Proceedings of the 28th International Conference on Intelligent User Interfaces","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-03-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124765141","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Categorical and Continuous Features in Counterfactual Explanations of AI Systems 人工智能系统反事实解释中的分类和连续特征
Pub Date : 2023-03-27 DOI: 10.1145/3581641.3584090
Greta Warren, R. Byrne, Markt. Keane
Recently, eXplainable AI (XAI) research has focused on the use of counterfactual explanations to address interpretability, algorithmic recourse, and bias in AI system decision-making. The proponents of these algorithms claim they meet users’ requirements for counterfactual explanations. For instance, many claim that the output of their algorithms work as explanations because they prioritise "plausible", "actionable" or "causally important" features in their generated counterfactuals. However, very few of these claims have been tested in controlled psychological studies, and we know very little about which aspects of counterfactual explanations help users to understand AI system decisions. Furthermore, we do not know whether counterfactual explanations are an advance on more traditional causal explanations that have a much longer history in AI (in explaining expert systems and decision trees). Accordingly, we carried out two user studies to (i) test a fundamental distinction in feature-types, between categorical and continuous features, and (ii) compare the relative effectiveness of counterfactual and causal explanations. The studies used a simulated, automated decision-making app that determined safe driving limits after drinking alcohol, based on predicted blood alcohol content, and user responses were measured objectively (users’ predictive accuracy) and subjectively (users’ satisfaction and trust judgments). Study 1 (N=127) showed that users understand explanations referring to categorical features more readily than those referring to continuous features. It also discovered a dissociation between objective and subjective measures: counterfactual explanations elicited higher accuracy of predictions than no-explanation control descriptions but no higher accuracy than causal explanations, yet counterfactual explanations elicited greater satisfaction and trust judgments than causal explanations. Study 2 (N=211) found that users were more accurate for categorically-transformed features compared to continuous ones, and also replicated the results of Study 1. The findings delineate important boundary conditions for current and future counterfactual explanation methods in XAI.
最近,可解释人工智能(XAI)研究的重点是使用反事实解释来解决人工智能系统决策中的可解释性、算法追索权和偏见。这些算法的支持者声称,它们满足了用户对反事实解释的要求。例如,许多人声称,他们的算法的输出可以作为解释,因为他们在生成的反事实中优先考虑“合理的”、“可操作的”或“因果重要的”特征。然而,这些说法很少在对照心理学研究中得到测试,我们对反事实解释的哪些方面有助于用户理解人工智能系统的决策知之甚少。此外,我们不知道反事实解释是否是传统因果解释的进步,后者在人工智能(解释专家系统和决策树)中有着更长的历史。因此,我们进行了两项用户研究,以(i)测试特征类型的基本区别,在分类和连续特征之间,以及(ii)比较反事实和因果解释的相对有效性。这些研究使用了一个模拟的自动化决策应用程序,该应用程序根据预测的血液酒精含量确定饮酒后的安全驾驶限制,并客观地测量用户的反应(用户的预测准确性)和主观地(用户的满意度和信任判断)。研究1 (N=127)表明,用户更容易理解涉及分类特征的解释,而不是涉及连续特征的解释。它还发现了客观和主观测量之间的分离:反事实解释比无解释控制描述引起更高的预测准确性,但不高于因果解释的准确性,然而反事实解释比因果解释引起更高的满意度和信任判断。研究2 (N=211)发现,与连续特征相比,用户对分类转换后的特征更准确,也重复了研究1的结果。这些发现为当前和未来的XAI反事实解释方法描绘了重要的边界条件。
{"title":"Categorical and Continuous Features in Counterfactual Explanations of AI Systems","authors":"Greta Warren, R. Byrne, Markt. Keane","doi":"10.1145/3581641.3584090","DOIUrl":"https://doi.org/10.1145/3581641.3584090","url":null,"abstract":"Recently, eXplainable AI (XAI) research has focused on the use of counterfactual explanations to address interpretability, algorithmic recourse, and bias in AI system decision-making. The proponents of these algorithms claim they meet users’ requirements for counterfactual explanations. For instance, many claim that the output of their algorithms work as explanations because they prioritise \"plausible\", \"actionable\" or \"causally important\" features in their generated counterfactuals. However, very few of these claims have been tested in controlled psychological studies, and we know very little about which aspects of counterfactual explanations help users to understand AI system decisions. Furthermore, we do not know whether counterfactual explanations are an advance on more traditional causal explanations that have a much longer history in AI (in explaining expert systems and decision trees). Accordingly, we carried out two user studies to (i) test a fundamental distinction in feature-types, between categorical and continuous features, and (ii) compare the relative effectiveness of counterfactual and causal explanations. The studies used a simulated, automated decision-making app that determined safe driving limits after drinking alcohol, based on predicted blood alcohol content, and user responses were measured objectively (users’ predictive accuracy) and subjectively (users’ satisfaction and trust judgments). Study 1 (N=127) showed that users understand explanations referring to categorical features more readily than those referring to continuous features. It also discovered a dissociation between objective and subjective measures: counterfactual explanations elicited higher accuracy of predictions than no-explanation control descriptions but no higher accuracy than causal explanations, yet counterfactual explanations elicited greater satisfaction and trust judgments than causal explanations. Study 2 (N=211) found that users were more accurate for categorically-transformed features compared to continuous ones, and also replicated the results of Study 1. The findings delineate important boundary conditions for current and future counterfactual explanation methods in XAI.","PeriodicalId":118159,"journal":{"name":"Proceedings of the 28th International Conference on Intelligent User Interfaces","volume":"58 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-03-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128663372","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
TransASL: A Smart Glass based Comprehensive ASL Recognizer in Daily Life TransASL:日常生活中基于智能玻璃的综合ASL识别器
Pub Date : 2023-03-27 DOI: 10.1145/3581641.3584071
Yincheng Jin, Seokmin Choi, Yang Gao, Jiyang Li, Zhengxiong Li, Zhanpeng Jin
Sign language is a primary language used by deaf and hard-of-hearing (DHH) communities. However, existing sign language translation solutions primarily focus on recognizing manual markers. The non-manual markers, such as negative head shaking, question markers, and mouthing, are critical grammatical and semantic components of sign language for better usability and generalizability. Considering the significant role of non-manual markers, we propose the TransASL, a real-time, end-to-end system for sign language recognition and translation. TransASL extracts feature from both manual markers and non-manual markers via a customized eyeglasses-style wearable device with two parallel sensing modalities. Manual marker information is collected by two pairs of outward-facing microphones and speakers mounted to the legs of the eyeglasses. In contrast, non-manual marker information is acquired from a pair of inward-facing microphones and speakers connected to the eyeglasses. Both manual and non-manual marker features undergo a multi-modal, multi-channel fusion network and are eventually recognized as comprehensible ASL content. We evaluate the recognition performance of various sign language expressions at both the word and sentence levels. Given 80 frequently used ASL words and 40 meaningful sentences consisting of manual and non-manual markers, TransASL can achieve the WER of 8.3% and 7.1%, respectively. Our proposed work reveals a great potential for convenient ASL recognition in daily communications between ASL signers and hearing people.
手语是聋人和听力障碍者(DHH)社区使用的主要语言。然而,现有的手语翻译解决方案主要侧重于识别手动标记。非手动标记,如消极摇头、问号和口型,是手语中重要的语法和语义成分,有助于提高可用性和泛化性。考虑到非手动标记的重要作用,我们提出了TransASL,一个实时的端到端手语识别和翻译系统。TransASL通过定制的眼镜式可穿戴设备从手动和非手动标记中提取特征,该设备具有两种并行传感模式。手动标记信息通过安装在眼镜腿上的两对向外的麦克风和扬声器收集。相比之下,非手动标记信息是从连接到眼镜的一对内向麦克风和扬声器中获取的。手动和非手动标记特征都经过多模态、多通道的融合网络,最终被识别为可理解的美国手语内容。我们在单词和句子两个层面上评估了各种手语表达的识别性能。在给定80个常用的ASL单词和40个由手动和非手动标记组成的有意义的句子的情况下,TransASL分别可以实现8.3%和7.1%的WER。我们的研究表明,在手语使用者和听障人士之间的日常交流中,手语识别具有很大的潜力。
{"title":"TransASL: A Smart Glass based Comprehensive ASL Recognizer in Daily Life","authors":"Yincheng Jin, Seokmin Choi, Yang Gao, Jiyang Li, Zhengxiong Li, Zhanpeng Jin","doi":"10.1145/3581641.3584071","DOIUrl":"https://doi.org/10.1145/3581641.3584071","url":null,"abstract":"Sign language is a primary language used by deaf and hard-of-hearing (DHH) communities. However, existing sign language translation solutions primarily focus on recognizing manual markers. The non-manual markers, such as negative head shaking, question markers, and mouthing, are critical grammatical and semantic components of sign language for better usability and generalizability. Considering the significant role of non-manual markers, we propose the TransASL, a real-time, end-to-end system for sign language recognition and translation. TransASL extracts feature from both manual markers and non-manual markers via a customized eyeglasses-style wearable device with two parallel sensing modalities. Manual marker information is collected by two pairs of outward-facing microphones and speakers mounted to the legs of the eyeglasses. In contrast, non-manual marker information is acquired from a pair of inward-facing microphones and speakers connected to the eyeglasses. Both manual and non-manual marker features undergo a multi-modal, multi-channel fusion network and are eventually recognized as comprehensible ASL content. We evaluate the recognition performance of various sign language expressions at both the word and sentence levels. Given 80 frequently used ASL words and 40 meaningful sentences consisting of manual and non-manual markers, TransASL can achieve the WER of 8.3% and 7.1%, respectively. Our proposed work reveals a great potential for convenient ASL recognition in daily communications between ASL signers and hearing people.","PeriodicalId":118159,"journal":{"name":"Proceedings of the 28th International Conference on Intelligent User Interfaces","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-03-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131668538","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
SlideSpecs: Automatic and Interactive Presentation Feedback Collation 幻灯片:自动和交互式演示反馈整理
Pub Date : 2023-03-27 DOI: 10.1145/3581641.3584035
Jeremy Warner, Amy Pavel, Tonya Nguyen, Maneesh Agrawala, Bjoern Hartmann
Presenters often collect audience feedback through practice talks to refine their presentations. In formative interviews, we find that although text feedback and verbal discussions allow presenters to receive feedback, organizing that feedback into actionable presentation revisions remains challenging. Feedback may lack context, be redundant, and be spread across various emails, notes, and conversations. To collate and contextualize both text and verbal feedback, we present SlideSpecs. SlideSpecs lets audience members provide text feedback (e.g., ‘font too small’) while attaching an automatically detected context, including relevant slides (e.g., ‘Slide 7’) or content tags (e.g., ‘slide design’). SlideSpecs also records and transcribes spoken group discussions that commonly occur after practice talks and facilitates linking text critiques to relevant discussion segments. Finally, presenters can use SlideSpecs to review all text and spoken feedback in a single contextually rich interface (e.g., relevant slides, topics, and follow-up discussions). We demonstrate the effectiveness of SlideSpecs by deploying it in eight practice talks with a range of topics and purposes and reporting our findings.
演讲者经常通过练习演讲来收集听众的反馈,以完善他们的演讲。在形成性访谈中,我们发现,尽管文本反馈和口头讨论允许演示者接收反馈,但将这些反馈组织成可操作的演示文稿修订仍然具有挑战性。反馈可能缺乏上下文,是多余的,并且分散在各种电子邮件、笔记和对话中。为了对文本和口头反馈进行整理和语境化,我们提供幻灯片。SlideSpecs允许观众提供文本反馈(例如,“字体太小”),同时附加一个自动检测的上下文,包括相关幻灯片(例如,“幻灯片7”)或内容标签(例如,“幻灯片设计”)。slidesecs还记录和转录了通常在练习演讲后发生的口头小组讨论,并促进了将文本评论链接到相关讨论部分。最后,演示者可以使用slidesecs在一个上下文丰富的界面(例如,相关的幻灯片、主题和后续讨论)中查看所有文本和语音反馈。我们通过在一系列主题和目的的八次实践会谈中部署幻灯片演示了它的有效性,并报告了我们的发现。
{"title":"SlideSpecs: Automatic and Interactive Presentation Feedback Collation","authors":"Jeremy Warner, Amy Pavel, Tonya Nguyen, Maneesh Agrawala, Bjoern Hartmann","doi":"10.1145/3581641.3584035","DOIUrl":"https://doi.org/10.1145/3581641.3584035","url":null,"abstract":"Presenters often collect audience feedback through practice talks to refine their presentations. In formative interviews, we find that although text feedback and verbal discussions allow presenters to receive feedback, organizing that feedback into actionable presentation revisions remains challenging. Feedback may lack context, be redundant, and be spread across various emails, notes, and conversations. To collate and contextualize both text and verbal feedback, we present SlideSpecs. SlideSpecs lets audience members provide text feedback (e.g., ‘font too small’) while attaching an automatically detected context, including relevant slides (e.g., ‘Slide 7’) or content tags (e.g., ‘slide design’). SlideSpecs also records and transcribes spoken group discussions that commonly occur after practice talks and facilitates linking text critiques to relevant discussion segments. Finally, presenters can use SlideSpecs to review all text and spoken feedback in a single contextually rich interface (e.g., relevant slides, topics, and follow-up discussions). We demonstrate the effectiveness of SlideSpecs by deploying it in eight practice talks with a range of topics and purposes and reporting our findings.","PeriodicalId":118159,"journal":{"name":"Proceedings of the 28th International Conference on Intelligent User Interfaces","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-03-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133959713","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
期刊
Proceedings of the 28th International Conference on Intelligent User Interfaces
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1