首页 > 最新文献

Proceedings of the 2nd ACM Multimedia Workshop on Multimodal Conversational AI最新文献

英文 中文
Towards Enriching Responses with Crowd-sourced Knowledge for Task-oriented Dialogue 面向任务的对话,用众包知识丰富回应
Pub Date : 2021-11-17 DOI: 10.1145/3475959.3485392
Ying He, Lizi Liao, Zheng Zhang, Tat-Seng Chua
Task-oriented dialogue agents are built to assist users in completing various tasks. Generating appropriate responses for satisfactory task completion is the ultimate goal. Hence, as a convenient and straightforward way, metrics such as success rate, inform rate etc., have been widely leveraged to evaluate the generated responses. However, beyond task completion, there are several other factors that largely affect user satisfaction, which remain under-explored. In this work, we focus on analyzing different agent behavior patterns that lead to higher user satisfaction scores. Based on the findings, we design a neural response generation model EnRG. It naturally combines the power of pre-trained GPT-2 in response semantic modeling and the merit of dual attention in making use of the external crowd-sourced knowledge. Equipped with two gates via explicit dialogue act modeling, it effectively controls the usage of external knowledge sources in the form of both text and image. We conduct extensive experiments. Both automatic and human evaluation results demonstrate that, beyond comparable task completion, our proposed method manages to generate responses gaining higher user satisfaction.
构建面向任务的对话代理是为了帮助用户完成各种任务。为满意地完成任务而产生适当的响应是最终目标。因此,作为一种方便和直接的方法,诸如成功率、通知率等指标已被广泛用于评估生成的响应。然而,除了任务完成之外,还有其他几个因素在很大程度上影响用户满意度,这些因素仍未得到充分探讨。在这项工作中,我们专注于分析不同的代理行为模式,从而获得更高的用户满意度分数。在此基础上,设计了神经反应生成模型EnRG。它自然地结合了预先训练的GPT-2在响应语义建模方面的能力和利用外部众包知识的双重关注的优点。通过显式对话行为建模设置两扇门,有效控制文本和图像两种形式的外部知识来源的使用。我们进行广泛的实验。自动和人工评估结果都表明,除了可比的任务完成情况外,我们提出的方法能够生成获得更高用户满意度的响应。
{"title":"Towards Enriching Responses with Crowd-sourced Knowledge for Task-oriented Dialogue","authors":"Ying He, Lizi Liao, Zheng Zhang, Tat-Seng Chua","doi":"10.1145/3475959.3485392","DOIUrl":"https://doi.org/10.1145/3475959.3485392","url":null,"abstract":"Task-oriented dialogue agents are built to assist users in completing various tasks. Generating appropriate responses for satisfactory task completion is the ultimate goal. Hence, as a convenient and straightforward way, metrics such as success rate, inform rate etc., have been widely leveraged to evaluate the generated responses. However, beyond task completion, there are several other factors that largely affect user satisfaction, which remain under-explored. In this work, we focus on analyzing different agent behavior patterns that lead to higher user satisfaction scores. Based on the findings, we design a neural response generation model EnRG. It naturally combines the power of pre-trained GPT-2 in response semantic modeling and the merit of dual attention in making use of the external crowd-sourced knowledge. Equipped with two gates via explicit dialogue act modeling, it effectively controls the usage of external knowledge sources in the form of both text and image. We conduct extensive experiments. Both automatic and human evaluation results demonstrate that, beyond comparable task completion, our proposed method manages to generate responses gaining higher user satisfaction.","PeriodicalId":346594,"journal":{"name":"Proceedings of the 2nd ACM Multimedia Workshop on Multimodal Conversational AI","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-11-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116802958","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
The Design of a Trust-based Game as a Conversational Component of Interactive Environment for a Human-agent Negotiation 基于信任的博弈作为人机协商交互环境中的会话组件的设计
Pub Date : 2021-11-17 DOI: 10.1145/3475959.3485393
Andrey V. Vlasov, O. Zinchenko, Zhenjie Zhao, Mansur Bakaev, Arsenjy Karavaev
This research shed the light on how humans interact with virtual partners (Figure 1; 3D view: https://p3d.in/bVJpq) in an interactive environment based on economic games and how this environment can be applied to the training process with immersive technologies. The designed system could be integrated as a tool and be a component of an e-learning platform with Conversational AI and human-agent-interactions which allows human users to play and learn. Scientifically, we have considered the trust problem from a different point of view - learning by doing (i.e., gaming), and proposed that individuals can wear "trust care" lenses on trained "golden eyes" while communicating with others. We explore how contextual trust can be used to promote any human-agent collaboration even in the domain of a competitive negotiation scenario. We present small-scale online testing via instant messaging in Telegram [@trudicbot] and prepare VR testing to demonstrate the potentials of the trust- based game approach.
这项研究揭示了人类如何与虚拟伙伴互动(图1;3D视图:https://p3d.in/bVJpq)在基于经济游戏的互动环境中,以及如何将这种环境应用于沉浸式技术的培训过程。设计的系统可以作为一个工具集成,并成为电子学习平台的一个组成部分,具有会话人工智能和人类代理交互,允许人类用户玩和学习。从科学的角度来看,我们从另一个角度考虑了信任问题——在实践中学习(即游戏),并提出个人在与他人交流时可以在训练有素的“金眼睛”上戴上“信任护理”镜片。我们探索如何使用上下文信任来促进任何人类-代理协作,即使是在竞争性谈判场景的领域。我们通过Telegram [@trudicbot]中的即时通讯进行小规模在线测试,并准备VR测试来展示基于信任的游戏方法的潜力。
{"title":"The Design of a Trust-based Game as a Conversational Component of Interactive Environment for a Human-agent Negotiation","authors":"Andrey V. Vlasov, O. Zinchenko, Zhenjie Zhao, Mansur Bakaev, Arsenjy Karavaev","doi":"10.1145/3475959.3485393","DOIUrl":"https://doi.org/10.1145/3475959.3485393","url":null,"abstract":"This research shed the light on how humans interact with virtual partners (Figure 1; 3D view: https://p3d.in/bVJpq) in an interactive environment based on economic games and how this environment can be applied to the training process with immersive technologies. The designed system could be integrated as a tool and be a component of an e-learning platform with Conversational AI and human-agent-interactions which allows human users to play and learn. Scientifically, we have considered the trust problem from a different point of view - learning by doing (i.e., gaming), and proposed that individuals can wear \"trust care\" lenses on trained \"golden eyes\" while communicating with others. We explore how contextual trust can be used to promote any human-agent collaboration even in the domain of a competitive negotiation scenario. We present small-scale online testing via instant messaging in Telegram [@trudicbot] and prepare VR testing to demonstrate the potentials of the trust- based game approach.","PeriodicalId":346594,"journal":{"name":"Proceedings of the 2nd ACM Multimedia Workshop on Multimodal Conversational AI","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-11-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124502171","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Towards a Real-time Measure of the Perception of Anthropomorphism in Human-robot Interaction 对人机交互中拟人化感知的实时测量
Pub Date : 2021-11-17 DOI: 10.1145/3475959.3485394
Maria Tsfasman, Avinash Saravanan, Dekel Viner, Daan Goslinga, Sarah de Wolf, Chirag Raman, C. Jonker, Catharine Oertel
How human-like do conversational robots need to look to enable long-term human-robot conversation? One essential aspect of long-term interaction is a human's ability to adapt to the varying degrees of a conversational partner's engagement and emotions. Prosodically, this can be achieved through (dis)entrainment. While speech-synthesis has been a limiting factor for many years, restrictions in this regard are increasingly mitigated. These advancements now emphasise the importance of studying the effect of robot embodiment on human entrainment. In this study, we conducted a between-subjects online human-robot interaction experiment in an educational use-case scenario where a tutor was either embodied through a human or a robot face. 43 English-speaking participants took part in the study for whom we analysed the degree of acoustic-prosodic entrainment to the human or robot face, respectively. We found that the degree of subjective and objective perception of anthropomorphism positively correlates with acoustic-prosodic entrainment.
对话机器人需要多像人类才能实现长期的人机对话?长期互动的一个重要方面是人类适应对话伙伴不同程度的参与和情绪的能力。在韵律上,这可以通过(非)娱乐来实现。虽然语音合成多年来一直是一个限制因素,但这方面的限制越来越少。这些进步现在强调了研究机器人化身对人类娱乐的影响的重要性。在这项研究中,我们在一个教育用例场景中进行了一个受试者之间的在线人机交互实验,其中导师要么通过人的脸体现,要么通过机器人的脸体现。43名说英语的参与者参加了这项研究,我们分别分析了他们对人类或机器人面部的声音韵律的影响程度。我们发现,拟人化的主观和客观感知程度与声韵律娱乐呈正相关。
{"title":"Towards a Real-time Measure of the Perception of Anthropomorphism in Human-robot Interaction","authors":"Maria Tsfasman, Avinash Saravanan, Dekel Viner, Daan Goslinga, Sarah de Wolf, Chirag Raman, C. Jonker, Catharine Oertel","doi":"10.1145/3475959.3485394","DOIUrl":"https://doi.org/10.1145/3475959.3485394","url":null,"abstract":"How human-like do conversational robots need to look to enable long-term human-robot conversation? One essential aspect of long-term interaction is a human's ability to adapt to the varying degrees of a conversational partner's engagement and emotions. Prosodically, this can be achieved through (dis)entrainment. While speech-synthesis has been a limiting factor for many years, restrictions in this regard are increasingly mitigated. These advancements now emphasise the importance of studying the effect of robot embodiment on human entrainment. In this study, we conducted a between-subjects online human-robot interaction experiment in an educational use-case scenario where a tutor was either embodied through a human or a robot face. 43 English-speaking participants took part in the study for whom we analysed the degree of acoustic-prosodic entrainment to the human or robot face, respectively. We found that the degree of subjective and objective perception of anthropomorphism positively correlates with acoustic-prosodic entrainment.","PeriodicalId":346594,"journal":{"name":"Proceedings of the 2nd ACM Multimedia Workshop on Multimodal Conversational AI","volume":"134 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-11-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133865334","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
iFetch iFetch
Pub Date : 2021-11-17 DOI: 10.1145/3475959.3485395
R. Sousa, Pedro Ferreira, P. Costa, Pedro Azevedo, João Costeira, Carlos Santiago, João Magalhães, David Semedo, Rafael Ferreira, Alexander I. Rudnicky, Alexander Hauptmann
Most of the interaction between large organizations and their users will be mediated by AI agents in the near future. This perception is becoming undisputed as online shopping dominates entire market segments, and the new "digitally-native" generations become consumers. iFetch is a new generation of task-oriented conversational agents that interact with users seamlessly using verbal and visual information. Through the conversation, iFetch provides targeted advice and a "physical store-like" experience while maintaining user engagement. This context entails the following vital components: 1) highly complex memory models that keep track of the conversation, 2) extraction of key semantic features from language and images that reveal user intent, 3) generation of multimodal responses that will keep users engaged in the conversation and 4) an interrelated knowledge base of products from which to extract relevant product lists.
{"title":"iFetch","authors":"R. Sousa, Pedro Ferreira, P. Costa, Pedro Azevedo, João Costeira, Carlos Santiago, João Magalhães, David Semedo, Rafael Ferreira, Alexander I. Rudnicky, Alexander Hauptmann","doi":"10.1145/3475959.3485395","DOIUrl":"https://doi.org/10.1145/3475959.3485395","url":null,"abstract":"Most of the interaction between large organizations and their users will be mediated by AI agents in the near future. This perception is becoming undisputed as online shopping dominates entire market segments, and the new \"digitally-native\" generations become consumers. iFetch is a new generation of task-oriented conversational agents that interact with users seamlessly using verbal and visual information. Through the conversation, iFetch provides targeted advice and a \"physical store-like\" experience while maintaining user engagement. This context entails the following vital components: 1) highly complex memory models that keep track of the conversation, 2) extraction of key semantic features from language and images that reveal user intent, 3) generation of multimodal responses that will keep users engaged in the conversation and 4) an interrelated knowledge base of products from which to extract relevant product lists.","PeriodicalId":346594,"journal":{"name":"Proceedings of the 2nd ACM Multimedia Workshop on Multimodal Conversational AI","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-11-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128009632","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Conversational AI Efforts within Facebook AI Applied Research Facebook AI应用研究中的会话AI努力
Pub Date : 2021-11-17 DOI: 10.1145/3475959.3478678
A. Geramifard
The goal of the conversational AI team at Facebook AI Applied Research team is to create AI driven dialog capabilities with the augmented/virtual reality product focus. This talk provides an overview of our recent efforts on data collection, multimodal dialog, pipelined model-based policies and end-to-end architectures.
Facebook AI应用研究团队的对话AI团队的目标是创建以增强/虚拟现实产品为重点的AI驱动对话功能。本次演讲概述了我们最近在数据收集、多模式对话、基于模型的流水线策略和端到端架构方面的努力。
{"title":"Conversational AI Efforts within Facebook AI Applied Research","authors":"A. Geramifard","doi":"10.1145/3475959.3478678","DOIUrl":"https://doi.org/10.1145/3475959.3478678","url":null,"abstract":"The goal of the conversational AI team at Facebook AI Applied Research team is to create AI driven dialog capabilities with the augmented/virtual reality product focus. This talk provides an overview of our recent efforts on data collection, multimodal dialog, pipelined model-based policies and end-to-end architectures.","PeriodicalId":346594,"journal":{"name":"Proceedings of the 2nd ACM Multimedia Workshop on Multimodal Conversational AI","volume":"2014 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-11-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132310985","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Proceedings of the 2nd ACM Multimedia Workshop on Multimodal Conversational AI 第二届ACM多模态会话AI多媒体研讨会论文集
{"title":"Proceedings of the 2nd ACM Multimedia Workshop on Multimodal Conversational AI","authors":"","doi":"10.1145/3475959","DOIUrl":"https://doi.org/10.1145/3475959","url":null,"abstract":"","PeriodicalId":346594,"journal":{"name":"Proceedings of the 2nd ACM Multimedia Workshop on Multimodal Conversational AI","volume":"73 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-11-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130069889","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Proceedings of the 2nd ACM Multimedia Workshop on Multimodal Conversational AI
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1