首页 > 最新文献

ICMI-MLMI '10最新文献

英文 中文
Vlogcast yourself: nonverbal behavior and attention in social media Vlogcast自己:社交媒体中的非语言行为和注意力
Pub Date : 2010-11-08 DOI: 10.1145/1891903.1891964
Joan-Isaac Biel, D. Gática-Pérez
We introduce vlogs as a type of rich human interaction which is multimodal in nature and suitable for new large-scale behavioral data analysis. The automatic analysis of vlogs is useful not only to study social media, but also remote communication scenarios, and requires the integration of methods for multimodal processing and for social media understanding. Based on works from social psychology and computing, we first propose robust audio, visual, and multimodal cues to measure the nonverbal behavior of vloggers in their videos. Then, we investigate the relation between behavior and the attention videos receive in YouTube. Our study shows significant correlations between some nonverbal behavioral cues and the average number of views per video.
我们将视频日志作为一种丰富的人类交互方式,它具有多模态的性质,适合于新的大规模行为数据分析。vlogs的自动分析不仅适用于社交媒体研究,也适用于远程传播场景,需要将多模态处理方法和社交媒体理解方法相结合。基于社会心理学和计算机的工作,我们首先提出了健壮的音频、视觉和多模态线索来衡量视频博主在其视频中的非语言行为。然后,我们研究了行为与视频在YouTube上获得的关注之间的关系。我们的研究表明,一些非语言行为线索与每个视频的平均观看次数之间存在显著的相关性。
{"title":"Vlogcast yourself: nonverbal behavior and attention in social media","authors":"Joan-Isaac Biel, D. Gática-Pérez","doi":"10.1145/1891903.1891964","DOIUrl":"https://doi.org/10.1145/1891903.1891964","url":null,"abstract":"We introduce vlogs as a type of rich human interaction which is multimodal in nature and suitable for new large-scale behavioral data analysis. The automatic analysis of vlogs is useful not only to study social media, but also remote communication scenarios, and requires the integration of methods for multimodal processing and for social media understanding. Based on works from social psychology and computing, we first propose robust audio, visual, and multimodal cues to measure the nonverbal behavior of vloggers in their videos. Then, we investigate the relation between behavior and the attention videos receive in YouTube. Our study shows significant correlations between some nonverbal behavioral cues and the average number of views per video.","PeriodicalId":181145,"journal":{"name":"ICMI-MLMI '10","volume":"176 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-11-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116506360","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 18
Real-time adaptive behaviors in multimodal human-avatar interactions 多模态人机交互中的实时自适应行为
Pub Date : 2010-11-08 DOI: 10.1145/1891903.1891909
Hui Zhang, Damian Fricker, Thomas G. Smith, Chen Yu
Multimodal interaction in everyday life seems so effortless. However, a closer look reveals that such interaction is indeed complex and comprises multiple levels of coordination, from high-level linguistic exchanges to low-level couplings of momentary bodily movements both within an agent and across multiple interacting agents. A better understanding of how these multimodal behaviors are coordinated can provide insightful principles to guide the development of intelligent multimodal interfaces. In light of this, we propose and implement a research framework in which human participants interact with a virtual agent in a virtual environment. Our platform allows the virtual agent to keep track of the user's gaze and hand movements in real time, and adjust his own behaviors accordingly. An experiment is designed and conducted to investigate adaptive user behaviors in a human-agent joint attention task. Multimodal data streams are collected in the study including speech, eye gaze, hand and head movements from both the human user and the virtual agent, which are then analyzed to discover various behavioral patterns. Those patterns show that human participants are highly sensitive to momentary multimodal behaviors generated by the virtual agent and they rapidly adapt their behaviors accordingly. Our results suggest the importance of studying and understanding real-time adaptive behaviors in human-computer multimodal interactions.
日常生活中的多模式互动似乎毫不费力。然而,仔细观察就会发现,这种互动确实很复杂,包括多个层面的协调,从一个代理内部和多个交互代理之间的高级语言交流到低级的瞬间身体运动耦合。更好地理解这些多模态行为是如何协调的,可以为指导智能多模态接口的开发提供有洞察力的原则。鉴于此,我们提出并实施了一个研究框架,其中人类参与者在虚拟环境中与虚拟代理进行交互。我们的平台允许虚拟代理实时跟踪用户的目光和手部运动,并相应地调整自己的行为。设计并实施了一项实验来研究人机联合注意任务中的自适应用户行为。该研究收集了来自人类用户和虚拟代理的多模态数据流,包括语音、眼睛注视、手和头部运动,然后对这些数据进行分析,以发现各种行为模式。这些模式表明,人类参与者对虚拟代理产生的瞬间多模态行为高度敏感,并能迅速调整自己的行为。我们的研究结果表明,研究和理解实时适应行为在人机多模态交互中的重要性。
{"title":"Real-time adaptive behaviors in multimodal human-avatar interactions","authors":"Hui Zhang, Damian Fricker, Thomas G. Smith, Chen Yu","doi":"10.1145/1891903.1891909","DOIUrl":"https://doi.org/10.1145/1891903.1891909","url":null,"abstract":"Multimodal interaction in everyday life seems so effortless. However, a closer look reveals that such interaction is indeed complex and comprises multiple levels of coordination, from high-level linguistic exchanges to low-level couplings of momentary bodily movements both within an agent and across multiple interacting agents. A better understanding of how these multimodal behaviors are coordinated can provide insightful principles to guide the development of intelligent multimodal interfaces. In light of this, we propose and implement a research framework in which human participants interact with a virtual agent in a virtual environment. Our platform allows the virtual agent to keep track of the user's gaze and hand movements in real time, and adjust his own behaviors accordingly. An experiment is designed and conducted to investigate adaptive user behaviors in a human-agent joint attention task. Multimodal data streams are collected in the study including speech, eye gaze, hand and head movements from both the human user and the virtual agent, which are then analyzed to discover various behavioral patterns. Those patterns show that human participants are highly sensitive to momentary multimodal behaviors generated by the virtual agent and they rapidly adapt their behaviors accordingly. Our results suggest the importance of studying and understanding real-time adaptive behaviors in human-computer multimodal interactions.","PeriodicalId":181145,"journal":{"name":"ICMI-MLMI '10","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-11-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126889034","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 18
Linearity and synchrony: quantitative metrics for slide-based presentation methodology 线性和同步:基于幻灯片的演示方法的定量度量
Pub Date : 2010-11-08 DOI: 10.1145/1891903.1891947
K. Kurihara, T. Mochizuki, Hiroki Oura, Mio Tsubakimoto, T. Nishimori, Jun Nakahara
In this paper we propose new quantitative metrics that express the characteristics of current general practices in slide-based presentation methodology. The proposed metrics are numerical expressions of: 'To what extent are the materials being presented in the prepared order?' and 'What is the degree of separation between the displays of the presenter and the audience?'. Through the use of these metrics, it becomes possible to quantitatively evaluate various extended methods designed to improve presentations. We illustrate examples of calculation and visualization for the proposed metrics.
在本文中,我们提出了新的量化指标,以表达当前基于幻灯片的演示方法的一般实践的特征。建议的度量标准是以下方面的数值表达:“材料在多大程度上按照准备好的顺序呈现?”以及“演示者和观众的展示之间的分离程度是多少?”通过使用这些指标,可以定量地评估旨在改进演示的各种扩展方法。我们举例说明了所建议的度量的计算和可视化。
{"title":"Linearity and synchrony: quantitative metrics for slide-based presentation methodology","authors":"K. Kurihara, T. Mochizuki, Hiroki Oura, Mio Tsubakimoto, T. Nishimori, Jun Nakahara","doi":"10.1145/1891903.1891947","DOIUrl":"https://doi.org/10.1145/1891903.1891947","url":null,"abstract":"In this paper we propose new quantitative metrics that express the characteristics of current general practices in slide-based presentation methodology. The proposed metrics are numerical expressions of: 'To what extent are the materials being presented in the prepared order?' and 'What is the degree of separation between the displays of the presenter and the audience?'. Through the use of these metrics, it becomes possible to quantitatively evaluate various extended methods designed to improve presentations. We illustrate examples of calculation and visualization for the proposed metrics.","PeriodicalId":181145,"journal":{"name":"ICMI-MLMI '10","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-11-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128091469","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Learning and evaluating response prediction models using parallel listener consensus 学习和评估使用平行听众共识的反应预测模型
Pub Date : 2010-11-08 DOI: 10.1145/1891903.1891908
I. D. Kok, Derya Ozkan, D. Heylen, Louis-Philippe Morency
Traditionally listener response prediction models are learned from pre-recorded dyadic interactions. Because of individual differences in behavior, these recordings do not capture the complete ground truth. Where the recorded listener did not respond to an opportunity provided by the speaker, another listener would have responded or vice versa. In this paper, we introduce the concept of parallel listener consensus where the listener responses from multiple parallel interactions are combined to better capture differences and similarities between individuals. We show how parallel listener consensus can be used for both learning and evaluating probabilistic prediction models of listener responses. To improve the learning performance, the parallel consensus helps identifying better negative samples and reduces outliers in the positive samples. We propose a new error measurement called fConsensus which exploits the parallel consensus to better define the concepts of exactness (mislabels) and completeness (missed labels) for prediction models. We present a series of experiments using the MultiLis Corpus where three listeners were tricked into believing that they had a one-on-one conversation with a speaker, while in fact they were recorded in parallel in interaction with the same speaker. In this paper we show that using parallel listener consensus can improve learning performance and represent better evaluation criteria for predictive models.
传统的听众反应预测模型是从预先录制的二元交互中学习的。由于个体行为的差异,这些记录并不能捕捉到完全的真实情况。如果录音的听众没有回应说话人提供的机会,另一个听众就会回应,反之亦然。在本文中,我们引入了平行听者共识的概念,将多个平行互动的听者反应结合起来,以更好地捕捉个体之间的差异和相似之处。我们展示了平行听者共识如何用于学习和评估听者反应的概率预测模型。为了提高学习性能,平行共识有助于识别更好的负样本,并减少正样本中的异常值。我们提出了一种新的误差测量方法,称为fConsensus,它利用并行共识来更好地定义预测模型的准确性(错误标签)和完整性(缺失标签)的概念。我们展示了一系列使用MultiLis语料库的实验,其中三名听众被骗相信他们与说话者进行了一对一的对话,而实际上他们与同一说话者的互动是并行记录的。在本文中,我们表明使用并行听众共识可以提高学习性能,并为预测模型提供更好的评估标准。
{"title":"Learning and evaluating response prediction models using parallel listener consensus","authors":"I. D. Kok, Derya Ozkan, D. Heylen, Louis-Philippe Morency","doi":"10.1145/1891903.1891908","DOIUrl":"https://doi.org/10.1145/1891903.1891908","url":null,"abstract":"Traditionally listener response prediction models are learned from pre-recorded dyadic interactions. Because of individual differences in behavior, these recordings do not capture the complete ground truth. Where the recorded listener did not respond to an opportunity provided by the speaker, another listener would have responded or vice versa. In this paper, we introduce the concept of parallel listener consensus where the listener responses from multiple parallel interactions are combined to better capture differences and similarities between individuals. We show how parallel listener consensus can be used for both learning and evaluating probabilistic prediction models of listener responses. To improve the learning performance, the parallel consensus helps identifying better negative samples and reduces outliers in the positive samples. We propose a new error measurement called fConsensus which exploits the parallel consensus to better define the concepts of exactness (mislabels) and completeness (missed labels) for prediction models. We present a series of experiments using the MultiLis Corpus where three listeners were tricked into believing that they had a one-on-one conversation with a speaker, while in fact they were recorded in parallel in interaction with the same speaker. In this paper we show that using parallel listener consensus can improve learning performance and represent better evaluation criteria for predictive models.","PeriodicalId":181145,"journal":{"name":"ICMI-MLMI '10","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-11-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132931055","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 19
Modelling and analyzing multimodal dyadic interactions using social networks 建模和分析使用社会网络的多模态二元互动
Pub Date : 2010-11-08 DOI: 10.1145/1891903.1891967
Sergio Escalera, P. Radeva, Jordi Vitrià, Xavier Baró, B. Raducanu
Social network analysis became a common technique used to model and quantify the properties of social interactions. In this paper, we propose an integrated framework to explore the characteristics of a social network extracted from multimodal dyadic interactions. First, speech detection is performed through an audio/visual fusion scheme based on stacked sequential learning. In the audio domain, speech is detected through clusterization of audio features. Clusters are modelled by means of an One-state Hidden Markov Model containing a diagonal covariance Gaussian Mixture Model. In the visual domain, speech detection is performed through differential-based feature extraction from the segmented mouth region, and a dynamic programming matching procedure. Second, in order to model the dyadic interactions, we employed the Influence Model whose states encode the previous integrated audio/visual data. Third, the social network is extracted based on the estimated influences. For our study, we used a set of videos belonging to New York Times' Blogging Heads opinion blog. The results are reported both in terms of accuracy of the audio/visual data fusion and centrality measures used to characterize the social network.
社会网络分析成为一种常用的技术,用于建模和量化社会互动的属性。在本文中,我们提出了一个综合框架来探索从多模态二元交互中提取的社会网络的特征。首先,通过基于堆叠顺序学习的视听融合方案进行语音检测。在音频领域,通过音频特征的聚类来检测语音。聚类模型采用含对角协方差高斯混合模型的单态隐马尔可夫模型进行建模。在视觉领域,语音检测是通过基于微分的特征提取,从分割的嘴区域,并动态规划匹配过程。其次,为了对二元交互建模,我们采用了影响模型,其状态编码了先前的集成音频/视觉数据。第三,根据估计的影响提取社会网络。在我们的研究中,我们使用了一组属于《纽约时报》博客观点博客的视频。结果报告了音频/视觉数据融合的准确性和用于表征社交网络的中心性度量。
{"title":"Modelling and analyzing multimodal dyadic interactions using social networks","authors":"Sergio Escalera, P. Radeva, Jordi Vitrià, Xavier Baró, B. Raducanu","doi":"10.1145/1891903.1891967","DOIUrl":"https://doi.org/10.1145/1891903.1891967","url":null,"abstract":"Social network analysis became a common technique used to model and quantify the properties of social interactions. In this paper, we propose an integrated framework to explore the characteristics of a social network extracted from multimodal dyadic interactions. First, speech detection is performed through an audio/visual fusion scheme based on stacked sequential learning. In the audio domain, speech is detected through clusterization of audio features. Clusters are modelled by means of an One-state Hidden Markov Model containing a diagonal covariance Gaussian Mixture Model. In the visual domain, speech detection is performed through differential-based feature extraction from the segmented mouth region, and a dynamic programming matching procedure. Second, in order to model the dyadic interactions, we employed the Influence Model whose states encode the previous integrated audio/visual data. Third, the social network is extracted based on the estimated influences. For our study, we used a set of videos belonging to New York Times' Blogging Heads opinion blog. The results are reported both in terms of accuracy of the audio/visual data fusion and centrality measures used to characterize the social network.","PeriodicalId":181145,"journal":{"name":"ICMI-MLMI '10","volume":"53 1-2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-11-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116590650","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Speak4it: multimodal interaction for local search Speak4it:本地搜索的多模式交互
Pub Date : 2010-11-08 DOI: 10.1145/1891903.1891917
Patrick Ehlen, Michael Johnston
Speak4itSM is a consumer-oriented mobile search application that leverages multimodal input and output to allow users to search for and act on local business information. It supports true multimodal integration where user inputs can be distributed over multiple input modes. In addition to specifying queries by voice (e.g., "bike repair shops near the golden gate bridge") users can combine speech and gesture. For example, "gas stations" + will return the gas stations along the specified route traced on the display. We provide interactive demonstrations of Speak4it on both the iPhone and iPad platforms and explain the underlying multimodal architecture and challenges of supporting multimodal interaction as a deployed mobile service.
Speak4itSM是一个面向消费者的移动搜索应用程序,它利用多模态输入和输出,允许用户搜索和处理本地商业信息。它支持真正的多模式集成,用户输入可以分布在多个输入模式上。除了通过语音指定查询(例如,“金门大桥附近的自行车修理店”)之外,用户还可以将语音和手势结合起来。例如,“gas stations”+将返回沿着显示器上跟踪的指定路线的加油站。我们提供了Speak4it在iPhone和iPad平台上的交互式演示,并解释了底层的多模式架构和作为部署的移动服务支持多模式交互的挑战。
{"title":"Speak4it: multimodal interaction for local search","authors":"Patrick Ehlen, Michael Johnston","doi":"10.1145/1891903.1891917","DOIUrl":"https://doi.org/10.1145/1891903.1891917","url":null,"abstract":"Speak4itSM is a consumer-oriented mobile search application that leverages multimodal input and output to allow users to search for and act on local business information. It supports true multimodal integration where user inputs can be distributed over multiple input modes. In addition to specifying queries by voice (e.g., \"bike repair shops near the golden gate bridge\") users can combine speech and gesture. For example, \"gas stations\" + <route drawn on display> will return the gas stations along the specified route traced on the display. We provide interactive demonstrations of Speak4it on both the iPhone and iPad platforms and explain the underlying multimodal architecture and challenges of supporting multimodal interaction as a deployed mobile service.","PeriodicalId":181145,"journal":{"name":"ICMI-MLMI '10","volume":"29 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-11-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121982464","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Does haptic feedback change the way we view touchscreens in cars? 触觉反馈是否会改变我们看待汽车触摸屏的方式?
Pub Date : 2010-11-08 DOI: 10.1145/1891903.1891952
Matthew J. Pitts, G. Burnett, M. Williams, Tom Wellings
Touchscreens are increasingly being used in mobile devices and in-vehicle systems. While the usability benefits of touchscreens are acknowledged, their use places significant visual demand on the user due to the lack of tactile and kinaesthetic feedback. Haptic feedback is shown to improve performance in mobile devices, but little objective data is available regarding touchscreen feedback in an automotive scenario. A study was conducted to investigate the effects of visual and haptic touchscreen feedback on driver visual behaviour and driving performance using a simulated driving environment. Results showed a significant interaction between visual and haptic feedback, with the presence of haptic feedback compensating for changes in visual feedback. Driving performance was unaffected by feedback condition but degraded from a baseline measure when touchscreen tasks were introduced. Subjective responses indicated an improved user experience and increased confidence when haptic feedback was enabled.
触摸屏越来越多地用于移动设备和车载系统。虽然触摸屏的可用性好处是公认的,但由于缺乏触觉和动觉反馈,它们的使用对用户的视觉需求很大。触觉反馈被证明可以提高移动设备的性能,但关于汽车场景中触摸屏反馈的客观数据很少。通过模拟驾驶环境,研究了视觉和触觉触摸屏反馈对驾驶员视觉行为和驾驶性能的影响。结果显示,视觉和触觉反馈之间存在显著的相互作用,触觉反馈的存在补偿了视觉反馈的变化。驾驶性能不受反馈条件的影响,但当引入触屏任务时,驾驶性能从基线测量下降。主观反应表明,当触觉反馈被启用时,用户体验得到改善,信心增加。
{"title":"Does haptic feedback change the way we view touchscreens in cars?","authors":"Matthew J. Pitts, G. Burnett, M. Williams, Tom Wellings","doi":"10.1145/1891903.1891952","DOIUrl":"https://doi.org/10.1145/1891903.1891952","url":null,"abstract":"Touchscreens are increasingly being used in mobile devices and in-vehicle systems. While the usability benefits of touchscreens are acknowledged, their use places significant visual demand on the user due to the lack of tactile and kinaesthetic feedback. Haptic feedback is shown to improve performance in mobile devices, but little objective data is available regarding touchscreen feedback in an automotive scenario. A study was conducted to investigate the effects of visual and haptic touchscreen feedback on driver visual behaviour and driving performance using a simulated driving environment. Results showed a significant interaction between visual and haptic feedback, with the presence of haptic feedback compensating for changes in visual feedback. Driving performance was unaffected by feedback condition but degraded from a baseline measure when touchscreen tasks were introduced. Subjective responses indicated an improved user experience and increased confidence when haptic feedback was enabled.","PeriodicalId":181145,"journal":{"name":"ICMI-MLMI '10","volume":"132 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-11-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133216007","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 16
Behavior and preference in minimal personality: a study on embodied conversational agents 最小人格中的行为与偏好:具身会话主体的研究
Pub Date : 2010-11-08 DOI: 10.1145/1891903.1891963
Yuting Chen, A. Naveed, R. Porzel
Endowing embodied conversational agent with personality affords more natural modalities for their interaction with human interlocutors. To bridge the personality gap between users and agents, we designed minimal two personalities for corresponding agents i.e. an introverted and an extroverted agent. Each features a combination of different verbal and non-verbal behaviors. In this paper, we present an examination of the effects of the speaking and behavior styles of the two agents and explore the resulting design factors pertinent for spoken dialogue systems. The results indicate that users prefer the extroverted agent to the introverted one. The personality traits of the agents influence the users' preferences, dialogues, and behavior. Statistically, it is highly significant that users are more talkative with the extroverted agent. We also investigate the spontaneous speech disfluency of the dialogues and demonstrate that the extroverted behavior model reduce the user's speech disfluency. Furthermore, users having different mental models behave differently with the agents. The results and findings show that the minimal personalities of agents maximally influence the interlocutors' behaviors.
赋予具身会话代理人格可以使其与人类对话者的互动更加自然。为了弥合用户和代理之间的性格差异,我们为相应的代理设计了最小的两种性格,即内向和外向的代理。每种语言都结合了不同的语言和非语言行为。在本文中,我们对两个智能体的说话和行为风格的影响进行了研究,并探讨了与口语对话系统相关的设计因素。结果表明,用户对外向型代理的偏好高于内向型代理。代理的人格特征影响用户的偏好、对话和行为。统计上,用户与外向的代理更健谈,这是非常显著的。我们还研究了对话中的自发语言不流畅性,并证明了外向行为模型减少了用户的语言不流畅性。此外,具有不同心智模型的用户对代理的行为也不同。研究结果表明,代理人的最小人格对对话者的行为影响最大。
{"title":"Behavior and preference in minimal personality: a study on embodied conversational agents","authors":"Yuting Chen, A. Naveed, R. Porzel","doi":"10.1145/1891903.1891963","DOIUrl":"https://doi.org/10.1145/1891903.1891963","url":null,"abstract":"Endowing embodied conversational agent with personality affords more natural modalities for their interaction with human interlocutors. To bridge the personality gap between users and agents, we designed minimal two personalities for corresponding agents i.e. an introverted and an extroverted agent. Each features a combination of different verbal and non-verbal behaviors. In this paper, we present an examination of the effects of the speaking and behavior styles of the two agents and explore the resulting design factors pertinent for spoken dialogue systems. The results indicate that users prefer the extroverted agent to the introverted one. The personality traits of the agents influence the users' preferences, dialogues, and behavior. Statistically, it is highly significant that users are more talkative with the extroverted agent. We also investigate the spontaneous speech disfluency of the dialogues and demonstrate that the extroverted behavior model reduce the user's speech disfluency. Furthermore, users having different mental models behave differently with the agents. The results and findings show that the minimal personalities of agents maximally influence the interlocutors' behaviors.","PeriodicalId":181145,"journal":{"name":"ICMI-MLMI '10","volume":"17 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-11-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133590184","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 17
Enhancing browsing experience of table and image elements in web pages 增强网页中表格和图像元素的浏览体验
Pub Date : 2010-11-08 DOI: 10.1145/1891903.1891935
Wenchang Xu, Xin Yang, Yuanchun Shi
As the popularity and diversification of both Internet and its access devices, users' browsing experience of web pages is in great need of improvement. Traditional browsing mode of web elements such as table and image is passive, which limits users' browsing efficiency of web pages. In this paper, we propose to enhance browsing experience of table and image elements in web pages by enabling real-time interactive access to web tables and images. We design new browsing modes that help users improve their browsing efficiency including operation mode, record mode for web tables and normal mode, starred mode, advanced mode for web images. We design and implement a plug-in for Microsoft Internet Explorer, called iWebWidget, which provides a customized user interface supporting real-time interactive access to web tables and images. Besides, we carry out a user study to testify the usefulness of iWebWidget. Experimental results show that users are satisfied and really enjoy the new browsing modes for both web tables and images.
随着互联网及其接入设备的普及和多样化,用户的网页浏览体验亟待改善。传统的表格、图片等网页元素的浏览方式是被动的,限制了用户对网页的浏览效率。在本文中,我们建议通过实现对网页表格和图像的实时交互访问来增强网页中表格和图像元素的浏览体验。我们设计了新的浏览模式,帮助用户提高浏览效率,包括对网页表格的操作模式、记录模式和对网页图片的正常模式、星号模式、高级模式。我们为Microsoft Internet Explorer设计并实现了一个名为iWebWidget的插件,它提供了一个定制的用户界面,支持对web表和图像的实时交互访问。此外,我们进行了一个用户研究来证明iWebWidget的有用性。实验结果表明,用户满意并真正喜欢新的网页表格和图片浏览模式。
{"title":"Enhancing browsing experience of table and image elements in web pages","authors":"Wenchang Xu, Xin Yang, Yuanchun Shi","doi":"10.1145/1891903.1891935","DOIUrl":"https://doi.org/10.1145/1891903.1891935","url":null,"abstract":"As the popularity and diversification of both Internet and its access devices, users' browsing experience of web pages is in great need of improvement. Traditional browsing mode of web elements such as table and image is passive, which limits users' browsing efficiency of web pages. In this paper, we propose to enhance browsing experience of table and image elements in web pages by enabling real-time interactive access to web tables and images. We design new browsing modes that help users improve their browsing efficiency including operation mode, record mode for web tables and normal mode, starred mode, advanced mode for web images. We design and implement a plug-in for Microsoft Internet Explorer, called iWebWidget, which provides a customized user interface supporting real-time interactive access to web tables and images. Besides, we carry out a user study to testify the usefulness of iWebWidget. Experimental results show that users are satisfied and really enjoy the new browsing modes for both web tables and images.","PeriodicalId":181145,"journal":{"name":"ICMI-MLMI '10","volume":"439 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-11-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125769860","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Automatic recognition of sign language subwords based on portable accelerometer and EMG sensors 基于便携式加速度计和肌电传感器的手语子词自动识别
Pub Date : 2010-11-08 DOI: 10.1145/1891903.1891926
Yun Li, Xiang Chen, Jianxun Tian, Xu Zhang, Kongqiao Wang, Jihai Yang
Sign language recognition (SLR) not only facilitates the communication between the deaf and hearing society, but also serves as a good basis for the development of gesture-based human-computer interaction (HCI). In this paper, the portable input devices based on accelerometers and surface electromyography (EMG) sensors worn on the forearm are presented, and an effective fusion strategy for combination of multi-sensor and multi-channel information is proposed to automatically recognize sign language at the subword classification level. Experimental results on the recognition of 121 frequently used Chinese sign language subwords demonstrate the feasibility of developing SLR system based on the presented portable input devices and that our proposed information fusion method is effective for automatic SLR. Our study will promote the realization of practical sign language recognizer and multimodal human-computer interfaces.
手语识别不仅方便了聋人与听人社会之间的交流,而且为基于手势的人机交互(HCI)的发展奠定了良好的基础。提出了一种基于加速度计和表面肌电传感器的便携式输入装置,并提出了一种多传感器、多通道信息相结合的有效融合策略,在子词分类层面实现手语自动识别。对121个常用汉语手语子词的识别实验结果表明,基于便携式输入设备开发单反系统是可行的,所提出的信息融合方法是实现自动单反的有效方法。我们的研究将促进实用手语识别器和多模态人机界面的实现。
{"title":"Automatic recognition of sign language subwords based on portable accelerometer and EMG sensors","authors":"Yun Li, Xiang Chen, Jianxun Tian, Xu Zhang, Kongqiao Wang, Jihai Yang","doi":"10.1145/1891903.1891926","DOIUrl":"https://doi.org/10.1145/1891903.1891926","url":null,"abstract":"Sign language recognition (SLR) not only facilitates the communication between the deaf and hearing society, but also serves as a good basis for the development of gesture-based human-computer interaction (HCI). In this paper, the portable input devices based on accelerometers and surface electromyography (EMG) sensors worn on the forearm are presented, and an effective fusion strategy for combination of multi-sensor and multi-channel information is proposed to automatically recognize sign language at the subword classification level. Experimental results on the recognition of 121 frequently used Chinese sign language subwords demonstrate the feasibility of developing SLR system based on the presented portable input devices and that our proposed information fusion method is effective for automatic SLR. Our study will promote the realization of practical sign language recognizer and multimodal human-computer interfaces.","PeriodicalId":181145,"journal":{"name":"ICMI-MLMI '10","volume":"28 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-11-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114464145","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 72
期刊
ICMI-MLMI '10
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1