首页 > 最新文献

GazeIn '14最新文献

英文 中文
Fusing Multimodal Human Expert Data to Uncover Hidden Semantics 融合多模态人类专家数据揭示隐藏语义
Pub Date : 2014-11-16 DOI: 10.1145/2666642.2666649
Xuan Guo, Qi Yu, Rui Li, Cecilia Ovesdotter Alm, Anne R. Haake
Problem solving in complex visual domains involves multiple levels of cognitive processing. Analyzing and representing these cognitive processes requires the elicitation and study of multimodal human data. We have developed methods for extracting experts' visual behaviors and verbal descriptions during medical image inspection. Now we address fusion of these data towards building a novel framework for organizing elements of expertise as a foundation for knowledge-dependent computational systems. In this paper, a multimodal graph-regularized non-negative matrix factorization approach is developed and used to fuse multimodal data collected during medical image inspection. Our experimental results on new data representation demonstrate the effectiveness of the proposed data fusion approach.
复杂视觉域的问题解决涉及多个层次的认知加工。分析和表示这些认知过程需要对多模态人类数据的启发和研究。我们开发了在医学图像检测过程中提取专家视觉行为和语言描述的方法。现在我们解决这些数据的融合,以建立一个新的框架来组织专业知识元素,作为知识依赖计算系统的基础。本文提出了一种多模态图正则化非负矩阵分解方法,用于医学图像检测过程中采集的多模态数据的融合。我们在新的数据表示上的实验结果证明了所提数据融合方法的有效性。
{"title":"Fusing Multimodal Human Expert Data to Uncover Hidden Semantics","authors":"Xuan Guo, Qi Yu, Rui Li, Cecilia Ovesdotter Alm, Anne R. Haake","doi":"10.1145/2666642.2666649","DOIUrl":"https://doi.org/10.1145/2666642.2666649","url":null,"abstract":"Problem solving in complex visual domains involves multiple levels of cognitive processing. Analyzing and representing these cognitive processes requires the elicitation and study of multimodal human data. We have developed methods for extracting experts' visual behaviors and verbal descriptions during medical image inspection. Now we address fusion of these data towards building a novel framework for organizing elements of expertise as a foundation for knowledge-dependent computational systems. In this paper, a multimodal graph-regularized non-negative matrix factorization approach is developed and used to fuse multimodal data collected during medical image inspection. Our experimental results on new data representation demonstrate the effectiveness of the proposed data fusion approach.","PeriodicalId":230150,"journal":{"name":"GazeIn '14","volume":"79 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-11-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134538946","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Analysis of Timing Structure of Eye Contact in Turn-changing 换向时眼神接触的时序结构分析
Pub Date : 2014-11-16 DOI: 10.1145/2666642.2666648
Ryo Ishii, K. Otsuka, Shiro Kumano, Junji Yamato
With the aim of constructing a model for predicting the next speaker and the start of the next utterance in multi-party meetings, we focus on the timing structure of the eye contact between the speaker, the listener, and the next speaker: who looks at whom first, who looks away first, and when the eye contact happens. We analyze the differences in the timing structure for the listener and next speaker in turn-changing and turn-keeping. The results of analysis show that the listeners in turn-keeping tend to look at the speaker more often first before the speaker looks at the listeners than the next speaker in turn-changing looks at the speaker first before the speaker looks at the next speaker when the eye contact with the speaker happens. The listeners in turn-keeping tend to look away from the speaker more often later after the speaker looks away from the listener than listeners and the next speaker in turn-changing looks away from the speaker later when the eye contact with the speaker happens. In addition, the interval between the end of eye contact, the end of the speaker's utterance, and the start of next speaker's utterance is different between the listener in turn-keeping, the listener in turn-changing, and the next speaker in turn-changing.
为了构建一个模型来预测多方会议中下一个发言者和下一个发言的开始,我们重点研究了发言者、听众和下一个发言者之间目光接触的时间结构:谁先看谁,谁先看向另一个发言者,以及目光接触发生的时间。我们分析了听者和下一位说话者在换轮和保持轮时的时间结构差异。分析结果表明,在演讲者看向听众之前,听众往往会先看向演讲者,而不是下一个演讲者,当与演讲者发生目光接触时,听众会在演讲者看向下一个演讲者之前先看向演讲者。在演讲者把目光从听众身上移开之后,听众比听众更倾向于把目光从演讲者身上移开,而下一个轮流的演讲者在与演讲者发生目光接触之后也会把目光从演讲者身上移开。此外,目光接触结束、说话人说话结束、下一个说话人开始说话的间隔时间,在听者保持回合、听者换回合、下一个说话人换回合之间是不同的。
{"title":"Analysis of Timing Structure of Eye Contact in Turn-changing","authors":"Ryo Ishii, K. Otsuka, Shiro Kumano, Junji Yamato","doi":"10.1145/2666642.2666648","DOIUrl":"https://doi.org/10.1145/2666642.2666648","url":null,"abstract":"With the aim of constructing a model for predicting the next speaker and the start of the next utterance in multi-party meetings, we focus on the timing structure of the eye contact between the speaker, the listener, and the next speaker: who looks at whom first, who looks away first, and when the eye contact happens. We analyze the differences in the timing structure for the listener and next speaker in turn-changing and turn-keeping. The results of analysis show that the listeners in turn-keeping tend to look at the speaker more often first before the speaker looks at the listeners than the next speaker in turn-changing looks at the speaker first before the speaker looks at the next speaker when the eye contact with the speaker happens. The listeners in turn-keeping tend to look away from the speaker more often later after the speaker looks away from the listener than listeners and the next speaker in turn-changing looks away from the speaker later when the eye contact with the speaker happens. In addition, the interval between the end of eye contact, the end of the speaker's utterance, and the start of next speaker's utterance is different between the listener in turn-keeping, the listener in turn-changing, and the next speaker in turn-changing.","PeriodicalId":230150,"journal":{"name":"GazeIn '14","volume":"121 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-11-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132225046","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Spatio-Temporal Event Selection in Basic Surveillance Tasks using Eye Tracking and EEG 利用眼动跟踪和脑电图在基本监控任务中进行时空事件选择
Pub Date : 2014-11-16 DOI: 10.1145/2666642.2666645
Jutta Hild, F. Putze, David Kaufman, Christian Kühnle, Tanja Schultz, J. Beyerer
In safety- and security-critical applications like video surveillance it is crucial that human operators detect task-relevant events in the continuous video streams and select them for report or dissemination to other authorities. Usually, the selection operation is performed using a manual input device like a mouse or a joystick. Due to the visually rich and dynamic input, the required high attention, the long working time, and the challenging manual selection of moving objects, it occurs that relevant events are missed. To alleviate this problem we propose adding another event selection process, using eye-brain input. Our approach is based on eye tracking and EEG, providing spatio-temporal event selection without any manual intervention. We report ongoing research, building on prior work where we showed the general feasibility of the approach. In this contribution, we extend our work testing the feasibility of the approach using more advanced and less artificial experimental paradigms simulating frequently occurring, basic types of real surveillance tasks. The paradigms are much closer to a real surveillance task in terms of the used visual stimuli, the more subtle cues for event indication, and the required viewing behavior. As a methodology we perform an experiment (N=10) with non-experts. The results confirm the feasibility of the approach for event selection in the advanced tasks. We achieve spatio-temporal event selection accuracy scores of up to 77% and 60% for different stages of event indication.
在视频监控等安全和安保关键应用中,人工操作员在连续视频流中检测与任务相关的事件并选择它们报告或传播给其他当局是至关重要的。通常,选择操作是使用手动输入设备,如鼠标或操纵杆来执行的。由于视觉上的丰富和动态输入,需要高度关注,工作时间长,以及具有挑战性的手动选择运动对象,因此会出现错过相关事件的情况。为了解决这个问题,我们建议增加另一个事件选择过程,使用眼脑输入。我们的方法是基于眼动追踪和脑电图,提供时空事件选择,而无需任何人工干预。我们报告正在进行的研究,以先前的工作为基础,我们展示了该方法的总体可行性。在这一贡献中,我们扩展了我们的工作,使用更先进和更少人为的实验范式来模拟频繁发生的基本类型的真实监视任务,测试该方法的可行性。在使用的视觉刺激、事件指示的更微妙的线索和所需的观看行为方面,范式更接近于真实的监视任务。作为一种方法,我们与非专家进行了一个实验(N=10)。结果证实了该方法在高级任务中事件选择的可行性。我们在事件指示的不同阶段实现了高达77%和60%的时空事件选择准确率分数。
{"title":"Spatio-Temporal Event Selection in Basic Surveillance Tasks using Eye Tracking and EEG","authors":"Jutta Hild, F. Putze, David Kaufman, Christian Kühnle, Tanja Schultz, J. Beyerer","doi":"10.1145/2666642.2666645","DOIUrl":"https://doi.org/10.1145/2666642.2666645","url":null,"abstract":"In safety- and security-critical applications like video surveillance it is crucial that human operators detect task-relevant events in the continuous video streams and select them for report or dissemination to other authorities. Usually, the selection operation is performed using a manual input device like a mouse or a joystick. Due to the visually rich and dynamic input, the required high attention, the long working time, and the challenging manual selection of moving objects, it occurs that relevant events are missed. To alleviate this problem we propose adding another event selection process, using eye-brain input. Our approach is based on eye tracking and EEG, providing spatio-temporal event selection without any manual intervention. We report ongoing research, building on prior work where we showed the general feasibility of the approach. In this contribution, we extend our work testing the feasibility of the approach using more advanced and less artificial experimental paradigms simulating frequently occurring, basic types of real surveillance tasks. The paradigms are much closer to a real surveillance task in terms of the used visual stimuli, the more subtle cues for event indication, and the required viewing behavior. As a methodology we perform an experiment (N=10) with non-experts. The results confirm the feasibility of the approach for event selection in the advanced tasks. We achieve spatio-temporal event selection accuracy scores of up to 77% and 60% for different stages of event indication.","PeriodicalId":230150,"journal":{"name":"GazeIn '14","volume":"69 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-11-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126097113","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Evaluating the Impact of Embodied Conversational Agents (ECAs) Attentional Behaviors on User Retention of Cultural Content in a Simulated Mobile Environment 在模拟移动环境中评估具身会话代理(eca)注意行为对用户保留文化内容的影响
Pub Date : 2014-11-16 DOI: 10.1145/2666642.2666650
Ioannis Doumanis, Serengul Smith
The paper presents an evaluation study of the impact of an ECA's attentional behaviors using a custom research method that combines facial expression analysis, eye-tracking and a retention test. The method provides additional channels to EEG-based methods (e.g., [8]) for the study of user attention and emotions. In order to validate the proposed approach, two tour guide applications were created with an embodied conversational agent (ECA) that presents cultural content about a real-tourist attraction. The agent simulates two attention-grabbing mechanisms - humorous and serious to attract the users' attention. A formal study was conducted to compare two tour guide applications in the lab. The data collected from the facial expression analysis and eye-tracking helped to explain particularly good and bad performances in retention tests. In terms of the study results, strong quantitative and qualitative evidence was found that an ECA should not attract more attention to itself than necessary, to avoid becoming a distraction from the flow of the content. It was also found that the ECA had an inverse effect on the retention performance of participants with different gender and their use on computer interfaces is not a good idea for elderly users.
本文采用一种结合面部表情分析、眼球追踪和记忆保留测试的定制研究方法,对ECA注意行为的影响进行了评估研究。该方法为基于脑电图的方法(如[8])研究用户注意力和情绪提供了额外的渠道。为了验证所提出的方法,我们用一个具体化的会话代理(ECA)创建了两个导游应用程序,该应用程序展示了有关真实旅游景点的文化内容。代理模拟幽默和严肃两种吸引注意力的机制来吸引用户的注意力。一项正式的研究进行了比较两种导游应用程序在实验室。从面部表情分析和眼球追踪中收集的数据有助于解释在记忆力测试中的表现。就研究结果而言,强有力的定量和定性证据表明,非洲经委会不应引起不必要的过多注意,以免分散对内容流动的注意力。研究还发现,ECA对不同性别参与者的留存表现有相反的影响,并且在计算机界面上使用ECA对老年用户来说不是一个好主意。
{"title":"Evaluating the Impact of Embodied Conversational Agents (ECAs) Attentional Behaviors on User Retention of Cultural Content in a Simulated Mobile Environment","authors":"Ioannis Doumanis, Serengul Smith","doi":"10.1145/2666642.2666650","DOIUrl":"https://doi.org/10.1145/2666642.2666650","url":null,"abstract":"The paper presents an evaluation study of the impact of an ECA's attentional behaviors using a custom research method that combines facial expression analysis, eye-tracking and a retention test. The method provides additional channels to EEG-based methods (e.g., [8]) for the study of user attention and emotions. In order to validate the proposed approach, two tour guide applications were created with an embodied conversational agent (ECA) that presents cultural content about a real-tourist attraction. The agent simulates two attention-grabbing mechanisms - humorous and serious to attract the users' attention. A formal study was conducted to compare two tour guide applications in the lab. The data collected from the facial expression analysis and eye-tracking helped to explain particularly good and bad performances in retention tests. In terms of the study results, strong quantitative and qualitative evidence was found that an ECA should not attract more attention to itself than necessary, to avoid becoming a distraction from the flow of the content. It was also found that the ECA had an inverse effect on the retention performance of participants with different gender and their use on computer interfaces is not a good idea for elderly users.","PeriodicalId":230150,"journal":{"name":"GazeIn '14","volume":"104 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-11-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131654480","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Study on Participant-controlled Eye Tracker Calibration Procedure 参与者控制眼动仪校准程序的研究
Pub Date : 2014-11-16 DOI: 10.1145/2666642.2666646
P. Kasprowski, Katarzyna Harężlak
The analysis of an eye movement signal, which can reveal o lot of information about the way human brain works, has recently attracted the attention of many researchers. The basis for such studies is data returned by specialized devices called eye-trackers. The first step of their usage is a calibration process, allowing to reflect an eye position to a point of regard. The main research problem analyzed in this paper is to check whether and how the chosen calibration scenario influences the calibration result (calibration errors). Based on this analysis of possible scenarios, a new user-controlled calibration procedure was developed. It was checked and compared with a classic approach during pilot studies using the Eye Tribe system as an eye-tracker device. The results obtained for both methods were examined in terms of provided accuracy.
眼动信号的分析可以揭示人类大脑工作方式的许多信息,最近引起了许多研究人员的注意。这类研究的基础是被称为眼球追踪器的专门设备传回的数据。他们使用的第一步是一个校准过程,允许反映一个眼睛的位置,以一个关注点。本文分析的主要研究问题是检查所选择的校准场景是否以及如何影响校准结果(校准误差)。基于对可能情况的分析,开发了一种新的用户控制校准程序。在试点研究中,使用Eye Tribe系统作为眼动仪,对该方法进行了检查并与经典方法进行了比较。根据所提供的准确性对两种方法得到的结果进行了检验。
{"title":"Study on Participant-controlled Eye Tracker Calibration Procedure","authors":"P. Kasprowski, Katarzyna Harężlak","doi":"10.1145/2666642.2666646","DOIUrl":"https://doi.org/10.1145/2666642.2666646","url":null,"abstract":"The analysis of an eye movement signal, which can reveal o lot of information about the way human brain works, has recently attracted the attention of many researchers. The basis for such studies is data returned by specialized devices called eye-trackers. The first step of their usage is a calibration process, allowing to reflect an eye position to a point of regard. The main research problem analyzed in this paper is to check whether and how the chosen calibration scenario influences the calibration result (calibration errors). Based on this analysis of possible scenarios, a new user-controlled calibration procedure was developed. It was checked and compared with a classic approach during pilot studies using the Eye Tribe system as an eye-tracker device. The results obtained for both methods were examined in terms of provided accuracy.","PeriodicalId":230150,"journal":{"name":"GazeIn '14","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-11-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123512559","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Analyzing Co-occurrence Patterns of Nonverbal Behaviors in Collaborative Learning 协作学习中非语言行为的共现模式分析
Pub Date : 2014-11-16 DOI: 10.1145/2666642.2666651
Sakiko Nihonyanagi, Yuki Hayashi, Y. Nakano
In collaborative learning, participants work on the learning task together. In this environment, linguistic information via speech as well as non-verbal information such as gaze and writing actions are important elements. It is expected that integrating the information from these behaviors will contribute to assessing the learning activity and characteristics of each participant in a more objective manner. With the objective of characterizing participants in the collaborative learning activity, this study analyzed the verbal and nonverbal behaviors and found that the gaze behaviors of individual participants and those between the participants provides useful information in distinguishing a leader of the group, one who follows the leader, or one who attends to other participants who do not appear to understand.
在协作学习中,参与者共同完成学习任务。在这种环境下,通过言语传递的语言信息以及凝视和书写动作等非语言信息都是重要的元素。整合这些行为的信息将有助于更客观地评估每个参与者的学习活动和特征。本研究分析了协作学习活动中参与者的语言和非语言行为,发现个体参与者的注视行为和参与者之间的注视行为提供了有用的信息,可以区分群体中的领导者,跟随领导者的人,或者关注其他似乎不理解的参与者。
{"title":"Analyzing Co-occurrence Patterns of Nonverbal Behaviors in Collaborative Learning","authors":"Sakiko Nihonyanagi, Yuki Hayashi, Y. Nakano","doi":"10.1145/2666642.2666651","DOIUrl":"https://doi.org/10.1145/2666642.2666651","url":null,"abstract":"In collaborative learning, participants work on the learning task together. In this environment, linguistic information via speech as well as non-verbal information such as gaze and writing actions are important elements. It is expected that integrating the information from these behaviors will contribute to assessing the learning activity and characteristics of each participant in a more objective manner. With the objective of characterizing participants in the collaborative learning activity, this study analyzed the verbal and nonverbal behaviors and found that the gaze behaviors of individual participants and those between the participants provides useful information in distinguishing a leader of the group, one who follows the leader, or one who attends to other participants who do not appear to understand.","PeriodicalId":230150,"journal":{"name":"GazeIn '14","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-11-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129648802","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Gaze-Based Virtual Task Predictor 基于注视的虚拟任务预测器
Pub Date : 2014-11-16 DOI: 10.1145/2666642.2666647
Çagla Çig, T. M. Sezgin
Pen-based systems promise an intuitive and natural interaction paradigm for tablet PCs and stylus-enabled phones. However, typical pen-based interfaces require users to switch modes frequently in order to complete ordinary tasks. Mode switching is usually achieved through hard or soft modifier keys, buttons, and soft-menus. Frequent invocation of these auxiliary mode switching elements goes against the goal of intuitive, fluid, and natural interaction. In this paper, we present a gaze-based virtual task prediction system that has the potential to alleviate dependence on explicit mode switching in pen-based systems. In particular, we show that a range of virtual manipulation commands, that would otherwise require auxiliary mode switching elements, can be issued with an 80% success rate with the aid of users' natural eye gaze behavior during pen-only interaction.
基于笔的系统为平板电脑和支持触控笔的手机提供了直观和自然的交互范例。然而,典型的基于笔的界面要求用户频繁切换模式以完成普通任务。模式切换通常通过硬或软修改键、按钮和软菜单来实现。频繁调用这些辅助模式切换元素违背了直观、流畅和自然交互的目标。在本文中,我们提出了一个基于注视的虚拟任务预测系统,它有可能减轻基于笔的系统对显式模式切换的依赖。特别是,我们展示了一系列虚拟操作命令,否则需要辅助模式切换元素,可以在用户的自然眼睛注视行为的帮助下,以80%的成功率在纯笔交互中发出。
{"title":"Gaze-Based Virtual Task Predictor","authors":"Çagla Çig, T. M. Sezgin","doi":"10.1145/2666642.2666647","DOIUrl":"https://doi.org/10.1145/2666642.2666647","url":null,"abstract":"Pen-based systems promise an intuitive and natural interaction paradigm for tablet PCs and stylus-enabled phones. However, typical pen-based interfaces require users to switch modes frequently in order to complete ordinary tasks. Mode switching is usually achieved through hard or soft modifier keys, buttons, and soft-menus. Frequent invocation of these auxiliary mode switching elements goes against the goal of intuitive, fluid, and natural interaction. In this paper, we present a gaze-based virtual task prediction system that has the potential to alleviate dependence on explicit mode switching in pen-based systems. In particular, we show that a range of virtual manipulation commands, that would otherwise require auxiliary mode switching elements, can be issued with an 80% success rate with the aid of users' natural eye gaze behavior during pen-only interaction.","PeriodicalId":230150,"journal":{"name":"GazeIn '14","volume":"13 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-11-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121025922","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Attention and Gaze in Situated Language Interaction 情境语言互动中的注意和凝视
Pub Date : 2014-11-16 DOI: 10.1145/2666642.2666643
D. Bohus
The ability to engage in natural language interaction in physically situated settings hinges on a set of competencies such as managing conversational engagement, turn taking, understanding, language and behavior generation, and interaction planning. In human-human interaction these are mixed-initiative, collaborative processes, that often involve a wide array of finely coordinated verbal and non-verbal actions. Eye gaze, and more generally attention, among many other channels, play a fundamental role. In this talk, I will discuss samples of research work we have conducted over the last few years on developing models for supporting physically situated dialog in relatively unconstrained environments. Throughout, I will highlight the role that gaze and attention play in these models. I will discuss and showcase several prototype systems that we have developed, and describe opportunities for reasoning about, interpreting and producing gaze signals in support of fluid, seamless spoken language interaction.
在物理环境中进行自然语言交互的能力取决于一系列能力,如管理会话参与、轮流、理解、语言和行为生成以及交互计划。在人与人之间的互动中,这些是混合主动的协作过程,通常涉及一系列精心协调的语言和非语言行为。在许多其他渠道中,眼睛的凝视,以及更普遍的注意力,起着基本的作用。在这次演讲中,我将讨论我们在过去几年中进行的研究工作的样本,这些研究工作是为了在相对不受约束的环境中开发支持物理位置对话的模型。贯穿全文,我将强调凝视和注意力在这些模型中所扮演的角色。我将讨论并展示我们开发的几个原型系统,并描述推理、解释和产生凝视信号的机会,以支持流畅、无缝的口语交互。
{"title":"Attention and Gaze in Situated Language Interaction","authors":"D. Bohus","doi":"10.1145/2666642.2666643","DOIUrl":"https://doi.org/10.1145/2666642.2666643","url":null,"abstract":"The ability to engage in natural language interaction in physically situated settings hinges on a set of competencies such as managing conversational engagement, turn taking, understanding, language and behavior generation, and interaction planning. In human-human interaction these are mixed-initiative, collaborative processes, that often involve a wide array of finely coordinated verbal and non-verbal actions. Eye gaze, and more generally attention, among many other channels, play a fundamental role. In this talk, I will discuss samples of research work we have conducted over the last few years on developing models for supporting physically situated dialog in relatively unconstrained environments. Throughout, I will highlight the role that gaze and attention play in these models. I will discuss and showcase several prototype systems that we have developed, and describe opportunities for reasoning about, interpreting and producing gaze signals in support of fluid, seamless spoken language interaction.","PeriodicalId":230150,"journal":{"name":"GazeIn '14","volume":"41 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-11-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124145281","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
GazeIn '14
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1