首页 > 最新文献

ICMI-MLMI '10最新文献

英文 中文
Gaze quality assisted automatic recognition of social contexts in collaborative Tetris 注视质量辅助协作俄罗斯方块社会情境的自动识别
Pub Date : 2010-11-08 DOI: 10.1145/1891903.1891914
Weifeng Li, Marc-Antoine Nüssli, Patrick Jermann
The use of dual eye-tracking is investigated in a collaborative game setting. Social context influences individual gaze and action during a collaborative Tetris game: results show that experts as well as novices adapt their playing style when interacting in mixed ability pairs. The long term goal of our work is to design adaptive gaze awareness tools that take the pair composition into account. We therefore investigate the automatic detection (or recognition) of pair composition using dual gaze-based as well as action-based multimodal features. We describe several methods for the improvement of detection (or recognition) and experimentally demonstrate their effectiveness, especially in the situations when the collected gaze data are noisy.
在协作游戏环境下,研究了双眼动追踪的使用。在合作玩俄罗斯方块游戏时,社会环境会影响个人的目光和行为:结果表明,专家和新手在混合能力的组合中互动时,会调整他们的游戏风格。我们工作的长期目标是设计考虑到对组合的自适应凝视感知工具。因此,我们研究了使用基于双凝视和基于动作的多模态特征对组合的自动检测(或识别)。我们描述了几种改进检测(或识别)的方法,并通过实验证明了它们的有效性,特别是在收集的凝视数据有噪声的情况下。
{"title":"Gaze quality assisted automatic recognition of social contexts in collaborative Tetris","authors":"Weifeng Li, Marc-Antoine Nüssli, Patrick Jermann","doi":"10.1145/1891903.1891914","DOIUrl":"https://doi.org/10.1145/1891903.1891914","url":null,"abstract":"The use of dual eye-tracking is investigated in a collaborative game setting. Social context influences individual gaze and action during a collaborative Tetris game: results show that experts as well as novices adapt their playing style when interacting in mixed ability pairs. The long term goal of our work is to design adaptive gaze awareness tools that take the pair composition into account. We therefore investigate the automatic detection (or recognition) of pair composition using dual gaze-based as well as action-based multimodal features. We describe several methods for the improvement of detection (or recognition) and experimentally demonstrate their effectiveness, especially in the situations when the collected gaze data are noisy.","PeriodicalId":181145,"journal":{"name":"ICMI-MLMI '10","volume":"6 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-11-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117165874","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Identifying emergent leadership in small groups using nonverbal communicative cues 利用非语言交际线索识别小组中的紧急领导
Pub Date : 2010-11-08 DOI: 10.1145/1891903.1891953
Dairazalia Sanchez-Cortes, O. Aran, M. S. Mast, D. Gática-Pérez
This paper addresses firstly an analysis on how an emergent leader is perceived in newly formed small-groups, and secondly, explore correlations between perception of leadership and automatically extracted nonverbal communicative cues. We hypothesize that the difference in individual nonverbal features between emergent leaders and non-emergent leaders is significant and measurable using speech activity. Our results on a new interaction corpus show that such an approach is promising, identifying the emergent leader with an accuracy of up to 80%.
本文首先分析了在新形成的小团体中如何感知一个突发领导者,然后探讨了领导力感知与自动提取的非语言交际线索之间的相关性。我们假设紧急领导者和非紧急领导者在个体非语言特征上的差异是显著的,并且可以用言语活动来衡量。我们在一个新的交互语料库上的结果表明,这种方法是有希望的,识别紧急领导者的准确率高达80%。
{"title":"Identifying emergent leadership in small groups using nonverbal communicative cues","authors":"Dairazalia Sanchez-Cortes, O. Aran, M. S. Mast, D. Gática-Pérez","doi":"10.1145/1891903.1891953","DOIUrl":"https://doi.org/10.1145/1891903.1891953","url":null,"abstract":"This paper addresses firstly an analysis on how an emergent leader is perceived in newly formed small-groups, and secondly, explore correlations between perception of leadership and automatically extracted nonverbal communicative cues. We hypothesize that the difference in individual nonverbal features between emergent leaders and non-emergent leaders is significant and measurable using speech activity. Our results on a new interaction corpus show that such an approach is promising, identifying the emergent leader with an accuracy of up to 80%.","PeriodicalId":181145,"journal":{"name":"ICMI-MLMI '10","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-11-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128142537","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 52
PhotoMagnets: supporting flexible browsing and searching in photo collections PhotoMagnets:支持在图片集中灵活浏览和搜索
Pub Date : 2010-11-08 DOI: 10.1145/1891903.1891936
Ya-Xi Chen, M. Reiter, A. Butz
People's activities around their photo collections are often highly dynamic and unstructured, such as casual browsing and searching or loosely structured storytelling. User interfaces to support such an exploratory behavior are a challenging research question. We explore ways to enhance the flexibility in dealing with photo collections and designed a system named PhotoMagnets. It uses a magnet metaphor in addition to more traditional interface elements in order to support a flexible combination of structured and unstructured photo browsing and searching. In an evaluation we received positive feedback especially on the flexibility provided by this approach.
人们围绕他们的照片收藏的活动通常是高度动态和非结构化的,比如随意的浏览和搜索,或者松散结构的讲故事。支持这种探索性行为的用户界面是一个具有挑战性的研究问题。我们探索了增强处理照片集合的灵活性的方法,并设计了一个名为PhotoMagnets的系统。为了支持结构化和非结构化照片浏览和搜索的灵活组合,除了使用更传统的界面元素外,它还使用了磁铁隐喻。在一次评估中,我们收到了积极的反馈,特别是关于这种方法提供的灵活性。
{"title":"PhotoMagnets: supporting flexible browsing and searching in photo collections","authors":"Ya-Xi Chen, M. Reiter, A. Butz","doi":"10.1145/1891903.1891936","DOIUrl":"https://doi.org/10.1145/1891903.1891936","url":null,"abstract":"People's activities around their photo collections are often highly dynamic and unstructured, such as casual browsing and searching or loosely structured storytelling. User interfaces to support such an exploratory behavior are a challenging research question. We explore ways to enhance the flexibility in dealing with photo collections and designed a system named PhotoMagnets. It uses a magnet metaphor in addition to more traditional interface elements in order to support a flexible combination of structured and unstructured photo browsing and searching. In an evaluation we received positive feedback especially on the flexibility provided by this approach.","PeriodicalId":181145,"journal":{"name":"ICMI-MLMI '10","volume":"35 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-11-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131203196","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Multi-modal computer assisted speech transcription 多模态计算机辅助语音转录
Pub Date : 2010-11-08 DOI: 10.1145/1891903.1891943
Luis Rodríguez, I. García-Varea, E. Vidal
Speech recognition systems are not typically able to produce error-free results in real scenarios. On account of this, human intervention is usually needed. This intervention can be included into the system by following the Computer Assisted Speech Transcription (CAST) approach, where the user constantly interacts with the system during the transcription process. In order to improve this user interaction, a speech multi-modal interface is proposed here. In addition, the user of word graphs within CAST aimed at facilitating the design of such interface as well as improving the system response time is also discussed
语音识别系统通常无法在真实场景中产生无错误的结果。因此,通常需要人为干预。这种干预可以通过遵循计算机辅助语音转录(CAST)方法纳入系统,在这种方法中,用户在转录过程中不断与系统交互。为了改善这种用户交互,本文提出了一种语音多模态界面。此外,本文还讨论了CAST中词图的使用,以方便这种界面的设计和提高系统的响应时间
{"title":"Multi-modal computer assisted speech transcription","authors":"Luis Rodríguez, I. García-Varea, E. Vidal","doi":"10.1145/1891903.1891943","DOIUrl":"https://doi.org/10.1145/1891903.1891943","url":null,"abstract":"Speech recognition systems are not typically able to produce error-free results in real scenarios. On account of this, human intervention is usually needed. This intervention can be included into the system by following the Computer Assisted Speech Transcription (CAST) approach, where the user constantly interacts with the system during the transcription process. In order to improve this user interaction, a speech multi-modal interface is proposed here. In addition, the user of word graphs within CAST aimed at facilitating the design of such interface as well as improving the system response time is also discussed","PeriodicalId":181145,"journal":{"name":"ICMI-MLMI '10","volume":"44 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-11-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132283466","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Active learning strategies for handwritten text transcription 手写文本抄写的主动学习策略
Pub Date : 2010-11-08 DOI: 10.1145/1891903.1891962
Nicolás Serrano, Adrià Giménez, A. Sanchís, Alfons Juan-Císcar
Active learning strategies are being increasingly used in a variety of real-world tasks, though their application to handwritten text transcription in old manuscripts remains nearly unexplored. The basic idea is to follow a sequential, line-byline transcription of the whole manuscript in which a continuously retrained system interacts with the user to efficiently transcribe each new line. This approach has been recently explored using a conventional strategy by which the user is only asked to supervise words that are not recognized with high confidence. In this paper, the conventional strategy is improved by also letting the system to recompute most probable hypotheses with the constraints imposed by user supervisions. In particular, two strategies are studied which differ in the frequency of hypothesis recomputation on the current line: after each (iterative) or all (delayed) user corrections. Empirical results are reported on two real tasks showing that these strategies outperform the conventional approach.
主动学习策略越来越多地应用于各种现实世界的任务中,尽管它们在旧手稿手写文本转录方面的应用几乎尚未被探索。基本思想是遵循整个手稿的顺序,行-署名转录,其中一个不断重新训练的系统与用户交互,以有效地转录每一个新的行。这种方法最近已经被探索了,使用一种传统的策略,通过这种策略,用户只被要求监督那些不能被高度自信地识别的单词。在本文中,通过让系统在用户监督的约束下重新计算最可能的假设,改进了传统的策略。特别地,研究了两种策略,它们在当前行的假设重新计算频率不同:在每次(迭代)或所有(延迟)用户更正之后。两个实际任务的实证结果表明,这些策略优于传统方法。
{"title":"Active learning strategies for handwritten text transcription","authors":"Nicolás Serrano, Adrià Giménez, A. Sanchís, Alfons Juan-Císcar","doi":"10.1145/1891903.1891962","DOIUrl":"https://doi.org/10.1145/1891903.1891962","url":null,"abstract":"Active learning strategies are being increasingly used in a variety of real-world tasks, though their application to handwritten text transcription in old manuscripts remains nearly unexplored. The basic idea is to follow a sequential, line-byline transcription of the whole manuscript in which a continuously retrained system interacts with the user to efficiently transcribe each new line. This approach has been recently explored using a conventional strategy by which the user is only asked to supervise words that are not recognized with high confidence. In this paper, the conventional strategy is improved by also letting the system to recompute most probable hypotheses with the constraints imposed by user supervisions. In particular, two strategies are studied which differ in the frequency of hypothesis recomputation on the current line: after each (iterative) or all (delayed) user corrections. Empirical results are reported on two real tasks showing that these strategies outperform the conventional approach.","PeriodicalId":181145,"journal":{"name":"ICMI-MLMI '10","volume":"12 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-11-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133185277","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Dynamic user interface distribution for flexible multimodal interaction 灵活的多模态交互的动态用户界面分布
Pub Date : 2010-11-08 DOI: 10.1145/1891903.1891930
Marco Blumendorf, Dirk Roscher, S. Albayrak
The availability of numerous networked interaction devices within smart environments makes the exploitation of these devices for innovative and more natural interaction possible. In our work we make use of TVs with remote controls, picture frames, mobile phones, touch screens, stereos and PCs to create multimodal user interfaces. The combination of the interaction capabilities of the different devices allows to achieve a more suitable interaction for a situation. Changing situations can then require the dynamic redistribution of the created interfaces and the alteration of the used modalities and devices to keep up the interaction. In this paper we describe our approach for dynamically (re-) distributing user interfaces at run-time. A distribution component is responsible for determining the devices for the interaction based on the (changing) environment situation and the user interface requirements. The component provides possibilities to the application developer and to the user to influence the distribution according to their needs. A user interface model describes the interaction and the modality relations according to the CARE properties (Complementarity, Assignment, Redundancy and Equivalency) and a context model gathers and provides information about the environment.
智能环境中大量网络交互设备的可用性使得利用这些设备进行创新和更自然的交互成为可能。在我们的工作中,我们利用带遥控器的电视、相框、移动电话、触摸屏、音响和个人电脑来创建多模式的用户界面。不同设备的交互能力的组合允许实现更适合于某种情况的交互。然后,不断变化的情况可能需要对创建的接口进行动态重新分配,并更改使用的模式和设备以保持交互。在本文中,我们描述了在运行时动态(重新)分配用户界面的方法。分布组件负责根据(变化的)环境情况和用户界面需求确定用于交互的设备。该组件为应用程序开发人员和用户提供了根据他们的需要影响分发的可能性。用户界面模型根据CARE属性(互补性、赋值性、冗余性和等效性)描述交互和模态关系,上下文模型收集并提供有关环境的信息。
{"title":"Dynamic user interface distribution for flexible multimodal interaction","authors":"Marco Blumendorf, Dirk Roscher, S. Albayrak","doi":"10.1145/1891903.1891930","DOIUrl":"https://doi.org/10.1145/1891903.1891930","url":null,"abstract":"The availability of numerous networked interaction devices within smart environments makes the exploitation of these devices for innovative and more natural interaction possible. In our work we make use of TVs with remote controls, picture frames, mobile phones, touch screens, stereos and PCs to create multimodal user interfaces. The combination of the interaction capabilities of the different devices allows to achieve a more suitable interaction for a situation. Changing situations can then require the dynamic redistribution of the created interfaces and the alteration of the used modalities and devices to keep up the interaction. In this paper we describe our approach for dynamically (re-) distributing user interfaces at run-time. A distribution component is responsible for determining the devices for the interaction based on the (changing) environment situation and the user interface requirements. The component provides possibilities to the application developer and to the user to influence the distribution according to their needs. A user interface model describes the interaction and the modality relations according to the CARE properties (Complementarity, Assignment, Redundancy and Equivalency) and a context model gathers and provides information about the environment.","PeriodicalId":181145,"journal":{"name":"ICMI-MLMI '10","volume":"32 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-11-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129942058","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 19
3D-press: haptic illusion of compliance when pressing on a rigid surface 3d按压:按压刚性表面时产生的顺应性触觉错觉
Pub Date : 2010-11-08 DOI: 10.1145/1891903.1891931
Johan Kildal
This paper reports a new intramodal haptic illusion. This illusion involves a person pressing on a rigid surface and perceiving that the surface is compliant, i.e. perceiving that the contact point displaces into the surface. The design process, method and conditions used to create this illusion are described in detail. A user study is also reported in which all participants using variants of the basic method experienced the illusion, demonstrating the effectiveness of the method. This study also offers an initial indication of the mechanical dimensions of illusory compliance that could be manipulated by varying the stimuli presented to the users. This method could be used to augment touch interaction with mobile devices, transcending the rigid two-dimensional tangible surface (touch display) currently found on them.
本文报道了一种新的模内触觉错觉。这种错觉涉及到一个人按压刚性表面并感知到表面是柔顺的,即感知到接触点移位到表面上。详细描述了创造这种错觉的设计过程、方法和条件。一项用户研究也被报道,其中所有使用基本方法变体的参与者都经历了错觉,证明了该方法的有效性。这项研究也提供了错觉依从性的机械维度的初步指示,可以通过改变呈现给用户的刺激来操纵。这种方法可以用来增强与移动设备的触摸交互,超越目前在移动设备上发现的刚性二维有形表面(触摸显示)。
{"title":"3D-press: haptic illusion of compliance when pressing on a rigid surface","authors":"Johan Kildal","doi":"10.1145/1891903.1891931","DOIUrl":"https://doi.org/10.1145/1891903.1891931","url":null,"abstract":"This paper reports a new intramodal haptic illusion. This illusion involves a person pressing on a rigid surface and perceiving that the surface is compliant, i.e. perceiving that the contact point displaces into the surface. The design process, method and conditions used to create this illusion are described in detail. A user study is also reported in which all participants using variants of the basic method experienced the illusion, demonstrating the effectiveness of the method. This study also offers an initial indication of the mechanical dimensions of illusory compliance that could be manipulated by varying the stimuli presented to the users. This method could be used to augment touch interaction with mobile devices, transcending the rigid two-dimensional tangible surface (touch display) currently found on them.","PeriodicalId":181145,"journal":{"name":"ICMI-MLMI '10","volume":"19 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-11-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132017596","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 63
The Ambient Spotlight: personal multimodal search without query 环境聚光灯:无查询的个人多模式搜索
Pub Date : 2010-11-08 DOI: 10.1145/1891903.1891919
J. Kilgour, J. Carletta, S. Renals
The Ambient Spotlight is a prototype system based on personal meeting capture using a laptop and a portable microphone array. The system automatically recognises and structures the meeting content using automatic speech recognition, topic segmentation and extractive summarisation. The recognised speech in the meeting is used to construct queries to automatically link meeting segments to other relevant material, both multimodal and textual. The interface to the system is constructed around a standard calendar interface, and it is integrated with the laptop's standard indexing, search and retrieval.
环境聚光灯是一个基于个人会议捕捉的原型系统,使用笔记本电脑和便携式麦克风阵列。系统采用自动语音识别、主题切分和抽取摘要等方法对会议内容进行自动识别和结构化。会议中识别的语音用于构建查询,以自动将会议片段链接到其他相关材料,包括多模态和文本。系统的接口是围绕标准日历接口构建的,它与笔记本电脑的标准索引、搜索和检索集成在一起。
{"title":"The Ambient Spotlight: personal multimodal search without query","authors":"J. Kilgour, J. Carletta, S. Renals","doi":"10.1145/1891903.1891919","DOIUrl":"https://doi.org/10.1145/1891903.1891919","url":null,"abstract":"The Ambient Spotlight is a prototype system based on personal meeting capture using a laptop and a portable microphone array. The system automatically recognises and structures the meeting content using automatic speech recognition, topic segmentation and extractive summarisation. The recognised speech in the meeting is used to construct queries to automatically link meeting segments to other relevant material, both multimodal and textual. The interface to the system is constructed around a standard calendar interface, and it is integrated with the laptop's standard indexing, search and retrieval.","PeriodicalId":181145,"journal":{"name":"ICMI-MLMI '10","volume":"29 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-11-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114207473","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Human-centered attention models for video summarization 视频摘要的以人为中心的注意力模型
Pub Date : 2010-11-08 DOI: 10.1145/1891903.1891938
Kaiming Li, Lei Guo, C. Faraco, Dajiang Zhu, Fan Deng, Tuo Zhang, Xi Jiang, Degang Zhang, Hanbo Chen, Xintao Hu, L. Miller, Tianming Liu
A variety of user attention models for video/audio streams have been developed for video summarization and abstraction, in order to facilitate efficient video browsing and indexing. Essentially, human brain is the end user and evaluator of multimedia content and representation, and its responses can provide meaningful guidelines for multimedia stream summarization. For example, video/audio segments that significantly activate the visual, auditory, language and working memory systems of the human brain should be considered more important than others. It should be noted that user experience studies could be useful for such evaluations, but are suboptimal in terms of their capability of accurately capturing the full-length dynamics and interactions of the brain's response. This paper presents our preliminary efforts in applying the brain imaging technique of functional magnetic resonance imaging (fMRI) to quantify and model the dynamics and interactions between multimedia streams and brain response, when the human subjects are presented with the multimedia clips, in order to develop human-centered attention models that can be used to guide and facilitate more effective and efficient multimedia summarization. Our initial results are encouraging.
为了方便高效的视频浏览和索引,各种视频/音频流的用户注意力模型已经被开发出来用于视频摘要和抽象。从本质上讲,人脑是多媒体内容和表示的最终用户和评估者,其响应可以为多媒体流总结提供有意义的指导。例如,能够显著激活人类大脑的视觉、听觉、语言和工作记忆系统的视频/音频片段应该被认为比其他片段更重要。值得注意的是,用户体验研究可能对这种评估有用,但就其准确捕捉大脑反应的全长动态和相互作用的能力而言,这是次优的。本文介绍了应用功能磁共振成像(fMRI)脑成像技术对多媒体流与大脑反应之间的动态和相互作用进行量化和建模的初步工作,以期建立以人为中心的注意力模型,用于指导和促进更有效和高效的多媒体总结。我们的初步结果令人鼓舞。
{"title":"Human-centered attention models for video summarization","authors":"Kaiming Li, Lei Guo, C. Faraco, Dajiang Zhu, Fan Deng, Tuo Zhang, Xi Jiang, Degang Zhang, Hanbo Chen, Xintao Hu, L. Miller, Tianming Liu","doi":"10.1145/1891903.1891938","DOIUrl":"https://doi.org/10.1145/1891903.1891938","url":null,"abstract":"A variety of user attention models for video/audio streams have been developed for video summarization and abstraction, in order to facilitate efficient video browsing and indexing. Essentially, human brain is the end user and evaluator of multimedia content and representation, and its responses can provide meaningful guidelines for multimedia stream summarization. For example, video/audio segments that significantly activate the visual, auditory, language and working memory systems of the human brain should be considered more important than others. It should be noted that user experience studies could be useful for such evaluations, but are suboptimal in terms of their capability of accurately capturing the full-length dynamics and interactions of the brain's response. This paper presents our preliminary efforts in applying the brain imaging technique of functional magnetic resonance imaging (fMRI) to quantify and model the dynamics and interactions between multimedia streams and brain response, when the human subjects are presented with the multimedia clips, in order to develop human-centered attention models that can be used to guide and facilitate more effective and efficient multimedia summarization. Our initial results are encouraging.","PeriodicalId":181145,"journal":{"name":"ICMI-MLMI '10","volume":"7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-11-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133601826","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Feedback is... late: measuring multimodal delays in mobile device touchscreen interaction 反馈是……延迟:测量移动设备触摸屏交互中的多模态延迟
Pub Date : 2010-11-08 DOI: 10.1145/1891903.1891907
Topi Kaaresoja, S. Brewster
Multimodal interaction is becoming common in many kinds of devices, particularly mobile phones. If care is not taken in design and implementation, there may be latencies in the timing of feedback in the different modalities may have unintended effects on users. This paper introduces an easy to implement multimodal latency measurement tool for touchscreen interaction. It uses off-the-shelf components and free software and is capable of measuring latencies accurately between different interaction events in different modalities. The tool uses a high-speed camera, a mirror, a microphone and an accelerometer to measure the touch, visual, audio and tactile feedback events that occur in touchscreen interaction. The microphone and the accelerometer are both interfaced with a standard PC soundcard that makes the measurement and analysis simple. The latencies are obtained by hand and eye using a slow-motion video player and an audio editor. To validate the tool, we measured four commercial mobile phones. Our results show that there are significant differences in latencies, not only between the devices, but also between different applications and modalities within one device. In this paper the focus is on mobile touchscreen devices, but with minor modifications our tool could be also used in other domains.
多模式交互在许多类型的设备中变得越来越普遍,尤其是移动电话。如果在设计和实现中不注意,在不同的模式中反馈的时间可能会有延迟,这可能会对用户产生意想不到的影响。本文介绍了一种易于实现的触摸屏交互多模态延迟测量工具。它使用现成的组件和免费软件,能够准确地测量不同模式下不同交互事件之间的延迟。该工具使用高速摄像机、镜子、麦克风和加速度计来测量触摸屏交互中发生的触摸、视觉、音频和触觉反馈事件。麦克风和加速度计都与一个标准的PC声卡接口,使测量和分析变得简单。使用慢动作视频播放器和音频编辑器通过手和眼获得延迟。为了验证这个工具,我们测量了四个商用手机。我们的结果表明,不仅在设备之间,而且在一个设备内的不同应用和模式之间,延迟也存在显着差异。在本文中,重点是移动触摸屏设备,但稍加修改,我们的工具也可以用于其他领域。
{"title":"Feedback is... late: measuring multimodal delays in mobile device touchscreen interaction","authors":"Topi Kaaresoja, S. Brewster","doi":"10.1145/1891903.1891907","DOIUrl":"https://doi.org/10.1145/1891903.1891907","url":null,"abstract":"Multimodal interaction is becoming common in many kinds of devices, particularly mobile phones. If care is not taken in design and implementation, there may be latencies in the timing of feedback in the different modalities may have unintended effects on users. This paper introduces an easy to implement multimodal latency measurement tool for touchscreen interaction. It uses off-the-shelf components and free software and is capable of measuring latencies accurately between different interaction events in different modalities. The tool uses a high-speed camera, a mirror, a microphone and an accelerometer to measure the touch, visual, audio and tactile feedback events that occur in touchscreen interaction. The microphone and the accelerometer are both interfaced with a standard PC soundcard that makes the measurement and analysis simple. The latencies are obtained by hand and eye using a slow-motion video player and an audio editor. To validate the tool, we measured four commercial mobile phones. Our results show that there are significant differences in latencies, not only between the devices, but also between different applications and modalities within one device. In this paper the focus is on mobile touchscreen devices, but with minor modifications our tool could be also used in other domains.","PeriodicalId":181145,"journal":{"name":"ICMI-MLMI '10","volume":"26 12","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-11-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114129927","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 44
期刊
ICMI-MLMI '10
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1