首页 > 最新文献

Proceedings of the 10th international conference on Intelligent user interfaces最新文献

英文 中文
CASIS: a context-aware speech interface system CASIS:上下文感知语音接口系统
Pub Date : 2005-01-10 DOI: 10.1145/1040830.1040880
H. Lee, Shinsuke Kobayashi, N. Koshizuka, K. Sakamura
In this paper, we propose a robust natural language interface called CASIS for controlling devices in an intelligent environment. CASIS is novel in a sense that it integrates physical context acquired from the sensors embedded in the environment with traditionally used context to reduce the system error rate and disambiguate deictic references and elliptical inputs. The n-best result of the speech recognizer is re-ranked by a score calculated using a Bayesian network consisting of information from the input utterance and context. In our prototype system that uses device states, brightness, speaker location, chair occupancy, speech direction and action history as context, the system error rate has been reduced by 41% compared to a baseline system that does not leverage on context information.
在本文中,我们提出了一种称为CASIS的鲁棒自然语言接口,用于控制智能环境中的设备。CASIS在某种意义上是新颖的,它将从嵌入环境中的传感器获取的物理环境与传统使用的环境相结合,以降低系统错误率,消除指示参考和椭圆输入的歧义。语音识别器的n个最佳结果通过使用由输入话语和上下文信息组成的贝叶斯网络计算的分数重新排名。在我们的原型系统中,使用设备状态、亮度、说话者位置、椅子占用、语音方向和动作历史作为上下文,与不利用上下文信息的基线系统相比,系统错误率降低了41%。
{"title":"CASIS: a context-aware speech interface system","authors":"H. Lee, Shinsuke Kobayashi, N. Koshizuka, K. Sakamura","doi":"10.1145/1040830.1040880","DOIUrl":"https://doi.org/10.1145/1040830.1040880","url":null,"abstract":"In this paper, we propose a robust natural language interface called CASIS for controlling devices in an intelligent environment. CASIS is novel in a sense that it integrates physical context acquired from the sensors embedded in the environment with traditionally used context to reduce the system error rate and disambiguate deictic references and elliptical inputs. The n-best result of the speech recognizer is re-ranked by a score calculated using a Bayesian network consisting of information from the input utterance and context. In our prototype system that uses device states, brightness, speaker location, chair occupancy, speech direction and action history as context, the system error rate has been reduced by 41% compared to a baseline system that does not leverage on context information.","PeriodicalId":376409,"journal":{"name":"Proceedings of the 10th international conference on Intelligent user interfaces","volume":"340 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124779167","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 26
Adaptive teaching strategy for online learning 网络学习的适应性教学策略
Pub Date : 2005-01-10 DOI: 10.1145/1040830.1040892
Jungsoon P. Yoo, Cen Li, C. Pettey
Finding the optimal teaching strategy for an individual student is difficult even for an experienced teacher. Identifying and incorporating multiple optimal teaching strategies for different students in a class is even harder. This paper presents an Adaptive tutor for online Learning, AtoL, for Computer Science laboratories that identifies and applies the appropriate teaching strategies for students on an individual basis. The optimal strategy for a student is identified in two steps. First, a basic strategy for a student is identified using rules learned from a supervised learning system. Then the basic strategy is refined to better fit the student using models learned using an unsupervised learning system that takes into account the temporal nature of the problem solving process. The learning algorithms as well as the initial experimental results are presented.
即使是经验丰富的教师,也很难找到适合个别学生的最佳教学策略。识别和整合针对不同学生的多种最佳教学策略就更难了。本文介绍了一种用于计算机科学实验室的在线学习自适应导师,即AtoL,它可以根据学生的个人情况确定并应用适当的教学策略。学生的最佳策略分为两个步骤。首先,使用从监督学习系统中学习到的规则来确定学生的基本策略。然后对基本策略进行细化,以便使用使用无监督学习系统学习的模型更好地适应学生,该系统考虑到问题解决过程的时间性质。给出了学习算法和初步实验结果。
{"title":"Adaptive teaching strategy for online learning","authors":"Jungsoon P. Yoo, Cen Li, C. Pettey","doi":"10.1145/1040830.1040892","DOIUrl":"https://doi.org/10.1145/1040830.1040892","url":null,"abstract":"Finding the optimal teaching strategy for an individual student is difficult even for an experienced teacher. Identifying and incorporating multiple optimal teaching strategies for different students in a class is even harder. This paper presents an Adaptive tutor for online Learning, AtoL, for Computer Science laboratories that identifies and applies the appropriate teaching strategies for students on an individual basis. The optimal strategy for a student is identified in two steps. First, a basic strategy for a student is identified using rules learned from a supervised learning system. Then the basic strategy is refined to better fit the student using models learned using an unsupervised learning system that takes into account the temporal nature of the problem solving process. The learning algorithms as well as the initial experimental results are presented.","PeriodicalId":376409,"journal":{"name":"Proceedings of the 10th international conference on Intelligent user interfaces","volume":"31 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124996224","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 12
Generating semantic contexts from spoken conversation in meetings 从会议中的口语会话中生成语义上下文
Pub Date : 2005-01-10 DOI: 10.1145/1040830.1040902
J. Ziegler, Zoulfa El Jerroudi, Karsten Böhm
SemanticTalk is a tool for supporting face-to-face meetings and discussions by automatically generating a semantic context from spoken conversations. We use speech recognition and topic extraction from a large terminological database to create a network of discussion topics in real-time. This network includes concepts explicitly addressed in the discussion as well as semantically associated terms, and is visualized to increase conversational awareness and creativity in the group.
semantic talk是一种支持面对面会议和讨论的工具,通过从口语对话中自动生成语义上下文。我们使用语音识别和从大型术语数据库中提取主题来实时创建讨论主题网络。这个网络包括讨论中明确提到的概念以及语义相关的术语,并通过可视化来提高小组中的对话意识和创造力。
{"title":"Generating semantic contexts from spoken conversation in meetings","authors":"J. Ziegler, Zoulfa El Jerroudi, Karsten Böhm","doi":"10.1145/1040830.1040902","DOIUrl":"https://doi.org/10.1145/1040830.1040902","url":null,"abstract":"SemanticTalk is a tool for supporting face-to-face meetings and discussions by automatically generating a semantic context from spoken conversations. We use speech recognition and topic extraction from a large terminological database to create a network of discussion topics in real-time. This network includes concepts explicitly addressed in the discussion as well as semantically associated terms, and is visualized to increase conversational awareness and creativity in the group.","PeriodicalId":376409,"journal":{"name":"Proceedings of the 10th international conference on Intelligent user interfaces","volume":"69 6 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125874408","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Conventions in human-human multi-threaded dialogues: a preliminary study 人-人多线程对话中的约定:初步研究
Pub Date : 2005-01-10 DOI: 10.1145/1040830.1040903
P. Heeman, Fan Yang, A. Kun, A. Shyrokov
In this paper, we explore the conventions that people use in managing multiple dialogue threads. In particular, we focus on where in a thread people interrupt when switching to another thread. We find that some subjects are able to vary where they switch depending on how urgent the interrupting task is. When time-allowed, they switched at the end of a discourse segment, which we hypothesize is less disruptive to the interrupted task when it is later resumed.
在本文中,我们探讨了人们在管理多个对话线程时使用的惯例。特别是,我们关注当切换到另一个线程时,人们在线程中的中断位置。我们发现,一些受试者能够根据中断任务的紧急程度改变他们切换的位置。在时间允许的情况下,他们会在一段话语结束时切换,我们假设这对被打断的任务的破坏性较小。
{"title":"Conventions in human-human multi-threaded dialogues: a preliminary study","authors":"P. Heeman, Fan Yang, A. Kun, A. Shyrokov","doi":"10.1145/1040830.1040903","DOIUrl":"https://doi.org/10.1145/1040830.1040903","url":null,"abstract":"In this paper, we explore the conventions that people use in managing multiple dialogue threads. In particular, we focus on where in a thread people interrupt when switching to another thread. We find that some subjects are able to vary where they switch depending on how urgent the interrupting task is. When time-allowed, they switched at the end of a discourse segment, which we hypothesize is less disruptive to the interrupted task when it is later resumed.","PeriodicalId":376409,"journal":{"name":"Proceedings of the 10th international conference on Intelligent user interfaces","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129154612","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 24
Person-independent estimation of emotional experiences from facial expressions 从面部表情对情绪体验的独立于人的估计
Pub Date : 2005-01-10 DOI: 10.1145/1040830.1040883
Timo Partala, Veikko Surakka, T. Vanhala
The aim of this research was to develop methods for the automatic person-independent estimation of experienced emotions from facial expressions. Ten subjects watched series of emotionally arousing pictures and videos, while the electromyographic (EMG) activity of two facial muscles: zygomaticus major (activated in smiling) and corrugator supercilii (activated in frowning) was registered. Based on the changes in the activity of these two facial muscles, it was possible to distinguish between ratings of positive and negative emotional experiences at a rate of almost 70% for pictures and over 80% for videos. Using these methods, the computer could adapt its behavior according to the user's emotions during human-computer interaction.
本研究的目的是开发从面部表情中自动独立于人的经验情绪估计方法。10名被试观看了一系列激动情绪的图片和视频,同时记录了两组面部肌肉的肌电图活动:颧大肌(微笑时激活)和皱眉肌(皱眉时激活)。根据这两块面部肌肉活动的变化,可以区分出积极和消极情绪体验的评分,图片的准确率接近70%,视频的准确率超过80%。利用这些方法,计算机可以在人机交互过程中根据用户的情绪调整其行为。
{"title":"Person-independent estimation of emotional experiences from facial expressions","authors":"Timo Partala, Veikko Surakka, T. Vanhala","doi":"10.1145/1040830.1040883","DOIUrl":"https://doi.org/10.1145/1040830.1040883","url":null,"abstract":"The aim of this research was to develop methods for the automatic person-independent estimation of experienced emotions from facial expressions. Ten subjects watched series of emotionally arousing pictures and videos, while the electromyographic (EMG) activity of two facial muscles: zygomaticus major (activated in smiling) and corrugator supercilii (activated in frowning) was registered. Based on the changes in the activity of these two facial muscles, it was possible to distinguish between ratings of positive and negative emotional experiences at a rate of almost 70% for pictures and over 80% for videos. Using these methods, the computer could adapt its behavior according to the user's emotions during human-computer interaction.","PeriodicalId":376409,"journal":{"name":"Proceedings of the 10th international conference on Intelligent user interfaces","volume":"38 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114815731","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 24
The UI pilot: a model-based tool to guide early interface design UI试点:一个基于模型的工具,用于指导早期的界面设计
Pub Date : 2005-01-10 DOI: 10.1145/1040830.1040877
A. Puerta, M. Micheletti, Alan Mak
In this paper, we introduce the User Interface Pilot, a model-based software tool that enables designers and engineers to create the initial specifications for the pages of a website, or for the screens of a desktop or mobile application. The tool guides the design of these specifications, commonly known as wireframes, in a user-centered fashion by framing the context of the design within the concepts of user tasks, user types, and data objects. Unlike previous model-based tools, the User Interface Pilot does not impose a rigid model-driven methodology and functions well within common software engineering development processes. The tool has been used in over twenty real-world user interface design projects.
在本文中,我们介绍了User Interface Pilot,这是一个基于模型的软件工具,它使设计师和工程师能够为网站页面或桌面或移动应用程序的屏幕创建初始规范。该工具通过在用户任务、用户类型和数据对象的概念中构建设计的上下文,以以用户为中心的方式指导这些规范(通常称为线框图)的设计。不像以前的基于模型的工具,User Interface Pilot没有强加一个严格的模型驱动的方法,并且在普通的软件工程开发过程中功能良好。该工具已在20多个实际用户界面设计项目中使用。
{"title":"The UI pilot: a model-based tool to guide early interface design","authors":"A. Puerta, M. Micheletti, Alan Mak","doi":"10.1145/1040830.1040877","DOIUrl":"https://doi.org/10.1145/1040830.1040877","url":null,"abstract":"In this paper, we introduce the User Interface Pilot, a model-based software tool that enables designers and engineers to create the initial specifications for the pages of a website, or for the screens of a desktop or mobile application. The tool guides the design of these specifications, commonly known as wireframes, in a user-centered fashion by framing the context of the design within the concepts of user tasks, user types, and data objects. Unlike previous model-based tools, the User Interface Pilot does not impose a rigid model-driven methodology and functions well within common software engineering development processes. The tool has been used in over twenty real-world user interface design projects.","PeriodicalId":376409,"journal":{"name":"Proceedings of the 10th international conference on Intelligent user interfaces","volume":"270 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127322081","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 33
Interactively building agents for consumer-side data mining 交互式地为消费者端数据挖掘构建代理
Pub Date : 2005-01-10 DOI: 10.1145/1040830.1040891
R. Tuchinda, Craig A. Knoblock
Integrating and mining data from different web sources can make end-users well-informed when they make decisions. One of many limitations that bars end-users from taking advantages of such process is the complexity in each of the steps required to gather, integrate, monitor, and mine data from different websites. We present the idea of combining the data integration, monitoring, and mining as one single process in the form of an intelligent assistant that guides end-users to specify their mining tasks by just answering questions. This easy-to-use approach, which trades off complexity in terms of available operations with the ease of use, has the ability to provide interesting insight into the data that would requires days of human effort to gather, combine, and mine manually from the web.
整合和挖掘来自不同web来源的数据可以使最终用户在做决定时获得充分的信息。阻止最终用户利用这种过程的许多限制之一是收集、集成、监视和挖掘来自不同网站的数据所需的每个步骤的复杂性。我们提出了将数据集成、监控和挖掘作为一个单一过程的想法,以智能助手的形式指导最终用户通过回答问题来指定他们的挖掘任务。这种易于使用的方法在可用操作的复杂性和易用性方面进行了权衡,能够提供对数据的有趣见解,而这些数据需要花费数天的人力从web中手动收集、组合和挖掘。
{"title":"Interactively building agents for consumer-side data mining","authors":"R. Tuchinda, Craig A. Knoblock","doi":"10.1145/1040830.1040891","DOIUrl":"https://doi.org/10.1145/1040830.1040891","url":null,"abstract":"Integrating and mining data from different web sources can make end-users well-informed when they make decisions. One of many limitations that bars end-users from taking advantages of such process is the complexity in each of the steps required to gather, integrate, monitor, and mine data from different websites. We present the idea of combining the data integration, monitoring, and mining as one single process in the form of an intelligent assistant that guides end-users to specify their mining tasks by just answering questions. This easy-to-use approach, which trades off complexity in terms of available operations with the ease of use, has the ability to provide interesting insight into the data that would requires days of human effort to gather, combine, and mine manually from the web.","PeriodicalId":376409,"journal":{"name":"Proceedings of the 10th international conference on Intelligent user interfaces","volume":"15 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124803351","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Communicating user's focus of attention by image processing as input for a mobile museum guide 通过图像处理作为移动博物馆导览的输入,传达用户关注的焦点
Pub Date : 2005-01-10 DOI: 10.1145/1040830.1040905
A. Albertini, R. Brunelli, O. Stock, M. Zancanaro
The paper presents a first prototype of a handheld museum guide delivering contextualized information based on the recognition of drawing details selected by the user through the guide camera. The resulting interaction modality has been analyzed and compared to previous approaches. Finally, alternative, more scalable, solutions are presented that preserve the most interesting features of the system described.
本文提出了一种手持式博物馆导览器的原型,该导览器通过导览相机对用户选择的绘画细节进行识别,从而提供情境化的信息。所得到的交互模式已被分析并与以前的方法进行了比较。最后,提出了可替代的、更具可扩展性的解决方案,这些解决方案保留了所描述系统的最有趣的特性。
{"title":"Communicating user's focus of attention by image processing as input for a mobile museum guide","authors":"A. Albertini, R. Brunelli, O. Stock, M. Zancanaro","doi":"10.1145/1040830.1040905","DOIUrl":"https://doi.org/10.1145/1040830.1040905","url":null,"abstract":"The paper presents a first prototype of a handheld museum guide delivering contextualized information based on the recognition of drawing details selected by the user through the guide camera. The resulting interaction modality has been analyzed and compared to previous approaches. Finally, alternative, more scalable, solutions are presented that preserve the most interesting features of the system described.","PeriodicalId":376409,"journal":{"name":"Proceedings of the 10th international conference on Intelligent user interfaces","volume":"17 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127288030","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 22
Designing interfaces for guided collection of knowledge about everyday objects from volunteers 为引导收集志愿者的日常物品知识设计界面
Pub Date : 2005-01-10 DOI: 10.1145/1040830.1040910
Timothy Chklovski
A new generation of intelligent applications can be enabled by broad-coverage knowledge repositories about everyday objects. We distill lessons in design of intelligent user interfaces which collect such broad-coverage knowledge from untrained volunteers. We motivate the knowledge-driven template-based approach adopted in Learner2, a second generation proactive acquisition interface for eliciting such knowledge. We present volume, accuracy, and recall of knowledge collected by fielding the system for 5 months. Learner2 has so far acquired 99,018 general statements, emphasizing knowledge about parts of and typical uses of objects.
新一代的智能应用程序可以通过广泛覆盖的日常物品知识库来实现。我们从未经训练的志愿者那里收集了如此广泛的知识,提炼出智能用户界面设计的经验教训。我们鼓励在Learner2中采用基于知识驱动模板的方法,这是第二代用于获取此类知识的主动获取界面。我们提出的数量,准确性和召回的知识收集的领域系统为5个月。到目前为止,Learner2已经掌握了99,018个一般语句,强调了关于物体的部分和典型用法的知识。
{"title":"Designing interfaces for guided collection of knowledge about everyday objects from volunteers","authors":"Timothy Chklovski","doi":"10.1145/1040830.1040910","DOIUrl":"https://doi.org/10.1145/1040830.1040910","url":null,"abstract":"A new generation of intelligent applications can be enabled by broad-coverage knowledge repositories about everyday objects. We distill lessons in design of intelligent user interfaces which collect such broad-coverage knowledge from untrained volunteers. We motivate the knowledge-driven template-based approach adopted in Learner2, a second generation proactive acquisition interface for eliciting such knowledge. We present volume, accuracy, and recall of knowledge collected by fielding the system for 5 months. Learner2 has so far acquired 99,018 general statements, emphasizing knowledge about parts of and typical uses of objects.","PeriodicalId":376409,"journal":{"name":"Proceedings of the 10th international conference on Intelligent user interfaces","volume":"155 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131072783","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Interaction techniques using prosodic features of speech and audio localization 使用语音和音频定位的韵律特征的交互技术
Pub Date : 2005-01-10 DOI: 10.1145/1040830.1040900
A. Olwal, Steven K. Feiner
We describe several approaches for using prosodic features of speech and audio localization to control interactive applications. This information can be applied to parameter control, as well as to speech disambiguation. We discuss how characteristics of spoken sentences can be exploited in the user interface; for example, by considering the speed with which a sentence is spoken and the presence of extraneous utterances. We also show how coarse audio localization can be used for low-fidelity gesture tracking, by inferring the speaker's head position.
我们描述了几种使用语音和音频定位的韵律特征来控制交互式应用程序的方法。这些信息可以应用于参数控制,以及语音消歧。我们讨论了如何在用户界面中利用口语句子的特征;例如,通过考虑一个句子的语速和外来话语的存在。我们还展示了粗糙的音频定位如何通过推断说话者的头部位置来用于低保真手势跟踪。
{"title":"Interaction techniques using prosodic features of speech and audio localization","authors":"A. Olwal, Steven K. Feiner","doi":"10.1145/1040830.1040900","DOIUrl":"https://doi.org/10.1145/1040830.1040900","url":null,"abstract":"We describe several approaches for using prosodic features of speech and audio localization to control interactive applications. This information can be applied to parameter control, as well as to speech disambiguation. We discuss how characteristics of spoken sentences can be exploited in the user interface; for example, by considering the speed with which a sentence is spoken and the presence of extraneous utterances. We also show how coarse audio localization can be used for low-fidelity gesture tracking, by inferring the speaker's head position.","PeriodicalId":376409,"journal":{"name":"Proceedings of the 10th international conference on Intelligent user interfaces","volume":"24 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121124557","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 24
期刊
Proceedings of the 10th international conference on Intelligent user interfaces
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1