首页 > 最新文献

MULTIMEDIA '04最新文献

英文 中文
Context for semantic metadata 语义元数据的上下文
Pub Date : 2004-10-10 DOI: 10.1145/1027527.1027574
K. Haase
This article argues for the growing importance of quality metadata and the equation of that quality with precision and semantic grounding. Such semantic grounding requires metadata that derives from intentional human intervention as well as mechanistic measurement of content media. In both cases, one chief problem in the automatic generation of semantic metadata is ambiguity leading to the overgeneration of inaccurate annotations. We look at a particular richly annotated image collection to show how context dramatically reduces the problem of ambiguity over this particular corpus. In particular, we consider both the abstract measurement of "contextual ambiguity" over the collection and the application of a particular disambiguation algorithm to synthesized keyword searches across the selection.
本文论证了高质量元数据日益增长的重要性,以及这种质量与精度和语义基础的关系。这种语义基础需要源自有意的人为干预以及内容媒体的机械测量的元数据。在这两种情况下,自动生成语义元数据的一个主要问题是歧义性导致过度生成不准确的注释。我们看一个特殊的丰富注释的图像集合,以显示上下文如何显着减少这个特定语料库的歧义问题。特别是,我们考虑了集合上“上下文歧义”的抽象测量和特定消歧义算法在整个选择中合成关键字搜索的应用。
{"title":"Context for semantic metadata","authors":"K. Haase","doi":"10.1145/1027527.1027574","DOIUrl":"https://doi.org/10.1145/1027527.1027574","url":null,"abstract":"This article argues for the growing importance of quality metadata and the equation of that quality with precision and semantic grounding. Such semantic grounding requires metadata that derives from intentional human intervention as well as mechanistic measurement of content media. In both cases, one chief problem in the automatic generation of semantic metadata is ambiguity leading to the overgeneration of inaccurate annotations. We look at a particular richly annotated image collection to show how context dramatically reduces the problem of ambiguity over this particular corpus. In particular, we consider both the abstract measurement of \"contextual ambiguity\" over the collection and the application of a particular disambiguation algorithm to synthesized keyword searches across the selection.","PeriodicalId":292207,"journal":{"name":"MULTIMEDIA '04","volume":"361 11","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2004-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"113956197","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 51
Video transport over wireless networks 通过无线网络传输视频
Pub Date : 2004-10-10 DOI: 10.1145/1027527.1027626
H. Garudadri, P. Sagetong, S. Nanda
In this paper, we propose an efficient scheme to transport video over wireless networks, specifically cdma2000® 1x. Speech transmission over cdma2000® uses a variable rate voice coder (vocoder) over a channel with multiple fixed rates. We apply these ideas to compressed video transmission over wireless IP networks. Explicit Bit Rate (EBR) video compression is designed to match the video encoder output to a set of fixed channel rates. We show that in comparison with VBR video transmission over a fixed rate wireless channel, EBR video transmission provides improved error resilience, reduced latency and improved efficiency.
在本文中,我们提出了一种通过无线网络传输视频的有效方案,特别是cdma2000®1x。cdma2000®上的语音传输使用可变速率语音编码器(vocoder)在多个固定速率的信道上传输。我们将这些思想应用于无线IP网络上的压缩视频传输。显式比特率(EBR)视频压缩是为了将视频编码器输出匹配到一组固定的信道速率。研究表明,与固定速率无线信道上的VBR视频传输相比,EBR视频传输具有更好的容错能力、更低的延迟和更高的效率。
{"title":"Video transport over wireless networks","authors":"H. Garudadri, P. Sagetong, S. Nanda","doi":"10.1145/1027527.1027626","DOIUrl":"https://doi.org/10.1145/1027527.1027626","url":null,"abstract":"In this paper, we propose an efficient scheme to transport video over wireless networks, specifically cdma2000® 1x. Speech transmission over cdma2000® uses a variable rate voice coder (vocoder) over a channel with multiple fixed rates. We apply these ideas to compressed video transmission over wireless IP networks. Explicit Bit Rate (EBR) video compression is designed to match the video encoder output to a set of fixed channel rates. We show that in comparison with VBR video transmission over a fixed rate wireless channel, EBR video transmission provides improved error resilience, reduced latency and improved efficiency.","PeriodicalId":292207,"journal":{"name":"MULTIMEDIA '04","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2004-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125177922","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Parsing and browsing tools for colonoscopy videos 结肠镜检查视频的解析和浏览工具
Pub Date : 2004-10-10 DOI: 10.1145/1027527.1027723
Yu Cao, Dalei Li, Wallapak Tavanapong, Jung-Hwan Oh, J. Wong, P. C. Groen
Colonoscopy is an important screening tool for colorectal cancer. During a colonoscopic procedure, a tiny video camera at the tip of the endoscope generates a video signal of the internal mucosa of the colon. The video data are displayed on a monitor for real-time analysis by the endoscopist. We call videos captured from colonoscopic procedures colonoscopy videos. Because these videos possess unique characteristics, new types of semantic units and parsing techniques are required. In this paper, we define new semantic units called operation shots, each is a segment of visual and audio data that correspond to a therapeutic or biopsy operation. We introduce a new spatio-temporal analysis technique to detect operation shots. Our experiments on colonoscopy videos demonstrate that the technique does not miss any meaningful operation shots and incurs a small number of false operation shots. Our prototype parsing software implements the operation shot detection technique along with our other techniques previously developed for colonoscopy videos. Our browsing tool enables users to quickly locate operation shots of interest. The proposed technique and software are useful (1) for post-procedure reviews and analyses for causes of complications due to biopsy or therapeutic operations, (2) for developing an effective content-based retrieval system for colonoscopy videos to facilitate endoscopic research and education, and (3) for development of a systematic approach to assess endoscopists' procedural skills.
结肠镜检查是结直肠癌的重要筛查手段。在结肠镜检查过程中,内窥镜尖端的微型摄像机产生结肠内部粘膜的视频信号。视频数据显示在监视器上,供内窥镜医师实时分析。我们把从结肠镜检查过程中捕获的视频称为结肠镜检查视频。由于这些视频具有独特的特征,因此需要新的语义单元类型和解析技术。在本文中,我们定义了新的语义单位,称为手术镜头,每个是一段视觉和音频数据,对应于治疗或活检手术。本文介绍了一种新的时空分析技术来检测手术镜头。我们对结肠镜视频的实验表明,该技术不会遗漏任何有意义的手术镜头,并且会产生少量的假手术镜头。我们的原型解析软件实现了操作镜头检测技术以及我们以前为结肠镜检查视频开发的其他技术。我们的浏览工具使用户能够快速定位感兴趣的操作镜头。所提出的技术和软件是有用的(1)用于术后检查和分析活检或治疗性手术引起的并发症的原因,(2)用于开发有效的基于内容的结肠镜检查视频检索系统,以促进内窥镜研究和教育,(3)用于开发评估内窥镜医师操作技能的系统方法。
{"title":"Parsing and browsing tools for colonoscopy videos","authors":"Yu Cao, Dalei Li, Wallapak Tavanapong, Jung-Hwan Oh, J. Wong, P. C. Groen","doi":"10.1145/1027527.1027723","DOIUrl":"https://doi.org/10.1145/1027527.1027723","url":null,"abstract":"Colonoscopy is an important screening tool for colorectal cancer. During a colonoscopic procedure, a tiny video camera at the tip of the endoscope generates a video signal of the internal mucosa of the colon. The video data are displayed on a monitor for real-time analysis by the endoscopist. We call videos captured from colonoscopic procedures <i>colonoscopy videos</i>. Because these videos possess unique characteristics, new types of semantic units and parsing techniques are required. In this paper, we define new semantic units called <i>operation shots</i>, each is a segment of visual and audio data that correspond to a therapeutic or biopsy operation. We introduce a new spatio-temporal analysis technique to detect operation shots. Our experiments on colonoscopy videos demonstrate that the technique does not miss any meaningful operation shots and incurs a small number of false operation shots. Our prototype parsing software implements the operation shot detection technique along with our other techniques previously developed for colonoscopy videos. Our browsing tool enables users to quickly locate operation shots of interest. The proposed technique and software are useful (1) for post-procedure reviews and analyses for causes of complications due to biopsy or therapeutic operations, (2) for developing an effective content-based retrieval system for colonoscopy videos to facilitate endoscopic research and education, and (3) for development of a systematic approach to assess endoscopists' procedural skills.","PeriodicalId":292207,"journal":{"name":"MULTIMEDIA '04","volume":"40 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2004-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125100538","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 37
Facial expression representation and recognition based on texture augmentation and topographic masking 基于纹理增强和地形掩蔽的面部表情表示与识别
Pub Date : 2004-10-10 DOI: 10.1145/1027527.1027580
L. Yin, J. Loi, Wei Xiong
The variation of facial texture and surface due to the change of expression is an important cue for analyzing and modeling facial expressions. In this paper, we propose a new approach to represent the facial expression by using a so-called topographic feature. In order to capture the variation of facial surface structure, facial textures are processed by increasing the resolution. The topographical structure of human face is analyzed based on the resolution-enhanced textures. We investigate the relationship between the facial expression and its topographic features, and propose to represent the facial expression by the topographic labels. The detected topographic facial surface and the expressive regions reflect the status of facial skin movement. Based on the observation that the facial texture and its topographic features change along with facial expressions, we compare the disparity of these features between the neutral face and the expressive face to distinguish a number of universal expressions. The experiment demonstrates the feasibility of the proposed approach for facial expression representation and recognition.
表情变化引起的面部纹理和表面变化是分析和建模面部表情的重要线索。在本文中,我们提出了一种新的方法,即使用所谓的地形特征来表示面部表情。为了捕捉人脸表面结构的变化,通过提高分辨率对人脸纹理进行处理。基于分辨率增强纹理对人脸的地形结构进行分析。研究了面部表情与其地形特征之间的关系,提出了用地形标签来表示面部表情的方法。检测到的地形面部表面和表情区域反映了面部皮肤运动的状态。在观察面部纹理及其地形特征随面部表情变化的基础上,我们比较了中性脸和表情脸在这些特征上的差异,以区分一些通用表情。实验证明了该方法在面部表情表征和识别方面的可行性。
{"title":"Facial expression representation and recognition based on texture augmentation and topographic masking","authors":"L. Yin, J. Loi, Wei Xiong","doi":"10.1145/1027527.1027580","DOIUrl":"https://doi.org/10.1145/1027527.1027580","url":null,"abstract":"The variation of facial texture and surface due to the change of expression is an important cue for analyzing and modeling facial expressions. In this paper, we propose a new approach to represent the facial expression by using a so-called topographic feature. In order to capture the variation of facial surface structure, facial textures are processed by increasing the resolution. The topographical structure of human face is analyzed based on the resolution-enhanced textures. We investigate the relationship between the facial expression and its topographic features, and propose to represent the facial expression by the topographic labels. The detected topographic facial surface and the expressive regions reflect the status of facial skin movement. Based on the observation that the facial texture and its topographic features change along with facial expressions, we compare the disparity of these features between the neutral face and the expressive face to distinguish a number of universal expressions. The experiment demonstrates the feasibility of the proposed approach for facial expression representation and recognition.","PeriodicalId":292207,"journal":{"name":"MULTIMEDIA '04","volume":"238 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2004-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131449945","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Location-aware projection with robust 3-D viewing point detection and fast image deformation 位置感知投影,具有鲁棒的三维观察点检测和快速图像变形
Pub Date : 2004-10-10 DOI: 10.1145/1027527.1027595
J. Shimamura, K. Arakawa
This paper describes a novel approach to the construction of a projector-based augmented reality environment. The approach is based on capturing the dynamic changes of surfaces and projecting the images within a large real environment using a system that includes a laser range finder and a projector, whose optical axes are integrated by mirrors. The proposed method offers two distinct advances: (1) robust 3-D viewing point detection from consecutive range images, and (2) fast view-driven image generation and presentation with view frustum clipping to measured surfaces. A prototype system is shown to confirm the feasibility of the method; it generates view-driven images to suit the user's viewing position that are then projected within dynamic real environment, in real-time.
本文描述了一种构建基于投影仪的增强现实环境的新方法。该方法基于捕捉表面的动态变化,并使用包括激光测距仪和投影仪在内的系统在大型真实环境中投射图像,其光轴由镜子集成。该方法提供了两个明显的进步:(1)从连续距离图像中进行鲁棒的三维观察点检测;(2)快速的视觉驱动图像生成和显示,并将视场截距裁剪到被测量表面。仿真结果验证了该方法的可行性;它生成适合用户观看位置的视图驱动图像,然后在动态真实环境中实时投影。
{"title":"Location-aware projection with robust 3-D viewing point detection and fast image deformation","authors":"J. Shimamura, K. Arakawa","doi":"10.1145/1027527.1027595","DOIUrl":"https://doi.org/10.1145/1027527.1027595","url":null,"abstract":"This paper describes a novel approach to the construction of a projector-based augmented reality environment. The approach is based on capturing the dynamic changes of surfaces and projecting the images within a large real environment using a system that includes a laser range finder and a projector, whose optical axes are integrated by mirrors. The proposed method offers two distinct advances: (1) robust 3-D viewing point detection from consecutive range images, and (2) fast view-driven image generation and presentation with view frustum clipping to measured surfaces. A prototype system is shown to confirm the feasibility of the method; it generates view-driven images to suit the user's viewing position that are then projected within dynamic real environment, in real-time.","PeriodicalId":292207,"journal":{"name":"MULTIMEDIA '04","volume":"204 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2004-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134308462","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
An EPIC enhanced meeting environment EPIC增强的会议环境
Pub Date : 2004-10-10 DOI: 10.1145/1027527.1027743
Qiong Liu, F. Zhao, John Doherty, Don Kimber
ePic is an integrated presentation authoring and playback system that makes it easy to use a wide range of devices installed in one or multiple multimedia venues.
ePic是一个集成的演示文稿创作和播放系统,可以轻松使用安装在一个或多个多媒体场所的各种设备。
{"title":"An EPIC enhanced meeting environment","authors":"Qiong Liu, F. Zhao, John Doherty, Don Kimber","doi":"10.1145/1027527.1027743","DOIUrl":"https://doi.org/10.1145/1027527.1027743","url":null,"abstract":"ePic is an integrated presentation authoring and playback system that makes it easy to use a wide range of devices installed in one or multiple multimedia venues.","PeriodicalId":292207,"journal":{"name":"MULTIMEDIA '04","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2004-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131486167","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Multi-model similarity propagation and its application for web image retrieval 多模型相似性传播及其在网络图像检索中的应用
Pub Date : 2004-10-10 DOI: 10.1145/1027527.1027746
Xin-Jing Wang, Wei-Ying Ma, Gui-Rong Xue, Xing Li
In this paper, we propose an iterative similarity propagation approach to explore the inter-relationships between Web images and their textual annotations for image retrieval. By considering Web images as one type of objects, their surrounding texts as another type, and constructing the links structure between them via webpage analysis, we can iteratively reinforce the similarities between images. The basic idea is that if two objects of the same type are both related to one object of another type, these two objects are similar; likewise, if two objects of the same type are related to two different, but similar objects of another type, then to some extent, these two objects are also similar. The goal of our method is to fully exploit the mutual reinforcement between images and their textual annotations. Our experiments based on 10,628 images crawled from the Web show that our proposed approach can significantly improve Web image retrieval performance.
在本文中,我们提出了一种迭代相似传播方法来探索Web图像及其文本注释之间的相互关系,用于图像检索。将网络图像视为一种对象,将其周围的文本视为另一种对象,通过网页分析构建它们之间的链接结构,可以迭代强化图像之间的相似性。基本思想是,如果同一类型的两个对象都与另一类型的一个对象相关,则这两个对象是相似的;同样,如果同一类型的两个对象与另一类型的两个不同但相似的对象相关联,那么这两个对象在某种程度上也是相似的。我们的方法的目标是充分利用图像和文本注释之间的相互强化。基于从Web上抓取的10,628张图像的实验表明,我们提出的方法可以显着提高Web图像检索性能。
{"title":"Multi-model similarity propagation and its application for web image retrieval","authors":"Xin-Jing Wang, Wei-Ying Ma, Gui-Rong Xue, Xing Li","doi":"10.1145/1027527.1027746","DOIUrl":"https://doi.org/10.1145/1027527.1027746","url":null,"abstract":"In this paper, we propose an iterative similarity propagation approach to explore the inter-relationships between Web images and their textual annotations for image retrieval. By considering Web images as one type of objects, their surrounding texts as another type, and constructing the links structure between them via webpage analysis, we can iteratively reinforce the similarities between images. The basic idea is that if two objects of the same type are both related to one object of another type, these two objects are similar; likewise, if two objects of the same type are related to two different, but similar objects of another type, then to some extent, these two objects are also similar. The goal of our method is to fully exploit the mutual reinforcement between images and their textual annotations. Our experiments based on 10,628 images crawled from the Web show that our proposed approach can significantly improve Web image retrieval performance.","PeriodicalId":292207,"journal":{"name":"MULTIMEDIA '04","volume":"19 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2004-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134527011","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 97
Interactive retrieval of 3D shape models using physical objects 使用物理对象的三维形状模型的交互式检索
Pub Date : 2004-10-10 DOI: 10.1145/1027527.1027685
Hiroyasu Ichida, Yuichi Itoh, Y. Kitamura, F. Kishino
We present a novel method for interactive retrieval of 3D shapes using ysical objects. Our method is based on simple ysical 3D interaction with a set of tangible blocks. As the user connects blocks, the system automatically recognizes the shape of the constructed ysical structure and picks similar 3D shape models from a preset model database, in real time. Our system fully supports interactive retrieval of 3D shape models in an extremely simple fashion, which is completely non-verbal and cross-cultural. These advantages make it an ideal interface for inexperienced users, previously barred from many applications that include 3D shape retrieval tasks.
我们提出了一种利用物理对象进行三维形状交互检索的新方法。我们的方法是基于简单的物理三维交互与一组有形的块。当用户连接块时,系统自动识别所构建的物理结构的形状,并实时从预设的模型数据库中选择相似的3D形状模型。我们的系统以一种非常简单的方式完全支持三维形状模型的交互式检索,这是完全非语言和跨文化的。这些优点使其成为经验不足的用户的理想界面,以前禁止许多应用程序,包括3D形状检索任务。
{"title":"Interactive retrieval of 3D shape models using physical objects","authors":"Hiroyasu Ichida, Yuichi Itoh, Y. Kitamura, F. Kishino","doi":"10.1145/1027527.1027685","DOIUrl":"https://doi.org/10.1145/1027527.1027685","url":null,"abstract":"We present a novel method for interactive retrieval of 3D shapes using ysical objects. Our method is based on simple ysical 3D interaction with a set of tangible blocks. As the user connects blocks, the system automatically recognizes the shape of the constructed ysical structure and picks similar 3D shape models from a preset model database, in real time. Our system fully supports interactive retrieval of 3D shape models in an extremely simple fashion, which is completely non-verbal and cross-cultural. These advantages make it an ideal interface for inexperienced users, previously barred from many applications that include 3D shape retrieval tasks.","PeriodicalId":292207,"journal":{"name":"MULTIMEDIA '04","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2004-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133506020","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 28
User-assisted tools for concurrency control in distributed multimedia collaborations 分布式多媒体协作中用于并发控制的用户辅助工具
Pub Date : 2004-10-10 DOI: 10.1145/1027527.1027652
A. Sabbir, K. Ravindran
In a distributed collaborative application, a key requirement is that all users see the same copy of a shared window object at any given point in time (WYSIWIS). In this paper, we study 'user-assisted causal ordering' of messages as the basis for achieving WYSIWIS. The approach requires specifying the synchronization constraints on accessing shared window objects in the form of an order in which messages need to be processed and object state updated. The specifications are made available to the window subsystem based on the user-level knowledge about the actions on objects and the current (shared) object state. In contrast with the current approaches employing transaction models, our approach allows flexibility in the programming of collaboration-style applications, and offers increased levels of concurrency.
在分布式协作应用程序中,一个关键需求是所有用户在任何给定时间点都能看到共享窗口对象的相同副本(所见即所得)。在本文中,我们研究了消息的“用户辅助因果排序”,作为实现WYSIWIS的基础。该方法需要以需要处理消息和更新对象状态的顺序的形式指定访问共享窗口对象的同步约束。这些规范是基于关于对象操作和当前(共享)对象状态的用户级知识提供给窗口子系统的。与使用事务模型的当前方法相比,我们的方法允许协作式应用程序编程的灵活性,并提供更高级别的并发性。
{"title":"User-assisted tools for concurrency control in distributed multimedia collaborations","authors":"A. Sabbir, K. Ravindran","doi":"10.1145/1027527.1027652","DOIUrl":"https://doi.org/10.1145/1027527.1027652","url":null,"abstract":"In a distributed collaborative application, a key requirement is that all users see the same copy of a shared window object at any given point in time (WYSIWIS). In this paper, we study 'user-assisted causal ordering' of messages as the basis for achieving WYSIWIS. The approach requires specifying the synchronization constraints on accessing shared window objects in the form of an order in which messages need to be processed and object state updated. The specifications are made available to the window subsystem based on the user-level knowledge about the actions on objects and the current (shared) object state. In contrast with the current approaches employing transaction models, our approach allows flexibility in the programming of collaboration-style applications, and offers increased levels of concurrency.","PeriodicalId":292207,"journal":{"name":"MULTIMEDIA '04","volume":"219 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2004-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133709607","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Grouping web image search result 分组网络图像搜索结果
Pub Date : 2004-10-10 DOI: 10.1145/1027527.1027632
Xin-Jing Wang, Wei-Ying Ma, Qi-Cai He, Xing Li
In this paper, we propose a Web image search result organizing method to facilitate user browsing. We formalize this problem as a salient image region pattern extraction problem. Given the images returned by Web search engine, we first segment the images into homogeneous regions and quantize the environmental regions into image codewords. The salient codeword "phrases" are then extracted and ranked based on a regression model learned from human labeled training data. According to the salient "phrases", images are assigned to different clusters, with the one nearest to the centroid as the entry for the corresponding cluster. Satisfying experimental results show the effectiveness of our proposed method.
本文提出了一种便于用户浏览的Web图像搜索结果组织方法。我们将此问题形式化为显著图像区域模式提取问题。针对网络搜索引擎返回的图像,首先将图像分割为均匀区域,并将环境区域量化为图像码字。然后根据从人类标记的训练数据中学习到的回归模型提取突出的码字“短语”并进行排名。根据突出的“短语”,将图像分配到不同的聚类中,最接近质心的图像作为相应聚类的入口。实验结果表明了该方法的有效性。
{"title":"Grouping web image search result","authors":"Xin-Jing Wang, Wei-Ying Ma, Qi-Cai He, Xing Li","doi":"10.1145/1027527.1027632","DOIUrl":"https://doi.org/10.1145/1027527.1027632","url":null,"abstract":"In this paper, we propose a Web image search result organizing method to facilitate user browsing. We formalize this problem as a salient image region pattern extraction problem. Given the images returned by Web search engine, we first segment the images into homogeneous regions and quantize the environmental regions into image codewords. The salient codeword \"phrases\" are then extracted and ranked based on a regression model learned from human labeled training data. According to the salient \"phrases\", images are assigned to different clusters, with the one nearest to the centroid as the entry for the corresponding cluster. Satisfying experimental results show the effectiveness of our proposed method.","PeriodicalId":292207,"journal":{"name":"MULTIMEDIA '04","volume":"5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2004-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133726025","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 30
期刊
MULTIMEDIA '04
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1