首页 > 最新文献

2015 IEEE International Symposium on Multimedia (ISM)最新文献

英文 中文
Endoscopic Video Retrieval: A Signature-Based Approach for Linking Endoscopic Images with Video Segments 内窥镜视频检索:一种基于特征的内窥镜图像与视频片段链接方法
Pub Date : 2015-12-01 DOI: 10.1109/ISM.2015.21
C. Beecks, Klaus Schöffmann, M. Lux, M. S. Uysal, T. Seidl
In the field of medical endoscopy more and more surgeons are changing over to record and store videos of their endoscopic procedures, such as surgeries and examinations, in long-term video archives. In order to support surgeons in accessing these endoscopic video archives in a content-based way, we propose a simple yet effective signature-based approach: the Signature Matching Distance based on adaptive-binning feature signatures. The proposed distance-based similarity model facilitates an adaptive representation of the visual properties of endoscopic images and allows for matching these properties efficiently. We conduct an extensive performance analysis with respect to the task of linking specific endoscopic images with video segments and show the high efficacy of our approach. We are able to link more than 88% of the endoscopic images to their corresponding correct video segments, which improves the current state of the art by one order of magnitude.
在医学内窥镜领域,越来越多的外科医生开始将手术和检查等内窥镜过程的视频记录和存储在长期视频档案中。为了支持外科医生以基于内容的方式访问这些内窥镜视频档案,我们提出了一种简单而有效的基于签名的方法:基于自适应分组特征签名的签名匹配距离。所提出的基于距离的相似性模型促进了内窥镜图像视觉属性的自适应表示,并允许有效地匹配这些属性。我们对连接特定内窥镜图像与视频片段的任务进行了广泛的性能分析,并显示了我们方法的高效率。我们能够将超过88%的内窥镜图像链接到相应的正确视频片段,这将目前的技术水平提高了一个数量级。
{"title":"Endoscopic Video Retrieval: A Signature-Based Approach for Linking Endoscopic Images with Video Segments","authors":"C. Beecks, Klaus Schöffmann, M. Lux, M. S. Uysal, T. Seidl","doi":"10.1109/ISM.2015.21","DOIUrl":"https://doi.org/10.1109/ISM.2015.21","url":null,"abstract":"In the field of medical endoscopy more and more surgeons are changing over to record and store videos of their endoscopic procedures, such as surgeries and examinations, in long-term video archives. In order to support surgeons in accessing these endoscopic video archives in a content-based way, we propose a simple yet effective signature-based approach: the Signature Matching Distance based on adaptive-binning feature signatures. The proposed distance-based similarity model facilitates an adaptive representation of the visual properties of endoscopic images and allows for matching these properties efficiently. We conduct an extensive performance analysis with respect to the task of linking specific endoscopic images with video segments and show the high efficacy of our approach. We are able to link more than 88% of the endoscopic images to their corresponding correct video segments, which improves the current state of the art by one order of magnitude.","PeriodicalId":250353,"journal":{"name":"2015 IEEE International Symposium on Multimedia (ISM)","volume":"77 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125094710","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 21
An Unified Image Tagging System Driven by Image-Click-Ads Framework 基于图像-点击-广告框架的统一图像标签系统
Pub Date : 2015-12-01 DOI: 10.1109/ISM.2015.12
Qiong Wu, P. Boulanger
With the exponential growth of web image data, image tagging is becoming crucial in many image based applications such as object recognition and content-based image retrieval. Despite the great progress achieved in automatic recognition technologies, none has yet provided a satisfactory solution to be widely useful in solving generic image recognition problems. So far, only manual tagging can provide reliable tagging results. However, such work is tedious, costly and workers have no motivation. In this paper, we propose an online image tagging system, EyeDentifyIt, driven by image-click-ads framework, which motivates crowdsourcing workers as well as general web users to tag images at high quality for low cost with low workload. A series of usability studies are presented to demonstrate how EyeDentifyIt provides improved user motivations and requires less workload, compared to state-of-the-art approaches.
随着网络图像数据的指数级增长,图像标注在许多基于图像的应用中变得至关重要,例如物体识别和基于内容的图像检索。尽管自动识别技术取得了很大的进步,但还没有一种令人满意的解决方案能够广泛应用于解决一般的图像识别问题。到目前为止,只有手动标注才能提供可靠的标注结果。然而,这样的工作是乏味的,昂贵的,工人没有动力。在本文中,我们提出了一个在线图像标记系统,EyeDentifyIt,由图像-点击-广告框架驱动,激励众包工作者和一般网络用户以低成本、低工作量的方式标记高质量的图像。一系列的可用性研究展示了EyeDentifyIt如何提供更好的用户动机,并且与最先进的方法相比,需要更少的工作量。
{"title":"An Unified Image Tagging System Driven by Image-Click-Ads Framework","authors":"Qiong Wu, P. Boulanger","doi":"10.1109/ISM.2015.12","DOIUrl":"https://doi.org/10.1109/ISM.2015.12","url":null,"abstract":"With the exponential growth of web image data, image tagging is becoming crucial in many image based applications such as object recognition and content-based image retrieval. Despite the great progress achieved in automatic recognition technologies, none has yet provided a satisfactory solution to be widely useful in solving generic image recognition problems. So far, only manual tagging can provide reliable tagging results. However, such work is tedious, costly and workers have no motivation. In this paper, we propose an online image tagging system, EyeDentifyIt, driven by image-click-ads framework, which motivates crowdsourcing workers as well as general web users to tag images at high quality for low cost with low workload. A series of usability studies are presented to demonstrate how EyeDentifyIt provides improved user motivations and requires less workload, compared to state-of-the-art approaches.","PeriodicalId":250353,"journal":{"name":"2015 IEEE International Symposium on Multimedia (ISM)","volume":"83 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133660355","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Location Specification and Representation in Multimedia Databases 多媒体数据库中的位置规范与表示
Pub Date : 2015-12-01 DOI: 10.1109/ISM.2015.128
H. Samet
Techniques for the specification and representation of the locational component of multimedia data are reviewed. The focus is on how the locational component is specified and also on how it is represented. For the specification component we also discuss textual specifications. For the representation component, the emphasis is on a sorting approach which yields an index to the locational component where the data includes both points as well as objects with a spatial extent.
回顾了多媒体数据的位置组件的规范和表示技术。重点是如何指定位置组件,以及如何表示位置组件。对于规范组件,我们还讨论了文本规范。对于表示组件,重点是排序方法,该方法生成位置组件的索引,其中的数据既包括点,也包括具有空间范围的对象。
{"title":"Location Specification and Representation in Multimedia Databases","authors":"H. Samet","doi":"10.1109/ISM.2015.128","DOIUrl":"https://doi.org/10.1109/ISM.2015.128","url":null,"abstract":"Techniques for the specification and representation of the locational component of multimedia data are reviewed. The focus is on how the locational component is specified and also on how it is represented. For the specification component we also discuss textual specifications. For the representation component, the emphasis is on a sorting approach which yields an index to the locational component where the data includes both points as well as objects with a spatial extent.","PeriodicalId":250353,"journal":{"name":"2015 IEEE International Symposium on Multimedia (ISM)","volume":"89 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131232263","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
A User-Based Framework for Group Re-Identification in Still Images 基于用户的静态图像群体再识别框架
Pub Date : 2015-12-01 DOI: 10.1109/ISM.2015.41
Nestor Z. Salamon, Julio C. S. Jacques Junior, S. Musse
In this work we propose a framework for group re-identification based on manually defined soft-biometric characteristics. Users are able to choose colors that describe the soft-biometric attributes of each person belonging to the searched group. Our technique matches these structured attributes against image databases using color distance metrics, a novel adaptive threshold selection and people's proximity high level feature. Experimental results show that the proposed approach is able to help the re-identification procedure ranking the most likely results without training data, and also being extensible to work without previous images.
在这项工作中,我们提出了一个基于手动定义的软生物特征的群体再识别框架。用户可以选择描述属于搜索组的每个人的软生物特征属性的颜色。我们的技术使用颜色距离度量、一种新的自适应阈值选择和人的接近高水平特征将这些结构化属性与图像数据库相匹配。实验结果表明,该方法能够在没有训练数据的情况下对最可能的结果进行排序,并且可以扩展到没有先验图像的情况下进行再识别。
{"title":"A User-Based Framework for Group Re-Identification in Still Images","authors":"Nestor Z. Salamon, Julio C. S. Jacques Junior, S. Musse","doi":"10.1109/ISM.2015.41","DOIUrl":"https://doi.org/10.1109/ISM.2015.41","url":null,"abstract":"In this work we propose a framework for group re-identification based on manually defined soft-biometric characteristics. Users are able to choose colors that describe the soft-biometric attributes of each person belonging to the searched group. Our technique matches these structured attributes against image databases using color distance metrics, a novel adaptive threshold selection and people's proximity high level feature. Experimental results show that the proposed approach is able to help the re-identification procedure ranking the most likely results without training data, and also being extensible to work without previous images.","PeriodicalId":250353,"journal":{"name":"2015 IEEE International Symposium on Multimedia (ISM)","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131651283","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Design and Development of a Cloud Based Cyber-Physical Architecture for the Internet-of-Things 基于云的物联网网络物理架构的设计与开发
Pub Date : 2015-12-01 DOI: 10.1109/ISM.2015.96
K. M. Alam, Alex Sopena, Abdulmotaleb El Saddik
Internet-of-Things (IoT) is considered as the next big disruptive technology field which main goal is to achieve social good by enabling collaboration among physical things or sensors. We present a cloud based cyber-physical architecture to leverage the Sensing as-a-Service (SenAS) model, where every physical thing is complemented by a cloud based twin cyber process. In this model, things can communicate using direct physical connections or through the cyber layer using peer-to-peer inter process communications. The proposed model offers simultaneous communication channels among groups of things by uniquely tagging each group with a relationship ID. An intelligent service layer ensures custom privacy and access rights management for the sensor owners. We also present the implementation details of an IoT platform and demonstrate its practicality by developing case study applications for the Internet-of-Vehicles (IoV) and the connected smart home.
物联网(IoT)被认为是下一个大的颠覆性技术领域,其主要目标是通过实现物理事物或传感器之间的协作来实现社会公益。我们提出了一个基于云的网络物理架构,以利用传感即服务(SenAS)模型,其中每个物理事物都由基于云的双网络过程补充。在这个模型中,事物可以使用直接的物理连接或通过网络层使用点对点进程间通信进行通信。该模型通过对每组进行关系ID的唯一标记,提供了事物组之间的同步通信通道。智能服务层确保传感器所有者的自定义隐私和访问权限管理。我们还介绍了物联网平台的实施细节,并通过开发车联网(IoV)和联网智能家居的案例研究应用来展示其实用性。
{"title":"Design and Development of a Cloud Based Cyber-Physical Architecture for the Internet-of-Things","authors":"K. M. Alam, Alex Sopena, Abdulmotaleb El Saddik","doi":"10.1109/ISM.2015.96","DOIUrl":"https://doi.org/10.1109/ISM.2015.96","url":null,"abstract":"Internet-of-Things (IoT) is considered as the next big disruptive technology field which main goal is to achieve social good by enabling collaboration among physical things or sensors. We present a cloud based cyber-physical architecture to leverage the Sensing as-a-Service (SenAS) model, where every physical thing is complemented by a cloud based twin cyber process. In this model, things can communicate using direct physical connections or through the cyber layer using peer-to-peer inter process communications. The proposed model offers simultaneous communication channels among groups of things by uniquely tagging each group with a relationship ID. An intelligent service layer ensures custom privacy and access rights management for the sensor owners. We also present the implementation details of an IoT platform and demonstrate its practicality by developing case study applications for the Internet-of-Vehicles (IoV) and the connected smart home.","PeriodicalId":250353,"journal":{"name":"2015 IEEE International Symposium on Multimedia (ISM)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131654296","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 15
Reconstructing Missing Areas in Facial Images 人脸图像缺失区域的重建
Pub Date : 2015-12-01 DOI: 10.1109/ISM.2015.68
Christoph Jansen, Radek Mackowiak, N. Hezel, Moritz Ufer, Gregor Altstadt, K. U. Barthel
In this paper, we present a novel approach to reconstruct missing areas in facial images by using a series of Restricted Boltzman Machines (RBMs). RBMs created with a low number of hidden neurons generalize well and are able to reconstruct basic structures in the missing areas. On the other hand networks with many hidden neurons tend to emphasize details, when using the reconstruction of the previous, more generalized RBMs, as their input. Since trained RBMs are fast in encoding and decoding data by design, our method is also suitable for processing video streams.
本文提出了一种利用一系列受限玻尔兹曼机(rbm)重建人脸图像缺失区域的新方法。隐藏神经元数量少的rbm泛化效果好,能够重建缺失区域的基本结构。另一方面,当使用先前更广义的rbm的重建作为输入时,具有许多隐藏神经元的网络倾向于强调细节。由于经过训练的rbm在编码和解码数据的设计上是快速的,因此我们的方法也适用于视频流的处理。
{"title":"Reconstructing Missing Areas in Facial Images","authors":"Christoph Jansen, Radek Mackowiak, N. Hezel, Moritz Ufer, Gregor Altstadt, K. U. Barthel","doi":"10.1109/ISM.2015.68","DOIUrl":"https://doi.org/10.1109/ISM.2015.68","url":null,"abstract":"In this paper, we present a novel approach to reconstruct missing areas in facial images by using a series of Restricted Boltzman Machines (RBMs). RBMs created with a low number of hidden neurons generalize well and are able to reconstruct basic structures in the missing areas. On the other hand networks with many hidden neurons tend to emphasize details, when using the reconstruction of the previous, more generalized RBMs, as their input. Since trained RBMs are fast in encoding and decoding data by design, our method is also suitable for processing video streams.","PeriodicalId":250353,"journal":{"name":"2015 IEEE International Symposium on Multimedia (ISM)","volume":"7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131684597","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Human-Based Video Browsing - Investigating Interface Design for Fast Video Browsing 基于人的视频浏览——探索快速视频浏览的界面设计
Pub Date : 2015-12-01 DOI: 10.1109/ISM.2015.104
Wolfgang Hürst, R. V. D. Werken
The Video Browser Showdown (VBS) is an annual event where researchers evaluate their video search systems in a competitive setting. Searching in videos is often a two-step process: first some sort of pre-filtering is done, where, for example, users query an indexed archive of files, followed by a human-based browsing, where users skim the returned result set in search for the relevant file or portion of it. The VBS aims at this whole search process, focusing in particular on its interactive aspects. Encouraged by previous years' results, we created a system that purely addresses the latter issue, i.e., interface and interaction design. By eliminating all kind of video indexing and query processing, we were aiming to demonstrate the importance of good interface design for video search and that its relevance is often underestimated by today's systems. This claim is clearly proven by the results our system achieved in the VBS 2015 competition, where our approach was on a par with the top performing ones. In this paper, we will describe our system along with related design decisions, present our results from the VBS event, and discuss them in further detail.
视频浏览器对决(VBS)是一年一度的活动,研究人员在竞争环境中评估他们的视频搜索系统。在视频中搜索通常是一个两步的过程:首先进行某种预过滤,例如,用户查询索引的文件存档,然后是基于人的浏览,用户浏览返回的结果集以搜索相关文件或其中的一部分。VBS的目标是整个搜索过程,特别关注其互动方面。受前几年成果的鼓舞,我们创建了一个纯粹解决后一个问题的系统,即界面和交互设计。通过消除所有类型的视频索引和查询处理,我们旨在证明良好的视频搜索界面设计的重要性,以及它的相关性经常被当今的系统所低估。我们的系统在2015年VBS竞赛中取得的成绩清楚地证明了这一点,我们的方法与表现最好的方法不相上下。在本文中,我们将描述我们的系统以及相关的设计决策,展示我们从VBS事件中得到的结果,并进一步详细讨论它们。
{"title":"Human-Based Video Browsing - Investigating Interface Design for Fast Video Browsing","authors":"Wolfgang Hürst, R. V. D. Werken","doi":"10.1109/ISM.2015.104","DOIUrl":"https://doi.org/10.1109/ISM.2015.104","url":null,"abstract":"The Video Browser Showdown (VBS) is an annual event where researchers evaluate their video search systems in a competitive setting. Searching in videos is often a two-step process: first some sort of pre-filtering is done, where, for example, users query an indexed archive of files, followed by a human-based browsing, where users skim the returned result set in search for the relevant file or portion of it. The VBS aims at this whole search process, focusing in particular on its interactive aspects. Encouraged by previous years' results, we created a system that purely addresses the latter issue, i.e., interface and interaction design. By eliminating all kind of video indexing and query processing, we were aiming to demonstrate the importance of good interface design for video search and that its relevance is often underestimated by today's systems. This claim is clearly proven by the results our system achieved in the VBS 2015 competition, where our approach was on a par with the top performing ones. In this paper, we will describe our system along with related design decisions, present our results from the VBS event, and discuss them in further detail.","PeriodicalId":250353,"journal":{"name":"2015 IEEE International Symposium on Multimedia (ISM)","volume":"11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131982040","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Go Green with EnVI: the Energy-Video Index 与EnVI一起走向绿色:能源视频指数
Pub Date : 2015-12-01 DOI: 10.1109/ISM.2015.50
Oche Ejembi, S. Bhatti
Video is the most prevalent traffic type on the Internet today. Significant research has been done on measuring user's Quality of Experience (QoE) through different metrics. We take the position that energy use must be incorporated into quality metrics for digital video. We present our novel, energy-aware QoE metric for video, the Energy-Video Index (EnVI). We present our EnVI measurements from the playback of a diverse set of online videos. We observe that 4K-UHD (2160p) video can use ~30% more energy on a client device compared to HD (720p), and up to ~600% more network bandwidth than FHD (1080p), without significant improvement in objective QoE measurements.
视频是当今互联网上最流行的流量类型。对于用户体验质量(QoE)的度量方法,已有大量的研究。我们的立场是,能源使用必须纳入数字视频的质量指标。我们提出了我们的新颖的,能源意识的视频QoE度量,能源-视频指数(EnVI)。我们通过播放一组不同的在线视频来展示我们的EnVI测量值。我们观察到,与高清(720p)相比,4K-UHD (2160p)视频可以在客户端设备上多使用约30%的能量,比FHD (1080p)多使用约600%的网络带宽,而客观QoE测量没有显着改善。
{"title":"Go Green with EnVI: the Energy-Video Index","authors":"Oche Ejembi, S. Bhatti","doi":"10.1109/ISM.2015.50","DOIUrl":"https://doi.org/10.1109/ISM.2015.50","url":null,"abstract":"Video is the most prevalent traffic type on the Internet today. Significant research has been done on measuring user's Quality of Experience (QoE) through different metrics. We take the position that energy use must be incorporated into quality metrics for digital video. We present our novel, energy-aware QoE metric for video, the Energy-Video Index (EnVI). We present our EnVI measurements from the playback of a diverse set of online videos. We observe that 4K-UHD (2160p) video can use ~30% more energy on a client device compared to HD (720p), and up to ~600% more network bandwidth than FHD (1080p), without significant improvement in objective QoE measurements.","PeriodicalId":250353,"journal":{"name":"2015 IEEE International Symposium on Multimedia (ISM)","volume":"21 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116758338","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Portable Lecture Capture that Captures the Complete Lecture 便携式讲座捕获捕获完整的讲座
Pub Date : 2015-12-01 DOI: 10.1109/ISM.2015.22
P. Dickson, Chris Kondrat, Ryan B. Szeto, W. R. Adrion, Tung T. Pham, Tim D. Richards
Lecture recording is not a new concept nor is high-resolution recording of multimedia presentations that include computer and whiteboard material. We describe a novel portable lecture capture system that captures not only computer content and video as do most modern lecture capture systems but also captures content from whiteboards. The white-board material is captured at high resolution and processed for clarity without the necessity for the electronic whiteboards required by many capture systems. Our presentation system also processes the entire lecture in real time. The system we present is the logical next step in lecture capture technology.
讲座录音并不是一个新概念,包括电脑和白板材料在内的多媒体演示的高分辨率录音也不是一个新概念。我们描述了一种新型的便携式讲座捕捉系统,它不仅能像大多数现代讲座捕捉系统那样捕捉计算机内容和视频,还能捕捉白板上的内容。白板材料以高分辨率捕获并处理以获得清晰度,而不需要许多捕获系统所需的电子白板。我们的演示系统也实时处理整个讲座。我们提出的系统是讲座捕捉技术合乎逻辑的下一步。
{"title":"Portable Lecture Capture that Captures the Complete Lecture","authors":"P. Dickson, Chris Kondrat, Ryan B. Szeto, W. R. Adrion, Tung T. Pham, Tim D. Richards","doi":"10.1109/ISM.2015.22","DOIUrl":"https://doi.org/10.1109/ISM.2015.22","url":null,"abstract":"Lecture recording is not a new concept nor is high-resolution recording of multimedia presentations that include computer and whiteboard material. We describe a novel portable lecture capture system that captures not only computer content and video as do most modern lecture capture systems but also captures content from whiteboards. The white-board material is captured at high resolution and processed for clarity without the necessity for the electronic whiteboards required by many capture systems. Our presentation system also processes the entire lecture in real time. The system we present is the logical next step in lecture capture technology.","PeriodicalId":250353,"journal":{"name":"2015 IEEE International Symposium on Multimedia (ISM)","volume":"35 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115448066","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Evaluation of Feature Detection in HDR Based Imaging Under Changes in Illumination Conditions 光照条件变化下基于HDR成像的特征检测评价
Pub Date : 2015-12-01 DOI: 10.1109/ISM.2015.58
A. Rana, G. Valenzise, F. Dufaux
High dynamic range (HDR) imaging enables to capture details in both dark and very bright regions of a scene, and is therefore supposed to provide higher robustness to illumination changes than conventional low dynamic range (LDR) imaging in tasks such as visual features extraction. However, it is not clear how much this gain is, and which are the best modalities of using HDR to obtain it. In this paper we evaluate the first block of the visual feature extraction pipeline, i.e., keypoint detection, using both LDR and different HDR-based modalities, when significant illumination changes are present in the scene. To this end, we captured a dataset with two scenes and a wide range of illumination conditions. On these images, we measure how the repeatability of either corner or blob interest points is affected with different LDR/HDR approaches. Our observations confirm the potential of HDR over conventional LDR acquisition. Moreover, extracting features directly from HDR pixel values is more effective than first tonemapping and then extracting features, provided that HDR luminance information is previously encoded to perceptually linear values.
高动态范围(HDR)成像能够捕获场景中黑暗和非常明亮区域的细节,因此在视觉特征提取等任务中,比传统的低动态范围(LDR)成像提供更高的光照变化鲁棒性。然而,目前尚不清楚这种增益有多大,以及使用HDR获得增益的最佳方式是什么。在本文中,我们评估了视觉特征提取管道的第一块,即关键点检测,使用LDR和不同的基于hdr的模式,当场景中存在显著的照明变化时。为此,我们捕获了一个具有两个场景和广泛照明条件的数据集。在这些图像上,我们测量了不同的LDR/HDR方法对角点或斑点兴趣点的可重复性的影响。我们的观察证实了HDR相对于传统LDR获取的潜力。此外,如果HDR亮度信息事先编码为感知线性值,则直接从HDR像素值中提取特征比先进行色调映射再提取特征更有效。
{"title":"Evaluation of Feature Detection in HDR Based Imaging Under Changes in Illumination Conditions","authors":"A. Rana, G. Valenzise, F. Dufaux","doi":"10.1109/ISM.2015.58","DOIUrl":"https://doi.org/10.1109/ISM.2015.58","url":null,"abstract":"High dynamic range (HDR) imaging enables to capture details in both dark and very bright regions of a scene, and is therefore supposed to provide higher robustness to illumination changes than conventional low dynamic range (LDR) imaging in tasks such as visual features extraction. However, it is not clear how much this gain is, and which are the best modalities of using HDR to obtain it. In this paper we evaluate the first block of the visual feature extraction pipeline, i.e., keypoint detection, using both LDR and different HDR-based modalities, when significant illumination changes are present in the scene. To this end, we captured a dataset with two scenes and a wide range of illumination conditions. On these images, we measure how the repeatability of either corner or blob interest points is affected with different LDR/HDR approaches. Our observations confirm the potential of HDR over conventional LDR acquisition. Moreover, extracting features directly from HDR pixel values is more effective than first tonemapping and then extracting features, provided that HDR luminance information is previously encoded to perceptually linear values.","PeriodicalId":250353,"journal":{"name":"2015 IEEE International Symposium on Multimedia (ISM)","volume":"15 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114292006","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 27
期刊
2015 IEEE International Symposium on Multimedia (ISM)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1