首页 > 最新文献

Proceedings of the 8th ACM on Multimedia Systems Conference最新文献

英文 中文
Localization of Acupoints using Augmented Reality 利用增强现实定位穴位
Pub Date : 2017-06-20 DOI: 10.1145/3083187.3083225
Yi-Zhang Chen, Corky Maigre, Min-Chun Hu, Kun-Chan Lan
An augmented reality (AR) system for acupuncture points localization is implemented on an Android smartphone. The user can utilize such a system to locate the relevant acupuncture point for the purpose of symptom relief (e.g. through acupressure).
在Android智能手机上实现了一种用于穴位定位的增强现实(AR)系统。用户可以利用这样的系统来定位相关的穴位以缓解症状(例如通过穴位按压)。
{"title":"Localization of Acupoints using Augmented Reality","authors":"Yi-Zhang Chen, Corky Maigre, Min-Chun Hu, Kun-Chan Lan","doi":"10.1145/3083187.3083225","DOIUrl":"https://doi.org/10.1145/3083187.3083225","url":null,"abstract":"An augmented reality (AR) system for acupuncture points localization is implemented on an Android smartphone. The user can utilize such a system to locate the relevant acupuncture point for the purpose of symptom relief (e.g. through acupressure).","PeriodicalId":123321,"journal":{"name":"Proceedings of the 8th ACM on Multimedia Systems Conference","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127236902","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Real Time Stable Haptic Rendering Of 3D Deformable Streaming Surface 三维可变形流表面的实时稳定触觉渲染
Pub Date : 2017-06-20 DOI: 10.1145/3083187.3083198
Yuan Tian, Chao Li, X. Guo, B. Prabhakaran
In recent years, many researches are focusing on the haptic interaction with streaming data like RGBD video / point cloud stream captured by commodity depth sensors. Most previous methods use partial streaming data from depth sensors and only investigate haptic rendering of the rigid surface without complex physics simulation. Many virtual reality and tele-immersive applications such as medical training, and art designing require the complete scene and physics simulation. In this paper, we propose a stable haptic rendering method capable of interacting with streaming deformable surface in real-time. Our method applies KinectFusion for real-time reconstruction of real-world object surface instead of incomplete surface. While construction, it simultaneously uses hierarchical shape matching (HSM) method to simulate the surface deformation in haptic-enabled interaction. We have demonstrated how to combine the fusion and physics simulation of deformation together, and proposed a continuous collision detection method based on Truncated Signed Distance Function (TSDF). Furthermore, we propose a fast TSDF warping method to update the deformation to TSDF, and a proxy finding method to find the proxy position. The proposed method is able to simulate the haptic-enabled deformation of the 3D fusion surface. Therefore it provides a novel haptic interaction for virtual reality and 3D tele-immersive applications. Experimental results show that the proposed approach provides stable haptic rendering and fast simulation of 3D deformable surface.
近年来,许多研究都集中在与流数据的触觉交互上,如由商品深度传感器捕获的RGBD视频/点云流。大多数以前的方法使用来自深度传感器的部分流数据,并且只研究刚性表面的触觉渲染,而没有复杂的物理模拟。许多虚拟现实和远程沉浸式应用,如医疗培训和艺术设计,需要完整的场景和物理模拟。在本文中,我们提出了一种稳定的触觉渲染方法,能够实时与流变形表面交互。该方法将KinectFusion应用于真实物体表面的实时重建,而不是不完整的表面。在构建过程中,同时采用层次形状匹配(HSM)方法模拟触觉交互过程中的表面变形。我们演示了如何将变形的融合和物理模拟结合在一起,并提出了一种基于截断符号距离函数(TSDF)的连续碰撞检测方法。此外,我们提出了一种快速的TSDF翘曲方法来更新变形到TSDF,并提出了一种代理查找方法来查找代理位置。该方法能够模拟三维融合表面的触觉变形。因此,它为虚拟现实和三维远程沉浸式应用提供了一种新的触觉交互方式。实验结果表明,该方法能够提供稳定的触觉渲染和快速的三维变形表面仿真。
{"title":"Real Time Stable Haptic Rendering Of 3D Deformable Streaming Surface","authors":"Yuan Tian, Chao Li, X. Guo, B. Prabhakaran","doi":"10.1145/3083187.3083198","DOIUrl":"https://doi.org/10.1145/3083187.3083198","url":null,"abstract":"In recent years, many researches are focusing on the haptic interaction with streaming data like RGBD video / point cloud stream captured by commodity depth sensors. Most previous methods use partial streaming data from depth sensors and only investigate haptic rendering of the rigid surface without complex physics simulation. Many virtual reality and tele-immersive applications such as medical training, and art designing require the complete scene and physics simulation. In this paper, we propose a stable haptic rendering method capable of interacting with streaming deformable surface in real-time. Our method applies KinectFusion for real-time reconstruction of real-world object surface instead of incomplete surface. While construction, it simultaneously uses hierarchical shape matching (HSM) method to simulate the surface deformation in haptic-enabled interaction. We have demonstrated how to combine the fusion and physics simulation of deformation together, and proposed a continuous collision detection method based on Truncated Signed Distance Function (TSDF). Furthermore, we propose a fast TSDF warping method to update the deformation to TSDF, and a proxy finding method to find the proxy position. The proposed method is able to simulate the haptic-enabled deformation of the 3D fusion surface. Therefore it provides a novel haptic interaction for virtual reality and 3D tele-immersive applications. Experimental results show that the proposed approach provides stable haptic rendering and fast simulation of 3D deformable surface.","PeriodicalId":123321,"journal":{"name":"Proceedings of the 8th ACM on Multimedia Systems Conference","volume":"17 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126842105","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
KVASIR: A Multi-Class Image Dataset for Computer Aided Gastrointestinal Disease Detection KVASIR:用于计算机辅助胃肠疾病检测的多类图像数据集
Pub Date : 2017-06-20 DOI: 10.1145/3083187.3083212
Konstantin Pogorelov, K. Randel, C. Griwodz, S. Eskeland, T. Lange, Dag Johansen, C. Spampinato, Duc-Tien Dang-Nguyen, M. Lux, P. Schmidt, M. Riegler, P. Halvorsen
Automatic detection of diseases by use of computers is an important, but still unexplored field of research. Such innovations may improve medical practice and refine health care systems all over the world. However, datasets containing medical images are hardly available, making reproducibility and comparison of approaches almost impossible. In this paper, we present KVASIR, a dataset containing images from inside the gastrointestinal (GI) tract. The collection of images are classified into three important anatomical landmarks and three clinically significant findings. In addition, it contains two categories of images related to endoscopic polyp removal. Sorting and annotation of the dataset is performed by medical doctors (experienced endoscopists). In this respect, KVASIR is important for research on both single- and multi-disease computer aided detection. By providing it, we invite and enable multimedia researcher into the medical domain of detection and retrieval.
利用计算机自动检测疾病是一个重要但尚未开发的研究领域。这些创新可能会改善全世界的医疗实践和完善卫生保健系统。然而,包含医学图像的数据集很难获得,使得方法的可重复性和比较几乎不可能。在本文中,我们提出了KVASIR,这是一个包含胃肠道内部图像的数据集。收集的图像分为三个重要的解剖标志和三个具有临床意义的发现。此外,它还包含两类与内镜息肉切除相关的图像。数据集的排序和注释由医生(经验丰富的内窥镜医师)执行。在这方面,KVASIR对于单疾病和多疾病计算机辅助检测的研究都是重要的。通过提供它,我们邀请并使多媒体研究人员进入医学检测和检索领域。
{"title":"KVASIR: A Multi-Class Image Dataset for Computer Aided Gastrointestinal Disease Detection","authors":"Konstantin Pogorelov, K. Randel, C. Griwodz, S. Eskeland, T. Lange, Dag Johansen, C. Spampinato, Duc-Tien Dang-Nguyen, M. Lux, P. Schmidt, M. Riegler, P. Halvorsen","doi":"10.1145/3083187.3083212","DOIUrl":"https://doi.org/10.1145/3083187.3083212","url":null,"abstract":"Automatic detection of diseases by use of computers is an important, but still unexplored field of research. Such innovations may improve medical practice and refine health care systems all over the world. However, datasets containing medical images are hardly available, making reproducibility and comparison of approaches almost impossible. In this paper, we present KVASIR, a dataset containing images from inside the gastrointestinal (GI) tract. The collection of images are classified into three important anatomical landmarks and three clinically significant findings. In addition, it contains two categories of images related to endoscopic polyp removal. Sorting and annotation of the dataset is performed by medical doctors (experienced endoscopists). In this respect, KVASIR is important for research on both single- and multi-disease computer aided detection. By providing it, we invite and enable multimedia researcher into the medical domain of detection and retrieval.","PeriodicalId":123321,"journal":{"name":"Proceedings of the 8th ACM on Multimedia Systems Conference","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122655207","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 356
Multimedia Sensor Dataset for the Analysis of Vehicle Movement 用于车辆运动分析的多媒体传感器数据集
Pub Date : 2017-06-20 DOI: 10.1145/3083187.3083217
Wonhee Cho, S. H. Kim
With applications ranging from basic trajectory calculations to complex autonomous vehicle operations, detailed vehicle movement analysis has been getting more attention in academia and industry. So far, real-data driven analysis, e.g., utilizing advanced machine-learning, has used data from sensors such as GPS and accelerometer. However, such research requires quality datasets to enable accurate analysis. To that end, we have collected real vehicle movement data, Multimedia Sensor Data, that contain synchronized sensor data in fine granularity such as GPS, accelerometer, digital compass, gyroscope, and, most importantly, matching real video images recorded at driving time. These real video images provide a way to accurately label the sensor data in generating a quality dataset, e.g., a training dataset. Then, we performed preprocessing steps to clean and refine the raw data, subsequently converted the results into csv files, which are compatible with a wide variety of analysis tools. We also provided sample cases to demonstrate methods of identifying abnormal driving patterns such as moving over a speed bump. This dataset will be useful for researchers refining their analyses of vehicle movements.
从基本的轨迹计算到复杂的自动驾驶车辆操作,详细的车辆运动分析越来越受到学术界和工业界的关注。到目前为止,实时数据驱动的分析,例如利用先进的机器学习,已经使用了来自GPS和加速度计等传感器的数据。然而,这样的研究需要高质量的数据集才能进行准确的分析。为此,我们收集了真实的车辆运动数据,多媒体传感器数据,其中包含精细粒度的同步传感器数据,如GPS,加速度计,数字罗盘,陀螺仪,最重要的是,匹配驾驶时记录的真实视频图像。这些真实的视频图像提供了一种准确的标签传感器数据生成质量的数据集,例如,一个训练数据集。然后,我们执行预处理步骤,对原始数据进行清理和细化,随后将结果转换为csv文件,该文件与各种分析工具兼容。我们还提供了示例案例来演示识别异常驾驶模式的方法,例如在减速带上移动。这个数据集将有助于研究人员改进他们对车辆运动的分析。
{"title":"Multimedia Sensor Dataset for the Analysis of Vehicle Movement","authors":"Wonhee Cho, S. H. Kim","doi":"10.1145/3083187.3083217","DOIUrl":"https://doi.org/10.1145/3083187.3083217","url":null,"abstract":"With applications ranging from basic trajectory calculations to complex autonomous vehicle operations, detailed vehicle movement analysis has been getting more attention in academia and industry. So far, real-data driven analysis, e.g., utilizing advanced machine-learning, has used data from sensors such as GPS and accelerometer. However, such research requires quality datasets to enable accurate analysis. To that end, we have collected real vehicle movement data, Multimedia Sensor Data, that contain synchronized sensor data in fine granularity such as GPS, accelerometer, digital compass, gyroscope, and, most importantly, matching real video images recorded at driving time. These real video images provide a way to accurately label the sensor data in generating a quality dataset, e.g., a training dataset. Then, we performed preprocessing steps to clean and refine the raw data, subsequently converted the results into csv files, which are compatible with a wide variety of analysis tools. We also provided sample cases to demonstrate methods of identifying abnormal driving patterns such as moving over a speed bump. This dataset will be useful for researchers refining their analyses of vehicle movements.","PeriodicalId":123321,"journal":{"name":"Proceedings of the 8th ACM on Multimedia Systems Conference","volume":"27 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125499980","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
A Measurement Study on Achieving Imperceptible Latency in Mobile Cloud Gaming 移动云游戏中实现不可察觉延迟的测量研究
Pub Date : 2017-06-20 DOI: 10.1145/3083187.3083191
Teemu Kämäräinen, M. Siekkinen, Antti Ylä-Jääski, Wenxiao Zhang, P. Hui
Cloud gaming is a relatively new paradigm in which the game is rendered in the cloud and is streamed to an end-user device through a thin client. Latency is a key challenge for cloud gaming. In order to optimize the end-to-end latency, it is first necessary to understand how the end-to-end latency builds up from the mobile device to the cloud gaming server. In this paper we dissect the delays occurring in the mobile device and measure access delays in various networks and network conditions. We also perform a Europe-wide latency measurement study to find the optimal server locations and see how the number of server locations affects the network delay. The results are compared to limits found for perceivable delays in recent human-computer interaction studies. We show that the limits can be achieved only with the latest mobile devices with specific control methods. In addition, we study the expected latency reduction by near future technological development and show that its potential impact is bigger on the end-to-end latency than that of replication of the service and server placement optimization.
云游戏是一种相对较新的模式,其中游戏在云中呈现,并通过瘦客户端流式传输到终端用户设备。延迟是云游戏面临的一个关键挑战。为了优化端到端延迟,首先有必要了解从移动设备到云游戏服务器的端到端延迟是如何形成的。在本文中,我们剖析了移动设备中出现的延迟,并测量了各种网络和网络条件下的访问延迟。我们还执行了欧洲范围内的延迟测量研究,以找到最佳的服务器位置,并查看服务器位置的数量如何影响网络延迟。结果与最近人机交互研究中发现的可感知延迟的限制进行了比较。我们表明,只有使用具有特定控制方法的最新移动设备才能达到限制。此外,我们研究了近期技术发展所带来的预期延迟减少,并表明其对端到端延迟的潜在影响大于服务复制和服务器放置优化。
{"title":"A Measurement Study on Achieving Imperceptible Latency in Mobile Cloud Gaming","authors":"Teemu Kämäräinen, M. Siekkinen, Antti Ylä-Jääski, Wenxiao Zhang, P. Hui","doi":"10.1145/3083187.3083191","DOIUrl":"https://doi.org/10.1145/3083187.3083191","url":null,"abstract":"Cloud gaming is a relatively new paradigm in which the game is rendered in the cloud and is streamed to an end-user device through a thin client. Latency is a key challenge for cloud gaming. In order to optimize the end-to-end latency, it is first necessary to understand how the end-to-end latency builds up from the mobile device to the cloud gaming server. In this paper we dissect the delays occurring in the mobile device and measure access delays in various networks and network conditions. We also perform a Europe-wide latency measurement study to find the optimal server locations and see how the number of server locations affects the network delay. The results are compared to limits found for perceivable delays in recent human-computer interaction studies. We show that the limits can be achieved only with the latest mobile devices with specific control methods. In addition, we study the expected latency reduction by near future technological development and show that its potential impact is bigger on the end-to-end latency than that of replication of the service and server placement optimization.","PeriodicalId":123321,"journal":{"name":"Proceedings of the 8th ACM on Multimedia Systems Conference","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130777093","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 32
A Dataset for Exploring User Behaviors in VR Spherical Video Streaming VR球形视频流中用户行为研究数据集
Pub Date : 2017-06-20 DOI: 10.1145/3083187.3083210
Chenglei Wu, Zhihao Tan, Zhi Wang, Shiqiang Yang
With Virtual Reality (VR) devices and content getting increasingly popular, understanding user behaviors in virtual environment is important for not only VR product design but also user experience improvement. In VR applications, the head movement is one of the most important user behaviors, which can reflect a user's visual attention, preference, and even unique motion pattern. However, to the best of our knowledge, no dataset containing this information is publicly available. In this paper, we present a head tracking dataset composed of 48 users (24 males and 24 females) watching 18 sphere videos from 5 categories. We carefully record how users watch the videos, how their heads move in each session, what directions they focus, and what content they can remember after each session. Based on this dataset, we show that people share certain common patterns in VR spherical video streaming, which are different from conventional video streaming. We believe the dataset can serve good resource for exploring user behavior patterns in VR applications.
随着虚拟现实(VR)设备和内容的日益普及,了解用户在虚拟环境中的行为不仅对VR产品设计很重要,而且对用户体验的改善也很重要。在VR应用中,头部运动是最重要的用户行为之一,它可以反映用户的视觉注意力、偏好,甚至是独特的运动模式。然而,据我们所知,没有包含这些信息的数据集是公开的。在本文中,我们提出了一个由48个用户(24男24女)组成的头部跟踪数据集,该数据集观看了来自5个类别的18个球体视频。我们仔细记录用户观看视频的方式,他们的头部在每次会话中如何移动,他们关注的方向,以及每次会话后他们能记住的内容。基于这个数据集,我们发现人们在VR球形视频流中有一些共同的模式,这与传统的视频流不同。我们相信该数据集可以为探索VR应用中的用户行为模式提供良好的资源。
{"title":"A Dataset for Exploring User Behaviors in VR Spherical Video Streaming","authors":"Chenglei Wu, Zhihao Tan, Zhi Wang, Shiqiang Yang","doi":"10.1145/3083187.3083210","DOIUrl":"https://doi.org/10.1145/3083187.3083210","url":null,"abstract":"With Virtual Reality (VR) devices and content getting increasingly popular, understanding user behaviors in virtual environment is important for not only VR product design but also user experience improvement. In VR applications, the head movement is one of the most important user behaviors, which can reflect a user's visual attention, preference, and even unique motion pattern. However, to the best of our knowledge, no dataset containing this information is publicly available. In this paper, we present a head tracking dataset composed of 48 users (24 males and 24 females) watching 18 sphere videos from 5 categories. We carefully record how users watch the videos, how their heads move in each session, what directions they focus, and what content they can remember after each session. Based on this dataset, we show that people share certain common patterns in VR spherical video streaming, which are different from conventional video streaming. We believe the dataset can serve good resource for exploring user behavior patterns in VR applications.","PeriodicalId":123321,"journal":{"name":"Proceedings of the 8th ACM on Multimedia Systems Conference","volume":"72 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134327314","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 207
A Holistic Multimedia System for Gastrointestinal Tract Disease Detection 胃肠道疾病检测的整体多媒体系统
Pub Date : 2017-06-20 DOI: 10.1145/3083187.3083189
Konstantin Pogorelov, S. Eskeland, T. Lange, C. Griwodz, K. Randel, H. Stensland, Duc-Tien Dang-Nguyen, C. Spampinato, Dag Johansen, M. Riegler, P. Halvorsen
Analysis of medical videos for detection of abnormalities and diseases requires both high precision and recall, but also real-time processing for live feedback and scalability for massive screening of entire populations. Existing work on this field does not provide the necessary combination of retrieval accuracy and performance.; AB@In this paper, a multimedia system is presented where the aim is to tackle automatic analysis of videos from the human gastrointestinal (GI) tract. The system includes the whole pipeline from data collection, processing and analysis, to visualization. The system combines filters using machine learning, image recognition and extraction of global and local image features. Furthermore, it is built in a modular way so that it can easily be extended. At the same time, it is developed for efficient processing in order to provide real-time feedback to the doctors. Our experimental evaluation proves that our system has detection and localisation accuracy at least as good as existing systems for polyp detection, it is capable of detecting a wider range of diseases, it can analyze video in real-time, and it has a low resource consumption for scalability.
对医学视频进行分析以检测异常和疾病,既需要高精度和召回率,也需要实时处理以进行实时反馈,还需要对整个人群进行大规模筛查的可扩展性。这一领域的现有工作没有提供检索精度和性能的必要结合。AB@In本文提出了一个多媒体系统,其目的是解决自动分析视频从人体胃肠道(GI)。该系统包括从数据采集、处理、分析到可视化的整个流程。该系统结合了使用机器学习、图像识别和提取全局和局部图像特征的过滤器。此外,它是以模块化的方式构建的,因此可以很容易地扩展。同时,为了向医生提供实时反馈,它被开发为高效处理。我们的实验评估证明,我们的系统具有至少与现有息肉检测系统一样好的检测和定位精度,能够检测更广泛的疾病,可以实时分析视频,并且具有低资源消耗的可扩展性。
{"title":"A Holistic Multimedia System for Gastrointestinal Tract Disease Detection","authors":"Konstantin Pogorelov, S. Eskeland, T. Lange, C. Griwodz, K. Randel, H. Stensland, Duc-Tien Dang-Nguyen, C. Spampinato, Dag Johansen, M. Riegler, P. Halvorsen","doi":"10.1145/3083187.3083189","DOIUrl":"https://doi.org/10.1145/3083187.3083189","url":null,"abstract":"Analysis of medical videos for detection of abnormalities and diseases requires both high precision and recall, but also real-time processing for live feedback and scalability for massive screening of entire populations. Existing work on this field does not provide the necessary combination of retrieval accuracy and performance.; AB@In this paper, a multimedia system is presented where the aim is to tackle automatic analysis of videos from the human gastrointestinal (GI) tract. The system includes the whole pipeline from data collection, processing and analysis, to visualization. The system combines filters using machine learning, image recognition and extraction of global and local image features. Furthermore, it is built in a modular way so that it can easily be extended. At the same time, it is developed for efficient processing in order to provide real-time feedback to the doctors. Our experimental evaluation proves that our system has detection and localisation accuracy at least as good as existing systems for polyp detection, it is capable of detecting a wider range of diseases, it can analyze video in real-time, and it has a low resource consumption for scalability.","PeriodicalId":123321,"journal":{"name":"Proceedings of the 8th ACM on Multimedia Systems Conference","volume":"55 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124224295","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 22
Unified Remix: a Server Side Solution for Adaptive Bit-Rate Streaming with Inserted and Edited Media Content 统一的混音:一个服务器端解决方案,自适应比特率流与插入和编辑媒体内容
Pub Date : 2017-06-20 DOI: 10.1145/3083187.3083227
Arjen Wagenaar, Dirk Griffioen, R. Mekuria
We present Unified Remix, our solution for adaptive bit-rate streaming of video presentations with inserted or edited content. The solution addresses three important challenges encountered when streaming personalized media presentations. First, it reduces vulnerability to ad blocking technologies and client-side playback deviations encountered when using manifest manipulation based methods. Second, it reduces storage and computational costs associated with alternative server side solutions such as brute force re-encoding or duplicate storage towards levels comparable to linear video streaming (VoD or Live). Third, it handles the multi-source, multi-DRM and multi-protocol aspects for modern video streaming natively in the workflow. The solution is based on a combination of existing proven streaming technologies such as Unified Origin and newly designed components such as the Remix MPEG-4 module. The framework uses standardized technologies such as MPEG-4 ISOBMFF, SMIL and MPEG-DASH. The components work together in a micro service architecture enabling flexible deployment using a (container) orchestration framework on premises or in the cloud. The solution is demonstrated in two use cases: content pre-/post/mid roll and Live Archive to VoD conversion. As many use cases can be implemented based upon Unified Remix, we envision it as a key component of professional video streaming platforms.
我们提出了统一的Remix,我们的解决方案,自适应比特率流视频演示与插入或编辑的内容。该解决方案解决了在流式传输个性化媒体演示时遇到的三个重要挑战。首先,它减少了对广告拦截技术的脆弱性,以及在使用基于清单操作的方法时遇到的客户端播放偏差。其次,它减少了与其他服务器端解决方案相关的存储和计算成本,例如暴力重新编码或重复存储,以达到与线性视频流(VoD或Live)相当的水平。第三,在工作流中原生地处理现代视频流的多源、多drm和多协议方面的问题。该解决方案基于现有的经过验证的流媒体技术(如Unified Origin)和新设计的组件(如Remix MPEG-4模块)的组合。该框架采用了MPEG-4 ISOBMFF、SMIL和MPEG-DASH等标准化技术。这些组件在微服务体系结构中协同工作,支持在本地或云中使用(容器)编排框架进行灵活部署。该解决方案在两个用例中进行了演示:内容前/后/中卷和Live Archive到VoD的转换。由于许多用例可以基于统一Remix实现,我们将其设想为专业视频流平台的关键组件。
{"title":"Unified Remix: a Server Side Solution for Adaptive Bit-Rate Streaming with Inserted and Edited Media Content","authors":"Arjen Wagenaar, Dirk Griffioen, R. Mekuria","doi":"10.1145/3083187.3083227","DOIUrl":"https://doi.org/10.1145/3083187.3083227","url":null,"abstract":"We present Unified Remix, our solution for adaptive bit-rate streaming of video presentations with inserted or edited content. The solution addresses three important challenges encountered when streaming personalized media presentations. First, it reduces vulnerability to ad blocking technologies and client-side playback deviations encountered when using manifest manipulation based methods. Second, it reduces storage and computational costs associated with alternative server side solutions such as brute force re-encoding or duplicate storage towards levels comparable to linear video streaming (VoD or Live). Third, it handles the multi-source, multi-DRM and multi-protocol aspects for modern video streaming natively in the workflow. The solution is based on a combination of existing proven streaming technologies such as Unified Origin and newly designed components such as the Remix MPEG-4 module. The framework uses standardized technologies such as MPEG-4 ISOBMFF, SMIL and MPEG-DASH. The components work together in a micro service architecture enabling flexible deployment using a (container) orchestration framework on premises or in the cloud. The solution is demonstrated in two use cases: content pre-/post/mid roll and Live Archive to VoD conversion. As many use cases can be implemented based upon Unified Remix, we envision it as a key component of professional video streaming platforms.","PeriodicalId":123321,"journal":{"name":"Proceedings of the 8th ACM on Multimedia Systems Conference","volume":"279 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121275103","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
A Scalable and Privacy-Aware IoT Service for Live Video Analytics 用于实时视频分析的可扩展和隐私感知物联网服务
Pub Date : 2017-06-20 DOI: 10.1145/3083187.3083192
Junjue Wang, Brandon Amos, Anupam Das, P. Pillai, N. Sadeh, M. Satyanarayanan
We present OpenFace, our new open-source face recognition system that approaches state-of-the-art accuracy. Integrating OpenFace with inter-frame tracking, we build RTFace, a mechanism for denaturing video streams that selectively blurs faces according to specified policies at full frame rates. This enables privacy management for live video analytics while providing a secure approach for handling retrospective policy exceptions. Finally, we present a scalable, privacy-aware architecture for large camera networks using RTFace.
我们介绍OpenFace,我们新的开源人脸识别系统,接近最先进的精度。将OpenFace与帧间跟踪相结合,我们构建了RTFace,这是一种变性视频流的机制,可以根据指定的策略在全帧速率下选择性地模糊人脸。这支持实时视频分析的隐私管理,同时为处理回顾性策略异常提供了一种安全的方法。最后,我们使用RTFace为大型摄像机网络提供了一个可扩展的隐私感知架构。
{"title":"A Scalable and Privacy-Aware IoT Service for Live Video Analytics","authors":"Junjue Wang, Brandon Amos, Anupam Das, P. Pillai, N. Sadeh, M. Satyanarayanan","doi":"10.1145/3083187.3083192","DOIUrl":"https://doi.org/10.1145/3083187.3083192","url":null,"abstract":"We present OpenFace, our new open-source face recognition system that approaches state-of-the-art accuracy. Integrating OpenFace with inter-frame tracking, we build RTFace, a mechanism for denaturing video streams that selectively blurs faces according to specified policies at full frame rates. This enables privacy management for live video analytics while providing a secure approach for handling retrospective policy exceptions. Finally, we present a scalable, privacy-aware architecture for large camera networks using RTFace.","PeriodicalId":123321,"journal":{"name":"Proceedings of the 8th ACM on Multimedia Systems Conference","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122407911","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 103
360° Video Viewing Dataset in Head-Mounted Virtual Reality 头戴式虚拟现实中的360°视频观看数据集
Pub Date : 2017-06-20 DOI: 10.1145/3083187.3083219
Wen-Chih Lo, Ching-Ling Fan, Jean Lee, Chun-Ying Huang, Kuan-Ta Chen, Cheng-Hsin Hsu
360° videos and Head-Mounted Displays (HMDs) are getting increasingly popular. However, streaming 360° videos to HMDs is challenging. This is because only video content in viewers' Field-of-Views (FoVs) is rendered, and thus sending complete 360° videos wastes resources, including network bandwidth, storage space, and processing power. Optimizing the 360° video streaming to HMDs is, however, highly data and viewer dependent, and thus dictates real datasets. However, to our best knowledge, such datasets are not available in the literature. In this paper, we present our datasets of both content data (such as image saliency maps and motion maps derived from 360° videos) and sensor data (such as viewer head positions and orientations derived from HMD sensors). We put extra efforts to align the content and sensor data using the timestamps in the raw log files. The resulting datasets can be used by researchers, engineers, and hobbyists to either optimize existing 360° video streaming applications (like rate-distortion optimization) and novel applications (like crowd-driven camera movements). We believe that our dataset will stimulate more research activities along this exciting new research direction.
360°视频和头戴式显示器(hmd)越来越受欢迎。然而,将360°视频流式传输到hmd是具有挑战性的。这是因为只有观众的视场(fov)中的视频内容被渲染,因此发送完整的360°视频浪费资源,包括网络带宽,存储空间和处理能力。然而,优化360°视频流到hmd是高度依赖于数据和观众的,因此决定了真实的数据集。然而,据我们所知,这样的数据集在文献中是不可用的。在本文中,我们展示了内容数据(如来自360°视频的图像显著性地图和运动地图)和传感器数据(如来自HMD传感器的观众头部位置和方向)的数据集。我们花了额外的精力来使用原始日志文件中的时间戳来对齐内容和传感器数据。由此产生的数据集可以被研究人员、工程师和业余爱好者用于优化现有的360°视频流应用程序(如率失真优化)和新应用程序(如人群驱动的摄像机移动)。我们相信我们的数据集将沿着这个令人兴奋的新研究方向激发更多的研究活动。
{"title":"360° Video Viewing Dataset in Head-Mounted Virtual Reality","authors":"Wen-Chih Lo, Ching-Ling Fan, Jean Lee, Chun-Ying Huang, Kuan-Ta Chen, Cheng-Hsin Hsu","doi":"10.1145/3083187.3083219","DOIUrl":"https://doi.org/10.1145/3083187.3083219","url":null,"abstract":"360° videos and Head-Mounted Displays (HMDs) are getting increasingly popular. However, streaming 360° videos to HMDs is challenging. This is because only video content in viewers' Field-of-Views (FoVs) is rendered, and thus sending complete 360° videos wastes resources, including network bandwidth, storage space, and processing power. Optimizing the 360° video streaming to HMDs is, however, highly data and viewer dependent, and thus dictates real datasets. However, to our best knowledge, such datasets are not available in the literature. In this paper, we present our datasets of both content data (such as image saliency maps and motion maps derived from 360° videos) and sensor data (such as viewer head positions and orientations derived from HMD sensors). We put extra efforts to align the content and sensor data using the timestamps in the raw log files. The resulting datasets can be used by researchers, engineers, and hobbyists to either optimize existing 360° video streaming applications (like rate-distortion optimization) and novel applications (like crowd-driven camera movements). We believe that our dataset will stimulate more research activities along this exciting new research direction.","PeriodicalId":123321,"journal":{"name":"Proceedings of the 8th ACM on Multimedia Systems Conference","volume":"12 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131885193","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 187
期刊
Proceedings of the 8th ACM on Multimedia Systems Conference
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1