首页 > 最新文献

2012 IEEE International Conference on Multimedia and Expo Workshops最新文献

英文 中文
Living the Past: Augmented Reality and Archeology 生活在过去:增强现实和考古学
Pub Date : 2012-07-09 DOI: 10.1109/ICMEW.2012.67
Andrea Bernardini, C. Delogu, E. Pallotti, Luca Costantini
Archeological remnants in urban areas tend to be included in the urban landscape or even remain hidden in subterranean locations which are not visible and, for these reasons, they are accessed with difficulty by visitors. In our previous experience, we developed a mobile application, which guided visitors in real time through various archaeological sites using texts, images, and videos. The results of an evaluation test which collected visitors' impressions and suggestions showed us that the mobile application allowed them to visit archeological remnants in a more participative way but that most visitors were unable to imagine what relation the archaeological remnants had with the ancient urban landscape. To solve this problem and improve the visitors' experience, we are now working at another application, which combines historical and archeological details with an immersive experience. The mobile application recognizes a cultural heritage element by image recognition or by positioning and it augments the interface with various layers of information. Furthermore, the application will provide not only information but it will offer to visitors an emotional experience.
城市地区的考古遗迹往往包含在城市景观中,甚至隐藏在不可见的地下位置,因此游客很难进入。根据我们之前的经验,我们开发了一个移动应用程序,它通过文本、图像和视频实时引导游客参观各种考古遗址。一项收集了游客印象和建议的评估测试结果表明,移动应用程序让他们以一种更具参与性的方式参观考古遗迹,但大多数游客无法想象考古遗迹与古代城市景观的关系。为了解决这个问题并改善游客的体验,我们现在正在开发另一个应用程序,它将历史和考古细节与身临其境的体验相结合。移动应用程序通过图像识别或定位来识别文化遗产元素,并通过各种信息层来增强界面。此外,该应用程序不仅提供信息,还将为访问者提供情感体验。
{"title":"Living the Past: Augmented Reality and Archeology","authors":"Andrea Bernardini, C. Delogu, E. Pallotti, Luca Costantini","doi":"10.1109/ICMEW.2012.67","DOIUrl":"https://doi.org/10.1109/ICMEW.2012.67","url":null,"abstract":"Archeological remnants in urban areas tend to be included in the urban landscape or even remain hidden in subterranean locations which are not visible and, for these reasons, they are accessed with difficulty by visitors. In our previous experience, we developed a mobile application, which guided visitors in real time through various archaeological sites using texts, images, and videos. The results of an evaluation test which collected visitors' impressions and suggestions showed us that the mobile application allowed them to visit archeological remnants in a more participative way but that most visitors were unable to imagine what relation the archaeological remnants had with the ancient urban landscape. To solve this problem and improve the visitors' experience, we are now working at another application, which combines historical and archeological details with an immersive experience. The mobile application recognizes a cultural heritage element by image recognition or by positioning and it augments the interface with various layers of information. Furthermore, the application will provide not only information but it will offer to visitors an emotional experience.","PeriodicalId":385797,"journal":{"name":"2012 IEEE International Conference on Multimedia and Expo Workshops","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2012-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123174379","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Depth Map Super-Resolution Using Synthesized View Matching for Depth-Image-Based Rendering 基于深度图像渲染的合成视图匹配深度图超分辨率
Pub Date : 2012-07-09 DOI: 10.1109/ICMEW.2012.111
Wei Hu, Gene Cheung, Xin Li, O. Au
In texture-plus-depth format of 3D visual data, texture and depth maps of multiple viewpoints are coded and transmitted at sender. At receiver, decoded texture and depth maps of two neighboring viewpoints are used to synthesize a desired intermediate view via depth-image-based rendering (DIBR). In this paper, to enable transmission of depth maps at low resolution for bit saving, we propose a novel super-resolution (SR) algorithm to increase the resolution of the received depth map at decoder to match the corresponding received high resolution texture map for DIBR. Unlike previous depth map SR techniques that only utilize the texture map of the same view 0 to interpolate missing depth pixels of view 0, we use texture maps of the same and neighboring viewpoints, 0 and 1, so that the error between the original texture map of view 1 and the synthesized image of view 1 (interpolated using texture and depth maps of view 0) can be used as a regularization term during depth map SR of view 0. Further, piecewise smoothness of the reconstructed depth map is enforced by computing only the lowest frequency coefficients in Graph based Transform (GBT) domain for each interpolated block. Experimental results show that our SR scheme out-performed a previous scheme by up to 1.7dB in synthesized view quality in PSNR.
在3D视觉数据的纹理加深度格式中,多视点的纹理图和深度图在发送端进行编码和传输。在接收端,使用解码后的相邻视点纹理图和深度图,通过基于深度图像的渲染(deep -based rendering, DIBR)合成所需的中间视图。为了实现低分辨率深度图的传输以节省比特,我们提出了一种新的超分辨率(SR)算法,以提高解码器接收深度图的分辨率,以匹配DIBR接收的相应高分辨率纹理图。深度与以前老地图技术,只有利用纹理映射相同的视图0插入缺失的像素的深度视图0,我们使用相同的纹理地图和邻近视点,0和1,这样误差的原始纹理映射视图和视图的合成图像1(插值使用纹理和深度的地图视图0)可以作为正则化项在深度地图SR的0。此外,通过对每个插值块仅计算基于图形变换(GBT)域的最低频率系数来保证重建深度图的分段平滑性。实验结果表明,在PSNR条件下,该方案的合成图像质量比现有方案提高了1.7dB。
{"title":"Depth Map Super-Resolution Using Synthesized View Matching for Depth-Image-Based Rendering","authors":"Wei Hu, Gene Cheung, Xin Li, O. Au","doi":"10.1109/ICMEW.2012.111","DOIUrl":"https://doi.org/10.1109/ICMEW.2012.111","url":null,"abstract":"In texture-plus-depth format of 3D visual data, texture and depth maps of multiple viewpoints are coded and transmitted at sender. At receiver, decoded texture and depth maps of two neighboring viewpoints are used to synthesize a desired intermediate view via depth-image-based rendering (DIBR). In this paper, to enable transmission of depth maps at low resolution for bit saving, we propose a novel super-resolution (SR) algorithm to increase the resolution of the received depth map at decoder to match the corresponding received high resolution texture map for DIBR. Unlike previous depth map SR techniques that only utilize the texture map of the same view 0 to interpolate missing depth pixels of view 0, we use texture maps of the same and neighboring viewpoints, 0 and 1, so that the error between the original texture map of view 1 and the synthesized image of view 1 (interpolated using texture and depth maps of view 0) can be used as a regularization term during depth map SR of view 0. Further, piecewise smoothness of the reconstructed depth map is enforced by computing only the lowest frequency coefficients in Graph based Transform (GBT) domain for each interpolated block. Experimental results show that our SR scheme out-performed a previous scheme by up to 1.7dB in synthesized view quality in PSNR.","PeriodicalId":385797,"journal":{"name":"2012 IEEE International Conference on Multimedia and Expo Workshops","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2012-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115282872","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 15
Statistical Color Model Based Adult Video Filter 基于统计色彩模型的成人视频滤镜
Pub Date : 2012-07-09 DOI: 10.1109/ICMEW.2012.66
Liang Yin, Mingzhi Dong, Weihong Deng, Jun Guo, Bin Zhang
This paper, guided by Statistical Color Models, proposes a real-time Adult Video detector to filter the adult content in the video. A generic color model is constructed by statistical analysis of the sample images containing adult pixels. We fully utilize the video continuity characteristics, i.e. preceding and following N frames considered in the classification. Our method, through experimental, displays a satisfactory performance for detecting adult content. The reminder of the paper addresses the application of real-time adult video filter that blocks adult content from kids.
本文以统计色彩模型为指导,提出了一种实时成人视频检测器,用于过滤视频中的成人内容。通过对包含成人像素的样本图像进行统计分析,构建了一个通用的颜色模型。我们充分利用了视频的连续性特征,即分类中考虑的前后N帧。通过实验,我们的方法对成人内容的检测效果令人满意。论文的提醒说明了实时成人视频过滤器的应用,该过滤器可以阻止儿童观看成人内容。
{"title":"Statistical Color Model Based Adult Video Filter","authors":"Liang Yin, Mingzhi Dong, Weihong Deng, Jun Guo, Bin Zhang","doi":"10.1109/ICMEW.2012.66","DOIUrl":"https://doi.org/10.1109/ICMEW.2012.66","url":null,"abstract":"This paper, guided by Statistical Color Models, proposes a real-time Adult Video detector to filter the adult content in the video. A generic color model is constructed by statistical analysis of the sample images containing adult pixels. We fully utilize the video continuity characteristics, i.e. preceding and following N frames considered in the classification. Our method, through experimental, displays a satisfactory performance for detecting adult content. The reminder of the paper addresses the application of real-time adult video filter that blocks adult content from kids.","PeriodicalId":385797,"journal":{"name":"2012 IEEE International Conference on Multimedia and Expo Workshops","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2012-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116748237","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Inter Prediction Based on Low-rank Matrix Completion 基于低秩矩阵补全的内部预测
Pub Date : 2012-07-09 DOI: 10.1109/ICMEW.2012.98
Yunhui Shi, He Li, Jin Wang, Wenpeng Ding, Baocai Yin
This paper proposes a new method of inter prediction based on low-rank matrix completion. By collection and rearrangement, image regions with high correlations can be used to generate a low-rank or approximately low-rank matrix. We view prediction values as the missing part in an incomplete low-rank matrix, and obtain the prediction by recovering the generated low-rank matrix. Taking advantage of exact recovery of incomplete matrix, the low-rank based prediction can exploit temporal correlation better. Our proposed prediction has the advantage of higher accuracy and less extra information, as the motion vector doesn't need to be encoded. Simulation results show that the bit-rate saving of the proposed scheme can reach up to 9.91% compared with H.264/AVC. Our scheme also outperforms the counterpart of the Template Matching Averaging (TMA) prediction by 8.06% at most.
提出了一种基于低秩矩阵补全的互预测新方法。通过收集和重排,具有高相关性的图像区域可以用来生成低秩或近似低秩矩阵。我们将预测值视为不完全低秩矩阵中缺失的部分,通过恢复生成的低秩矩阵得到预测值。利用不完全矩阵的精确恢复,低秩预测可以更好地利用时间相关性。我们提出的预测具有精度高、额外信息少的优点,因为运动矢量不需要编码。仿真结果表明,与H.264/AVC相比,该方案的码率节省可达9.91%。我们的方案也优于模板匹配平均(TMA)预测最多8.06%。
{"title":"Inter Prediction Based on Low-rank Matrix Completion","authors":"Yunhui Shi, He Li, Jin Wang, Wenpeng Ding, Baocai Yin","doi":"10.1109/ICMEW.2012.98","DOIUrl":"https://doi.org/10.1109/ICMEW.2012.98","url":null,"abstract":"This paper proposes a new method of inter prediction based on low-rank matrix completion. By collection and rearrangement, image regions with high correlations can be used to generate a low-rank or approximately low-rank matrix. We view prediction values as the missing part in an incomplete low-rank matrix, and obtain the prediction by recovering the generated low-rank matrix. Taking advantage of exact recovery of incomplete matrix, the low-rank based prediction can exploit temporal correlation better. Our proposed prediction has the advantage of higher accuracy and less extra information, as the motion vector doesn't need to be encoded. Simulation results show that the bit-rate saving of the proposed scheme can reach up to 9.91% compared with H.264/AVC. Our scheme also outperforms the counterpart of the Template Matching Averaging (TMA) prediction by 8.06% at most.","PeriodicalId":385797,"journal":{"name":"2012 IEEE International Conference on Multimedia and Expo Workshops","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2012-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122627128","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Dynamic Resource Allocation for Event Processing in Surveillance Systems 监视系统中事件处理的动态资源分配
Pub Date : 2012-07-09 DOI: 10.1109/ICMEW.2012.74
D. Ahmed
Allocating computing resources to different tasks of surveillance systems has always been a big challenge. The problem becomes complicated when it requires dealing with real-time computation and decision making as the system cannot afford of processing all sensory feeds and execute computationally expensive algorithms. In multi-modal surveillance systems, real-time event detection and understanding of a situation is crucial. So, the proper use of computing resources is necessary to control and manage an area of surveillance. This paper introduces a dynamic task scheduling technique considering available computing resources and real-time requirement according to the current surveillance context. The task scheduler determines the importance of each sensor with respect to its observation and surrounding context. The scheduler dynamically allocates CPU clock to data streams of each sensor so that it can minimize event detection time from the time of its occurrence. The simulation results reveal that the task scheduler can offer proper resource utilization which is valuable for surveillance systems.
将计算资源分配到监控系统的不同任务中一直是一个巨大的挑战。当需要处理实时计算和决策时,由于系统无法承担处理所有感官反馈和执行计算昂贵的算法,问题变得复杂。在多模式监测系统中,实时事件检测和对情况的理解至关重要。因此,正确使用计算资源对于控制和管理监视区域是必要的。针对当前监控环境,提出了一种考虑可用计算资源和实时性要求的动态任务调度技术。任务调度程序确定每个传感器相对于其观察和周围环境的重要性。调度器动态地为每个传感器的数据流分配CPU时钟,以便它可以从事件发生的时间开始最小化事件检测时间。仿真结果表明,该任务调度程序能够提供合理的资源利用率,对监控系统具有一定的参考价值。
{"title":"Dynamic Resource Allocation for Event Processing in Surveillance Systems","authors":"D. Ahmed","doi":"10.1109/ICMEW.2012.74","DOIUrl":"https://doi.org/10.1109/ICMEW.2012.74","url":null,"abstract":"Allocating computing resources to different tasks of surveillance systems has always been a big challenge. The problem becomes complicated when it requires dealing with real-time computation and decision making as the system cannot afford of processing all sensory feeds and execute computationally expensive algorithms. In multi-modal surveillance systems, real-time event detection and understanding of a situation is crucial. So, the proper use of computing resources is necessary to control and manage an area of surveillance. This paper introduces a dynamic task scheduling technique considering available computing resources and real-time requirement according to the current surveillance context. The task scheduler determines the importance of each sensor with respect to its observation and surrounding context. The scheduler dynamically allocates CPU clock to data streams of each sensor so that it can minimize event detection time from the time of its occurrence. The simulation results reveal that the task scheduler can offer proper resource utilization which is valuable for surveillance systems.","PeriodicalId":385797,"journal":{"name":"2012 IEEE International Conference on Multimedia and Expo Workshops","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2012-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122733632","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Minimizing Video Retransmission Delay and Energy Consumption with Caching Routers 最小化视频重传延迟和缓存路由器的能耗
Pub Date : 2012-07-09 DOI: 10.1109/ICMEW.2012.25
M. Mcgarry, Jesus Hernandez, R. Ferzli, V. Syrotiuk
We investigated the use of caching of packets containing video at intermediary routers to reduce the delay and energy consumption of Automatic Repeat reQuest (ARQ) error recovery. We modeled the two mathematical programs that select the optimal set of routers to have caching ability, one to minimize energy consumption and the other to minimize retransmission delay. Both of these mathematical programs have identical structure. We then solve these mathematical programs with a dynamic programming solution whose execution time growth is polynomial in the size of the input parameters. Our performance analysis indicates that the optimal solution significantly outperforms several heuristic solutions.
我们研究了在中间路由器上使用包含视频的数据包缓存来减少自动重复请求(ARQ)错误恢复的延迟和能耗。我们对两个数学程序进行建模,以选择具有缓存能力的最优路由器集,一个是最小化能耗,另一个是最小化重传延迟。这两个数学程序具有相同的结构。然后,我们用一个动态规划解来求解这些数学程序,该解的执行时间增长是输入参数大小的多项式。我们的性能分析表明,最优解决方案显著优于几个启发式解决方案。
{"title":"Minimizing Video Retransmission Delay and Energy Consumption with Caching Routers","authors":"M. Mcgarry, Jesus Hernandez, R. Ferzli, V. Syrotiuk","doi":"10.1109/ICMEW.2012.25","DOIUrl":"https://doi.org/10.1109/ICMEW.2012.25","url":null,"abstract":"We investigated the use of caching of packets containing video at intermediary routers to reduce the delay and energy consumption of Automatic Repeat reQuest (ARQ) error recovery. We modeled the two mathematical programs that select the optimal set of routers to have caching ability, one to minimize energy consumption and the other to minimize retransmission delay. Both of these mathematical programs have identical structure. We then solve these mathematical programs with a dynamic programming solution whose execution time growth is polynomial in the size of the input parameters. Our performance analysis indicates that the optimal solution significantly outperforms several heuristic solutions.","PeriodicalId":385797,"journal":{"name":"2012 IEEE International Conference on Multimedia and Expo Workshops","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2012-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116219501","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
A New Texture Feature for Improved Food Recognition Accuracy in a Mobile Phone Based Dietary Assessment System 一种新的纹理特征提高了基于手机的膳食评估系统的食物识别精度
Pub Date : 2012-07-09 DOI: 10.1109/ICMEW.2012.79
M. Rahman, M. Pickering, D. Kerr, C. Boushey, E. Delp
Poor diet is one of the key determinants of an individual's risk of developing chronic diseases. Assessing what people eat is fundamental to establishing the link between diet and disease. Food records are considered the best approach for assessing energy intake however paper-based food recording is cumbersome and often inaccurate. Researchers have begun to explore how mobile devices can be used to reduce the burden of recording nutritional intake. The integrated camera in a mobile phone can be used for capturing images of food consumed. These images are then processed to automatically identify the food items for record keeping purposes. In such systems, the accurate classification of food items in these images is vital to the success of such a system. In this paper we will present a new method for generating texture features from food images and demonstrate that this new feature provides greater food classification accuracy for a mobile phone based dietary assessment system.
不良饮食习惯是个人患慢性病风险的关键决定因素之一。评估人们的饮食是建立饮食与疾病之间联系的基础。食物记录被认为是评估能量摄入的最佳方法,但纸质食物记录既麻烦又不准确。研究人员已经开始探索如何使用移动设备来减轻记录营养摄入量的负担。手机中的集成摄像头可以用来捕捉食物的图像。然后对这些图像进行处理,以自动识别食品,以便保存记录。在这样的系统中,在这些图像中对食物进行准确分类对这样一个系统的成功至关重要。在本文中,我们将提出一种从食物图像中生成纹理特征的新方法,并证明这种新特征为基于手机的饮食评估系统提供了更高的食物分类精度。
{"title":"A New Texture Feature for Improved Food Recognition Accuracy in a Mobile Phone Based Dietary Assessment System","authors":"M. Rahman, M. Pickering, D. Kerr, C. Boushey, E. Delp","doi":"10.1109/ICMEW.2012.79","DOIUrl":"https://doi.org/10.1109/ICMEW.2012.79","url":null,"abstract":"Poor diet is one of the key determinants of an individual's risk of developing chronic diseases. Assessing what people eat is fundamental to establishing the link between diet and disease. Food records are considered the best approach for assessing energy intake however paper-based food recording is cumbersome and often inaccurate. Researchers have begun to explore how mobile devices can be used to reduce the burden of recording nutritional intake. The integrated camera in a mobile phone can be used for capturing images of food consumed. These images are then processed to automatically identify the food items for record keeping purposes. In such systems, the accurate classification of food items in these images is vital to the success of such a system. In this paper we will present a new method for generating texture features from food images and demonstrate that this new feature provides greater food classification accuracy for a mobile phone based dietary assessment system.","PeriodicalId":385797,"journal":{"name":"2012 IEEE International Conference on Multimedia and Expo Workshops","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2012-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125033846","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 21
A Dense 3D Reconstruction Approach from Uncalibrated Video Sequences 从未校准的视频序列密集三维重建方法
Pub Date : 2012-07-09 DOI: 10.1109/ICMEW.2012.108
L. Ling, I. Burnett, E. Cheng
Current approaches for 3D reconstruction from feature points of images are classed as sparse and dense techniques. However, the sparse approaches are insufficient for surface reconstruction since only sparsely distributed feature points are presented. Further, existing dense reconstruction approaches require pre-calibrated camera orientation, which limits the applicability and flexibility. This paper proposes a one-stop 3D reconstruction solution that reconstructs a highly dense surface from an uncalibrated video sequence, the camera orientations and surface reconstruction are simultaneously computed from new dense point features using an approach motivated by Structure from Motion (SfM) techniques. Further, this paper presents a flexible automatic method with the simple interface of 'videos to 3D model'. These improvements are essential to practical applications in 3D modeling and visualization. The reliability of the proposed algorithm has been tested on various data sets and the accuracy and performance are compared with both sparse and dense reconstruction benchmark algorithms.
目前基于图像特征点的三维重建方法分为稀疏和密集两类。然而,稀疏方法只提供稀疏分布的特征点,不足以进行表面重建。此外,现有的密集重建方法需要预先校准相机方向,这限制了适用性和灵活性。本文提出了一种一站式3D重建解决方案,该方案从未校准的视频序列中重建高密度表面,使用基于运动结构(SfM)技术的方法从新的密集点特征中同时计算摄像机方向和表面重建。此外,本文还提出了一种灵活的“视频到三维模型”的自动化方法。这些改进对于3D建模和可视化的实际应用至关重要。在各种数据集上测试了该算法的可靠性,并与稀疏重构基准算法和密集重构基准算法进行了精度和性能比较。
{"title":"A Dense 3D Reconstruction Approach from Uncalibrated Video Sequences","authors":"L. Ling, I. Burnett, E. Cheng","doi":"10.1109/ICMEW.2012.108","DOIUrl":"https://doi.org/10.1109/ICMEW.2012.108","url":null,"abstract":"Current approaches for 3D reconstruction from feature points of images are classed as sparse and dense techniques. However, the sparse approaches are insufficient for surface reconstruction since only sparsely distributed feature points are presented. Further, existing dense reconstruction approaches require pre-calibrated camera orientation, which limits the applicability and flexibility. This paper proposes a one-stop 3D reconstruction solution that reconstructs a highly dense surface from an uncalibrated video sequence, the camera orientations and surface reconstruction are simultaneously computed from new dense point features using an approach motivated by Structure from Motion (SfM) techniques. Further, this paper presents a flexible automatic method with the simple interface of 'videos to 3D model'. These improvements are essential to practical applications in 3D modeling and visualization. The reliability of the proposed algorithm has been tested on various data sets and the accuracy and performance are compared with both sparse and dense reconstruction benchmark algorithms.","PeriodicalId":385797,"journal":{"name":"2012 IEEE International Conference on Multimedia and Expo Workshops","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2012-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122951378","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Creative Transformations of Personal Photographs 个人照片的创造性转变
Pub Date : 2012-07-09 DOI: 10.1109/ICMEW.2012.87
Yi Wu, K. Seshadrinathan, Wei Sun, M. E. Choubassi, J. Ratcliff, I. Kozintsev
The popularity of mobile photography paves the way to create new ways of viewing, interacting and enabling a user's creative expression with personal media. In this paper, we describe an instantaneous and automatic method to localize the camera and enable segmentation of foreground objects such as people from an input image, assuming knowledge of the environment in which the image was taken. Camera localization is performed by comparing multiple views of the 3D environment against the uncalibrated input image. Following localization, selected views of the 3D environment are aligned, color-mapped and compared against the input image to segment the foreground content. We demonstrate results using our proposed system in two illustrative applications: a virtual game played between multiple users involving virtual projectiles and a group shot of multiple people who may not be available simultaneously at the same time or place created against a background of their choice.
移动摄影的普及为创造新的观看、互动方式铺平了道路,并使用户能够通过个人媒体进行创造性的表达。在本文中,我们描述了一种即时和自动的方法来定位相机,并从输入图像中分割前景物体(如人),假设图像拍摄环境的知识。相机定位是通过将3D环境的多个视图与未校准的输入图像进行比较来执行的。定位后,将对选定的3D环境视图进行对齐、颜色映射,并与输入图像进行比较,以分割前景内容。我们在两个说明性应用程序中使用我们提出的系统来演示结果:在多个用户之间进行的虚拟游戏,涉及虚拟投射物,以及在他们选择的背景下创建的多个人可能无法在同一时间或地点同时可用的集体照。
{"title":"Creative Transformations of Personal Photographs","authors":"Yi Wu, K. Seshadrinathan, Wei Sun, M. E. Choubassi, J. Ratcliff, I. Kozintsev","doi":"10.1109/ICMEW.2012.87","DOIUrl":"https://doi.org/10.1109/ICMEW.2012.87","url":null,"abstract":"The popularity of mobile photography paves the way to create new ways of viewing, interacting and enabling a user's creative expression with personal media. In this paper, we describe an instantaneous and automatic method to localize the camera and enable segmentation of foreground objects such as people from an input image, assuming knowledge of the environment in which the image was taken. Camera localization is performed by comparing multiple views of the 3D environment against the uncalibrated input image. Following localization, selected views of the 3D environment are aligned, color-mapped and compared against the input image to segment the foreground content. We demonstrate results using our proposed system in two illustrative applications: a virtual game played between multiple users involving virtual projectiles and a group shot of multiple people who may not be available simultaneously at the same time or place created against a background of their choice.","PeriodicalId":385797,"journal":{"name":"2012 IEEE International Conference on Multimedia and Expo Workshops","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2012-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121846949","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Virtual interactions: Can EEG Help Make the Difference with Real Interaction? 虚拟交互:脑电图能帮助实现与真实交互的不同吗?
Pub Date : 2012-07-09 DOI: 10.1109/ICMEW.2012.33
J. Rzepecki, Jonathan Delcourt, Matthieu Perreira Da Silva, P. Callet
Science and technology progress fast, but mouse and keyboard are still used to control multimedia devices. One of the limiting factors of gesture based HCIs adoption is the detection of the user's intention to interact. This study tries to make a step in that direction with use of consumer EEG sensor headset. EEG headset records in real-time data that can help to identify intention of the user based on his emotional state. For each subject EEG responses for different stimuli are recorded. Acquiring these data allows to determine the potential of EEG based intention detection. The findings are promising and with proper implementation should allow to building a new type of HCI devices.
科技进步很快,但是控制多媒体设备仍然需要使用鼠标和键盘。采用基于手势的hci的限制因素之一是对用户交互意图的检测。本研究试图在这个方向上迈出一步,使用消费者脑电图传感器耳机。脑电图耳机记录实时数据,可以根据用户的情绪状态帮助识别用户的意图。记录每位受试者对不同刺激的脑电图反应。获取这些数据可以确定基于EEG的意图检测的潜力。这些发现是有希望的,通过适当的实施,应该可以建立一种新型的HCI设备。
{"title":"Virtual interactions: Can EEG Help Make the Difference with Real Interaction?","authors":"J. Rzepecki, Jonathan Delcourt, Matthieu Perreira Da Silva, P. Callet","doi":"10.1109/ICMEW.2012.33","DOIUrl":"https://doi.org/10.1109/ICMEW.2012.33","url":null,"abstract":"Science and technology progress fast, but mouse and keyboard are still used to control multimedia devices. One of the limiting factors of gesture based HCIs adoption is the detection of the user's intention to interact. This study tries to make a step in that direction with use of consumer EEG sensor headset. EEG headset records in real-time data that can help to identify intention of the user based on his emotional state. For each subject EEG responses for different stimuli are recorded. Acquiring these data allows to determine the potential of EEG based intention detection. The findings are promising and with proper implementation should allow to building a new type of HCI devices.","PeriodicalId":385797,"journal":{"name":"2012 IEEE International Conference on Multimedia and Expo Workshops","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2012-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122795308","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
期刊
2012 IEEE International Conference on Multimedia and Expo Workshops
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1