首页 > 最新文献

2012 IEEE International Conference on Multimedia and Expo Workshops最新文献

英文 中文
A Novel SVM Based Food Recognition Method for Calorie Measurement Applications 一种基于支持向量机的食物识别方法在卡路里测量中的应用
Pub Date : 2012-07-09 DOI: 10.1109/ICMEW.2012.92
P. Pouladzadeh, G. Villalobos, R. Almaghrabi, S. Shirmohammadi
Emerging food classification methods play an important role in nowadays food recognition applications. For this purpose, a new recognition algorithm for food is presented, considering its shape, color, size, and texture characteristics. Using various combinations of these features, a better classification will be achieved. Based on our simulation results, the proposed algorithm recognizes food categories with an approval recognition rate of 92.6%, in average.
新兴的食品分类方法在当今的食品识别应用中发挥着重要作用。为此,提出了一种考虑食物形状、颜色、大小和质地特征的新的食物识别算法。使用这些特征的各种组合,将实现更好的分类。仿真结果表明,该算法对食品类别的平均认可识别率为92.6%。
{"title":"A Novel SVM Based Food Recognition Method for Calorie Measurement Applications","authors":"P. Pouladzadeh, G. Villalobos, R. Almaghrabi, S. Shirmohammadi","doi":"10.1109/ICMEW.2012.92","DOIUrl":"https://doi.org/10.1109/ICMEW.2012.92","url":null,"abstract":"Emerging food classification methods play an important role in nowadays food recognition applications. For this purpose, a new recognition algorithm for food is presented, considering its shape, color, size, and texture characteristics. Using various combinations of these features, a better classification will be achieved. Based on our simulation results, the proposed algorithm recognizes food categories with an approval recognition rate of 92.6%, in average.","PeriodicalId":385797,"journal":{"name":"2012 IEEE International Conference on Multimedia and Expo Workshops","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2012-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122983860","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 48
Improvisational Construction of a Context for Dynamic Implementation of Arbitrary Smart Object Federation Scenarios 动态实现任意智能对象联合场景的上下文的临时构建
Pub Date : 2012-07-09 DOI: 10.1109/ICMEW.2012.45
Jérémie Julia, Yuzuru Tanaka
This paper extends a proposed new approach to deal with smart objects or smart mobile devices by proposing a middleware framework inspired by RNA mechanisms in molecular biology. This framework represents complex application scenarios of proximity-based federation of smart objects as catalytic reaction networks. Each catalytic reaction is modeled as an RNA expression from a DNA. We introduce smart object subtype polymorphism and port subtype polymorphism in this framework to simplify the description. We also add a new condition for the description of the rules. This new condition allows us to describe physical user interactions with smart objects. This approach is used to describe new rules allowing the user to build the context improvisation ally for a reaction by physically interacting with simple mobile smart objects.
本文通过提出一个受分子生物学中RNA机制启发的中间件框架,扩展了一种处理智能对象或智能移动设备的新方法。该框架代表了基于邻近度的智能对象联合作为催化反应网络的复杂应用场景。每个催化反应都以DNA的RNA表达为模型。在此框架中引入智能对象子类型多态性和端口子类型多态性来简化描述。我们还为规则的描述添加了一个新条件。这个新条件允许我们描述用户与智能对象的物理交互。这种方法用于描述新规则,允许用户通过与简单的移动智能对象进行物理交互来即兴构建反应的上下文。
{"title":"Improvisational Construction of a Context for Dynamic Implementation of Arbitrary Smart Object Federation Scenarios","authors":"Jérémie Julia, Yuzuru Tanaka","doi":"10.1109/ICMEW.2012.45","DOIUrl":"https://doi.org/10.1109/ICMEW.2012.45","url":null,"abstract":"This paper extends a proposed new approach to deal with smart objects or smart mobile devices by proposing a middleware framework inspired by RNA mechanisms in molecular biology. This framework represents complex application scenarios of proximity-based federation of smart objects as catalytic reaction networks. Each catalytic reaction is modeled as an RNA expression from a DNA. We introduce smart object subtype polymorphism and port subtype polymorphism in this framework to simplify the description. We also add a new condition for the description of the rules. This new condition allows us to describe physical user interactions with smart objects. This approach is used to describe new rules allowing the user to build the context improvisation ally for a reaction by physically interacting with simple mobile smart objects.","PeriodicalId":385797,"journal":{"name":"2012 IEEE International Conference on Multimedia and Expo Workshops","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2012-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115704635","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
An Overview of Perceptual Processing for Digital Pictures 数字图像的感知处理综述
Pub Date : 2012-07-09 DOI: 10.1109/ICMEW.2012.27
H. Wu, Weisi Lin, Lina Karam
This paper presents an overview of state-of-the-art technologies for perceptual processing of digital pictures, as well as a discussion of the issues related to their implementation, optimization and testing. The paper begins with a brief description of the main computational modules that are used as part of a perceptual-based visual signal processing framework. Then, a number of perceptual-based visual processing techniques and applications to which perceptual models are presented, including image/video compression, visual signal quality evaluation, and computer graphics. The most significant research efforts are highlighted for each topic, and a number of issues and views are put forward regarding the related research and opportunities.
本文概述了数字图像感知处理的最新技术,并讨论了与它们的实施、优化和测试相关的问题。本文首先简要介绍了作为基于感知的视觉信号处理框架的一部分的主要计算模块。然后,介绍了一些基于感知的视觉处理技术和应用,包括图像/视频压缩、视觉信号质量评估和计算机图形学。突出了每个主题最重要的研究成果,并就相关研究和机会提出了一些问题和观点。
{"title":"An Overview of Perceptual Processing for Digital Pictures","authors":"H. Wu, Weisi Lin, Lina Karam","doi":"10.1109/ICMEW.2012.27","DOIUrl":"https://doi.org/10.1109/ICMEW.2012.27","url":null,"abstract":"This paper presents an overview of state-of-the-art technologies for perceptual processing of digital pictures, as well as a discussion of the issues related to their implementation, optimization and testing. The paper begins with a brief description of the main computational modules that are used as part of a perceptual-based visual signal processing framework. Then, a number of perceptual-based visual processing techniques and applications to which perceptual models are presented, including image/video compression, visual signal quality evaluation, and computer graphics. The most significant research efforts are highlighted for each topic, and a number of issues and views are put forward regarding the related research and opportunities.","PeriodicalId":385797,"journal":{"name":"2012 IEEE International Conference on Multimedia and Expo Workshops","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2012-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131706504","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 20
Real-Time Pitch Training System for Violin Learners 实时音高训练系统小提琴学习者
Pub Date : 2012-07-09 DOI: 10.1109/ICMEW.2012.35
Jian-Heng Wang, Siang-An Wang, Wen-Chieh Chen, Ken-Ning Chang, Herng-Yow Chen
This paper specifically targets violin learners who are working on their pitch accuracy. We employ a pitch tracking algorithm to extract the pitch played. Through volume thresholding and region detection, only parts of frames are processed. So our system can provide real-time feedback to show violin learners whether they played the right pitch. The system also provides the major scales and arpeggio scores as teaching materials, and violin learners can choose different tempos to practice, depending on their level. The user-friendly system interface allows violin learners to easily perceive the pitch differential between the pitch of the target note and the pitch played, allowing users to precisely adjust their playing. The statistical feedback records progress and analyzes error patterns, enabling violin teachers to evaluate student progress precisely, and correct common error patterns effectively.
本文专门针对小提琴学习者谁是工作在他们的音准。我们使用音调跟踪算法来提取播放的音调。通过体积阈值和区域检测,只处理部分帧。因此,我们的系统可以提供实时反馈,向小提琴学习者展示他们是否弹对了音高。该系统还提供了大调音阶和琶音乐谱作为教材,小提琴学习者可以根据自己的水平选择不同的节奏进行练习。用户友好的系统界面允许小提琴学习者很容易地感知到目标音符的音高和演奏的音高之间的音差,允许用户精确地调整他们的演奏。统计反馈记录进度,分析错误模式,使小提琴教师能够准确评估学生的进度,并有效纠正常见的错误模式。
{"title":"Real-Time Pitch Training System for Violin Learners","authors":"Jian-Heng Wang, Siang-An Wang, Wen-Chieh Chen, Ken-Ning Chang, Herng-Yow Chen","doi":"10.1109/ICMEW.2012.35","DOIUrl":"https://doi.org/10.1109/ICMEW.2012.35","url":null,"abstract":"This paper specifically targets violin learners who are working on their pitch accuracy. We employ a pitch tracking algorithm to extract the pitch played. Through volume thresholding and region detection, only parts of frames are processed. So our system can provide real-time feedback to show violin learners whether they played the right pitch. The system also provides the major scales and arpeggio scores as teaching materials, and violin learners can choose different tempos to practice, depending on their level. The user-friendly system interface allows violin learners to easily perceive the pitch differential between the pitch of the target note and the pitch played, allowing users to precisely adjust their playing. The statistical feedback records progress and analyzes error patterns, enabling violin teachers to evaluate student progress precisely, and correct common error patterns effectively.","PeriodicalId":385797,"journal":{"name":"2012 IEEE International Conference on Multimedia and Expo Workshops","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2012-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125355159","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 15
A Demonstration of a Hierarchical Multi-Layout 3D Video Browser 分层多布局3D视频浏览器的演示
Pub Date : 2012-07-09 DOI: 10.1109/ICMEW.2012.121
Christopher Müller, Martin Smole, Klaus Schöffmann
This paper demonstrates a novel 3D Video Browser (3VB) that enables interactive search within a single video as well as video collections by utilizing 3D projection and an intuitive interaction. The browsing approach is based on hierarchical search, which means that the user can split a video into several segments. The 3VB disposes a convenient interface that allows flexible arrangement of video segments in the 3D space. It allows for concurrent playback of video segments and flexible inspection of these segments at any desired level of detail through convenient user interaction.
本文介绍了一种新型的3D视频浏览器(3VB),该浏览器利用3D投影和直观的交互功能,可以在单个视频和视频集合中进行交互式搜索。浏览方法基于分层搜索,这意味着用户可以将视频分成几个部分。3VB配置了一个方便的接口,允许在3D空间中灵活地安排视频片段。它允许视频片段的并发播放,并通过方便的用户交互在任何所需的细节级别灵活地检查这些片段。
{"title":"A Demonstration of a Hierarchical Multi-Layout 3D Video Browser","authors":"Christopher Müller, Martin Smole, Klaus Schöffmann","doi":"10.1109/ICMEW.2012.121","DOIUrl":"https://doi.org/10.1109/ICMEW.2012.121","url":null,"abstract":"This paper demonstrates a novel 3D Video Browser (3VB) that enables interactive search within a single video as well as video collections by utilizing 3D projection and an intuitive interaction. The browsing approach is based on hierarchical search, which means that the user can split a video into several segments. The 3VB disposes a convenient interface that allows flexible arrangement of video segments in the 3D space. It allows for concurrent playback of video segments and flexible inspection of these segments at any desired level of detail through convenient user interaction.","PeriodicalId":385797,"journal":{"name":"2012 IEEE International Conference on Multimedia and Expo Workshops","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2012-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124756678","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Infrared and Intertial Tracking in the Immersive Audio Environment for Enhanced Military Training 增强军事训练的沉浸式音频环境中的红外和间隔跟踪
Pub Date : 2012-07-09 DOI: 10.1109/ICMEW.2012.38
Pratik Shah, A. Faza, Raghavendra Nimmala, S. Grant, W. Chapin, Robert Montgomery
The Immersive Audio Environment (IAE) was designed to provide an effective military training facility. It's efficacy at synthesizing sounds from desired directions and also the ability to synthesize moving sounds has been previously reported. This paper discusses the addition of a tracking system to evaluate subject training performance. Numerous tracking systems have been developed for tracking in immersive environments. Some examples include using head mounted web cams, visible light cameras mounted on the support structure, or even single camera tracking as in commercially available entertainment. Our system uses a combination of an existing infrared tracking system and a specially designed system of inertial tracking. This paper presents tests and results to evaluate the accuracy of the tracking system with respect to our application and verifies the efficacy of using the IAE for training enhancement.
沉浸式音频环境(IAE)的设计是为了提供有效的军事训练设施。它在合成所需方向的声音方面的功效,以及合成移动声音的能力,此前已有报道。本文讨论了增加跟踪系统来评估主体培训绩效的方法。为了在沉浸式环境中进行跟踪,已经开发了许多跟踪系统。一些例子包括使用头戴式网络摄像头,安装在支撑结构上的可见光摄像头,甚至是商业娱乐中使用的单摄像头跟踪。我们的系统结合了现有的红外跟踪系统和专门设计的惯性跟踪系统。本文给出了测试和结果,以评估我们的应用跟踪系统的准确性,并验证了使用IAE进行训练增强的有效性。
{"title":"Infrared and Intertial Tracking in the Immersive Audio Environment for Enhanced Military Training","authors":"Pratik Shah, A. Faza, Raghavendra Nimmala, S. Grant, W. Chapin, Robert Montgomery","doi":"10.1109/ICMEW.2012.38","DOIUrl":"https://doi.org/10.1109/ICMEW.2012.38","url":null,"abstract":"The Immersive Audio Environment (IAE) was designed to provide an effective military training facility. It's efficacy at synthesizing sounds from desired directions and also the ability to synthesize moving sounds has been previously reported. This paper discusses the addition of a tracking system to evaluate subject training performance. Numerous tracking systems have been developed for tracking in immersive environments. Some examples include using head mounted web cams, visible light cameras mounted on the support structure, or even single camera tracking as in commercially available entertainment. Our system uses a combination of an existing infrared tracking system and a specially designed system of inertial tracking. This paper presents tests and results to evaluate the accuracy of the tracking system with respect to our application and verifies the efficacy of using the IAE for training enhancement.","PeriodicalId":385797,"journal":{"name":"2012 IEEE International Conference on Multimedia and Expo Workshops","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2012-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128319227","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Hidden Markov Model for Event Photo Stream Segmentation 事件照片流分割的隐马尔可夫模型
Pub Date : 2012-07-09 DOI: 10.1109/ICMEW.2012.12
Jesse Prabawa Gozali, Min-Yen Kan, H. Sundaram
A photo stream is a chronological sequence of photos. Most existing photo stream segmentation methods assume that a photo stream comprises of photos from multiple events and their goal is to produce groups of photos, each corresponding to an event, i.e. they perform automatic albuming. Even if these photos are grouped by event, sifting through the abundance of photos in each event is cumbersome. To help make photos of each event more manageable, we propose a photo stream segmentation method for an event photo stream - the chronological sequence of photos of a single event - to produce groups of photos, each corresponding to a photo-worthy moment in the event. Our method is based on a hidden Markov model with parameters learned from time, EXIF metadata, and visual information from 1) training data of unlabelled, unsegmented event photo streams and 2) the event photo stream we want to segment. In an experiment with over 5000 photos from 28 personal photo sets, our method outperformed all six baselines with statistical significance (p <; 0.10 with the best baseline and p <; 0.005 with the others).
照片流是按时间顺序排列的照片。大多数现有的照片流分割方法假设一个照片流由来自多个事件的照片组成,它们的目标是生成一组照片,每组照片对应一个事件,即它们执行自动分类。即使这些照片是按事件分组的,从每个事件的大量照片中筛选也是很麻烦的。为了使每个事件的照片更易于管理,我们提出了一个事件照片流的照片流分割方法-单个事件的照片按时间顺序排列-生成一组照片,每个照片对应于事件中值得拍照的时刻。我们的方法是基于一个隐马尔可夫模型,该模型的参数来自时间、EXIF元数据和视觉信息,这些信息来自1)未标记、未分割的事件照片流的训练数据和2)我们想要分割的事件照片流。在对来自28个个人照片集的5000多张照片进行的实验中,我们的方法优于所有6个基线,具有统计学显著性(p <;0.10,最佳基线,p <;0.005与其他)。
{"title":"Hidden Markov Model for Event Photo Stream Segmentation","authors":"Jesse Prabawa Gozali, Min-Yen Kan, H. Sundaram","doi":"10.1109/ICMEW.2012.12","DOIUrl":"https://doi.org/10.1109/ICMEW.2012.12","url":null,"abstract":"A photo stream is a chronological sequence of photos. Most existing photo stream segmentation methods assume that a photo stream comprises of photos from multiple events and their goal is to produce groups of photos, each corresponding to an event, i.e. they perform automatic albuming. Even if these photos are grouped by event, sifting through the abundance of photos in each event is cumbersome. To help make photos of each event more manageable, we propose a photo stream segmentation method for an event photo stream - the chronological sequence of photos of a single event - to produce groups of photos, each corresponding to a photo-worthy moment in the event. Our method is based on a hidden Markov model with parameters learned from time, EXIF metadata, and visual information from 1) training data of unlabelled, unsegmented event photo streams and 2) the event photo stream we want to segment. In an experiment with over 5000 photos from 28 personal photo sets, our method outperformed all six baselines with statistical significance (p <; 0.10 with the best baseline and p <; 0.005 with the others).","PeriodicalId":385797,"journal":{"name":"2012 IEEE International Conference on Multimedia and Expo Workshops","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2012-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121194676","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 17
Motion Segmentation Based on 3D Histogram and Temporal Mode Selection 基于三维直方图和时间模式选择的运动分割
Pub Date : 2012-07-09 DOI: 10.1109/ICMEW.2012.90
D. Mukherjee, Q. M. J. Wu
Motion segmentation has been a well explored research topic due to its vast application area. This work proposes a real-time motion segmentation method based on 3D histogram and temporal mode selection. The temporal distribution of a video sequence consists of the motion in the foreground and the relatively immobile background. A 3D histogram provides a short-term memory of the aforementioned distribution. The temporal mode selection process involves identifying the most frequent values in the distribution and construct the background thereafter. This work provides a detailed analysis of the proposed method along with an easy-to-implement algorithm. A number of experimental results and comparisons with some of the leading algorithms are provided to show that the proposed method can provide real-time, robust and highly accurate results.
运动分割由于其广泛的应用领域,一直是一个被广泛探索的研究课题。本文提出了一种基于三维直方图和时间模式选择的实时运动分割方法。视频序列的时间分布由前景中的运动和相对不动的背景组成。3D直方图提供了上述分布的短期记忆。时间模式选择过程包括识别分布中最频繁的值并构建其后的背景。这项工作提供了一个详细的分析所提出的方法以及一个易于实现的算法。实验结果表明,该方法具有实时性、鲁棒性和较高的精度。
{"title":"Motion Segmentation Based on 3D Histogram and Temporal Mode Selection","authors":"D. Mukherjee, Q. M. J. Wu","doi":"10.1109/ICMEW.2012.90","DOIUrl":"https://doi.org/10.1109/ICMEW.2012.90","url":null,"abstract":"Motion segmentation has been a well explored research topic due to its vast application area. This work proposes a real-time motion segmentation method based on 3D histogram and temporal mode selection. The temporal distribution of a video sequence consists of the motion in the foreground and the relatively immobile background. A 3D histogram provides a short-term memory of the aforementioned distribution. The temporal mode selection process involves identifying the most frequent values in the distribution and construct the background thereafter. This work provides a detailed analysis of the proposed method along with an easy-to-implement algorithm. A number of experimental results and comparisons with some of the leading algorithms are provided to show that the proposed method can provide real-time, robust and highly accurate results.","PeriodicalId":385797,"journal":{"name":"2012 IEEE International Conference on Multimedia and Expo Workshops","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2012-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116538628","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Depth Extraction from Monocular Video Using Bidirectional Energy Minimization and Initial Depth Segmentation 基于双向能量最小化和初始深度分割的单目视频深度提取
Pub Date : 2012-07-09 DOI: 10.1109/ICMEW.2012.94
Chunyu Lin, J. D. Cock, Jürgen Slowack, P. Lambert, R. Walle
In this paper, we propose to extract depth information from a monocular video sequence. When estimating the depth of the current frame, the bidirectional energy minimization in our scheme considers both the previous frame and next frame, which promises a much more robust depth map and reduces the problems associated with occlusion to a certain extent. After getting an initial depth map from bidirectional energy minimization, we further refine the depth map using segmentation by assuming similar depth values in one segmented region. Different from other segmentation algorithms, we use initial depth information together with the original color image to get more reliable segmented regions. Finally, detecting the sky region using a dark channel prior is employed to correct some possibly wrong depth values for outdoor video. The experimental results are much more accurate compared with the state-of-the-art algorithms.
本文提出了从单目视频序列中提取深度信息的方法。在估计当前帧的深度时,我们的方案中的双向能量最小化同时考虑了前一帧和下一帧,这保证了一个更加鲁棒的深度图,并在一定程度上减少了与遮挡相关的问题。在从双向能量最小化中得到初始深度图之后,我们通过在一个分割区域中假设相似的深度值来进一步细化深度图。与其他分割算法不同的是,我们将初始深度信息与原始彩色图像结合使用,得到更可靠的分割区域。最后,利用暗通道先验检测天空区域,修正一些可能错误的户外视频深度值。与现有算法相比,实验结果更加准确。
{"title":"Depth Extraction from Monocular Video Using Bidirectional Energy Minimization and Initial Depth Segmentation","authors":"Chunyu Lin, J. D. Cock, Jürgen Slowack, P. Lambert, R. Walle","doi":"10.1109/ICMEW.2012.94","DOIUrl":"https://doi.org/10.1109/ICMEW.2012.94","url":null,"abstract":"In this paper, we propose to extract depth information from a monocular video sequence. When estimating the depth of the current frame, the bidirectional energy minimization in our scheme considers both the previous frame and next frame, which promises a much more robust depth map and reduces the problems associated with occlusion to a certain extent. After getting an initial depth map from bidirectional energy minimization, we further refine the depth map using segmentation by assuming similar depth values in one segmented region. Different from other segmentation algorithms, we use initial depth information together with the original color image to get more reliable segmented regions. Finally, detecting the sky region using a dark channel prior is employed to correct some possibly wrong depth values for outdoor video. The experimental results are much more accurate compared with the state-of-the-art algorithms.","PeriodicalId":385797,"journal":{"name":"2012 IEEE International Conference on Multimedia and Expo Workshops","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2012-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115818953","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Surround Sound Using Variable-Ambisonics and Variable-Polar Pattern Theories 利用变混响和变极模式理论的环绕声
Pub Date : 2012-07-09 DOI: 10.1109/ICMEW.2012.124
M. J. Morrell, J. Reiss, Sonia Wilkie
This paper details the technologies displayed by the authors at ICME 2012. The technologies are those of Variable-Ambisonics and Variable-Polar Pattern Reproduction. These technologies are demonstrated using virtual 2-dimensional speaker layouts via binaural headphone reproduction. The technologies offer benefits over standard pair-wise panning of surround sound by offering on-the-fly changing of speaker count, source width control and per sound source rendering. The technologies also overcome problems of Ambisonics and Ambisonic based reproduction methods by allowing each sound source to be individually rendered by decoder type, of an arbitrary rather than fixed order and the mixing of various orders within the same reproduction system.
本文详细介绍了作者在ICME 2012上展示的技术。这两种技术分别是变双音技术和变极型复制技术。这些技术通过双耳耳机再现使用虚拟二维扬声器布局进行演示。这些技术通过提供实时改变扬声器数量、声源宽度控制和每个声源渲染,比标准的双对环绕声平移更有优势。该技术还通过允许每个声源按解码器类型单独呈现,任意而不是固定顺序,以及在同一复制系统中混合各种顺序,克服了双声和基于双声的复制方法的问题。
{"title":"Surround Sound Using Variable-Ambisonics and Variable-Polar Pattern Theories","authors":"M. J. Morrell, J. Reiss, Sonia Wilkie","doi":"10.1109/ICMEW.2012.124","DOIUrl":"https://doi.org/10.1109/ICMEW.2012.124","url":null,"abstract":"This paper details the technologies displayed by the authors at ICME 2012. The technologies are those of Variable-Ambisonics and Variable-Polar Pattern Reproduction. These technologies are demonstrated using virtual 2-dimensional speaker layouts via binaural headphone reproduction. The technologies offer benefits over standard pair-wise panning of surround sound by offering on-the-fly changing of speaker count, source width control and per sound source rendering. The technologies also overcome problems of Ambisonics and Ambisonic based reproduction methods by allowing each sound source to be individually rendered by decoder type, of an arbitrary rather than fixed order and the mixing of various orders within the same reproduction system.","PeriodicalId":385797,"journal":{"name":"2012 IEEE International Conference on Multimedia and Expo Workshops","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2012-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125163223","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
2012 IEEE International Conference on Multimedia and Expo Workshops
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1