首页 > 最新文献

2013 IEEE International Symposium on Broadband Multimedia Systems and Broadcasting (BMSB)最新文献

英文 中文
An i-Vector Representation of Acoustic Environments for Audio-Based Video Event Detection on User Generated Content 基于用户生成内容的基于音频的视频事件检测声学环境的i向量表示
Benjamin Elizalde, Howard Lei, G. Friedland
Audio-based video event detection (VED) on user-generated content (UGC) aims to find videos that show an observable event such as a wedding ceremony or birthday party rather than a sound, such as music, clapping or singing. The difficulty of video content analysis on UGC lies in the acoustic variability and lack of structure of the data. The UGC task has been explored mainly by computer vision, but can be benefited by the used of audio. The i-vector system is state-of-the-art in Speaker Verification, and is outperforming a conventional Gaussian Mixture Model (GMM)-based approach. The system compensates for undesired acoustic variability and extracts information from the acoustic environment, making it a meaningful choice for detection on UGC. This paper employs the i-vector-based system for audio-based VED on UGC and expands the understanding of the system on the task. It also includes a performance comparison with the conventional GMM-based and state-of-the-art Random Forest (RF)-based systems. The i-vector system aids audio-based event detection by addressing UGC audio characteristics. It outperforms the GMM-based system, and is competitive with the RF-based system in terms of the Missed Detection (MD) rate at 4% and 2.8% False Alarm (FA) rates, and complements the RF-based system by demonstrating slightly improvement in combination over the standalone systems.
针对用户生成内容(UGC)的基于音频的视频事件检测(VED)旨在发现显示可观察到的事件(如婚礼或生日派对)的视频,而不是音乐、鼓掌或唱歌等声音。UGC视频内容分析的难点在于数据的声学变异性和缺乏结构化。UGC任务主要是通过计算机视觉来探索的,但音频的使用也可以从中受益。i向量系统是最先进的说话人验证,并优于传统的高斯混合模型(GMM)为基础的方法。该系统补偿了不期望的声学变异性,并从声学环境中提取信息,使其成为检测UGC的有意义的选择。本文采用基于i向量的系统实现基于UGC的基于音频的视频生成,扩展了系统对任务的理解。它还包括与传统的基于gmm和最先进的基于随机森林(RF)的系统的性能比较。i-vector系统通过处理UGC音频特征来辅助基于音频的事件检测。它优于基于gmm的系统,并且在4%的未检出率(MD)和2.8%的误报率(FA)方面与基于rf的系统具有竞争力,并且通过与独立系统的组合略有改进来补充基于rf的系统。
{"title":"An i-Vector Representation of Acoustic Environments for Audio-Based Video Event Detection on User Generated Content","authors":"Benjamin Elizalde, Howard Lei, G. Friedland","doi":"10.1109/ISM.2013.27","DOIUrl":"https://doi.org/10.1109/ISM.2013.27","url":null,"abstract":"Audio-based video event detection (VED) on user-generated content (UGC) aims to find videos that show an observable event such as a wedding ceremony or birthday party rather than a sound, such as music, clapping or singing. The difficulty of video content analysis on UGC lies in the acoustic variability and lack of structure of the data. The UGC task has been explored mainly by computer vision, but can be benefited by the used of audio. The i-vector system is state-of-the-art in Speaker Verification, and is outperforming a conventional Gaussian Mixture Model (GMM)-based approach. The system compensates for undesired acoustic variability and extracts information from the acoustic environment, making it a meaningful choice for detection on UGC. This paper employs the i-vector-based system for audio-based VED on UGC and expands the understanding of the system on the task. It also includes a performance comparison with the conventional GMM-based and state-of-the-art Random Forest (RF)-based systems. The i-vector system aids audio-based event detection by addressing UGC audio characteristics. It outperforms the GMM-based system, and is competitive with the RF-based system in terms of the Missed Detection (MD) rate at 4% and 2.8% False Alarm (FA) rates, and complements the RF-based system by demonstrating slightly improvement in combination over the standalone systems.","PeriodicalId":6311,"journal":{"name":"2013 IEEE International Symposium on Broadband Multimedia Systems and Broadcasting (BMSB)","volume":"61 1","pages":"114-117"},"PeriodicalIF":0.0,"publicationDate":"2013-12-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82634096","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 17
A Two-Phase Generation Model for Automatic Image Annotation 图像自动标注的两阶段生成模型
Liang Xie, Peng Pan, Yansheng Lu, Shixun Wang, Tong Zhu, Haijiao Xu, Deng Chen
Automatic image annotation is an important task for multimedia retrieval. By allocating relevant words to un-annotated images, these images can be retrieved in response to textual queries. There are many researches on the problem of image annotation and most of them construct models based on joint probability or posterior probabilities of words. In this paper we estimate the probabilities that words generate the images, and propose a two-phase generation model for the generation procedure. Each word first generates its related words, then these words generate an un-annotated image, and the relation between the words and the un-annotated image is obtained by the probability of the two-phase generation. The textual words usually contain more semantic information than visual content of images, thus the probabilities that words generate images is more reliable than the probability that images generate words. As a result, our model estimates the more reliable probability than other probabilistic methods for image annotation. The other advantage of our model is the relation of words is taken into consideration. The experimental results on Corel 5K and MIR Flickr demonstrate that our model performs better than other previous methods. And two-phase generation which considering word's relation for annotation is better than one-phase generation which only consider the relation between words and images. Moreover, the methods which estimate the generative probability obtain better performance than SVM which estimates the posterior probability.
图像自动标注是多媒体检索的重要任务。通过将相关单词分配给未注释的图像,可以在响应文本查询时检索这些图像。关于图像标注问题的研究很多,大多数都是基于单词的联合概率或后验概率来构建模型。本文估计了文字生成图像的概率,提出了一种两阶段生成模型。每个单词首先生成与其相关的单词,然后这些单词生成一个未注释的图像,通过两阶段生成的概率得到单词与未注释图像之间的关系。文本单词通常比图像的视觉内容包含更多的语义信息,因此单词生成图像的概率比图像生成单词的概率更可靠。因此,我们的模型估计的概率比其他概率方法更可靠。我们模型的另一个优点是考虑了单词之间的关系。在Corel 5K和MIR Flickr上的实验结果表明,该模型的性能优于以往的方法。考虑词与图像关系的两阶段生成比只考虑词与图像关系的一阶段生成效果更好。此外,估计生成概率的方法比估计后验概率的支持向量机获得了更好的性能。
{"title":"A Two-Phase Generation Model for Automatic Image Annotation","authors":"Liang Xie, Peng Pan, Yansheng Lu, Shixun Wang, Tong Zhu, Haijiao Xu, Deng Chen","doi":"10.1109/ISM.2013.33","DOIUrl":"https://doi.org/10.1109/ISM.2013.33","url":null,"abstract":"Automatic image annotation is an important task for multimedia retrieval. By allocating relevant words to un-annotated images, these images can be retrieved in response to textual queries. There are many researches on the problem of image annotation and most of them construct models based on joint probability or posterior probabilities of words. In this paper we estimate the probabilities that words generate the images, and propose a two-phase generation model for the generation procedure. Each word first generates its related words, then these words generate an un-annotated image, and the relation between the words and the un-annotated image is obtained by the probability of the two-phase generation. The textual words usually contain more semantic information than visual content of images, thus the probabilities that words generate images is more reliable than the probability that images generate words. As a result, our model estimates the more reliable probability than other probabilistic methods for image annotation. The other advantage of our model is the relation of words is taken into consideration. The experimental results on Corel 5K and MIR Flickr demonstrate that our model performs better than other previous methods. And two-phase generation which considering word's relation for annotation is better than one-phase generation which only consider the relation between words and images. Moreover, the methods which estimate the generative probability obtain better performance than SVM which estimates the posterior probability.","PeriodicalId":6311,"journal":{"name":"2013 IEEE International Symposium on Broadband Multimedia Systems and Broadcasting (BMSB)","volume":"15 1","pages":"155-162"},"PeriodicalIF":0.0,"publicationDate":"2013-12-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82759233","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Adaptive HTTP Streaming Utilizing Temporal Sub-layers of High Efficiency Video Coding (HEVC) 利用高效视频编码(HEVC)时间子层的自适应HTTP流
S. Deshpande
The newly approved High Efficiency Video Coding Standard (HEVC) includes temporal sub-layering feature which provides temporal scalability. Two types of pictures - Temporal Sub-layer Access Pictures and Step-wise Temporal Sub-layer Access Pictures are provided for this purpose. This paper utilizes the temporal scalability of HEVC to provide bandwidth adaptive HTTP streaming to clients. We describe our HTTP streaming algorithm, which is media timeline aware. Temporal sub-layers are switched on the server side dynamically. We performed subjective tests to determine user perception regarding acceptable frame rates when using temporal scalability of HEVC. These results are used to control the algorithm's temporal switching behavior to provide a good quality of experience to the user. We applied Internet and 3GPP error-delay patterns to validate the performance of our algorithm.
新批准的高效视频编码标准(High Efficiency Video Coding Standard, HEVC)包含了时间子分层特性,提供了时间可扩展性。为此提供了两种类型的图片——时序子层访问图片和分步时序子层访问图片。本文利用HEVC的时间可扩展性为客户端提供带宽自适应的HTTP流。我们描述了我们的HTTP流算法,它是媒体时间轴感知的。时间子层在服务器端动态切换。我们进行了主观测试,以确定用户在使用HEVC的时间可扩展性时对可接受帧率的感知。这些结果用于控制算法的时间切换行为,为用户提供高质量的体验。我们应用互联网和3GPP的错误延迟模式来验证我们的算法的性能。
{"title":"Adaptive HTTP Streaming Utilizing Temporal Sub-layers of High Efficiency Video Coding (HEVC)","authors":"S. Deshpande","doi":"10.1109/ISM.2013.73","DOIUrl":"https://doi.org/10.1109/ISM.2013.73","url":null,"abstract":"The newly approved High Efficiency Video Coding Standard (HEVC) includes temporal sub-layering feature which provides temporal scalability. Two types of pictures - Temporal Sub-layer Access Pictures and Step-wise Temporal Sub-layer Access Pictures are provided for this purpose. This paper utilizes the temporal scalability of HEVC to provide bandwidth adaptive HTTP streaming to clients. We describe our HTTP streaming algorithm, which is media timeline aware. Temporal sub-layers are switched on the server side dynamically. We performed subjective tests to determine user perception regarding acceptable frame rates when using temporal scalability of HEVC. These results are used to control the algorithm's temporal switching behavior to provide a good quality of experience to the user. We applied Internet and 3GPP error-delay patterns to validate the performance of our algorithm.","PeriodicalId":6311,"journal":{"name":"2013 IEEE International Symposium on Broadband Multimedia Systems and Broadcasting (BMSB)","volume":"13 1","pages":"384-390"},"PeriodicalIF":0.0,"publicationDate":"2013-12-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"86585977","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Constraint Satisfaction Programming for Video Summarization 视频摘要的约束满足规划
Sid-Ahmed Berrani, Haykel Boukadida, P. Gros
This paper addresses the problem of automatic video summarization. The proposed solution relies on constraint satisfaction programming (CSP). Summary generation rules are expressed as constraints and the summary is created using the CSP solver given the input video, its audio-visual features and possibly user parameters (like the desired duration). The solution clearly separates production rules from the generation algorithm, which in practice allows users to easily express their constraints and preferences and also to modify them w.r.t. the target application. The solution is extensively evaluated in the context of tennis match summarization.
本文主要研究视频的自动摘要问题。提出的解决方案依赖于约束满足规划(CSP)。摘要生成规则表示为约束,摘要是在给定输入视频、其视听特征和可能的用户参数(如期望的持续时间)的情况下使用CSP求解器创建的。该解决方案清楚地将生产规则与生成算法分离开来,这在实践中允许用户轻松地表达他们的约束和偏好,并在目标应用程序中修改它们。该解决方案在网球比赛总结的背景下进行了广泛的评估。
{"title":"Constraint Satisfaction Programming for Video Summarization","authors":"Sid-Ahmed Berrani, Haykel Boukadida, P. Gros","doi":"10.1109/ISM.2013.38","DOIUrl":"https://doi.org/10.1109/ISM.2013.38","url":null,"abstract":"This paper addresses the problem of automatic video summarization. The proposed solution relies on constraint satisfaction programming (CSP). Summary generation rules are expressed as constraints and the summary is created using the CSP solver given the input video, its audio-visual features and possibly user parameters (like the desired duration). The solution clearly separates production rules from the generation algorithm, which in practice allows users to easily express their constraints and preferences and also to modify them w.r.t. the target application. The solution is extensively evaluated in the context of tennis match summarization.","PeriodicalId":6311,"journal":{"name":"2013 IEEE International Symposium on Broadband Multimedia Systems and Broadcasting (BMSB)","volume":"29 1","pages":"195-202"},"PeriodicalIF":0.0,"publicationDate":"2013-12-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"89376710","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
Development and Preliminary Evaluation of an Interactive System to Support CAD Teaching CAD教学交互式系统的开发与初步评价
S. Akhtar, S. Warburton, W. Xu
It has been a goal for many researchers to make education more enjoyable, attractive and effective through the use of multimedia technology [1]. Achieving this goal requires rich interactive communication between students and tutors and a clear understanding of the educational environment. Despite the availability of wide range of commercial systems that can support multimedia within the classroom there remains a gap for an innovative system that can provide a blended approach to providing support for live teaching sessions. This paper introduces Surrey Connect, a bespoke system, designed to enhance the teaching and learning experience in large classroom settings. It provides lecture recording with selective replay, implicit and explicit responses and multiple tutors support. In addition to its interactive user interface, Surrey Connect acts an early warning system by monitoring learners' in-class behaviour, presenting it in an interactive dashboard to the tutor and suggesting interventions in the light of rule-based programmable knowledge. The system has been tested in a real classroom environment of Computer Aided Design course with more than 150 students and received over 60% positive feedback from the students.
通过使用多媒体技术,使教育更有趣、更有吸引力和更有效,一直是许多研究者的目标。实现这一目标需要学生和导师之间丰富的互动交流以及对教育环境的清晰理解。尽管有各种各样的商业系统可以支持课堂内的多媒体,但仍然缺乏一种创新的系统,可以提供混合的方法来为现场教学会议提供支持。本文介绍了Surrey Connect,一个定制系统,旨在提高教学和学习的经验,在大型教室设置。它提供讲座录音与选择性重播,隐式和明确的反应和多个导师的支持。除了交互式用户界面之外,Surrey Connect还充当了一个预警系统,通过监测学习者的课堂行为,将其呈现在交互式仪表板中给导师,并根据基于规则的可编程知识提出干预建议。该系统已在150多名学生参加的计算机辅助设计课程的真实课堂环境中进行了测试,学生的积极反馈率超过60%。
{"title":"Development and Preliminary Evaluation of an Interactive System to Support CAD Teaching","authors":"S. Akhtar, S. Warburton, W. Xu","doi":"10.1109/ISM.2013.92","DOIUrl":"https://doi.org/10.1109/ISM.2013.92","url":null,"abstract":"It has been a goal for many researchers to make education more enjoyable, attractive and effective through the use of multimedia technology [1]. Achieving this goal requires rich interactive communication between students and tutors and a clear understanding of the educational environment. Despite the availability of wide range of commercial systems that can support multimedia within the classroom there remains a gap for an innovative system that can provide a blended approach to providing support for live teaching sessions. This paper introduces Surrey Connect, a bespoke system, designed to enhance the teaching and learning experience in large classroom settings. It provides lecture recording with selective replay, implicit and explicit responses and multiple tutors support. In addition to its interactive user interface, Surrey Connect acts an early warning system by monitoring learners' in-class behaviour, presenting it in an interactive dashboard to the tutor and suggesting interventions in the light of rule-based programmable knowledge. The system has been tested in a real classroom environment of Computer Aided Design course with more than 150 students and received over 60% positive feedback from the students.","PeriodicalId":6311,"journal":{"name":"2013 IEEE International Symposium on Broadband Multimedia Systems and Broadcasting (BMSB)","volume":"36 1","pages":"480-485"},"PeriodicalIF":0.0,"publicationDate":"2013-12-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"85687627","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Intelligent and Selective Video Frames Discarding Policies for Improving Video Quality over Wired/Wireless Networks 提高有线/无线网络视频质量的智能选择性视频弃帧策略
Khalid A. Darabkh, Abeer M. Awad, A. Khalifeh
Although IEEE 802.11 Wireless LAN (WLAN) is of a great interest nowadays, it lacks the efficient capability of supporting real-time streaming which is due to mainly the contention nature of wireless media. In this paper, we extend our earlier work concerning improving video traffic over wireless networks through effectively studying the dependencies between video frames and their implications on the overall network performance. In other words, we propose very efficient and novel algorithms that aim to minimize the cost of possible losses by intelligently and selectively discard frames based on their contribution to picture quality, namely, partial and intelligent-partial frame discarding policies considering the dependencies between video frames. The performance metrics that are employed to evaluate the performance of the proposed algorithms include the rate of non-decodable frames and peak signal-to-noise ratio (PSNR). Our results are promising and show significant improvements in the perceived video quality over what is relevant in the current literature.
虽然目前IEEE 802.11无线局域网(WLAN)备受关注,但由于无线媒体的争用特性,它缺乏支持实时流的有效能力。在本文中,我们通过有效地研究视频帧之间的依赖关系及其对整体网络性能的影响,扩展了我们关于改善无线网络视频流量的早期工作。换句话说,我们提出了非常高效和新颖的算法,旨在通过基于对图像质量的贡献智能和选择性地丢弃帧来最小化可能损失的成本,即考虑视频帧之间依赖关系的部分和智能部分帧丢弃策略。用于评估所提出算法性能的性能指标包括不可解码帧率和峰值信噪比(PSNR)。我们的结果是有希望的,并且在感知视频质量方面显示出与当前文献相关的显着改进。
{"title":"Intelligent and Selective Video Frames Discarding Policies for Improving Video Quality over Wired/Wireless Networks","authors":"Khalid A. Darabkh, Abeer M. Awad, A. Khalifeh","doi":"10.1109/ISM.2013.57","DOIUrl":"https://doi.org/10.1109/ISM.2013.57","url":null,"abstract":"Although IEEE 802.11 Wireless LAN (WLAN) is of a great interest nowadays, it lacks the efficient capability of supporting real-time streaming which is due to mainly the contention nature of wireless media. In this paper, we extend our earlier work concerning improving video traffic over wireless networks through effectively studying the dependencies between video frames and their implications on the overall network performance. In other words, we propose very efficient and novel algorithms that aim to minimize the cost of possible losses by intelligently and selectively discard frames based on their contribution to picture quality, namely, partial and intelligent-partial frame discarding policies considering the dependencies between video frames. The performance metrics that are employed to evaluate the performance of the proposed algorithms include the rate of non-decodable frames and peak signal-to-noise ratio (PSNR). Our results are promising and show significant improvements in the perceived video quality over what is relevant in the current literature.","PeriodicalId":6311,"journal":{"name":"2013 IEEE International Symposium on Broadband Multimedia Systems and Broadcasting (BMSB)","volume":"17 1","pages":"297-300"},"PeriodicalIF":0.0,"publicationDate":"2013-12-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"80183755","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Dynamic Recombination of Evolving Guitar Sounds (DREGS): A Genetic Algorithm Approach to Guitar Synthesizer Control 演化吉他声音的动态重组(DREGS):一种吉他合成器控制的遗传算法
Timothy M. Walker, Sean Whalen
A system is described which integrates multiple hardware interfaces and software packages in order to control the parameters of a guitar synthesizer in real time. An interactive genetic algorithm is developed in order to create and explore parameter settings, and a mobile device wirelessly sets the fitness values. The synthesizer parameters are represented as genes within an individual, and individuals dynamically interact within a population as the user rates the resulting sounds by changing orientation.
介绍了一种集成多个硬件接口和软件包的系统,以实现对吉他合成器参数的实时控制。开发了一种交互式遗传算法来创建和探索参数设置,并通过移动设备无线设置适应度值。合成器参数表示为个体内的基因,当用户通过改变方向对产生的声音进行评级时,个体在群体内动态交互。
{"title":"Dynamic Recombination of Evolving Guitar Sounds (DREGS): A Genetic Algorithm Approach to Guitar Synthesizer Control","authors":"Timothy M. Walker, Sean Whalen","doi":"10.1109/ISM.2013.47","DOIUrl":"https://doi.org/10.1109/ISM.2013.47","url":null,"abstract":"A system is described which integrates multiple hardware interfaces and software packages in order to control the parameters of a guitar synthesizer in real time. An interactive genetic algorithm is developed in order to create and explore parameter settings, and a mobile device wirelessly sets the fitness values. The synthesizer parameters are represented as genes within an individual, and individuals dynamically interact within a population as the user rates the resulting sounds by changing orientation.","PeriodicalId":6311,"journal":{"name":"2013 IEEE International Symposium on Broadband Multimedia Systems and Broadcasting (BMSB)","volume":"73 9 1","pages":"248-254"},"PeriodicalIF":0.0,"publicationDate":"2013-12-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"76381438","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Longitudinal Characterization of Breast Morphology during Reconstructive Surgery 乳房重建手术中乳房形态的纵向特征
Lijuan Zhao, Shishir K. Shah, F. Merchant
Quantitative analysis of breast morphology facilitates pre-operative planning and post-operative outcome assessments in breast reconstruction. Our project is developing algorithms to quantify changes in local breast morphology occurring over time. The project encompasses three topics: (1) Three-dimensional (3D) images registration, (2) Breast contour detection, and (3) Quantitative analysis of local breast morphology changes. We developed a semi-automated 3D image registration algorithm. We have also developed an approach to directly compute breast contour on 3D images. In the future, we will improve existing and develop additional algorithms to fulfill our project goals.
乳房形态学的定量分析有助于乳房重建的术前计划和术后结果评估。我们的项目是开发算法来量化局部乳房形态随时间的变化。该项目包括三个主题:(1)三维(3D)图像配准;(2)乳房轮廓检测;(3)局部乳房形态变化的定量分析。我们开发了一种半自动三维图像配准算法。我们还开发了一种直接在3D图像上计算乳房轮廓的方法。在未来,我们将改进现有的并开发额外的算法来实现我们的项目目标。
{"title":"Longitudinal Characterization of Breast Morphology during Reconstructive Surgery","authors":"Lijuan Zhao, Shishir K. Shah, F. Merchant","doi":"10.1109/ISM.2013.79","DOIUrl":"https://doi.org/10.1109/ISM.2013.79","url":null,"abstract":"Quantitative analysis of breast morphology facilitates pre-operative planning and post-operative outcome assessments in breast reconstruction. Our project is developing algorithms to quantify changes in local breast morphology occurring over time. The project encompasses three topics: (1) Three-dimensional (3D) images registration, (2) Breast contour detection, and (3) Quantitative analysis of local breast morphology changes. We developed a semi-automated 3D image registration algorithm. We have also developed an approach to directly compute breast contour on 3D images. In the future, we will improve existing and develop additional algorithms to fulfill our project goals.","PeriodicalId":6311,"journal":{"name":"2013 IEEE International Symposium on Broadband Multimedia Systems and Broadcasting (BMSB)","volume":"9 1","pages":"407-408"},"PeriodicalIF":0.0,"publicationDate":"2013-12-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"85869288","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
VideoTopic: Content-Based Video Recommendation Using a Topic Model VideoTopic:使用主题模型的基于内容的视频推荐
Qiusha Zhu, M. Shyu, Haohong Wang
Most video recommender systems limit the content to the metadata associated with the videos, which could lead to poor results since metadata is not always available or correct. Meanwhile, the visual information of videos is typically not fully explored, which is especially important for recommending new items with limited metadata information. In this paper, a novel content-based video recommendation framework, called Video Topic, that utilizes a topic model is proposed. It decomposes the recommendation process into video representation and recommendation generation. It aims to capture user interests in videos by using a topic model to represent the videos, and then generates recommendations by finding those videos that most fit to the topic distribution of the user interests. Experimental results on the Movie Lens dataset validate the effectiveness of Video Topic by evaluating each of its components and the whole framework.
大多数视频推荐系统将内容限制在与视频相关的元数据上,这可能会导致糟糕的结果,因为元数据并不总是可用或正确的。同时,视频的视觉信息通常没有被充分挖掘,这对于元数据信息有限的新项目推荐尤为重要。本文利用主题模型,提出了一种新的基于内容的视频推荐框架——视频主题。将推荐过程分解为视频表示和推荐生成。它的目的是通过使用主题模型来表示视频来捕获用户对视频的兴趣,然后通过找到最适合用户兴趣主题分布的视频来生成推荐。在Movie Lens数据集上的实验结果通过评估视频主题的每个组成部分和整个框架来验证视频主题的有效性。
{"title":"VideoTopic: Content-Based Video Recommendation Using a Topic Model","authors":"Qiusha Zhu, M. Shyu, Haohong Wang","doi":"10.1109/ISM.2013.41","DOIUrl":"https://doi.org/10.1109/ISM.2013.41","url":null,"abstract":"Most video recommender systems limit the content to the metadata associated with the videos, which could lead to poor results since metadata is not always available or correct. Meanwhile, the visual information of videos is typically not fully explored, which is especially important for recommending new items with limited metadata information. In this paper, a novel content-based video recommendation framework, called Video Topic, that utilizes a topic model is proposed. It decomposes the recommendation process into video representation and recommendation generation. It aims to capture user interests in videos by using a topic model to represent the videos, and then generates recommendations by finding those videos that most fit to the topic distribution of the user interests. Experimental results on the Movie Lens dataset validate the effectiveness of Video Topic by evaluating each of its components and the whole framework.","PeriodicalId":6311,"journal":{"name":"2013 IEEE International Symposium on Broadband Multimedia Systems and Broadcasting (BMSB)","volume":"1 1","pages":"219-222"},"PeriodicalIF":0.0,"publicationDate":"2013-12-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"86446020","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 42
Contextualized Privacy Filters in Video Surveillance Using Crowd Density Maps 基于人群密度图的视频监控情境化隐私过滤器
H. Fradi, A. Melle, J. Dugelay
The widespread growth in the adoption of digital video surveillance systems emphasizes the need for privacy preservation video analytics techniques. While these privacy aspects have shown big interest in recent years, little importance has been given to the concept of context-aware privacy protection filters. In this paper, we specifically focus on the dependency between privacy preservation and crowd density. We show that additional information about the crowd density in the scene can be used in order to adjust the level of privacy protection according to the local needs. This additional information cue consists of modeling time-varying dynamics of the crowd density using local features as an observation of a probabilistic crowd function. It also involves a feature tracking step which enables excluding feature points on the background. This process is favourable for the later density function estimation since the influence of features irrelevant to the underlying crowd density is removed. Then, the protection level of personal privacy in videos is adapted according to the crowd density. Afterwards, a framework for objective evaluation of the contextualized protection filters is proposed. The effectiveness of the proposed context-aware privacy filters has been demonstrated by assessing the intelligibility vs. privacy trade-off using videos from different crowd datasets.
数字视频监控系统的广泛采用强调了对隐私保护视频分析技术的需求。虽然近年来这些隐私方面表现出了很大的兴趣,但上下文感知隐私保护过滤器的概念却很少受到重视。在本文中,我们特别关注隐私保护与人群密度之间的依赖关系。我们表明,可以使用场景中人群密度的附加信息,以便根据当地需求调整隐私保护水平。这个额外的信息线索包括使用局部特征作为概率人群函数的观察来建模人群密度的时变动态。它还包括一个特征跟踪步骤,可以排除背景上的特征点。这个过程有利于后期的密度函数估计,因为与底层人群密度无关的特征的影响被消除了。然后,根据人群密度调整视频中个人隐私的保护级别。在此基础上,提出了一种情境化保护滤波器的客观评价框架。通过使用来自不同人群数据集的视频评估可理解性与隐私权衡,证明了所提出的上下文感知隐私过滤器的有效性。
{"title":"Contextualized Privacy Filters in Video Surveillance Using Crowd Density Maps","authors":"H. Fradi, A. Melle, J. Dugelay","doi":"10.1109/ISM.2013.23","DOIUrl":"https://doi.org/10.1109/ISM.2013.23","url":null,"abstract":"The widespread growth in the adoption of digital video surveillance systems emphasizes the need for privacy preservation video analytics techniques. While these privacy aspects have shown big interest in recent years, little importance has been given to the concept of context-aware privacy protection filters. In this paper, we specifically focus on the dependency between privacy preservation and crowd density. We show that additional information about the crowd density in the scene can be used in order to adjust the level of privacy protection according to the local needs. This additional information cue consists of modeling time-varying dynamics of the crowd density using local features as an observation of a probabilistic crowd function. It also involves a feature tracking step which enables excluding feature points on the background. This process is favourable for the later density function estimation since the influence of features irrelevant to the underlying crowd density is removed. Then, the protection level of personal privacy in videos is adapted according to the crowd density. Afterwards, a framework for objective evaluation of the contextualized protection filters is proposed. The effectiveness of the proposed context-aware privacy filters has been demonstrated by assessing the intelligibility vs. privacy trade-off using videos from different crowd datasets.","PeriodicalId":6311,"journal":{"name":"2013 IEEE International Symposium on Broadband Multimedia Systems and Broadcasting (BMSB)","volume":"1 1","pages":"92-99"},"PeriodicalIF":0.0,"publicationDate":"2013-12-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"88620978","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
期刊
2013 IEEE International Symposium on Broadband Multimedia Systems and Broadcasting (BMSB)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1