Frame clustering technique towards single video summarization

Priyamvada R. Sachan, Keshaveni
{"title":"Frame clustering technique towards single video summarization","authors":"Priyamvada R. Sachan, Keshaveni","doi":"10.1109/CCIP.2016.7802877","DOIUrl":null,"url":null,"abstract":"Recent advances in technology, multimedia and social networking sites have led to a massive growth in web video content available for the general population. This results in information overload and management problem of the same. In this context, video summarization plays an important role as it aims to reduce the content size of video and yet present the important semantic concepts in the video. This gives an opportunity to reorganize video content in most succinct form for efficient and on- demand user consumption. Video summarization in its true sense is a hard problem as it involves domain specific semantic understanding of video content and user expectations. Most of the existing approaches relies on segmenting video into contiguous shots & selecting one or more keyframes from each shot and present these keyframes as summary. Such approaches may work well if independent concepts in video appear only once. However, in videos where same concepts are repeated multiple times, these existing approaches may pick repeating summary frames belonging to same concepts. In this paper, we present a novel frame clustering approach for generating very concise summaries by grouping all frames of similar concepts together irrespective of their occurrence sequence. The proposed approach is aimed towards large videos in domains like travel guide, documentaries, dramas where video revolves around few repeating concepts. The approach utilizes multiple video features in a generic way for frame-similarity determination and is extensible for multi-video summarization. Experimental comparative results substantiate the efficiency of the proposed approach in generating concise video summaries on videos with repeating concepts.","PeriodicalId":354589,"journal":{"name":"2016 Second International Conference on Cognitive Computing and Information Processing (CCIP)","volume":"15 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2016-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"4","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2016 Second International Conference on Cognitive Computing and Information Processing (CCIP)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/CCIP.2016.7802877","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 4

Abstract

Recent advances in technology, multimedia and social networking sites have led to a massive growth in web video content available for the general population. This results in information overload and management problem of the same. In this context, video summarization plays an important role as it aims to reduce the content size of video and yet present the important semantic concepts in the video. This gives an opportunity to reorganize video content in most succinct form for efficient and on- demand user consumption. Video summarization in its true sense is a hard problem as it involves domain specific semantic understanding of video content and user expectations. Most of the existing approaches relies on segmenting video into contiguous shots & selecting one or more keyframes from each shot and present these keyframes as summary. Such approaches may work well if independent concepts in video appear only once. However, in videos where same concepts are repeated multiple times, these existing approaches may pick repeating summary frames belonging to same concepts. In this paper, we present a novel frame clustering approach for generating very concise summaries by grouping all frames of similar concepts together irrespective of their occurrence sequence. The proposed approach is aimed towards large videos in domains like travel guide, documentaries, dramas where video revolves around few repeating concepts. The approach utilizes multiple video features in a generic way for frame-similarity determination and is extensible for multi-video summarization. Experimental comparative results substantiate the efficiency of the proposed approach in generating concise video summaries on videos with repeating concepts.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
面向单个视频摘要的帧聚类技术
最近技术、多媒体和社交网站的进步导致了大众可获得的网络视频内容的大量增长。这就造成了信息过载和管理问题。在这种情况下,视频摘要扮演着重要的角色,因为它旨在减少视频的内容大小,同时又能呈现视频中重要的语义概念。这提供了一个以最简洁的形式重组视频内容的机会,以实现高效和按需的用户消费。真正意义上的视频摘要是一个难题,因为它涉及到对视频内容和用户期望的特定领域的语义理解。大多数现有的方法依赖于将视频分割成连续的镜头&从每个镜头中选择一个或多个关键帧,并将这些关键帧作为摘要呈现。如果视频中的独立概念只出现一次,这种方法可能会很有效。然而,在多次重复相同概念的视频中,这些现有方法可能会选择属于相同概念的重复摘要帧。在本文中,我们提出了一种新的框架聚类方法,通过将所有相似概念的框架分组在一起,而不管它们的出现顺序如何,从而生成非常简洁的摘要。所提出的方法是针对像旅游指南、纪录片、戏剧等领域的大型视频,这些视频围绕着很少重复的概念。该方法以一种通用的方式利用多个视频特征来确定帧相似度,并可扩展到多视频摘要。实验对比结果证实了该方法在对重复概念的视频生成简洁的视频摘要方面的有效性。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Hilbert Huang transform based speech recognition An Integrated curvature and Convex Hull based concave point detection in Histopathological images Study of Complex-valued Learning algorithms for Post-surgery survival prediction Optimized Radial Basis Function Neural Network model for wind power prediction Early detection of grapes diseases using machine learning and IoT
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1