Generative Adversarial Networks for Video Summarization Based on Key-frame Selection

IF 2 4区 计算机科学 Q3 AUTOMATION & CONTROL SYSTEMS Information Technology and Control Pub Date : 2023-03-28 DOI:10.5755/j01.itc.52.1.32278
Xiayun Hu, Xiaobin Hu, Jingxian Li, Kun You
{"title":"Generative Adversarial Networks for Video Summarization Based on Key-frame Selection","authors":"Xiayun Hu, Xiaobin Hu, Jingxian Li, Kun You","doi":"10.5755/j01.itc.52.1.32278","DOIUrl":null,"url":null,"abstract":"Video summarization based on generative adversarial networks (GANs) has been shown to easily produce more realistic results. However, most summary videos are composed of multiple key components. If the selection of some video frames changes during the training process, the information carried by these frames may not be reasonably reflected in the identification results. In this paper, we propose a video summarization method based on selecting keyframes over GANs. The novelty of the proposed method is the discriminator not only identifies the completeness of the video, but also takes into account the value judgment of the candidate keyframes, thus enabling the influence of keyframes on the result value. Given GANs are mainly designed to generate continuous real values, it is generally challenging to generate discrete symbol sequences during the summarization process directly. However, if the generated sample is based on discrete symbols, the slight guidance change of the discrimination network may be meaningless. To better use the advantages of GANs, the study also adopts the video summarization optimization method of GANs under a collaborative reinforcement learning strategy. Experimental results show the proposed method gets a significant summarization effect and character compared with the existing cutting-edge methods.","PeriodicalId":54982,"journal":{"name":"Information Technology and Control","volume":"5 1","pages":"185-198"},"PeriodicalIF":2.0000,"publicationDate":"2023-03-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Information Technology and Control","FirstCategoryId":"94","ListUrlMain":"https://doi.org/10.5755/j01.itc.52.1.32278","RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"AUTOMATION & CONTROL SYSTEMS","Score":null,"Total":0}
引用次数: 0

Abstract

Video summarization based on generative adversarial networks (GANs) has been shown to easily produce more realistic results. However, most summary videos are composed of multiple key components. If the selection of some video frames changes during the training process, the information carried by these frames may not be reasonably reflected in the identification results. In this paper, we propose a video summarization method based on selecting keyframes over GANs. The novelty of the proposed method is the discriminator not only identifies the completeness of the video, but also takes into account the value judgment of the candidate keyframes, thus enabling the influence of keyframes on the result value. Given GANs are mainly designed to generate continuous real values, it is generally challenging to generate discrete symbol sequences during the summarization process directly. However, if the generated sample is based on discrete symbols, the slight guidance change of the discrimination network may be meaningless. To better use the advantages of GANs, the study also adopts the video summarization optimization method of GANs under a collaborative reinforcement learning strategy. Experimental results show the proposed method gets a significant summarization effect and character compared with the existing cutting-edge methods.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
基于关键帧选择的视频摘要生成对抗网络
基于生成对抗网络(GANs)的视频摘要易于产生更真实的结果。然而,大多数摘要视频由多个关键组件组成。如果在训练过程中改变了一些视频帧的选择,这些帧所携带的信息可能无法在识别结果中得到合理的反映。在本文中,我们提出了一种基于gan选择关键帧的视频摘要方法。该方法的新颖之处在于,该鉴别器不仅识别视频的完整性,而且考虑了候选关键帧的值判断,从而实现了关键帧对结果值的影响。由于gan主要用于生成连续实值,因此在总结过程中直接生成离散符号序列通常具有挑战性。但是,如果生成的样本是基于离散符号的,那么识别网络的微小制导变化可能是没有意义的。为了更好地发挥gan的优势,本研究还采用了协同强化学习策略下的gan视频摘要优化方法。实验结果表明,与现有的前沿方法相比,该方法具有显著的总结效果和特点。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
Information Technology and Control
Information Technology and Control 工程技术-计算机:人工智能
CiteScore
2.70
自引率
9.10%
发文量
36
审稿时长
12 months
期刊介绍: Periodical journal covers a wide field of computer science and control systems related problems including: -Software and hardware engineering; -Management systems engineering; -Information systems and databases; -Embedded systems; -Physical systems modelling and application; -Computer networks and cloud computing; -Data visualization; -Human-computer interface; -Computer graphics, visual analytics, and multimedia systems.
期刊最新文献
Model construction of big data asset management system for digital power grid regulation Melanoma Diagnosis Using Enhanced Faster Region Convolutional Neural Networks Optimized by Artificial Gorilla Troops Algorithm A Scalable and Stacked Ensemble Approach to Improve Intrusion Detection in Clouds Traffic Sign Detection Algorithm Based on Improved Yolox Apply Physical System Model and Computer Algorithm to Identify Osmanthus Fragrans Seed Vigor Based on Hyperspectral Imaging and Convolutional Neural Network
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1