Narrative Dataset: Towards Goal-Driven Narrative Generation

Karen Stephen, Rishabh Sheoran, Satoshi Yamazaki
{"title":"Narrative Dataset: Towards Goal-Driven Narrative Generation","authors":"Karen Stephen, Rishabh Sheoran, Satoshi Yamazaki","doi":"10.1145/3552463.3557021","DOIUrl":null,"url":null,"abstract":"In this paper, we propose a new dataset called the Narrative dataset, which is a work in progress, towards generating video and text narratives of complex daily events from long videos, captured from multiple cameras. As most of the existing datasets are collected from publicly available videos such as YouTube videos, there are no datasets targeted towards the task of narrative summarization of complex videos which contains multiple narratives. Hence, we create story plots and conduct video shooting with hired actors to create complex video sets where 3 to 4 narratives happen in each video. In the story plot, a narrative composes of multiple events corresponding to video clips of key human activities. On top of the shot video sets and the story plot, the narrative dataset contains dense annotation of actors, objects, and their relationships for each frame as the facts of narratives. Therefore, narrative dataset richly contains holistic and hierarchical structure of facts, events, and narratives. Moreover, Narrative Graph, a collection of scene graphs of narrative events with their causal relationships, is introduced for bridging the gap between the collection of facts and generation of the summary sentences of a narrative. Beyond related subtasks such as scene graph generation, narrative dataset potentially provide challenges of subtasks for bridging human event clips to narratives.","PeriodicalId":293267,"journal":{"name":"Proceedings of the 1st Workshop on User-centric Narrative Summarization of Long Videos","volume":"25 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 1st Workshop on User-centric Narrative Summarization of Long Videos","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3552463.3557021","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1

Abstract

In this paper, we propose a new dataset called the Narrative dataset, which is a work in progress, towards generating video and text narratives of complex daily events from long videos, captured from multiple cameras. As most of the existing datasets are collected from publicly available videos such as YouTube videos, there are no datasets targeted towards the task of narrative summarization of complex videos which contains multiple narratives. Hence, we create story plots and conduct video shooting with hired actors to create complex video sets where 3 to 4 narratives happen in each video. In the story plot, a narrative composes of multiple events corresponding to video clips of key human activities. On top of the shot video sets and the story plot, the narrative dataset contains dense annotation of actors, objects, and their relationships for each frame as the facts of narratives. Therefore, narrative dataset richly contains holistic and hierarchical structure of facts, events, and narratives. Moreover, Narrative Graph, a collection of scene graphs of narrative events with their causal relationships, is introduced for bridging the gap between the collection of facts and generation of the summary sentences of a narrative. Beyond related subtasks such as scene graph generation, narrative dataset potentially provide challenges of subtasks for bridging human event clips to narratives.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
叙事数据集:朝着目标驱动的叙事生成
在本文中,我们提出了一个新的数据集,称为叙事数据集,这是一项正在进行的工作,旨在从多个摄像机捕获的长视频中生成复杂日常事件的视频和文本叙述。由于大多数现有数据集都是从公开可用的视频(如YouTube视频)中收集的,因此没有针对包含多个叙事的复杂视频的叙事摘要任务的数据集。因此,我们创建故事情节,并聘请演员进行视频拍摄,制作复杂的视频集,每个视频中发生3到4个故事。在故事情节中,一个叙事由多个事件组成,这些事件对应于人类关键活动的视频片段。在镜头视频集和故事情节之上,叙事数据集包含演员、对象及其每帧关系的密集注释,作为叙事的事实。因此,叙事数据集丰富地包含了事实、事件和叙事的整体和层次结构。此外,本文还引入了叙事图(Narrative Graph),它是叙事事件及其因果关系的场景图集合,用于弥合事实收集与叙事总结句生成之间的差距。除了场景图生成等相关子任务之外,叙事数据集还可能为将人类事件剪辑连接到叙事提供子任务的挑战。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Video Summarization in the Deep Learning Era: Current Landscape and Future Directions Panel Discussion: Emerging Topics on Video Summarization Contrastive Representation Learning for Expression Recognition from Masked Face Images Learning, Understanding and Interaction in Videos Soccer Game Summarization using Audio Commentary, Metadata, and Captions
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1