Defocus video matting

M. McGuire, W. Matusik, H. Pfister, J. Hughes, F. Durand
{"title":"Defocus video matting","authors":"M. McGuire, W. Matusik, H. Pfister, J. Hughes, F. Durand","doi":"10.1145/1186822.1073231","DOIUrl":null,"url":null,"abstract":"Video matting is the process of pulling a high-quality alpha matte and foreground from a video sequence. Current techniques require either a known background (e.g., a blue screen) or extensive user interaction (e.g., to specify known foreground and background elements). The matting problem is generally under-constrained, since not enough information has been collected at capture time. We propose a novel, fully autonomous method for pulling a matte using multiple synchronized video streams that share a point of view but differ in their plane of focus. The solution is obtained by directly minimizing the error in filter-based image formation equations, which are over-constrained by our rich data stream. Our system solves the fully dynamic video matting problem without user assistance: both the foreground and background may be high frequency and have dynamic content, the foreground may resemble the background, and the scene is lit by natural (as opposed to polarized or collimated) illumination.","PeriodicalId":211118,"journal":{"name":"ACM SIGGRAPH 2005 Papers","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2005-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"171","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"ACM SIGGRAPH 2005 Papers","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/1186822.1073231","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 171

Abstract

Video matting is the process of pulling a high-quality alpha matte and foreground from a video sequence. Current techniques require either a known background (e.g., a blue screen) or extensive user interaction (e.g., to specify known foreground and background elements). The matting problem is generally under-constrained, since not enough information has been collected at capture time. We propose a novel, fully autonomous method for pulling a matte using multiple synchronized video streams that share a point of view but differ in their plane of focus. The solution is obtained by directly minimizing the error in filter-based image formation equations, which are over-constrained by our rich data stream. Our system solves the fully dynamic video matting problem without user assistance: both the foreground and background may be high frequency and have dynamic content, the foreground may resemble the background, and the scene is lit by natural (as opposed to polarized or collimated) illumination.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
散焦视频抠图
视频抠图是从视频序列中提取高质量的阿尔法哑光和前景的过程。当前的技术要么需要一个已知的背景(例如,蓝屏),要么需要大量的用户交互(例如,指定已知的前景和背景元素)。由于在捕获时没有收集到足够的信息,因此抠图问题通常是受限的。我们提出了一种新颖的、完全自主的方法,使用多个同步视频流来拉哑光,这些视频流共享一个观点,但在焦点平面上不同。该方法通过直接最小化基于滤波的图像形成方程的误差来获得,而这些方程受到丰富数据流的过度约束。我们的系统在没有用户帮助的情况下解决了完全动态的视频抠图问题:前景和背景都可能是高频的,并且具有动态内容,前景可能与背景相似,场景由自然照明(而不是偏光或准直)照亮。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Session details: I3D (symposium on interactive 3D graphics) Session details: Mesh manipulation Session details: Texture synthesis Session details: Precomputed light transport Session details: Hardware rendering
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1