Foreground segmentation using adaptive mixture models in color and depth

M. Harville, G. Gordon, J. Woodfill
{"title":"Foreground segmentation using adaptive mixture models in color and depth","authors":"M. Harville, G. Gordon, J. Woodfill","doi":"10.1109/EVENT.2001.938860","DOIUrl":null,"url":null,"abstract":"Segmentation of novel or dynamic objects in a scene, often referred to as \"background subtraction\" or foreground segmentation\", is a critical early in step in most computer vision applications in domains such as surveillance and human-computer interaction. All previously described, real-time methods fail to handle properly one or more common phenomena, such as global illumination changes, shadows, inter-reflections, similarity of foreground color to background and non-static backgrounds (e.g. active video displays or trees waving in the wind). The advent of hardware and software for real-time computation of depth imagery makes better approaches possible. We propose a method for modeling the background that uses per-pixel, time-adaptive, Gaussian mixtures in the combined input space of depth and luminance-invariant color. This combination in itself is novel, but we further improve it by introducing the ideas of (1) modulating the background model learning rate based on scene activity, and (2) making color-based segmentation criteria dependent on depth observations. Our experiments show that the method possesses much greater robustness to problematic phenomena than the prior state-of-the-art, without sacrificing real-time performance, making it well-suited for a wide range of practical applications in video event detection and recognition.","PeriodicalId":375539,"journal":{"name":"Proceedings IEEE Workshop on Detection and Recognition of Events in Video","volume":"12 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2001-07-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"264","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings IEEE Workshop on Detection and Recognition of Events in Video","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/EVENT.2001.938860","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 264

Abstract

Segmentation of novel or dynamic objects in a scene, often referred to as "background subtraction" or foreground segmentation", is a critical early in step in most computer vision applications in domains such as surveillance and human-computer interaction. All previously described, real-time methods fail to handle properly one or more common phenomena, such as global illumination changes, shadows, inter-reflections, similarity of foreground color to background and non-static backgrounds (e.g. active video displays or trees waving in the wind). The advent of hardware and software for real-time computation of depth imagery makes better approaches possible. We propose a method for modeling the background that uses per-pixel, time-adaptive, Gaussian mixtures in the combined input space of depth and luminance-invariant color. This combination in itself is novel, but we further improve it by introducing the ideas of (1) modulating the background model learning rate based on scene activity, and (2) making color-based segmentation criteria dependent on depth observations. Our experiments show that the method possesses much greater robustness to problematic phenomena than the prior state-of-the-art, without sacrificing real-time performance, making it well-suited for a wide range of practical applications in video event detection and recognition.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
前景分割使用自适应混合模型在颜色和深度
场景中新物体或动态物体的分割,通常被称为“背景减去”或“前景分割”,是大多数计算机视觉应用领域(如监视和人机交互)的关键早期步骤。所有先前描述的实时方法都不能正确处理一个或多个常见现象,例如全局照明变化、阴影、相互反射、前景颜色与背景和非静态背景的相似性(例如活动视频显示或风中摇曳的树木)。用于深度图像实时计算的硬件和软件的出现使更好的方法成为可能。我们提出了一种在深度和亮度不变颜色的组合输入空间中使用逐像素、时间自适应、高斯混合的背景建模方法。这种组合本身是新颖的,但我们通过引入(1)基于场景活动调节背景模型学习率和(2)基于深度观察的基于颜色的分割标准的思想进一步改进了它。我们的实验表明,该方法在不牺牲实时性能的情况下,对问题现象具有比现有技术更强的鲁棒性,使其非常适合视频事件检测和识别中的广泛实际应用。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Multimodal 3-D tracking and event detection via the particle filter Segmentation and recognition of continuous human activity Hierarchical unsupervised learning of facial expression categories View-invariant representation and learning of human action Detecting independently moving objects and their interactions in georeferenced airborne video
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1