首页 > 最新文献

2011 IEEE Workshop on Person-Oriented Vision最新文献

英文 中文
A system for automatic and interactive detection of static objects 一种用于自动和交互检测静态物体的系统
Pub Date : 2011-02-14 DOI: 10.1109/POV.2011.5712365
Rubén Heras Evangelio, Michael Pátzold, T. Sikora
Designing static object detection systems that are able to incorporate user interaction conveys a great benefit in many surveillance applications, since some correctly detected static objects can be considered to have no interest by a human operator. Interactive systems allow the user to include these decisions into the system, making automated surveillance systems more attractive and comfortable to use. In this paper we present a system for the detection of static objects that, based on the detection of a dual background model, classifies pixels by means of a finite-state machine. The state machine provides the meaning for the interpretation of the results obtained from background subtraction and it can be optionally used to integrate user input. The system can thus be used both in an automatic and an interactive manner without requiring any expert knowledge from the user. We successfully validated the system with several public datasets.
设计能够结合用户交互的静态对象检测系统在许多监视应用中具有很大的好处,因为一些正确检测到的静态对象可以被认为是人类操作员不感兴趣的。交互式系统允许用户将这些决策纳入系统,使自动化监控系统更有吸引力,使用更舒适。在本文中,我们提出了一种基于双背景模型检测的静态目标检测系统,该系统利用有限状态机对像素进行分类。状态机为解释从背景减法中获得的结果提供了意义,并且可以选择性地使用它来集成用户输入。因此,该系统可以以自动和交互的方式使用,而不需要用户的任何专业知识。我们成功地用几个公共数据集验证了该系统。
{"title":"A system for automatic and interactive detection of static objects","authors":"Rubén Heras Evangelio, Michael Pátzold, T. Sikora","doi":"10.1109/POV.2011.5712365","DOIUrl":"https://doi.org/10.1109/POV.2011.5712365","url":null,"abstract":"Designing static object detection systems that are able to incorporate user interaction conveys a great benefit in many surveillance applications, since some correctly detected static objects can be considered to have no interest by a human operator. Interactive systems allow the user to include these decisions into the system, making automated surveillance systems more attractive and comfortable to use. In this paper we present a system for the detection of static objects that, based on the detection of a dual background model, classifies pixels by means of a finite-state machine. The state machine provides the meaning for the interpretation of the results obtained from background subtraction and it can be optionally used to integrate user input. The system can thus be used both in an automatic and an interactive manner without requiring any expert knowledge from the user. We successfully validated the system with several public datasets.","PeriodicalId":197184,"journal":{"name":"2011 IEEE Workshop on Person-Oriented Vision","volume":"82 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-02-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126040344","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
A benchmark for interactive image segmentation algorithms 交互式图像分割算法的基准
Pub Date : 2011-02-14 DOI: 10.1109/POV.2011.5712366
Yibiao Zhao, Xiaohan Nie, Y. Duan, Yaping Huang, Siwei Luo
This paper proposes a general benchmark for interactive segmentation algorithms. The main contribution can be summarized as follows: (I) A new dataset of fifty images is released. These images are categorized into five groups: animal, artifact, human, building and plant. They cover several major challenges for the interactive image segmentation task, including fuzzy boundary, complex texture, cluttered background, shading effect, sharp corner, and overlapping color. (II) We propose two types of schemes, point-process and boundary-process, to generate user scribbles automatically. The point-process simulates the human interaction process that users incrementally draw scribbles to some major components of the image. The boundary-process simulates the refining process that users place more scribbles near the segment boundaries to refine the details of result segments. (III) We then apply two precision measures to quantitatively evaluate the result segments of different algorithm. The region precision measures how many pixels are correctly classified, and the boundary precision measures how close is the segment boundary to the real boundary. This benchmark offered a tentative way to guarantee evaluation fairness of person-oriented tasks. Based on the benchmark, five state-of-the-art interactive segmentation algorithms are evaluated. All the images, synthesized user scribbles, running results are publicly available on the webpage1.
本文提出了交互式分割算法的通用基准。主要贡献如下:(1)发布了包含50幅图像的新数据集。这些图像被分为五组:动物、人工制品、人类、建筑和植物。它们涵盖了交互式图像分割任务的几个主要挑战,包括模糊边界、复杂纹理、杂乱背景、阴影效果、尖锐角和重叠颜色。(二)提出了两种自动生成用户涂鸦的方案:点流程和边界流程。点过程模拟了用户逐渐在图像的一些主要组件上涂鸦的人类交互过程。boundary-process模拟了用户在段边界附近放置更多涂鸦以精炼结果段细节的精炼过程。(III)应用两种精度度量对不同算法的结果片段进行定量评价。区域精度衡量有多少像素被正确分类,边界精度衡量分割边界与实际边界的接近程度。该基准为保证以人为本的任务评价公平性提供了一种尝试性的方法。在此基础上,对五种最先进的交互式分割算法进行了评价。所有的图像、合成的用户涂鸦、运行结果都在网页上公开提供。
{"title":"A benchmark for interactive image segmentation algorithms","authors":"Yibiao Zhao, Xiaohan Nie, Y. Duan, Yaping Huang, Siwei Luo","doi":"10.1109/POV.2011.5712366","DOIUrl":"https://doi.org/10.1109/POV.2011.5712366","url":null,"abstract":"This paper proposes a general benchmark for interactive segmentation algorithms. The main contribution can be summarized as follows: (I) A new dataset of fifty images is released. These images are categorized into five groups: animal, artifact, human, building and plant. They cover several major challenges for the interactive image segmentation task, including fuzzy boundary, complex texture, cluttered background, shading effect, sharp corner, and overlapping color. (II) We propose two types of schemes, point-process and boundary-process, to generate user scribbles automatically. The point-process simulates the human interaction process that users incrementally draw scribbles to some major components of the image. The boundary-process simulates the refining process that users place more scribbles near the segment boundaries to refine the details of result segments. (III) We then apply two precision measures to quantitatively evaluate the result segments of different algorithm. The region precision measures how many pixels are correctly classified, and the boundary precision measures how close is the segment boundary to the real boundary. This benchmark offered a tentative way to guarantee evaluation fairness of person-oriented tasks. Based on the benchmark, five state-of-the-art interactive segmentation algorithms are evaluated. All the images, synthesized user scribbles, running results are publicly available on the webpage1.","PeriodicalId":197184,"journal":{"name":"2011 IEEE Workshop on Person-Oriented Vision","volume":"44 3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-02-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131312219","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 22
Finding lost children 寻找走失儿童
Pub Date : 2011-02-14 DOI: 10.1109/POV.2011.5712362
Ashley M. Eden, C. M. Christoudias, Trevor Darrell
During a disaster, children may be quickly wrenched from their families. Research shows that children in such circumstances are often unable or unwilling to give their names or other identifying information. Currently in the US, there is no existing system in the public health infrastructure that effectively expedites reunification when children can't be identified. Working with the Children's Hospital Boston, we have engineered a system to speed reunification of children with their families, should they get separated in a disaster. Our system is based on a Content Based Image Retrieval and attribute search. In this paper we will describe the system and a series of evaluations, including a realistic disaster drill set up and run jointly with the Children's Hospital.
在一场灾难中,孩子们可能很快就被从他们的家庭中夺走。研究表明,在这种情况下,儿童往往不能或不愿说出自己的姓名或其他身份信息。目前在美国,公共卫生基础设施中没有现有的系统可以有效地加快无法识别儿童的团聚。我们与波士顿儿童医院合作,设计了一个系统,以便在灾难中走散的儿童与家人快速团聚。我们的系统是基于内容的图像检索和属性搜索。在本文中,我们将描述该系统和一系列评估,包括与儿童医院共同建立和运行的现实灾难演习。
{"title":"Finding lost children","authors":"Ashley M. Eden, C. M. Christoudias, Trevor Darrell","doi":"10.1109/POV.2011.5712362","DOIUrl":"https://doi.org/10.1109/POV.2011.5712362","url":null,"abstract":"During a disaster, children may be quickly wrenched from their families. Research shows that children in such circumstances are often unable or unwilling to give their names or other identifying information. Currently in the US, there is no existing system in the public health infrastructure that effectively expedites reunification when children can't be identified. Working with the Children's Hospital Boston, we have engineered a system to speed reunification of children with their families, should they get separated in a disaster. Our system is based on a Content Based Image Retrieval and attribute search. In this paper we will describe the system and a series of evaluations, including a realistic disaster drill set up and run jointly with the Children's Hospital.","PeriodicalId":197184,"journal":{"name":"2011 IEEE Workshop on Person-Oriented Vision","volume":"83 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-02-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130319709","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Augmented reality for immersive remote collaboration 增强现实沉浸式远程协作
Pub Date : 2011-02-14 DOI: 10.1109/POV.2011.5712368
Dan Gelb, A. Subramanian, K. Tan
Video conferencing systems are designed to deliver a collaboration experience that is as close as possible to actually meeting in person. Current systems, however, do a poor job of integrating video streams presenting the users with shared collaboration content. Real and virtual content are unnaturally separated, leading to problems with nonverbal communication and the overall conference experience. Methods of interacting with shared content are typically limited to pointing with a mouse, which is not a natural component of face-to-face human conversation. This paper presents a natural and intuitive method for sharing digital content within a meeting using augmented reality and computer vision. Real and virtual content is seamlessly integrated into the collaboration space. We develop new vision based methods for interacting with inserted digital content including target finding and gesture based control. These improvements let us deliver an immersive collaboration experience using natural gesture and object based interaction.
视频会议系统旨在提供尽可能接近实际面对面会议的协作体验。然而,当前的系统在集成视频流向用户展示共享协作内容方面做得很差。真实内容和虚拟内容不自然地分开,导致非语言交流和整体会议体验的问题。与共享内容交互的方法通常仅限于用鼠标指向,这不是面对面的人类对话的自然组成部分。本文提出了一种利用增强现实和计算机视觉在会议中共享数字内容的自然和直观的方法。真实和虚拟内容无缝集成到协作空间中。我们开发了新的基于视觉的方法来与插入的数字内容进行交互,包括目标查找和基于手势的控制。这些改进使我们能够使用自然手势和基于对象的交互提供身临其境的协作体验。
{"title":"Augmented reality for immersive remote collaboration","authors":"Dan Gelb, A. Subramanian, K. Tan","doi":"10.1109/POV.2011.5712368","DOIUrl":"https://doi.org/10.1109/POV.2011.5712368","url":null,"abstract":"Video conferencing systems are designed to deliver a collaboration experience that is as close as possible to actually meeting in person. Current systems, however, do a poor job of integrating video streams presenting the users with shared collaboration content. Real and virtual content are unnaturally separated, leading to problems with nonverbal communication and the overall conference experience. Methods of interacting with shared content are typically limited to pointing with a mouse, which is not a natural component of face-to-face human conversation. This paper presents a natural and intuitive method for sharing digital content within a meeting using augmented reality and computer vision. Real and virtual content is seamlessly integrated into the collaboration space. We develop new vision based methods for interacting with inserted digital content including target finding and gesture based control. These improvements let us deliver an immersive collaboration experience using natural gesture and object based interaction.","PeriodicalId":197184,"journal":{"name":"2011 IEEE Workshop on Person-Oriented Vision","volume":"39 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-02-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133851509","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 18
Active inference for retrieval in camera networks 基于摄像机网络的主动推理检索
Pub Date : 2011-02-14 DOI: 10.1109/POV.2011.5712363
Daozheng Chen, M. Bilgic, L. Getoor, D. Jacobs, Lilyana Mihalkova, Tom Yeh
We address the problem of searching camera network videos to retrieve frames containing specified individuals. We show the benefit of utilizing a learned probabilistic model that captures dependencies among the cameras. In addition, we develop an active inference framework that can request human input at inference time, directing human attention to the portions of the videos whose correct annotation would provide the biggest performance improvements. Our primary contribution is to show that by mapping video frames in a camera network onto a graphical model, we can apply collective classification and active inference algorithms to significantly increase the performance of the retrieval system, while minimizing the number of human annotations required.
我们解决了搜索摄像机网络视频以检索包含特定个体的帧的问题。我们展示了利用学习概率模型捕获相机之间的依赖关系的好处。此外,我们开发了一个主动推理框架,可以在推理时请求人工输入,将人类的注意力引导到视频的正确注释将提供最大性能改进的部分。我们的主要贡献是通过将摄像机网络中的视频帧映射到图形模型上,我们可以应用集体分类和主动推理算法来显着提高检索系统的性能,同时最小化所需的人工注释的数量。
{"title":"Active inference for retrieval in camera networks","authors":"Daozheng Chen, M. Bilgic, L. Getoor, D. Jacobs, Lilyana Mihalkova, Tom Yeh","doi":"10.1109/POV.2011.5712363","DOIUrl":"https://doi.org/10.1109/POV.2011.5712363","url":null,"abstract":"We address the problem of searching camera network videos to retrieve frames containing specified individuals. We show the benefit of utilizing a learned probabilistic model that captures dependencies among the cameras. In addition, we develop an active inference framework that can request human input at inference time, directing human attention to the portions of the videos whose correct annotation would provide the biggest performance improvements. Our primary contribution is to show that by mapping video frames in a camera network onto a graphical model, we can apply collective classification and active inference algorithms to significantly increase the performance of the retrieval system, while minimizing the number of human annotations required.","PeriodicalId":197184,"journal":{"name":"2011 IEEE Workshop on Person-Oriented Vision","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-02-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127915955","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
User oriented language model for face detection 面向用户的人脸检测语言模型
Pub Date : 2011-02-14 DOI: 10.1109/POV.2011.5712364
Daesik Jang, G. Miller, S. Fels, S. Oldridge
This paper provides a novel approach for a user oriented language model for face detection. Even though there are many open source or commercial libraries to solve the problem of face detection, they are still hard to use because they require specific knowledge on details of algorithmic techniques. This paper proposes a high-level language model for face detection with which users can develop systems easily and even without specific knowledge on face detection theories and algorithms. Important conditions are firstly considered to categorize the large problem space of face detection. The conditions identified here are then represented as expressions in terms of a language model so that developers can use them to express various problems. Once the conditions are expressed by users, the proposed associated interpreter interprets the conditions to find and organize the best algorithms to solve the represented problem with corresponding conditions. We show a proof-of-concept implementation and some test and analyze example problems to show the ease of use and usability.
本文提出了一种面向用户的人脸检测语言模型。尽管有许多开源或商业库来解决人脸检测问题,但它们仍然很难使用,因为它们需要对算法技术的细节有专门的了解。本文提出了一种用于人脸检测的高级语言模型,用户可以使用该模型轻松开发系统,甚至不需要特定的人脸检测理论和算法知识。首先考虑重要条件,对人脸检测的大问题空间进行分类。这里确定的条件然后用语言模型表示为表达式,以便开发人员可以使用它们来表示各种问题。一旦用户表达了条件,所提出的关联解释器对条件进行解释,找到并组织最佳算法,用相应的条件来解决所表示的问题。我们展示了一个概念验证实现和一些测试和分析示例问题,以显示易用性和可用性。
{"title":"User oriented language model for face detection","authors":"Daesik Jang, G. Miller, S. Fels, S. Oldridge","doi":"10.1109/POV.2011.5712364","DOIUrl":"https://doi.org/10.1109/POV.2011.5712364","url":null,"abstract":"This paper provides a novel approach for a user oriented language model for face detection. Even though there are many open source or commercial libraries to solve the problem of face detection, they are still hard to use because they require specific knowledge on details of algorithmic techniques. This paper proposes a high-level language model for face detection with which users can develop systems easily and even without specific knowledge on face detection theories and algorithms. Important conditions are firstly considered to categorize the large problem space of face detection. The conditions identified here are then represented as expressions in terms of a language model so that developers can use them to express various problems. Once the conditions are expressed by users, the proposed associated interpreter interprets the conditions to find and organize the best algorithms to solve the represented problem with corresponding conditions. We show a proof-of-concept implementation and some test and analyze example problems to show the ease of use and usability.","PeriodicalId":197184,"journal":{"name":"2011 IEEE Workshop on Person-Oriented Vision","volume":"90 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-02-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123348145","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 19
期刊
2011 IEEE Workshop on Person-Oriented Vision
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1