首页 > 最新文献

2022 IEEE International Conference on Multisensor Fusion and Integration for Intelligent Systems (MFI)最新文献

英文 中文
Vision-based Fall Detection in Aircraft Maintenance Environment with Pose Estimation 基于姿态估计的飞机维修环境视觉坠落检测
Adeyemi Osigbesan, Solene Barrat, Harkeerat Singh, Dongzi Xia, Siddharth Singh, Yang Xing, Weisi Guo, A. Tsourdos
Fall-related injuries at the workplace account for a fair percentage of the global accident at work claims according to Health and Safety Executive (HSE). With a significant percentage of these being fatal, industrial and maintenance workshops have great potential for injuries that can be associated with slips, trips, and other types of falls, owing to their characteristic fast-paced workspaces. Typically, the short turnaround time expected for aircraft undergoing maintenance increases the risk of workers falling, and thus makes a good case for the study of more contemporary methods for the detection of work-related falls in the aircraft maintenance environment. Advanced development in human pose estimation using computer vision technology has made it possible to automate real-time detection and classification of human actions by analyzing body part motion and position relative to time. This paper attempts to combine the analysis of body silhouette bounding box with body joint position estimation to detect and categorize in real-time, human motion captured in continuous video feeds into a fall or a non-fall event. We proposed a standard wide-angle camera, installed at a diagonal ceiling position in an aircraft hangar for our visual data input, and a three-dimensional convolutional neural network with Long Short-Term Memory (LSTM) layers using a technique we referred to as Region Key point (Reg-Key) repartitioning for visual pose estimation and fall detection.
根据健康与安全执行局(HSE)的数据,工作场所与跌倒有关的伤害占全球工作事故索赔的相当比例。其中很大一部分是致命的,工业和维修车间由于其快节奏的工作空间特点,有很大的潜在伤害,可能与滑倒、绊倒和其他类型的跌倒有关。通常情况下,飞机维修所需的短周转时间增加了工人摔倒的风险,因此研究更现代的方法来检测飞机维修环境中与工作有关的摔倒是一个很好的例子。利用计算机视觉技术进行人体姿态估计的先进发展,使得通过分析身体部位的运动和相对于时间的位置来自动实时检测和分类人体动作成为可能。本文试图将身体轮廓边界盒分析与身体关节位置估计相结合,实时检测和分类连续视频馈送中捕获的人体运动,将其分为跌倒事件和非跌倒事件。我们提出了一个标准的广角摄像机,安装在机库的对角线天花板位置,用于视觉数据输入,以及一个具有长短期记忆(LSTM)层的三维卷积神经网络,使用我们称为区域关键点(Reg-Key)重新划分的技术,用于视觉姿态估计和跌倒检测。
{"title":"Vision-based Fall Detection in Aircraft Maintenance Environment with Pose Estimation","authors":"Adeyemi Osigbesan, Solene Barrat, Harkeerat Singh, Dongzi Xia, Siddharth Singh, Yang Xing, Weisi Guo, A. Tsourdos","doi":"10.1109/MFI55806.2022.9913877","DOIUrl":"https://doi.org/10.1109/MFI55806.2022.9913877","url":null,"abstract":"Fall-related injuries at the workplace account for a fair percentage of the global accident at work claims according to Health and Safety Executive (HSE). With a significant percentage of these being fatal, industrial and maintenance workshops have great potential for injuries that can be associated with slips, trips, and other types of falls, owing to their characteristic fast-paced workspaces. Typically, the short turnaround time expected for aircraft undergoing maintenance increases the risk of workers falling, and thus makes a good case for the study of more contemporary methods for the detection of work-related falls in the aircraft maintenance environment. Advanced development in human pose estimation using computer vision technology has made it possible to automate real-time detection and classification of human actions by analyzing body part motion and position relative to time. This paper attempts to combine the analysis of body silhouette bounding box with body joint position estimation to detect and categorize in real-time, human motion captured in continuous video feeds into a fall or a non-fall event. We proposed a standard wide-angle camera, installed at a diagonal ceiling position in an aircraft hangar for our visual data input, and a three-dimensional convolutional neural network with Long Short-Term Memory (LSTM) layers using a technique we referred to as Region Key point (Reg-Key) repartitioning for visual pose estimation and fall detection.","PeriodicalId":344737,"journal":{"name":"2022 IEEE International Conference on Multisensor Fusion and Integration for Intelligent Systems (MFI)","volume":"178 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-09-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124423667","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
PIPO: Policy Optimization with Permutation-Invariant Constraint for Distributed Multi-Robot Navigation 基于排列不变约束的分布式多机器人导航策略优化
Ruiqi Zhang, Guang Chen, Jing Hou, Zhijun Li, Alois Knoll
For large-scale multi-agent systems (MAS), ensuring the safety and effectiveness of navigation in complicated scenarios is a challenging task. With the agent scale increasing, most existing centralized methods lose their magic for the lack of scalability, and the popular decentralized approaches are hampered by high latency and computing requirements. In this research, we offer PIPO, a novel policy optimization algorithm for decentralized MAS navigation with permutation-invariant constraints. To conduct navigation and avoid un-necessary exploration in the early episodes, we first defined a guide-policy. Then, we introduce the permutation invariant property in decentralized multi-agent systems and leverage the graph convolution network to produce the same output under shuffled observations. Our approach can be easily scaled to an arbitrary number of agents and used in large-scale systems for its decentralized training and execution. We also provide extensive experiments to demonstrate that our PIPO significantly outperforms the baselines of multi-agent reinforcement learning algorithms and other leading methods in variant scenarios.
对于大规模多智能体系统(MAS)来说,确保复杂场景下导航的安全性和有效性是一项具有挑战性的任务。随着智能体规模的增加,大多数现有的集中式方法由于缺乏可扩展性而失去了魔力,而流行的分散方法则受到高延迟和计算需求的阻碍。在本研究中,我们提出了一种新的具有排列不变约束的分散MAS导航策略优化算法PIPO。为了在早期章节中进行导航并避免不必要的探索,我们首先定义了一个指南策略。然后,我们引入了分散多智能体系统的排列不变性,并利用图卷积网络在洗牌观测下产生相同的输出。我们的方法可以很容易地扩展到任意数量的代理,并用于大规模系统的分散训练和执行。我们还提供了大量的实验来证明我们的PIPO在不同场景下显著优于多智能体强化学习算法和其他领先方法的基线。
{"title":"PIPO: Policy Optimization with Permutation-Invariant Constraint for Distributed Multi-Robot Navigation","authors":"Ruiqi Zhang, Guang Chen, Jing Hou, Zhijun Li, Alois Knoll","doi":"10.1109/MFI55806.2022.9913862","DOIUrl":"https://doi.org/10.1109/MFI55806.2022.9913862","url":null,"abstract":"For large-scale multi-agent systems (MAS), ensuring the safety and effectiveness of navigation in complicated scenarios is a challenging task. With the agent scale increasing, most existing centralized methods lose their magic for the lack of scalability, and the popular decentralized approaches are hampered by high latency and computing requirements. In this research, we offer PIPO, a novel policy optimization algorithm for decentralized MAS navigation with permutation-invariant constraints. To conduct navigation and avoid un-necessary exploration in the early episodes, we first defined a guide-policy. Then, we introduce the permutation invariant property in decentralized multi-agent systems and leverage the graph convolution network to produce the same output under shuffled observations. Our approach can be easily scaled to an arbitrary number of agents and used in large-scale systems for its decentralized training and execution. We also provide extensive experiments to demonstrate that our PIPO significantly outperforms the baselines of multi-agent reinforcement learning algorithms and other leading methods in variant scenarios.","PeriodicalId":344737,"journal":{"name":"2022 IEEE International Conference on Multisensor Fusion and Integration for Intelligent Systems (MFI)","volume":"41 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-09-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115808769","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Distributed Extended Object Tracking Filter Through Embedded ADMM Technique 基于嵌入式ADMM技术的分布式扩展目标跟踪滤波器
Zhifei Li, Hongyan Wang, Shi Yan, Hongxia Zou, Mingyang Du
This work is concerned with the distributed extended object tracking system over a realistic network, where all nodes are required to achieve consensus on both the extent and kinematics. To this end, we first exploit an aligned velocity model to establish a tight relation between the orientation and velocity vector. Then, we use the moment-matching method to give two separate models to match the information filter (IF) framework. Later, we resort to the two models to propose a centralized IF and extend it to the distributed scenario based on the embedded alternating direction method of multipliers (ADMM) technique. To keep an agreement between nodes, an optimization function is given, followed by a consensus-based constraint. Numerical simulation together with theoretical analysis verifies the convergence and consensus of the proposed filter.
本文研究了现实网络上的分布式扩展目标跟踪系统,要求所有节点在范围和运动学上达成一致。为此,我们首先利用对准速度模型来建立方向和速度矢量之间的紧密关系。然后,我们使用矩匹配方法给出两个独立的模型来匹配信息过滤(IF)框架。随后,我们利用这两个模型提出了一个集中式中频,并基于嵌入式乘法器交替方向方法(ADMM)技术将其扩展到分布式场景。为了保持节点之间的一致性,给出了一个优化函数,然后给出了一个基于共识的约束。数值仿真和理论分析验证了该滤波器的收敛性和一致性。
{"title":"Distributed Extended Object Tracking Filter Through Embedded ADMM Technique","authors":"Zhifei Li, Hongyan Wang, Shi Yan, Hongxia Zou, Mingyang Du","doi":"10.1109/MFI55806.2022.9913876","DOIUrl":"https://doi.org/10.1109/MFI55806.2022.9913876","url":null,"abstract":"This work is concerned with the distributed extended object tracking system over a realistic network, where all nodes are required to achieve consensus on both the extent and kinematics. To this end, we first exploit an aligned velocity model to establish a tight relation between the orientation and velocity vector. Then, we use the moment-matching method to give two separate models to match the information filter (IF) framework. Later, we resort to the two models to propose a centralized IF and extend it to the distributed scenario based on the embedded alternating direction method of multipliers (ADMM) technique. To keep an agreement between nodes, an optimization function is given, followed by a consensus-based constraint. Numerical simulation together with theoretical analysis verifies the convergence and consensus of the proposed filter.","PeriodicalId":344737,"journal":{"name":"2022 IEEE International Conference on Multisensor Fusion and Integration for Intelligent Systems (MFI)","volume":"26 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-09-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122676957","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Encrypted Fast Covariance Intersection Without Leaking Fusion Weights 不泄漏融合权值的加密快速协方差相交
Marko Ristic, B. Noack
State estimate fusion is a common requirement in distributed sensor networks and can be complicated by untrusted participants or network eavesdroppers. We present a method for computing the common Fast Covariance Intersection fusion algorithm on an untrusted cloud without disclosing individual estimates or the fused result. In an existing solution to this problem, fusion weights corresponding to estimate errors are leaked to the cloud to perform the fusion. In this work, we present a method that guarantees no data identifying estimators or their estimated values is leaked to the cloud by requiring an additional computation step by the party querying the cloud for the fused result. The Paillier encryption scheme is used to homomorphically compute separate parts of the computation that can be combined after decryption. This encrypted Fast Covariance Intersection algorithm can be used in scenarios where the fusing cloud is not trusted and any information on estimator performances must remain confidential.
状态估计融合是分布式传感器网络中的一种常见需求,但不可信参与者或网络窃听者可能会使其复杂化。提出了一种在不可信云上不泄露个体估计和融合结果的快速协方差交叉融合算法的计算方法。在该问题的现有解决方案中,将与估计误差相对应的融合权重泄露给云来执行融合。在这项工作中,我们提出了一种方法,通过要求向云查询融合结果的一方进行额外的计算步骤,保证没有识别估计器或其估计值的数据泄露到云中。Paillier加密方案用于同态计算可在解密后组合的计算的独立部分。这种加密的快速协方差相交算法可用于融合云不可信且任何关于估计器性能的信息必须保密的场景。
{"title":"Encrypted Fast Covariance Intersection Without Leaking Fusion Weights","authors":"Marko Ristic, B. Noack","doi":"10.1109/MFI55806.2022.9913840","DOIUrl":"https://doi.org/10.1109/MFI55806.2022.9913840","url":null,"abstract":"State estimate fusion is a common requirement in distributed sensor networks and can be complicated by untrusted participants or network eavesdroppers. We present a method for computing the common Fast Covariance Intersection fusion algorithm on an untrusted cloud without disclosing individual estimates or the fused result. In an existing solution to this problem, fusion weights corresponding to estimate errors are leaked to the cloud to perform the fusion. In this work, we present a method that guarantees no data identifying estimators or their estimated values is leaked to the cloud by requiring an additional computation step by the party querying the cloud for the fused result. The Paillier encryption scheme is used to homomorphically compute separate parts of the computation that can be combined after decryption. This encrypted Fast Covariance Intersection algorithm can be used in scenarios where the fusing cloud is not trusted and any information on estimator performances must remain confidential.","PeriodicalId":344737,"journal":{"name":"2022 IEEE International Conference on Multisensor Fusion and Integration for Intelligent Systems (MFI)","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-09-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128416574","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Event-based Driver Distraction Detection and Action Recognition 基于事件的驾驶员分心检测与动作识别
Chu Yang, Peigen Liu, Guang Chen, Zhengfa Liu, Ya Wu, Alois Knoll
Driver distraction is one of the important factors leading to traffic accidents. With the development of mobile infotainment and the overestimation of immature autonomous driving technology, this phenomenon has become more and more serious. However, most existing distraction detection algorithms can not achieve satisfactory performance due to the complex in-cabin light condition and limited computing resource of edge devices. To this end, we introduce a light weight and flexible event-based system to monitor driver state. Compared with frame-based camera, the event camera responds to pixel wise light intensity changes asynchronously and has several promising advantages, including high dynamic range, high temporal resolution, low latency and low data redundant, which makes it suitable for the mobile terminal applications. The system first denoises the events stream and encode it into a sequence of 3D tensors. Then, the voxel features at different time steps are extracted using efficient net and fed into LSTM to establish temporal model, based on which, the driver distraction is detected. In addition, we extend the proposed architecture to recognise driver action and adopt transfer learning strategy to improve the detection performance. Extensive experiments are conducted on both simulated dataset (transform from Drive&Act) and real event dataset (collected by ourselves). The experimental results shows the advantages of the system on accuracy and efficient for driver state monitoring.
驾驶员注意力分散是导致交通事故的重要因素之一。随着移动信息娱乐的发展和对不成熟的自动驾驶技术的高估,这一现象越来越严重。然而,由于舱内光线条件复杂,边缘设备的计算资源有限,现有的大多数分心检测算法都不能达到令人满意的效果。为此,我们引入了一个轻量级且灵活的基于事件的系统来监控驱动程序状态。与基于帧的相机相比,事件相机对像素级光强变化的响应是异步的,具有高动态范围、高时间分辨率、低延迟和低数据冗余等优点,适合移动端应用。该系统首先对事件流去噪,并将其编码为三维张量序列。然后,利用高效网络提取不同时间步长的体素特征,并将其输入LSTM中建立时间模型,在此基础上检测驾驶员分心;此外,我们扩展了所提出的架构来识别驾驶员动作,并采用迁移学习策略来提高检测性能。在模拟数据集(从Drive&Act转换)和真实事件数据集(自己收集)上进行了大量的实验。实验结果表明,该系统具有准确、高效的驾驶状态监测优势。
{"title":"Event-based Driver Distraction Detection and Action Recognition","authors":"Chu Yang, Peigen Liu, Guang Chen, Zhengfa Liu, Ya Wu, Alois Knoll","doi":"10.1109/MFI55806.2022.9913871","DOIUrl":"https://doi.org/10.1109/MFI55806.2022.9913871","url":null,"abstract":"Driver distraction is one of the important factors leading to traffic accidents. With the development of mobile infotainment and the overestimation of immature autonomous driving technology, this phenomenon has become more and more serious. However, most existing distraction detection algorithms can not achieve satisfactory performance due to the complex in-cabin light condition and limited computing resource of edge devices. To this end, we introduce a light weight and flexible event-based system to monitor driver state. Compared with frame-based camera, the event camera responds to pixel wise light intensity changes asynchronously and has several promising advantages, including high dynamic range, high temporal resolution, low latency and low data redundant, which makes it suitable for the mobile terminal applications. The system first denoises the events stream and encode it into a sequence of 3D tensors. Then, the voxel features at different time steps are extracted using efficient net and fed into LSTM to establish temporal model, based on which, the driver distraction is detected. In addition, we extend the proposed architecture to recognise driver action and adopt transfer learning strategy to improve the detection performance. Extensive experiments are conducted on both simulated dataset (transform from Drive&Act) and real event dataset (collected by ourselves). The experimental results shows the advantages of the system on accuracy and efficient for driver state monitoring.","PeriodicalId":344737,"journal":{"name":"2022 IEEE International Conference on Multisensor Fusion and Integration for Intelligent Systems (MFI)","volume":"19 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-09-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128156644","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Enhancing Event-based Structured Light Imaging with a Single Frame 单帧增强基于事件的结构光成像
Huijiao Wang, Tangbo Liu, Chu He, Cheng Li, Jian-zhuo Liu, Lei Yu
Benefiting from the extremely low latency, events have been used for Structured Light Imaging (SLI) to predict the depth surface. However, existing methods only focus on improving scanning speeds but neglect perturbations from event noise and timestamp jittering for depth estimation. In this paper, we build a hybrid SLI system equipped with an event camera, a high-resolution frame camera, and a digital light projector, where a single intensity frame is adopted as a guidance to enhance the event-based SLI quality. To achieve this end, we propose a Multi-Modal Feature Fusion Network (MFFN) consisting of a feature fusion module and an upscale module to simultaneously fuse events and a single intensity frame, suppress event perturbations, and reconstruct a high-quality depth surface. Further, for training MFFN, we build a new Structured Light Imaging based on Event and Frame cameras (EF-SLI) dataset collected from the hybrid SLI system, containing paired inputs composed of a set of synchronized events and one single corresponding frame, and ground-truth references obtained by a high-quality SLI approach. Experiments demonstrate that our proposed MFFN outperforms state-of-the-art event-based SLI approaches in terms of accuracy at different scanning speeds.
得益于极低的延迟,事件已被用于结构光成像(SLI)来预测深度表面。然而,现有的方法只注重提高扫描速度,而忽略了事件噪声和时间戳抖动对深度估计的影响。在本文中,我们构建了一个由事件相机、高分辨率帧相机和数字光投影仪组成的混合SLI系统,其中采用单强度帧作为指导,以提高基于事件的SLI质量。为了实现这一目标,我们提出了一个由特征融合模块和高级模块组成的多模态特征融合网络(MFFN),以同时融合事件和单个强度帧,抑制事件扰动,重建高质量的深度表面。此外,为了训练MFFN,我们基于从混合SLI系统中收集的事件和帧相机(EF-SLI)数据集构建了新的结构光成像,该数据集包含由一组同步事件和单个对应帧组成的配对输入,以及由高质量SLI方法获得的真地参考。实验表明,我们提出的MFFN在不同扫描速度下的精度优于最先进的基于事件的SLI方法。
{"title":"Enhancing Event-based Structured Light Imaging with a Single Frame","authors":"Huijiao Wang, Tangbo Liu, Chu He, Cheng Li, Jian-zhuo Liu, Lei Yu","doi":"10.1109/MFI55806.2022.9913845","DOIUrl":"https://doi.org/10.1109/MFI55806.2022.9913845","url":null,"abstract":"Benefiting from the extremely low latency, events have been used for Structured Light Imaging (SLI) to predict the depth surface. However, existing methods only focus on improving scanning speeds but neglect perturbations from event noise and timestamp jittering for depth estimation. In this paper, we build a hybrid SLI system equipped with an event camera, a high-resolution frame camera, and a digital light projector, where a single intensity frame is adopted as a guidance to enhance the event-based SLI quality. To achieve this end, we propose a Multi-Modal Feature Fusion Network (MFFN) consisting of a feature fusion module and an upscale module to simultaneously fuse events and a single intensity frame, suppress event perturbations, and reconstruct a high-quality depth surface. Further, for training MFFN, we build a new Structured Light Imaging based on Event and Frame cameras (EF-SLI) dataset collected from the hybrid SLI system, containing paired inputs composed of a set of synchronized events and one single corresponding frame, and ground-truth references obtained by a high-quality SLI approach. Experiments demonstrate that our proposed MFFN outperforms state-of-the-art event-based SLI approaches in terms of accuracy at different scanning speeds.","PeriodicalId":344737,"journal":{"name":"2022 IEEE International Conference on Multisensor Fusion and Integration for Intelligent Systems (MFI)","volume":"57 3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-09-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116834161","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
UPC-Faster-RCNN: A Dynamic Self-Labeling Algorithm for Open-Set Object Detection Based on Unknown Proposal Clustering UPC-Faster-RCNN:一种基于未知建议聚类的开集目标检测动态自标记算法
Yujun Liao, Y. Wu, Y. Mo, Feilin Liu, Yufei He, Junqiao Zhao
To promote the development of object detection in a more realistic world, efforts have been made to a new task named open-set object detection. This task aims to increase the model’s ability to recognize unknown classes. In this work, we propose a novel dynamic self-labeling algorithm, named UPC-Faster-RCNN. The wisdom of DBSCAN is applied to build our unknown proposal clustering algorithm, which aims to filter and cluster the unknown objects proposals. An effective dynamic self-labeling algorithm is proposed to generate high-quality pseudo labels from clustered proposals. We evaluate UPC-Faster-RCNN on a composite dataset of PASCAL VOC and COCO. The extensive experiments show that UPC-Faster-RCNN effectively increases the ability upon Faster-RCNN baseline to detect unknown target, while keeping the ability to detect known targets. Specifically, UPC-Faster-RCNN decreases the WI by 23.8%, decreases the A-OSE by 6542, and slightly increase the mAP in known classes by 0.3%.
为了促进目标检测在更现实世界中的发展,人们提出了一种新的任务——开集目标检测。该任务旨在提高模型识别未知类的能力。在这项工作中,我们提出了一种新的动态自标记算法,命名为UPC-Faster-RCNN。利用DBSCAN的智慧构建未知建议聚类算法,对未知对象建议进行过滤聚类。提出了一种有效的动态自标记算法,从聚类建议中生成高质量的伪标签。我们在PASCAL VOC和COCO的复合数据集上评估UPC-Faster-RCNN。大量实验表明,UPC-Faster-RCNN在保持已知目标检测能力的同时,有效地提高了在Faster-RCNN基线上检测未知目标的能力。具体而言,UPC-Faster-RCNN使已知类别的WI降低了23.8%,A-OSE降低了6542,mAP略微提高了0.3%。
{"title":"UPC-Faster-RCNN: A Dynamic Self-Labeling Algorithm for Open-Set Object Detection Based on Unknown Proposal Clustering","authors":"Yujun Liao, Y. Wu, Y. Mo, Feilin Liu, Yufei He, Junqiao Zhao","doi":"10.1109/MFI55806.2022.9913863","DOIUrl":"https://doi.org/10.1109/MFI55806.2022.9913863","url":null,"abstract":"To promote the development of object detection in a more realistic world, efforts have been made to a new task named open-set object detection. This task aims to increase the model’s ability to recognize unknown classes. In this work, we propose a novel dynamic self-labeling algorithm, named UPC-Faster-RCNN. The wisdom of DBSCAN is applied to build our unknown proposal clustering algorithm, which aims to filter and cluster the unknown objects proposals. An effective dynamic self-labeling algorithm is proposed to generate high-quality pseudo labels from clustered proposals. We evaluate UPC-Faster-RCNN on a composite dataset of PASCAL VOC and COCO. The extensive experiments show that UPC-Faster-RCNN effectively increases the ability upon Faster-RCNN baseline to detect unknown target, while keeping the ability to detect known targets. Specifically, UPC-Faster-RCNN decreases the WI by 23.8%, decreases the A-OSE by 6542, and slightly increase the mAP in known classes by 0.3%.","PeriodicalId":344737,"journal":{"name":"2022 IEEE International Conference on Multisensor Fusion and Integration for Intelligent Systems (MFI)","volume":"77 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-09-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126254691","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Probabilistic Information Matrix Fusion in a Multi-Object Environment 多目标环境下的概率信息矩阵融合
Kaipei Yang, Y. Bar-Shalom
In distributed sensor fusion systems, each of the local sensors has its own tracker processing local measurements for measurement-to-track association and state estimation. Only the processed data, local tracks (LT) comprising state vector estimates and their covariance matrices are transmitted to the fusion center (FC). In this work, a multi-object hybrid probabilistic information matrix fusion (MO-HPIMF) is derived taking into account all association hypotheses. In MO-HPIMF, the association carried out is between the FC track states (prediction) and the LT state estimates from local sensors. When having a large number of objects and sensors in fusion, only the m-best FC-track-to-LT association hypotheses should be incorporated in MO-HPIMF to reduce the computational complexity. A Sequential m-best 2-D method is used for solving the multidimensional assignment problem in this work. It is shown in the simulations that MO-HPIMF can successfully track all targets of interest and is superior to track-to-track fusion (T2TF, a commonly used approach in distributed sensor fusion system) which relies on hard association decisions.
在分布式传感器融合系统中,每个本地传感器都有自己的跟踪器处理本地测量,用于测量-跟踪关联和状态估计。只有经过处理的数据、包含状态向量估计及其协方差矩阵的局部轨迹(LT)才被传输到融合中心(FC)。本文提出了一种考虑所有关联假设的多目标混合概率信息矩阵融合(MO-HPIMF)。在MO-HPIMF中,FC轨迹状态(预测)与来自局部传感器的LT状态估计之间进行了关联。当融合对象和传感器数量较大时,MO-HPIMF中只应采用m-best fc -track- lt关联假设,以降低计算复杂度。本文采用序列m-最优二维方法求解多维分配问题。仿真结果表明,MO-HPIMF可以成功地跟踪所有感兴趣的目标,并且优于依赖于硬关联决策的分布式传感器融合系统中常用的航迹到航迹融合(T2TF)。
{"title":"Probabilistic Information Matrix Fusion in a Multi-Object Environment","authors":"Kaipei Yang, Y. Bar-Shalom","doi":"10.1109/MFI55806.2022.9913872","DOIUrl":"https://doi.org/10.1109/MFI55806.2022.9913872","url":null,"abstract":"In distributed sensor fusion systems, each of the local sensors has its own tracker processing local measurements for measurement-to-track association and state estimation. Only the processed data, local tracks (LT) comprising state vector estimates and their covariance matrices are transmitted to the fusion center (FC). In this work, a multi-object hybrid probabilistic information matrix fusion (MO-HPIMF) is derived taking into account all association hypotheses. In MO-HPIMF, the association carried out is between the FC track states (prediction) and the LT state estimates from local sensors. When having a large number of objects and sensors in fusion, only the m-best FC-track-to-LT association hypotheses should be incorporated in MO-HPIMF to reduce the computational complexity. A Sequential m-best 2-D method is used for solving the multidimensional assignment problem in this work. It is shown in the simulations that MO-HPIMF can successfully track all targets of interest and is superior to track-to-track fusion (T2TF, a commonly used approach in distributed sensor fusion system) which relies on hard association decisions.","PeriodicalId":344737,"journal":{"name":"2022 IEEE International Conference on Multisensor Fusion and Integration for Intelligent Systems (MFI)","volume":"116 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-09-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129456474","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Self-Supervised Learning and Multi-Task Pre-Training Based Single-Channel Acoustic Denoising 基于自监督学习和多任务预训练的单通道声学去噪
Yi Li, Yang Sun, S. M. Naqvi
In self-supervised learning-based single-channel speech denoising problem, it is challenging to reduce the gap between the denoising performance on the estimated and target speech signals with existed pre-tasks. In this paper, we propose a multi-task pre-training method to improve the speech denoising performance within self-supervised learning. In the proposed pre-training autoencoder (PAE), only a very limited set of unpaired and unseen clean speech signals are required to learn speech latent representations. Meanwhile, to solve the limitation of existing single pre-task, the proposed masking module exploits the dereverberated mask and estimated ratio mask to denoise the mixture as the new pre-task. The downstream task autoencoder (DAE) utilizes unlabeled and unseen reverberant mixtures to generate the estimated mixtures. The DAE is trained to share a latent representation with the clean examples from the learned representation in the PAE. Experimental results on a benchmark dataset demonstrate that the proposed method outperforms the state-of-the-art approaches.
在基于自监督学习的单通道语音去噪问题中,如何利用已有的预任务减小估计语音信号与目标语音信号去噪性能之间的差距是一个挑战。在本文中,我们提出了一种多任务预训练方法来提高自监督学习中的语音去噪性能。在本文提出的预训练自编码器(PAE)中,只需要一组非常有限的未配对和未见过的干净语音信号来学习语音潜在表征。同时,为了解决现有单一预任务的局限性,本文提出的掩模模块利用脱脱掩模和估计比例掩模作为新的预任务对混合信号进行去噪。下游任务自编码器(DAE)利用未标记和不可见的混响混合物来产生估计的混合物。DAE被训练成与PAE中学习到的表示中的干净示例共享潜在表示。在一个基准数据集上的实验结果表明,该方法优于目前最先进的方法。
{"title":"Self-Supervised Learning and Multi-Task Pre-Training Based Single-Channel Acoustic Denoising","authors":"Yi Li, Yang Sun, S. M. Naqvi","doi":"10.1109/MFI55806.2022.9913855","DOIUrl":"https://doi.org/10.1109/MFI55806.2022.9913855","url":null,"abstract":"In self-supervised learning-based single-channel speech denoising problem, it is challenging to reduce the gap between the denoising performance on the estimated and target speech signals with existed pre-tasks. In this paper, we propose a multi-task pre-training method to improve the speech denoising performance within self-supervised learning. In the proposed pre-training autoencoder (PAE), only a very limited set of unpaired and unseen clean speech signals are required to learn speech latent representations. Meanwhile, to solve the limitation of existing single pre-task, the proposed masking module exploits the dereverberated mask and estimated ratio mask to denoise the mixture as the new pre-task. The downstream task autoencoder (DAE) utilizes unlabeled and unseen reverberant mixtures to generate the estimated mixtures. The DAE is trained to share a latent representation with the clean examples from the learned representation in the PAE. Experimental results on a benchmark dataset demonstrate that the proposed method outperforms the state-of-the-art approaches.","PeriodicalId":344737,"journal":{"name":"2022 IEEE International Conference on Multisensor Fusion and Integration for Intelligent Systems (MFI)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-09-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130250025","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
A Spatio-Temporal-Semantic Environment Representation for Autonomous Mobile Robots equipped with various Sensor Systems 基于不同传感器系统的自主移动机器人时空语义环境表征
Mark Niemeyer, Sebastian Pütz, J. Hertzberg
The large amount of high resolution sensor data, both temporal and spatial, that autonomous mobile robots collect in today’s systems requires structured and efficient management and storage during the robot mission. In response, we present SEEREP: A Spatio-Temporal-Semantic Environment Representation for Autonomous Mobile Robots. SEEREP handles various types of data at once and provides an efficient query interface for all three modalities that can be combined for high-level analyses. It supports common robotic sensor data types such as images and point clouds, as well as sensor and robot coordinate frames changing over time. Furthermore, SEEREP provides an efficient HDF5-based storage system running on the robot during operation, compatible with ROS and the corresponding sensor message definitions. The compressed HDF5 data backend can be transferred efficiently to an application server with a running SEEREP query server providing gRPC interfaces with Protobuf and Flattbuffer message types. The query server can support high-level planning and reasoning systems in e.g. agricultural environments, or other partially unstructured environments that change over time. In this paper we show that SEEREP is much better suited for these tasks than a traditional GIS, which cannot handle the different types of robotic sensor data.
在当今的系统中,自主移动机器人收集的大量高分辨率传感器数据,包括时间和空间,需要在机器人任务期间进行结构化和有效的管理和存储。作为回应,我们提出了SEEREP:自主移动机器人的时空语义环境表示。SEEREP可以同时处理各种类型的数据,并为所有三种模式提供有效的查询接口,这些模式可以组合起来进行高级分析。它支持常见的机器人传感器数据类型,如图像和点云,以及传感器和机器人坐标帧随时间变化。此外,SEEREP提供了一个高效的基于hdf5的存储系统,在机器人操作期间运行,与ROS和相应的传感器消息定义兼容。压缩的HDF5数据后端可以高效地传输到具有运行SEEREP查询服务器的应用服务器,该服务器提供带有Protobuf和Flattbuffer消息类型的gRPC接口。查询服务器可以支持高级规划和推理系统,例如农业环境,或其他随时间变化的部分非结构化环境。在本文中,我们表明SEEREP比传统的GIS更适合这些任务,传统的GIS不能处理不同类型的机器人传感器数据。
{"title":"A Spatio-Temporal-Semantic Environment Representation for Autonomous Mobile Robots equipped with various Sensor Systems","authors":"Mark Niemeyer, Sebastian Pütz, J. Hertzberg","doi":"10.1109/MFI55806.2022.9913873","DOIUrl":"https://doi.org/10.1109/MFI55806.2022.9913873","url":null,"abstract":"The large amount of high resolution sensor data, both temporal and spatial, that autonomous mobile robots collect in today’s systems requires structured and efficient management and storage during the robot mission. In response, we present SEEREP: A Spatio-Temporal-Semantic Environment Representation for Autonomous Mobile Robots. SEEREP handles various types of data at once and provides an efficient query interface for all three modalities that can be combined for high-level analyses. It supports common robotic sensor data types such as images and point clouds, as well as sensor and robot coordinate frames changing over time. Furthermore, SEEREP provides an efficient HDF5-based storage system running on the robot during operation, compatible with ROS and the corresponding sensor message definitions. The compressed HDF5 data backend can be transferred efficiently to an application server with a running SEEREP query server providing gRPC interfaces with Protobuf and Flattbuffer message types. The query server can support high-level planning and reasoning systems in e.g. agricultural environments, or other partially unstructured environments that change over time. In this paper we show that SEEREP is much better suited for these tasks than a traditional GIS, which cannot handle the different types of robotic sensor data.","PeriodicalId":344737,"journal":{"name":"2022 IEEE International Conference on Multisensor Fusion and Integration for Intelligent Systems (MFI)","volume":"27 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-09-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128819491","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
期刊
2022 IEEE International Conference on Multisensor Fusion and Integration for Intelligent Systems (MFI)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1