首页 > 最新文献

2022 International Workshop on Intelligent Systems (IWIS)最新文献

英文 中文
TASuRe: Text Aware Super-Resolution TASuRe:文本感知超分辨率
Pub Date : 2022-08-17 DOI: 10.1109/IWIS56333.2022.9920903
Elena Filonenko, A. Filonenko, K. Jo
Recognition of text on low-resolution (LR) images is a challenging task. Traditional interpolation methods, as well as general super-resolution approaches, do not recover the shape of text character robustly. In this work, we propose a text-aware super- resolution neural network called TASuRe. Text awareness is interwoven in the proposed network that contains a text rectification part and text recognition auxiliary module. The training procedure is built around character shape restoration by adding a binary mask to the input image and using a specialized loss that penalties the network for missing gradients on the border of characters. Experiments on the real LR images have shown that the proposed network can deal with hard cases better than convolutional competitors.
低分辨率图像上的文本识别是一项具有挑战性的任务。传统的插值方法和一般的超分辨率插值方法都不能很好地恢复文本字符的形状。在这项工作中,我们提出了一个文本感知的超分辨率神经网络称为TASuRe。该网络包含文本纠错模块和文本识别辅助模块。训练过程是围绕字符形状恢复建立的,通过向输入图像添加二进制掩码,并使用专门的损失来惩罚网络在字符边界上丢失的梯度。在真实LR图像上的实验表明,该网络可以比卷积网络更好地处理复杂情况。
{"title":"TASuRe: Text Aware Super-Resolution","authors":"Elena Filonenko, A. Filonenko, K. Jo","doi":"10.1109/IWIS56333.2022.9920903","DOIUrl":"https://doi.org/10.1109/IWIS56333.2022.9920903","url":null,"abstract":"Recognition of text on low-resolution (LR) images is a challenging task. Traditional interpolation methods, as well as general super-resolution approaches, do not recover the shape of text character robustly. In this work, we propose a text-aware super- resolution neural network called TASuRe. Text awareness is interwoven in the proposed network that contains a text rectification part and text recognition auxiliary module. The training procedure is built around character shape restoration by adding a binary mask to the input image and using a specialized loss that penalties the network for missing gradients on the border of characters. Experiments on the real LR images have shown that the proposed network can deal with hard cases better than convolutional competitors.","PeriodicalId":340399,"journal":{"name":"2022 International Workshop on Intelligent Systems (IWIS)","volume":"147 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-08-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115277824","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Attribute estimation using pedestrian clothing and autonomous mobile robot navigation based on it 基于行人服装的属性估计及其自主移动机器人导航
Pub Date : 2022-08-17 DOI: 10.1109/IWIS56333.2022.9920723
Hidenao Takahashi, Nami Takino, H. Hashimoto, D. Chugo, S. Muramatsu, S. Yokota, Jin-Hua She, H. Hashimoto
This paper proposes a new method for estimating the attributes of a pedestrian, e.g. an office worker or a student, from the pedestrian's clothing and, based on the estimation results, a method for efficiently navigating an autonomous mobile robot while safely avoiding pedestrians. The robot can estimate the organization the pedestrians belong to by learning in advance the clothing of pedestrians that frequently appear in the city in which the is travelling, e.g. the uniforms of certain high school students. Furthermore, the robot is given the direction in which the person in that organization is likely to be heading at the current time, so it uses this information to estimate the future walking trajectory of that pedestrian and design a running path that is less likely to prevent that pedestrian from walking. We use the artificial potential method to design the robot's running trajectory. Even if pedestrians with different attributes are detected in the robot's running path, our proposed robot can efficiently reach its destination while avoiding those pedestrians appropriately in real time. The effectiveness of the proposed method was confirmed by computer simulations and subjects' experiments with our prototype robot.
本文提出了一种从行人的服装中估计行人(如上班族或学生)属性的新方法,并基于估计结果,提出了一种有效导航自主移动机器人同时安全避开行人的方法。机器人通过提前学习在其所行经的城市中经常出现的行人的服装,例如某高中生的制服,来估计行人属于哪个组织。此外,机器人被告知该组织中人员当前可能走向的方向,因此它使用这些信息来估计该行人未来的行走轨迹,并设计一条不太可能阻止该行人行走的跑步路径。采用人工势法设计机器人的运行轨迹。即使在机器人的运行路径中检测到不同属性的行人,我们提出的机器人也能在实时避开这些行人的同时有效地到达目的地。通过计算机仿真和实验验证了该方法的有效性。
{"title":"Attribute estimation using pedestrian clothing and autonomous mobile robot navigation based on it","authors":"Hidenao Takahashi, Nami Takino, H. Hashimoto, D. Chugo, S. Muramatsu, S. Yokota, Jin-Hua She, H. Hashimoto","doi":"10.1109/IWIS56333.2022.9920723","DOIUrl":"https://doi.org/10.1109/IWIS56333.2022.9920723","url":null,"abstract":"This paper proposes a new method for estimating the attributes of a pedestrian, e.g. an office worker or a student, from the pedestrian's clothing and, based on the estimation results, a method for efficiently navigating an autonomous mobile robot while safely avoiding pedestrians. The robot can estimate the organization the pedestrians belong to by learning in advance the clothing of pedestrians that frequently appear in the city in which the is travelling, e.g. the uniforms of certain high school students. Furthermore, the robot is given the direction in which the person in that organization is likely to be heading at the current time, so it uses this information to estimate the future walking trajectory of that pedestrian and design a running path that is less likely to prevent that pedestrian from walking. We use the artificial potential method to design the robot's running trajectory. Even if pedestrians with different attributes are detected in the robot's running path, our proposed robot can efficiently reach its destination while avoiding those pedestrians appropriately in real time. The effectiveness of the proposed method was confirmed by computer simulations and subjects' experiments with our prototype robot.","PeriodicalId":340399,"journal":{"name":"2022 International Workshop on Intelligent Systems (IWIS)","volume":"48 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-08-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123064458","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Sensor Fusion of Camera and 2D LiDAR for Self-Driving Automobile in Obstacle Avoidance Scenarios 自动驾驶汽车避障场景中摄像头与2D激光雷达传感器融合
Pub Date : 2022-08-17 DOI: 10.1109/IWIS56333.2022.9920917
Tan-Thien-Nien Nguyen, Thanh-Danh Phan, Minh-Thien Duong, Chi-Tam Nguyen, Hong-Phong Ly, M. Le
Obstacle dodging and overtaking are the pivotal tasks ensuring safety for self-driving automobiles. Multi-sensors fusion is the must-required condition to explore the entire surrounding information. This paper proposes a novel frontal dynamic car dodging strategy for automobiles with the left-hand side steering wheel by fusion of a camera and 2D LiDAR features. To begin with, we improve the LiteSeg model to extract the segmented mask, which can determine the drivable area and the avoiding direction. In addition to a camera, 2D LiDAR is used to check the scene information on the right side, which the camera's range cannot cover. As for point clouds output of 2D LiDAR, we adopt the Adaptive Breakpoint Detection (defined as ABD) algorithm to cluster the objects in a scanning plane. Subsequently, the RANSAC algorithm forms a straight line from the clustered point clouds to determine the boundary of the right-side obstacle. Besides, we compute the distance from LiDAR to the estimated straight line to maintain a safe distance when overtaking. Last but not least, the post-processing results of the two devices are fused to decide on the obstacle dodging and overtaking. The comprehensive experiments reveal that our self-driving automobile could perform well on the university campus in diverse scenarios.
避障和超车是确保自动驾驶汽车安全的关键任务。多传感器融合是探索整个周围信息的必要条件。本文提出了一种融合摄像头和二维激光雷达特征的左侧方向盘汽车正面动态闪避策略。首先,我们对LiteSeg模型进行改进,提取出分割后的掩码,从而确定可行驶区域和避开方向。除了摄像头之外,2D LiDAR还用于检查右侧的场景信息,这是摄像头无法覆盖的。对于2D LiDAR的点云输出,我们采用自适应断点检测(Adaptive Breakpoint Detection,定义为ABD)算法对扫描平面内的目标进行聚类。随后,RANSAC算法从聚类点云形成一条直线来确定右侧障碍物的边界。此外,我们还计算了激光雷达到估计直线的距离,以便在超车时保持安全距离。最后,将两个设备的后处理结果进行融合,以确定避障和超车。综合实验表明,我们的自动驾驶汽车可以在大学校园的各种场景中表现良好。
{"title":"Sensor Fusion of Camera and 2D LiDAR for Self-Driving Automobile in Obstacle Avoidance Scenarios","authors":"Tan-Thien-Nien Nguyen, Thanh-Danh Phan, Minh-Thien Duong, Chi-Tam Nguyen, Hong-Phong Ly, M. Le","doi":"10.1109/IWIS56333.2022.9920917","DOIUrl":"https://doi.org/10.1109/IWIS56333.2022.9920917","url":null,"abstract":"Obstacle dodging and overtaking are the pivotal tasks ensuring safety for self-driving automobiles. Multi-sensors fusion is the must-required condition to explore the entire surrounding information. This paper proposes a novel frontal dynamic car dodging strategy for automobiles with the left-hand side steering wheel by fusion of a camera and 2D LiDAR features. To begin with, we improve the LiteSeg model to extract the segmented mask, which can determine the drivable area and the avoiding direction. In addition to a camera, 2D LiDAR is used to check the scene information on the right side, which the camera's range cannot cover. As for point clouds output of 2D LiDAR, we adopt the Adaptive Breakpoint Detection (defined as ABD) algorithm to cluster the objects in a scanning plane. Subsequently, the RANSAC algorithm forms a straight line from the clustered point clouds to determine the boundary of the right-side obstacle. Besides, we compute the distance from LiDAR to the estimated straight line to maintain a safe distance when overtaking. Last but not least, the post-processing results of the two devices are fused to decide on the obstacle dodging and overtaking. The comprehensive experiments reveal that our self-driving automobile could perform well on the university campus in diverse scenarios.","PeriodicalId":340399,"journal":{"name":"2022 International Workshop on Intelligent Systems (IWIS)","volume":"221 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-08-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116420187","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
A Fast Real-time Face Gender Detector on CPU using Superficial Network with Attention Modules 基于关注模块的CPU表面网络快速实时人脸性别检测
Pub Date : 2022-08-17 DOI: 10.1109/IWIS56333.2022.9920714
Adri Priadana, M. D. Putro, Changhyun Jeong, K. Jo
A gender detector has become an essential part of digital signage to support the decision to provide relevant ads for each audience. Application installed in digital signage must be capable of running on low-cost or CPU devices to minimize system costs. This study proposed a fast face gender detector (Gender-CPU) that can sprint in real-time on CPU devices implemented on digital signage. The proposed architecture contains a superficial network with attention modules (SufiaNet). This architecture only consists of three convolution layers, making it super shallow and generating a small number of parameters. In order to redeem the lack of a super shallow network, the global attention module is assigned to improve the quality of the feature map resulting from the previous convolution layers. In the experiment, the training and validation process is conducted on the UTKFace, the Adience Benchmark, and the Labeled Faces in the Wild (LFW) datasets. The SufiaNet gains competitive accuracy compared to other common and light architectures on the three datasets. Moreover, the detector can run 84.97 frames per second on a CPU device, which is fast to run in real-time.
性别探测器已经成为数字标牌的重要组成部分,以支持为每个受众提供相关广告的决定。安装在数字标牌中的应用程序必须能够在低成本或CPU设备上运行,以最大限度地降低系统成本。本研究提出了一种快速的人脸性别检测器(gender -CPU),可以在数字标牌上实现CPU设备的实时冲刺。所提出的体系结构包含一个具有注意力模块的表层网络(SufiaNet)。该架构仅由三个卷积层组成,使其非常浅,并且生成少量参数。为了弥补超浅网络的不足,分配了全局关注模块,以提高前几层卷积得到的特征映射的质量。在实验中,在UTKFace、Adience Benchmark和Labeled Faces In the Wild (LFW)数据集上进行了训练和验证过程。在这三个数据集上,SufiaNet获得了与其他常见和轻量级架构相比具有竞争力的精度。此外,检测器可以在CPU设备上每秒运行84.97帧,实时运行速度很快。
{"title":"A Fast Real-time Face Gender Detector on CPU using Superficial Network with Attention Modules","authors":"Adri Priadana, M. D. Putro, Changhyun Jeong, K. Jo","doi":"10.1109/IWIS56333.2022.9920714","DOIUrl":"https://doi.org/10.1109/IWIS56333.2022.9920714","url":null,"abstract":"A gender detector has become an essential part of digital signage to support the decision to provide relevant ads for each audience. Application installed in digital signage must be capable of running on low-cost or CPU devices to minimize system costs. This study proposed a fast face gender detector (Gender-CPU) that can sprint in real-time on CPU devices implemented on digital signage. The proposed architecture contains a superficial network with attention modules (SufiaNet). This architecture only consists of three convolution layers, making it super shallow and generating a small number of parameters. In order to redeem the lack of a super shallow network, the global attention module is assigned to improve the quality of the feature map resulting from the previous convolution layers. In the experiment, the training and validation process is conducted on the UTKFace, the Adience Benchmark, and the Labeled Faces in the Wild (LFW) datasets. The SufiaNet gains competitive accuracy compared to other common and light architectures on the three datasets. Moreover, the detector can run 84.97 frames per second on a CPU device, which is fast to run in real-time.","PeriodicalId":340399,"journal":{"name":"2022 International Workshop on Intelligent Systems (IWIS)","volume":"50 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-08-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128504360","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Safe Cooperative Intersection of Autonomous and Connected Robots: Trajectory and Schedule Optimization 自主与互联机器人的安全协同交叉口:轨迹与调度优化
Pub Date : 2022-08-17 DOI: 10.1109/IWIS56333.2022.9920759
Wendan Du, A. Abbas-Turki, A. Koukam, K. Jo
Conventional intersection managements, such as signalized intersections, may not necessarily be the optimal strategies when it comes to connected and automated vehicles (CAV) environment. Cooperative intersection management (CIM) is tailored for CAVs aiming at replacing the conventional traffic control strategies. This paper focuses on the safety constraint in the case of conflicting movements. The constraint is needed for both optimization: robot trajectory (speed profile) and sequence (which robot crosses the intersection first, which is the second one and so on). The main objective is to safely minimize the lost time between two conflicting robots while saving energy. The paper studies the safety condition of the cyber-physical system and derives the optimal control point that saves time. Quadratic optimization is used to derive the speed trajectory of the robot in order to save energy. Simulations show that the proposed approach allows safe and efficient cooperative intersection.
传统的交叉口管理,如信号交叉口,可能不一定是最优策略,当涉及到连接和自动驾驶汽车(CAV)的环境。协同交叉口管理(CIM)是为自动驾驶汽车量身定制的,旨在取代传统的交通控制策略。本文主要研究冲突运动情况下的安全约束问题。机器人轨迹优化(速度剖面)和顺序优化(哪个机器人先过路口,哪个是第二个,等等)都需要约束。主要目标是在节省能量的同时,最大限度地减少两个冲突机器人之间的时间损失。研究了网络物理系统的安全状况,导出了节省时间的最优控制点。采用二次优化方法推导机器人的速度轨迹,以节省能量。仿真结果表明,该方法能够实现安全、高效的协同交叉口。
{"title":"Safe Cooperative Intersection of Autonomous and Connected Robots: Trajectory and Schedule Optimization","authors":"Wendan Du, A. Abbas-Turki, A. Koukam, K. Jo","doi":"10.1109/IWIS56333.2022.9920759","DOIUrl":"https://doi.org/10.1109/IWIS56333.2022.9920759","url":null,"abstract":"Conventional intersection managements, such as signalized intersections, may not necessarily be the optimal strategies when it comes to connected and automated vehicles (CAV) environment. Cooperative intersection management (CIM) is tailored for CAVs aiming at replacing the conventional traffic control strategies. This paper focuses on the safety constraint in the case of conflicting movements. The constraint is needed for both optimization: robot trajectory (speed profile) and sequence (which robot crosses the intersection first, which is the second one and so on). The main objective is to safely minimize the lost time between two conflicting robots while saving energy. The paper studies the safety condition of the cyber-physical system and derives the optimal control point that saves time. Quadratic optimization is used to derive the speed trajectory of the robot in order to save energy. Simulations show that the proposed approach allows safe and efficient cooperative intersection.","PeriodicalId":340399,"journal":{"name":"2022 International Workshop on Intelligent Systems (IWIS)","volume":"38 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-08-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134279472","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Lightweight Bird Eye View Detection Network with Bridge Block Based on YOLOv5 基于YOLOv5的轻型桥块鸟瞰图检测网络
Pub Date : 2022-08-17 DOI: 10.1109/IWIS56333.2022.9920755
Jehwan Choi, Kanghyun Jo
In this paper, The network with a faster detection speed than the original YOLOv5 nano model is proposed. The network defined as a bridge module reduced the number of channels and changed the speed quickly by applying pixel-wise operation instead of using a convolution layer. Especially, element-wise addition operation of each output feature maps is the main method. As a result, the detection speed is faster than the original detection method about 30 35%. On the other hand, mAP (mean average precision) is recorded at 50.7%, which is 1.4% lower than the original detection method. However, the original detection method showed good results in 3 classes and the proposed method showed good results in 5 classes. And the proposed method detected more objects in a detection result image. Therefore, the proposed method is a more efficient object detection network.
本文提出了比原来的YOLOv5纳米模型检测速度更快的网络。定义为桥接模块的网络通过应用逐像素操作而不是使用卷积层来减少通道数量并快速改变速度。其中,各输出特征映射的逐元加法运算是主要方法。因此,检测速度比原来的检测方法快约30 - 35%。另一方面,mAP (mean average precision)为50.7%,比原检测方法降低了1.4%。然而,原检测方法在3个类别中显示出良好的结果,而本文方法在5个类别中显示出良好的结果。该方法在一幅检测结果图像中检测出更多的目标。因此,本文提出的方法是一种更高效的目标检测网络。
{"title":"Lightweight Bird Eye View Detection Network with Bridge Block Based on YOLOv5","authors":"Jehwan Choi, Kanghyun Jo","doi":"10.1109/IWIS56333.2022.9920755","DOIUrl":"https://doi.org/10.1109/IWIS56333.2022.9920755","url":null,"abstract":"In this paper, The network with a faster detection speed than the original YOLOv5 nano model is proposed. The network defined as a bridge module reduced the number of channels and changed the speed quickly by applying pixel-wise operation instead of using a convolution layer. Especially, element-wise addition operation of each output feature maps is the main method. As a result, the detection speed is faster than the original detection method about 30 35%. On the other hand, mAP (mean average precision) is recorded at 50.7%, which is 1.4% lower than the original detection method. However, the original detection method showed good results in 3 classes and the proposed method showed good results in 5 classes. And the proposed method detected more objects in a detection result image. Therefore, the proposed method is a more efficient object detection network.","PeriodicalId":340399,"journal":{"name":"2022 International Workshop on Intelligent Systems (IWIS)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-08-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117330738","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Automatic Feature Detection and Classification for Watermelon (Citrillus lanatus) 西瓜特征自动检测与分类
Pub Date : 2022-08-17 DOI: 10.1109/IWIS56333.2022.9920868
J. Sánchez-Galán, Anel Henry Royo, K. Jo, Danilo Cáceres Hernández
This document focuses on the contributions made in the development and advances achieved in the task of automatic Feature Detection for Watermelon (Citrillus lanatus). A special interest is given to feature-based methods such as: morphological and adaptive threshold approaches, that work by extracting color and texture information. A first hand example about how these two methods can be applied to a data set comprised in export level watermelons coming from Panama is provided. Limitations of the method are discussed and a final conclusion about the field and recent avenues of work with ensemble methods is given. The importance of this document is that it helps the automatic understanding of watermelon patterns with computer vision.
本文主要介绍了西瓜(Citrillus lanatus)特征自动检测的发展和取得的进展。特别关注基于特征的方法,如形态学和自适应阈值方法,它们通过提取颜色和纹理信息来工作。提供了一个关于如何将这两种方法应用于来自巴拿马的出口级西瓜数据集的第一手示例。讨论了该方法的局限性,最后总结了集成方法的研究领域和最新研究方向。该文档的重要性在于它有助于计算机视觉对西瓜图案的自动理解。
{"title":"Automatic Feature Detection and Classification for Watermelon (Citrillus lanatus)","authors":"J. Sánchez-Galán, Anel Henry Royo, K. Jo, Danilo Cáceres Hernández","doi":"10.1109/IWIS56333.2022.9920868","DOIUrl":"https://doi.org/10.1109/IWIS56333.2022.9920868","url":null,"abstract":"This document focuses on the contributions made in the development and advances achieved in the task of automatic Feature Detection for Watermelon (Citrillus lanatus). A special interest is given to feature-based methods such as: morphological and adaptive threshold approaches, that work by extracting color and texture information. A first hand example about how these two methods can be applied to a data set comprised in export level watermelons coming from Panama is provided. Limitations of the method are discussed and a final conclusion about the field and recent avenues of work with ensemble methods is given. The importance of this document is that it helps the automatic understanding of watermelon patterns with computer vision.","PeriodicalId":340399,"journal":{"name":"2022 International Workshop on Intelligent Systems (IWIS)","volume":"96 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-08-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117315190","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
A Study on Efficient Multi-task Networks for Multiple Object Tracking 多目标跟踪的高效多任务网络研究
Pub Date : 2022-08-17 DOI: 10.1109/IWIS56333.2022.9920879
Xuan-Thuy Vo, T. Tran, Duy-Linh Nguyen, K. Jo
Multiple object tracking involves multi-task learning to handle object detection and data association tasks concurrently. Conventionally, object detection consists of object classification and object localization (e.g., object regression) tasks, and data association is treated as a classification task. However, various tasks can cause inconsistent learning due to that the learning targets of object detection and data association tasks are different. Object detection focuses on positional information of objects while data association requires strong semantic information to identify same object target. Besides, advantageous character of multi-task learning is the correlation between tasks, and adopting such character in learning the networks can result in better generalization performance. However, existing multiple object tracking methods learn this information by treating multi-task branches independently. To understand the behaviours of multi-task networks in multiple object tracking, in this paper, we explore task-dependent representations through empirical experiments and observe that multi-task branches in multiple object tracking are complementary. To better learn such information, we introduce a novel Correlation Estimation (CE) module to estimate the correlation between object classification and bounding box regression based on statistical features of box regression quality. Finally, extensive experiments are conducted on the benchmark dataset MOT17. As a result, our method outperforms state-of-the-art online trackers without requiring additional training datasets.
多目标跟踪涉及多任务学习,可以同时处理目标检测和数据关联任务。通常,目标检测包括目标分类和目标定位(例如,目标回归)任务,数据关联被视为分类任务。然而,由于对象检测和数据关联任务的学习目标不同,不同的任务会导致学习不一致。目标检测关注的是目标的位置信息,而数据关联需要很强的语义信息来识别相同的目标。此外,多任务学习的优点是任务间的相关性,在学习网络时采用这种特性可以获得更好的泛化性能。然而,现有的多目标跟踪方法通过独立处理多任务分支来学习这些信息。为了理解多任务网络在多目标跟踪中的行为,本文通过实证实验探索任务相关表征,并观察到多目标跟踪中的多任务分支是互补的。为了更好地学习这些信息,我们引入了一种新的Correlation Estimation (CE)模块,基于盒回归质量的统计特征来估计目标分类与边界盒回归之间的相关性。最后,在基准数据集MOT17上进行了大量的实验。因此,我们的方法优于最先进的在线跟踪器,而不需要额外的训练数据集。
{"title":"A Study on Efficient Multi-task Networks for Multiple Object Tracking","authors":"Xuan-Thuy Vo, T. Tran, Duy-Linh Nguyen, K. Jo","doi":"10.1109/IWIS56333.2022.9920879","DOIUrl":"https://doi.org/10.1109/IWIS56333.2022.9920879","url":null,"abstract":"Multiple object tracking involves multi-task learning to handle object detection and data association tasks concurrently. Conventionally, object detection consists of object classification and object localization (e.g., object regression) tasks, and data association is treated as a classification task. However, various tasks can cause inconsistent learning due to that the learning targets of object detection and data association tasks are different. Object detection focuses on positional information of objects while data association requires strong semantic information to identify same object target. Besides, advantageous character of multi-task learning is the correlation between tasks, and adopting such character in learning the networks can result in better generalization performance. However, existing multiple object tracking methods learn this information by treating multi-task branches independently. To understand the behaviours of multi-task networks in multiple object tracking, in this paper, we explore task-dependent representations through empirical experiments and observe that multi-task branches in multiple object tracking are complementary. To better learn such information, we introduce a novel Correlation Estimation (CE) module to estimate the correlation between object classification and bounding box regression based on statistical features of box regression quality. Finally, extensive experiments are conducted on the benchmark dataset MOT17. As a result, our method outperforms state-of-the-art online trackers without requiring additional training datasets.","PeriodicalId":340399,"journal":{"name":"2022 International Workshop on Intelligent Systems (IWIS)","volume":"7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-08-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116060485","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Using Disparity Map for Moving Object Position Estimation in Pan Tilt Camera Images 利用视差图估计平移倾斜相机图像中的运动目标位置
Pub Date : 2022-08-17 DOI: 10.1109/IWIS56333.2022.9920838
Tomasz Kocejko, J. Rumiński, K. Jo
In this paper we present the algorithm for rapid moving object position estimation in an images acquired from pan tilt camera. Detection of a moving object in a image acquired from a moving camera might be quite challenging. Standard methods that relay on analyzing two consecutive frames are not applicable due to the changing background. To overtake this problem we decided to evaluate the possibility of calculating a disparity map based on this consecutive images. As a result we were able to obtain approximate position of moving object in real time. Using cpu the average detection time was below 0.17 second.
本文提出了一种快速估计平移相机图像中运动目标位置的算法。从移动摄像机获取的图像中检测移动物体可能相当具有挑战性。由于背景的变化,依赖于分析两个连续帧的标准方法不适用。为了解决这个问题,我们决定评估基于这些连续图像计算视差图的可能性。因此,我们能够实时获得运动物体的大致位置。使用cpu时,平均检测时间低于0.17秒。
{"title":"Using Disparity Map for Moving Object Position Estimation in Pan Tilt Camera Images","authors":"Tomasz Kocejko, J. Rumiński, K. Jo","doi":"10.1109/IWIS56333.2022.9920838","DOIUrl":"https://doi.org/10.1109/IWIS56333.2022.9920838","url":null,"abstract":"In this paper we present the algorithm for rapid moving object position estimation in an images acquired from pan tilt camera. Detection of a moving object in a image acquired from a moving camera might be quite challenging. Standard methods that relay on analyzing two consecutive frames are not applicable due to the changing background. To overtake this problem we decided to evaluate the possibility of calculating a disparity map based on this consecutive images. As a result we were able to obtain approximate position of moving object in real time. Using cpu the average detection time was below 0.17 second.","PeriodicalId":340399,"journal":{"name":"2022 International Workshop on Intelligent Systems (IWIS)","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-08-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124889885","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Graph-based Attribute-aware Unsupervised Person Re-identification with Contrastive learning 基于图的属性感知无监督人再识别与对比学习
Pub Date : 2022-08-17 DOI: 10.1109/IWIS56333.2022.9920894
Ge Cao, Qing Tang, Kanghyun Jo
This paper is employed on the unsupervised per-son re-identification (Re-ID) task which does not leverage any annotation provided by the target dataset and other datasets. Previous works have investigated the effectiveness of applying self-supervised contrastive learning, which adopts the cluster-based method to generate the pseudo label and split each cluster into multiple proxies by camera ID. This paper applies the Attribute Enhancement Module (AEM), which utilizes Graph Convolutional Network to integrate the correlations between attributes, human body parts features, and the extracted dis-criminative feature. And the experiments are implemented to demonstrate the great performance of the proposed Attribute Enhancement Contrastive Learning (AECL) in camera-agnostic version and camera-aware version on two large-scale datasets, including Market-1501 and DukeMTMC-ReID. Compared with the baseline and the state-of-the-art, the proposed framework achieves competitive results.
本文用于无监督个人重新识别(Re-ID)任务,该任务不利用目标数据集和其他数据集提供的任何注释。之前的工作已经研究了应用自监督对比学习的有效性,该方法采用基于聚类的方法生成伪标签,并根据相机ID将每个聚类分成多个代理。本文采用属性增强模块(Attribute Enhancement Module, AEM),该模块利用图卷积网络(Graph Convolutional Network)对属性、人体部位特征和提取的判别特征之间的相关性进行整合。并在Market-1501和DukeMTMC-ReID两个大规模数据集上进行了实验,验证了所提出的属性增强对比学习(AECL)在相机不可知版本和相机感知版本上的优异性能。与基线和最先进的框架相比,所提出的框架具有竞争力。
{"title":"Graph-based Attribute-aware Unsupervised Person Re-identification with Contrastive learning","authors":"Ge Cao, Qing Tang, Kanghyun Jo","doi":"10.1109/IWIS56333.2022.9920894","DOIUrl":"https://doi.org/10.1109/IWIS56333.2022.9920894","url":null,"abstract":"This paper is employed on the unsupervised per-son re-identification (Re-ID) task which does not leverage any annotation provided by the target dataset and other datasets. Previous works have investigated the effectiveness of applying self-supervised contrastive learning, which adopts the cluster-based method to generate the pseudo label and split each cluster into multiple proxies by camera ID. This paper applies the Attribute Enhancement Module (AEM), which utilizes Graph Convolutional Network to integrate the correlations between attributes, human body parts features, and the extracted dis-criminative feature. And the experiments are implemented to demonstrate the great performance of the proposed Attribute Enhancement Contrastive Learning (AECL) in camera-agnostic version and camera-aware version on two large-scale datasets, including Market-1501 and DukeMTMC-ReID. Compared with the baseline and the state-of-the-art, the proposed framework achieves competitive results.","PeriodicalId":340399,"journal":{"name":"2022 International Workshop on Intelligent Systems (IWIS)","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-08-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132224061","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
期刊
2022 International Workshop on Intelligent Systems (IWIS)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1