首页 > 最新文献

IS&T International Symposium on Electronic Imaging最新文献

英文 中文
Generative adversarial networks (GANs) and object tracking (OT) for vehicle accident detection 生成对抗网络(GANs)和目标跟踪(OT)在车辆事故检测中的应用
Pub Date : 2023-01-16 DOI: 10.2352/ei.2023.35.3.mobmu-364
Taraka Rama Krishna Kanth Kannuri, Kirsnaragavan Arudpiragasam, Klaus Schwarz, Michael Hartmann, Reiner Creutzburg
Accident detection is one of the biggest challenges as there are various anomalies, occlusions, and objects in the image at different times. Therefore, this paper focuses on detecting traffic accidents through a combination of Object Tracking (OT) and image generation using GAN with variants such as skip connection, residual, and attention connection. The background removal techniques will be applied to reduce the background variation in the frame. Later, YOLO-R is used to detect objects, followed by DeepSort tracking of objects in the frame. Finally, the distance error metric and the adversarial error are determined using the Kalman filter and the GAN approach and help to decide accidents in videos.
事故检测是最大的挑战之一,因为在不同的时间图像中存在各种异常、遮挡和物体。因此,本文的重点是通过结合目标跟踪(OT)和使用GAN的图像生成来检测交通事故,其中包含跳跃连接、残差和注意连接等变体。背景去除技术将被用于减少背景变化的框架。然后使用YOLO-R对目标进行检测,然后对帧内的目标进行深度排序跟踪。最后,利用卡尔曼滤波和GAN方法确定距离误差度量和对抗误差,以帮助确定视频中的事故。
{"title":"Generative adversarial networks (GANs) and object tracking (OT) for vehicle accident detection","authors":"Taraka Rama Krishna Kanth Kannuri, Kirsnaragavan Arudpiragasam, Klaus Schwarz, Michael Hartmann, Reiner Creutzburg","doi":"10.2352/ei.2023.35.3.mobmu-364","DOIUrl":"https://doi.org/10.2352/ei.2023.35.3.mobmu-364","url":null,"abstract":"Accident detection is one of the biggest challenges as there are various anomalies, occlusions, and objects in the image at different times. Therefore, this paper focuses on detecting traffic accidents through a combination of Object Tracking (OT) and image generation using GAN with variants such as skip connection, residual, and attention connection. The background removal techniques will be applied to reduce the background variation in the frame. Later, YOLO-R is used to detect objects, followed by DeepSort tracking of objects in the frame. Finally, the distance error metric and the adversarial error are determined using the Kalman filter and the GAN approach and help to decide accidents in videos.","PeriodicalId":73514,"journal":{"name":"IS&T International Symposium on Electronic Imaging","volume":"138 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-01-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135694714","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Optical flow for autonomous driving: Applications, challenges and improvements 自动驾驶的光流:应用、挑战和改进
Pub Date : 2023-01-16 DOI: 10.2352/ei.2023.35.16.avm-128
Shihao Shen, Louis Kerofsky, Senthil Yogamani
Estimating optical flow presents unique challenges in AV applications: large translational motion, wide variations in depth of important objects, strong lens distortion in commonly used fisheye cameras and rolling shutter artefacts in dynamic scenes. Even simple translational motion can produce complicated optical flow fields. Lack of ground truth data also creates a challenge. We evaluate recent optical flow methods on fisheye imagery found in AV applications. We explore various training techniques in challenging scenarios and domain adaptation for transferring models trained on synthetic data where ground truth is available to real-world data. We propose novel strategies that facilitate learning robust representations efficiently to address low-light degeneracies. Finally, we discuss the main challenges and open problems in this problem domain.
估计光流在AV应用中提出了独特的挑战:大的平移运动,重要物体深度的广泛变化,常用鱼眼相机的强烈镜头畸变以及动态场景中的滚动快门伪影。即使是简单的平移运动也会产生复杂的光流场。地面真实数据的缺乏也带来了挑战。我们评估了最近在AV应用中发现的鱼眼图像的光流方法。我们在具有挑战性的场景和领域适应中探索各种训练技术,以转移在合成数据上训练的模型,其中地面真相可用于真实世界数据。我们提出了新的策略,促进学习鲁棒表示有效地解决弱光退化。最后,讨论了该问题领域的主要挑战和有待解决的问题。
{"title":"Optical flow for autonomous driving: Applications, challenges and improvements","authors":"Shihao Shen, Louis Kerofsky, Senthil Yogamani","doi":"10.2352/ei.2023.35.16.avm-128","DOIUrl":"https://doi.org/10.2352/ei.2023.35.16.avm-128","url":null,"abstract":"Estimating optical flow presents unique challenges in AV applications: large translational motion, wide variations in depth of important objects, strong lens distortion in commonly used fisheye cameras and rolling shutter artefacts in dynamic scenes. Even simple translational motion can produce complicated optical flow fields. Lack of ground truth data also creates a challenge. We evaluate recent optical flow methods on fisheye imagery found in AV applications. We explore various training techniques in challenging scenarios and domain adaptation for transferring models trained on synthetic data where ground truth is available to real-world data. We propose novel strategies that facilitate learning robust representations efficiently to address low-light degeneracies. Finally, we discuss the main challenges and open problems in this problem domain.","PeriodicalId":73514,"journal":{"name":"IS&T International Symposium on Electronic Imaging","volume":"128 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-01-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135644696","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
FastPoints: A state-of-the-art point cloud renderer for Unity FastPoints:最先进的Unity点云渲染器
Pub Date : 2023-01-16 DOI: 10.2352/ei.2023.35.1.vda-394
Elias Neuman-Donihue, Michael Jarvis, Yuhao Zhu
In this paper, we introduce FastPoints, a state-of-the-art point cloud renderer for the Unity game development platform. Our program supports standard unprocessed point cloud formats with non-programmatic, drag-and-drop support, and creates an out-of-core data structure for large clouds without requiring an explicit preprocessing step; instead, the software renders a decimated point cloud immediately and constructs a shallow octree online, during which time the Unity editor remains fully interactive.
在本文中,我们将介绍用于Unity游戏开发平台的最先进的点云渲染器FastPoints。我们的程序支持标准的未处理的点云格式与非编程,拖放支持,并创建大型云的核心外数据结构,而不需要明确的预处理步骤;相反,软件会立即渲染一个毁坏的点云,并在线构建一个浅八叉树,在此期间,Unity编辑器保持完全交互。
{"title":"FastPoints: A state-of-the-art point cloud renderer for Unity","authors":"Elias Neuman-Donihue, Michael Jarvis, Yuhao Zhu","doi":"10.2352/ei.2023.35.1.vda-394","DOIUrl":"https://doi.org/10.2352/ei.2023.35.1.vda-394","url":null,"abstract":"In this paper, we introduce FastPoints, a state-of-the-art point cloud renderer for the Unity game development platform. Our program supports standard unprocessed point cloud formats with non-programmatic, drag-and-drop support, and creates an out-of-core data structure for large clouds without requiring an explicit preprocessing step; instead, the software renders a decimated point cloud immediately and constructs a shallow octree online, during which time the Unity editor remains fully interactive.","PeriodicalId":73514,"journal":{"name":"IS&T International Symposium on Electronic Imaging","volume":"12 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-01-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135694179","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Practical OSINT investigation - Similarity calculation using Reddit user profile data 实用OSINT调查-相似度计算使用Reddit用户资料数据
Pub Date : 2023-01-16 DOI: 10.2352/ei.2023.35.3.mobmu-356
Valeria Vishnevskaya, Klaus Schwarz, Reiner Creutzburg
This paper presents a practical Open Source Intelligence (OSINT) use case for user similarity measurements with the use of open profile data from the Reddit social network. This PoC work combines the open data from Reddit and the part of the state-of-the-art BERT model. Using the PRAW Python library, the project fetches comments and posts of users. Then these texts are converted into a feature vector - representation of all user posts and comments. The main idea here is to create a comparable user's pair similarity score based on their comments and posts. For example, if we fix one user and calculate scores of all mutual pairs with other users, we will produce a total order on the set of all mutual pairs with that user. This total order can be described as a degree of written similarity with this chosen user. A set of "similar" users for one particular user can be used to recommend to the user interesting for him people. The similarity score also has a "transitive property": if $user_1$ is "similar" to $user_2$ and $user_2$ is similar to $user_3$ then inner properties of our model guarantees that $user_1$ and $user_3$ are pretty "similar" too. In this way, this score can be used to cluster a set of users into sets of "similar" users. It could be used in some recommendation algorithms or tune already existing algorithms to consider a cluster's peculiarities. Also, we can extend our model and calculate feature vectors for subreddits. In that way, we can find similar to the user's subreddits and recommend them to him.
本文提出了一个实用的开源智能(OSINT)用例,用于使用来自Reddit社交网络的开放个人资料数据进行用户相似性测量。这项PoC工作结合了Reddit的开放数据和最先进的BERT模型的一部分。该项目使用PRAW Python库获取用户的评论和帖子。然后将这些文本转换成一个特征向量-所有用户帖子和评论的表示。这里的主要思想是根据用户的评论和帖子创建一个可比较的用户配对相似度评分。例如,如果我们固定一个用户并计算与其他用户的所有互对的分数,我们将在与该用户的所有互对的集合上生成一个总排序。这个总顺序可以用与所选用户的书写相似度来描述。一个特定用户的一组“相似”用户可以用来向用户推荐他感兴趣的人。相似度得分也有一个“传递属性”:如果$user_1$与$user_2$“相似”,$user_2$与$user_3$相似,那么我们模型的内部属性保证$user_1$和$user_3$也非常“相似”。这样,这个分数就可以用来将一组用户聚类为“相似”用户集。它可以用在一些推荐算法中,或者调优已经存在的算法来考虑集群的特性。此外,我们可以扩展我们的模型并计算子reddit的特征向量。这样,我们就可以找到与用户相似的subreddits并推荐给他。
{"title":"Practical OSINT investigation - Similarity calculation using Reddit user profile data","authors":"Valeria Vishnevskaya, Klaus Schwarz, Reiner Creutzburg","doi":"10.2352/ei.2023.35.3.mobmu-356","DOIUrl":"https://doi.org/10.2352/ei.2023.35.3.mobmu-356","url":null,"abstract":"This paper presents a practical Open Source Intelligence (OSINT) use case for user similarity measurements with the use of open profile data from the Reddit social network. This PoC work combines the open data from Reddit and the part of the state-of-the-art BERT model. Using the PRAW Python library, the project fetches comments and posts of users. Then these texts are converted into a feature vector - representation of all user posts and comments. The main idea here is to create a comparable user's pair similarity score based on their comments and posts. For example, if we fix one user and calculate scores of all mutual pairs with other users, we will produce a total order on the set of all mutual pairs with that user. This total order can be described as a degree of written similarity with this chosen user. A set of \"similar\" users for one particular user can be used to recommend to the user interesting for him people. The similarity score also has a \"transitive property\": if $user_1$ is \"similar\" to $user_2$ and $user_2$ is similar to $user_3$ then inner properties of our model guarantees that $user_1$ and $user_3$ are pretty \"similar\" too. In this way, this score can be used to cluster a set of users into sets of \"similar\" users. It could be used in some recommendation algorithms or tune already existing algorithms to consider a cluster's peculiarities. Also, we can extend our model and calculate feature vectors for subreddits. In that way, we can find similar to the user's subreddits and recommend them to him.","PeriodicalId":73514,"journal":{"name":"IS&T International Symposium on Electronic Imaging","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-01-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135694712","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A qualitative study of LiDAR technologies and their application areas 激光雷达技术及其应用领域的定性研究
Pub Date : 2023-01-16 DOI: 10.2352/ei.2023.35.3.mobmu-368
Daniel Jaster, Reiner Creutzburg, Eberhard Hasche
In this work, the most relevant 3D LiDAR technologies and their applications in 2022 were investigated. For this purpose, applications of LiDAR systems were classified into the typical application areas "3D modeling", "smart city", "robotics", "smart automotive" and "consumer goods". The investigation has shown that neither "mechanical" LiDAR technologies, nor so-called solid-state LiDAR technologies, nor "hybrid" LiDAR technologies can be evaluated as optimal for the typical application areas. In none of the application areas could all of the elaborated requirements be met. However, the "hybrid" LiDAR technologies such as sequential MEMS LiDAR technology and sequential flash LiDAR technology proved to be among the most suitable for most typical application areas. However, other technologies also tended to be suitable for individual typical application areas. Finally, it was found that several of the LiDAR technologies investigated are currently equally suitable for some typical application areas. To evaluate the suitability, concrete LiDAR systems - of different technologies and properties - were compared with the specific requirements of exemplary applications of an application area. The results of the investigation provide an orientation as to which LiDAR technology is promising for which application area.
本文对2022年最相关的3D激光雷达技术及其应用进行了研究。为此,将激光雷达系统的应用分为“3D建模”、“智慧城市”、“机器人”、“智能汽车”和“消费品”等典型应用领域。调查表明,无论是“机械”激光雷达技术,还是所谓的固态激光雷达技术,还是“混合”激光雷达技术,都不能被评估为典型应用领域的最佳技术。在任何一个应用领域中,都不可能满足所有详细阐述的需求。然而,“混合”激光雷达技术,如顺序MEMS激光雷达技术和顺序闪存激光雷达技术被证明是最适合大多数典型应用领域的技术。然而,其他技术也倾向于适合个别典型的应用领域。最后,研究发现,目前所研究的几种激光雷达技术同样适用于一些典型的应用领域。为了评估其适用性,将不同技术和特性的具体激光雷达系统与应用领域示例应用的具体要求进行了比较。研究结果为激光雷达技术在哪些应用领域有前景提供了方向。
{"title":"A qualitative study of LiDAR technologies and their application areas","authors":"Daniel Jaster, Reiner Creutzburg, Eberhard Hasche","doi":"10.2352/ei.2023.35.3.mobmu-368","DOIUrl":"https://doi.org/10.2352/ei.2023.35.3.mobmu-368","url":null,"abstract":"In this work, the most relevant 3D LiDAR technologies and their applications in 2022 were investigated. For this purpose, applications of LiDAR systems were classified into the typical application areas \"3D modeling\", \"smart city\", \"robotics\", \"smart automotive\" and \"consumer goods\". The investigation has shown that neither \"mechanical\" LiDAR technologies, nor so-called solid-state LiDAR technologies, nor \"hybrid\" LiDAR technologies can be evaluated as optimal for the typical application areas. In none of the application areas could all of the elaborated requirements be met. However, the \"hybrid\" LiDAR technologies such as sequential MEMS LiDAR technology and sequential flash LiDAR technology proved to be among the most suitable for most typical application areas. However, other technologies also tended to be suitable for individual typical application areas. Finally, it was found that several of the LiDAR technologies investigated are currently equally suitable for some typical application areas. To evaluate the suitability, concrete LiDAR systems - of different technologies and properties - were compared with the specific requirements of exemplary applications of an application area. The results of the investigation provide an orientation as to which LiDAR technology is promising for which application area.","PeriodicalId":73514,"journal":{"name":"IS&T International Symposium on Electronic Imaging","volume":"208 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-01-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135694715","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Image Processing: Algorithms and Systems XXI Conference Overview and Papers Program 图像处理:算法和系统第21届会议综述和论文计划
Pub Date : 2023-01-16 DOI: 10.2352/ei.2023.35.9.ipas-a09
Abstract Image Processing: Algorithms and Systems continues the tradition of the past conference, Nonlinear Image Processing and Pattern Analysis, in exploring new image processing algorithms. Specifically, the conference aims at highlighting the importance of the interaction between transform-, model-, and learning-based approaches for creating effective algorithms and building modern imaging systems for new and emerging applications. It also reverberates the growing call for integration of the theoretical research on image processing algorithms with the more applied research on image processing systems.
图像处理:算法与系统会议延续了以往的传统,非线性图像处理与模式分析,在探索新的图像处理算法。具体来说,会议旨在强调基于转换、模型和学习的方法之间的相互作用的重要性,这些方法可以为新的和新兴的应用创建有效的算法和构建现代成像系统。这也反映了将图像处理算法的理论研究与图像处理系统的应用研究相结合的呼声。
{"title":"Image Processing: Algorithms and Systems XXI Conference Overview and Papers Program","authors":"","doi":"10.2352/ei.2023.35.9.ipas-a09","DOIUrl":"https://doi.org/10.2352/ei.2023.35.9.ipas-a09","url":null,"abstract":"Abstract Image Processing: Algorithms and Systems continues the tradition of the past conference, Nonlinear Image Processing and Pattern Analysis, in exploring new image processing algorithms. Specifically, the conference aims at highlighting the importance of the interaction between transform-, model-, and learning-based approaches for creating effective algorithms and building modern imaging systems for new and emerging applications. It also reverberates the growing call for integration of the theoretical research on image processing algorithms with the more applied research on image processing systems.","PeriodicalId":73514,"journal":{"name":"IS&T International Symposium on Electronic Imaging","volume":"10 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-01-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135695208","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Intelligent Robotics and Industrial Applications using Computer Vision 2023 Conference Overview and Papers Program 使用计算机视觉的智能机器人和工业应用2023会议综述和论文计划
Pub Date : 2023-01-16 DOI: 10.2352/ei.2023.35.5.iriacv-a05
Abstract This conference brings together real-world practitioners and researchers in intelligent robots and computer vision to share recent applications and developments. Topics of interest include the integration of imaging sensors supporting hardware, computers, and algorithms for intelligent robots, manufacturing inspection, characterization, and/or control. The decreased cost of computational power and vision sensors has motivated the rapid proliferation of machine vision technology in a variety of industries, including aluminum, automotive, forest products, textiles, glass, steel, metal casting, aircraft, chemicals, food, fishing, agriculture, archaeological products, medical products, artistic products, etc. Other industries, such as semiconductor and electronics manufacturing, have been employing machine vision technology for several decades. Machine vision supporting handling robots is another main topic. With respect to intelligent robotics another approach is sensor fusion – combining multi-modal sensors in audio, location, image and video data for signal processing, machine learning and computer vision, and additionally other 3D capturing devices. There is a need for accurate, fast, and robust detection of objects and their position in space. Their surface, background, and illumination are uncontrolled, and in most cases the objects of interest are within a bulk of many others. For both new and existing industrial users of machine vision, there are numerous innovative methods to improve productivity, quality, and compliance with product standards. There are several broad problem areas that have received significant attention in recent years. For example, some industries are collecting enormous amounts of image data from product monitoring systems. New and efficient methods are required to extract insight and to perform process diagnostics based on this historical record. Regarding the physical scale of the measurements, microscopy techniques are nearing resolution limits in fields such as semiconductors, biology, and other nano-scale technologies. Techniques such as resolution enhancement, model-based methods, and statistical imaging may provide the means to extend these systems beyond current capabilities. Furthermore, obtaining real-time and robust measurements in-line or at-line in harsh industrial environments is a challenge for machine vision researchers, especially when the manufacturer cannot make significant changes to their facility or process.
本次会议汇集了智能机器人和计算机视觉领域的实践者和研究人员,分享了最新的应用和发展。感兴趣的主题包括集成成像传感器支持硬件,计算机和算法的智能机器人,制造检查,表征,和/或控制。计算能力和视觉传感器成本的降低推动了机器视觉技术在各种行业的快速扩散,包括铝、汽车、林产品、纺织、玻璃、钢铁、金属铸造、飞机、化工、食品、渔业、农业、考古产品、医疗产品、艺术产品等。其他行业,如半导体和电子制造业,几十年来一直在使用机器视觉技术。支持搬运机器人的机器视觉是另一个主要主题。对于智能机器人,另一种方法是传感器融合——将音频、位置、图像和视频数据中的多模态传感器结合起来,用于信号处理、机器学习和计算机视觉,以及其他3D捕获设备。需要对物体及其在空间中的位置进行准确、快速和可靠的检测。它们的表面、背景和照明都是不受控制的,在大多数情况下,感兴趣的物体都在许多其他物体的中间。对于机器视觉的新用户和现有的工业用户来说,有许多创新的方法来提高生产力、质量和产品标准的合规性。近年来,有几个广泛的问题领域受到了极大的关注。例如,一些行业正在从产品监控系统中收集大量的图像数据。需要新的和有效的方法来提取洞察力,并基于此历史记录执行过程诊断。关于测量的物理尺度,显微镜技术在半导体、生物学和其他纳米尺度技术等领域的分辨率已经接近极限。诸如分辨率增强、基于模型的方法和统计成像等技术可以提供扩展这些系统的手段,使其超出当前的能力。此外,对于机器视觉研究人员来说,在恶劣的工业环境中获得实时和可靠的在线或在线测量是一个挑战,特别是当制造商无法对其设施或工艺进行重大更改时。
{"title":"Intelligent Robotics and Industrial Applications using Computer Vision 2023 Conference Overview and Papers Program","authors":"","doi":"10.2352/ei.2023.35.5.iriacv-a05","DOIUrl":"https://doi.org/10.2352/ei.2023.35.5.iriacv-a05","url":null,"abstract":"Abstract This conference brings together real-world practitioners and researchers in intelligent robots and computer vision to share recent applications and developments. Topics of interest include the integration of imaging sensors supporting hardware, computers, and algorithms for intelligent robots, manufacturing inspection, characterization, and/or control. The decreased cost of computational power and vision sensors has motivated the rapid proliferation of machine vision technology in a variety of industries, including aluminum, automotive, forest products, textiles, glass, steel, metal casting, aircraft, chemicals, food, fishing, agriculture, archaeological products, medical products, artistic products, etc. Other industries, such as semiconductor and electronics manufacturing, have been employing machine vision technology for several decades. Machine vision supporting handling robots is another main topic. With respect to intelligent robotics another approach is sensor fusion – combining multi-modal sensors in audio, location, image and video data for signal processing, machine learning and computer vision, and additionally other 3D capturing devices. There is a need for accurate, fast, and robust detection of objects and their position in space. Their surface, background, and illumination are uncontrolled, and in most cases the objects of interest are within a bulk of many others. For both new and existing industrial users of machine vision, there are numerous innovative methods to improve productivity, quality, and compliance with product standards. There are several broad problem areas that have received significant attention in recent years. For example, some industries are collecting enormous amounts of image data from product monitoring systems. New and efficient methods are required to extract insight and to perform process diagnostics based on this historical record. Regarding the physical scale of the measurements, microscopy techniques are nearing resolution limits in fields such as semiconductors, biology, and other nano-scale technologies. Techniques such as resolution enhancement, model-based methods, and statistical imaging may provide the means to extend these systems beyond current capabilities. Furthermore, obtaining real-time and robust measurements in-line or at-line in harsh industrial environments is a challenge for machine vision researchers, especially when the manufacturer cannot make significant changes to their facility or process.","PeriodicalId":73514,"journal":{"name":"IS&T International Symposium on Electronic Imaging","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-01-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135695223","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Immersive security personnel training module for active shooter events 身临其境的安全人员培训模块为主动射击事件
Pub Date : 2023-01-16 DOI: 10.2352/ei.2023.35.12.ervr-217
Sharad Sharma, JeeWoong Park, Brendan Tran Morris
There is a need to prepare for emergencies such as active shooter events. Emergency response training drills and exercises are necessary to train for such events as we are unable to predict when emergencies do occur. There has been progress in understanding human behavior, unpredictability, human motion synthesis, crowd dynamics, and their relationships with active shooter events, but challenges remain. This paper presents an immersive security personnel training module for active shooter events in an indoor building. We have created an experimental platform for conducting active shooter drills for training that gives a fully immersive feel of the situation and allow one to perform virtual evacuation drills. The security personnel training module also incorporates four sub-modules namely 1) Situational assessment module, 2) Individual officer intervention module, 3) Team Response Module, and 4) Rescue Task Force module. We have developed an immersive virtual reality training module for active shooter events using an Oculus for course of action, visualization, and situational awareness for active shooter events as shown in Fig.1. The immersive security personnel training module aims to get information about the emergency situation inside the building. The dispatched officer will verify the active shooter situation in the building. The security personnel should find a safe zone in the building and secure the people in that area. The security personnel should also find the number and location of persons in possible jeopardy. Upon completion of the initial assessment, the first security personnel shall advise communications and request resources as deemed necessary. This will allow determining whether to take immediate action alone or with another officer or wait until additional resources are available. After successfully gathering the information, the personnel needs to update the info to their officer through a communication device.
有必要为紧急情况做好准备,比如活跃的枪手事件。紧急反应训练演习和演习是必要的,因为我们无法预测紧急情况何时发生。在理解人类行为、不可预测性、人类运动合成、人群动力学及其与主动射击事件的关系方面取得了进展,但挑战依然存在。提出了一种针对室内建筑枪击案的沉浸式安全人员培训模块。我们已经创建了一个实验平台,用于进行主动射击演习的训练,让人完全身临其境的感觉,并允许一个人进行虚拟疏散演习。安保人员培训模块还包括四个子模块,即1)态势评估模块、2)单兵干预模块、3)团队应对模块和4)救援工作队模块。我们已经开发了一个身临其境的虚拟现实训练模块为主动射击事件使用Oculus的行动过程,可视化和态势感知主动射击事件如图1所示。沉浸式安全人员培训模块旨在获取有关建筑物内紧急情况的信息。派去的警官会核实大楼里的枪手情况。保安人员应该在大楼里找到一个安全区域,并保护该区域的人员。安全人员还应查明可能处于危险中的人员的数量和位置。初步评估完成后,第一批保安人员应提供必要的通信建议和资源请求。这将决定是单独行动,还是与另一名警官一起行动,还是等到有更多资源可用时再行动。在成功收集信息后,工作人员需要通过通信设备将信息更新给他们的主管。
{"title":"Immersive security personnel training module for active shooter events","authors":"Sharad Sharma, JeeWoong Park, Brendan Tran Morris","doi":"10.2352/ei.2023.35.12.ervr-217","DOIUrl":"https://doi.org/10.2352/ei.2023.35.12.ervr-217","url":null,"abstract":"There is a need to prepare for emergencies such as active shooter events. Emergency response training drills and exercises are necessary to train for such events as we are unable to predict when emergencies do occur. There has been progress in understanding human behavior, unpredictability, human motion synthesis, crowd dynamics, and their relationships with active shooter events, but challenges remain. This paper presents an immersive security personnel training module for active shooter events in an indoor building. We have created an experimental platform for conducting active shooter drills for training that gives a fully immersive feel of the situation and allow one to perform virtual evacuation drills. The security personnel training module also incorporates four sub-modules namely 1) Situational assessment module, 2) Individual officer intervention module, 3) Team Response Module, and 4) Rescue Task Force module. We have developed an immersive virtual reality training module for active shooter events using an Oculus for course of action, visualization, and situational awareness for active shooter events as shown in Fig.1. The immersive security personnel training module aims to get information about the emergency situation inside the building. The dispatched officer will verify the active shooter situation in the building. The security personnel should find a safe zone in the building and secure the people in that area. The security personnel should also find the number and location of persons in possible jeopardy. Upon completion of the initial assessment, the first security personnel shall advise communications and request resources as deemed necessary. This will allow determining whether to take immediate action alone or with another officer or wait until additional resources are available. After successfully gathering the information, the personnel needs to update the info to their officer through a communication device.","PeriodicalId":73514,"journal":{"name":"IS&T International Symposium on Electronic Imaging","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-01-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135694149","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Engineering Reality of Virtual Reality 2023 Conference Overview and Papers Program 2023虚拟现实工程现实会议综述和论文计划
Pub Date : 2023-01-16 DOI: 10.2352/ei.2023.35.12.ervr-a12
Abstract Virtual and augmented reality systems are evolving. In addition to research, the trend toward content building continues and practitioners find that technologies and disciplines must be tailored and integrated for specific visualization and interactive applications. This conference serves as a forum where advances and practical advice toward both creative activity and scientific investigation are presented and discussed. Research results can be presented and applications can be demonstrated.
虚拟现实和增强现实系统正在不断发展。除了研究之外,内容构建的趋势还在继续,从业者发现技术和学科必须针对特定的可视化和交互式应用进行定制和集成。这次会议是一个论坛,在这里,对创造性活动和科学研究的进展和实际建议进行了介绍和讨论。研究成果可以展示和应用可以演示。
{"title":"Engineering Reality of Virtual Reality 2023 Conference Overview and Papers Program","authors":"","doi":"10.2352/ei.2023.35.12.ervr-a12","DOIUrl":"https://doi.org/10.2352/ei.2023.35.12.ervr-a12","url":null,"abstract":"Abstract Virtual and augmented reality systems are evolving. In addition to research, the trend toward content building continues and practitioners find that technologies and disciplines must be tailored and integrated for specific visualization and interactive applications. This conference serves as a forum where advances and practical advice toward both creative activity and scientific investigation are presented and discussed. Research results can be presented and applications can be demonstrated.","PeriodicalId":73514,"journal":{"name":"IS&T International Symposium on Electronic Imaging","volume":"68 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-01-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135695217","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Improvement of vehicles accident detection using object tracking with U-Net 基于U-Net的目标跟踪改进车辆事故检测
Pub Date : 2023-01-16 DOI: 10.2352/ei.2023.35.3.mobmu-363
Kirsnaragavan Arudpiragasam, Taraka Rama Krishna Kanth Kannuri, Klaus Schwarz, Michael Hartmann, Reiner Creutzburg
Over the past decade, researchers have suggested many methods to find anomalies. However, none of the studies has applied frame reconstruction with Object Tracking (OT) to detect anomalies. Therefore, this study focuses on road accident detection using a combination of OT and U-Net associated with variants such as skip, skip residual and attention connections. The U-Net algorithm is developed for reconstructing the images using the UFC-Crime dataset. Furthermore, YOLOV4 and DeepSort are used for object detection and tracking within the frames. Finally, the Mahalanobis distance and the reconstruction error (RCE) are determined using a Kalman filter and the U-Net model.
在过去的十年里,研究人员提出了许多发现异常的方法。然而,目前还没有研究将帧重建与目标跟踪(OT)相结合来检测异常。因此,本研究的重点是使用OT和U-Net相结合的方法进行道路事故检测,并结合诸如跳过、跳过残余和注意连接等变体。U-Net算法用于使用UFC-Crime数据集重建图像。此外,YOLOV4和DeepSort用于帧内的目标检测和跟踪。最后,利用卡尔曼滤波和U-Net模型确定了马氏距离和重建误差。
{"title":"Improvement of vehicles accident detection using object tracking with U-Net","authors":"Kirsnaragavan Arudpiragasam, Taraka Rama Krishna Kanth Kannuri, Klaus Schwarz, Michael Hartmann, Reiner Creutzburg","doi":"10.2352/ei.2023.35.3.mobmu-363","DOIUrl":"https://doi.org/10.2352/ei.2023.35.3.mobmu-363","url":null,"abstract":"Over the past decade, researchers have suggested many methods to find anomalies. However, none of the studies has applied frame reconstruction with Object Tracking (OT) to detect anomalies. Therefore, this study focuses on road accident detection using a combination of OT and U-Net associated with variants such as skip, skip residual and attention connections. The U-Net algorithm is developed for reconstructing the images using the UFC-Crime dataset. Furthermore, YOLOV4 and DeepSort are used for object detection and tracking within the frames. Finally, the Mahalanobis distance and the reconstruction error (RCE) are determined using a Kalman filter and the U-Net model.","PeriodicalId":73514,"journal":{"name":"IS&T International Symposium on Electronic Imaging","volume":"26 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-01-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135695029","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IS&T International Symposium on Electronic Imaging
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1