首页 > 最新文献

2023 IEEE International Conference on Smart Computing (SMARTCOMP)最新文献

英文 中文
On Learning Data-Driven Models For In-Flight Drone Battery Discharge Estimation From Real Data 基于实际数据的无人机电池放电估计数据驱动模型学习研究
Pub Date : 2023-06-01 DOI: 10.1109/SMARTCOMP58114.2023.00038
Austin Coursey, Marcos Quiñones-Grueiro, G. Biswas
Accurate estimation of the battery state of charge (SOC) for unmanned aerial vehicles (UAV) in-flight monitoring is essential for the safety and survivability of the system. Successful physics-based models of the battery have been developed in the past, however, these models do not take into account the effects of mission profile and environmental conditions during flight on the battery power consumption. Recently, data-driven methods have become popular given their ease of use and scalability. Yet, most benchmarking experiments have been conducted on simulated battery datasets. In this work, we compare different data-driven models for battery SOC estimation of a hexacopter UAV system using real flight data. We analyze the importance of a number of flight variables under different environmental conditions to determine the factors that affect battery SOC over the course of the flight. Our experiments demonstrate that additional flight variables are necessary to create an accurate SOC estimation model through data-driven methods.
准确估计无人机(UAV)飞行监测中电池的充电状态(SOC)对系统的安全性和生存性至关重要。过去已经开发了成功的基于物理的电池模型,然而,这些模型没有考虑任务剖面和飞行过程中环境条件对电池功耗的影响。最近,数据驱动的方法由于其易用性和可伸缩性而变得流行起来。然而,大多数基准测试实验都是在模拟电池数据集上进行的。在这项工作中,我们比较了使用真实飞行数据进行六旋翼无人机系统电池荷电状态估计的不同数据驱动模型。我们分析了不同环境条件下一些飞行变量的重要性,以确定在飞行过程中影响电池SOC的因素。我们的实验表明,通过数据驱动的方法创建准确的SOC估计模型需要额外的飞行变量。
{"title":"On Learning Data-Driven Models For In-Flight Drone Battery Discharge Estimation From Real Data","authors":"Austin Coursey, Marcos Quiñones-Grueiro, G. Biswas","doi":"10.1109/SMARTCOMP58114.2023.00038","DOIUrl":"https://doi.org/10.1109/SMARTCOMP58114.2023.00038","url":null,"abstract":"Accurate estimation of the battery state of charge (SOC) for unmanned aerial vehicles (UAV) in-flight monitoring is essential for the safety and survivability of the system. Successful physics-based models of the battery have been developed in the past, however, these models do not take into account the effects of mission profile and environmental conditions during flight on the battery power consumption. Recently, data-driven methods have become popular given their ease of use and scalability. Yet, most benchmarking experiments have been conducted on simulated battery datasets. In this work, we compare different data-driven models for battery SOC estimation of a hexacopter UAV system using real flight data. We analyze the importance of a number of flight variables under different environmental conditions to determine the factors that affect battery SOC over the course of the flight. Our experiments demonstrate that additional flight variables are necessary to create an accurate SOC estimation model through data-driven methods.","PeriodicalId":163556,"journal":{"name":"2023 IEEE International Conference on Smart Computing (SMARTCOMP)","volume":"280 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122092828","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
FactionFormer: Context-Driven Collaborative Vision Transformer Models for Edge Intelligence FactionFormer:边缘智能的上下文驱动协同视觉转换模型
Pub Date : 2023-06-01 DOI: 10.1109/SMARTCOMP58114.2023.00084
Sumaiya Tabassum Nimi, Md. Adnan Arefeen, M. Y. S. Uddin, Biplob K. Debnath, S. Chakradhar
Edge Intelligence has received attention in the recent times for its potential towards improving responsiveness, reducing the cost of data transmission, enhancing security and privacy, and enabling autonomous decisions by edge devices. However, edge devices lack the power and compute resources necessary to execute most Al models. In this paper, we present FactionFormer, a novel method to deploy resource-intensive deep-learning models, such as vision transformers (ViT), on resource-constrained edge devices. Our method is based on a key observation: edge devices are often deployed in settings where they encounter only a subset of the classes that the resource-intensive Al model is trained to classify, and this subset changes across deployments. Therefore, we automatically identify this subset as a faction, devise on-the fly a bespoke resource-efficient ViT called a modelette for the faction, and set up an efficient processing pipeline consisting of a modelette on the device, a wireless network such as 5G, and the resource-intensive ViT model on an edge server, all of which work collaboratively to do the inference. For several ViT models pre-trained on benchmark datasets, FactionFormer’s modelettes are up to 4× smaller than the corresponding baseline models in terms of the number of parameters, and they can infer up to 2.5× faster than the baseline setup where every input is processed by the resource-intensive ViT on the edge server. Our work is the first of its kind to propose a device-edge collaborative inference framework where bespoke deep learning models for the device are automatically devised on-the-fly for most frequently encountered subset of classes.
近年来,边缘智能因其在提高响应能力、降低数据传输成本、增强安全性和隐私性以及实现边缘设备自主决策方面的潜力而受到关注。然而,边缘设备缺乏执行大多数人工智能模型所需的能力和计算资源。在本文中,我们提出了FactionFormer,这是一种在资源受限的边缘设备上部署资源密集型深度学习模型(如视觉变压器(ViT))的新方法。我们的方法基于一个关键的观察:边缘设备通常部署在它们只遇到资源密集型人工智能模型训练分类的类的子集的设置中,并且这个子集在部署中会发生变化。因此,我们自动将该子集识别为一个派别,动态地为该派别设计一个定制的资源高效的ViT,称为modelette,并建立一个高效的处理管道,该管道由设备上的modelette、无线网络(如5G)和边缘服务器上的资源密集型ViT模型组成,所有这些都协同工作以进行推理。对于在基准数据集上预先训练的几个ViT模型,就参数数量而言,FactionFormer的模型比相应的基线模型小4倍,并且它们的推断速度比基线设置快2.5倍,其中每个输入都由边缘服务器上的资源密集型ViT处理。我们的工作首次提出了一种设备边缘协作推理框架,在该框架中,为设备定制的深度学习模型会自动为最常遇到的类子集动态设计。
{"title":"FactionFormer: Context-Driven Collaborative Vision Transformer Models for Edge Intelligence","authors":"Sumaiya Tabassum Nimi, Md. Adnan Arefeen, M. Y. S. Uddin, Biplob K. Debnath, S. Chakradhar","doi":"10.1109/SMARTCOMP58114.2023.00084","DOIUrl":"https://doi.org/10.1109/SMARTCOMP58114.2023.00084","url":null,"abstract":"Edge Intelligence has received attention in the recent times for its potential towards improving responsiveness, reducing the cost of data transmission, enhancing security and privacy, and enabling autonomous decisions by edge devices. However, edge devices lack the power and compute resources necessary to execute most Al models. In this paper, we present FactionFormer, a novel method to deploy resource-intensive deep-learning models, such as vision transformers (ViT), on resource-constrained edge devices. Our method is based on a key observation: edge devices are often deployed in settings where they encounter only a subset of the classes that the resource-intensive Al model is trained to classify, and this subset changes across deployments. Therefore, we automatically identify this subset as a faction, devise on-the fly a bespoke resource-efficient ViT called a modelette for the faction, and set up an efficient processing pipeline consisting of a modelette on the device, a wireless network such as 5G, and the resource-intensive ViT model on an edge server, all of which work collaboratively to do the inference. For several ViT models pre-trained on benchmark datasets, FactionFormer’s modelettes are up to 4× smaller than the corresponding baseline models in terms of the number of parameters, and they can infer up to 2.5× faster than the baseline setup where every input is processed by the resource-intensive ViT on the edge server. Our work is the first of its kind to propose a device-edge collaborative inference framework where bespoke deep learning models for the device are automatically devised on-the-fly for most frequently encountered subset of classes.","PeriodicalId":163556,"journal":{"name":"2023 IEEE International Conference on Smart Computing (SMARTCOMP)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124326734","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
BeautyNet: A Makeup Activity Recognition Framework using Wrist-worn Sensor BeautyNet:一个使用腕带传感器的化妆活动识别框架
Pub Date : 2023-06-01 DOI: 10.1109/SMARTCOMP58114.2023.00072
Fatimah Albargi, Naima Khan, Indrajeet Ghosh, Ahana Roy
The significance of enhancing facial features has grown increasingly in popularity among all groups of people bringing a surge in makeup activities. The makeup market is one of the most profitable and founding sectors in the fashion industry which involves product retailing and demands user training. Makeup activities imply exceptionally delicate hand movements and require much training and practice for perfection. However, the only available choices in learning makeup activities are hands-on workshops by professional instructors or, at most, video-based visual instructions. None of these exhibits immense benefits to beginners, or visually impaired people. One can consistently watch and listen to the best of their abilities, but to precisely practice, perform, and reach makeup satisfaction, recognition from an IoT (Internet-of-Things) device with results and feedback would be the utmost support. In this work, we propose a makeup activity recognition framework, BeautyNet which detects different makeup activities from wrist-worn sensor data collected from ten participants of different age groups in two experimental setups. Our framework employs a LSTM-autoencoder based classifier to extract features from the sensor data and classifies five makeup activities (i.e., applying cream, lipsticks, blusher, eyeshadow, and mascara) in controlled and uncontrolled environment. Empirical results indicate that BeautyNet achieves 95% and 93% accuracy for makeup activity detection in controlled and uncontrolled settings, respectively. In addition, we evaluate BeautyNet with various traditional machine learning algorithms using our in-house dataset and noted an increase in accuracy by ≈ 4-7%.
增强面部特征的重要性在所有人群中越来越受欢迎,带来了化妆活动的激增。化妆品市场是时尚行业中最赚钱、最基础的行业之一,涉及产品零售,需要用户培训。化妆活动需要非常精细的手部动作,需要大量的训练和练习才能达到完美。然而,学习化妆活动的唯一选择是专业教师的动手讲习班,或者最多是基于视频的视觉指导。这些对初学者或视障人士都没有多大好处。人们可以不断地观看和倾听他们的最佳能力,但要精确地练习,表演并达到化妆满意度,来自物联网(IoT)设备的识别结果和反馈将是最大的支持。在这项工作中,我们提出了一个化妆活动识别框架,BeautyNet,它从两个实验设置中收集的来自不同年龄组的10名参与者的腕戴传感器数据中检测不同的化妆活动。我们的框架采用基于lstm自编码器的分类器从传感器数据中提取特征,并在受控和非受控环境中对五种化妆活动(即涂面霜、口红、腮红、眼影和睫毛膏)进行分类。实证结果表明,BeautyNet在受控和非受控设置下的化妆活动检测准确率分别达到95%和93%。此外,我们使用内部数据集使用各种传统机器学习算法对BeautyNet进行了评估,并注意到准确率提高了约4-7%。
{"title":"BeautyNet: A Makeup Activity Recognition Framework using Wrist-worn Sensor","authors":"Fatimah Albargi, Naima Khan, Indrajeet Ghosh, Ahana Roy","doi":"10.1109/SMARTCOMP58114.2023.00072","DOIUrl":"https://doi.org/10.1109/SMARTCOMP58114.2023.00072","url":null,"abstract":"The significance of enhancing facial features has grown increasingly in popularity among all groups of people bringing a surge in makeup activities. The makeup market is one of the most profitable and founding sectors in the fashion industry which involves product retailing and demands user training. Makeup activities imply exceptionally delicate hand movements and require much training and practice for perfection. However, the only available choices in learning makeup activities are hands-on workshops by professional instructors or, at most, video-based visual instructions. None of these exhibits immense benefits to beginners, or visually impaired people. One can consistently watch and listen to the best of their abilities, but to precisely practice, perform, and reach makeup satisfaction, recognition from an IoT (Internet-of-Things) device with results and feedback would be the utmost support. In this work, we propose a makeup activity recognition framework, BeautyNet which detects different makeup activities from wrist-worn sensor data collected from ten participants of different age groups in two experimental setups. Our framework employs a LSTM-autoencoder based classifier to extract features from the sensor data and classifies five makeup activities (i.e., applying cream, lipsticks, blusher, eyeshadow, and mascara) in controlled and uncontrolled environment. Empirical results indicate that BeautyNet achieves 95% and 93% accuracy for makeup activity detection in controlled and uncontrolled settings, respectively. In addition, we evaluate BeautyNet with various traditional machine learning algorithms using our in-house dataset and noted an increase in accuracy by ≈ 4-7%.","PeriodicalId":163556,"journal":{"name":"2023 IEEE International Conference on Smart Computing (SMARTCOMP)","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132117443","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Cyber Framework for Steering and Measurements Collection Over Instrument-Computing Ecosystems 仪器计算生态系统转向和测量收集的网络框架
Pub Date : 2023-06-01 DOI: 10.1109/SMARTCOMP58114.2023.00046
Anees Al-Najjar, Nageswara S. V. Rao, R. Sankaran, H. Zandi, Debangshu Mukherjee, M. Ziatdinov, Craig Bridges
We propose a framework to develop cyber solutions to support remote steering of science instruments and measurements collection over instrument-computing ecosystems. It is based on provisioning separate data and control connections at the network level, and developing software modules consisting of Python wrappers for instrument commands and Pyro server-client codes that make them available across the ecosystem network. We demonstrate automated measurement transfers and remote steering operations in a microscopy use case for materials research over an ecosystem of Nion microscopes and computing platforms connected over site networks. The proposed framework is currently under further refinement and being adopted to science workflows with automated remote experiments steering for autonomous chemistry laboratories and smart energy grid simulations.
我们提出了一个框架来开发网络解决方案,以支持在仪器计算生态系统上远程控制科学仪器和测量收集。它基于在网络级别提供单独的数据和控制连接,以及开发由仪器命令的Python包装器和Pyro服务器-客户端代码组成的软件模块,使它们在整个生态系统网络中可用。我们在一个显微镜用例中演示了自动测量传输和远程转向操作,该用例用于材料研究,该用例涉及通过现场网络连接的Nion显微镜和计算平台的生态系统。所提出的框架目前正在进一步完善,并被采用到科学工作流程中,为自主化学实验室和智能能源网格模拟提供自动化远程实验指导。
{"title":"Cyber Framework for Steering and Measurements Collection Over Instrument-Computing Ecosystems","authors":"Anees Al-Najjar, Nageswara S. V. Rao, R. Sankaran, H. Zandi, Debangshu Mukherjee, M. Ziatdinov, Craig Bridges","doi":"10.1109/SMARTCOMP58114.2023.00046","DOIUrl":"https://doi.org/10.1109/SMARTCOMP58114.2023.00046","url":null,"abstract":"We propose a framework to develop cyber solutions to support remote steering of science instruments and measurements collection over instrument-computing ecosystems. It is based on provisioning separate data and control connections at the network level, and developing software modules consisting of Python wrappers for instrument commands and Pyro server-client codes that make them available across the ecosystem network. We demonstrate automated measurement transfers and remote steering operations in a microscopy use case for materials research over an ecosystem of Nion microscopes and computing platforms connected over site networks. The proposed framework is currently under further refinement and being adopted to science workflows with automated remote experiments steering for autonomous chemistry laboratories and smart energy grid simulations.","PeriodicalId":163556,"journal":{"name":"2023 IEEE International Conference on Smart Computing (SMARTCOMP)","volume":"13 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130749601","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
NextGenGW - a Software Framework Based on MQTT and Semantic Definition Format NextGenGW——基于MQTT和语义定义格式的软件框架
Pub Date : 2023-06-01 DOI: 10.1109/SMARTCOMP58114.2023.00035
Carlos Resende, Waldir Moreira, Luís Almeida
To access all the potential value present in IoT, the IoT devices need to be interoperable. Some works in the literature target this issue, but it is not yet entirely solved, mainly because the proposed solutions are not standard-based at the semantic level. This paper presents the detailed implementation of our standard-based software framework targeting IoT interoperability, named NextGenGW. With NextGenGW, we propose the first integration of IETF SDF with the MQTT protocol. We define an evaluation baseline for validating IoT gateway performance while focusing on interoperability. Our evaluation results show the NextGenGW suitability for deployment in devices with reduced resources and for use cases that require high scalability both in terms of connected IoT end nodes and the number of requests per time interval.
为了获得物联网中存在的所有潜在价值,物联网设备需要具有互操作性。文献中的一些工作针对这个问题,但尚未完全解决,主要是因为提出的解决方案在语义层面上不是基于标准的。本文介绍了我们针对物联网互操作性的基于标准的软件框架NextGenGW的详细实现。通过NextGenGW,我们提出了IETF SDF与MQTT协议的首次集成。我们定义了一个评估基线,用于验证物联网网关的性能,同时关注互操作性。我们的评估结果显示,NextGenGW适合部署在资源较少的设备中,以及在连接的物联网终端节点和每个时间间隔的请求数量方面需要高可扩展性的用例。
{"title":"NextGenGW - a Software Framework Based on MQTT and Semantic Definition Format","authors":"Carlos Resende, Waldir Moreira, Luís Almeida","doi":"10.1109/SMARTCOMP58114.2023.00035","DOIUrl":"https://doi.org/10.1109/SMARTCOMP58114.2023.00035","url":null,"abstract":"To access all the potential value present in IoT, the IoT devices need to be interoperable. Some works in the literature target this issue, but it is not yet entirely solved, mainly because the proposed solutions are not standard-based at the semantic level. This paper presents the detailed implementation of our standard-based software framework targeting IoT interoperability, named NextGenGW. With NextGenGW, we propose the first integration of IETF SDF with the MQTT protocol. We define an evaluation baseline for validating IoT gateway performance while focusing on interoperability. Our evaluation results show the NextGenGW suitability for deployment in devices with reduced resources and for use cases that require high scalability both in terms of connected IoT end nodes and the number of requests per time interval.","PeriodicalId":163556,"journal":{"name":"2023 IEEE International Conference on Smart Computing (SMARTCOMP)","volume":"439 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123505824","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Multi-modal AI Systems for Human and Animal Pose Estimation in Challenging Conditions 挑战性条件下人类和动物姿态估计的多模态人工智能系统
Pub Date : 2023-06-01 DOI: 10.1109/SMARTCOMP58114.2023.00060
Qianyi Deng
This paper explores the development of multi-modal AI systems for pose estimation in challenging conditions for both humans and animals. Existing single-modality approaches struggle in challenging scenarios such as emergency response and wildlife observation due to factors like smoke, low light, obstacles, and long-distance observations. To address these challenges, this research proposes integrating multiple sensor modalities and leveraging the strengths of different sensors to enhance accuracy and robustness in pose estimation.
本文探讨了在人类和动物具有挑战性的条件下进行姿态估计的多模态人工智能系统的发展。由于烟雾、低光、障碍物和远距离观测等因素,现有的单模态方法在应急响应和野生动物观测等具有挑战性的情况下难以实现。为了解决这些挑战,本研究提出了集成多种传感器模式并利用不同传感器的优势来提高姿态估计的准确性和鲁棒性。
{"title":"Multi-modal AI Systems for Human and Animal Pose Estimation in Challenging Conditions","authors":"Qianyi Deng","doi":"10.1109/SMARTCOMP58114.2023.00060","DOIUrl":"https://doi.org/10.1109/SMARTCOMP58114.2023.00060","url":null,"abstract":"This paper explores the development of multi-modal AI systems for pose estimation in challenging conditions for both humans and animals. Existing single-modality approaches struggle in challenging scenarios such as emergency response and wildlife observation due to factors like smoke, low light, obstacles, and long-distance observations. To address these challenges, this research proposes integrating multiple sensor modalities and leveraging the strengths of different sensors to enhance accuracy and robustness in pose estimation.","PeriodicalId":163556,"journal":{"name":"2023 IEEE International Conference on Smart Computing (SMARTCOMP)","volume":"12 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116796725","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
BITS 2023 Welcome Message from General Chairs and TPC Chairs 总主席和TPC主席欢迎辞
Pub Date : 2023-06-01 DOI: 10.1109/smartcomp58114.2023.00010
{"title":"BITS 2023 Welcome Message from General Chairs and TPC Chairs","authors":"","doi":"10.1109/smartcomp58114.2023.00010","DOIUrl":"https://doi.org/10.1109/smartcomp58114.2023.00010","url":null,"abstract":"","PeriodicalId":163556,"journal":{"name":"2023 IEEE International Conference on Smart Computing (SMARTCOMP)","volume":"83 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131036738","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Vision Transformer-based Real-Time Camouflaged Object Detection System at Edge 基于视觉变换的边缘伪装目标实时检测系统
Pub Date : 2023-06-01 DOI: 10.1109/SMARTCOMP58114.2023.00029
Rohan Putatunda, Azim Khan, A. Gangopadhyay, Jianwu Wang, Carl E. Busart, R. Erbacher
Camouflaged object detection is a challenging task in computer vision that involves identifying objects that are intentionally or unintentionally hidden in their surrounding environment. Vision Transformer mechanisms play a critical role in improving the performance of deep learning models by focusing on the most relevant features that help object detection under camouflaged conditions. In this paper, we utilized a vision transformer (VT) in two phases, a) By integrating VT with a deep learning architecture for efficient monocular depth map generation for camouflaged objects and b) By embedding VT multiclass object detection model with multimodal feature input (RGB with RGB-D) that increases the visual cues and provides more representational information to the model for performance enhancement. Additionally, we performed an ablation study to understand the role of the vision transformer in camouflaged object detection and incorporated GRAD-CAM on top of the model to visualize the performance improvement achieved by embedding the VT in the model architecture. We deployed the model on resource-constrained edge devices for real-time object detection to realistically test the performance of the trained model.
伪装物体检测是计算机视觉中的一项具有挑战性的任务,它涉及识别有意或无意隐藏在周围环境中的物体。视觉转换机制在提高深度学习模型的性能方面发挥着关键作用,它专注于在伪装条件下帮助目标检测的最相关特征。在本文中,我们分两个阶段使用视觉转换器(VT), a)将VT与深度学习架构集成,以有效地生成伪装对象的单目深度图;b)通过嵌入具有多模态特征输入(RGB与RGB- d)的VT多类目标检测模型,增加视觉线索并为模型提供更多的代表性信息,以增强性能。此外,我们进行了消融研究,以了解视觉转换器在伪装目标检测中的作用,并在模型上集成了GRAD-CAM,以可视化通过在模型架构中嵌入VT实现的性能改进。我们将模型部署在资源受限的边缘设备上进行实时目标检测,以真实地测试训练模型的性能。
{"title":"Vision Transformer-based Real-Time Camouflaged Object Detection System at Edge","authors":"Rohan Putatunda, Azim Khan, A. Gangopadhyay, Jianwu Wang, Carl E. Busart, R. Erbacher","doi":"10.1109/SMARTCOMP58114.2023.00029","DOIUrl":"https://doi.org/10.1109/SMARTCOMP58114.2023.00029","url":null,"abstract":"Camouflaged object detection is a challenging task in computer vision that involves identifying objects that are intentionally or unintentionally hidden in their surrounding environment. Vision Transformer mechanisms play a critical role in improving the performance of deep learning models by focusing on the most relevant features that help object detection under camouflaged conditions. In this paper, we utilized a vision transformer (VT) in two phases, a) By integrating VT with a deep learning architecture for efficient monocular depth map generation for camouflaged objects and b) By embedding VT multiclass object detection model with multimodal feature input (RGB with RGB-D) that increases the visual cues and provides more representational information to the model for performance enhancement. Additionally, we performed an ablation study to understand the role of the vision transformer in camouflaged object detection and incorporated GRAD-CAM on top of the model to visualize the performance improvement achieved by embedding the VT in the model architecture. We deployed the model on resource-constrained edge devices for real-time object detection to realistically test the performance of the trained model.","PeriodicalId":163556,"journal":{"name":"2023 IEEE International Conference on Smart Computing (SMARTCOMP)","volume":"85 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116101298","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Keynotes 主题演讲
Pub Date : 2023-06-01 DOI: 10.1109/smartcomp58114.2023.00008
{"title":"Keynotes","authors":"","doi":"10.1109/smartcomp58114.2023.00008","DOIUrl":"https://doi.org/10.1109/smartcomp58114.2023.00008","url":null,"abstract":"","PeriodicalId":163556,"journal":{"name":"2023 IEEE International Conference on Smart Computing (SMARTCOMP)","volume":"68 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115285304","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Calibrating Real-World City Traffic Simulation Model Using Vehicle Speed Data 使用车速数据校准真实城市交通仿真模型
Pub Date : 2023-06-01 DOI: 10.1109/SMARTCOMP58114.2023.00076
Seyedmehdi Khaleghian, H. Neema, Mina Sartipi, Toan V. Tran, Rishav Sen, Abhishek Dubey
Large-scale traffic simulations are necessary for the planning, design, and operation of city-scale transportation systems. These simulations enable novel and complex transportation technology and services such as optimization of traffic control systems, supporting on-demand transit, and redesigning regional transit systems for better energy efficiency and emissions. For a city-wide simulation model, big data from multiple sources such as Open Street Map (OSM), traffic surveys, geo-location traces, vehicular traffic data, and transit details are integrated to create a unique and accurate representation. However, in order to accurately identify the model structure and have reliable simulation results, these traffic simulation models must be thoroughly calibrated and validated against real-world data. This paper presents a novel calibration approach for a city-scale traffic simulation model based on limited real-world speed data. The simulation model runs a microscopic and mesoscopic realistic traffic simulation from Chattanooga, TN (US) for a 24-hour period and includes various transport modes such as transit buses, passenger cars, and trucks. The experiment results presented demonstrate the effectiveness of our approach for calibrating large-scale traffic networks using only real-world speed data. This paper presents our proposed calibration approach that utilizes 2160 real-world speed data points, performs sensitivity analysis of the simulation model to input parameters, and genetic algorithm for optimizing the model for calibration.
大规模的交通模拟对于城市规模交通系统的规划、设计和运行是必要的。这些模拟实现了新颖和复杂的运输技术和服务,例如优化交通控制系统,支持按需运输,以及重新设计区域运输系统以提高能源效率和排放。对于城市范围的模拟模型,来自多个来源的大数据,如开放街道地图(OSM),交通调查,地理位置痕迹,车辆交通数据和交通细节被整合,以创建一个独特而准确的表示。然而,为了准确识别模型结构并获得可靠的仿真结果,必须对这些流量仿真模型进行彻底的校准,并针对实际数据进行验证。本文提出了一种基于有限真实速度数据的城市尺度交通仿真模型标定方法。仿真模型对美国田纳西州查塔努加市24小时内微观和中观的真实交通进行仿真,包括公交、客车和卡车等多种运输方式。实验结果表明,我们的方法仅使用真实世界的速度数据来校准大规模交通网络是有效的。本文提出了一种利用2160个真实世界速度数据点的校准方法,对仿真模型进行灵敏度分析以输入参数,并使用遗传算法对模型进行优化以进行校准。
{"title":"Calibrating Real-World City Traffic Simulation Model Using Vehicle Speed Data","authors":"Seyedmehdi Khaleghian, H. Neema, Mina Sartipi, Toan V. Tran, Rishav Sen, Abhishek Dubey","doi":"10.1109/SMARTCOMP58114.2023.00076","DOIUrl":"https://doi.org/10.1109/SMARTCOMP58114.2023.00076","url":null,"abstract":"Large-scale traffic simulations are necessary for the planning, design, and operation of city-scale transportation systems. These simulations enable novel and complex transportation technology and services such as optimization of traffic control systems, supporting on-demand transit, and redesigning regional transit systems for better energy efficiency and emissions. For a city-wide simulation model, big data from multiple sources such as Open Street Map (OSM), traffic surveys, geo-location traces, vehicular traffic data, and transit details are integrated to create a unique and accurate representation. However, in order to accurately identify the model structure and have reliable simulation results, these traffic simulation models must be thoroughly calibrated and validated against real-world data. This paper presents a novel calibration approach for a city-scale traffic simulation model based on limited real-world speed data. The simulation model runs a microscopic and mesoscopic realistic traffic simulation from Chattanooga, TN (US) for a 24-hour period and includes various transport modes such as transit buses, passenger cars, and trucks. The experiment results presented demonstrate the effectiveness of our approach for calibrating large-scale traffic networks using only real-world speed data. This paper presents our proposed calibration approach that utilizes 2160 real-world speed data points, performs sensitivity analysis of the simulation model to input parameters, and genetic algorithm for optimizing the model for calibration.","PeriodicalId":163556,"journal":{"name":"2023 IEEE International Conference on Smart Computing (SMARTCOMP)","volume":"72 7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133770359","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
期刊
2023 IEEE International Conference on Smart Computing (SMARTCOMP)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1