{"title":"用于无人机运动规划的短期视觉- imu融合记忆代理","authors":"Zhihan Xue, T. Gonsalves","doi":"10.1109/ICIET55102.2022.9778991","DOIUrl":null,"url":null,"abstract":"This paper provides a deep reinforcement learning motion planning algorithm that uses time series image information as input. In this paper, we use a special visual state structure. This structure is composed of two kinds of sensor data. We combine the monocular visual information compressed by the variational autoencoder (VAE) and the inertial measurement unit (IMU) data. Then we use four consecutive combinations of the sensor data as the state of the reinforcement learning model. This method greatly simplifies the neural network complexity of continuous action space deep reinforcement learning on visual control tasks. After training in a drone simulation environment, this method has been proven to enable the drone to learn to avoid obstacles of various shapes and obstacles on the side blind corner of the camera.","PeriodicalId":371262,"journal":{"name":"2022 10th International Conference on Information and Education Technology (ICIET)","volume":"15 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-04-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Short-Term Visual-IMU Fusion Memory Agent For Drone's Motion Planning\",\"authors\":\"Zhihan Xue, T. Gonsalves\",\"doi\":\"10.1109/ICIET55102.2022.9778991\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"This paper provides a deep reinforcement learning motion planning algorithm that uses time series image information as input. In this paper, we use a special visual state structure. This structure is composed of two kinds of sensor data. We combine the monocular visual information compressed by the variational autoencoder (VAE) and the inertial measurement unit (IMU) data. Then we use four consecutive combinations of the sensor data as the state of the reinforcement learning model. This method greatly simplifies the neural network complexity of continuous action space deep reinforcement learning on visual control tasks. After training in a drone simulation environment, this method has been proven to enable the drone to learn to avoid obstacles of various shapes and obstacles on the side blind corner of the camera.\",\"PeriodicalId\":371262,\"journal\":{\"name\":\"2022 10th International Conference on Information and Education Technology (ICIET)\",\"volume\":\"15 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2022-04-09\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2022 10th International Conference on Information and Education Technology (ICIET)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICIET55102.2022.9778991\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 10th International Conference on Information and Education Technology (ICIET)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICIET55102.2022.9778991","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Short-Term Visual-IMU Fusion Memory Agent For Drone's Motion Planning
This paper provides a deep reinforcement learning motion planning algorithm that uses time series image information as input. In this paper, we use a special visual state structure. This structure is composed of two kinds of sensor data. We combine the monocular visual information compressed by the variational autoencoder (VAE) and the inertial measurement unit (IMU) data. Then we use four consecutive combinations of the sensor data as the state of the reinforcement learning model. This method greatly simplifies the neural network complexity of continuous action space deep reinforcement learning on visual control tasks. After training in a drone simulation environment, this method has been proven to enable the drone to learn to avoid obstacles of various shapes and obstacles on the side blind corner of the camera.