Joongheon Kim, Yu-Cheol Lee, Jun Hwan Lee, Jin Seek Choi
{"title":"关于无人驾驶飞行器及其应用的特刊","authors":"Joongheon Kim, Yu-Cheol Lee, Jun Hwan Lee, Jin Seek Choi","doi":"10.4218/etr2.12628","DOIUrl":null,"url":null,"abstract":"<p>Recently, research on autonomous mobility control has been actively and widely conducted for various applications. In particular, autonomous mobility control for unmanned aerial and ground vehicles has been our research interest because it has considerable challenges, such as time-consuming and high-delay computations, complicated functionalities, and dangerous tasks that were previously performed by humans. Furthermore, fully autonomous unmanned aerial and ground vehicles are barely practical and have various operational limitations, such as high-precision sensing, high computational complexity, low autonomy, and restricted mobility. To develop the required technologies to overcome these limitations and achieve full autonomy for unmanned aerial and ground vehicles, various studies have addressed aspects such as precise pose estimation, environment mapping, path planning, trajectory optimization, and 2D/3D object tracking and detection.</p><p>With fully autonomous operation and functionalities for unmanned aerial and ground vehicles, emerging applications will become more diverse and include autonomous artificial-intelligence-based surveillance, autonomous disaster prevention broadcasting and control, mobile autonomous aerial and ground wireless/cellular access service provisioning, autonomous multirobot coordination, and cooperation for smart factory management in smart city applications, for which a skilled human operator must currently intervene throughout operation.</p><p>For this special issue, we selected 11 key studies on (1) communication, networks, and mobility [<span>1-5</span>] and (2) object detection and tracking in autonomous driving [<span>6-11</span>].</p><p>In [<span>1</span>], surveys and discussions are presented on recent deep-learning-based developments to achieve autonomous mobility control and efficient resource management of autonomous vehicles including unmanned aerial vehicles (UAVs). The developments include multiagent reinforcement learning and neural Myerson auction. We believe that integrating multiagent reinforcement learning and neural Myerson auction will be critical for efficient and trustworthy autonomous mobility services.</p><p>In [<span>2</span>], a safe landing algorithm is introduced for urban drone delivery. The proposed algorithm generates a safe and efficient vertical landing path for drones, allowing them to avoid obstacles commonly found in urban environments, such as trees, streetlights, utility poles, and wires. To this end, landing-angle control is implemented to land vertically, and a rapidly-exploring random tree (RRT) is used in a collision avoidance algorithm. This combination of methods enables precise and reliable drone delivery in urban settings.</p><p>In [<span>3</span>], a loosely coupled relative position estimation method is proposed based on a decentralized ultrawideband global navigation support system and inertial navigation system for flight controllers. Key obstacles to multi-drone collaboration are noted and include relative positional errors and the absence of communication devices. To address such problems, an extended Kalman filter (EKF) is adopted to correct distance errors by fusing ultrawideband data acquired through random communications using a novel UWB communication module.</p><p>Unmanned vehicles are being increasingly used for time-consuming, complicated, and dangerous tasks that were previously performed by humans. However, they have limitations for applications like establishing high-speed wireless networks. In [<span>4</span>], a 3D geometry-based stochastic model for UAV multiple-input multiple-output (MIMO) channels is presented. The UAV flying direction and location have a significant impact on MIMO performance. This innovative model of 3D navigation and scattering environments is closely related to the scope of this special issue.</p><p>In [<span>5</span>], UAVs are considered essential components in non-terrestrial networks, especially in 5G-and-beyond communication systems. Employing UAVs operated in conjunction with a 4G/5G base station has proven to be a practical solution for providing cellular network services in areas where conventional communication infrastructures are unavailable. This paper introduces the uncrewed aerial vehicle–base station system that utilizes a high-capacity wireless backhaul operating in millimeter wave frequency bands.</p><p>In [<span>6</span>], advanced video analytics for tasks such as moving object detection and segmentation are presented, thereby increasing the demand for such methods in unmanned aerial and ground vehicle applications. A novel zero-shot video object segmentation is introduced to focus on the discovery of moving objects in challenging scenarios. This method employs a background memory model for training from sparse annotations over time by using temporal modeling of the background to accurately detect moving objects. In addition, the method addresses the limitations of existing state-of-the-art solutions for detecting salient objects within images regardless of their motion.</p><p>In [<span>7</span>], an adaptive UAV-assisted object-recognition algorithm is introduced for urban surveillance scenarios. In a UAV-assisted surveillance system, UAVs are equipped with learning-based object recognition models and can collect surveillance images. Owing to UAV limitations (for example, limited battery and computational capabilities), adaptive control considering these limitations is devised to maximize the time-averaged recognition performance subject to stability through Lyapunov optimization.</p><p>In [<span>8</span>], modern semantic segmentation frameworks combining low- and high-level context information are used to improve performance. In addition, post-level context information is considered in a context refinement network (CRFNet). Training for improving the semantic segmentation predictions proceeds through an encoder–decoder structure. This study directly considers the relation between spatially neighboring pixels of a label map using methods such as Markov and conditional random fields.</p><p>In [<span>9</span>], real-time accurate 3D multi-pedestrian detection and tracking are achieved using 3D LiDAR point clouds from crowded environments. Pedestrian detection segments a sparse 3D point cloud into individual pedestrians using a lightweight convolutional autoencoder and connected component labeling. Multi-pedestrian tracking associates the same pedestrians by considering motion and appearance cues in continuous frames. In addition, the dynamic movements of pedestrians are estimated with various patterns by adaptively mixing heterogeneous motion models.</p><p>In [<span>10</span>], sensor-fusion-based object detection and classification are presented. The proposed method operates in real time, rendering it suitable for integration into autonomous vehicles. It performs well on a custom dataset and publicly available datasets, demonstrating its effectiveness in real-world road environments. In addition, a 3D moving object detection dataset called ETRI 3D MOD, is constructed.</p><p>In [<span>11</span>], three techniques for combining information from multiple cameras are proposed, namely, feature, early, and late fusion techniques. Extensive experiments were conducted on pedestrian-view intersection classification. The proposed model with feature fusion provides an area under the curve and an F1-score of 82.00 and 46.48, respectively, outperforming a model trained using only real three-camera data and one-camera models by a large margin.</p><p>The authors declare that there are no conflicts of interest.</p>","PeriodicalId":11901,"journal":{"name":"ETRI Journal","volume":"45 5","pages":"731-734"},"PeriodicalIF":1.3000,"publicationDate":"2023-10-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Special issue on autonomous unmanned aerial/ground vehicles and their applications\",\"authors\":\"Joongheon Kim, Yu-Cheol Lee, Jun Hwan Lee, Jin Seek Choi\",\"doi\":\"10.4218/etr2.12628\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p>Recently, research on autonomous mobility control has been actively and widely conducted for various applications. In particular, autonomous mobility control for unmanned aerial and ground vehicles has been our research interest because it has considerable challenges, such as time-consuming and high-delay computations, complicated functionalities, and dangerous tasks that were previously performed by humans. Furthermore, fully autonomous unmanned aerial and ground vehicles are barely practical and have various operational limitations, such as high-precision sensing, high computational complexity, low autonomy, and restricted mobility. To develop the required technologies to overcome these limitations and achieve full autonomy for unmanned aerial and ground vehicles, various studies have addressed aspects such as precise pose estimation, environment mapping, path planning, trajectory optimization, and 2D/3D object tracking and detection.</p><p>With fully autonomous operation and functionalities for unmanned aerial and ground vehicles, emerging applications will become more diverse and include autonomous artificial-intelligence-based surveillance, autonomous disaster prevention broadcasting and control, mobile autonomous aerial and ground wireless/cellular access service provisioning, autonomous multirobot coordination, and cooperation for smart factory management in smart city applications, for which a skilled human operator must currently intervene throughout operation.</p><p>For this special issue, we selected 11 key studies on (1) communication, networks, and mobility [<span>1-5</span>] and (2) object detection and tracking in autonomous driving [<span>6-11</span>].</p><p>In [<span>1</span>], surveys and discussions are presented on recent deep-learning-based developments to achieve autonomous mobility control and efficient resource management of autonomous vehicles including unmanned aerial vehicles (UAVs). The developments include multiagent reinforcement learning and neural Myerson auction. We believe that integrating multiagent reinforcement learning and neural Myerson auction will be critical for efficient and trustworthy autonomous mobility services.</p><p>In [<span>2</span>], a safe landing algorithm is introduced for urban drone delivery. The proposed algorithm generates a safe and efficient vertical landing path for drones, allowing them to avoid obstacles commonly found in urban environments, such as trees, streetlights, utility poles, and wires. To this end, landing-angle control is implemented to land vertically, and a rapidly-exploring random tree (RRT) is used in a collision avoidance algorithm. This combination of methods enables precise and reliable drone delivery in urban settings.</p><p>In [<span>3</span>], a loosely coupled relative position estimation method is proposed based on a decentralized ultrawideband global navigation support system and inertial navigation system for flight controllers. Key obstacles to multi-drone collaboration are noted and include relative positional errors and the absence of communication devices. To address such problems, an extended Kalman filter (EKF) is adopted to correct distance errors by fusing ultrawideband data acquired through random communications using a novel UWB communication module.</p><p>Unmanned vehicles are being increasingly used for time-consuming, complicated, and dangerous tasks that were previously performed by humans. However, they have limitations for applications like establishing high-speed wireless networks. In [<span>4</span>], a 3D geometry-based stochastic model for UAV multiple-input multiple-output (MIMO) channels is presented. The UAV flying direction and location have a significant impact on MIMO performance. This innovative model of 3D navigation and scattering environments is closely related to the scope of this special issue.</p><p>In [<span>5</span>], UAVs are considered essential components in non-terrestrial networks, especially in 5G-and-beyond communication systems. Employing UAVs operated in conjunction with a 4G/5G base station has proven to be a practical solution for providing cellular network services in areas where conventional communication infrastructures are unavailable. This paper introduces the uncrewed aerial vehicle–base station system that utilizes a high-capacity wireless backhaul operating in millimeter wave frequency bands.</p><p>In [<span>6</span>], advanced video analytics for tasks such as moving object detection and segmentation are presented, thereby increasing the demand for such methods in unmanned aerial and ground vehicle applications. A novel zero-shot video object segmentation is introduced to focus on the discovery of moving objects in challenging scenarios. This method employs a background memory model for training from sparse annotations over time by using temporal modeling of the background to accurately detect moving objects. In addition, the method addresses the limitations of existing state-of-the-art solutions for detecting salient objects within images regardless of their motion.</p><p>In [<span>7</span>], an adaptive UAV-assisted object-recognition algorithm is introduced for urban surveillance scenarios. In a UAV-assisted surveillance system, UAVs are equipped with learning-based object recognition models and can collect surveillance images. Owing to UAV limitations (for example, limited battery and computational capabilities), adaptive control considering these limitations is devised to maximize the time-averaged recognition performance subject to stability through Lyapunov optimization.</p><p>In [<span>8</span>], modern semantic segmentation frameworks combining low- and high-level context information are used to improve performance. In addition, post-level context information is considered in a context refinement network (CRFNet). Training for improving the semantic segmentation predictions proceeds through an encoder–decoder structure. This study directly considers the relation between spatially neighboring pixels of a label map using methods such as Markov and conditional random fields.</p><p>In [<span>9</span>], real-time accurate 3D multi-pedestrian detection and tracking are achieved using 3D LiDAR point clouds from crowded environments. Pedestrian detection segments a sparse 3D point cloud into individual pedestrians using a lightweight convolutional autoencoder and connected component labeling. Multi-pedestrian tracking associates the same pedestrians by considering motion and appearance cues in continuous frames. In addition, the dynamic movements of pedestrians are estimated with various patterns by adaptively mixing heterogeneous motion models.</p><p>In [<span>10</span>], sensor-fusion-based object detection and classification are presented. The proposed method operates in real time, rendering it suitable for integration into autonomous vehicles. It performs well on a custom dataset and publicly available datasets, demonstrating its effectiveness in real-world road environments. In addition, a 3D moving object detection dataset called ETRI 3D MOD, is constructed.</p><p>In [<span>11</span>], three techniques for combining information from multiple cameras are proposed, namely, feature, early, and late fusion techniques. Extensive experiments were conducted on pedestrian-view intersection classification. The proposed model with feature fusion provides an area under the curve and an F1-score of 82.00 and 46.48, respectively, outperforming a model trained using only real three-camera data and one-camera models by a large margin.</p><p>The authors declare that there are no conflicts of interest.</p>\",\"PeriodicalId\":11901,\"journal\":{\"name\":\"ETRI Journal\",\"volume\":\"45 5\",\"pages\":\"731-734\"},\"PeriodicalIF\":1.3000,\"publicationDate\":\"2023-10-29\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"ETRI Journal\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://onlinelibrary.wiley.com/doi/10.4218/etr2.12628\",\"RegionNum\":4,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q3\",\"JCRName\":\"ENGINEERING, ELECTRICAL & ELECTRONIC\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"ETRI Journal","FirstCategoryId":"94","ListUrlMain":"https://onlinelibrary.wiley.com/doi/10.4218/etr2.12628","RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"ENGINEERING, ELECTRICAL & ELECTRONIC","Score":null,"Total":0}
Special issue on autonomous unmanned aerial/ground vehicles and their applications
Recently, research on autonomous mobility control has been actively and widely conducted for various applications. In particular, autonomous mobility control for unmanned aerial and ground vehicles has been our research interest because it has considerable challenges, such as time-consuming and high-delay computations, complicated functionalities, and dangerous tasks that were previously performed by humans. Furthermore, fully autonomous unmanned aerial and ground vehicles are barely practical and have various operational limitations, such as high-precision sensing, high computational complexity, low autonomy, and restricted mobility. To develop the required technologies to overcome these limitations and achieve full autonomy for unmanned aerial and ground vehicles, various studies have addressed aspects such as precise pose estimation, environment mapping, path planning, trajectory optimization, and 2D/3D object tracking and detection.
With fully autonomous operation and functionalities for unmanned aerial and ground vehicles, emerging applications will become more diverse and include autonomous artificial-intelligence-based surveillance, autonomous disaster prevention broadcasting and control, mobile autonomous aerial and ground wireless/cellular access service provisioning, autonomous multirobot coordination, and cooperation for smart factory management in smart city applications, for which a skilled human operator must currently intervene throughout operation.
For this special issue, we selected 11 key studies on (1) communication, networks, and mobility [1-5] and (2) object detection and tracking in autonomous driving [6-11].
In [1], surveys and discussions are presented on recent deep-learning-based developments to achieve autonomous mobility control and efficient resource management of autonomous vehicles including unmanned aerial vehicles (UAVs). The developments include multiagent reinforcement learning and neural Myerson auction. We believe that integrating multiagent reinforcement learning and neural Myerson auction will be critical for efficient and trustworthy autonomous mobility services.
In [2], a safe landing algorithm is introduced for urban drone delivery. The proposed algorithm generates a safe and efficient vertical landing path for drones, allowing them to avoid obstacles commonly found in urban environments, such as trees, streetlights, utility poles, and wires. To this end, landing-angle control is implemented to land vertically, and a rapidly-exploring random tree (RRT) is used in a collision avoidance algorithm. This combination of methods enables precise and reliable drone delivery in urban settings.
In [3], a loosely coupled relative position estimation method is proposed based on a decentralized ultrawideband global navigation support system and inertial navigation system for flight controllers. Key obstacles to multi-drone collaboration are noted and include relative positional errors and the absence of communication devices. To address such problems, an extended Kalman filter (EKF) is adopted to correct distance errors by fusing ultrawideband data acquired through random communications using a novel UWB communication module.
Unmanned vehicles are being increasingly used for time-consuming, complicated, and dangerous tasks that were previously performed by humans. However, they have limitations for applications like establishing high-speed wireless networks. In [4], a 3D geometry-based stochastic model for UAV multiple-input multiple-output (MIMO) channels is presented. The UAV flying direction and location have a significant impact on MIMO performance. This innovative model of 3D navigation and scattering environments is closely related to the scope of this special issue.
In [5], UAVs are considered essential components in non-terrestrial networks, especially in 5G-and-beyond communication systems. Employing UAVs operated in conjunction with a 4G/5G base station has proven to be a practical solution for providing cellular network services in areas where conventional communication infrastructures are unavailable. This paper introduces the uncrewed aerial vehicle–base station system that utilizes a high-capacity wireless backhaul operating in millimeter wave frequency bands.
In [6], advanced video analytics for tasks such as moving object detection and segmentation are presented, thereby increasing the demand for such methods in unmanned aerial and ground vehicle applications. A novel zero-shot video object segmentation is introduced to focus on the discovery of moving objects in challenging scenarios. This method employs a background memory model for training from sparse annotations over time by using temporal modeling of the background to accurately detect moving objects. In addition, the method addresses the limitations of existing state-of-the-art solutions for detecting salient objects within images regardless of their motion.
In [7], an adaptive UAV-assisted object-recognition algorithm is introduced for urban surveillance scenarios. In a UAV-assisted surveillance system, UAVs are equipped with learning-based object recognition models and can collect surveillance images. Owing to UAV limitations (for example, limited battery and computational capabilities), adaptive control considering these limitations is devised to maximize the time-averaged recognition performance subject to stability through Lyapunov optimization.
In [8], modern semantic segmentation frameworks combining low- and high-level context information are used to improve performance. In addition, post-level context information is considered in a context refinement network (CRFNet). Training for improving the semantic segmentation predictions proceeds through an encoder–decoder structure. This study directly considers the relation between spatially neighboring pixels of a label map using methods such as Markov and conditional random fields.
In [9], real-time accurate 3D multi-pedestrian detection and tracking are achieved using 3D LiDAR point clouds from crowded environments. Pedestrian detection segments a sparse 3D point cloud into individual pedestrians using a lightweight convolutional autoencoder and connected component labeling. Multi-pedestrian tracking associates the same pedestrians by considering motion and appearance cues in continuous frames. In addition, the dynamic movements of pedestrians are estimated with various patterns by adaptively mixing heterogeneous motion models.
In [10], sensor-fusion-based object detection and classification are presented. The proposed method operates in real time, rendering it suitable for integration into autonomous vehicles. It performs well on a custom dataset and publicly available datasets, demonstrating its effectiveness in real-world road environments. In addition, a 3D moving object detection dataset called ETRI 3D MOD, is constructed.
In [11], three techniques for combining information from multiple cameras are proposed, namely, feature, early, and late fusion techniques. Extensive experiments were conducted on pedestrian-view intersection classification. The proposed model with feature fusion provides an area under the curve and an F1-score of 82.00 and 46.48, respectively, outperforming a model trained using only real three-camera data and one-camera models by a large margin.
The authors declare that there are no conflicts of interest.
期刊介绍:
ETRI Journal is an international, peer-reviewed multidisciplinary journal published bimonthly in English. The main focus of the journal is to provide an open forum to exchange innovative ideas and technology in the fields of information, telecommunications, and electronics.
Key topics of interest include high-performance computing, big data analytics, cloud computing, multimedia technology, communication networks and services, wireless communications and mobile computing, material and component technology, as well as security.
With an international editorial committee and experts from around the world as reviewers, ETRI Journal publishes high-quality research papers on the latest and best developments from the global community.