Yangte Gao;Zhihao Che;Lin Li;Jianfeng Gao;Fukun Bi
{"title":"Deep Spatial Feature Transformation for Oriented Aerial Object Detection","authors":"Yangte Gao;Zhihao Che;Lin Li;Jianfeng Gao;Fukun Bi","doi":"10.1109/JMASS.2023.3234076","DOIUrl":null,"url":null,"abstract":"Object detection in aerial images has received extensive attention in the field of computer vision. Different from natural images, the aerial objects are usually distributed in any direction. Therefore, the existing detector usually needs more parameters to encode the direction information, resulting in a large number of redundant calculations. In addition, because an ordinary convolution neural network (CNN) does not effectively model the direction change, a large amount of the rotated data is required for the aerial detector. To solve these problems, we propose a deep spatial feature transformation network (DSFT-Net), which includes a spatial feature extraction module and a feature selection module. Specifically, we add the rotation convolution kernel to the detector to extract the directional feature of the rotated target to accurately predict the direction of the model. Then, we build a dual pyramid to separate the features in the classification and regression tasks. Finally, the polarization function is proposed to construct the critical features that are suitable for their respective tasks, achieving feature selection and more refined detection. Experiments on public remote sensing benchmarks (e.g., DOTA, HRSC2016, and UCAS-AOD) have proved the effectiveness of our detector.","PeriodicalId":100624,"journal":{"name":"IEEE Journal on Miniaturization for Air and Space Systems","volume":"4 2","pages":"93-99"},"PeriodicalIF":0.0000,"publicationDate":"2023-01-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Journal on Miniaturization for Air and Space Systems","FirstCategoryId":"1085","ListUrlMain":"https://ieeexplore.ieee.org/document/10005798/","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Object detection in aerial images has received extensive attention in the field of computer vision. Different from natural images, the aerial objects are usually distributed in any direction. Therefore, the existing detector usually needs more parameters to encode the direction information, resulting in a large number of redundant calculations. In addition, because an ordinary convolution neural network (CNN) does not effectively model the direction change, a large amount of the rotated data is required for the aerial detector. To solve these problems, we propose a deep spatial feature transformation network (DSFT-Net), which includes a spatial feature extraction module and a feature selection module. Specifically, we add the rotation convolution kernel to the detector to extract the directional feature of the rotated target to accurately predict the direction of the model. Then, we build a dual pyramid to separate the features in the classification and regression tasks. Finally, the polarization function is proposed to construct the critical features that are suitable for their respective tasks, achieving feature selection and more refined detection. Experiments on public remote sensing benchmarks (e.g., DOTA, HRSC2016, and UCAS-AOD) have proved the effectiveness of our detector.