Citation: | Dai Weicong, Jin Longxu, Li Guoning, et al. Real-time airplane detection algorithm in remote-sensing images based on improved YOLOv3[J]. Opto-Electronic Engineering, 2018, 45(12): 180350. |
[1] | Girshick R, Donahue J, Darrell T, et al. Rich feature hierarchies for accurate object detection and semantic segmentation[C]// Proceedings of 2014 IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 2014: 580-587. |
[2] | Girshick R. Fast R-CNN[C]//IEEE International Conference on Computer Vision. IEEE, 2015: 1440-1448. |
[3] |
Ren S Q, He K M, Girshick R, et al. Faster R-CNN: towards real-time object detection with region proposal networks[C]// Proceedings of the 28th International Conference on Neural Information Processing Systems, Montreal, Canada, 2015: 91-99. |
[4] | He K, Gkioxari G, Dollár P, et al. Mask R-CNN[C]//IEEE International Conference on Computer Vision. IEEE, 2017: 2980-2988. |
[5] | Redmon J, Divvala S, Girshick R, et al. You Only Look Once: Unified, Real-Time Object Detection[C]//Computer Vision and Pattern Recognition. IEEE, 2016: 779-788. |
[6] | Liu W, Anguelov D, Erhan D, et al. SSD: Single Shot MultiBox Detector[C]//European Conference on Computer Vision. Springer International Publishing, 2016: 21-37. |
[7] | Redmon J, Farhadi A. YOLO9000: Better, Faster, Stronger[C]// IEEE Conference on Computer Vision and Pattern Recognition. IEEE, 2017: 6517-6525. |
[8] | Redmon J, Farhadi A. YOLOv3: An Incremental Improvement[J]. arXiv preprint arXiv: 1804.02767, 2018. |
[9] | 薛月菊, 黄宁, 涂淑琴, 等.未成熟芒果的改进YOLOv2识别方法[J].农业工程学报, 2018, 34(7): 173-179. Xue Y J, Huang N, Tu S Q, et al. Immature mango detection based on improved YOLOv2[J]. Transactions of the Chinese Society of Agricultural Engineering, 2018, 34(7): 173-179. |
[10] | 王思雨, 高鑫, 孙皓, 等.基于卷积神经网络的高分辨率SAR图像飞机目标检测方法[J].雷达学报, 2017, 6(2): 195-203. Wang S Y, Gao X, Sun H, et al. An aircraft detection method based on convolutional neural networks in high-resolution SAR images[J]. Journal of Radars, 2017, 6(2): 195-203. |
[11] | 周敏, 史振威, 丁火平.遥感图像飞机目标分类的卷积神经网络方法[J].中国图象图形学报, 2017, 22(5): 702-708. Zhou M, Shi Z W, Ding H P. Aircraft classification in remote-sensing images using convolutional neural networks[J]. Journal of Image and Graphics, 2017, 22(5): 702-708. |
[12] | 谷雨, 徐英.基于随机卷积特征和集成超限学习机的快速SAR目标识别[J].光电工程, 2018, 45(1): 170432. doi: 10.12086/oee.2018.170432 Gu Y, Xu Y. Fast SAR target recognition based on random convolution features and ensemble extreme learning machines[J]. Opto-Electronic Engineering, 2018, 45(1): 170432. doi: 10.12086/oee.2018.170432 |
[13] |
Huang G, Liu Z, Maaten L V D, et al. Densely Connected Convolutional Networks[C]//IEEE Conference on Computer Vision and Pattern Recognition. IEEE Computer Society, 2017: 2261-2269. |
[14] | Lin T Y, Dollar P, Girshick R, et al. Feature Pyramid Networks for Object Detection[C]//IEEE Conference on Computer Vision and Pattern Recognition. IEEE Computer Society, 2017: 936-944. |
Overview: The detection of airplanes in remote-sensing images has many important applications in many domains. However, limited to the performance of traditional machine learning methods, the airplanes in remote-sensing images are difficult to be detected. Recently, deep convolutional neural networks are employed to solve object detection problem and reach excellent accuracy. YOLO is one of the most famous real-time object detection algorithms based on regression. Compared with other algorithms, YOLO is more generalized when applied to many domains. Focusing on the airplanes in remote-sensing images, a real-time algorithm based on improved YOLOv3 is proposed to detect airplanes in remote-sensing images. Firstly, a convolutional neural network that consists of 49 convolutional layers is proposed to detect airplanes in remote-sensing images specifically. In the transition blocks of proposed convolutional neural network, we employ 1×1 convolution kernels to further reduce the parameters. Secondly, dense connection is employed on proposed convolutional neural network, and the maxpool is employed to enhance the feature transmit between two dense blocks. In this way, the feature transmit between two dense blocks is reconnected after a undersampling convolutional layer. The dense connection in proposed convolutional neural network enable the network to avoid over-fitting and reach high accuracy although the network is trained by relative few training data. Finally, to deal with the fact that airplanes in remote-sensing images are small targets mainly, we propose to increase the scale detections from 3 to 4 and employ dense connection to merge feature map among different scales. The anchor boxes in our work are obtained by running k-means clustering on the training set bounding boxes. The algorithm is trained and tested on the designed airplane dataset, which have 990 remote-sensing images. The qualitative experiment results show that our algorithm has stronger robustness than other existing algorithms, and our algorithm also shows especially high recall on small targets. The quantitative experiment results show that our algorithm obtains 96.26% on precision, 93.81% on recall and 89.31% on AP. Our algorithm reaches a relative improvement of 13.1% with respect to the YOLOv3 on AP. The detector proposed in this study is proven to perform real-time speed of more than 58.3 frames per second on a 1070Ti GPU. This study demonstrates the high effectiveness and accuracy of deep convolutional neural network in detecting airplanes on remote-sensing images. Meanwhile, the research also shows the fact that the performance of convolutional neural networks is decided by their structure and the number of training data.
An illustration of predicted bounding boxes on 13x13 grids of YOLOv3
An illustration of transition module
The relationship between the number of anchor boxes and average IOU
Multi-scale detection with dense connection
Three samples of airplane dataset
The detection results of YOLOv3-air、YOLOv3-tiny、YOLOv3 in order. (a) P883; (b) P902; (c) P903; (d) P909; (e) P866; (f) P867