Abstract

Object detection in aerial images, particularly of vehicles, is highly important in remote sensing applications including traffic management, urban planning, parking space utilization, surveillance, and search and rescue. In this article, we investigate the ability of three-dimensional (3D) feature maps to improve the performance of deep neural network (DNN) for vehicle detection. First, we propose a DNN based on YOLOv3 with various base networks, including DarkNet-53, SqueezeNet, MobileNet-v2, and DenseNet-201. We assessed the base networks and their performance in combination with YOLOv3 on efficiency, processing time, and the memory that each architecture required. In the second part, 3D depth maps were generated using pairs of aerial images and their parallax displacement. Next, a fully connected neural network (fcNN) was trained on 3D feature maps of trucks, semi-trailers and trailers. A cascade of these networks was then proposed to detect vehicles in aerial images. Upon the DNN detecting a region, coordinates and confidence levels were used to extract the corresponding 3D features. The fcNN used 3D features as the input to improve the DNN performance. The data set used in this work was acquired from numerous flights of an unmanned aerial vehicle (UAV) across two industrial harbors over two years. The experimental results show that 3D features improved the precision of DNNs from 88.23 % to 96.43 % and from 97.10 % to 100 % when using DNN confidence thresholds of 0.01 and 0.05, respectively. Accordingly, the proposed system was able to successfully remove 72.22 % to 100 % of false positives from the DNN outputs. These results indicate the importance of 3D features utilization to improve object detection in aerial images for future research.

Highlights

  • V EHICLE detection in aerial images is a key factor in understanding complex transportation systems [1]

  • Data sets from 14 flights were used to develop the deep neural network (DNN) to detect the target regions; data from three flights were used to build the fully connected neural network (fcNN) based on the 3D depth maps

  • The results show that the DNN containing DarkNet-53 achieved the highest average precision (AP); it was selected as the extractor network for the proposed system

Read more

Summary

Introduction

V EHICLE detection in aerial images is a key factor in understanding complex transportation systems [1]. Urban planning, parking space utilization, surveillance, and search and rescue are among a wide variety of applications that require vehicle detection [2]–[6]. Development and availability of unmanned aerial vehicle (UAV) platforms have made aerial imaging relatively inexpensive, convenient, and suitable for real-time performance applications [7], [8]. Detecting vehicles in aerial images both accurately and quickly is challenging. As aerial images are taken from altitude with a top-down view, vehicles appear relatively small, and a single image may contain many vehicles [10]. In comparison to video analysis from a stationary camera, as UAV moves, the background varies, which hinders accurate detection [1]

Objectives
Methods
Results
Conclusion
Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.