Abstract

Abstract. Dense image matching is essential to photogrammetry applications, including Digital Surface Model (DSM) generation, three dimensional (3D) reconstruction, and object detection and recognition. The development of an efficient and robust method for dense image matching has been one of the technical challenges due to high variations in illumination and ground features of aerial images of large areas. Nowadays, due to the development of deep learning technology, deep neural network-based algorithms outperform traditional methods on a variety of tasks such as object detection, semantic segmentation and stereo matching. The proposed network includes cost-volume computation, cost-volume aggregation, and disparity prediction. It starts with a pre-trained VGG-16 network as a backend and using the U-net architecture with nine layers for feature map extraction and a correlation layer for cost volume calculation, after that a guided filter based cost aggregation is adopted for cost volume filtering and finally the soft Argmax function is utilized for disparity prediction. The experimental conducted on a UAV dataset demonstrated that the proposed method achieved the RMSE (root mean square error) of the reprojection error better than 1 pixel in image coordinate and in-ground positioning accuracy within 2.5 ground sample distance. The comparison experiments on KITTI 2015 dataset shows the proposed unsupervised method even comparably with other supervised methods.

Highlights

  • Dense image matching is essential to photogrammetry applications, including Digital Surface Model (DSM) generation, three dimensional (3D) reconstruction, and object detection and recognition (Xu et al, 2017)

  • Zbontar and LeCun (2015) describe patch based matching as a binary classification problem to determine the pixel-wise correspondence use deep neural networks

  • The common supervised learning methods can use a large amount of ground truth data for loss function determination; for unsupervised learning, only a reasonable loss function can lead to good prediction results

Read more

Summary

INTRODUCTION

Dense image matching is essential to photogrammetry applications, including Digital Surface Model (DSM) generation, three dimensional (3D) reconstruction, and object detection and recognition (Xu et al, 2017). The development of an efficient and robust method for dense image matching has been one of the technical challenges due to high variations in illumination and ground features of aerial images of large areas. Due to the development of deep learning technology, deep neural networkbased algorithms outperform traditional methods on a variety of tasks such as object detection, semantic segmentation and stereo matching. On the stereo matching benchmark KITTI, the top 50 methods on the rank are deep learning-based, which show a significant advantage by utilizing deep neural networks for dense image matching tasks (Geiger, 2012). In real 3D data production, the accurate 3D points cloud ground truth is usually obtained by LiDAR system, which is really expensive for handling a large survey area (Yuan et al, 2019). To tackle the above problems, in this paper, we present an end-toend unsupervised multi-constraint Deep Neural Network for aerial image-based dense image matching

RELATED WORKS
Multi-Constraint Loss Function
METHOD
Feature extraction
Cost-Volume Computation
Disparity Prediction
Smoothness Loss
Quality assessment
Analysis
CONCLUSION
Findings
Comparison experiments
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call