Abstract
Compared to RGB cameras, thermal sensors are more reliable at night. However, in the stereo matching task, it is difficult to estimate reliable depth simply depending on the binocular thermal cameras, due to their monochrome and low-contrast images. This paper proposed a thermal-visible stereo matching method based on multi-modal autoencoder (MANet), the key to which is to extract the modal-invariant features between thermal and visible images, and meanwhile, retain the characteristic features in terms of the optical properties. Besides, we designed a cross-reconstruction constraint to improve the scene feature extraction capability of MANet. In the stereo matching process, we analyze the thermal and visible properties and adopt optical attention mechanism to improve night-scene stereo matching. We established a stereo matching system with thermal-visible camera based on LiDAR calibration, and provided the dataset for evaluation. The experiment demonstrates that the proposed MANet can effectively extract the modal-invariant features and can be applied on thermal-visible stereo matching.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.