Abstract
3D time-of-flight (ToF) imaging is used in a variety of applications such as augmented reality (AR), computer interfaces, robotics and autonomous systems. Single-photon avalanche diodes (SPADs) are one of the enabling technologies providing accurate depth data even over long ranges. By developing SPADs in array format with integrated processing combined with pulsed, flood-type illumination, high-speed 3D capture is possible. However, array sizes tend to be relatively small, limiting the lateral resolution of the resulting depth maps and, consequently, the information that can be extracted from the image for applications such as object detection. In this paper, we demonstrate that these limitations can be overcome through the use of convolutional neural networks (CNNs) for high-performance object detection. We present outdoor results from a portable SPAD camera system that outputs 16-bin photon timing histograms with 64×32 spatial resolution, with each histogram containing thousands of photons. The results, obtained with exposure times down to 2 ms (equivalent to 500 FPS) and in signal-to-background (SBR) ratios as low as 0.05, point to the advantages of providing the CNN with full histogram data rather than point clouds alone. Alternatively, a combination of point cloud and active intensity data may be used as input, for a similar level of performance. In either case, the GPU-accelerated processing time is less than 1 ms per frame, leading to an overall latency (image acquisition plus processing) in the millisecond range, making the results relevant for safety-critical computer vision applications which would benefit from faster than human reaction times.
Highlights
Sensitive avalanche photo-diodes (APD) or single-photon avalanche diodes (SPAD) are commonly used as detectors. direct ToF (dToF) systems traditionally relied on single-point detectors, complemented by optical scanning to cover a given field of view (FoV), at the expense of a low acquisition rate
We evaluate the performance of object detection for the data types offered by the SPAD camera: depth, SPC-256, SPC-64, Act_I-D and histogram
A total of 6388 training examples are fed into the neural network to create a model and a total of 574 examples are used for testing with a SBR < 0.5 on average
Summary
DToF systems traditionally relied on single-point detectors, complemented by optical scanning to cover a given field of view (FoV), at the expense of a low acquisition rate. Advances in SPAD technology have resulted in array format sensors with integrated timing electronics. SPAD-based dToF devices in line [15] and image sensor format [16,17] using blade-scanned and flood illumination sources respectively, have been developed. These devices enable high-speed 3D imaging, especially when featuring on-chip processing such as histogramming [18].
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.